If you have an event to announce, please contact us and we will post it here.

Thursday, October 25, 2012

Shay Cohen: Consistent and Efficient Algorithms for Latent-Variable PCFGs

Shay Cohen will present a talk at Rutgers titled "Consistent and Efficient Algorithms for Latent-Variable PCFGs"

Date: Friday, November 2
Time:  11:00am
Location: Rutgers School of Communication and Information, 4 Huntington St., New Brunswick, NJ (Faculty Lounge, Room 323) (map)


ABSTRACT:


In the past few years, there has been an increased interest in the machine learning community in spectral algorithms for estimating models with latent variables. Examples include algorithms for estimating mixture of Gaussians or for estimating the parameters of a hidden Markov model.

Until the introduction of spectral algorithms, the EM algorithm has been the mainstay for estimation with latent variables. Still, with EM there is no guarantee of convergence to the global maximum of the likelihood function, and therefore EM generally does not provide consistent estimates for the model parameters. Spectral algorithms, on the other hand, are often shown to be consistent.

In this talk, I am interested in presenting a spectral algorithm for latent-variable PCFGs, a model widely used for parsing in the NLP community. This model augments the nonterminals in a PCFG grammar with latent states.  These latent states refine the nonterminal category in order to capture subtle syntactic nuances in the data.  This model has been successfully implemented in state-of-the-art parsers such as the Berkeley parser.

The algorithm developed is considerably faster than EM, as it makes only one pass over the data. Statistics are collected from the data in this pass, and singular value decomposition is performed on matrices containing these statistics. Our algorithm is also provably consistent in the sense that, given enough samples, it will estimate probabilities for test trees close to their true probabilities under the latent-variable PCFG model.

If time permits, I will also present a method aimed at improving the efficiency of parsing with latent variable PCFGs.  This method relies on tensor decomposition of the latent-variable PCFG. This tensor decomposition is approximate, and therefore the new parser is an approximate parser as well. Still, the quality of approximation can be theoretically guaranteed by inspecting how errors from the approximation propagate in the parse trees. interests span a range of topics in natural language processing and machine learning. He is especially interested in developing algorithms and methods for the use of probabilistic grammars.


BIO:

ShayCohen is a postdoctoral research scientist in the Department of Computer Science at Columbia University. He is currently a Computing Innovation fellow (NSF/CRA). He received his B.Sc. and M.Sc. from Tel Aviv University in 2000 and 2004, and his Ph.D. from Carnegie Mellon University in 2011. His research

No comments:

Post a Comment