On the Accuracy of Self-Normalized Log-Linear Models
暂无分享,去创建一个
Dan Klein | Jacob Andreas | Michael I. Jordan | Maxim Rabinovich | D. Klein | Jacob Andreas | Maxim Rabinovich
[1] Andrew McCallum,et al. Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data , 2001, ICML.
[2] Ashish Vaswani,et al. Decoding with Large-Scale Neural Language Models Improves Translation , 2013, EMNLP.
[3] Peter L. Bartlett,et al. Neural Network Learning - Theoretical Foundations , 1999 .
[4] Eric R. Ziegel,et al. Generalized Linear Models , 2002, Technometrics.
[5] Richard M. Schwartz,et al. Fast and Robust Neural Network Joint Models for Statistical Machine Translation , 2014, ACL.
[6] Peter L. Bartlett,et al. The Sample Complexity of Pattern Classification with Neural Networks: The Size of the Weights is More Important than the Size of the Network , 1998, IEEE Trans. Inf. Theory.
[7] Aapo Hyvärinen,et al. Noise-contrastive estimation: A new estimation principle for unnormalized statistical models , 2010, AISTATS.
[8] Dan Klein,et al. When and why are log-linear models self-normalizing? , 2015, NAACL.
[9] Pradeep Ravikumar,et al. Graphical Models via Generalized Linear Models , 2012, NIPS.
[10] Axthonv G. Oettinger,et al. IEEE Transactions on Information Theory , 1998 .
[11] Nando de Freitas,et al. An Introduction to Sequential Monte Carlo Methods , 2001, Sequential Monte Carlo Methods in Practice.