Parallel Randomized Block Coordinate Descent for Neural Probabilistic Language Model with High-Dimensional Output Targets
Training a large probabilistic neural network language model,with typical high-dimensional output is excessively timeconsuming,which is one of the main reasons that more simplified models such as n-gram is often more popular despite the inferior performance.In this paper a Chinese neural probabilistic language model is trained using the Fudan Chinese Language Corpus.As hundreds of thousands of distinct words have been tokenized from the raw corpus,the model contains tens of millions of parameters.To address the challenge,popular parallel computing platform MPI(Message Passing Interface)based on cluster is employed to implement the parallel neural network language model.Specifically,we propose a new method termed as Parallel Randomized Block Coordinate Descent(PRBCD)to train this model cost-effectively.Different from traditional coordinate descent method,our new method could be employed in network with multiple layers,allowing scaling up the gradients with respect to hidden units proportionally based on sampled parameters.We empirically show that our PRBCD is stable and is well suited for language models,which contain only a few layers while often have a large amount of parameters and extremely high-dimensional output targets.
Language model Stochastic optimization Parallel computing
Xin Liu Junchi Yan Xiangfeng Wang Hongyuan Zha
East China Normal University,Shanghai,China East China Normal University,Shanghai,China;IBM Research - China,Shanghai,China East China Normal University,Shanghai,China;Georgia Institute of Technology,Atlanta,USA
国际会议
第七届全国模式识别学术会议(The 7th Chinese Conference on Pattern Recognition,CCPR2016)
成都
英文
334-348
2016-11-03(万方平台首次上网日期,不代表论文的发表时间)