期刊文献+

Tuning the Learning Rate for Stochastic Variational Inference

Tuning the Learning Rate for Stochastic Variational Inference
原文传递
导出
摘要 Stochastic variational inference (SVI) can learn topic models with very big corpora. It optimizes the variational objective by using the stochastic natural gradient algorithm with a decreasing learning rate. This rate is crucial for SVI; however, it is often tuned by hand in real applications. To address this, we develop a novel algorithm, which tunes the learning rate of each iteration adaptively. The proposed algorithm uses the Kullback-Leibler (KL) divergence to measure the similarity between the variational distribution with noisy update and that with batch update, and then optimizes the learning rates by minimizing the KL divergence. We apply our algorithm to two representative topic models: latent Dirichlet allocation and hierarchical Dirichlet process. Experimental results indicate that our algorithm performs better and converges faster than commonly used learning rates. Stochastic variational inference (SVI) can learn topic models with very big corpora. It optimizes the variational objective by using the stochastic natural gradient algorithm with a decreasing learning rate. This rate is crucial for SVI; however, it is often tuned by hand in real applications. To address this, we develop a novel algorithm, which tunes the learning rate of each iteration adaptively. The proposed algorithm uses the Kullback-Leibler (KL) divergence to measure the similarity between the variational distribution with noisy update and that with batch update, and then optimizes the learning rates by minimizing the KL divergence. We apply our algorithm to two representative topic models: latent Dirichlet allocation and hierarchical Dirichlet process. Experimental results indicate that our algorithm performs better and converges faster than commonly used learning rates.
出处 《Journal of Computer Science & Technology》 SCIE EI CSCD 2016年第2期428-436,共9页 计算机科学技术学报(英文版)
基金 This work was supported by the National Natural Science Foundation of China under Grant Nos. 61170092, 61133011 and 61103091.
关键词 stochastic variational inference online learning adaptive learning rate topic model stochastic variational inference, online learning, adaptive learning rate, topic model
  • 相关文献

参考文献1

二级参考文献29

  • 1Hotho A, Staab S, Stumme G. Wordnet improves text docu- ment clustering. In Proc. SIGIR 2003 Semantic Web Work- shop, Toronto, Canada, Aug. 1, 2003.
  • 2Hu J, Fang L, Cao Y, Zeng H J, Li H, Yang Q, Chen Z. En- hancing text clustering by leveraging Wikipedia semantics. In Proc. SIGIR 2008, Singapore, Jul. 20-24, 2008, pp.179-186.
  • 3Heymann P, Koutrika G, Garcia-Molina H. Can social book- marking improve web search? In Proc. WSDM2008, PaloAlto, USA, Feb. 11-12, 2008, pp.195-206.
  • 4Ramage D, Heymann P, Manning C D, Garcia-Molina H. Clustering the tagged web. In Proc. WSDM2009, Barcelona, Spain, Feb. 9-12, 2009, pp.54-63.
  • 5http: / /www.dai-labor.de/en/ competence_centers/ irml/ data- sets/, April 2010.
  • 6Li X, Guo L, Zhao Y E. Tag-based social interest discovery. In Proc. WWW2008, Beijing, China, Apr. 21-25, 2008, pp.675- 684.
  • 7Wetzker R, Zimmermann C, Bauckhage C. Analyzing so- cial bookmaxking systems: A del.icio.us cookbook. In Proc. ECAI 2008 Mining Social Data Workshop, Patras, Greece, Jul. 21-25, 2008, pp.26-30.
  • 8Griffiths T L, Steyvers M. Finding scientific topics. In Proc. National Academy of Sciences, 2004, 101(Suppl.1): 5228- 5235.
  • 9Blei D M, Ng A Y, Jordan M I. Latent dirichlet allocation. Journal of Machine Learning Research, 2003, 3: 993-1022.
  • 10Lu C, Chen X, Park E K. Exploit the tripartite network of social tagging for web clustering. In Proc. CIKM2009, Hong Kong, China, Nov. 2-6, 2009, pp.1545-1548.

共引文献4

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部