Regularized Latent Semantic Indexing: A New Approach to Large-Scale Topic Modeling

被引:31
|
作者
Wang, Quan [1 ]
Xu, Jun [2 ]
Li, Hang [2 ]
Craswell, Nick [3 ]
机构
[1] Peking Univ, MOE Microsoft Key Lab Stat & Informat Technol, Beijing 100871, Peoples R China
[2] Microsoft Res Asia, Beijing, Peoples R China
[3] Microsoft Corp, Bellevue, WA 98004 USA
关键词
Algorithms; Experimentation; Topic modeling; regularization; sparse methods; distributed learning; online learning;
D O I
10.1145/2414782.2414787
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Topic modeling provides a powerful way to analyze the content of a collection of documents. It has become a popular tool in many research areas, such as text mining, information retrieval, natural language processing, and other related fields. In real-world applications, however, the usefulness of topic modeling is limited due to scalability issues. Scaling to larger document collections via parallelization is an active area of research, but most solutions require drastic steps, such as vastly reducing input vocabulary. In this article we introduce Regularized Latent Semantic Indexing (RLSI)-including a batch version and an online version, referred to as batch RLSI and online RLSI, respectively-to scale up topic modeling. Batch RLSI and online RLSI are as effective as existing topic modeling techniques and can scale to larger datasets without reducing input vocabulary. Moreover, online RLSI can be applied to stream data and can capture the dynamic evolution of topics. Both versions of RLSI formalize topic modeling as a problem of minimizing a quadratic loss function regularized by l(1) and/or l(2) norm. This formulation allows the learning process to be decomposed into multiple suboptimization problems which can be optimized in parallel, for example, via MapReduce. We particularly propose adopting l(1) norm on topics and l(2) norm on document representations to create a model with compact and readable topics and which is useful for retrieval. In learning, batch RLSI processes all the documents in the collection as a whole, while online RLSI processes the documents in the collection one by one. We also prove the convergence of the learning of online RLSI. Relevance ranking experiments on three TREC datasets show that batch RLSI and online RLSI perform better than LSI, PLSI, LDA, and NMF, and the improvements are sometimes statistically significant. Experiments on a Web dataset containing about 1.6 million documents and 7 million terms, demonstrate a similar boost in performance.
引用
收藏
页数:44
相关论文
共 50 条
  • [41] Hierarchical Semantic Indexing for Large Scale Image Retrieval
    Jia Deng
    Berg, Alexander C.
    Li Fei-Fei
    [J]. 2011 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2011, : 785 - 792
  • [42] Fast Large-Scale Multimedia Indexing and Searching
    Mohamed, Hisham
    Osipyan, Hasmik
    Marchand-Maillet, Stephane
    [J]. 2014 12TH INTERNATIONAL WORKSHOP ON CONTENT-BASED MULTIMEDIA INDEXING (CBMI), 2014,
  • [43] Modeling and analysis method of semantic description requirements of large-scale users
    Cheng F.
    Yu S.
    Chu J.
    Fan J.
    Hu Y.
    [J]. Jisuanji Jicheng Zhizao Xitong/Computer Integrated Manufacturing Systems, CIMS, 2022, 28 (02): : 601 - 611
  • [44] Approach of image modeling for large-scale geometric models
    Tang, Xiaoan
    Chen, Min
    Sun, Maoyin
    [J]. Jisuanji Fuzhu Sheji Yu Tuxingxue Xuebao/Journal of Computer-Aided Design and Computer Graphics, 2002, 14 (06): : 509 - 512
  • [45] Topic Modeling Twitter Data Using Latent Dirichlet Allocation and Latent Semantic Analysis
    Qomariyah, Siti
    Iriawan, Nur
    Fithriasari, Kartika
    [J]. 2ND INTERNATIONAL CONFERENCE ON SCIENCE, MATHEMATICS, ENVIRONMENT, AND EDUCATION, 2019, 2019, 2194
  • [46] A Large-scale Approach to Modeling Molecular Biosignatures: The Diatomics
    Cross, Thomas M.
    Benoit, David M.
    Pignatari, Marco
    Gibson, Brad K.
    [J]. ASTROPHYSICAL JOURNAL, 2022, 925 (01):
  • [47] A new approach to large-scale effectiveness evaluation
    Spencer, John
    Pill, Charles
    Curtis, Sian
    Kunyanga, Edward
    [J]. LANCET, 2011, 377 (9774): : 1317 - 1317
  • [48] A New Approach to build a Geographical Taxonomy of Adjacency Automatically Using the Latent Semantic Indexing Method
    El Midaoui, Omar
    El Qadi, Abderrahim
    Rahmani, Moulay Driss
    Aboutajdine, Driss
    [J]. 2015 INTELLIGENT SYSTEMS AND COMPUTER VISION (ISCV), 2015,
  • [49] A Joint MLE Approach to Large-Scale Structured Latent Attribute Analysis
    Gu, Yuqi
    Xu, Gongjun
    [J]. JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2023, 118 (541) : 746 - 760
  • [50] The New Clustering Strategy and Algorithm Based on Latent Semantic Indexing
    Yan, Bing
    Du, YaJun
    Li, ZhanShen
    [J]. ICNC 2008: FOURTH INTERNATIONAL CONFERENCE ON NATURAL COMPUTATION, VOL 1, PROCEEDINGS, 2008, : 486 - 490