Article: Three level weight for latent semantic analysis: an efficient approach to find enhanced semantic themes Journal: International Journal of Knowledge and Learning IJKL 2023 Vol 16 No.1 pp.56 72 Abstract: Latent semantic analysis is a prominent semantic themes detection and topic modelling technique. In this paper, we have designed a three-level weight for latent semantic analysis for creating an optimised semantic space for large collection of documents. Using this novel approach, an efficient latent semantic space is created, in which terms in documents comes closer to each other, which appear far away in actual document collection. In this approach, authors used two dataset: first is a synthetic dataset consists of small stories collected by the authors; second is benchmark BBC-news dataset used in text mining applications. These proposed three level weight models assign weight at term level, document level, and at a corpus level. These weight models are known as: 1 NPC; 2 NTC; 3 APC; 4 ATC. These weight models are tested on both the dataset, compared with state of the art term frequency and it has shown significant improved performances in term set correlation, document set correlation and has also shown highest correlation in semantic similarity of terms in semantic space generated through these three level weights. Our approach also shows automatic context clustering generated in dataset through three level weights. Inderscience Publishers linking academia, business and industry through research

Neuroinformatics and Semantic Representations: Theory and Applications It provides efficient implementations of popular algorithms such as Latent Semantic Analysis (LSA), Latent Dirichlet Allocation (LDA), and Word2Vec. Gensim is well-suited for tasks like document clustering, topic extraction, and word embeddings. However, it may not offer the breadth of functionality for general-purpose NLP tasks compared to other […]