skip to main content
research-article
Open access

Transformer-Based Topic Modeling for Urdu Translations of the Holy Quran

Published: 24 October 2024 Publication History

Abstract

Topic modeling enables the discovery of concealed themes and patterns in extensive text collections. It facilitates a thorough examination of the messages present in religious texts. Topic modeling for Quranic verses is a trending study area, with various translations already explored including Bahasa, English, and Arabic. Yet, there is a need for further research, particularly in Urdu translations of the Quran. In this study, we propose applying the BERTopic framework to Urdu translations of the Holy Quran. By leveraging the BERTopic approach, which incorporates a fine-tuned BERT model, we aim to capture the contextual nuances and linguistic complexities unique to the Quran. In this study, we utilized existing Urdu translations of the Quran from eight different translators sourced from Tanzil, a renowned resource for Quranic text and translations. We assessed the performance of our proposed BERTopic model compared to traditional techniques like LDA and NMF, using coherence and diversity metrics. The results indicate that our BERT-based approach outperforms these conventional methods, achieving an average coherence improvement of 0.03 and a diversity score of 0.83. These findings highlight the effectiveness of BERTopic in extracting meaningful topics from Urdu translations of the Holy Quran and contribute to the computational analysis of religious texts, supporting scholarly endeavors in comparative studies of Quranic translations in Urdu.

1 Introduction

The Quran, a religious text revealed to the Prophet Muhammad (Peace Be Upon Him) over 14 centuries ago, holds immense significance. Muslims widely read it for its spiritual importance and comprehensive guidance in addressing various challenges across social, economic, and religious aspects of life. Religion, embraced by diverse nations speaking different languages, translates its religious texts into numerous languages to guide its followers. However, translating a spiritual text can lead to varying interpretations due to the background and understanding of the translator.
Scholars delve into Quranic verses based on their areas of interest. For instance, some scholars focus on “mercy” in the Quran to identify verses, passages, and interpretations that emphasize mercy. Manually identifying topics and conducting in-depth studies from different Quranic translations is time-consuming because the meanings and ideas of different translators overlap from verse to verse and from chapter to chapter [Al Ghamdi and Khan 2022]. Hence, an automatic method is needed to identify topics from multiple translations of religious texts. One solution to this problem is topic modeling, which extracts hidden topics from an extensive collection of documents [Alhawarat 2015]. It allows scholars to gain valuable insights for focused research [Siddiqui et al. 2013]. Topic modeling has been applied to Quranic translations in various languages such as English [Al Ghamdi and Khan 2022], Arabic [Alshammeri et al. 2021b], and Indonesian [Rolliawati et al. 2020]. Still, such techniques are not available for Urdu translations of religious texts.
Urdu, a language rich in grammar but lacking in resources, poses challenges due to its numerous word derivations and inflections [Khalid et al. 2017]. Unlike English, Urdu is written from right to left and follows the Nastaliq writing style, similar to Arabic [Alhawarat and Hegazi 2018; Al Qudah et al. 2022]. Therefore, many essential resources and accurate text processing toolkits developed for other languages cannot be directly applied to Urdu [Daud et al. 2017]. Limited research on Urdu topic modeling has mainly focused on classical approaches [Shakeel et al. 2018; Rehman et al. 2018; Amin et al. 2020]. Topic modeling on religious texts is more challenging than on general texts because a single topic may repeat in different places in the Quran with varying details and contexts. Additionally, one verse may contain multiple topics, making it even more challenging due to the variations in Quranic translations. Although traditional techniques such as Latent Semantic Analysis [Landauer et al. 1998], Probabilistic Latent Semantic Analysis [Hofmann 1999], Nonnegative Matrix Factorization (NMF) [Lee and Seung 2000], and Latent Dirichlet Allocation (LDA) [Blei et al. 2003] are widely used and have provided valuable insights, they often struggle to capture contextual nuances and semantic representations present in Urdu translations of the Quranic text [Vayansky and Kumar 2020]. In recent years, Bidirectional Transformers for Language Understanding (BERT) [Devlin et al. 2019] has shown promise for context-aware language understanding, successfully applied in English and Arabic. It highlights the need for further research to leverage the potential of such models in Urdu translations of the Quran.
In this work, by leveraging BERT contextual understanding, we have proposed a topic modeling technique to analyze themes, concepts, and underlying meanings in the Urdu translations of the Quran. Our contributions lie in exploring the adaptability of BERT-based models to the intricacies of the Urdu language and Quranic text and in capturing the nuanced semantics of Urdu Quranic translations, thus offering a valuable resource for scholars, researchers, and practitioners of religious studies. Through rigorous experimentation, we demonstrate the potential of these models, therefore advancing the state of the art in text analysis for religious texts in Urdu while addressing the existing research gap.
The contributions of this work are outlined as follows:
We develop the Quran Urdu Topic Modeling Corpus (Quran-UTM), which comprises Urdu translations of the Quran, including eight esteemed translations by widely recognized translators. To focus on the actual content for topic modeling, the irrelevant information such as metadata, punctuation, and diacritics are removed from each Urdu translation. Corpus and source code are publicly available for the research community.1
We have developed a framework for context-aware topic modeling using BERT.
We have assessed the performance of our BERT-based model against traditional methods, such as LDA and NMF. The results show that our BERT-based approach outperforms these conventional methods, achieving significantly higher coherence and diversity scores.
This approach has the potential to bridge the existing gap in Quranic topic modeling for Urdu translations and unlock new insights into the Quran’s profound message in the Urdu language.
The rest of the article is organized as follows. Section 2 presents a comparative review of existing work in topic modeling. Section 3 presents the Quran-UTM dataset. Section 4 presents the proposed methodology, and Section 5 discusses the experimental setup and presents the obtained results. Section 6 provides a discussion and analysis of the results, and finally, Section 7 concludes the study and offers suggestions for future research directions.

2 Related Work

Several studies have been done on topic modeling of Quranic text and its translations in recent years. While most of these studies have focused on the original Arabic script [Abuzayed and Al-Khalifa 2021] and English translations [Alshammeri et al. 2021b] of the Holy Quran, some have also explored other domains. This section provides an overview of previous works that deal with traditional topic modeling techniques for Quranic text and its translations and the application of context-aware advanced methods.

2.1 Classical Techniques for Topic Modeling

Classical techniques for topic modeling have been employed for text in multiple languages, including English, Arabic, and Urdu.
Rahman et al. [2018] aim to classify topics in each verse of Surah-Al-Baqarah from the Quran using machine learning techniques, particularly employing Support Vector Machine (SVM) as the primary algorithm. A comparative analysis with other classification algorithms is conducted. The research contributes to developing a computational environment for text mining the Quran and enabling efficient topic identification. However, there is a need for further investigation to explore additional methods or approaches that can enhance the accuracy and comprehensiveness of topic classification in Islamic literature.
Siddiqui et al. [2015] apply a probabilistic topic modeling algorithm, LDA, to discover the thematic structure of the Quran. By considering each Surah as a document, the paper aims to automatically identify the hidden thematic structure of the Quran, which contains vast information addressing various aspects of human life. Using the Arabic Quran as the corpus, the study successfully identifies significant themes in the Surahs and determines the most important terms associated with these themes.
Seifollahi et al. [2021] introduce a two-stage algorithm for topic modeling, addressing the limitations of traditional methods. While LDA relies on simplistic bag-of-words assumptions, this innovative approach leverages word embeddings and co-occurrence patterns. The first stage determines topic–word distributions by soft-clustering embedded n-grams from documents. In the second stage, it computes document–topic distributions by sampling topics from the topic–word distributions. This method capitalizes on word embeddings, providing a more accurate representation. Nonetheless, the article does not explicitly address potential constraints or difficulties linked to the proposed approach, leaving an opportunity for further investigation in future research.
Alshammeri et al. [2021a] employ the Natural Language Processing (NLP) method Doc2vec to detect semantic-based similarity between verses of the Quran. By mapping Arabic Quranic verses to numerical vectors encoding their semantic properties, the model achieves 76% accuracy in similarity detection compared to annotated textual similarity datasets. These findings provide a foundation for further research in Quranic semantic analysis, although improvements are needed to enhance the model’s performance and robustness.

2.2 Context-aware Techniques for Topic Modeling

In recent years, researchers have focused on using context-aware techniques for topic modeling for text in various languages.
Abuzayed and Al-Khalifa [2021] conduct an experimental study on BERT for Arabic Topic Modeling, specifically focusing on the effectiveness of the BERTopic technique. The study aims to explore the application of BERTopic using different Pre-trained Arabic Language Models as embeddings. By evaluating the results against traditional techniques like LDA and NMF, the study highlights the superior performance of BERTopic in uncovering latent topics in Arabic text. The evaluation metric was the Normalized Pointwise Mutual Information (NPMI) measure. This research contributes to the literature by emphasizing the potential of BERT-based techniques for Arabic topic modeling and addressing a research gap in this area.
Alhaj et al. [2022] propose a deep-learning-based approach using BERTopic to improve the classification of cognitive distortions in Arabic content on Twitter. By enriching text representation with latent topics derived from BERTopic, the study achieves improved performance compared to baseline models. However, more research is needed to understand the specific impact of BERTopic on cognitive distortion classification in Arabic on social media platforms. Further investigation is required to explore this aspect and provide deeper insights.
Alsaleh et al. [2021] introduce AraBERT, a powerful language model based on BERT that has demonstrated remarkable performance in various Natural Language Processing tasks. In this study, the authors leverage AraBERT to classify pairs of Quranic verses from the QurSim dataset as semantically related. The authors pre-processed the dataset and created three subsets for comparison. AraBERT, AraBERTv0.2, and AraBERTv2 are evaluated to determine the optimal version for our datasets. AraBERTv0.2 achieves an impressive accuracy score of 92% on a dataset containing labels “2” and “–1,” the latter generated outside the QurSim dataset. While this paper successfully explores Quranic verse semantic relatedness using AraBERT, there is still room for further research and improvement to better understand the contextual nuances within the Quranic text.
Alshammeri et al. [2021c] introduce a novel approach for detecting semantic similarity in the Quran using a Siamese transformer-based architecture. The study focuses on the significance of semantic similarity detection in various natural language comprehension tasks and its relevance to NLP applications. Leveraging Arabic pre-trained contextual representations, the proposed model generates verse embeddings that capture semantically meaningful information. The twin transformer networks are fine-tuned on a Quranic semantic similarity dataset, resulting in improved performance compared to previous studies. However, the article does not explicitly address any potential limitations or challenges associated with the proposed approach, leaving an opportunity for further investigation in future research.
The article of [Yang et al. 2023] introduces sDTM, a novel supervised topic modeling approach that enhances traditional methods like LDA by incorporating auxiliary data such as star ratings or post categories. Combining a neural variational autoencoder and a recurrent neural network, sDTM improves empirical estimation and predictive performance. While making valuable contributions to the field, the article could benefit from further investigation into scalability and practical limitations, ensuring its applicability to large datasets and real-world text analytics tasks. In their study, George and Sumathy [2023] incorporated LDA with clustering, BERT, and dimensionality reduction techniques (PCA, t-SNE, UMAP) to enhance topic modeling. The experiments were conducting using the open-source, cross-platform Integrated Development Environment (IDE) Spyder. The machine learning experiments were performed using Scikit-learn and NLTK tools for analysis. While the results demonstrated improved topic extraction on benchmark datasets, there is a need for further quantitative analysis of topic quality enhancement, as well as considerations for scalability and real-world deployment challenges.
Table 1 provides a quick comparison of existing research categorized by language (English, Arabic, Indonesian, and Urdu), arranged in reverse chronological order within each language group based on their publication year. Notable observations from the existing studies include:
Table 1.
ReferenceLanguageSentence SemanticsWord SemanticsContext-aware TechniquesCovers Quranic Domain
Egger and Yu [2022]EnglishYesYesYesNo
Al Ghamdi and Khan [2022]EnglishNoYesNoYes
Weng et al. [2022]EnglishNoNoNoNo
Silveira et al. [2021]EnglishNoNoNoYes
Aftar et al. [2024]ArabicYesYesYesNo
Abdelrazek et al. [2022]ArabicYesYesYesNo
Alshammeri et al. [2021b]ArabicYesYesNoYes
Abuzayed and Al-Khalifa [2021]ArabicYesYesYesNo
Hutama and Suhartono [2022]IndonesianYesYesYesYes
Rolliawati et al. [2020]IndonesianNoNoNoYes
Mustafa et al. [2023]UrduNoYesNoNo
Zoya et al. [2021]UrduNoNoNoNo
Mustafa et al. [2021]UrduNoNoNoNo
Amin et al. [2020]UrduNoNoNoNo
Munir et al. [2019]UrduNoNoNoNo
Rehman et al. [2018]UrduNoNoNoNo
Shakeel et al. [2018]UrduNoNoNoNo
Table 1. Comparison of Various Aspects
Limited Research on Al-Quran Urdu Translation. As shown in Table 1, there are few studies on Urdu topic modeling between 2018 and 2023, and these studies do not cover Quranic text, which indicates the research community’s limited focus on this specific NLP task. Conversely, topic modeling on religious texts has been conducted in various languages such as English, Arabic, and Indonesian, highlighting the need for more comprehensive investigations into Al-Quran Urdu translation.
Scarcity of Context-aware Topic Modeling Techniques for Urdu. Table 1 reveals that many studies in English and Arabic have applied context-aware techniques, especially transformer-based language models like BERT and its variations. Most of these studies employing context-aware topic modeling techniques have been published between 2021 and 2024, demonstrating the rapid adoption of these models in the research community. In contrast, only one recent study in Urdu utilized static word2vec embeddings [Mustafa et al. 2023]. Due to variations in Quranic translation, identifying topics from different Quranic translations is more challenging. Static embeddings assign fixed vectors to words regardless of context, lacking adaptability, while transformer-based contextual embeddings dynamically capture nuanced meanings by contextualizing words within various Quranic Urdu translations. Consequently, there is a need to evaluate the effectiveness of context-aware topic modeling techniques in the Urdu translation of the Quran, which is the focus of this research.
In conclusion, context-aware techniques have exhibited remarkable effectiveness in topic modeling for religious texts, particularly in English and Arabic, when applied to the Holy Quran. However, utilizing these techniques for topic modeling in Urdu within the religious context is an under-explored research area. Our study aims to bridge this gap by leveraging BERT-based approaches for text analysis and topic modeling for Urdu translations of the Quran. Our contributions lie in not only exploring the adaptability of BERT-based models to the intricacies of the Urdu language and Quranic text but also capturing the nuanced semantics of Urdu Quranic translations, thus offering a valuable resource for scholars, researchers, and practitioners in the field of religious studies. Through rigorous experimentation, we demonstrate the potential of these models, thus advancing the state-of-the-art text analysis for religious texts in Urdu while addressing this pressing research gap.

3 Quran-UTM: Quran Urdu Topic Modeling Corpus

To address the gap in topic modeling for Al-Quran Urdu translations, we have developed a new corpus called Quran-UTM. Figure 1 provides an overview of the Quran-UTM corpus creation process. The subsequent sections present detailed descriptions of each step involved in developing our dataset and its statistical analysis.
Fig. 1.
Fig. 1. Overview of Quran-UTM dataset.

3.1 Data Acquisition

We gathered the different Quran Urdu translations from the Tanzil project, accessible at https://tanzil.net/trans/. Only 8 of the 115 Quran translations available on Tanzil are in Urdu. Our study utilized all available Urdu translations, specifically those by Junagarhi, Jalandhry, Jawadi, Ahmed Ali, Kanzuliman, Maududi, Najafi, and Qadri. These translations offer diverse perspectives and interpretations, providing a comprehensive collection of Quranic texts in the Urdu language. The translations were manually downloaded from the Tanzil platform, emphasizing the process’s transparency and validity. The acquired dataset ensures the availability of trustworthy and valuable resources for the subsequent stages of the topic modeling research by sourcing the translations through manual download from Tanzil. Each translation encompasses 6,236 verses from 114 Surahs.

3.2 Data Cleaning

After collecting the eight Quran Urdu translations, we observed that these translations had some irrelevant information, such as metadata, punctuation, and diacritics. Eliminating noise and irrelevant information enhances the quality and relevance of the Quran Urdu translations. It not only reduces computational overhead but also makes it easier to extract meaningful topics, ultimately leading to better model performance and valuable insights in topic modeling. One example of cleaning the verse of Ahmad Ali’s translation is given in Figure 2.
Fig. 2.
Fig. 2. Overview of cleaning of Al-Quran Urdu translation.
Metadata Removal. To ensure consistency, we removed the metadata present at the beginning and end of each translation’s text file. This step helped streamline the data and focus solely on the actual content of the translations. Python’s Pandas2 library was utilized to accomplish these tasks efficiently.
Punctuation Removal. Punctuation marks can pose challenges during text tokenization and do not contribute to retrieval. Hence, punctuation marks were removed from the text to mitigate these issues. The Regex3 library is employed for this task.
Diacritics Removal. In Urdu, diacritics are used to indicate pronunciation, but they do not change the fundamental meaning of words. To standardize the Urdu translations, we removed the diacritics using the Urduhack4 library that ensures words are treated consistently by the model, which is crucial for accurate topic detection.

3.3 Quran-UTM Corpus Statistics

The statistics of the Quran-UTM dataset are provided in Table 2. It can be observed from the table that the Qadri translation contains the most significant number of words and characters, with 232,744 words and 1,030,375 characters. The Najafi translation has the second-largest count, with 206,847 words and 902,665 characters. On the other hand, the Ahmad Ali translation has the smallest number of unique words, consisting of 9,011 words. The detailed statistics of the Quran-UTM dataset are provided in Table 2.
Table 2.
TranslatorWordsCharactersUnique Words
Junagarhi187,353802,13611,262
Jalandhry188,667807,89811,424
Jawadi186,652799,2359,337
Ahmed Ali171,124716,8869,011
Kanzuliman159,685679,8599,918
Maududi204,593864,16811,807
Najafi206,847902,66512,929
Qadri232,7441,030,37515,136
Table 2. Statistics of Quran-UTM Dataset

4 Proposed Methodology

This section presents our proposed BERTopic model [Grootendorst 2022] for the Quranic Urdu translations topic modeling. To capture the contextual nuances of the Quran in Urdu, we utilize BERT embeddings, which convert the Urdu translations into semantic representations in vectors. Afterward, we reduce the dimensionality of the vector representation through Uniform Manifold Approximation and Projection (UMAP), followed by clustering to group verses with similar semantics. We use descriptive analysis and evaluation metrics to infer and assess the discovered topics. Figure 3 provides a workflow of our proposed methodology. The details of each step are provided in subsequent sections.
Fig. 3.
Fig. 3. An overview of the proposed methodology.

4.1 BERT Embeddings

First, we transform our documents into numerical representations using BERT embedding, which differentiates the different meanings of the same word and various words with the same meaning based on their context within a sentence.
For example, as illustrated in Figure 4, the word in the Surah Taha represents light and guidance. It symbolizes the divine illumination and the hope for spiritual insight that Prophet Musa sought when he approached the fire on Mount Sinai. Conversely, in Surah Al-Baqrah verse number 24 denotes Hellfire, representing the concept of punishment and suffering in the afterlife. These two contexts show how the same word, , diverges into vastly different topics in topic modeling. One context is associated with light, guidance, and knowledge, while the other pertains to punishment, retribution, and the concept of Hell. It illustrates the richness of language in the Quran, where a single term can encapsulate multiple layers of meaning depending on its usage.
Fig. 4.
Fig. 4. Same word “fire” in different context.
Different translators choose different words to translate the original text based on their understanding. For example, in Figure 5, the words and in these translations serve the same purpose of conveying a successful outcome for those who follow divine guidance. These words belong to the same topic. Traditional techniques failed to capture the contextual nuances and semantic representations in different Urdu translations of the Quranic text. BERT achieves this through bidirectional analysis, considering preceding and following words to generate context-specific embeddings.
Fig. 5.
Fig. 5. Two different words in same context.
To apply BERT embedding, we utilized the pre-trained sentence transformer model “paraphrase-multilingual-MiniLM-L12-v2” from Hugging Face [Reimers and Gurevych 2019]. Hugging Face is a popular library and platform that provides a wide range of pre-trained models for natural language processing tasks. Sentence transformers, a type of deep learning model, have gained significant attention for their ability to capture semantic representations of sentences and paragraphs. The sentence transformer model we employed, specifically the “paraphrase-multilingual-MiniLM-L12-v2,” is designed to map sentences and paragraphs to a 384-dimensional dense vector space, enabling clustering or semantic search tasks. By leveraging transformer architectures, sentence transformers encode contextual information and the semantic meaning of the text into these dense vector representations.
In our case, the sentence transformer model was employed to generate BERT embeddings of the proposed Quran-UTM dataset. These embeddings serve as high-dimensional representations of the text, capturing the intricate relationships between words and phrases. By leveraging the contextual information encoded in the embeddings, we can understand the finer details and subtle aspects of the Quranic text in Urdu, uncovering its hidden meanings and more profound significance.

4.2 Dimensionality Reduction

When using BERT embeddings, each word or token of the Quran-UTM dataset is represented by a high-dimensional vector. This high dimensionality can pose challenges regarding computational complexity, memory requirements, and the interpretability of the embeddings. A dimensionality reduction technique is employed to handle the high dimensionality of BERT embeddings. We utilized the UMAP model for dimensionality reduction [McInnes et al. 2018]. There are traditional approaches to reduce the dimensionality of document embedding, such as t-Distributed Stochastic Neighbor Embedding (t-SNE), which emphasizes preserving local similarities [Van der Maaten and Hinton 2008], and Principal Component Analysis (PCA), which is a linear dimensionality reduction technique that focuses on the global structure of the data by projecting it into a lower-dimensional space [Pearson 1901]. In contrast, UMAP offers a balance between preserving global and local structure in data, making it suitable for effectively visualizing and analyzing complex datasets [McInnes et al. 2018]. Additionally, research has shown that reducing high-dimensional embedding with UMAP can enhance the performance of clustering algorithms like K-means and Hierarchical Density-Based Spatial Clustering of Applications with Noise (HDBSCAN) in terms of accuracy and processing time [Allaoui et al. 2020]. UMAP uses a cross-entropy-based loss function to optimize the low-dimensional representation. The loss function is defined in Equation (1):
\begin{equation} L = \sum _{i,j} w_{ij} \left[ d_{ij} \log \left(\frac{d_{ij}}{\sigma _{ij}}\right) - (1 - d_{ij}) \log \left(\frac{1 - d_{ij}}{1 - \sigma _{ij}}\right) \right], \end{equation}
(1)
where:
\(i, j\) represent pairs of data points.
\(d_{ij}\) is the distance between data points i and j in the high-dimensional space.
\(\sigma _{ij}\) is the similarity between data points i and j in the low-dimensional space (after applying a Gaussian kernel).
\(w_{ij}\) is a weight that adjusts the importance of each pair of data points based on their distance in the nearest neighbor graph.
UMAP works by mapping the high-dimensional space of BERT embeddings onto a lower-dimensional space, typically two or three dimensions, which can be easily visualized. The mapping process considers the underlying manifold structure of the data, capturing the essential similarities and differences between the verses. By reducing the dimensionality, we can simplify the data representation, making it more manageable for further analysis and interpretation.

4.3 Clustering

Clustering plays a crucial role in the BERTopic model as it enables grouping similar Quran verses based on their semantic similarities. This step is essential for organizing the vast corpus of Quran-UTM into meaningful clusters or groups, each representing a potential topic. A suitable algorithm is employed for clustering, such as K-means or HDBSCAN. By default, the BERTopic framework uses HDBSCAN clustering [McInnes et al. 2017]. HDBSCAN is a density-based clustering algorithm that builds upon the traditional DBSCAN algorithm by incorporating a hierarchical approach. It has the ability to handle clusters of varying densities and identify outliers. Sometimes, HDBSCAN produces numerous outliers, potentially leading to excessive outlier clusters, making it difficult to identify meaningful clusters accurately. So, in this specific study, we choose the K-means algorithm because it prevents the generation of excessive outliers. It is a partition-based clustering algorithm that iteratively assigns data points to the nearest cluster centroid to minimize the sum of squared distances within each cluster.
By applying K-means to the reduced-dimensional BERT embeddings, we discover distinct clusters of Quran verses that exhibit similar semantic characteristics. These clusters represent topics or themes that emerge from the Quran Urdu translations. By grouping verses with shared semantic similarities, we gain valuable insights into the underlying structure and content of the text, facilitating the exploration and analysis of the translations. Clustering helps to organize the Quran Urdu translations into coherent groups, making it easier to identify patterns, trends, and topics within the text. It provides a framework for understanding the relationships between different verses and unveiling the underlying themes present in the translations.

4.4 Tokenizer

To ensure modularity in BERTopic’s algorithm, a flexible topic representation technique is essential, one that avoids assumptions about cluster structures. We employ a bag-of-words approach, where all documents within a cluster are combined into a single document to count word frequencies, generating a cluster-level representation. This process uses the default tokenizer from CountVectorizer in the sklearn library,5 which breaks down the text into individual words or tokens and handles tasks like stop word removal.
Stopwords, being less meaningful and the most frequent words in any text, connect the significant parts of a sentence and build context. However, our approach uses sentence transformers that extract contextual information from sentences, making the removal of stopwords in the initial stages unnecessary.
Eliminating stopwords during the post-processing phase was crucial for enhancing the quality and relevance of the identified topics by removing less informative words and focusing on more meaningful content in the Quran-UTM dataset. In our proposed framework, we utilized a freely available, pre-existing stopwords list [Shafi et al. 2023].

4.5 Topic Inference

After tokenizing and removing the stopwords within a cluster, the next step is to analyze the verses within each cluster to identify the most representative words and themes. Topic inference is the process of identifying and extracting meaningful words from clusters. It involves analyzing the content and patterns within the data to uncover underlying themes, concepts, or subjects prevalent across the clusters of verses.
Descriptive analysis techniques are applied to the verses within each cluster during topic inference. It involves calculating the importance or significance of words within a cluster using Class-based Term Frequency-Inverse Document Frequency (C-TF-IDF).
The C-TF-IDF weight for a term t in a document d belonging to a cluster c is given by Equation (2):
\begin{equation} \text{C-TF-IDF}(t, d, c) = \text{TF}(t, d) \times \text{IDF}(t, D) \times \log \left(\frac{\text{Total number of documents in } D}{\text{No. of docs containing term $t$ in cluster $c$}}\right). \end{equation}
(2)
By examining the C-TF-IDF scores of words within a cluster, we can identify the most characteristic and informative words for that particular topic. These words provide insights into the primary themes and content represented by the cluster, effectively inferring the topics present in the Quran-UTM dataset.

4.6 Evaluation Measures

This study uses two well-known topic modeling evaluation metrics—coherence and diversity—to evaluate the quality and performance of topic models [Lau et al. 2014; Röder et al. 2015; Terragni et al. 2021]. The coherence measure quantifies the interpretability and meaningfulness of topics by considering the co-occurrence of terms within topics. We employ two coherence metrics, NPMI, [Lau et al. 2014] and Coherence Value (\(C_V\)), [Röder et al. 2015], and a diversity metric, Inverted Rank-Biased Overlap (IRBO) [Terragni et al. 2021], to assess the model’s performance. The details of each metric are as follows:

4.6.1 NPMI Score.

The first coherence metric we use to evaluate the topics is NPMI [Bouma 2009], which has been shown to correlate with human judgment [Lau et al. 2014]. It ranges from (–1,1), where –1 signifies that the words never appear together, 0 suggests that their occurrences are entirely independent, and 1 indicates a perfect and consistent co-occurrence of the words. The NPMI score between two words \(\left(x_{i}, x_{j}\right)\) is calculated as follows:
\begin{equation} \operatorname{NPMI}\left(x_{i}, x_{j}\right)=\frac{\log \frac{p\left(x_{i}, x_{j}\right)+\epsilon }{p\left(x_{i}\right) p\left(x_{j}\right)}}{\log p\left(x_{i}, x_{j}\right)+\epsilon }. \end{equation}
(3)
The probabilities \(p\left(x_{i}\right)\) and \(p\left(x_{i}, x_{j}\right)\) represent the occurrence frequencies of word \(x_i\) and the co-occurrence of word pair \(\left(x_{i}, x_{j}\right)\), respectively. These probabilities are estimated based on their frequencies in a corpus. Here, \(\epsilon\) is added to avoid division by zero. The NPMI score calculates the normalized mutual information between two words and then averages the values across all word pairs in all topics.

4.6.2 \(C_V\) Score.

\(C_V\) is a coherence evaluation metric that assesses the quality of topics generated by topic modeling algorithms. It utilizes a sliding window, a one-set segmentation of the top words, and an indirect confirmation measure involving NPMI and cosine similarity. Given the top T words of a topic, \(\left(x_1, x_2, \ldots , x_T\right)\). The calculation of the \(C_V\) coherence score is as follows:
\begin{equation} C_V=\frac{1}{T} \sum _{i=1}^T \cos \left(\mathbf {v}_{\mathrm{NPMI}}\left(x_i\right), \mathbf {v}_{\mathrm{NPMI}}\left(\left\lbrace x_i\right\rbrace _{i=1}^T\right)\right). \end{equation}
(4)
The NPMI score is computed using Equation (3). Likewise, the average \(C_V\) score is obtained by considering the scores of all topics. This approach has been highly influential in evaluating topic quality, as evidenced by extensive comparisons with other commonly used coherence measures. The \(C_V\) metric yields the best scores that closely align with human evaluations of topic quality (see Röder et al. [2015] for experimental results).

4.6.3 IRBO Score.

The diversity metric indicates the semantic diversity of the generated topics. Previous studies have proposed different diversity metrics to determine the distinctiveness of the generated topics [Nan et al. 2019; Burkhardt and Kramer 2019; Dieng et al. 2020]. In this study, we use the recently proposed IRBO metric to quantify the diversity of topics [Terragni et al. 2021]. It assigns a score of 0 when the topics are identical and 1 when they are entirely dissimilar. Suppose we have a set \(\aleph\) containing T topics. Each topic is represented as a list of words, and the order of words within each list reflects their importance or ranking within the topic. The IRBO of these topics is calculated as follows:
\begin{equation} \operatorname{IRBO}(\aleph)=1-\frac{\sum _{i=2}^T \sum _{j=1}^{i1} \operatorname{RBO}\left(l_i, l_j\right)}{n}, \end{equation}
(5)
where \(n= (_2^T)\) represents the number of pairs of lists to be compared. The term \(RBO\left(l_i, l_j\right)\) refers to the traditional Rank-Biased Overlap measure between two ranked lists \(l_i\) and \(l_j\) [Webber et al. 2010]. IRBO enables the comparison of lists even when they do not necessarily share the same items and may not cover all items in a given domain. When two lists (topics) share common words, the IRBO score is lower if the shared words are found at the top than when they are located lower in the ranked lists.

5 Experimental Setup and Results

In this section, we describe the experimental setup and the results.

5.1 Experimental Setup

We compare the proposed BERTopic model with two classical models, LDA and NMF. We implement LDA and NMF using the Gensim6 library. We also employed the Python library Gensim, specifically its ”CoherenceModel” class, to calculate coherence scores. This class allows us to evaluate the coherence of topics generated by different models, including BERTopic, LDA, and NMF. The Coherence measure provided by Gensim considers the degree to which words tend to appear together within the same topic and how distinct they are from words in other topics. This assessment provides a measure of topic coherence that reflects the semantic consistency and relevance of the topics generated by our models.
One of the advantages of BERTopic is that it does not require specifying the number of topics in advance. However, to ensure a fair comparison across all models, we standardize by setting the same number of topics and extracting the top 10 words within each topic. We determine the optimal number of topics by iterating the LDA model over a range from 5 to 50 topics, in increments of 5, based on the \(C_V\) score, as shown in Figure 6. From the figure, it can be seen that the LDA model achieves the highest \(C_V\) score with five topics. Therefore, we compare all models using this value, setting the number of topics to five.
Fig. 6.
Fig. 6. \(C_V\) score of LDA by increasing number of topics.
The diverse dimensions of SBERT models exhibit varying capabilities in capturing semantic information, which affects the quality of embeddings. Therefore, the default UMAP dimension of 2 may sometimes yield optimal results. To address this, we experimented with different dimensions for the Maududi translation, ranging from 2 to 14, in increments of 1. The \(C_V\) scores for each dimension are shown in Figure 7. As observed, dimension 4 achieves the highest \(C_V\) score, so we set the UMAP model’s dimension value to 4. UMAP’s stochastic nature might cause variations in BERTopic results, so we set a random generator value of 42 for reproducibility. Additionally, we enabled low-memory mode in UMAP to avoid memory issues.
Fig. 7.
Fig. 7. \(C_V\) score on Maududi translation with different dimensions.
We set calculate-probability to true when instantiating BERTopic to obtain a document–topic probability matrix. This matrix provides the probability distribution of each document across different topics, allowing for a more fine-grained analysis and interpretation of the topic model results.

5.2 Results

Table 3 presents the results of three topic modeling techniques—LDA, NMF, and the proposed BERTopic—applied to eight different translations of the Quran: Junagarhi, Jalandhry, Jawadi, Ahmed Ali, Kanzuliman, Maududi, Najafi, and Qadri. The models are evaluated based on CV, NPMI, and IRBO scores.
Table 3.
ModelsMetricsJunagarhiJalandhryJawadiAhmed AliKanzulimanMaududiNajafiQadri
LDACV0.460.410.480.460.510.540.430.44
NPMI\(-\)0.01\(-\)0.05\(-\)0.02\(-\)0.010\(-\)0.03\(-\)0.01\(-\)0.02
Diversity0.510.540.650.470.540.540.490.36
NMFCV0.550.570.510.540.560.590.550.56
NPMI0.010.010.010.020.01\(-\)0.0100.03
Diversity0.830.770.840.790.750.750.780.86
BERTopicCV0.520.510.490.470.510.620.550.53
NPMI0.020.02\(-\)0.01\(-\)0.020.020.10.020.04
Diversity0.870.790.880.710.80.830.860.87
Table 3. Comparison of Proposed Model (BERTopic) with State-of-the-art Models
The table shows that BERTopic achieves the highest scores on the Maududi translation, with a CV of 0.62 and an NPMI of 0.1. Additionally, the Jawadi translation achieves a high IRBO score of 0.88 using the BERTopic model. The NPMI scores of the LDA model on each translation are generally negative or zero, suggesting that the word co-occurrence patterns might not align well with the human judgment of topic coherence. In contrast, the NPMI scores for the NMF and BERTopic models are mostly positive, indicating that their topics are more coherent. The varying results across translations highlight the richness and diversity of themes discovered in the Quran’s Urdu translations.
Figure 8 presents the average scores for each translation. The figure shows that the average NPMI and IRBO scores for BERTopic, 0.03 and 0.83, respectively, are higher than those for the classical models, LDA and NMF. Although NMF achieves the highest average CV score of 0.55, BERTopic performs well in the remaining two metrics.
Fig. 8.
Fig. 8. Average scores over all translations.
Figure 9 displays the word scores for each topic derived from the BERTopic model applied to the Kanzuliman translation from the Quran-UTM dataset. These word scores indicate the importance or relevance of each word within a specific topic. The figure demonstrates that topics are represented by numerous words, beginning with the most representative ones. Each word is assigned a c-TF-IDF score, with higher scores indicating greater representativeness within the topic. For example, in the first topic of the Kanzuliman translation, the word “” has the highest representation, while “” is the least represented word.
Fig. 9.
Fig. 9. Topic word scores of Kanzuliman translation.

6 Discussion and Analysis

In our study, we performed topic modeling on all of these translations to comprehensively understand the variations in their content. For example, we focused on exploring the topic of “Hellfire” across these translations and made an interesting observation. In Qadri’s translation, this topic was found in 126 different places within the Quranic text, whereas in Kanzuliman’s translation, it appeared only 84 times, as shown in Figure 10. It’s important to note that this difference in the frequency of topic occurrence does not imply that these translations are inherently different, as the Quran is considered the word of Allah. However, it does suggest that specific translations may be more explanatory than others.
Fig. 10.
Fig. 10. Comparison of the topic “Hellfire” in various translations.
Upon manual inspection of Qadri’s translation, we found that it provides additional explanations for the translated verses by including parenthetical explanations. This explanatory translation style may contribute to the higher frequency of topics identified in Qadri’s translation. Figure 11 showcases this feature in Qadri’s translation, highlighting how the inclusion of explanations enhances the richness of the translated text. This simple example shows that topic modeling can be effectively utilized for comparative analysis of the Quran and other religious texts.
Fig. 11.
Fig. 11. Manual exploration of two translations showing difference in word frequencies in them.
Tables 4 and 5 present topics from the Kanzuliman and Ahmed Ali translations, respectively. The variations in word choice and phrasing between the two translations reflect subtle differences in interpretation. For instance, Topic 4 in the Kanzuliman translation frequently references specific figures like and , highlighting historical or narrative contexts. In contrast, in Topic 3 of the Ahmed Ali translation is discussed within a broader context, including terms like (prophets) or (messengers). Despite addressing divine entities using different terms, such as , , and , the topics maintain similar meanings in context. It can also be observed that Topic 5 in both translations pertains to the same overarching theme, yet the words and their relevance scores differ. It demonstrates that our BERT-based context-aware topic modeling technique effectively captures the nuances of different translations by considering the surrounding context of words. This approach ensures a more accurate and meaningful extraction of themes, preserving the complexity of the text. This method aids scholars in delving deeper into these topics, examining the variations and nuances in translations, and uncovering further insights. Topic modeling facilitates rigorous analysis, allowing researchers to explore religious texts’ diverse interpretations and perspectives. It promotes meaningful dialogue and advances our understanding of these sacred texts.
Table 4.
Table 4. Topics Extracted from Kanzuliman Translation with Their Top Keywords
Table 5.
Table 5. Topics Extracted from Ahmed Ali Translation with Their Top Keywords

7 Conclusion and Future Work

Topic modeling in Urdu translations of the Quran presents significant challenges. Manually identifying topics and conducting comprehensive studies across different translations is highly time-consuming due to the overlapping meanings and ideas between verses and chapters. An automatic method for identifying topics from various translations is needed. While topic modeling has been applied to Quranic translations in English, Arabic, and Indonesian languages, similar techniques for Urdu translations of religious texts are not yet available. In this study, we introduce an approach for topic modeling in Urdu translations of the Quran utilizing BERT-based techniques. Our results reveal the potential of BERT embeddings for topic extraction and analysis. We applied the BERTopic model to eight different Urdu translations of the Quran and compared its performance with two traditional models, LDA and NMF, using coherence and diversity metrics. Our findings show that the BERT-based method surpasses these conventional techniques, with an average coherence improvement of 0.03 and a diversity score of 0.83. These results highlight the effectiveness of BERTopic in identifying meaningful topics from Urdu translations of the Holy Quran and contribute to the computational analysis of religious texts, aiding scholarly research in comparative studies of Quranic translations in Urdu. Despite these promising results, there is room for improvement. One area of improvement is the need for a dedicated Urdu-specific BERT model. Monolingual BERT models exist for various languages, but none specifically for the Urdu language. In future work, developing an Urdu-specific BERT model will significantly enhance the results and improve topic modeling performance in Urdu translations.

Footnotes

References

[1]
Aly Abdelrazek, Walaa Medhat, Eman Gawish, and Ahmed Hassan. 2022. Topic modeling on Arabic language dataset: Comparative study. International Conference on Model and Data Engineering (MEDI’22). Springer, 61–71. DOI:
[2]
A. Abuzayed and H. Al-Khalifa. 2021. BERT for Arabic topic modeling: An experimental study on BERTopic technique. In Procedia CIRP. Elsevier B.V., 191–194. DOI:
[3]
Sania Aftar, Luca Gagliardelli, Amina El Ganadi, Federico Ruozzi, and Sonia Bergamaschi. 2024. A novel methodology for topic identification in Hadith. In Proceedings of the 20th Conference on Information and Research Science Connecting to Digital and Library Science (IRCDL’24).
[4]
Norah Mohammad Al Ghamdi and Muhammad Badruddin Khan. 2022. Assessment of performance of machine learning based similarities calculated for different English translations of Holy Quran. International Journal of Computer Science & Network Security 22, 4 (2022), 111–118. DOI:
[5]
Islam Al Qudah, Ibrahim Hashem, Abdelaziz Soufyane, Weisi Chen, and Tarek Merabtene. 2022. Applying latent Dirichlet allocation technique to classify topics on sustainability using Arabic text. Science and Information Conference (SAI’22). Springer, 630–638. DOI:
[6]
Fatima Alhaj, Ali Al-Haj, Ahmad Sharieh, and Riad Jabri. 2022. Improving Arabic cognitive distortion classification in Twitter using BERTopic. International Journal of Advanced Computer Science and Applications 13, 1 (2022), 854–860. DOI:
[7]
Mohammad Alhawarat. 2015. Extracting topics from the holy Quran using generative models. International Journal of Advanced Computer Science and Applications 6, 12 (2015), 288–294. DOI:
[8]
Mohammad Alhawarat and M. Hegazi. 2018. Revisiting k-means and topic modeling, a comparison study to cluster arabic documents. IEEE Access 6 (2018), 42740–42749. DOI:
[9]
Mebarka Allaoui, Mohammed Lamine Kherfi, and Abdelhakim Cheriet. 2020. Considerably improving clustering algorithms using UMAP dimensionality reduction technique: A comparative study. International Conference on Image and Signal Processing (ICISPC’20). Springer, 317–325. DOI:
[10]
Abdullah N. Alsaleh, Eric Atwell, and Abdulrahman Altahhan. 2021. Quranic verses semantic relatedness using AraBert. In Proceedings of the 6th Arabic Natural Language Processing Workshop. Leeds, 185–190.
[11]
M. Alshammeri, E. Atwell, and M. A. Alsalka. 2021a. Detecting semantic-based similarity between verses of the Quran with Doc2vec. In Procedia CIRP. Elsevier B.V., 351–358. DOI:
[12]
M. Alshammeri, E. Atwell, and M. A. Alsalka. 2021b. Quranic topic modelling using paragraph vectors. In Advances in Intelligent Systems and Computing. Springer, 218–230. DOI:
[13]
Menwa Alshammeri, Eric Atwell, and Mhd Ammar Alsalka. 2021c. A Siamese transformer-based architecture for detecting semantic similarity in the Quran. International Journal on Islamic Applications in Computer Science and Technology (IJASAT) 9 (2021), 1–11.
[14]
Ahmad Amin, Toqir A. Rana, Natash Ali Mian, Muhammad Waseem Iqbal, Abbas Khalid, Tahir Alyas, and Mohammad Tubishat. 2020. TOP-rank: A novel unsupervised approach for topic prediction using keyphrase extraction for Urdu documents. IEEE Access 8 (2020), 212675–212686. DOI:
[15]
David M. Blei, Andrew Y. Ng, and Michael I. Jordan. 2003. Latent Dirichlet allocation. Journal of Machine Learning Research 3 (Jan. 2003), 993–1022.
[16]
Gerlof Bouma. 2009. Normalized (pointwise) mutual information in collocation extraction. Proceedings of GSCL 30 (2009), 31–40.
[17]
Sophie Burkhardt and Stefan Kramer. 2019. Decoupling sparsity and smoothness in the Dirichlet variational autoencoder topic model. Journal of Machine Learning Research 20, 131 (2019), 1–27. http://jmlr.org/papers/v20/18-569.html
[18]
Ali Daud, Wahab Khan, and Dunren Che. 2017. Urdu language processing: A survey. Artificial Intelligence Review 47 (2017), 279–311. DOI:
[19]
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In North American Chapter of the Association for Computational Linguistics. DOI:
[20]
Adji B. Dieng, Francisco J. R. Ruiz, and David M. Blei. 2020. Topic modeling in embedding spaces. Transactions of the Association for Computational Linguistics 8 (2020), 439–453. DOI:
[21]
R. Egger and J. Yu. 2022. A topic modeling comparison between LDA, NMF, Top2Vec, and BERTopic to demystify Twitter posts. Frontiers in Sociology 7 (May2022), 886498. DOI:
[22]
Lijimol George and P. Sumathy. 2023. An integrated clustering and BERT framework for improved topic modeling. International Journal of Information Technology 15 (2023), 1–9. DOI:
[23]
Maarten Grootendorst. 2022. BERTopic: Neural topic modeling with a class-based TF-IDF procedure. arXiv preprint arXiv:2203.05794 (2022).
[24]
Thomas Hofmann. 1999. Probabilistic latent semantic indexing. In Proceedings of the 22nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’99). 50–57. DOI:
[25]
L. B. Hutama and D. Suhartono. 2022. Indonesian hoax news classification with multilingual transformer model and BERTopic. Informatica (Slovenia) 46, 8 (2022), 81–90. DOI:
[26]
Komal Khalid, Hammad Afzal, Faiza Moqaddas, Naima Iltaf, Ahmed Muqeem Sheri, and Raheel Nawaz. 2017. Extension of semantic based Urdu linguistic resources using natural language processing. In 2017 IEEE 15th International Conference on Dependable, Autonomic and Secure Computing, 15th International Conference on Pervasive Intelligence and Computing, 3rd International Conference on Big Data Intelligence and Computing and Cyber Science and Technology Congress (DASC/PiCom/DataCom/CyberSciTech ’17). IEEE, 1322–1325. DOI:
[27]
Thomas K. Landauer, Peter W. Foltz, and Darrell Laham. 1998. An introduction to latent semantic analysis. Discourse Processes 25, 2–3 (1998), 259–284. DOI:
[28]
Jey Han Lau, David Newman, and Timothy Baldwin. 2014. Machine reading tea leaves: Automatically evaluating topic coherence and topic model quality. In Proceedings of the 14th Conference of the European Chapter of the Association for Computational Linguistics (EACL’14). 530–539. DOI:
[29]
Daniel Lee and H. Sebastian Seung. 2000. Algorithms for non-negative matrix factorization. Advances in Neural Information Processing Systems 13 (2000), 205.
[30]
L. McInnes, J. Healy, and S. Astels. 2017. hdbscan: Hierarchical density based clustering. Journal of Open Source Software 2, 11 (2017), 861. DOI:
[31]
L. McInnes, J. Healy, N. Saul, and L. Großberger. 2018. UMAP: Uniform manifold approximation and projection. Journal of Open Source Software 3, 29 (2018), 1–7. DOI:
[32]
Siraj Munir, Shaukat Wasi, and Syed Imran Jami. 2019. A comparison of topic modelling approaches for Urdu text. Indian Journal of Science and Technology 12 (2019), 45. DOI:
[33]
Mubashar Mustafa, Feng Zeng, Hussain Ghulam, and Wenjia Li. 2021. Discovering coherent topics from Urdu text. In ATAIT. 68–80.
[34]
Mubashar Mustafa, Feng Zeng, Usama Manzoor, and Lin Meng. 2023. Discovering coherent topics from Urdu text: A comparative study of statistical models, clustering techniques and word embedding. In 2023 6th International Conference on Information and Computer Technologies (ICICT ’23). IEEE, 127–131. DOI:
[35]
Feng Nan, Ran Ding, Ramesh Nallapati, and Bing Xiang. 2019. Topic modeling with Wasserstein autoencoders. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 6345–6381. DOI:
[36]
Karl Pearson. 1901. LIII. On lines and planes of closest fit to systems of points in space. London, Edinburgh, and Dublin Philosophical Magazine and Journal of Science 2, 11 (1901), 559–572. DOI:
[37]
M. I. Rahman, N. A. Samsudin, A. Mustapha, and A. Abdullahi. 2018. Comparative analysis for topic classification in Juz Al-Baqarah. Indonesian Journal of Electrical Engineering and Computer Science 12, 1 (Oct.2018), 406–411. DOI:
[38]
Anwar Ur Rehman, Zobia Rehman, Junaid Akram, Waqar Ali, Munam Ali Shah, and Muhammad Salman. 2018. Statistical topic modeling for Urdu text articles. In 2018 24th International Conference on Automation and Computing (ICAC ’18). IEEE, 1–6. DOI:
[39]
N. Reimers and I. Gurevych. 2019. Sentence-BERT: Sentence embeddings using siamese BERT-networks. InProceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP ’19), 3982–3992. DOI:
[40]
Michael Röder, Andreas Both, and Alexander Hinneburg. 2015. Exploring the space of topic coherence measures. In Proceedings of the Eighth ACM International Conference on Web Search and Data Mining (WSDM’15). 399–408. DOI:
[41]
Dwi Rolliawati, Indri Rozas, and Khalid Khalid. 2020. Text mining approach for topic modeling of corpus Al Qur’an in Indonesian translation. In Proceedings of the 2nd International Conference on Quran and Hadith Studies Information Technology and Media in Conjunction with the 1st International Conference on Islam, Science and Technology (ICONQUHAS & ICONIST ’20). DOI:
[42]
Sattar Seifollahi, Massimo Piccardi, and Alireza Jolfaei. 2021. An embedding-based topic model for document classification. Transactions on Asian and Low-resource Language Information Processing 20, 3 (2021), 1–13. DOI:
[43]
Jawad Shafi, Hafiz Rizwan Iqbal, Rao Muhammad Adeel Nawab, and Paul Rayson. 2023. UNLT: Urdu natural language toolkit. Natural Language Engineering 29, 4 (2023), 942–977. DOI:
[44]
Khadija Shakeel, Ghulam Rasool Tahir, Irsha Tehseen, and Mubashir Ali. 2018. A framework of Urdu topic modeling using latent dirichlet allocation (LDA). In 2018 IEEE 8th Annual Computing and Communication Workshop and Conference (CCWC ’18). IEEE, 117–123. DOI:
[45]
Muazzam Ahmed Siddiqui, Syed Muhammad Faraz, and Sohail Abdul Sattar. 2013. Discovering the thematic structure of the Quran using probabilistic topic model. 2013 Taibah University International Conference on Advances in Information Technology for the Holy Quran and Its Sciences (NOORIC’13). IEEE, 234–239. DOI:
[46]
M. A. Siddiqui, S. M. Faraz, and S. A. Sattar. 2015. Discovering the thematic structure of the Quran using probabilistic topic model. In Proceedings of the 2013 Taibah University International Conference on Advances in Information Technology for the Holy Quran and Its Sciences (NOORIC ’13). Institute of Electrical and Electronics Engineers Inc., 234–239. DOI:
[47]
R. Silveira, C. G. O. Fernandes, J. A. Monteiro Neto, V. Furtado, and J. E. P. Filho. 2021. Topic modelling of legal documents via LEGAL-BERT. In Proceedings of the 1st International Workshop RELATED—Relations in the Legal Domain 2021. DOI:
[48]
Silvia Terragni, Elisabetta Fersini, and Enza Messina. 2021. Word embedding-based topic similarity measures. International Conference on Applications of Natural Language to Information Systems (NLDB’21). 33–45. DOI:
[49]
Laurens Van der Maaten and Geoffrey Hinton. 2008. Visualizing data using t-SNE. Journal of Machine Learning Research 9, 86 (2008), 2579–2605. http://jmlr.org/papers/v9/vandermaaten08a.html
[50]
Ike Vayansky and Sathish A. P. Kumar. 2020. A review of topic modeling methods. Information Systems 94 (2020), 101582. DOI:
[51]
William Webber, Alistair Moffat, and Justin Zobel. 2010. A similarity measure for indefinite rankings. ACM Transactions on Information Systems (TOIS) 28, 4 (2010), 1–38. DOI:
[52]
M.-H. Weng, S. Wu, and M. Dyer. 2022. Identification and visualization of key topics in scientific publications with transformer-based language models and document clustering methods. Applied Sciences 12, 21 (Nov.2022), 11220. DOI:
[53]
Yi Yang, Kunpeng Zhang, and Yangyang Fan. 2023. sdtm: A supervised Bayesian deep topic model for text analytics. Information Systems Research 34, 1 (2023), 137–156. DOI:
[54]
Zoya, S. Latif, F. Shafait, and R. Latif. 2021. Analyzing LDA and NMF topic models for Urdu Tweets via automatic labeling. IEEE Access 9 (2021), 127531–127547. DOI:

Cited By

View all
  • (2024)Topic Modeling of Quranic Verses using Latent Dirichlet Allocation with English LanguageVFAST Transactions on Software Engineering10.21015/vtse.v12i4.194612:4(239-251)Online publication date: 31-Dec-2024

Index Terms

  1. Transformer-Based Topic Modeling for Urdu Translations of the Holy Quran

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Transactions on Asian and Low-Resource Language Information Processing
    ACM Transactions on Asian and Low-Resource Language Information Processing  Volume 23, Issue 10
    October 2024
    189 pages
    EISSN:2375-4702
    DOI:10.1145/3613658
    • Editor:
    • Imed Zitouni
    Issue’s Table of Contents

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 24 October 2024
    Online AM: 09 September 2024
    Accepted: 29 August 2024
    Revised: 20 August 2024
    Received: 09 October 2023
    Published in TALLIP Volume 23, Issue 10

    Check for updates

    Author Tags

    1. Topic modeling
    2. Quranic topic modeling
    3. Urdu Quranic NLP
    4. Urdu translations of the Holy Quran

    Qualifiers

    • Research-article

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)702
    • Downloads (Last 6 weeks)169
    Reflects downloads up to 02 Mar 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Topic Modeling of Quranic Verses using Latent Dirichlet Allocation with English LanguageVFAST Transactions on Software Engineering10.21015/vtse.v12i4.194612:4(239-251)Online publication date: 31-Dec-2024

    View Options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Login options

    Full Access

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media