Vol. 6, No. 12, 2015 Extracting Topics From The Holy Quran .

2y ago
19 Views
2 Downloads
482.16 KB
7 Pages
Last View : 1m ago
Last Download : 3m ago
Upload by : Roy Essex
Transcription

(IJACSA) International Journal of Advanced Computer Science and Applications,Vol. 6, No. 12, 2015Extracting Topics from the Holy QuranUsing Generative ModelsMohammad AlhawaratDepartment of Computer Science,Prince Sattam Bin Abdulaziz University,Alkharj, Saudi ArabiAbstract—The holy Quran is one of the Holy Books of God.It is considered one of the main references for an estimated 1.6billion of Muslims around the world. The Holy Quran language isArabic. Specialized as well as non-specialized people in religionneed to search and lookup certain information from the HolyQuran. Most research projects concentrate on the translation ofthe holy Quran in different languages. Nevertheless, few researchprojects pay attention to original text of the holy Quran in Arabiclanguage. Keyword search is one of the Information Retrieval (IR)methods but will retrieve what is called exact search. Semanticsearch aims at finding deeper meanings of a text, and it is ahot field of study in Natural Language Processing (NLP). Inthis paper topic modeling techniques are explored to setup aframework for semantic search in the holy Quran. As the HolyQuran is the word of God, its meanings are unlimited. In thispaper the words of chapter Joseph (Peace Be Upon Him (PBUH))from the Holy Quran is analyzed based on topic modelingtechniques as a case study. Latent Dirichlet Allocation (LDA)topic modeling technique has been applied in this paper intotwo structures (Hizb Quarters and verses) of Joseph chapter as:words, roots and stems. The log-Likelihood has been calculatedfor the two structures of the chapter. Results show that the beststructure to use is verses, which gives the least energy for data.Some of the results of the attained topics are shown. These resultssuggest that topic modeling techniques failed to capture in anaccurate manner the coherent topics of the chapter.Keywords—Statistical models; Latent Dirichlet Analysis (LDA);Holy Quran; Unsupervised LearningI.I NTRODUCTIONThe holy Quran is considered an essential reference forMuslims where they read in a regular basis. They usually needto search it and retrieve relevant information based on morethan just simple keyword search techniques.Dealing with the holy Quran is different from dealingwith regular Arabic corpora that is usually extracted fromNewspapers and speeches, and hence is the word of human.The holy Quran is the word of God and the meanings ofits words are unlimited. The sequence of text is differentfrom human words. For example, one topic could repeat indifferent places in the holy Quran with different details andsometimes in different contexts. Also, one chapter usuallyhas many topics. While one topic might be started in oneverse, another topic may starts immediately in the next verse.Also, one verse may have different topics. Moreover, there aredifferent authentic interpretations for the verses of the holyQuran; therefore it is very hard for a computer to managethem in the way scholars do especially in situations wheremeanings are seem opposite to each other. Finally, there ismuch relevant information that is found in prophet Mohammad(PBUH) sayings (Hadith) that interpret many verses of the holyQuran. For all of these reasons, it sometimes hard to resolvea disambiguation if a word has many synonyms and differentsenses.Research in Arabic NLP still young and have many challenges [1]. This is because that Arabic language is differentfrom many other natural languages [2], [3]. Words in Arabic language have many derivations and have also complexDiglossia (modern and colloquial) [4]. Also, Arabic lettersappear in different shapes according to their position in theword. Another characteristic of the Arabic language is thediacritic. Some of these diacritical marks are usually notwritten, but is understood by Arabic readers. Therefore, twoexact written words without diacritical marks have totallydifferent meanings. All of these and other characteristics ofthe Arabic language should be taken in consideration whenprocessing Arabic text.The holy Quran can be considered as a ”Golden Text” touse in Text mining and NLP fields. This might be true fordifferent reasons: it’s the word of God, it’s limited in terms oftext size and it has many translations and many interpretations.These all together encourage building a semantic comprehensive source for the holy Quran that will allow advancedsemantic search and knowledge extraction.Searching in the holy Quran is an essential task for Muslims as well as non-Muslims who study it. Many applicationshave been built to allow search in the holy Quran. Most ofthese search engines allow simple search techniques wheresome of them are mentioned in [5]. However, few researchprojects are concerned with advanced search in the holy Quranusing some NLP techniques such as the papers presentedin The holy Quran and new technology workshop that heldby King Fahad Complex for printing the holy Quran in AlMadinah Al-Munawwarah, Saudi Arabia in 2008. The workshop participants discussed different issues related to the holyQuran including searching techniques. Also more papers arepresented in another event in Taibah University InternationalConference on Advances in Information Technology for theHoly Quran and Its Sciences that held in Al-Madinah Al288 P a g ewww.ijacsa.thesai.org

(IJACSA) International Journal of Advanced Computer Science and Applications,Vol. 6, No. 12, 2015Munawwarah, Saudi Arabia in 2013. The presented papers arerelated to a wide range of topics concerning the holy Quranincluding natural processing issues, security, education andmany more.Similar work has been carried out to extract verses fromthe holy Quran using an expert system that use Web OntologyLanguage (OWL) [21]. Again the work use English translationof the holy Quran and not Arabic language.There are different approaches to model and cluster topicsin text documents such as LDA, Latent Semantic Analysis(LSA) and traditional clustering techniques such as K-means.In this research LDA is used for several reasons includingaccuracy, scalability and comprehension [6], [7], [8].Another work explored the structure of a simple domainQuran ontology for birds and animals that are mentioned inthe holy Quran [22]. The authors propose a framework forsemantic search in the holy Quran using their domain ontologyand they have evaluated it using SPARQL query language. Thiswork uses English translation of the holy Quran.LDA has been developed to extract topics from text usingstatistical methods [9], [10], [11], [12], [13]. LDA is one of thetechniques that belongs to a large family called probabilisticmodeling. The basic intuition behind LDA is that a textdocument has multiple topics where each topic is defined as adistribution over a set of words. There are many flavors of theLDA model; a thorough review of the LDA topic modelingtechniques can be found in [14]. Topic modeling has beenapplied to many field of study such as Information RetrievalIR, geographical IR, computational linguistics and NLP [15],[16], [17], [18], [19], [20].This paper aims to build up the first stage in a frameworkthat will allow possible semantic search in the holy Quran.This is done by applying LDA topic modeling to chapterJoseph of the holy Quran as a case study. This chapter hasbeen chosen because it includes relative topics regarding storyof the prophet Joseph (PBUH). The LDA topic modeling hasbeen applied to words, roots and stems of that chapter. Nextstages might include: studying the topics of the whole holyQuran, linking the text of the holy Quran to both authenticateinterpretation of the holy Quran and the related Sayings of theprophet Mohammad (PBUH). These might be achieved usingmachine learning, text mining as well as NLP techniques.It should be stated explicitly here that this research is not areligious study; rather it is a statistical study that might resultin information that would guide specialized religious peopleto understand more about the word of God.The paper is organized as follows: in section II relatedwork is presented, in section III topic modeling is introduced,in section IV the methodology as well as preparation of theData Set is explained, in section V experimental setups areexplained, section VI includes discussion of the results attainedin the paper and finally section VII contains conclusion.II.R ELATED W ORKShoaib et. al. [5] have proposed a simple WordNet forthe English translation of the second chapter of the holyQuran (Al-Baqrah). They have created topic-synonym relationsbetween the words in that chapter with different priorities.They have defined different relations that are used in traditional WordNet such as: synonymy, polysemy, hyperonymy,hyponymy, holonymy and meronymy. Then they developeda semantic search algorithm that will fetch all verses thatcontains the query word and its synonyms with high priority.It is not clear how the authors build their simple WordNet.In similar studies, usually authentic religion references shouldbe used such as interpretation of the holy Quran or meaningsof the words of the holy Quran. However, the results showthat the developed semantic search outperform simple searchalgorithms.Data mining techniques such as SVM and nave Bayesianclassifiers are used cluster chapters of the holy Quran based onMajor Phases of Prophet Mohammads (PBUH) Messengership[23]. This work classifies chapters of the holy Quran ratherthan verses or words of the holy Quran.LDA topic modeling technique has been used to extracttopics from an Arabic corpora composed of Newspapers [24].The authors have developed a preprocessing lemma-basedstemming algorithm and then applied the LDA technique onArabic processed text.In [25] author has used clustering techniques in machinelearning to extract topics of the holy Quran. The extraction oftopics was based on a corpus that is composed of the versesof the holy Quran using nonnegative matrix factorization. Theauthor used Buckwalter code for Arabic letters [3]. Topicsare visualized and related verses for each topic are shownfor selected topics based on the topic main keywords. Oneof the shortcoming of his work is that verses are dealt withseparately as each as a document. The author claims that hehas extracted and identified the underlying topics of the holyQuran. However, this claim is far from reality as no one couldidentify the underlying topics of the holy Quran even wellknown scholars of Quran studies. Also, the it is totally unclearhow he has linked the keywords of each topic with the relatedverses that correspond to topic keywords. Nevertheless, thefindings are promising and might help in revealing deepermeanings of the holy Quran by specialized people in Quranicstudies.LDA technique has been compared LDA with K-meansclustering technique [8]. The authors have applied both LDAand K-means technique on a set of Arabic documents fromOSAC (Open Source Arabic Corpora). The results show thatLDA outperforms K-means in most instances.III.T OPIC M ODELINGTopic modeling is a hot field of study in both machinelearning and NLP. Topic models are generative models thatare based on probability distributions of multiple topics in adocument over a set of words. Such models basically dependon term-frequencies in a document. One of these models isLDA. As mentioned previously, LDA is better than othermodels such as LSA for several reasons[6], [7], [8]. LDAoutperforms LSA in many applications including semanticrepresentation [12] and have been used in different fields in thelast decade or so including NLP [15], [16], [17]. It is used byresearchers to extract important and hot topics; usually fromlarge corpora.289 P a g ewww.ijacsa.thesai.org

(IJACSA) International Journal of Advanced Computer Science and Applications,Vol. 6, No. 12, 2015The basic intuition behind LDA is that a set of wordsof documents are randomly pre-assigned with probability distributions that would represent multiple-topic latent structureon those documents. After that, latent structure of the topicsof documents is inferred statistically in a reverse-engineeringmanner.the text of the holy Quran is the word of God, there is nomargin for errors in the process of extracting both roots andstems. Therefore, the roots and stems have been constructedmanually; based on two web sites [28], [29] and verified bythe authors according to their experience in Arabic languageand as native speakers.Initially, a number of topics T should be specified. Then,a term distribution ϕ over a parameter β is chosen for eachtopic. After that, ratios θ of topic distribution for document dare specified. Then, a topic zi is chosen and after that a wordis chosen conditioned on that topic over a parameter α. Bothϕ and θ are Dirichlet distributions.These data sets will be used as the input for the implantation of the LDA to reveal the main topics for the text ofthe chapter of Joseph (PBUH). Different experimental setupsare prepared to compute the topic models for the text of thatchapter based on the aforementioned structures.The probability of the ith word in a specific document isgiven by:P (wi ) T P (wi zi j)P (zi j)(1)j 1where zi represents a latent variable that designates thetopic for the drawn ith word. P (wi zi j) represents theprobability of the word wi under topic j. P (zi j) representsthe probability of a word from topic j of a document.Note that P (w z) can be represented by a multinomialdistributions ϕ over a term distribution such that P (w z (j)j) ϕw and P (z) can be represented by a multinomialdistributions θ over a topic distribution over D documents such(d)that P (z j) θj .Then an estimation method is used to infer the latent structure of the topics of documents. Different estimation methodscan be used in this context including: Variational ExpectationMaximization (VEM) method and Gibbs sampling. For moreinformation about details of these methods please refer to [10],[11], [13], [26].V.Both packages tm and topicmodels of R are used inexperiments (a practical guide for topicmodels can befound in [30]). First, the tm package will be used for textpreparation and processing as building the corpus, removingstop words and building the Document Term matrix (DTM).Second, the topicmodels package will be used to buildand fit LDA model for all structures of the text with the threeshapes of word.The text with two structures has been processed where thestop words are removed. Then, three DTMs have been builtfor text as: words, roots and stems. The content of the DTMis basically calculated using Term Frequencies (TF) measure.After that, the tf-idf measure has been applied on each DTM toremove frequent terms that appears on most documents, andhence are not recognized as important terms. This has beendone by calculating the median and choosing high-frequentterms with frequency more than the calculated median.TABLE II: The number of topics along with the log-Likelihoodfor the fitted topic models for the Joseph chapter estimated byGibbs sampling with 10-fold cross-validationBesides LDA, Correlated Topic Model (CTM) can beused to extract correlated topics from documents. CTM isan extension of LDA. LDA usually uses Gibbs sampling formodel estimation.IV.E XPERIMENTSHizb quartersTopics No./Log-LikelihoodVersesTopics 169Stem27/-86019/-172DATA S ET P REPARATION AND M ETHODOLOGYThe text of chapter Joseph in the format of CP1256 hasbeen taken from [27] in the shape of two structures: Hizbquarters and verses, all without diacritic. The frequency detailsof these selected structures are shown in table I. For moreinformation about the text structure of the holy Quran pleaserefer to [27].TABLE I: The number of documents for the Joseph chapterbased on different structures and for words, roots and stemsafter applying tf-idf measure on DTMsNo. of Hizb quarters/termsNo. of Verses/terms111/721Original No. (TF)6/721With TF-IDF (words)6/29989/323With TF-IDF (roots)6/327108/163With TF-IDF (stems)6/398103/193These two structures will be used in the topic modelingprocess in three shapes: words, roots and stems. BecauseAfter that, different experimental setups are prepared tofind the main topics in the chapter of Joseph (PBUH). Theseare found first using TF measure and then using different estimation techniques for LDA besides Correlated Topics Model(CTM)-where CTM can use VEM only: VEM. VEM with fixed α. GibbsThen, a validation technique that is based on the logLikelihood of the data set is calculated. This is performed tofind the best number of topics for each structure of that chapter.The best number of the topics is calculated using 10-foldcross-validation technique for the two structures with the threeterm shapes, results of log-likelihood and number of topics areshown in table II. Then the topics are recorded for all casesusing the best topic numbers that are calculated according tothe aforementioned technique. In some cases different topics290 P a g ewww.ijacsa.thesai.org

(IJACSA) International Journal of Advanced Computer Science and Applications,Vol. 6, No. 12, 2015number is chosen because the energy-based topics number islarge. The main parameters are set as suggested by [11] whereα 50/k (where k is the number of topics) and β 0.1. Inmany of the experiment setups, the seed parameter of the LDAand CTM models are set to the number of terms according totable I.Samples of the results of the topics are shown in figures 1 13 for the two structures with three shapes of the terms: words,roots and stems. Figures 1 - 9 represent the Verses structurewhere figures 1 - 3 are for words, figures 4 - 6 are for stemsand figures 1 - 3 are for roots. Figures 10 - 13 represent theHizb Quarters structure for words, roots and stems.Fig. 3: Sample of topics for words based on Verses where TFis used (Topics Number is 17)Fig. 1: Sample of topics for words based on Verses whereGibbs sampling is used (Topics Number is 17)Fig. 4: Sample of topics for stems based on Verses whereVEM is used (Topics Number is 19)Fig. 2: Sample of topics for words based on Verses whereCTM is used (Topics Number is 17)Fig. 5: Sample of topics for stems based on Verses whereVEM with fixed α is used (Topics Number is 19)291 P a g ewww.ijacsa.thesai.org

(IJACSA) International Journal of Advanced Computer Science and Applications,Vol. 6, No. 12, 2015Fig. 6: Sample of topics for stems based on Verses where TFis used (Topics Number is 19)Fig. 9: Sample of topics for roots based on Verses where TFis used (Topics Number is 15)VI.R ESULTS AND D ISCUSSIONThe number of documents and terms of the chapter ofJoseph (PBUH) is shown in table I. Both TF and TF-IDFmeasures are used and then the number of documents andterms are recorded for words, roots and stems. The resultsof applying LDA model to the text of the chapter with Gibbssampling technique is shown in table II. Note that the term withlow energy are roots and stems compared with high energy forwords.Many experiment setups have been carried out with different parameter settings apart from the aforementioned setupsin section V. Sample of the results are shown in figures 1 13. All the results of all the experiments show that most of theresulted topics are a mix of more than one topic. However, veryfew topics form one coherent topic such as topic number threeof figure 3 and topic number three and fourteen of figure 5.Fig. 7: Sample of topics for roots based on Verses where Gibbssampling is used (Topics Number is 5)Some topics include a mix of two to may be five topics.In some cases all of the terms of the topic are coherent exceptone or two words such as topic number 12 of figure 10.Regarding the shapes of the word; on one hand the rootsare considered problematic as there are many shared wordsbetween topics such as the topics that appear in figure 12.One of the reasons behind this is that there are some differentwords in meaning but their root in Arabic language is the same.On the other hand, both words and stems show better resultsas it appear in most of the figures. For words it is obvious thateach word has usually its own semantic in one context. Forstems, although there is more than a word with the same stembut they have the same semantic in similar contexts.The estimation methods that are used in this study showdifferent ”percentage of successful” with different shapes ofwords. For example, TF measure gives better results than TFIDF measure in certain cases. On another occasion, CTM givesbetter results. The same is true for VEM, VEM with fixed αand Gibbs sampling.Fig. 8: Sample of topics for roots based on Verses where VEMis used (Topics Number is 15)Also, it is important to mention that all of the numerical results including best number of topics as well as logLikelihood of the data are based on the seed parameter292 P a g ewww.ijacsa.thesai.org

(IJACSA) International Journal of Advanced Computer Science and Applications,Vol. 6, No. 12, 2015for LDA and CTM models. However, many experiments areexecuted with different values for seed parameter withoutaffecting the quality of the resulted topics.In other set of experiments, the parameter alpha is set tosmaller numbers than that suggested by [11] where α 50/k(k is the number of topics). When α is set to 1/k, the resultsshow topics with slightly better quality.Although the topic modeling techniques used in this studyfailed to extract coherent topics, still the results are promisingas some topics are coherent even that they are very few.Fig. 12: Sample of topics for roots based on Hizb Quarterswhere VEM is used (Topics Number is 44)Fig. 10: Sample of topics for words based on Hizb Quarterswhere VEM is used (Topics Number is 17)Fig. 13: Sample of topics for roots based on Hizb Quarterswhere TF is used (Topics Number is 15)VII.Fig. 11: Sample of topics for stems based on Hizb Quarterswhere Gibbs sampling is used (Topics Number is 26)C ONCLUSIONThe topicmodels R package has been used to analysethe underlying topics of the chapter Joseph (PBUH). Firstthe best number of topics for the two structures have beencalculated for the three shapes of words and the results areshown in table I. After that, several experiment setups areexecuted for both of the document structures with three termshapes: word, root and stem. Then, results are recorded andsamples of the result are shown in figures 1 - 13. The resultsare evaluated based on understanding of the meanings andinterpretation of the chapter of Joseph (PBUH). The resultssuggest that verses structure is better than Hizb quarters onein forming more coherent topics. Most of the resulted topicsinclude a mix of more than one topic out of the main topics ofthe chapter of Joseph (PBUH). However, few of the resultedtopics contain one coherent topic.Semantic search in the holy Quran can be supportedby finding accurate coherent topics which helps in finding293 P a g ewww.ijacsa.thesai.org

(IJACSA) International Journal of Advanced Computer Science and Applications,Vol. 6, No. 12, 2015contextual terms related to the user search terms. The holyQuran contains hundreds of topics if not thousands. Whileone verse may contain multiple topics, another set of versesmay comprise one topic. Also, one topic may repeat in severalcontexts and in more than one chapter. If the results areenhanced by combining LDA with another technique thenthey can be then used together to search for relevant wordsaccording to the distribution of topics over words.The results of this study strongly suggests that whilestatistical methods succeeded in extracting important topicsfrom text corpora of humans -as many studies show, it failedto achieve the same results with the word of God. This isobvious because the words of God are unlimited in meaningand are one of the attributes/characters of God.Future work may include exploring more statistical methods and/or combining the methods used in this study with otherdata mining techniques. Also, if the text of the holy Quranwould be linked to one of its authentic interpretations, thentopic modeling might find coherent topics because interpretations are the word of human.R 3][14]A. Farghaly and K. Shaalan, “Arabic natural language processing:Challenges and solutions,” vol. 8, no. 4, pp. 14:1–14:22, Dec. 2009.[Online]. Available: http://doi.acm.org/10.1145/1644879.1644881M. Saad and W. Ashour, “Arabic morphological tools for text mining,”in 6th International Symposium on Electrical and Electronics Engineering and Computer Science, European University of Lefke, Cyprus, 2010,2010, p. 112117.N. Y. Habash, Introduction to Arabic Natural Language Processing,G. Hirst, Ed. Morgan and Claypool Publishers, 2010.M. DIAB and N. HABASH, “Arabic dialect tutorial,” in In Proceedingsof the Human Language Technology Conference of the North AmericanChapter of the Association for Computational Linguistics (NAACL07),2007, pp. 29–34.M. Shoaib, M. Nadeem Yasin, U. Hikmat, M. Saeed, and M. Khiyal,“Relational wordnet model for semantic search in holy quran,” inInternational Conference on Emerging Technologies, 2009. ICET 2009.,Oct 2009, pp. 29–34.I. Biro, “Document classification with latent dirichlet allocation,” Ph.D.dissertation, Eötvös Loránd University, 2009.P. Crossno, A. Wilson, T. Shead, and D. Dunlavy, “Topicview: Visuallycomparing topic models of text collections,” in Tools with ArtificialIntelligence (ICTAI), 2011 23rd IEEE International Conference on, Nov2011, pp. 936–943.A. Kelaiaia and H. Merouani, “Clustering with probabilistic topicmodels on arabic texts,” in Modeling Approaches and Algorithms forAdvanced Computer Applications, ser. Studies in Computational Intelligence, A. Amine, A. M. Otmane, and L. Bellatreche, Eds. SpringerInternational Publishing, 2013, vol. 488, pp. 65–74.D. Blei and J. Lafferty, “Topic models,” in Text Mining: Theory andApplications, Srivastava and M. Sahami, Eds. Taylor and Francis,2006.D. M. Blei, A. Y. Ng, and M. I. Jordan, “Latent dirichlet allocation,”J. Mach. Learn. Res., vol. 3, pp. 993–1022, Mar. 2003.T. L. Griffiths and M. Steyvers, “Finding scientific topics,” Proceedingsof the National Academy of Sciences of the United States of America,vol. 101, no. Suppl 1, pp. 5228–5235, Apr. 2004.T. L. Griffiths, J. B. Tenenbaum, and M. Steyvers, “Topics in semanticrepresentation,” Psychological Review, vol. 114, p. 2007, 2007.M. Steyvers and T. Griffiths, “Probabilistic topic models,” in LatentSemantic Analysis: A Road to Meaning., T. Landauer, D. Mcnamara,S. Dennis, and W. Kintsch, Eds. Laurence Erlbaum, 2006.D. M. Blei, “Probabilistic topic models,” Communications of the ACM,vol. 55, no. 4, pp. 77–84, 2012.[15] G. K. Gerber, R. D. Dowell, T. Jaakkola, and D. K. Gifford, “Automateddiscovery of functional generality of human gene expression programs.”PLoS Computational Biology, vol. 3, no. 8, 2007.[16] J. Boyd-Graber, D. M. Blei, and X. Zhu, “A topic model for wordsense disambiguation,” in Empirical Methods in Natural LanguageProcessing, 2007.[17] S. Gerrish and D. M. Blei, “Predicting legislative roll calls from text.” inICML, L. Getoor and T. Scheffer, Eds. Omnipress, 2011, pp. 489–496.[18] X. Wei and W. B. Croft, “Lda-based document models for ad-hocretrieval,” in Proceedings of the 29th Annual International ACM SIGIRConference on Research and Development in Information Retrieval.ACM, 2006, pp. 178–185.[19] Z. Li, C. Wang, X. Xie, X. Wang, and W.-Y. Ma, “Exploring lda-baseddocument model for geographic information retrieval,” in Advances inMultilingual and Multimodal Information Retrieval, ser. Lecture Notesin Computer Science, C. Peters, V. Jijkoun, T. Mandl, H. Mller, D. Oard,A. Peas, V. Petras, and D. Santos, Eds. Springer Berlin Heidelberg,2008, vol. 5152, pp. 842–849.[20] D. Hall, D. Jurafsky, and C. D. Manning, “Studying the historyof ideas using topic models,” in Proceedings of the Conference onEmpirical Methods in Natural Language Processing, ser. EMNLP ’08.Association for Computational Linguistics, 2008, pp. 363–371.[21] A. A. Aliyu Rufai Yauri, Rabiah Abdul Kadir and M. A. A. Murad,“Quranic verse extraction base on concepts using owl-dl ontology.”vol. 6, no. 23, pp. 4492–4498, 2013.[22] M. S. Hikmat Ullah Khan, Syed Muhammad Saqlain and M. Sher,“Ontology-based semantic search in holy quran,” vol. 2, no. 6, pp. 562–566, 2013.[23] M. Nassourou, “Using machine learning algorithms for categorizingquranic chapters by major phases of prophet mohammads messengership,” vol. 2, no. 11, pp. 863–871, 2012.[24] A. Brahmi, A. Ech-Cherif, and A. Benyettou, “An arabic lemma-basedstemmer for latent topic modeling,” Int. Arab J. Inf. Technol., vol. 10,no. 2, pp. 160–168, 2013.[25] M. H. Panju, “Statistical extraction and visualization of topics in thequr’an corpus,” Master’s thesis, University of Waterloo, 2014.[26] W. M. Darling, “A theoretical and practical implementation tutorial ontopic modeling and gibbs sampling,” in Proceedings of the 49th AnnualMeeting of the Association for Computational Linguistics: HumanLanguage Technologies, 2011, pp. 642–647.[27] M. Alhawarat, M. Hegazi, and A. Hilal, “Processing the text of theholy quran: a text mining study,” International Journal of AdvancedComputer Science and Applications(IJACSA), vol. 6, no. 2, pp. 262–267, February 2015.[28] Mushafqatar.com. (2015) Mushaf qatar. [Online]. up gather[29] //www.almaany.com/quran-b/[30] B. Grn, J. Kepler, U. Linz, K. Hornik, and W. W. Wien, “topicmodels:An r package for fitting topic models,” Journal of Statistical Software,vol. 3, no. 8, 2011.294 P a g ewww.ijacsa.thesai.org

The Holy Quran language is Arabic. Specialized as well as non-specialized people in religion need to search and lookup certain information from the Holy Quran. Most research projects concentrate on the translation of the holy Quran in different languages. Nevertheless, few research projects pay attention to original text of t

Related Documents:

Menschen Pagina 20 Schritte international Neu Pagina 22 Motive Pagina 24 Akademie Deutsch Pagina 25 Starten wir! Pagina 26 Themen aktuell Pagina 28 em neu Pagina 29 Sicher! Pagina 30 Vol A1 1 Vol A1 Vol 1 Vol 1 2 Vol unico Vol 1 Volume 1 Volume 1 Vol 1 Vol 1 1 Vol A1 2 Vol 2 Vol 1 2 Vol A2 1 Vol A2 Vol 3 Vol

Akenson, Donald Harman Vol 8: 10 Alan, Radous, at Agincourt Vol 12: 1 Albert, King Vol 7: 45, 47 Albert, Prince Vol 12: 17; Vol 14: 1 Alden, John Vol 5: 34; Vol 9: 18 Alexander III Vol 13: 24 Aleyn, John, at Agincourt Vol 12: 1 Allen, Pat Vol 10: 44 Alling Vol 4: 26 Amore, Shirley Vol 12: 3 Anderson, Robert Vol 10: 46 Anderson, Virginia DeJohn .

2015 2015 2015 2015 2015 2015 2015 2015 2015 2015 2015 2015 2015 2015 2015 2015 . Removal handle Sound output / wax protection system. 11 Virto V-10 Custom made shell Battery door Volume control (optional) Push button Removal handle . Before using

Accreditation Programme for Nursing and Midwifery . Date of submission of report to Bangladesh Nursing and Midwifery Council_ 2) The Review Team During the site visit, the review team members validate the self-assessment for each of the criteria. . as per DGNM guideline. Yes ⃝No

Canadian Journal of Mathematics, Vol.2 (1950) to Vcl.19 Canadian J. (1967) (Canada) Makh. Comptes Rendus, Des Seances de l'Acaddmie des Sciences. Comptes Paris, Vol.230 (1950) to Vol.265 (1967) (France) Rendus Crod Science, Vol.1 (1961) to Vol.7 (1967) (U.S.) Crop Sci. Current Science, Vol.19 (1950) to Vol.36 (1967) (India) Current Scd. Der .

Vino Nobile Di Montepulciano Riserva Primitivo Di Manduria I.G.T Nero D’Avola I.G.T., Sicilia Salice Salentino Riserva Peppoli Antinori 2013, Chianti classico 13,5 % Vol 14 % Vol 13,5 % Vol 13,5 % V 13,5 % Vol 14 % Vol 13 % Vol Tignanello 201313 % Vol 29 34 38 26,5 29 39 39 235 24. 28. 30

communication briefings vol 22:4 feb 2003 - 30:7 may 2011 comparative literature vol 13:1 winter 1961 - 17:4 fall 1965 computers in education journal vol 6:1 jan/feb/mar 2015 - 8:2 apr/may/june 2017 concordia theological monthly vol 8:1 jan 1937 - 19:12 dec 1948 confessional presbyterian vol 15 2019 - consumer reports latest 1 year current issues

Alter Metal Recycling . 13 . 9/21/2015 156.73 9/24/2015 66.85 9/27/2015 22.24 9/30/2015 35.48 10/3/2015 31.36 10/6/2015 62.97 10/9/2015 36.17 10/12/2015 80.48 10/15/2015 84.99 10/18/2015 90.93 10/21/2015 82.