Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
An Integrated Approach to Automatic Synonym Detection in Turkish Corpus Tuğba Yıldız1 , Savaş Yıldırım1 , and Banu Diri2 1 Department of Computer Engineering, Istanbul Bilgi University, Eski Silahtarağa Elektrik Santrali, Kazım Karabekir Cad. No: 2/13, 34060 Eyüp, Istanbul, Turkey {tdalyan,savasy}@bilgi.edu.tr 2 Department of Computer Engineering, Yildiz Technical University, Davutpasa, 34349 Istanbul, Turkey banu@ce.yildiz.edu.tr Abstract. In this study, we designed a model to determine synonymy. Our main assumption is that synonym pairs show similar semantic and dependency relation by the definition. They share same meronym/holonym and hypernym/hyponym relations. Contrary to synonymy, hypernymy and meronymy relations can probably be acquired by applying lexico-syntactic patterns to a big corpus. Such acquisition might be utilized and ease detection of synonymy. Likewise, we utilized some particular dependency relations such as object/subject of a verb, etc. Machine learning algorithms were applied on all these acquired features. The first aim is to find out which dependency and semantic features are the most informative and contribute most to the model. Performance of each feature is individually evaluated with cross validation. The model that combines all features shows promising results and successfully detects synonymy relation. The main contribution of the study is to integrate both semantic and dependency relation within distributional aspect. Second contribution is considered as being first major attempt for Turkish synonym identification based on corpus-driven approach. Keywords: Synonym, relations. 1 near-synonym, pattern-based, dependency Introduction As one of the most well-known semantic relations, synonymy has been subject to numerous studies. By the definition, synonyms are words with identical or similar meanings. The discovery of synonym relations may help to address various Natural Language Processing (NLP) applications, such as information retrieval and question answering [1–3], automatic thesaurus construction [4,5], automatic text summarization [6], language generation [7], English lexical substitution task [8], lexical entailment acquisition [9]. A. Przepiórkowski and M. Ogrodniczuk (Eds.): PolTAL 2014, LNAI 8686, pp. 116–127, 2014. c Springer International Publishing Switzerland 2014  An Integrated Approach to Automatic Synonym Detection 117 Various methods have been proposed for automatic synonym acquisition. Recent studies were generally based on distributional similarity and pattern-based approach. General idea behind distributional similarity is to capture the semantically related words. Distributional similarity of words sharing a large number of contexts could be informative [10]. Pattern-based approach is the most precise acquisition methodology earlier applied by Hearst [11] and relies on lexicosyntactic patterns (LSPs). On the other hand, these methodologies themselves can be ambiguous and insufficient. Distributional similarity approach can cover other semantically related words and might not distinguish between synonyms and other relations. For example, list of top-10 distributionally similar words for orange is: yellow, lemon, peach, pink, lime, purple, tomato, onion, mango, lavender [12]. In addition, the pattern-based approach tends to capture hyponymy and meronymy relations as well, whereas it is apparently incompatible for synonyms detection. Thus, pattern-based approach or external features such as grammatical relations can be integrated into distributional similarity approach for identifying synonyms by narrowing distributional context. Although some studies have showed that classical distributional methods always have a higher recall than pattern-based techniques in this area [13], integrating two or more approaches were reported that system performance was improved [9, 13–15]. In this study, overall objective is to determine synonym nouns in a Turkish Corpus by relying on distributional similarity that is based on syntactic features (obtained by dependency relations) and semantic features obtained by syntactic patterns and LSPs respectively. The features of the proposed model consist of co-occurrence statistics, four semantic relations and ten syntactic dependency relations where a pair of words are represented with fifteen different features and a target class (SYN/NONSYN). One of the main contributions of the study is that the system first obtains acquirable semantic relations such as hypernymy, meronymy from corpus by LSPs to extract subtle relations such as synonymy. The second contribution of the study is considered to be the first major attempt for Turkish synonym identification based on corpus-driven approach. 2 Related Works A variety of methods have been proposed to automatically or semi-automatically detect synonyms from text source, dictionaries, wikipedia, search engines. Among them, the most popular methods are based on distributional hypothesis [10] which states that semantically similar words share similar contexts. The process of this approach was as follows: co-occurrence, syntactic information, grammatical relations of the words surrounding the target word are extracted as a first step. Afterwards target word is represented as a vector with these contextual features. At the second step, the semantic similarity of two terms is evaluated by applying a similarity measure between their vectors. The words can be ranked by their both semantic and syntactic similarity. Finally, top candidates are selected as the most similar words from ranked list. 118 T. Yıldız, S. Yıldırım, and B. Diri There have been various studies [4,16,17] which used distributional similarity to the automatic extraction of semantically related words from large corpora. Distributional approaches have been applied into monolingual [4,18,19], monolingual parallel [20, 21], bilingual corpora [20, 22], multilingual parallel corpora [23] and monolingual dictionary [24,25], bilingual dictionaries [12]. Some of the studies [26–29] were relied on multiple-choice synonym questions such as SAT analogy questions, TOEFL synonym questions, ESL synonym-antonym questions. These studies fell into different types with respect to weighting scheme, similarity measurement, grammatical relations, etc. However most of these studies are not individually sufficient for synonyms. Because this approach also covers near-synonyms and does not distinguish between synonyms and other relations, hence, recent studies used different strategies: integrating two independent approaches such as distributional similarity and pattern-based approach, utilizing external features or ensemble method with combining the results to obtain more accuracy. Mirkin [9] integrated pattern-based and distributional similarity methods to acquire lexical entailment. Firstly, they extracted candidate entailment pairs for the input term by these methods. Another study [31] emphasized that selection of useful contextual information was important for the performance of synonym acquisition. Therefore, they extracted three kinds of word relationships from corpora: dependency, sentence co-occurrence, and proximity. They utilized vector space model(VSM), tf-idf weighting scheme and cosine similarity. Dependency and proximity performed relatively well by themselves. The combination performance of all contextual information gave the best result. Other study of Hagiwara (2008) [14] proposed a synonym extraction method by using supervised learning based on distributional and/or pattern-based features. They constructed five synonym classifiers: Distributional Similarity (DSIM), Distributional Features (DFEAT), Pattern-based Features (PAT), Distributional Similarity and Pattern-based Features (DSIMPAT) and Distributional and Pattern-based Features (DFEAT-PAT). Other study [15] used three vector-based models to detect semantically related nouns in Dutch. They analyzed the impact of three linguistic properties of the nouns. They compared results from a dependency-based model with context feature with 1st and 2nd order bag-of-words model. They examined the effect of the nouns’ frequency, semantic specificity and semantic class. In one of the recent studies, [30], graded relevance ranking problem was applied to discover and rank the quality of the target term’s potential synonyms. The model used supervised learning method; linear regression with three contextual features and one string similarity feature. The method was compared to two different methods [14, 27]. As a result, proposed methods outperformed the existing ones. In Turkish, recent studies on synonym relations are based on dictionary definition TDK1 and Wiktionary2 [32, 34, 35]. Within this framework, the main 1 2 Türk Dil Kurumu (The Turkish Language Association) Vikisözlük: Özgür Sözlük An Integrated Approach to Automatic Synonym Detection 119 contribution of our work is its corpus-driven characteristics and it relies on both dependency and semantic relations. 3 3.1 Methodology Data The methodology employed here is to identify the synonym pairs from a large Turkish corpus of 500M tokens. A Turkish morphological parser, which is based on a two-level morphology [33], was used. A good way to evaluate system performance is to compare the results to a gold standard. First, as gold standard, human judgments about the similarity of pairs of word are used. We manually and randomly selected 200 synonym pairs and 200 non-synonym pairs to build a training data set. Secondly, non-synonym pairs are especially selected from associated (relevant) pairs such as tree-leaf, student-school, computer-game, etc. Otherwise, selection of irrelevant pairs for negative examples can lead to false induction. The model is considered accurate if it can distinguish correct synonym pairs from relevant or strongly associated ones. 3.2 Similarity Measurement and Representation Synonym pairs were gathered on the basis of co-occurrence statistics, semantic and grammatical relations. In order to compute the similarity between concepts and eliminate incorrect candidates, we used the cosine similarity measurement based on the word space model which is a representational Vector Space. In this study, words space was derived from a specialized context obtained by dependency patterns. Vector representation of words gives strong distributional indication for synonymy detection. Similarity measurement between two vectors sometimes needs term weighting. Weighting scheme for context vectors might be normalization, pmi, dice, jaccard or raw frequency. The scheme can vary depending on the problem, therefore, it must be tested on the domain. Since we do not observe any significant improvements between the weighting formula, raw frequency is used for context vectors. 3.3 Features Our methodology relies on the assumption that synonym pairs mostly show similar dependency and semantic characteristics in corpus. They share the same meronym/holonym relations, same particular list of governing verbs, adjective modification profile and so on, by definition. Even though it is no-use applying LSPs to extract synonymy, acquisition of other semantic relations such as meronymy could be easily done by simple string matching utilization and morphological analysis. By means of the acquisitions, the proposed model can determine if 120 T. Yıldız, S. Yıldırım, and B. Diri a given word pair is synonym or not. All attributes are based on relation measurements between pairs. For each synonym pair, 15 different features are extracted from different models: co-occurrence, semantic relations based on LSPs and grammatical relations based on syntactic patterns and head-modifier relation. Feature 1: Co-occurrence. The first feature is gathered statistics about the co-occurrence of word pairs with a broad context (window size is equal to 8 from left and right) from corpora. Contrary to hypernymy and meronymy relation, it is seems impossible to directly extract synonym pairs by applying LSPs to a big corpus. Synonym pairs are not likely to co-occur together in same context and specific patterns at the same time. Therefore, first-order distributional similarity does not work for synonyms. At least, second order representation is needed. Simple co-occurrence measure might not be used for synonymy but non-synonymy. Their co-occurrence could be lower than relevant pairs. We experimentally selected dice metric to measure co-occurring feature. It is computed by roughly dividing the number of co-occurrences by summation of marginal frequencies of words. Features 2/3: Meronym/Holonym. Detection of meronymy/holonymy is used to detect synonymy relation. After applying LSPs, some elimination assumption and measurement metrics such as chi or pmi to acquire meronym/holonym relation, we obtain a big matrix in which rows depict whole candidates, columns depict part candidates and cells represent the possibility of that corresponding whole and part are in meronymy relation. To measure the similarity of meronymy profile of two given words, cosine function is applied on two rows indexed by two given words. Applying cosine function on two columns gives the similarity of holonym profile. For the relation, three different clusters of LSPs are analyzed in Turkish corpus; General (GP), Dictionary-based (TDK-P) and Bootstrapped patterns (BP) [37, 38]. First cluster is based on widely used general patterns. These patterns are collected from some pioneer studies and analyzed in Turkish. Second one is based on dictionary patterns that are extracted from TDK and Wiktionary. We adopted both types of patterns to extract the sentences that include part-whole relations from a Turkish corpus. Third cluster is based on bootstrapping. Some manually prepared seeds were used to induce and score LPSs. Based on reliability scores, we decided to filter out some generated patterns and finally obtained six different significant patterns. Once all three pattern clusters have been evaluated, third cluster of patterns (BP) showed significant performance. Table 1 shows six example patterns in third cluster(BP). All of the experiments in the studies, [37, 38], indicate that proposed methods have good indicative capacity. Features 4/5: Hyponym/Hypernym. Same procedure in meronymy acquisition holds true for hypernymy and hyponymy relation. One relation matrix is built for hypernymy/hyponymy by applying LSPs and same procedure is carried out. The most important LSPs for Turkish [36] are as follows: An Integrated Approach to Automatic Synonym Detection 1. 2. 3. 4. 121 “NPs gibi CLASS” (CLASS such as NPs), “NPs ve diğer CLASS” (NPs and other CLASS) “CLASS lArdAn NPs” (NPs from CLASS) “NPs ve benzeri CLASS” (NPs and similar CLASS) First pattern gives strong indication of is-a hierarchy. Given the syntactic patterns above, the algorithm extracts the candidate list of hyponyms for a hypernym. The method had a good capacity to get higher precision, such as 72.5% [36]. Table 1. Bootstrapped Patterns and Examples Patterns NPy+gen NPx+pos NPy+nom NPx+pos NPy+Gen (N-ADJ) NPx+Pos NPy of one-of NPxs NPx whose NPy NPxs with NPy Examples door of the house / evin kapısı house door / ev kapısı back garden gate of the house / evin arka bahçe kapısı the door of one of the houses / evlerden birinin kapısı the house whose door is locked / kapısı kilitli olan ev the house with garden and pool / bahçeli ve havuzlu ev Features 6–15. The dependency relations are obtained by syntactic patterns (or regular expression). For example, for auto and car pair, possible governing verbs bearing direct-object relations might be drive, design, produce, use, etc. The dimension of word-space model of direct-object syntactic relation consists of verbs and the cells indicate the number of times the selected noun is governed by corresponding verb. The more they are governed by the similar verb profile, the more likely they are synonyms. Likewise, the process is naturally applicable for other syntactic features. The more they are modified by same adjectives, the more likely they are synonym. Although 36 different patterns were extracted, eight were eliminated because of the poor results. Then we grouped them according to their syntactic structures. Representation of groups, number of patterns and examples in English/Turkish are given in Table 2. The essential problem we face in the experiments is the lack of features of some words. Particularly, rare words cannot be represented due to lack of corpus evidence. Even in the corpus that contains about 500M words, all instances of use of Turkish language may not be present. Thus, those instances in train data that do not occur in any of dependency and semantic relations are eliminated. Especially the pairs including low frequent word cannot be represented and evaluated by means of the methodology as the number of missing values in many features increases. Out of 400 instances, about 40–50 are discarded from training data due to insufficiency. 3.4 Binary Classification for Synonym Finally, train data turns out to contain balanced number of negative and positive examples with fifteen attributes. All the cells contain real value between 0–1. We 122 T. Yıldız, S. Yıldırım, and B. Diri know and accept that all features but co-occurrence feature have positive linear relationship with target class. Therefore, the data is considered to exhibit linear dependency. As a consequence of linearity, linear regression is an excellent and simple approach for such a classification. It has been widely used in statistical applications. The most suitable algorithm is logistic regression which can easily be used for binary classification in the domains with numeric attributes and nominal target class. Contrary to the linear regression, it builds a linear model based on a transformed target variable. Another model would be perceptron. If the data can be separated perfectly into two groups using a threshold value or a function, it is said to be linearly separable. The perceptron learning rule is capable of finding a separating hyperplane on linearly separable data. However, our problem looks more suitable for logistic regression (transformed linear regression) than perceptron. Table 2. Dependency Features Features Dependency relation direct object of verb G1 G2 G3 G4 G5 G6 G7 G8 G9 G10 4 # of Patterns Examples 13 I drive a car araba sürüyorum subject of verb 3 waiting car bekleyen araba direct object/subject of verb 3 modified by adjective+(with/without) 2 car with gasoline benzinli araba modified by inf 1 swimming pool yüzme havuzu modified by noun 1 toy car oyuncak araba modified by adjective 1 red car kırmızı araba modified by acronym locations 1 the cars in ABD ABD’deki arabalar modified by proper noun locations 1 the cars in Istanbul Istanbul’daki arabalar modified by locations 2 the car at parking lot otoparktaki araba Results and Discussion To evaluate the impact of semantic and dependency relations in finding synonyms, first, we look at their individual performances in terms of cross-validation. Picking up each feature one by one with target class, we evaluated the performance of logistic regression on the projected data. As long as the averaged fmeasured score of the corresponding feature is higher than 50%, it is considered a useful feature otherwise, independent feature. An Integrated Approach to Automatic Synonym Detection 123 The first aim is to find out which feature is the most informative for detecting synonymy and contributes most to the overall success of the model. When evaluating the result as shown in Table 3, the semantic features are notably better than syntactic dependency models in finding true synonyms. They are called to be good indicators. Table 3. F-Measure of Semantic Relations (SRs) Features F-Measure co-occurrence hyponym hypernym meronym holonym 62.5 60.5 60 68.7 73.7 Among semantic relations, the most powerful attributes are meronymy and holonymy features with f-measure of 68.7% and 73.7%, respectively. The possible reason for the success seems to be the sufficient number of cases matched by lexico-syntactic and syntactic pattern from which semantic and syntactic features are constructed. For example, the model utilizing meronymy relations has a good production capacity and success. The Table 4 shows that meronymyholonymy matrix has the size of 17K x 18K. The total number of instance is 1.7M. Average number of instances for each meronym is 102 and for each holonym is 96. They also show good performance. The averaged number of instances for hypernymy and hyponymy are 50 and 8, respectively. As a result of insufficient data volume, hypernymy/hyponymy semantic relation is relatively weaker than meronymy. Table 4. Statistics for features : Mero:Meronym, Hypo: Hyponym, AVG cpr: average case per row, AVG cpc: average case per column #ofrow #ofcol #ofcases AVG cpr AVG cpc G1 16K 1.7K 3.3M 206 2010 G2 G3 G4 G5 G6 G7 G8 G9 G10 Mero Hypo 18K 10K 13K 7K 13K 20K 6K 1.7K 13K 17K 4.3K 1.7K 1.4K 5K 1.6K 13K 5.6K 1.6K 0.2K 5K 18K 29K 3M 0.5M 1.6M 1M 5.3M 12M 0.1M 0.01M 1M 1.7M 0.2M 164 47 128 140 391 590 23 7 75 102 50 1783 341 319 621 405 2106 86 51 195 96 8 Among dependency relations, G1, G4 and G7 have better performance as shown in the Table 5. Also their production capacities are sufficient as well. The poorest groups, G8 and G9, have low production capacity and their performances are worse. As a consequence of the poor results, they are called independent and useless variables. Co-occurrence feature has negative linear relation with target class and its individual performance is 62.5%. It is acceptable as a useful feature. The successful features are linearly dependent on target class. The most suitable machine learning algorithm is the logistic regression. After aggregating all useful features which have better than the individual performances, the machine learning process was carried out and evaluated. The achievement of aggregated 124 T. Yıldız, S. Yıldırım, and B. Diri Table 5. F-measure of Dependency Relations Features G1 G2 G3 G4 G5 G6 G7 G8 G9 G10 F-Measure 64.7 58 60.5 65 61.6 58.8 63 49.4 48.3 62.6 model was evaluated in terms of cross validation. On the aggregated data where all useful features are considered, the performance of logistic regression is fmeasure of 80.3% and that of voted perceptron is 74%. The achieved score is better than the individual performance of each feature. The number of useful features is obviously the main factor to get higher scores. The proposed model utilizes only a huge corpus and morphological analyzer and it receives an acceptable score. Moreover, other useful resources might be integrated into the model to obtain better result. Dictionary definitions, WordNet, and other useful resources could be used and evaluated in future work. 5 Conclusion In this study, synonym pairs were determined on the basis of co-occurrence statistics, semantic and dependency relations within distributional aspect. Contrary to hypernymy and meronymy relation, simply applying LSPs does not extract synonym pairs from a big corpus. Instead, we extracted other semantic relations to ease detection of synonymy. Our methodology relies on some assumptions. One is that the synonym pairs mostly show similar semantic characteristics by definition. They share the same meronym/holonym and hypernym/hyponym relations. Particular lexico-syntactic patterns can be used to initiate the acquisition process of those semantic features. Secondly, a pair of synonym words mostly shares a particular list of governing verbs and modifying adjectives. The more a pair of words are governed by similar verb profile and modified by similar adjectives, the more likely they are synonym. We built ten groups of syntactic patterns according to their syntactic structures. To apply machine learning algorithm, three annotators manually and randomly selected 200 synonym pairs and 200 non-synonyms. Non-synonym pairs were especially selected from associated (relevant) pairs such as tree-leaf, appleorange, school-student. Otherwise, such negative example selection could lead to false inference. The main challenge faced in the experiments is the lack of features of some words due to their corpus evidence. Thus, such instances were eliminated. Remaining instances was classified by the most suitable algorithm which is the logistic regression. It can easily be used for binary classification in domains with numeric attributes and nominal target class. As long as individual performance of any feature is higher than f-measure of 50%, it is considered as useful features or considered independent feature from target class. The aim was to find out which features are the most informative for detecting synonymy and contribute most to the overall success of the model. When comparing the results, it was clearly observed that the semantic features are notably better than syntactic dependency models in finding true An Integrated Approach to Automatic Synonym Detection 125 synonyms. The most effective attributes are meronymy and holonymy features with weighted average f-measure of 68.7% and 73.7% respectively. The analysis indicated that the possible reason for the success is sufficiency in the number of cases from which semantic and dependency features are constructed. As a consequence of insufficient data volume, hypernymy/hyponymy relation is relatively worse than meronymy. Among dependency relations, G1, G4 and G7 outperformed the others. Likewise, it was also observed that sufficiency in the number of cases was the strong factor. After aggregating all useful features, the same learning process was carried out. The aggregated model shows promising results and performance. Regression model achieved an acceptable f-measure of 80.3%. One of the main contributions of the study is that the system first obtains acquirable semantic relations such as hypernymy, meronymy from corpus by lexico-syntactic patterns to extract subtle relations such as synonymy. The second contribution of the study is considered to be the first major attempt for Turkish synonym identification based on corpus-driven approach. References 1. Mandala, R., Tokunaga, T., Tanaka, H.: Combining Multiple Evidence from Different Types of Thesaurus for Query Expansion. In: 22nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Berkeley, CA, USA, pp. 191–197 (1999) 2. Bai, J., Song, D., Bruza, P., Nie, J., Cao, G.: Query Expansion Using Term Relationships in Language Models for Information Retrieval. In: 14th ACM International Conference on Information and Knowledge Management, Bremen, Germany, pp. 688–695 (2005) 3. Stefan, R., Liu, Y., Vasserman, A.: Translating Queries into Snippets for Improved Query Expansion. In: 22nd International Conference on Computational Linguistics, COLING 2008, Manchester, UK, pp. 737–744 (2008) 4. Lin, D.: Automatic Retrieval and Clustering of Similar Words. In: 36th Annual Meeting of the Association for Computational Linguistics and 17th International Conference on Computational Linguistics, Montreal, Quebec, Canada, pp. 768–774 (1998) 5. Inkpen, D.: A Statistical Model for Near-synonym Choice. ACM Transactions on Speech and Language Processing 4(1), 1–17 (2007) 6. Barzilay, R., Elhadad, M.: Using Lexical Chains for Text Summarization. In: Proceedings of the ACL Workshop on Intelligent Scalable Text Summarization, Madrid, Spain, pp. 10–17 (1997) 7. Inkpen, D.Z., Hirst, G.: Near-synonym Choice in Natural Language Generation. In: Recent Advances in Natural Language Processing III, Selected Papers from RANLP 2003, Borovets, Bulgaria, pp. 141–152 (2003) 8. McCarthy, D., Navigli, R.: The English Lexical Substitution Task. Language Resources and Evaluation 43(2), 139–159 (2009) 9. Mirkin, S., Dagan, I., Geffet, M.: Integrating Pattern-Based and Distributional Similarity Methods for Lexical Entailment Acquisition. In: Proceedings of the COLING/ACL 2006 on Main Conference Poster Sessions, Sydney, Austraila, pp. 579–586 (2006) 10. Harris, Z.: Distributional Structure. Word 10(23), 146–162 (1954) 126 T. Yıldız, S. Yıldırım, and B. Diri 11. Hearst, M.A.: Automatic Acquisition of Hyponyms from Large Text Corpora. In: 14th International Conference on Computational Linguistics, COLING 1992, Nantes, France, pp. 539–545 (1992) 12. Lin, D., Zhao, S., Qin, L., Zhou, M.: Identifying Synonyms among Distributionally Similar Words. In: IJCAI 2003, Proceedings of the Eighteenth International Joint Conference on Artificial Intelligence, Acapulco, Mexico, pp. 1492–1493 (2003) 13. Wang, W., Thomas, C., Sheth, A.P., Chan, V.: Pattern-based Synonym and Antonym Extraction. In: Proceedings of the 48th Annual Southeast Regional Conference, Oxford, MS, USA, p. 64 (2010) 14. Hagiwara, M.: A Supervised Learning Approach to Automatic Synonym Identification Based on Distributional Features. In: 46th Annual Meeting of the Association for Computational Linguistics on Human Language Technologies: Student Research Workshop, Columbus, OH, pp. 1–6 (2008) 15. Heylen, K., Peirsman, Y., Geeraerts, D., Speelman, D.: Modelling Word Similarity: an Evaluation of Automatic Synonymy Extraction Algorithms. In: International Conference on Language Resources and Evaluation, LREC (2008) 16. Hindle, D.: Noun Classification from Predicate-Argument Structures. In: 28th Annual Meeting of the Association for Computational Linguistics, Pittsburgh, Pennsylvania, USA, pp. 268–275 (1990) 17. Gasperin, C., Gamallo, P., Agustini, A., Lopes, G., Lima, V.: Using Syntactic Contexts for Measuring Word Similarity. In: Workshop on Knowledge Acquisition and Categorization, ESSLLI (2001) 18. Curran, J.R., Moens, M.: Improvements in Automatic Thesaurus Extraction. In: ACL 2002 Workshop on Unsupervised Lexical Acquisition, Philadelphia, USA, pp. 59–66 (2002) 19. van der Plas, L., Bouma, G.: Syntactic Contexts for Finding Semantically Related Words. In: Meeting of Computational Linguistics in the Netherlands (CLIN), Amsterdam, pp. 173–186 (2005) 20. Barzilay, R., McKeown, K.: Extracting Paraphrases from a Parallel Corpus. In: 39th Annual Meeting and 10th Conference of the European Chapter, Proceedings of the Conference, Toulouse, France, pp. 50–57 (2001) 21. Ibrahim, A., Katz, B., Lin, J.: Extracting Structural Paraphrases from Aligned Monolingual Corpora. In: The Second International Workshop on Paraphrasing: Paraphrase Acquisition and Applications, Sapporo, Japan, pp. 57–64 (2003) 22. Shimohata, M., Sumita, E.: Automatic Paraphrasing Based on Parallel Corpus for Normalization. In: Third International Conference on Language Resources and Evaluation, Las Palmas, Canary Islands, Spain, pp. 453–457 (2002) 23. van der Plas, L., Tiedemann, J.: Finding Synonyms Using Automatic Word Alignment and Measures of Distributional Similarity. In: 21st International Conference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguistics, Sydney, Australia, pp. 866–873 (2006) 24. Blondel, V.D., Sennelart, P.: Automatic Extraction of Synonyms in a Dictionary. In: SIAM Workshop on Text Mining, Arlington, VA (2002) 25. Wang, T., Hirst, G.: Exploring Patterns in Dictionary Definitions for Synonym Extraction. Natural Language Engineering 18(3), 313–342 (2012) 26. Freitag, D., Blume, M., Byrnes, J., Chow, E., Kapadia, S., Rohwer, R., Wang, Z.: New Experiments in Distributional Representations of Synonymy. In: Ninth Conference on Computational Natural Language Learning (CoNLL), Ann Arbor, Michigan, pp. 25–32 (2005) 27. Terra, E., Clarke, C.: Frequency Estimates for Statistical Word Similarity Measures. In: HTL/NAACL 2003, Edmonton, Canada, pp. 165–172 (2003) An Integrated Approach to Automatic Synonym Detection 127 28. Turney, P.D., Littman, M.L., Bigham, J., Shnayder, V.: Combining Independent Modules in Lexical Multiple-choice Problems. In: Recent Advances in Natural Language Processing III, Selected Papers from RANLP 2003, Borovets, Bulgaria, pp. 101–110 (2003) 29. Turney, P.D.: A Uniform Approach to Analogies, Synonyms, Antonyms, and Associations. In: 22nd International Conference on Computational Linguistics, Coling 2008, Manchester, UK, pp. 905–912 (2008) 30. Yates, A., Goharian, N., Frieder, O.: Graded Relevance Ranking for Synonym Discovery. In: 22nd International World Wide Web Conference, WWW 2013, Rio de Janeiro, Brazil, pp. 139–140 (2013) 31. Hagiwara, M., Ogawa, Y., Toyama, K.: Selection of Effective Contextual Information for Automatic Synonym Acquisition. In: 21st International Conference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguistics, Sydney, Australia, pp. 353–360 (2006) 32. Yazici, E., Amasyali, M.F.: Automatic Extraction of Semantic Relationships using Turkish Dictionary Definitions. In: EMO Bilimsel Dergi, Istanbul (2011) 33. Sak, H., Güngör, T., Saraçlar, M.: Turkish Language Resources: Morphological Parser, Morphological Disambiguator and Web Corpus. In: Nordström, B., Ranta, A. (eds.) GoTAL 2008. LNCS (LNAI), vol. 5221, pp. 417–427. Springer, Heidelberg (2008) 34. Serbetci, A., Orhan, Z., Pehlivan, I.: Extraction of Semantic Word Relations in Turkish from Dictionary Definitions. In: ACL 2011 Workshop on Relational Models of Semantics, Portland, pp. 11–18 (2011) 35. Orhan, Z., Pehlivan, I., Uslan, V., Onder, P.: Automated Extraction of Semantic Word Relations in Turkish Lexicon. Mathematical and Computational Applications (1), 13–22 (2011) 36. Yildirim, S., Yildiz, T.: Automatic Extraction of Turkish Hypernym-Hyponym Pairs From Large Corpus. In: COLING (Demos), pp. 493–500 (2012) 37. Yildiz, T., Diri, B., Yildirim, S.: Analysis of Lexico-syntactic Patterns for Meronym Extraction from a Turkish Corpus. In: 6th Language and Technology Conference: Human Language Technologies as a Challenge for Computer Science and Linguistics, pp. 126–138 (2013) 38. Yıldız, T., Yıldırım, S., Diri, B.: Extraction of Part-Whole Relations from Turkish Corpora. In: Gelbukh, A. (ed.) CICLing 2013, Part I. LNCS, vol. 7816, pp. 126–138. Springer, Heidelberg (2013)