Academia.edu no longer supports Internet Explorer.
To browse Academia.edu and the wider internet faster and more securely, please take a few seconds to upgrade your browser.
1995
In this dissertation, I apply statistical techniques to the formidable natural language processing task of word-sense disambiguation. In particular, I develop probabilistic classi ers | systems that perform disambiguation by assigning, out of a set of word meaning designations, the one that is most probable according to a probabilistic model. The model expresses the relationships among the classi cation variable (in this case, the variable representing the sense tag of the ambiguous word) and variables that correspond to properties of the ambiguous word and the
1991
We describe a statistical technique for assigning senses to words. An instance of a word is assigned a sense by asking a question about the context in which the word appears. The question is constructed to have high mutual information with the translation of that instance in another language. When we incorporated this method of assigning senses into our statistical machine translation system, the error rate of the system decreased by thirteen percent.
1997
The Naive Mix is a new supervised learning algorithm that is based on a sequential method for selecting probabilistic models. The usual objective of model selection is to nd a single model that adequately characterizes the data in a training sample. However, during model selection a sequence of models is generated that consists of the best{ tting model at each level of model complexity. The Naive Mix utilizes this sequence of models to de ne a probabilistic model which is then used as a probabilistic classi er to perform word{sense disambiguation. The models in this sequence are restricted to the class of decomposable log{linear models. This class of models o ers a number of computational advantages. Experiments disambiguating twelve di erent words show that a Naive Mix formulated with a forward sequential search and Akaike's Information Criteria rivals established supervised learning algorithms such as decision trees (C4.5), rule induction (CN2) and nearest{neighbor classi cation (PEBLS).
This paper proposes and assesses a new possibilistic approach for automatic monolingual word sense disambiguation (WSD). In fact, in spite of their advantages, the traditional dictionaries suffer from the lack of accurate information useful for WSD. Moreover, there exists a lack of high-coverage semantically labeled corpora on which methods of learning could be trained. For these multiple reasons, it became important to use a semantic dictionary of contexts (SDC) ensuring the machine learning in a semantic platform of WSD. Our approach combines traditional dictionaries and labeled corpora to build a SDC and identify the sense of a word by using a possibilistic matching model. Besides, we present and evaluate a second new probabilistic approach for automatic monolingual WSD. This approach uses and extends an existing probabilistic semantic distance to compute similarities between words by exploiting a semantic graph of a traditional dictionary and the SDC. To assess and compare these two approaches, we performed experiments on the standard ROMANSEVAL test collection and we compared our results to some existing French monolingual WSD systems. Experiments showed an encouraging improvement in terms of disambiguation rates of French words. These results reveal the contribution of possibility theory as a mean to treat imprecision in information systems.
2012
Artificial intelligence (AI) has been a major research area in the later quarter of 20 th century and is likely to be even more so in the 21 st century. A key part of AI is Word Sense Disambiguation (WSD) which deals with choosing the correct sense of a word in the given text. All human languages have words with multiple meaning and selecting the intended sense is important. This paper briefly describes various methods presently used for WSD and their relative effectiveness. WSD applications currently find application in Information Retrieval, Information Extraction, Automated Answering Machine, Speech Reorganization, Machine Translation among many others. WSD has promise for the future in taking AI to the next higher level. Kywords: Natural Language Processing (NLP), Artificial Intelligence (AI), Word Sense Disambiguation (WSD), Knowledge Based Methods, Supervised/Unsupervised Methods.
Lecture Notes in Computer Science, 2002
This paper presents a method to combine two unsupervised methods (Specification Marks, Conceptual Density) and one supervised (Maximum Entropy) for the automatic resolution of lexical ambiguity of nouns in English texts. The main objective is to improved the accuracy of knowledge-based methods with statistical information supplied by the corpus-based method. We explore a way of combining the classification results of the three methods: "voting" is the way we have chosen to combine the three methods in one unique decision. These three methods have been applied both individually as in a combined way to disambiguate a set of polysemous words. Our results show that a combination of different knowledge-based methods and the addition of statistical information from a corpus-based method might eventually lead to improve accuracy of first ones.
Knowledge and Information Systems
This paper proposes and assesses a new possibilistic approach for automatic monolingual word sense disambiguation (WSD). In fact, in spite of their advantages, the traditional dictionaries suffer from the lack of accurate information useful for WSD. Moreover, there exists a lack of high-coverage semantically labeled corpora on which methods of learning could be trained. For these multiple reasons, it became important to use a semantic dictionary of contexts (SDC) ensuring the machine learning in a semantic platform of WSD. Our approach combines traditional dictionaries and labeled corpora to build a SDC and identify the sense of a word by using a possibilistic matching model. Besides, we present and evaluate a second new probabilistic approach for automatic monolingual WSD. This approach uses and extends an existing probabilistic semantic distance to compute similarities between words by exploiting a semantic graph of a traditional dictionary and the SDC. To assess and compare these...
Revista Espanola De Linguistica Aplicada, 2009
This paper presents an algorithm based on collocational data for word sense disambiguation (WSD). The aim of this algorithm is to maximize efficiency by minimizing (1) computational costs and (2) linguistic tagging/annotation. The formalization of our WSD algorithm is based on discriminant function analysis (DFA). This statistical technique allows us to parameterize each collocational item with its meaning, using just bare text. The parameterized data allow us to classify cases (sentences with an ambiguous word) into the values of a categorical dependent (each of the meanings of the ambiguous word). To evaluate the validity and efficiency of our WSD algorithm, we previously hand sense-tagged all the sentences containing ambiguous words and then cross-validated the hand sense-tagged data with the automatic WSD performance. Finally, we present the global results of our algorithm after applying it to a limited set of words in both languages: Spanish and English, highlighting the points...
Proceedings of the 32nd annual meeting on Association for Computational Linguistics -, 1994
Most probabilistic classifiers used for word-sense disambiguation have either been based on only one contextual feature or have used a model that is simply assumed to characterize the interdependencies among multiple contextual features. In this paper, a different approach to formulating a probabilistic model is presented along with a case study of the performance of models produced in this manner for the disambiguation of the noun interest. We describe a method for formulating probabilistic models that use multiple contextual features for word-sense disambiguation, without requiring untested assumptions regarding the form of the model. Using this approach, the joint distribution of all variables is described by only the most systematic variable interactions, thereby limiting the number of parameters to be estimated, supporting computational efficiency, and providing an understanding of the data.
2012
This paper presents and experiments a new approach for automatic word sense disambiguation (WSD) applied for French texts. First, we are inspired from possibility theory by taking advantage of a double relevance measure (possibility and necessity) between words and their contexts. Second, we propose, analyze and compare two different training methods: judgment and dictionary based training. Third, we summarize and discuss the overall performance of the various performed tests in a global analysis way. In order to assess and compare our approach with similar WSD systems we performed experiments on the standard ROMANSEVAL test collection.
International Journal of Recent Technology and Engineering, 2014
Word sense disambiguation is a technique in the field of natural language processing where the main task is to find the correct sense in which a word occurs in a particular context. It is found to be of vital help to applications such as question answering, machine translation, text summarization, text classification, information retrieval etc. This has resulted in excessive interest in approaches based on machine learning which performs classification of word senses automatically. The main motivation behind word sense disambiguation is to allow the users to make ample use of the available technologies because ambiguities present in any language provide great difficulty in the use of information technology as words in human language that occur in a particular context can be interpreted in more than one way depending on the context. In this paper we put forward a survey of supervised, unsupervised and knowledge based approaches and algorithms available in word sense disambiguation (WSD). Index Terms-Machine readable dictionary, Machine translation, Natural language processing, Wordnet, Word sense disambiguation.
2009
Abstract Word sense disambiguation (WSD) is the ability to identify the meaning of words in context in a computational manner. WSD is considered an AI-complete problem, that is, a task whose solution is at least as hard as the most difficult problems in artificial intelligence. We introduce the reader to the motivations for solving the ambiguity of words and provide a description of the task. We overview supervised, unsupervised, and knowledge-based approaches.
Proceedings of the workshop on Human Language Technology - HLT '94, 1994
This paper presents and evaluates models created according to a schema that provides a description of the joint distribution of the values of sense tags and contextual features that is potentially applicable to a wide range of content words. The models are evaluated through a series of experiments, the results of which suggest that the schema is particularly well suited to nouns but that it is also applicable to words in other syntactic categories.
The word sense disambiguation (WSD) is the task ofautomatically selecting the correct sense given a context and it helps in solving many ambiguity problems inherently existing in all natural languages.Statistical Natural Language Processing (NLP),which is based on probabilistic, stochastic and statistical methods, has been used to solve many NLP problems.The Naive Bayes algorithm which is one of the supervised learning techniques has worked well in many classification problems. In the present work, WSD task to disambiguate the senses of different words from the standard corpora available in the " 1998 SENSEVAL Word Sense Disambiguation (WSD) shared task " is performed by applying Naïve Bayes machine learning technique. It is observed that senses of ambiguous word having lesser number of part-of-speeches are disambiguated more correctly. Other key observation is that with lesser number of senses to be disambiguated, the chances of words being disambiguated with correct senses are more. I. INTRODUCTION The ambiguity in the senses of the words of different languages does exist inherently in all natural languages used by humans. There are many words in every language which carry more than one meaning for the same word. For example, the word ―chair‖ has one sense which means a piece of furniture and other sense of it means a person chairing say some session. So obviously we need some context to select the correct sense given a situation. Automatically selecting the correct sense given a context is in the core of solving many ambiguity problems. The word sense disambiguation (WSD) is the task to automatically determine which of the senses of an ambiguous (target) word is chosen in the specific use of the word by taking into consideration the context of word's use [1,2]. Having an accurate and reliable word sense disambiguation has been the target of natural language community since long. The motivation and belief behind performing word sense disambiguation is that many tasks which are performed under the umbrella of NLP are highly benefitted with properly disambiguated word senses.Statistical NLP, a special approach of NLP based onthe probabilistic, stochastic and statistical methods, uses machine learning algorithms to solve many NLP problems. AS a branch ofartificial intelligence, machine learning involves computationallylearning patterns from given data, and applying to new or unseen data the pattern which were learned earlier. Machine learning is defined by Tom M.Mitchell as ―A computer program is said to learn from experience E with respect to some class of tasksT and performance measure P, if its performance at tasks in T,as measured by P, improves withexperience E [3].‖ Learning algorithms can be generally classified into three types: supervised learning, semi-supervised learning and unsupervised learning. Supervised learning technique is based on the idea of studying the features of positive and negative examples over a large collection of annotated corpus. Semi-supervised learning uses both labeled data and unlabeled data for the learning process to reduce the dependence on training data. In the unsupervised learning, decisions are made on the basis of unlabeled data. The methods of unsupervised learning are mostly built upon clustering techniques, similarity based functions and distribution statistics. For automatic WSD,supervised learningis one ofthe most successfulapproaches.
WSEAS Transactions on …, 2003
In this paper we propose and discuss a method for Word Sense Disambiguation. A Lexicon approach is presented based on the use of the WordNet. More precisely, the context and the senses of the ambiguous word are represented as vectors of weighted terms, in a vector space model, using WordNet definitions and the rich hypernymy relations. Calculating the conditional probabilities (relative frequencies) for these terms we can measure the similarity of the target word with a sense. Hence, the ambiguous word in the context is assigned to the most similar sense. Our algorithm does not need any training and is tested on the entire Semantic Concordance Corpus (Semcor). The estimated performance of the algorithm is 78,13% .
Abstract Word Sense Disambiguation (WSD) is one of the most important open problems in Natural Language Processing. One of the most successful current lines of research in WSD is the corpus-based approach, in which machine learning algorithms are applied to learn statistical models or classifiers from corpora. When a machine learning approach learns from previously semantically annotated corpora it is said to be supervised, whereas when it does not use sense tagged data during training it is called unsupervised.
SpringerReference
This paper describes a program that disambignates English word senses in unrestricted text using statistical models of the major Roget's Thesaurus categories. Roget's categories serve as approximations of conceptual classes. The categories listed for a word in Roger's index tend to correspond to sense distinctions; thus selecting the most likely category provides a useful level of sense disambiguatiou. The selection of categories is accomplished by identifying and weighting words that are indicative of each category when seen in context, using a Bayesian theoretical framework.
Dictionary-based Method. And by using the WordNet, we extract concepts of each of the words and Compare them with each other. And by scoring on each of the concepts of the ambiguous word, we chose the correct concept. Keywords: Word Sense Disambiguation, Supervised Method, Unsupervised Method, Semi-Supervised Method, Knowledge base and Dictionary-based Method, Lesk Algorithm.
Proceedings of the 2nd workshop on Learning language in logic and the 4th conference on Computational natural language learning -, 2000
This paper describes a set of comparative experiments, including cross-corpus evaluation, between five alternative algorithms for supervised Word Sense Disambiguation (WSD), namely Naive Bayes, Exemplar-based learning, SNoW, Decision Lists, and Boosting. Two main conclusions can be drawn: 1) The LazyBoosting algorithm outperforms the other four state-of-theart algorithms in terms of accuracy and ability to tune to new domains; 2) The domain dependence of WSD systems seems very strong and suggests that some kind of adaptation or tuning is required for cross-corpus application.
Loading Preview
Sorry, preview is currently unavailable. You can download the paper by clicking the button above.