Academia.edu no longer supports Internet Explorer.
To browse Academia.edu and the wider internet faster and more securely, please take a few seconds to upgrade your browser.
The paper is an attempt to utilize the componential analysis theory, the distinctive feature matrices and the conceptual frames as tools for representing human knowledge. It also presents an automatic disambiguation system (ADS) that accounts for textual ambiguities at all the linguistic levels. The traditional methods of tackling ambiguity are not adequate since they specify a few constrains assisted by clues derived from the textual information. The proposed ADS makes extensive use of the linguistic constrains LC's in the disambiguation process (DP). It also utilizes the statistical guide which is, in essence, a combination of the traditional method and the recent statistical methods. It has been found out that this intermarriage between the old (ruled base) and the new (statistical) can help solving a lot of textual analysis problems such as those related to error detection and knowledge inference. Furthermore, the model can serve other linguistic applications such as translation.
Proceedings of the 36th annual meeting on Association for Computational Linguistics -, 1998
In natural language processing, many methods have been proposed to solve the ambiguity problems. In this paper, we propose a technique to combine a method of interactive disambiguation and automatic one for alnbiguous words. The characteristic of our method is that the accuracy of the interactive disambiguation is considered. The method solves the two following problems when combining those disambiguation lnethods: (1) when should the interactive disambiguation be executed? (2) which ambiguous word should be disambiguated when more than one ambiguous words exist in a sentence? Our method defines the condition of executing the interaction with users and the order of disambiguation based on the strategy where the accuracy of the result. is maximized, considering the accuracy of the interactive disambiguation and automatic one. Using this lnethod, user interaction can be controlled while holding the accuracy of results.
Proceedings of the Conference on Recent Advances in …, 2003
This paper describes an unsupervised approach for natural language disambiguation, applicable to ambiguity problems where classes of equivalence can be defined over the set of words in a lexicon. Lexical knowledge is induced from non-ambiguous words via classes of equivalence, and enables the automatic generation of annotated corpora. The only requirements are a lexicon and a raw textual corpus. The method was tested on two natural language ambiguity tasks in several languages: part of speech tagging (English, Swedish, Chinese), and word sense disambiguation (English, Romanian). Classifiers trained on automatically constructed corpora were found to have a performance comparable with classifiers that learn from expensive manually annotated data.
2020
The aim of this paper is to present a combination of NLP and Multiple Criteria Decision-Aid (MCDA) in order to reach an effective analysis when dealing with linguistic data from various sources. The coexistence of these two concepts has allowed us, based on a set of actions and criteria, to develop a coherent system that integrates the entire process of textual data analysis (no-voweled Arabic texts) into decision making in case of ambiguity. Our solution is based on decision theory and an MCDA approach with a TOPSIS technique. This method allows the multi-scenario classification of morphosyntactical ambiguity cases in order to come out with the best performance and reduce the number of candidate scenarios.
SSRN, 2022
In present era, Natural Language Processing (NLP) is critical for improving human-machine communication. It is a broad interest to process textual data and gathers valuable and exact information from these texts. NLP compiles the text and sends the data to a computer for further processing. The current state of NLP's mathematical model for proper understanding of word meaning is unclear, and the meaning of words in context is unclear, evoking multiple senses. The spread and improvement of Natural Language Processing applications are being hampered by ambiguity in interpreting the precise meaning of texts such as machine translation (MT), Human-Machine interfaces, and so on. The approach of discovering the correct interpretation of ambiguous word in a given sentence is accepted as Word Sense Disambiguation (WSD).WSD is recognized as being one of natural language processing's more challenging and unsolved problems. Many ambiguities in natural languages are apparent, and researchers are offering to solve the problem in a variety of languages to achieve good disambiguation. These ambiguities must be solved in order to make sense including its texts and advance NLP processing and applications. WSD has a number of NLP applications for which it could be a problem, such as Machine Translation (MT), Information Retrieval (IR), Dialogues, Speech Synthesis (SS), and Question Answering (QA). The effectiveness of many strategies directly applied to WSD, such as Dictionary and Knowledge-based, Supervised, Semi-Supervised and Unsupervised approach, is compared in this study.
Proceedings of the 16th conference on Computational linguistics -, 1996
In many contexts, automatic analyzers cannot fully disambiguate a sentence or an utterance reliably, but can produce ambiguous results containing the correct interpretation. It is useful to study vatious properties of these ambiguities in the view of subsequent total or partial interactive disambiguation. We have proposed a technique for labelling ambiguities in texts and in dialogue transcriptions, and experimented it on multilingual data. It has been first necessary to define formally the very notion of ambiguity relative to a representation system, as well as associated concepts such as ambiguity kernel, ambiguity scope, ambiguity occurrence.
2003
After a brief introduction and overview of some remarkable studies about Natural Language Processing and Word Sense Disambiguation, the authors describe a general purpose statistic method for the disambiguation of polysemous words in English. Unlike other methods, the one here introduced needs no linguistic or grammatical restrictions to produce effective results. Such a method consists in an algorithm based on the co-occurrence and frequency of words in the Internet, and avails itself of WordNet 1.6 and Altavista Search Engine. The results of this study are also presented and discussed. In the final section of the paper, possible future applications for the devised algorithm are described.
In today's era most of the people are depended on the web to search some contents. At the time of searching they never bother about ambiguities that exist between words. An ambiguous word is a word that has multiple meaning in different contexts. The sense of the word is determined by the context in which the ambiguous word appears. When the user performs the search related to ambiguous word, web displays all the results related to senses of the word. Some of them are relevant and some are irrelevant according to user perspective. Word Sense Disambiguation (WSD) is the process of identifying the senses of word in textual context, when word has multiple meanings. The purpose of the research is to elaborate the methodology, approaches of WSD that can handle all issues with better performance and accuracy. In this paper the authors are discussing both the approaches and their roles in various applications like IR, MT, IE, KM etc.
Machine Translation (MT) is a crucial application of (NLP) Natural language Processing. This MT technique automatic and based on computers. One of the most modern techniques adopted in MT is machine learning (ML). Over the past few years, ML has grown in popularity during MT process among researchers. Ambiguity is a major challenge in MT. Word Sense Disambiguation (WSD) is a common technique for solving the ambiguity problem. ML approaches are commonly used for the WSD techniques and are used for training and testing purposes. The outcome prediction of the test data gives encouraging results. Text classification is one of the most significant techniques for resolving the WSD. In this paper, we have analyzed some common supervised ML text classification algorithms and also proposed a “hybrid model” called “AmbiF.” We have compared the results of all analyzed algorithms with the proposed model “AmbiF. The analyzed supervised algorithms are Decision Tree, Bayesian Network, Support Vect...
International Journal of Recent Technology and Engineering, 2014
Word sense disambiguation is a technique in the field of natural language processing where the main task is to find the correct sense in which a word occurs in a particular context. It is found to be of vital help to applications such as question answering, machine translation, text summarization, text classification, information retrieval etc. This has resulted in excessive interest in approaches based on machine learning which performs classification of word senses automatically. The main motivation behind word sense disambiguation is to allow the users to make ample use of the available technologies because ambiguities present in any language provide great difficulty in the use of information technology as words in human language that occur in a particular context can be interpreted in more than one way depending on the context. In this paper we put forward a survey of supervised, unsupervised and knowledge based approaches and algorithms available in word sense disambiguation (WSD). Index Terms-Machine readable dictionary, Machine translation, Natural language processing, Wordnet, Word sense disambiguation.
2012
This paper presents and experiments a new approach for automatic word sense disambiguation (WSD) applied for French texts. First, we are inspired from possibility theory by taking advantage of a double relevance measure (possibility and necessity) between words and their contexts. Second, we propose, analyze and compare two different training methods: judgment and dictionary based training. Third, we summarize and discuss the overall performance of the various performed tests in a global analysis way. In order to assess and compare our approach with similar WSD systems we performed experiments on the standard ROMANSEVAL test collection.
Proceedings of the 16th conference on Computational linguistics -, 1996
In this paper we sketch a decidable inference-based procedure for lexical disambiguation which operates on semantic representations of discourse and conceptual knowledge, In contrast to other approaches which use a classical logic for the disambiguating inferences and run into decidability problems, we argue on the basis of empirical evidence that the underlying iifference mechanism has to be essentially incomplete in order to be (cognitively) adequate. Since our conceptual knowledge can be represented in a rather restricted representation language, it is then possible to show that the restrictions satisfied by the conceptual knowledge and the inferences ensure in an empirically adequate ww the decidability of the problem, although a fully expressive language is used to represent discourse.
Word sense ambiguity resolution is one of the major issues in the process of machine translation. Statistical and example-based methods are usually applied for this purpose. In statistical methods, ambiguity resolution is mostly carried out by making use of some statistics extracted from previously translated documents or dual corpora of source and target languages. In this paper, we look at the problem from a different viewpoint. The proposed system consists of two main parts. The first part includes a data mining algorithm which runs offline and extracts some useful knowledge about the cooccurrences of the words. The second part of the system is an expert system whose knowledge base includes the set of association rules generated by the first part. For the inference engine of the expert system, we propose an efficient algorithm based on forward chaining in order to deduce the correct senses of the words. The performance of the system in terms of applicability and precision will be analyzed and discussed through a set of experiments.
In this article, we present an experiment of linguistic parameter tuning in the representation of the semantic space of polysemous words. We evaluate quantitatively the influence of some basic linguistic knowledge (lemmas, multi-word expressions, grammatical tags and syntactic relations) on the performances of a similarity-based Word-Sense disambiguation method. The question we try to answer, by this experiment, is which kinds of linguistic knowledge are most useful for the semantic disambiguation of polysemous words, in a multilingual framework. The experiment is about 20 French polysemous words (16 nouns and 4 verbs) and we make use of the French-English part of the sentence-aligned EuroParl Corpus for training and testing. Our results show a strong correlation between the system accuracy and the degree of precision of the linguistic features used, particularly the syntactic dependency relations. Furthermore, the lemma-based approach absolutely outperforms the word form-based approach. The best accuracy achieved by our system amounts to 90%.
Relationship Analysis System and Method for Semantic Disambiguation of Natural Language, 2007
Current approaches to natural language understanding involve example-based statistical analyses or Latent Semantic Indexing to interpret the contextual meaning of messages. However, Any Language Communications has developed a novel system that uses the innate relationships of the words in a sensible message to determine the true contextual meaning of the message. This patented methodology is called “Relationship Analysis” and includes a class/category structure of language concepts, a weighted inheritance system, a number language word conversion, and a tailored genetic algorithm to select the best of the possible word meanings. Relationship Analysis is a powerful language-independent method that has been tested using machine translations with English, French, and Arabic as source languages and English, French, German, Hindi, and Russian as target languages. A simplified form of Relationship Analysis does sophisticated text analyses, in which concepts in the text are recognized irrespective of the text language. Such analyses have been demonstrated using English and Arabic texts, with applications that include concept searches, email routing, semantic tagging, and semantic metadata indexing. In addition, a class/category data analysis provides machine-readable codes suitable for further computer system processing.
Dictionary-based Method. And by using the WordNet, we extract concepts of each of the words and Compare them with each other. And by scoring on each of the concepts of the ambiguous word, we chose the correct concept. Keywords: Word Sense Disambiguation, Supervised Method, Unsupervised Method, Semi-Supervised Method, Knowledge base and Dictionary-based Method, Lesk Algorithm.
1995
In this dissertation, I apply statistical techniques to the formidable natural language processing task of word-sense disambiguation. In particular, I develop probabilistic classi ers | systems that perform disambiguation by assigning, out of a set of word meaning designations, the one that is most probable according to a probabilistic model. The model expresses the relationships among the classi cation variable (in this case, the variable representing the sense tag of the ambiguous word) and variables that correspond to properties of the ambiguous word and the
Revista Electrónica de Lingüística Aplicada, 2022
Babelfy is an online tool, developed in the context of Natural Language Processing. When an item with more than one meaning is introduced in Babelfy, it chooses the appropriate meaning considering the context. The objective of this research study is to test the Word Sense Disambiguation skills of Babelfy in Spanish from a linguistic approach. To do so, a descriptive and comparative study between Babelfy and native Spanish speakers was carried out. Twenty-two pairs of sentences with an ambiguous word were designed, the first sentence of the pair had a neutral context and the second one a facilitating context. These sentence-pairs were introduced in Babelfy to check which meaning of the ambiguous word was selected and to explore whether there were differences depending on the type of context. These results were then compared to the answers of sixty-two Spanish native speakers. The data show that the behaviour of speakers when encountering an ambiguous word is not equivalent to the way Babelfy performs Word Sense Disambiguation, especially when the context is neutral, and the word has related meanings.
Archives of Acoustics
The paper presents an approach to semantic disambiguation in a transfer MT system based on a bilingual dictionary, i.e. the dictionary built from source-target pairs. The approach assumes building a concept ontology for nouns. The ontological concepts are applied as semantic values in lexical rules for verbs, adjectives and prepositions. A set of semantic accommodation rules is developed. The set is consulted during the process of semantic disambiguation that follows syntactical parsing. The approach has been applied in a commercial bi-directional Polish-English MT system called Translatica.
System, 1998
Reference: LEFFA, Vilson Jose. Textual Constraints In L2 Lexical Disambiguation. System, England, v. 26, n. 2, p. 183-194, 1998.
2019
In rule-based machine translation, semantic disambiguation is perhaps the most difficult phase to implement. Whereas the morphological disambiguation is largely based on tags and set names, in semantic disambiguation this is only partially possible. In this report we will discuss various possibilities for writing rules for semantic disambiguation
Loading Preview
Sorry, preview is currently unavailable. You can download the paper by clicking the button above.