Academia.edu no longer supports Internet Explorer.
To browse Academia.edu and the wider internet faster and more securely, please take a few seconds to upgrade your browser.
2009, Proceedings of the 12th Conference of the European …
On behalf of the Programme Committee, we are pleased to present the proceedings of the Student Research Workshop held at the 12th Conference of the European Chapter of the Association for Computational Linguistics. Following the tradition of providing a forum for student researchers and the success of the previous workshops held in Bergen (1999), Toulouse (2001), Budapest (2003 and Trento (2006), a panel of senior researchers will take part in the presentation of the papers, providing detailed comments on the work of the authors.
2000
In this paper we argue in favour of an integration between statistically and syntactically based parsing, where syntax is intended in terms of shallow parsing with elementary trees. None of the statistically based analyses produce an accuracy level comparable to the one obtained by means of linguistic rules . Of course their data are strictly referred to English, with the exception of [2, 3, 4]. As to Italian, purely statistically based approaches are inefficient basically due to great sparsity of tag distribution -50% or less of unambiguous tags when punctuation is subtracted from the total count as reported by . We shall discuss our general statistical and syntactic framework and then we shall report on an experiment with four different setups: the first two approaches are bottom-up driven, i.e. from local tag combinations: A. Statistics only tag disambiguation; B. Stastistics plus syntactic biases; C. Syntactic-driven disambiguation with no statistics; D. Syntactic-driven disambiguation with conditional probabilities computed on syntactic constituents. The second two approaches are top-down driven, i.e. driven from syntactic structural cues in terms of elementary trees: In a preliminary experiment we made with automatic tagger, we obtained 99% accuracy in the training set and 98% in the test set using combined approaches: data derived from statistical tagging is well below 95% even when referred to the training set, and the same applies to syntactic tagging.
2005
My thesis work would not have been possible without the help of my advisor, other collaborators, and fellow students. I am especially fortunate to have been advised by Chris Manning. Firstly, I am grateful to him for teaching me almost everything I know about doing research and being part of the academic community. Secondly, I deeply appreciate his constant support and advice on many levels. The work in this thesis was profoundly shaped by numerous insightful discussions with him. I am also very happy to have been able to collaborate with Andrew Ng on random walk models for word dependency distributions. It has been a source of inspiration to interact with someone having such far-reaching research goals and being an endless source of ideas. He also gave me valuable advice on multiple occasions. Many thanks to Dan Jurafsky for initially bringing semantic role labeling to my attention as an interesting domain my research would fit in, contributing useful ideas, and helping with my dissertation on a short notice. Thanks also to the other members of my thesis defense committee-Trevor Hastie and Francine Chen. I am also grateful to Dan Flickinger and Stephan Oepen for numerous discussions on my work in hpsg parsing. Being part of the NLP and the greater AI group at Stanford has been extremely stimulating and fun. I am very happy to have shared an office with Dan Klein and Roger Levy for several years, and with Bill McCartney for one year. Dan taught me, among other things, to always aim to win when entering a competition, and to understand things from first principles. Roger was an example of how to be an excellent researcher while maintaining a balanced life with many outside interests. vi I will miss the heated discussions about research and the fun at NLP lunch. And thanks to Jenny for the Quasi-Newton numerical optimization code. Thanks to Aria Haghighi for collaborating on semantic role labeling models and for staying up very early in the morning to finish writing up papers. I would also like to take the chance to express my gratitude to my undergraduate advisor Galia Angelova and my high-school math teacher Georgi Nikov. Although they did not contribute directly to the current effort, I wouldn't be writing this without them. I am indebted in many ways to Penka Markova-most importantly, for being my friend for many years and for teaching me optimism and self-confidence, and additionally, for collaborating with me on the hpsg parsing work. I will miss the foosball games and green tea breaks with Rajat Raina and Mayur Naik. Thanks also to Jason Townsend and Haiyan Liu for being my good friends. Thanks to Galen Andrew for making my last year at Stanford the happiest. Finally, many thanks to my parents Diana and Nikola, my sister Maria, and my nieces Iana and Diana, for their support, and for bringing me great peace and joy. I gratefully acknowledge the financial support through the ROSIE project funded by Scottish Enterprise under the Stanford-Edinburgh link programme and through ARDA's Advanced Question Answering for Intelligence (AQUAINT) program.
Arxiv preprint cs/9906006, 1999
Late in an evening in November 1993, I received a bizarre phone-call concerning a research position in a project on parsing natural language. I was told that the project is about resolving ambiguity, that it is for two years only (stressing that a PhD is not the goal) and that it pays better than being a PhD-student (an "immoral" approach :-)). It sounded like adventure because I had already met Remko Scha a couple of times the year before, when I was writing my Master's thesis on ambiguity. During one of these times I asked Remko "how do you people in natural language processing get rid of ambiguity from a natural language grammar", Remko answered tersely "we are not interested in making natural language grammars unambiguous". As a computer scientist I was puzzled; I felt that Computer Science is a "safer" place to be than those "ambiguous linguistic environments". In an interview for the job I also met Rens Bod and Steven Krauwer, who was the intended project leader. The week before the interview I had read the papers on DOP. Because I was told that there were no polynomial-time parsing algorithms for DOP, I sat down and designed such an algorithm. During the interview I explained some of the details of the algorithm, Remko and Steven were interested in seeing this written down first, Rens was surprised and did not believe it was possible. Despite of that, I was hired to develop a parser for DOP in a two year project called CLASK. Meanwhile, Remko and his group were involved in a national project ("OVIS") of the Netherlands organization for Scientific Research (NWO). The results of CLASK constituted my "visa" for joining "OVIS" for one year. After that, Remko and I decided that it is time to concentrate on writing a thesis; NWO and the Foundation for Language and Speech (STT) decided to support this proposal. This thesis exists thanks to various project proposals submitted together with Remko Scha. Without the support of Remko Scha (ILLC), Steven Krauwer (STT), Jan Landsbergen (OTS) and Alice Dijkstra (NWO), this thesis would have remained virtual. Our proposals would not have become projects without additional support from Loe Boves, Martin Everaart, Gertjan van Noord, Eric Reuland, and the STT-board. I am grateful to my promoters for the involvement and the supervision. They listened, discussed, read, commented and corrected always with so much patience. I am especially indebted to Christer Samuelsson and Remko Bonnema who read and commented on earlier versions of all chapters; in particular, Christer detected and suggested corrections to a serious error in the original paper that led to chapter 3. I thank also Ameen Abu-Hanna, ix Yaser Yacoob and Yoad Winter for reading and commenting on earlier versions. Apart from the aforementioned people, this thesis benefited from discussions with Erik Aarts,
2001
Structural ambiguity is one of the most difficult problems in natural language processing. Two disambiguation mechanisms for unrestricted text analysis are commonly used: lexical knowledge and context considerations. Our parsing method includes three different mechanisms to reveal syntactic struc-tures and an additional voting module to obtain the most probable structures for a sentence. The developed tools do not require any tagging or syntactic marking of texts.
1993
Inthis paper we will showt hat Grammatical Inference is applicable to Natural Language Processing. Given the wide and complexrange of structures appearing in an unrestricted Natural Language likeE nglish, full Grammatical Inference, yielding a comprehensive syntactic and semantic definition of English, is too much to hope for at present. Instead, we focus on techniques for dealing with ambiguity resolution by probabilistic ranking; this does not require a full formal Chomskyan grammar.W eg iv e as hort overviewo ft he different levels and methods being investigated at CCALAS for probabilistic ranking of candidates in ambiguous English input. Grammatical Inference from English corpora. An earlier title for this paper was "Overviewo fg rammar acquisition research at CCALAS, Leeds University", but this was modified to avoid the impression of an incoherent set of research strands with no integrated, focussed common techniques or applications. The researchers in our group have nod etailed development plan imposed 'from above', but are working on independent PhD programmes; however, there are common theoretical tennets, ideas, and potential applications linking individual projects. In fact, preparing for the Colloquia on Grammatical Inference has helped us to appreciate these overarching, linking themes, as we realised that the definitions stated in the Programme clearly applied to our own work at CCALAS: 'Grammatical Inference ... has suffered from the lack of a focused research community ... Simply stated, the grammatical inference problem is to learn an efficient description that captures the essence of a set of data. This description may be used subsequently to classify data, or to generate further examples of similar data.' The data in our case is unrestricted English input, as exemplified by a Corpus or large collection of text samples. This renders a much harder challenge to Grammatical Inference than artificial languages, or selected examples of wellformed English sentences. The range of lexical items and grammatical constructs appearing in an unrestricted English Corpus is very large; and the problem is not just one of scale. The Corpus-based approach carries with it a blurring of the classical Chomskyan distinction between 'grammatical' and 'ungrammatical' English sentences. Indeed, [Sampson 87] went to the extreme of positing that there is NO boundary between grammatical and ungrammatical sentences in English; this might seem to imply that it is hopeless and eveni nv alid to attempt to infer a grammar for English. Furthermore, the Corpus-based approach eschews the use of 'intuitively constructed' examples in training: a learning algorithm should be trained with 'real' sentences from a Corpus. It would seem to followf rom this that we are also proscribed from artificially constructing negative counterexamples for our learning algorithms: we cannot guarantee that such counterexamples are truly illegal.
1996
In this paper, an integrated score function is proposed to resolve the ambiguity of deepstructure, which includes the cases of constituents and the senses of words. With the integrated score function, different knowledge sources, including part-of-speech, syntax and semantics, are integrated in a uniform formulation. Based on this formulation, different models for case identification and word-sense disambiguation are derived. In the baseline system, the values of parameters are estimated by using the maximum likelihood estimation method. The accuracy rates of 56.3% for parse tree, 77.5% for case and 86.2% for word sense are obtained when the baseline system is tested on a corpus of 800 sentences. Afterwards, to reduce the estimation error caused by the maximum likelihood estimation, the Good-Turing's smoothing method is applied. In addition, a robust discriminative learning algorithm is also derived to minimize the testing set error rate. By applying these algorithms, the accura...
The Second Workshop on …, 2011
Welcome to the second workshop on Statistical Parsing of Morphologically Rich Languages! Following the warm reception of the first official SPMRL workshop at NAACL-HLT 2010, our aim with the second workshop is to build upon the success of the first and offer a platform to the growing community of people who are interested in developing tools and resources for parsing MRLs. We decided to collocate with the International Workshop on Parsing Technologies (IWPT), both because the themes of the two events are so closely related and because the seeds of the SPMRL workshop were planted during IWPT 2009 in Paris. The warm welcome and support of the IWPT community made it our unequivocal choice, and we are honored and pleased to collocate our second SPMRL workshop with this year's IWPT event Fourteen papers were submitted in total to the workshop. After two withdrawals,we chose to accept four long papers and four short papers, giving an acceptance rate of 66%. Our goal during the selection process was to produce a varied, balanced and interesting program without compromising on quality, and we believe that we have achieved this goal. This year's papers cover a broad range of languages (Arabic, Basque, French, German, Hindi, Korean,Turkish) and are concerned with the most pressing issues (handling discontinuity, incorporating morphological information, the problems of real-world text) over a range of parsing approaches (discriminative and generative, constituency and dependency) We believe that they will result in a lively and productive workshop.
If a sentence is ambiguous, it often happens that the correct reading is the one which can most easily be incorporated into the discourse context. In this paper we present a simple method for implementing this intuition using the mechanism of presupposition resolution. The basic idea is that we can choose between the alternative readings of an ambiguous sentence by picking the reading which has the greatest number of satistified presuppositions. We present two uses of the disambiguation algorithm in our bilingual human-machine dialogue system.
1993
Programme Committee received a large number of submissions (5 page extended abstracts) from all over the world. The general quality of the submissions was high. Out of a total of 229 submissions, 47 were accepted, including 7 reserve papers. Every abstract submitted was reviewed by one member of the Programme Committee and three referees (see pages v and vi). Electronic submission and reviewing procedures helped to speed up this process and turned out not to cause an unreasonable work load at our centre. We trust that the resulting programme offers an inspiring cross-section of excellent work in the field. The programme features invited talks and thematic sessions around two prominent themes in contemporary research: the relations between logic and computational linguistics, and the use of data-oriented methods in CL. The thematic orientation is further developed in the tutorial sessions which are scheduled the days preceding the conference (19-20 April 1993). New elements compared ...
Proceedings of the 17th international conference on Computational linguistics -, 1998
This paper proposes a new class-based method to estimate the strength of association in word co-occurrence for the purpose of structural disambiguation. To deal with sparseness of data, we use a conceptual dictionary as the source for acquiring upper classes of the words related in the co-occurrence, and then use t-scores to determine a pair of classes to be employed for calculating the strength of association. We have applied our method to determining dependency relations in Japanese and prepositional phrase attachments in English. The experimental results show that the method is sound, effective and useful in resolving structural ambiguities.
Procesamiento de Lenguaje Natural, 2010
Resumen: Entre el método clásico y simbólico de desambiguación de sentidos (WSD) que utiliza representaciones semánticas profundas de oraciones y textos, y el método estadístico que utiliza información relativa a la co-ocurrencia de palabras, existe una tendencia reciente a usar métodos híbridos. De manera similar a la llamada semántica lightweight (Marek, 2009), en este artículo se propone hacer uso de escasa información semántica. Describimos un modelo de aproximación sobre la base de Flat Underspecified Discourse Representation Structures (FUDRSs, cf. Eberle 2004) que valora conocimiento sobre estructura contextual, restricciones de semántica léxica e interpretaciones preferenciales. Presentamos una guía de anotación para la anotación por humanos de textos con los correspondientes indicadores. Mediante su uso, la fiabilidad de la herramienta que implementa el modelo puede ser testada con respecto a la precisión de anotación y a la predicción de desambiguación, y cómo ambas pueden ser mejoradas mediante el bootstrapping del conocimiento del sistema usando información de corpus. Para el corpus set de test considerado, la tasa de reconocimiento de la lectura preferida es del 80-90% (dependiendo de la compensación de errores de análisis sintáctico).
Linguistik online, 2003
Natural Language is highly ambiguous, on every level. This article describes a fast broadcoverage state-of-the-art parser that uses a carefully handwritten grammar and probabilitybased machine learning approaches on the syntactic level. It is shown in detail which statistical learning models based on Maximum-Likelihood Estimation (MLE) can support a highly developed linguistic grammar in the disambiguation process.
2009
Abstract This paper evaluates two semi-supervised techniques for the adaptation of a parse selection model to Wikipedia domains. The techniques examined are Structural Correspondence Learning (SCL)(Blitzer et al., 2006) and Self-training (Abney, 2007; McClosky et al., 2006). A preliminary evaluation favors the use of SCL over the simpler self-training techniques.
Arxiv preprint cs/0205025, 2002
The candidate confirms that the work submitted is his own and the appropriate credit has been given where reference has been made to the work of others. Abstract . . . refined and abstract meanings largely grow out of more concrete meanings.
2009
Abstract This paper presents a novel approach of incorporating fine-grained treebanking decisions made by human annotators as discriminative features for automatic parse disambiguation. To our best knowledge, this is the first work that exploits treebanking decisions for this task. The advantage of this approach is that use of human judgements is made. The paper presents comparative analyses of the performance of discriminative models built using treebanking decisions and state-of-the-art features.
1998
Abstract A method of syntactic disambiguation based on proper prepositional phrase attachment, or, more generally, attachment of the clauses in specific grammatical cases, is described. The research was based on Spanish and Russian material. The data set built and used by the procedure is a kind of a syntactic government patterns dictionary. The algorithm requires a morphological and a syntactic parser and assigns probability weights to the variants built by the parser. No manual markup is required.
Computational and Corpus-Based Phraseology, 2019
Multi-word terms pose many challenges in Natural Language Processing (NLP) because of their structure ambiguity. Although the structural disambiguation of multi-word expressions, also known as bracketing, has been widely studied, no definitive solution has as yet been found. Although linguists, terminologists, and translators must deal with bracketing problems, they generally must resolve problems without using advanced NLP systems. This paper describes a series of manual steps for the bracketing of multi-word terms (MWTs) based on their linguistic properties and recent advances in NLP. After analyzing 100 three-and four-term combinations, a set of criteria for MWT bracketing was devised and arranged in a step-by-step protocol based on frequency and reliability. Also presented is a case study that illustrates the procedure.
IEEE Transactions on Learning Technologies, 2002
2009
This volume contains the Proceedings of the RASLAN (RASLAN 2009), coorganized by the the Center of Natural Language Processing at the Faculty of Informatics, Masaryk University and held on December 4 th-6 th 2009 in Karlova Studánka, Sporthotel Kurzovní, Jeseníky, Czech Republic. The RASLAN Workshop is an event dedicated to exchange of information between research teams working on the projects of computer processing of Slavonic languages and related areas going on in the Centre. RASLAN is focused on theoretical as well as technical aspects of the project work, presentations of verified methods are welcomed together with descriptions of development trends. The workshop also serves as a place for discussion about new ideas. The intention is to have it as a forum for presentation and discussion of the latest developments in the the field of language engineering, especially for undergraduates and postgraduates affiliated to the NLP Center at FI MU. Topics of the Workshop include (but are not limited to): * text corpora and tagging * syntactic parsing * sense disambiguation * machine translation, computer lexicography * semantic networks and ontologies * semantic web * knowledge representation * applied systems and software for NLP
Loading Preview
Sorry, preview is currently unavailable. You can download the paper by clicking the button above.