Syntax – Ixa Group. Language Technology. https://www.ehu.eus/ehusfera/ixa News from the Ixa Group in the University of the Basque Country Wed, 07 Jan 2015 16:56:43 +0000 en-US hourly 1 https://wordpress.org/?v=5.6.4 Ninth Workshop on Syntax, Semantics and Structure in Statistical Translation (SSST-9) https://www.ehu.eus/ehusfera/ixa/2015/01/07/ninth-workshop-on-syntax-semantics-and-structure-in-statistical-translation-ssst-9/ https://www.ehu.eus/ehusfera/ixa/2015/01/07/ninth-workshop-on-syntax-semantics-and-structure-in-statistical-translation-ssst-9/#comments Wed, 07 Jan 2015 16:52:24 +0000 http://www.ehu.eus/ehusfera/ixa/?p=2103 Eneko AGIRRE and Nora ARANBERRI, Ixa Group from University of the Basque Country, together with Dekai Wu, Hong Kong University of Science and Technology (HKUST) and Marine Carpuat, National Research Council (NRC) Canada, are the organizers of the SSST-9 – ”Ninth Workshop on Syntax, Semantics and Structure in Statistical Translation” that takes place in Denver, [...]]]> Eneko AGIRRE and Nora ARANBERRI, Ixa Group from University of the Basque Country,  together with Dekai Wu, Hong Kong University of Science and Technology (HKUST) and Marine Carpuat, National Research Council (NRC) Canada, are the organizers of the SSST-9 – ”Ninth Workshop on Syntax, Semantics and Structure in Statistical Translation” that takes place in Denver, Colorado, USA, in Jun, 4, 2015. (NAACL HLT 2015 / SIGMT / SIGLEX)

This Workshop seeks to bring together a large number of researchers working on diverse aspects of structure, semantics and representation in relation to statistical machine translation. Since its first edition in 2006, its program each year has comprised high-quality papers discussing current work spanning topics including: new grammatical models of translation; new learning methods for syntax- and semantics-based models; formal properties of synchronous/transduction grammars (hereafter S/TGs); discriminative training of models incorporating linguistic features; using S/TGs for semantics and generation; and syntax- and semantics-based evaluation of machine translation.

QTLeap_LogoQTLeap project Best Paper Award

This year SSST-9 will award a best paper award among papers which advance MT using semantics and deep language processing.

This award is sponsored by the European Union QTLeap project.

IXA talde is a partner in QTLeap (Quality Translation by Deep Language Engineering Approaches),  a project that is run by an European consortium with other seven partners: Bulgarian Academy of Sciences, Charles University in Prague, German Research Center for Artificial Intelligence, Higher Functions Lda., Humboldt University in Berlin, University of the Basque Country, University of Groningen and University of Lisbon. For more information and contact details please visit: qtleap.eu.

Organizers

Important Dates

Submission deadline for papers and extended abstracts: 8 Mar 2015
Notification to authors: 24 Mar 2015
Camera copy deadline: 3 Apr 2015

]]>
https://www.ehu.eus/ehusfera/ixa/2015/01/07/ninth-workshop-on-syntax-semantics-and-structure-in-statistical-translation-ssst-9/feed/ 1
Talk: Going to the Roots of Dependency Parsing (Miguel Ballesteros, 2014/12/16) https://www.ehu.eus/ehusfera/ixa/2014/12/15/talk-going-to-the-roots-of-dependency-parsing-miguel-ballesteros-20141216/ https://www.ehu.eus/ehusfera/ixa/2014/12/15/talk-going-to-the-roots-of-dependency-parsing-miguel-ballesteros-20141216/#comments Mon, 15 Dec 2014 23:13:42 +0000 http://www.ehu.eus/ehusfera/ixa/?p=2096 Speaker: Miguel Ballesteros (Universitat Pompeu Fabra)

Miguel is a Visiting lecturer – Postdoc in Pompeu Fabra University, Barcelona. He works on natural language processing and machine learning with a special interest on linguistic structure prediction problems, such as dependency parsing and phrase structure parsing. He completed his BsC, MsC and PhD at the [...]]]> Speaker:  Miguel Ballesteros (Universitat Pompeu Fabra)

Miguel is a Visiting lecturer – Postdoc in Pompeu Fabra University, Barcelona. He works on natural language processing and machine learning with a special interest on linguistic structure prediction problems, such as dependency parsing and phrase structure parsing.  He completed his BsC,  MsC and PhD at the Universidad Complutense de Madrid. During the last years, he was a Visiting Researcher in Uppsala University, Singapore University of Technology and Design and Carnegie Mellon University.

Data: December 16th 2014, Tuesday
Time: 15:30-16:30
Room: 3.2 room. Faculty of Informatics (UPV/EHU)

Title: “Going to the Roots of Dependency Parsing.”

Abstract:

I will first introduce transition-based dependency parsing and present the conclusions extracted from a set of experiments on the root node of dependency analysis,
besides I’m going to sum up my current, past and future research collaboration projects with some new results and developments.

Dependency trees used in syntactic parsing often include a root node representing a dummy
word prefixed or suffixed to the sentence, a device that is generally considered a mere technical
convenience and is tacitly assumed to have no impact on empirical results. We demonstrate that
this assumption is false and that the accuracy of data-driven dependency parsers can in fact be
sensitive to the existence and placement of the dummy root node. In particular, we show that
a greedy, left-to-right, arc-eager transition-based parser consistently performs worse when the
dummy root node is placed at the beginning of the sentence (following the current convention
in data-driven dependency parsing) than when it is placed at the end or omitted completely.
Control experiments with an arc-standard transition-based parser and an arc-factored graph-
based parser reveal no consistent preferences but nevertheless exhibit considerable variation in
results depending on root placement. We conclude that the treatment of dummy root nodes in
data-driven dependency parsing is an underestimated source of variation in experiments and
may also be a parameter worth tuning for some parsers.

Papers:

Miguel Ballesteros and Joakim Nivre.  2013. Going to the Roots of Dependency Parsing. Computational Linguistics 39:1

Miguel Ballesteros
and Bernd Bohnet   2014.  Automatic Feature Selection for Agenda-Based Dependency Parsing . 25th International Conference on Computational Linguistics (COLING 2014) Dublin, Ireland

Isabel Ortiz, Miguel Ballesteros and Yue Zhang  2014.  ViZPar: A GUI for ZPar with Manual Feature Selection . Demonstrations of the 30th Spanish Conference on Natural Language Processing (SEPLN 2014) Girona, Spain

Miguel Ballesteros and Joakim Nivre.  2014.  MaltOptimizer: Fast and Effective Parser Optimization. Natural Language Engineering.

Miguel Ballesteros, Bernd Bohnet, Simon Mille and Leo Wanner  2014. Deep-Syntactic Parsing. 25th International Conference on Computational Linguistics (COLING 2014) Dublin, Ireland

]]> https://www.ehu.eus/ehusfera/ixa/2014/12/15/talk-going-to-the-roots-of-dependency-parsing-miguel-ballesteros-20141216/feed/ 1
Talk: A Vector Space Model approach to discover constructions (Toni Martí, 2014-12-15) https://www.ehu.eus/ehusfera/ixa/2014/12/12/talk-a-vector-space-model-approach-to-discover-constructions-toni-marti-2014-12-15/ https://www.ehu.eus/ehusfera/ixa/2014/12/12/talk-a-vector-space-model-approach-to-discover-constructions-toni-marti-2014-12-15/#comments Fri, 12 Dec 2014 11:03:11 +0000 http://www.ehu.eus/ehusfera/ixa/?p=2090 Speaker: Toni Martí, Universitat de Barcelona, General Linguistics, Data: December 15th 2014, Monday Time: 11:00 – 12:30 Room: 3.2 computer Science Faculty (UPV/EHU)

Title: “A Vector Space Model approach to discover constructions”

Abstract:

In cognitive models, a construction is a conventional symbolic unit that involves a pairing of form and meaning. [...]]]>

Speaker: Toni Martí,
Universitat de Barcelona, General Linguistics,
Data: December 15th 2014, Monday
Time: 11:00 – 12:30
Room: 3.2 computer Science Faculty (UPV/EHU)

Title: “A Vector Space Model approach to discover constructions”

Abstract:

In cognitive models, a construction is a conventional symbolic unit that involves a pairing of form and meaning. These units can be of different types depending on their complexity -morphemes, words, compound words, collocates, idioms and more abstract/schematic patterns. Cognitive Linguistics assumes the hypothesis that these constructions are learned from usage and stored in the human memory, where they are accessed during both the production and comprehension of language. Therefore, constructions are fundamental linguistic units for infering the structure of language, for infering speakers’ knowledge of language, and their identification is crucial for language understanding. Although a broad range of these linguistic structures have been subjected to linguistic analysis, we assume that there exist a huge number of constructions that are still to be discovered. There are different approaches to the task of identifying and discovering them, depending on the type of construction we are looking for or dealing with. This fact allows for a wide range of methods and approaches aiming at the treatment of this kind of linguistic units.

From the point of view of the methodology and knowledge applied in the automatic detection of constructions, we can distinguish two main approaches: those that have been guided by previously gathered empirical data and those that apply methods oriented to learning constructions from plain text or automatically annotated text. That is, those methods that do not use manually annotated data nor are based on ad hoc linguistic knowledge.

Our proposal is based on the Harris distributional hypothesis, which states that semantically related words (or other linguistic units) will share the same context. We propose a new specific hypothesis within the family of distributional hypotheses the pattern-construction hypothesis, which states that those contexts that are relevant to the definition of a cluster of semantically related words, tend to be (part of) a lexico-syntactic construction. Following this hypothesis, we implemented a methodology that uses Vector Space Models (VSM) to discover candidates for consideration as constructions from a large automatically processed corpus. This approach is in line with the idea proposed by Landauer et al. 2007, who states that VSMs are plausible models of some aspects of human cognition.

]]> https://www.ehu.eus/ehusfera/ixa/2014/12/12/talk-a-vector-space-model-approach-to-discover-constructions-toni-marti-2014-12-15/feed/ 1