<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://wiki.nlpl.eu/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Sara</id>
	<title>Nordic Language Processing Laboratory - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://wiki.nlpl.eu/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Sara"/>
	<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/Special:Contributions/Sara"/>
	<updated>2026-05-16T04:47:26Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.31.10</generator>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/home&amp;diff=923</id>
		<title>Parsing/home</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/home&amp;diff=923"/>
		<updated>2020-01-15T07:18:55Z</updated>

		<summary type="html">&lt;p&gt;Sara: /* Parsing Systems */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Background =&lt;br /&gt;
&lt;br /&gt;
An experimentation environment for data-driven dependency parsing is maintained for NLPL under the coordination of Uppsala University (UU).&lt;br /&gt;
The data is available on the Norwegian Saga cluster and on the Finnish Puhti cluster.&lt;br /&gt;
The software is available on the Norwegian Saga cluster&lt;br /&gt;
&lt;br /&gt;
Initially, software and data were commissioned on the Norwegian Abel supercluster, see [http://wiki.nlpl.eu/index.php/Parsing/abel The Abel page] for legacy information.&lt;br /&gt;
&lt;br /&gt;
= Preprocessing Tools =&lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe]&lt;br /&gt;
&lt;br /&gt;
Additionally, a variety of tools for sentence splitting, tokenization, lemmatization, et al.&lt;br /&gt;
are available through the NLPL installations of the&lt;br /&gt;
[http://nltk.org Natural Language Processing Toolkit (NLTK)] and the&lt;br /&gt;
[https://spacy.io spaCy: Natural Language Processing in Python] tools.&lt;br /&gt;
&lt;br /&gt;
= Parsing Systems =&lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/uuparser The Uppsala Parser]&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe] &lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/turboparser TurboParser]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Additionallly, parsers are available in several toolkits installed by nlpl: [http://wiki.nlpl.eu/index.php/Parsing/stanfordnlp StanfordNLP], [https://www.nltk.org/ NLTK], [https://spacy.io/ spaCy].&lt;br /&gt;
&lt;br /&gt;
= Training and Evaluation Data = &lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/ud Universal Dependencies v2.0–2.5]&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/sdp Semantic Dependency Parsing]&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/home&amp;diff=922</id>
		<title>Parsing/home</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/home&amp;diff=922"/>
		<updated>2020-01-15T07:17:50Z</updated>

		<summary type="html">&lt;p&gt;Sara: /* Parsing Systems */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Background =&lt;br /&gt;
&lt;br /&gt;
An experimentation environment for data-driven dependency parsing is maintained for NLPL under the coordination of Uppsala University (UU).&lt;br /&gt;
The data is available on the Norwegian Saga cluster and on the Finnish Puhti cluster.&lt;br /&gt;
The software is available on the Norwegian Saga cluster&lt;br /&gt;
&lt;br /&gt;
Initially, software and data were commissioned on the Norwegian Abel supercluster, see [http://wiki.nlpl.eu/index.php/Parsing/abel The Abel page] for legacy information.&lt;br /&gt;
&lt;br /&gt;
= Preprocessing Tools =&lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe]&lt;br /&gt;
&lt;br /&gt;
Additionally, a variety of tools for sentence splitting, tokenization, lemmatization, et al.&lt;br /&gt;
are available through the NLPL installations of the&lt;br /&gt;
[http://nltk.org Natural Language Processing Toolkit (NLTK)] and the&lt;br /&gt;
[https://spacy.io spaCy: Natural Language Processing in Python] tools.&lt;br /&gt;
&lt;br /&gt;
= Parsing Systems =&lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/uuparser The Uppsala Parser]&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe] &lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/turboparser TurboParser]&lt;br /&gt;
* Additional parsers: [http://wiki.nlpl.eu/index.php/Parsing/stanfordnlp StanfordNLP], [https://www.nltk.org/ NLTK], [https://spacy.io/ spaCy]. For the parsers in these toolkits we refer to the official documentation.&lt;br /&gt;
&lt;br /&gt;
= Training and Evaluation Data = &lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/ud Universal Dependencies v2.0–2.5]&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/sdp Semantic Dependency Parsing]&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Infrastructure/software/catalogue&amp;diff=920</id>
		<title>Infrastructure/software/catalogue</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Infrastructure/software/catalogue&amp;diff=920"/>
		<updated>2020-01-14T21:34:50Z</updated>

		<summary type="html">&lt;p&gt;Sara: /* Activity C: Data-Driven Parsing */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Background =&lt;br /&gt;
&lt;br /&gt;
This page provides a high-level summary of NLPL-specific software installed on either of our two systems.&lt;br /&gt;
As a rule of thumb, NLPL aims to build on generic software installations provided by the&lt;br /&gt;
system maintainers (e.g. development tools and libraries that are not discipline-specific),&lt;br /&gt;
using the [http://modules.sourceforge.net/ &amp;lt;tt&amp;gt;module&amp;lt;/tt&amp;gt;s infrastructure].&lt;br /&gt;
For example, an environment like OpenNMT is unlikely to be used by other disciplines,&lt;br /&gt;
and NLPL stands to gain from in-house, shared expertise that comes with maintaining&lt;br /&gt;
a project-specific installation.&lt;br /&gt;
On the other hand, the CUDA libraries are general extensions to the operating system&lt;br /&gt;
that most users of deep learning frameworks on gpus will want to use; hence, CUDA is&lt;br /&gt;
most appropriately installed by the core system maintainers.&lt;br /&gt;
Frameworks like PyTorch and TensorFlow, arguably, present a middle ground to this&lt;br /&gt;
rule of thumb:&lt;br /&gt;
In principle, they are not discipline-specific, but in mid-2018 at least the demand for&lt;br /&gt;
installations of these frameworks is strong within NLPL, and the project will likely&lt;br /&gt;
benefit from growing its competencies in this area.&lt;br /&gt;
&lt;br /&gt;
= Module Catalogue =&lt;br /&gt;
&lt;br /&gt;
The discipline-specific modules maintained by NLPL are not activated by default.&lt;br /&gt;
To make available the NLPL community directory of software modules, on top of the&lt;br /&gt;
pre-configured, system-wide modules, one needs to execute the following&lt;br /&gt;
(on Abel, Puhti, or Taito):&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
module use -a /proj*/nlpl/software/modules/etc&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
For Saga, the NLPL community directory is in a different location:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
module use -a /cluster/shared/nlpl/software/modules/etc&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We will at times assume a shell variable &amp;lt;tt&amp;gt;$NLPLROOT&amp;lt;/tt&amp;gt; that points to the&lt;br /&gt;
top-level project directory, i.e. &amp;lt;tt&amp;gt;/projects/nlpl/&amp;lt;/tt&amp;gt; (on Abel),&lt;br /&gt;
&amp;lt;tt&amp;gt;/proj/nlpl/&amp;lt;/tt&amp;gt; (on Taito),&lt;br /&gt;
&amp;lt;tt&amp;gt;/projappl/nlpl/&amp;lt;/tt&amp;gt; (on Puhti), and&lt;br /&gt;
&amp;lt;tt&amp;gt;/cluster/shared/nlpl/&amp;lt;/tt&amp;gt; (on Saga).&lt;br /&gt;
&lt;br /&gt;
For NLPL users, we recommend that one adds the above &amp;lt;tt&amp;gt;module use&amp;lt;/tt&amp;gt; command&lt;br /&gt;
to the shell start-up script, e.g. &amp;lt;tt&amp;gt;.bashrc&amp;lt;/tt&amp;gt; in the user home directory.&lt;br /&gt;
&lt;br /&gt;
To inspect what is available, one can use the &amp;lt;tt&amp;gt;avail&amp;lt;/tt&amp;gt; sub-command&lt;br /&gt;
(on Abel), e.g.&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
module avail 2&amp;gt;&amp;amp;1 | grep nlpl&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
= User-Installed Software =&lt;br /&gt;
&lt;br /&gt;
Even if NLPL strives to make available a comprehensive set of ready-to-run sofware modules,&lt;br /&gt;
users will at times want to install their own add-on components.&lt;br /&gt;
For Python add-on components, some&lt;br /&gt;
[http://wiki.nlpl.eu/index.php/Infrastructure/software/user emerging instructions] are available.&lt;br /&gt;
&lt;br /&gt;
= Activity A: Basic Infrastructure =&lt;br /&gt;
&lt;br /&gt;
Interoperability of NLPL installations with each other, as well as with system-wide&lt;br /&gt;
software that is maintained by the core operations teams for Abel and Taito, is no&lt;br /&gt;
small challenge; neither is parallelism across the two systems, for example in&lt;br /&gt;
available software (and versions) and techniques for ‘mixing and matching’.&lt;br /&gt;
These challenges are discussed in some more detail with regard to the&lt;br /&gt;
[http://wiki.nlpl.eu/index.php/Infrastructure/software/python Python programming environment]&lt;br /&gt;
and with regard to&lt;br /&gt;
[http://wiki.nlpl.eu/index.php/Infrastructure/software/frameworks common Deep Learning frameworks].&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Module Name/Version !! Description !! System !! Install Date !! Maintainer&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-cupy/5.4.0 || Matrix Library Accelerated by CUDA || Abel (3.7) || May 2018 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-cython/0.29.3 || C Extensions for Python || Abel (3.5, 3.7) || December 2018 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-dynet/2.1 || DyNet Dynamic Neural Network Toolkit (CPU) || Abel (3.5, 3.7) || February 2019 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Infrastructure/software/nltk nlpl-nltk/3.3] || Natural Language Toolkit (NLTK) || Abel, Taito || September 2018 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Infrastructure/software/pytorch nlpl-pytorch/0.4.1] || PyTorch Deep Learning Framework (CPU and GPU) || Abel, Taito || September 2018 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Infrastructure/software/pytorch nlpl-pytorch/1.0.0] || PyTorch Deep Learning Framework (CPU and GPU) || Abel (3.5, 3.7) || January 2019 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Infrastructure/software/pytorch nlpl-pytorch/1.1.0] || PyTorch Deep Learning Framework (CPU and GPU) || Abel (3.5, 3.7) || May 2019 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Infrastructure/software/spacy nlpl-spacy/2.0.12] || spaCy: Natural Language Processing in Python || Abel, Taito || October 2018 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Infrastructure/software/python nlpl-scipy/201901] || SciPy Ecosystem of Python Add-Ons || Abel (3.5, 3.7) || January 2019 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Infrastructure/software/tensorflow nlpl-tensorflow/1.11] || TensorFlow Deep Learning Framework (CPU and GPU) || Abel, Taito || September 2018 || Stephan Oepen&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
= Activity B: Statistical and Neural Machine Translation =&lt;br /&gt;
&lt;br /&gt;
=== On Saga and Puhti ===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Module Name/Version !! Description !! System !! Install Date !! Maintainer&lt;br /&gt;
|-&lt;br /&gt;
|  [http://wiki.nlpl.eu/index.php/Translation/home#Using_the_Moses_module nlpl-moses/4.0-a89691f]  || Moses SMT system, including GIZA++, MGIZA, fast_align || Puhti, Saga || December 2019 || Yves Scherrer&lt;br /&gt;
|-&lt;br /&gt;
|  [http://wiki.nlpl.eu/index.php/Translation/home#Using_the_Efmaral_module nlpl-efmaral/0.1_20191218] || efmaral and eflomal word alignment tools      || Puhti, Saga || December 2019 || Yves Scherrer&lt;br /&gt;
|-&lt;br /&gt;
|  [http://wiki.nlpl.eu/index.php/Translation/mttools nlpl-mttools/20191218] || A collection of preprocessing and evaluation scripts for machine translation      || Puhti, Saga || December 2019 || Yves Scherrer&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-opennmt-py/1.0.0rc2/3.7 || OpenNMT Python Library || Saga || October 2019 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-opennmt-py/1.0.0 || OpenNMT Python Library || Puhti || December 2019 || Yves Scherrer&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-marian-nmt/1.8.0-eba7aed || Marian neural machine translation system      || Puhti, Saga || December 2019 || Jörg Tiedemann&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== On Abel and Taito ===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Module Name/Version !! Description !! System !! Install Date !! Maintainer&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Translation/taito_abel#Using_the_Moses_module nlpl-moses/mmt-mvp-v0.12.1-2739-gdc42bcb]      || Moses SMT system, including GIZA++, MGIZA, fast_align || Taito || July 2017 || Yves Scherrer&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Translation/taito_abel#Using_the_Moses_module nlpl-moses/4.0-65c75ff]      || Moses SMT System Release 4.0, including GIZA++, MGIZA, fast_align, SALM&amp;lt;br/&amp;gt;Some minor fixes added to existing install 2/2018.&amp;lt;br/&amp;gt; Should not break compatibility except when using tokenizer.perl for Finnish or Swedish. || Taito, Abel || November 2017 || Yves Scherrer&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Translation/taito_abel#Using_the_Efmaral_module nlpl-efmaral/0.1_2017_07_20] || efmaral and eflomal word alignment tools      || Taito || July 2017 || Yves Scherrer&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Translation/taito_abel#Using_the_Efmaral_module nlpl-efmaral/0.1_2017_11_24] || efmaral and eflomal word alignment tools      || Taito, Abel || November 2017 || Yves Scherrer&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Translation/taito_abel#Using_the_Efmaral_module nlpl-efmaral/0.1_2018_12_13/17] || efmaral and eflomal word alignment tools      || Taito, Abel || December 2018 || Yves Scherrer&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Translation/taito_abel#Using_the_HNMT_module nlpl-hnmt/1.0.1] || HNMT neural machine translation system      || Taito || March 2018 || Yves Scherrer&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Translation/opennmt-py nlpl-opennmt-py/0.2.1] || OpenNMT Python Library || Abel, Taito || September 2018 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Translation/taito_abel#Using_the_Marian_module nlpl-marian/1.2.0] || Marian neural machine translation system      || Taito || March 2018 || Yves Scherrer&lt;br /&gt;
|-&lt;br /&gt;
| marian/1.5 || Marian neural machine translation system      || Taito || June 2018 || CSC staff&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Translation/taito_abel#Using_the_mttools_module nlpl-mttools/2018_12_23] || A collection of preprocessing and evaluation scripts for machine translation      || Taito, Abel || December 2018 || Yves Scherrer&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
= Activity C: Data-Driven Parsing =&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Module Name/Version !! Description !! System !! Install Date !! Maintainer&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-corenlp/3.9.2 || Stanford CoreNLP Suite (Including All Models) || Abel || May 2019 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Parsing/dozat nlpl-dozat/201812] || Stanford Graph-Based Parser by Tim Dozat (v3) || Abel || December 2018 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Parsing/stanfordnlp nlpl-stanfordnlp/0.1.1] || Stanford NLP Neural Pipeline || Abel || February 2019 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Parsing/stanfordnlp nlpl-stanfordnlp/0.2.0] || Stanford NLP Neural Pipeline || Saga || ? || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Parsing/uuparser nlpl-uuparser/2.3.1] || Uppsala Parser || Saga,Abel || December 2019 || Sara Stymne ||&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Parsing/turboparser nlpl-turboparser/2.3.0] || TurboParser || Saga|| January 2020 || Sara Stymne ||&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Parsing/udpipe nlpl-udpipe/1.2.1-devel] || UDPipe 1.2 with Pre-Trained Models      || Saga, Puhti,Taito, Abel || November 2017 || Jörg Tiedemann&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Parsing/udpipe nlpl-udpipe_future/3.7] || UDPipe Future || Abel || June 2019 || Andrey Kutuzov&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Parsing/repp nlpl-repp/201812] || REPP Tokenizer (and Sentence Splitter) || Abel || December 2018 || Stephan Oepen&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
= Activity E: Pre-Trained Word Embeddings =&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Module Name/Version !! Description !! System !! Install Date !! Maintainer&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-gensim/3.6.0 || Topic Modeling and Word Vectors Library || Taito, Abel || October 2018 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-gensim/3.7.0 || Topic Modeling and Word Vectors Library || Abel (3.5, 3.7) || December 2018 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-gensim/3.7.3 || Topic Modeling and Word Vectors Library || Abel (3.5, 3.7) || May 2018 || Stephan Oepen&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
= Activity G: OPUS Parallel Corpus =&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Module Name/Version !! Description !! System !! Install Date !! Maintainer&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-cwb/3.4.12 || Corpus Work Bench (CWB)      || Taito, Abel || November 2017 || Jörg Tiedemann&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-opus/0.1 || Various OPUS Tools      || Taito, Abel || November 2017 || Jörg Tiedemann&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-opus/0.2 || Various OPUS Tools      || Taito, Abel || 2018 || Jörg Tiedemann&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-opus/201901 || Various OPUS Tools      || Taito, Abel || January 2019 || Jörg Tiedemann&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-uplug/0.3.8dev || UPlug Parallel Corpus Tools      || Taito, Abel || November 2017 || Jörg Tiedemann&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/home&amp;diff=919</id>
		<title>Parsing/home</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/home&amp;diff=919"/>
		<updated>2020-01-14T21:29:54Z</updated>

		<summary type="html">&lt;p&gt;Sara: /* Parsing Systems */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Background =&lt;br /&gt;
&lt;br /&gt;
An experimentation environment for data-driven dependency parsing is maintained for NLPL under the coordination of Uppsala University (UU).&lt;br /&gt;
The data is available on the Norwegian Saga cluster and on the Finnish Puhti cluster.&lt;br /&gt;
The software is available on the Norwegian Saga cluster&lt;br /&gt;
&lt;br /&gt;
Initially, software and data were commissioned on the Norwegian Abel supercluster, see [http://wiki.nlpl.eu/index.php/Parsing/abel The Abel page] for legacy information.&lt;br /&gt;
&lt;br /&gt;
= Preprocessing Tools =&lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe]&lt;br /&gt;
&lt;br /&gt;
Additionally, a variety of tools for sentence splitting, tokenization, lemmatization, et al.&lt;br /&gt;
are available through the NLPL installations of the&lt;br /&gt;
[http://nltk.org Natural Language Processing Toolkit (NLTK)] and the&lt;br /&gt;
[https://spacy.io spaCy: Natural Language Processing in Python] tools.&lt;br /&gt;
&lt;br /&gt;
= Parsing Systems =&lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/uuparser The Uppsala Parser]&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe] &lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/turboparser TurboParser]&lt;br /&gt;
* Additional parsers: StanfordNLP, NLTK, spaCy&lt;br /&gt;
&lt;br /&gt;
= Training and Evaluation Data = &lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/ud Universal Dependencies v2.0–2.5]&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/sdp Semantic Dependency Parsing]&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/udpipe&amp;diff=917</id>
		<title>Parsing/udpipe</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/udpipe&amp;diff=917"/>
		<updated>2020-01-14T21:03:18Z</updated>

		<summary type="html">&lt;p&gt;Sara: /* Running UDPipe */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= UDPipe =&lt;br /&gt;
&lt;br /&gt;
UDPipe is an end-to-end system for morphosyntactic parsing in the UD framework developed by Milan Straka. It was used as the baseline system in the CoNLL shared tasks on universal dependency parsing in 2017 and 2018. This page only contains a brief introduction, for the full capability of UDPipe, please see the official  [http://ufal.mff.cuni.cz/udpipe UDPipe Web page] including the [http://ufal.mff.cuni.cz/udpipe/users-manual UDPipe User's Manual] and&lt;br /&gt;
&lt;br /&gt;
== Using UDPipe on Puhti and Saga ==&lt;br /&gt;
&lt;br /&gt;
UDPipe is available as a module on Puhti and Saga. It was installed as part of the OPUS activity.&lt;br /&gt;
&lt;br /&gt;
How to use UDPipe on Saga:&lt;br /&gt;
* Log into Saga&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /custer/shared/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the udpipe module:&lt;br /&gt;
 module load nlpl-udpipe/1.2.1-devel&lt;br /&gt;
&lt;br /&gt;
How to use UDPipe on Puhti:&lt;br /&gt;
* Log into Puhti&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /projappl/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the udpipe module:&lt;br /&gt;
 module load nlpl-udpipe/1.2.1-devel&lt;br /&gt;
&lt;br /&gt;
== Pre-trained models ==&lt;br /&gt;
&lt;br /&gt;
There are pre-trained models available for all languages in Universal Dependencies 2.4:&lt;br /&gt;
 /cluster/shared/nlpl/software/modules/udpipe/latest/models (Saga)&lt;br /&gt;
 /projappl/nlpl/software/modules/udpipe/latest/models (Puhti)&lt;br /&gt;
&lt;br /&gt;
== Running UDPipe ==&lt;br /&gt;
&lt;br /&gt;
To run UDPipe on raw text (in TESTFILE) run the following command:&lt;br /&gt;
 udpipe --tokenize --tag --parse MODEL_DIR/MODEL TESTFILE&lt;br /&gt;
&lt;br /&gt;
where MODEL_DIR is specified above, and MODEL is the model for the language (treebank) in question, e.g. swedich_talbanken-ud-2.4-190531.udpipe, for Swedish, tained on Talbanken.&lt;br /&gt;
&lt;br /&gt;
The above command performs segmentation, tokenization, POS-tagging and parsing. If only a subset of these tasks are needed, remove some of the flags. The output file is in CoNLLU format as default.&lt;br /&gt;
&lt;br /&gt;
If you want to parse with uuparser, run the following command (--tag is optional, depending on if you want POS-tags). Then you can run uuparser on the resulting file, for parsing.&lt;br /&gt;
 udpipe --tokenize --tag MODEL_DIR/MODEL TESTFILE&lt;br /&gt;
&lt;br /&gt;
UDPipe can handle several input and output formats and other variations. It is also possible to train new models. See the official manual for more details.&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/udpipe&amp;diff=915</id>
		<title>Parsing/udpipe</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/udpipe&amp;diff=915"/>
		<updated>2020-01-14T20:51:30Z</updated>

		<summary type="html">&lt;p&gt;Sara: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= UDPipe =&lt;br /&gt;
&lt;br /&gt;
UDPipe is an end-to-end system for morphosyntactic parsing in the UD framework developed by Milan Straka. It was used as the baseline system in the CoNLL shared tasks on universal dependency parsing in 2017 and 2018. This page only contains a brief introduction, for the full capability of UDPipe, please see the official  [http://ufal.mff.cuni.cz/udpipe UDPipe Web page] including the [http://ufal.mff.cuni.cz/udpipe/users-manual UDPipe User's Manual] and&lt;br /&gt;
&lt;br /&gt;
== Using UDPipe on Puhti and Saga ==&lt;br /&gt;
&lt;br /&gt;
UDPipe is available as a module on Puhti and Saga. It was installed as part of the OPUS activity.&lt;br /&gt;
&lt;br /&gt;
How to use UDPipe on Saga:&lt;br /&gt;
* Log into Saga&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /custer/shared/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the udpipe module:&lt;br /&gt;
 module load nlpl-udpipe/1.2.1-devel&lt;br /&gt;
&lt;br /&gt;
How to use UDPipe on Puhti:&lt;br /&gt;
* Log into Puhti&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /projappl/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the udpipe module:&lt;br /&gt;
 module load nlpl-udpipe/1.2.1-devel&lt;br /&gt;
&lt;br /&gt;
== Pre-trained models ==&lt;br /&gt;
&lt;br /&gt;
There are pre-trained models available for all languages in Universal Dependencies 2.4:&lt;br /&gt;
 /cluster/shared/nlpl/software/modules/udpipe/latest/models (Saga)&lt;br /&gt;
 /projappl/nlpl/software/modules/udpipe/latest/models (Puhti)&lt;br /&gt;
&lt;br /&gt;
== Running UDPipe ==&lt;br /&gt;
&lt;br /&gt;
To run UDPipe on raw text (in TESTFILE) run the following command:&lt;br /&gt;
 udpipe --tokenize --tag --parse MODEL_DIR/MODEL TESTFILE&lt;br /&gt;
&lt;br /&gt;
where MODEL_DIR is specified above, and MODEL is the model for the language (treebank) in question, e.g. swedich_talbanken-ud-2.4-190531.udpipe, for Swedish, tained on Talbanken.&lt;br /&gt;
&lt;br /&gt;
The above command performs segmentation, tokenization, POS-tagging and parsing. If only a subset of these tasks are needed, remove some of the flags. The output file is in CoNLLU format as default.&lt;br /&gt;
&lt;br /&gt;
UDPipe can handle several input and output formats and other variations. It is also possible to train new models. See the official manual for more details.&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=914</id>
		<title>Parsing/uuparser</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=914"/>
		<updated>2020-01-14T20:41:08Z</updated>

		<summary type="html">&lt;p&gt;Sara: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= The Uppsala Parser =&lt;br /&gt;
&lt;br /&gt;
The Uppsala Parser is a neural transition-based dependency parser based on bist-parser by Eli Kiperwasser and Yoav Goldberg and developed primarily in the context of the CoNLL shared tasks on universal dependency parsing in 2017 and 2018. The Uppsala Parser is publicly available at https://github.com/UppsalaNLP/uuparser. &lt;br /&gt;
Note that the version installed here may exhibit some slight differences, designed to improve ease of use.&lt;br /&gt;
&lt;br /&gt;
For the full documentation, see the github page.&lt;br /&gt;
&lt;br /&gt;
== Using the Uppsala Parser on Saga ==&lt;br /&gt;
&lt;br /&gt;
* Log into Saga&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /cluster/shared/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load nlpl-uuparser/2.3.1&lt;br /&gt;
&lt;br /&gt;
== Training a parsing model ==&lt;br /&gt;
&lt;br /&gt;
To train a set of parsing models on treebanks from Universal Dependencies (v2.2 or later):&lt;br /&gt;
&lt;br /&gt;
 uuparser --include [languages to include denoted by their treebank id] --outdir my-output-directory  --datadir ud-treebank-dir&lt;br /&gt;
&lt;br /&gt;
For example:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments  --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
will train separate models on UD Swedish-Talbanken, UD English-ParTUT and UD Russian-SynTagRus and store the results in the &amp;lt;code&amp;gt;experiments&amp;lt;/code&amp;gt; folder in your home directory. Model selection is included in the training process by default; that is, at each epoch the current model is evaluated on the UD dev data, and at the end of training the best performing model for each language is selected.&lt;br /&gt;
&lt;br /&gt;
== Predicting with a pre-trained parsing model ==&lt;br /&gt;
&lt;br /&gt;
To predict on UD test data with the models trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
To predict on other texts, prepare the data into the CoNLLU format (see below) and run the following command, assuming the model trained above, for the file TESTFILE:&lt;br /&gt;
&lt;br /&gt;
 uuparser --testfile TESTFILE --outdir ~/experiments --modeldir ~/experiments/sv_talbanken --predict&lt;br /&gt;
&lt;br /&gt;
== Options ==&lt;br /&gt;
&lt;br /&gt;
The parser has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 uuparser --help | less&lt;br /&gt;
&lt;br /&gt;
We recommend you set the &amp;lt;code&amp;gt;--dynet-mem&amp;lt;/code&amp;gt; option to a larger number when running the full training procedure on larger treebanks. Commonly used values are 5000 and 10000 (in MB). Dynet is the neural network library on which the parser is built.&lt;br /&gt;
&lt;br /&gt;
Note that due to random initialization and other non-deterministic elements in the training process, you will not obtain the same results even when training twice under exactly the same circumstances (e.g. languages, number of epochs etc.). To ensure identical results between two runs, we recommend setting the &amp;lt;code&amp;gt;--dynet-seed&amp;lt;/code&amp;gt; option to the same value both times (e.g. &amp;lt;code&amp;gt; --dynet-seed 123456789&amp;lt;/code&amp;gt;) and adding the &amp;lt;code&amp;gt;--use-default-seed&amp;lt;/code&amp;gt; flag. This ensures that Python's random number generator and Dynet both produce the same sequence of random numbers.&lt;br /&gt;
&lt;br /&gt;
== Training a multitreebank parsing model ==&lt;br /&gt;
&lt;br /&gt;
Our parser supports multitreebank parsing, that is a single parsing model for one or more treebanks, which could be from the same or form different languages. To train a ''multitreebank'' model for the three languages in the examples above, we simply add the &amp;lt;code&amp;gt;--multiling&amp;lt;/code&amp;gt; flag when training (note, though, that these models have so far been found to work best for groups of related languages)&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
In this case, instead of creating three separate models in the language-specific subdirectories within ~/experiments, a single model will be created directly in this folder. Predicting on UD test data is then:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Note that if you want to have different output directories for training and predicting, the &amp;lt;code&amp;gt;--modeldir&amp;lt;/code&amp;gt; option can be specified when predicting to tell the parser where the pre-trained model can be found.&lt;br /&gt;
&lt;br /&gt;
If you want to use the parser for a language or treebank that was not among the treebanks that the parser was trained for, this can be done in two ways:&lt;br /&gt;
 &lt;br /&gt;
 uuparser --include &amp;quot;sv_pud:sv_talbanken&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_pud&amp;quot; --forced-tbank-emb sv_talbanken --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Where in both cases, sv_pud will be parsed using sv_talbanken as a proxy (i.e. parsing sv_pud as if it is is sv_talbanken). With the first option, multiple treebanks and proxies can be defined separated by space.&lt;br /&gt;
&lt;br /&gt;
To predict on non-UD texts, the command is similar to the single treebank case, the mian difference being the use of the --multiling flag, and specifying a proxy treebank:&lt;br /&gt;
&lt;br /&gt;
 uuparser --testfile TESTFILE --outdir ~/experiments --modeldir MODELDIR --predict --multiling --forced-tbank-emb sv_talbanken&lt;br /&gt;
&lt;br /&gt;
This commands runs the parser on TESTFILE (in CoNLLU format), using a model found in MODELDIR, and prints the resulting parse in a file in ~/experiments. In the example sv_talbanken is used as the proxy treebank.&lt;br /&gt;
&lt;br /&gt;
== Pre-trained models ==&lt;br /&gt;
&lt;br /&gt;
There are three multilingual pre-trained models available, covering English and most of the Nordic languages. The models can be found in /cluster/shared/nlpl/software/modules/uuparser/2.3.1/models ($MODEL_DIR)&lt;br /&gt;
&lt;br /&gt;
The English model is found in the subdirectory $MODEL_DIR/en. It is trained on the four English UD treebanks: en_gum en_partut en_ewt en_lines. For English web texts it is recommended to use en_ewt as a proxy. For more formal texts, en_partut typically works well.&lt;br /&gt;
&lt;br /&gt;
The Scandinavian model is found in the subdirectory $MODEL_DIR/scandinavian. It is trained on six treebanks in Danish, Norwegian and Swedish: sv_talbanken sv_lines no_bokmaal no_nynorsk no_nynorsklia da_ddt. For Swedish, using sv_talbanken typically gives the best results, sv_lines might be better for fiction. For Norwegian, use no_bokmaal for Bokmål, no_nynorsk for general Nynorsk and no_nynorsklia for spoken Nynorsk. We have also had reasonable results when parsing Faroese with no_nynorsk as proxy.&lt;br /&gt;
&lt;br /&gt;
The Uralic model is found in the subdirectory $MODEL_DIR/uralic. It is trained on five treebanks in Estonian, Finnish and North Sàmi: fi_ftb fi_tdt et_edt et_ewt sme_giella. For Finnish we recommend using fi_tdt as a proxy. For Estonian et_edt is probably most useful for general texts and et_ewt might be good for web texts.&lt;br /&gt;
&lt;br /&gt;
To run with any of these models, use the same commands as above, exemplified with the Scandinavian model and no_nynorsk as a proxy treebank.&lt;br /&gt;
&lt;br /&gt;
 uuparser --testfile TESTFILE --outdir ~/experiments --modeldir MODEL_DIR/scandinavian --predict --multiling --forced-tbank-emb no_nynorsk&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== POS-tags ==&lt;br /&gt;
&lt;br /&gt;
The uuparser gives good results without POS-tags, mainly thanks to the use of character embeddings, and this is the default setting. It can be used with predicted POS-tags. If you wish to do this, we recommend using UDPipe to predict tags. To activate a POS-tag embedding, use the flag --pos-emb-size N, where N is the size of the embedding (12 has been a useful value).&lt;br /&gt;
&lt;br /&gt;
== Segmentation ==&lt;br /&gt;
&lt;br /&gt;
In the above examples, we assume pre-segmented input data already in the [http://universaldependencies.org/format.html CONLL-U] format. If your input is raw text, we recommend using UDPipe to segment and convert the format first. The UDPipe module can be loaded using &amp;lt;code&amp;gt;module load nlpl-udpipe&amp;lt;/code&amp;gt; and then run by typing &amp;lt;code&amp;gt;udpipe&amp;lt;/code&amp;gt; at the command line, see [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe].&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/udpipe&amp;diff=912</id>
		<title>Parsing/udpipe</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/udpipe&amp;diff=912"/>
		<updated>2020-01-14T20:23:59Z</updated>

		<summary type="html">&lt;p&gt;Sara: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= UDPipe =&lt;br /&gt;
&lt;br /&gt;
UDPipe is an end-to-end system for morphosyntactic parsing in the UD framework developed by Milan Straka. It was used as the baseline system in the CoNLL shared tasks on universal dependency parsing in 2017 and 2018. This page only contains a brief introduction, for the full capability of UDPipe, please see the official  [http://ufal.mff.cuni.cz/udpipe UDPipe Web page] including the [http://ufal.mff.cuni.cz/udpipe/users-manual UDPipe User's Manual] and&lt;br /&gt;
&lt;br /&gt;
== Using UDPipe on Puhti and Saga ==&lt;br /&gt;
&lt;br /&gt;
UDPipe is available as a module on Puhti and Saga. It was installed as part of the OPUS activity.&lt;br /&gt;
&lt;br /&gt;
How to use UDPipe on Saga:&lt;br /&gt;
* Log into Saga&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /custer/shared/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the udpipe module:&lt;br /&gt;
 module load nlpl-udpipe/1.2.1-devel&lt;br /&gt;
&lt;br /&gt;
How to use UDPipe on Puhti:&lt;br /&gt;
* Log into Puhti&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /projappl/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the udpipe module:&lt;br /&gt;
 module load nlpl-udpipe/1.2.1-devel&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=911</id>
		<title>Parsing/uuparser</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=911"/>
		<updated>2020-01-14T20:16:50Z</updated>

		<summary type="html">&lt;p&gt;Sara: /* Using the Uppsala Parser on Saga */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= The Uppsala Parser =&lt;br /&gt;
&lt;br /&gt;
The Uppsala Parser is a neural transition-based dependency parser based on bist-parser by Eli Kiperwasser and Yoav Goldberg and developed primarily in the context of the CoNLL shared tasks on universal dependency parsing in 2017 and 2018. The Uppsala Parser is publicly available at https://github.com/UppsalaNLP/uuparser. &lt;br /&gt;
Note that the version installed here may exhibit some slight differences, designed to improve ease of use.&lt;br /&gt;
&lt;br /&gt;
For the full documentation, see the github page.&lt;br /&gt;
&lt;br /&gt;
== Using the Uppsala Parser on Saga ==&lt;br /&gt;
&lt;br /&gt;
* Log into Saga&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /cluster/shared/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load nlpl-uuparser/2.3.1&lt;br /&gt;
&lt;br /&gt;
== Training a parsing model ==&lt;br /&gt;
&lt;br /&gt;
To train a set of parsing models on treebanks from Universal Dependencies (v2.2 or later):&lt;br /&gt;
&lt;br /&gt;
 uuparser --include [languages to include denoted by their treebank id] --outdir my-output-directory  --datadir ud-treebank-dir&lt;br /&gt;
&lt;br /&gt;
For example:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments  --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
will train separate models on UD Swedish-Talbanken, UD English-ParTUT and UD Russian-SynTagRus and store the results in the &amp;lt;code&amp;gt;experiments&amp;lt;/code&amp;gt; folder in your home directory. Model selection is included in the training process by default; that is, at each epoch the current model is evaluated on the UD dev data, and at the end of training the best performing model for each language is selected.&lt;br /&gt;
&lt;br /&gt;
== Predicting with a pre-trained parsing model ==&lt;br /&gt;
&lt;br /&gt;
To predict on UD test data with the models trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
To predict on other texts, prepare the data into the CoNLLU format (see below) and run the following command, assuming the model trained above, for the file TESTFILE:&lt;br /&gt;
&lt;br /&gt;
 uuparser --testfile TESTFILE --outdir ~/experiments --modeldir ~/experiments/sv_talbanken --predict&lt;br /&gt;
&lt;br /&gt;
== Options ==&lt;br /&gt;
&lt;br /&gt;
The parser has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 uuparser --help | less&lt;br /&gt;
&lt;br /&gt;
We recommend you set the &amp;lt;code&amp;gt;--dynet-mem&amp;lt;/code&amp;gt; option to a larger number when running the full training procedure on larger treebanks. Commonly used values are 5000 and 10000 (in MB). Dynet is the neural network library on which the parser is built.&lt;br /&gt;
&lt;br /&gt;
Note that due to random initialization and other non-deterministic elements in the training process, you will not obtain the same results even when training twice under exactly the same circumstances (e.g. languages, number of epochs etc.). To ensure identical results between two runs, we recommend setting the &amp;lt;code&amp;gt;--dynet-seed&amp;lt;/code&amp;gt; option to the same value both times (e.g. &amp;lt;code&amp;gt; --dynet-seed 123456789&amp;lt;/code&amp;gt;) and adding the &amp;lt;code&amp;gt;--use-default-seed&amp;lt;/code&amp;gt; flag. This ensures that Python's random number generator and Dynet both produce the same sequence of random numbers.&lt;br /&gt;
&lt;br /&gt;
== Training a multitreebank parsing model ==&lt;br /&gt;
&lt;br /&gt;
Our parser supports multitreebank parsing, that is a single parsing model for one or more treebanks, which could be from the same or form different languages. To train a ''multitreebank'' model for the three languages in the examples above, we simply add the &amp;lt;code&amp;gt;--multiling&amp;lt;/code&amp;gt; flag when training (note, though, that these models have so far been found to work best for groups of related languages)&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
In this case, instead of creating three separate models in the language-specific subdirectories within ~/experiments, a single model will be created directly in this folder. Predicting on UD test data is then:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Note that if you want to have different output directories for training and predicting, the &amp;lt;code&amp;gt;--modeldir&amp;lt;/code&amp;gt; option can be specified when predicting to tell the parser where the pre-trained model can be found.&lt;br /&gt;
&lt;br /&gt;
If you want to use the parser for a language or treebank that was not among the treebanks that the parser was trained for, this can be done in two ways:&lt;br /&gt;
 &lt;br /&gt;
 uuparser --include &amp;quot;sv_pud:sv_talbanken&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_pud&amp;quot; --forced-tbank-emb sv_talbanken --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Where in both cases, sv_pud will be parsed using sv_talbanken as a proxy (i.e. parsing sv_pud as if it is is sv_talbanken). With the first option, multiple treebanks and proxies can be defined separated by space.&lt;br /&gt;
&lt;br /&gt;
To predict on non-UD texts, the command is similar to the single treebank case, the mian difference being the use of the --multiling flag, and specifying a proxy treebank:&lt;br /&gt;
&lt;br /&gt;
 uuparser --testfile TESTFILE --outdir ~/experiments --modeldir MODELDIR --predict --multiling --forced-tbank-emb sv_talbanken&lt;br /&gt;
&lt;br /&gt;
This commands runs the parser on TESTFILE (in CoNLLU format), using a model found in MODELDIR, and prints the resulting parse in a file in ~/experiments. In the example sv_talbanken is used as the proxy treebank.&lt;br /&gt;
&lt;br /&gt;
== Pre-trained models ==&lt;br /&gt;
&lt;br /&gt;
There are three multilingual pre-trained models available, covering English and most of the Nordic languages. The models can be found in /cluster/shared/nlpl/software/modules/uuparser/2.3.1/models ($MODEL_DIR)&lt;br /&gt;
&lt;br /&gt;
The English model is found in the subdirectory $MODEL_DIR/en. It is trained on the four English UD treebanks: en_gum en_partut en_ewt en_lines. For English web texts it is recommended to use en_ewt as a proxy. For more formal texts, en_partut typically works well.&lt;br /&gt;
&lt;br /&gt;
The Scandinavian model is found in the subdirectory $MODEL_DIR/scandinavian. It is trained on six treebanks in Danish, Norwegian and Swedish: sv_talbanken sv_lines no_bokmaal no_nynorsk no_nynorsklia da_ddt. For Swedish, using sv_talbanken typically gives the best results, sv_lines might be better for fiction. For Norwegian, use no_bokmaal for Bokmål, no_nynorsk for general Nynorsk and no_nynorsklia for spoken Nynorsk. We have also had reasonable results when parsing Faroese with no_nynorsk as proxy.&lt;br /&gt;
&lt;br /&gt;
The Uralic model is found in the subdirectory $MODEL_DIR/uralic. It is trained on five treebanks in Estonian, Finnish and North Sàmi: fi_ftb fi_tdt et_edt et_ewt sme_giella. For Finnish we recommend using fi_tdt as a proxy. For Estonian et_edt is probably most useful for general texts and et_ewt might be good for web texts.&lt;br /&gt;
&lt;br /&gt;
To run with any of these models, use the same commands as above, exemplified with the Scandinavian model and no_nynorsk as a proxy treebank.&lt;br /&gt;
&lt;br /&gt;
 uuparser --testfile TESTFILE --outdir ~/experiments --modeldir MODEL_DIR/scandinavian --predict --multiling --forced-tbank-emb no_nynorsk&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Segmentation ==&lt;br /&gt;
&lt;br /&gt;
In the above examples, we assume pre-segmented input data already in the [http://universaldependencies.org/format.html CONLL-U] format. If your input is raw text, we recommend using UDPipe to segment and convert the format first. The UDPipe module can be loaded using &amp;lt;code&amp;gt;module load nlpl-udpipe&amp;lt;/code&amp;gt; and then run by typing &amp;lt;code&amp;gt;udpipe&amp;lt;/code&amp;gt; at the command line, see [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe].&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=910</id>
		<title>Parsing/uuparser</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=910"/>
		<updated>2020-01-14T20:09:25Z</updated>

		<summary type="html">&lt;p&gt;Sara: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= The Uppsala Parser =&lt;br /&gt;
&lt;br /&gt;
The Uppsala Parser is a neural transition-based dependency parser based on bist-parser by Eli Kiperwasser and Yoav Goldberg and developed primarily in the context of the CoNLL shared tasks on universal dependency parsing in 2017 and 2018. The Uppsala Parser is publicly available at https://github.com/UppsalaNLP/uuparser. &lt;br /&gt;
Note that the version installed here may exhibit some slight differences, designed to improve ease of use.&lt;br /&gt;
&lt;br /&gt;
For the full documentation, see the github page.&lt;br /&gt;
&lt;br /&gt;
== Using the Uppsala Parser on Saga ==&lt;br /&gt;
&lt;br /&gt;
* Log into Saga&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /cluster/shared/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load nlpl-uuparser/2.3.0&lt;br /&gt;
&lt;br /&gt;
== Training a parsing model ==&lt;br /&gt;
&lt;br /&gt;
To train a set of parsing models on treebanks from Universal Dependencies (v2.2 or later):&lt;br /&gt;
&lt;br /&gt;
 uuparser --include [languages to include denoted by their treebank id] --outdir my-output-directory  --datadir ud-treebank-dir&lt;br /&gt;
&lt;br /&gt;
For example:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments  --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
will train separate models on UD Swedish-Talbanken, UD English-ParTUT and UD Russian-SynTagRus and store the results in the &amp;lt;code&amp;gt;experiments&amp;lt;/code&amp;gt; folder in your home directory. Model selection is included in the training process by default; that is, at each epoch the current model is evaluated on the UD dev data, and at the end of training the best performing model for each language is selected.&lt;br /&gt;
&lt;br /&gt;
== Predicting with a pre-trained parsing model ==&lt;br /&gt;
&lt;br /&gt;
To predict on UD test data with the models trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
To predict on other texts, prepare the data into the CoNLLU format (see below) and run the following command, assuming the model trained above, for the file TESTFILE:&lt;br /&gt;
&lt;br /&gt;
 uuparser --testfile TESTFILE --outdir ~/experiments --modeldir ~/experiments/sv_talbanken --predict&lt;br /&gt;
&lt;br /&gt;
== Options ==&lt;br /&gt;
&lt;br /&gt;
The parser has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 uuparser --help | less&lt;br /&gt;
&lt;br /&gt;
We recommend you set the &amp;lt;code&amp;gt;--dynet-mem&amp;lt;/code&amp;gt; option to a larger number when running the full training procedure on larger treebanks. Commonly used values are 5000 and 10000 (in MB). Dynet is the neural network library on which the parser is built.&lt;br /&gt;
&lt;br /&gt;
Note that due to random initialization and other non-deterministic elements in the training process, you will not obtain the same results even when training twice under exactly the same circumstances (e.g. languages, number of epochs etc.). To ensure identical results between two runs, we recommend setting the &amp;lt;code&amp;gt;--dynet-seed&amp;lt;/code&amp;gt; option to the same value both times (e.g. &amp;lt;code&amp;gt; --dynet-seed 123456789&amp;lt;/code&amp;gt;) and adding the &amp;lt;code&amp;gt;--use-default-seed&amp;lt;/code&amp;gt; flag. This ensures that Python's random number generator and Dynet both produce the same sequence of random numbers.&lt;br /&gt;
&lt;br /&gt;
== Training a multitreebank parsing model ==&lt;br /&gt;
&lt;br /&gt;
Our parser supports multitreebank parsing, that is a single parsing model for one or more treebanks, which could be from the same or form different languages. To train a ''multitreebank'' model for the three languages in the examples above, we simply add the &amp;lt;code&amp;gt;--multiling&amp;lt;/code&amp;gt; flag when training (note, though, that these models have so far been found to work best for groups of related languages)&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
In this case, instead of creating three separate models in the language-specific subdirectories within ~/experiments, a single model will be created directly in this folder. Predicting on UD test data is then:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Note that if you want to have different output directories for training and predicting, the &amp;lt;code&amp;gt;--modeldir&amp;lt;/code&amp;gt; option can be specified when predicting to tell the parser where the pre-trained model can be found.&lt;br /&gt;
&lt;br /&gt;
If you want to use the parser for a language or treebank that was not among the treebanks that the parser was trained for, this can be done in two ways:&lt;br /&gt;
 &lt;br /&gt;
 uuparser --include &amp;quot;sv_pud:sv_talbanken&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_pud&amp;quot; --forced-tbank-emb sv_talbanken --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Where in both cases, sv_pud will be parsed using sv_talbanken as a proxy (i.e. parsing sv_pud as if it is is sv_talbanken). With the first option, multiple treebanks and proxies can be defined separated by space.&lt;br /&gt;
&lt;br /&gt;
To predict on non-UD texts, the command is similar to the single treebank case, the mian difference being the use of the --multiling flag, and specifying a proxy treebank:&lt;br /&gt;
&lt;br /&gt;
 uuparser --testfile TESTFILE --outdir ~/experiments --modeldir MODELDIR --predict --multiling --forced-tbank-emb sv_talbanken&lt;br /&gt;
&lt;br /&gt;
This commands runs the parser on TESTFILE (in CoNLLU format), using a model found in MODELDIR, and prints the resulting parse in a file in ~/experiments. In the example sv_talbanken is used as the proxy treebank.&lt;br /&gt;
&lt;br /&gt;
== Pre-trained models ==&lt;br /&gt;
&lt;br /&gt;
There are three multilingual pre-trained models available, covering English and most of the Nordic languages. The models can be found in /cluster/shared/nlpl/software/modules/uuparser/2.3.1/models ($MODEL_DIR)&lt;br /&gt;
&lt;br /&gt;
The English model is found in the subdirectory $MODEL_DIR/en. It is trained on the four English UD treebanks: en_gum en_partut en_ewt en_lines. For English web texts it is recommended to use en_ewt as a proxy. For more formal texts, en_partut typically works well.&lt;br /&gt;
&lt;br /&gt;
The Scandinavian model is found in the subdirectory $MODEL_DIR/scandinavian. It is trained on six treebanks in Danish, Norwegian and Swedish: sv_talbanken sv_lines no_bokmaal no_nynorsk no_nynorsklia da_ddt. For Swedish, using sv_talbanken typically gives the best results, sv_lines might be better for fiction. For Norwegian, use no_bokmaal for Bokmål, no_nynorsk for general Nynorsk and no_nynorsklia for spoken Nynorsk. We have also had reasonable results when parsing Faroese with no_nynorsk as proxy.&lt;br /&gt;
&lt;br /&gt;
The Uralic model is found in the subdirectory $MODEL_DIR/uralic. It is trained on five treebanks in Estonian, Finnish and North Sàmi: fi_ftb fi_tdt et_edt et_ewt sme_giella. For Finnish we recommend using fi_tdt as a proxy. For Estonian et_edt is probably most useful for general texts and et_ewt might be good for web texts.&lt;br /&gt;
&lt;br /&gt;
To run with any of these models, use the same commands as above, exemplified with the Scandinavian model and no_nynorsk as a proxy treebank.&lt;br /&gt;
&lt;br /&gt;
 uuparser --testfile TESTFILE --outdir ~/experiments --modeldir MODEL_DIR/scandinavian --predict --multiling --forced-tbank-emb no_nynorsk&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Segmentation ==&lt;br /&gt;
&lt;br /&gt;
In the above examples, we assume pre-segmented input data already in the [http://universaldependencies.org/format.html CONLL-U] format. If your input is raw text, we recommend using UDPipe to segment and convert the format first. The UDPipe module can be loaded using &amp;lt;code&amp;gt;module load nlpl-udpipe&amp;lt;/code&amp;gt; and then run by typing &amp;lt;code&amp;gt;udpipe&amp;lt;/code&amp;gt; at the command line, see [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe].&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=909</id>
		<title>Parsing/uuparser</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=909"/>
		<updated>2020-01-14T20:08:55Z</updated>

		<summary type="html">&lt;p&gt;Sara: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= The Uppsala Parser =&lt;br /&gt;
&lt;br /&gt;
The Uppsala Parser is a neural transition-based dependency parser based on bist-parser by Eli Kiperwasser and Yoav Goldberg and developed primarily in the context of the CoNLL shared tasks on universal dependency parsing in 2017 and 2018. The Uppsala Parser is publicly available at https://github.com/UppsalaNLP/uuparser. &lt;br /&gt;
Note that the version installed here may exhibit some slight differences, designed to improve ease of use.&lt;br /&gt;
&lt;br /&gt;
For the full documentation, see the github page.&lt;br /&gt;
&lt;br /&gt;
== Using the Uppsala Parser on Saga ==&lt;br /&gt;
&lt;br /&gt;
* Log into Saga&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /cluster/shared/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load nlpl-uuparser/2.3.0&lt;br /&gt;
&lt;br /&gt;
== Training a parsing model ==&lt;br /&gt;
&lt;br /&gt;
To train a set of parsing models on treebanks from Universal Dependencies (v2.2 or later):&lt;br /&gt;
&lt;br /&gt;
 uuparser --include [languages to include denoted by their treebank id] --outdir my-output-directory  --datadir ud-treebank-dir&lt;br /&gt;
&lt;br /&gt;
For example:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments  --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
will train separate models on UD Swedish-Talbanken, UD English-ParTUT and UD Russian-SynTagRus and store the results in the &amp;lt;code&amp;gt;experiments&amp;lt;/code&amp;gt; folder in your home directory. Model selection is included in the training process by default; that is, at each epoch the current model is evaluated on the UD dev data, and at the end of training the best performing model for each language is selected.&lt;br /&gt;
&lt;br /&gt;
== Predicting with a pre-trained parsing model ==&lt;br /&gt;
&lt;br /&gt;
To predict on UD test data with the models trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
To predict on other texts, prepare the data into the CoNLLU format (see below) and run the following command, assuming the model trained above, for the file TESTFILE:&lt;br /&gt;
&lt;br /&gt;
 uuparser --testfile TESTFILE --outdir ~/experiments --modeldir ~/experiments/sv_talbanken --predict&lt;br /&gt;
&lt;br /&gt;
== Options ==&lt;br /&gt;
&lt;br /&gt;
The parser has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 uuparser --help | less&lt;br /&gt;
&lt;br /&gt;
We recommend you set the &amp;lt;code&amp;gt;--dynet-mem&amp;lt;/code&amp;gt; option to a larger number when running the full training procedure on larger treebanks. Commonly used values are 5000 and 10000 (in MB). Dynet is the neural network library on which the parser is built.&lt;br /&gt;
&lt;br /&gt;
Note that due to random initialization and other non-deterministic elements in the training process, you will not obtain the same results even when training twice under exactly the same circumstances (e.g. languages, number of epochs etc.). To ensure identical results between two runs, we recommend setting the &amp;lt;code&amp;gt;--dynet-seed&amp;lt;/code&amp;gt; option to the same value both times (e.g. &amp;lt;code&amp;gt; --dynet-seed 123456789&amp;lt;/code&amp;gt;) and adding the &amp;lt;code&amp;gt;--use-default-seed&amp;lt;/code&amp;gt; flag. This ensures that Python's random number generator and Dynet both produce the same sequence of random numbers.&lt;br /&gt;
&lt;br /&gt;
== Training a multitreebank parsing model ==&lt;br /&gt;
&lt;br /&gt;
Our parser supports multitreebank parsing, that is a single parsing model for one or more treebanks, which could be from the same or form different languages. To train a ''multitreebank'' model for the three languages in the examples above, we simply add the &amp;lt;code&amp;gt;--multiling&amp;lt;/code&amp;gt; flag when training (note, though, that these models have so far been found to work best for groups of related languages)&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
In this case, instead of creating three separate models in the language-specific subdirectories within ~/experiments, a single model will be created directly in this folder. Predicting on UD test data is then:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Note that if you want to have different output directories for training and predicting, the &amp;lt;code&amp;gt;--modeldir&amp;lt;/code&amp;gt; option can be specified when predicting to tell the parser where the pre-trained model can be found.&lt;br /&gt;
&lt;br /&gt;
If you want to use the parser for a language or treebank that was not among the treebanks that the parser was trained for, this can be done in two ways:&lt;br /&gt;
 &lt;br /&gt;
 uuparser --include &amp;quot;sv_pud:sv_talbanken&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_pud&amp;quot; --forced-tbank-emb sv_talbanken --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Where in both cases, sv_pud will be parsed using sv_talbanken as a proxy (i.e. parsing sv_pud as if it is is sv_talbanken). With the first option, multiple treebanks and proxies can be defined separated by space.&lt;br /&gt;
&lt;br /&gt;
To predict on non-UD texts, the command is similar to the single treebank case, the mian difference being the use of the --multiling flag, and specifying a proxy treebank:&lt;br /&gt;
&lt;br /&gt;
 uuparser --testfile TESTFILE --outdir ~/experiments --modeldir MODELDIR --predict --multiling --forced-tbank-emb sv_talbanken&lt;br /&gt;
&lt;br /&gt;
This commands runs the parser on TESTFILE (in CoNLLU format), using a model found in MODELDIR, and prints the resulting parse in a file in ~/experiments. In the example sv_talbanken is used as the proxy treebank.&lt;br /&gt;
&lt;br /&gt;
== Pre-trained models&lt;br /&gt;
&lt;br /&gt;
There are three multilingual pre-trained models available, covering English and most of the Nordic languages. The models can be found in /cluster/shared/nlpl/software/modules/uuparser/2.3.1/models ($MODEL_DIR)&lt;br /&gt;
&lt;br /&gt;
The English model is found in the subdirectory $MODEL_DIR/en. It is trained on the four English UD treebanks: en_gum en_partut en_ewt en_lines. For English web texts it is recommended to use en_ewt as a proxy. For more formal texts, en_partut typically works well.&lt;br /&gt;
&lt;br /&gt;
The Scandinavian model is found in the subdirectory $MODEL_DIR/scandinavian. It is trained on six treebanks in Danish, Norwegian and Swedish: sv_talbanken sv_lines no_bokmaal no_nynorsk no_nynorsklia da_ddt. For Swedish, using sv_talbanken typically gives the best results, sv_lines might be better for fiction. For Norwegian, use no_bokmaal for Bokmål, no_nynorsk for general Nynorsk and no_nynorsklia for spoken Nynorsk. We have also had reasonable results when parsing Faroese with no_nynorsk as proxy.&lt;br /&gt;
&lt;br /&gt;
The Uralic model is found in the subdirectory $MODEL_DIR/uralic. It is trained on five treebanks in Estonian, Finnish and North Sàmi: fi_ftb fi_tdt et_edt et_ewt sme_giella. For Finnish we recommend using fi_tdt as a proxy. For Estonian et_edt is probably most useful for general texts and et_ewt might be good for web texts.&lt;br /&gt;
&lt;br /&gt;
To run with any of these models, use the same commands as above, exemplified with the Scandinavian model and no_nynorsk as a proxy treebank.&lt;br /&gt;
&lt;br /&gt;
 uuparser --testfile TESTFILE --outdir ~/experiments --modeldir MODEL_DIR/scandinavian --predict --multiling --forced-tbank-emb no_nynorsk&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Segmentation ==&lt;br /&gt;
&lt;br /&gt;
In the above examples, we assume pre-segmented input data already in the [http://universaldependencies.org/format.html CONLL-U] format. If your input is raw text, we recommend using UDPipe to segment and convert the format first. The UDPipe module can be loaded using &amp;lt;code&amp;gt;module load nlpl-udpipe&amp;lt;/code&amp;gt; and then run by typing &amp;lt;code&amp;gt;udpipe&amp;lt;/code&amp;gt; at the command line, see [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe].&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=908</id>
		<title>Parsing/uuparser</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=908"/>
		<updated>2020-01-14T19:58:17Z</updated>

		<summary type="html">&lt;p&gt;Sara: /* Segmentation */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= The Uppsala Parser =&lt;br /&gt;
&lt;br /&gt;
The Uppsala Parser is a neural transition-based dependency parser based on bist-parser by Eli Kiperwasser and Yoav Goldberg and developed primarily in the context of the CoNLL shared tasks on universal dependency parsing in 2017 and 2018. The Uppsala Parser is publicly available at https://github.com/UppsalaNLP/uuparser. &lt;br /&gt;
Note that the version installed here may exhibit some slight differences, designed to improve ease of use.&lt;br /&gt;
&lt;br /&gt;
For the full documentation, see the github page.&lt;br /&gt;
&lt;br /&gt;
== Using the Uppsala Parser on Saga ==&lt;br /&gt;
&lt;br /&gt;
* Log into Saga&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /cluster/shared/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load nlpl-uuparser/2.3.0&lt;br /&gt;
&lt;br /&gt;
== Training a parsing model ==&lt;br /&gt;
&lt;br /&gt;
To train a set of parsing models on treebanks from Universal Dependencies (v2.2 or later):&lt;br /&gt;
&lt;br /&gt;
 uuparser --include [languages to include denoted by their treebank id] --outdir my-output-directory  --datadir ud-treebank-dir&lt;br /&gt;
&lt;br /&gt;
For example:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments  --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
will train separate models on UD Swedish-Talbanken, UD English-ParTUT and UD Russian-SynTagRus and store the results in the &amp;lt;code&amp;gt;experiments&amp;lt;/code&amp;gt; folder in your home directory. Model selection is included in the training process by default; that is, at each epoch the current model is evaluated on the UD dev data, and at the end of training the best performing model for each language is selected.&lt;br /&gt;
&lt;br /&gt;
== Predicting with a pre-trained parsing model ==&lt;br /&gt;
&lt;br /&gt;
To predict on UD test data with the models trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
To predict on other texts, prepare the data into the CoNLLU format (see below) and run the following command, assuming the model trained above, for the file TESTFILE:&lt;br /&gt;
&lt;br /&gt;
 uuparser --testfile TESTFILE --outdir ~/experiments --modeldir ~/experiments/sv_talbanken --predict&lt;br /&gt;
&lt;br /&gt;
== Options ==&lt;br /&gt;
&lt;br /&gt;
The parser has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 uuparser --help | less&lt;br /&gt;
&lt;br /&gt;
We recommend you set the &amp;lt;code&amp;gt;--dynet-mem&amp;lt;/code&amp;gt; option to a larger number when running the full training procedure on larger treebanks. Commonly used values are 5000 and 10000 (in MB). Dynet is the neural network library on which the parser is built.&lt;br /&gt;
&lt;br /&gt;
Note that due to random initialization and other non-deterministic elements in the training process, you will not obtain the same results even when training twice under exactly the same circumstances (e.g. languages, number of epochs etc.). To ensure identical results between two runs, we recommend setting the &amp;lt;code&amp;gt;--dynet-seed&amp;lt;/code&amp;gt; option to the same value both times (e.g. &amp;lt;code&amp;gt; --dynet-seed 123456789&amp;lt;/code&amp;gt;) and adding the &amp;lt;code&amp;gt;--use-default-seed&amp;lt;/code&amp;gt; flag. This ensures that Python's random number generator and Dynet both produce the same sequence of random numbers.&lt;br /&gt;
&lt;br /&gt;
== Training a multitreebank parsing model ==&lt;br /&gt;
&lt;br /&gt;
Our parser supports multitreebank parsing, that is a single parsing model for one or more treebanks, which could be from the same or form different languages. To train a ''multitreebank'' model for the three languages in the examples above, we simply add the &amp;lt;code&amp;gt;--multiling&amp;lt;/code&amp;gt; flag when training (note, though, that these models have so far been found to work best for groups of related languages)&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
In this case, instead of creating three separate models in the language-specific subdirectories within ~/experiments, a single model will be created directly in this folder. Predicting on UD test data is then:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Note that if you want to have different output directories for training and predicting, the &amp;lt;code&amp;gt;--modeldir&amp;lt;/code&amp;gt; option can be specified when predicting to tell the parser where the pre-trained model can be found.&lt;br /&gt;
&lt;br /&gt;
If you want to use the parser for a language or treebank that was not among the treebanks that the parser was trained for, this can be done in two ways:&lt;br /&gt;
 &lt;br /&gt;
 uuparser --include &amp;quot;sv_pud:sv_talbanken&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_pud&amp;quot; --forced-tbank-emb sv_talbanken --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Where in both cases, sv_pud will be parsed using sv_talbanken as a proxy (i.e. parsing sv_pud as if it is is sv_talbanken). With the first option, multiple treebanks and proxies can be defined separated by space.&lt;br /&gt;
&lt;br /&gt;
To predict on non-UD texts, the command is similar to the single treebank case, the mian difference being the use of the --multiling flag, and specifying a proxy treebank:&lt;br /&gt;
&lt;br /&gt;
 uuparser --testfile TESTFILE --outdir ~/experiments --modeldir MODELDIR --predict --multiling --forced-tbank-emb sv_talbanken&lt;br /&gt;
&lt;br /&gt;
This commands runs the parser on TESTFILE (in CoNLLU format), using a model found in MODELDIR, and prints the resulting parse in a file in ~/experiments. In the example sv_talbanken is used as the proxy treebank.&lt;br /&gt;
&lt;br /&gt;
== Segmentation ==&lt;br /&gt;
&lt;br /&gt;
In the above examples, we assume pre-segmented input data already in the [http://universaldependencies.org/format.html CONLL-U] format. If your input is raw text, we recommend using UDPipe to segment and convert the format first. The UDPipe module can be loaded using &amp;lt;code&amp;gt;module load nlpl-udpipe&amp;lt;/code&amp;gt; and then run by typing &amp;lt;code&amp;gt;udpipe&amp;lt;/code&amp;gt; at the command line, see [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe].&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=907</id>
		<title>Parsing/uuparser</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=907"/>
		<updated>2020-01-14T19:52:32Z</updated>

		<summary type="html">&lt;p&gt;Sara: /* Training a multitreebank parsing model */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= The Uppsala Parser =&lt;br /&gt;
&lt;br /&gt;
The Uppsala Parser is a neural transition-based dependency parser based on bist-parser by Eli Kiperwasser and Yoav Goldberg and developed primarily in the context of the CoNLL shared tasks on universal dependency parsing in 2017 and 2018. The Uppsala Parser is publicly available at https://github.com/UppsalaNLP/uuparser. &lt;br /&gt;
Note that the version installed here may exhibit some slight differences, designed to improve ease of use.&lt;br /&gt;
&lt;br /&gt;
For the full documentation, see the github page.&lt;br /&gt;
&lt;br /&gt;
== Using the Uppsala Parser on Saga ==&lt;br /&gt;
&lt;br /&gt;
* Log into Saga&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /cluster/shared/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load nlpl-uuparser/2.3.0&lt;br /&gt;
&lt;br /&gt;
== Training a parsing model ==&lt;br /&gt;
&lt;br /&gt;
To train a set of parsing models on treebanks from Universal Dependencies (v2.2 or later):&lt;br /&gt;
&lt;br /&gt;
 uuparser --include [languages to include denoted by their treebank id] --outdir my-output-directory  --datadir ud-treebank-dir&lt;br /&gt;
&lt;br /&gt;
For example:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments  --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
will train separate models on UD Swedish-Talbanken, UD English-ParTUT and UD Russian-SynTagRus and store the results in the &amp;lt;code&amp;gt;experiments&amp;lt;/code&amp;gt; folder in your home directory. Model selection is included in the training process by default; that is, at each epoch the current model is evaluated on the UD dev data, and at the end of training the best performing model for each language is selected.&lt;br /&gt;
&lt;br /&gt;
== Predicting with a pre-trained parsing model ==&lt;br /&gt;
&lt;br /&gt;
To predict on UD test data with the models trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
To predict on other texts, prepare the data into the CoNLLU format (see below) and run the following command, assuming the model trained above, for the file TESTFILE:&lt;br /&gt;
&lt;br /&gt;
 uuparser --testfile TESTFILE --outdir ~/experiments --modeldir ~/experiments/sv_talbanken --predict&lt;br /&gt;
&lt;br /&gt;
== Options ==&lt;br /&gt;
&lt;br /&gt;
The parser has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 uuparser --help | less&lt;br /&gt;
&lt;br /&gt;
We recommend you set the &amp;lt;code&amp;gt;--dynet-mem&amp;lt;/code&amp;gt; option to a larger number when running the full training procedure on larger treebanks. Commonly used values are 5000 and 10000 (in MB). Dynet is the neural network library on which the parser is built.&lt;br /&gt;
&lt;br /&gt;
Note that due to random initialization and other non-deterministic elements in the training process, you will not obtain the same results even when training twice under exactly the same circumstances (e.g. languages, number of epochs etc.). To ensure identical results between two runs, we recommend setting the &amp;lt;code&amp;gt;--dynet-seed&amp;lt;/code&amp;gt; option to the same value both times (e.g. &amp;lt;code&amp;gt; --dynet-seed 123456789&amp;lt;/code&amp;gt;) and adding the &amp;lt;code&amp;gt;--use-default-seed&amp;lt;/code&amp;gt; flag. This ensures that Python's random number generator and Dynet both produce the same sequence of random numbers.&lt;br /&gt;
&lt;br /&gt;
== Training a multitreebank parsing model ==&lt;br /&gt;
&lt;br /&gt;
Our parser supports multitreebank parsing, that is a single parsing model for one or more treebanks, which could be from the same or form different languages. To train a ''multitreebank'' model for the three languages in the examples above, we simply add the &amp;lt;code&amp;gt;--multiling&amp;lt;/code&amp;gt; flag when training (note, though, that these models have so far been found to work best for groups of related languages)&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
In this case, instead of creating three separate models in the language-specific subdirectories within ~/experiments, a single model will be created directly in this folder. Predicting on UD test data is then:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Note that if you want to have different output directories for training and predicting, the &amp;lt;code&amp;gt;--modeldir&amp;lt;/code&amp;gt; option can be specified when predicting to tell the parser where the pre-trained model can be found.&lt;br /&gt;
&lt;br /&gt;
If you want to use the parser for a language or treebank that was not among the treebanks that the parser was trained for, this can be done in two ways:&lt;br /&gt;
 &lt;br /&gt;
 uuparser --include &amp;quot;sv_pud:sv_talbanken&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_pud&amp;quot; --forced-tbank-emb sv_talbanken --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Where in both cases, sv_pud will be parsed using sv_talbanken as a proxy (i.e. parsing sv_pud as if it is is sv_talbanken). With the first option, multiple treebanks and proxies can be defined separated by space.&lt;br /&gt;
&lt;br /&gt;
To predict on non-UD texts, the command is similar to the single treebank case, the mian difference being the use of the --multiling flag, and specifying a proxy treebank:&lt;br /&gt;
&lt;br /&gt;
 uuparser --testfile TESTFILE --outdir ~/experiments --modeldir MODELDIR --predict --multiling --forced-tbank-emb sv_talbanken&lt;br /&gt;
&lt;br /&gt;
This commands runs the parser on TESTFILE (in CoNLLU format), using a model found in MODELDIR, and prints the resulting parse in a file in ~/experiments. In the example sv_talbanken is used as the proxy treebank.&lt;br /&gt;
&lt;br /&gt;
== Segmentation ==&lt;br /&gt;
&lt;br /&gt;
In the above examples, we assume pre-segmented input data already in the [http://universaldependencies.org/format.html CONLL-U] format. If your input is raw text, we recommend using UDPipe to segment first. The UDPipe module can be loaded using &amp;lt;code&amp;gt;module load nlpl-udpipe&amp;lt;/code&amp;gt; and then run by typing &amp;lt;code&amp;gt;udpipe&amp;lt;/code&amp;gt; at the command line, see [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe].&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=906</id>
		<title>Parsing/uuparser</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=906"/>
		<updated>2020-01-14T19:48:26Z</updated>

		<summary type="html">&lt;p&gt;Sara: /* Predicting with a pre-trained parsing model */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= The Uppsala Parser =&lt;br /&gt;
&lt;br /&gt;
The Uppsala Parser is a neural transition-based dependency parser based on bist-parser by Eli Kiperwasser and Yoav Goldberg and developed primarily in the context of the CoNLL shared tasks on universal dependency parsing in 2017 and 2018. The Uppsala Parser is publicly available at https://github.com/UppsalaNLP/uuparser. &lt;br /&gt;
Note that the version installed here may exhibit some slight differences, designed to improve ease of use.&lt;br /&gt;
&lt;br /&gt;
For the full documentation, see the github page.&lt;br /&gt;
&lt;br /&gt;
== Using the Uppsala Parser on Saga ==&lt;br /&gt;
&lt;br /&gt;
* Log into Saga&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /cluster/shared/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load nlpl-uuparser/2.3.0&lt;br /&gt;
&lt;br /&gt;
== Training a parsing model ==&lt;br /&gt;
&lt;br /&gt;
To train a set of parsing models on treebanks from Universal Dependencies (v2.2 or later):&lt;br /&gt;
&lt;br /&gt;
 uuparser --include [languages to include denoted by their treebank id] --outdir my-output-directory  --datadir ud-treebank-dir&lt;br /&gt;
&lt;br /&gt;
For example:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments  --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
will train separate models on UD Swedish-Talbanken, UD English-ParTUT and UD Russian-SynTagRus and store the results in the &amp;lt;code&amp;gt;experiments&amp;lt;/code&amp;gt; folder in your home directory. Model selection is included in the training process by default; that is, at each epoch the current model is evaluated on the UD dev data, and at the end of training the best performing model for each language is selected.&lt;br /&gt;
&lt;br /&gt;
== Predicting with a pre-trained parsing model ==&lt;br /&gt;
&lt;br /&gt;
To predict on UD test data with the models trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
To predict on other texts, prepare the data into the CoNLLU format (see below) and run the following command, assuming the model trained above, for the file TESTFILE:&lt;br /&gt;
&lt;br /&gt;
 uuparser --testfile TESTFILE --outdir ~/experiments --modeldir ~/experiments/sv_talbanken --predict&lt;br /&gt;
&lt;br /&gt;
== Options ==&lt;br /&gt;
&lt;br /&gt;
The parser has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 uuparser --help | less&lt;br /&gt;
&lt;br /&gt;
We recommend you set the &amp;lt;code&amp;gt;--dynet-mem&amp;lt;/code&amp;gt; option to a larger number when running the full training procedure on larger treebanks. Commonly used values are 5000 and 10000 (in MB). Dynet is the neural network library on which the parser is built.&lt;br /&gt;
&lt;br /&gt;
Note that due to random initialization and other non-deterministic elements in the training process, you will not obtain the same results even when training twice under exactly the same circumstances (e.g. languages, number of epochs etc.). To ensure identical results between two runs, we recommend setting the &amp;lt;code&amp;gt;--dynet-seed&amp;lt;/code&amp;gt; option to the same value both times (e.g. &amp;lt;code&amp;gt; --dynet-seed 123456789&amp;lt;/code&amp;gt;) and adding the &amp;lt;code&amp;gt;--use-default-seed&amp;lt;/code&amp;gt; flag. This ensures that Python's random number generator and Dynet both produce the same sequence of random numbers.&lt;br /&gt;
&lt;br /&gt;
== Training a multitreebank parsing model ==&lt;br /&gt;
&lt;br /&gt;
Our parser supports multitreebank parsing, that is a single parsing model for one or more treebanks, which could be from the same or form different languages. To train a ''multitreebank'' model for the three languages in the examples above, we simply add the &amp;lt;code&amp;gt;--multiling&amp;lt;/code&amp;gt; flag when training (note, though, that these models have so far been found to work best for groups of related languages)&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
In this case, instead of creating three separate models in the language-specific subdirectories within ~/experiments, a single model will be created directly in this folder. Predicting on test data is then as easy as:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Note that if you want to have different output directories for training and predicting, the &amp;lt;code&amp;gt;--modeldir&amp;lt;/code&amp;gt; option can be specified when predicting to tell the parser where the pre-trained model can be found.&lt;br /&gt;
&lt;br /&gt;
If you want to use the parser for a language or treebank that was not among the treebanks that the parser was trained for, this can be done in two ways:&lt;br /&gt;
 &lt;br /&gt;
 uuparser --include &amp;quot;sv_pud:sv_talbanken&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_pud&amp;quot; --forced-tbank-emb sv_talbanken --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Where in both cases, sv_pud will be parsed using sv_talbanken as a proxy (i.e. parsing sv_pud as if it is is sv_talbanken). With the first option, multiple treebanks and proxies can be defined separated by space.&lt;br /&gt;
&lt;br /&gt;
== Segmentation ==&lt;br /&gt;
&lt;br /&gt;
In the above examples, we assume pre-segmented input data already in the [http://universaldependencies.org/format.html CONLL-U] format. If your input is raw text, we recommend using UDPipe to segment first. The UDPipe module can be loaded using &amp;lt;code&amp;gt;module load nlpl-udpipe&amp;lt;/code&amp;gt; and then run by typing &amp;lt;code&amp;gt;udpipe&amp;lt;/code&amp;gt; at the command line, see [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe].&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=905</id>
		<title>Parsing/uuparser</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=905"/>
		<updated>2020-01-14T19:48:00Z</updated>

		<summary type="html">&lt;p&gt;Sara: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= The Uppsala Parser =&lt;br /&gt;
&lt;br /&gt;
The Uppsala Parser is a neural transition-based dependency parser based on bist-parser by Eli Kiperwasser and Yoav Goldberg and developed primarily in the context of the CoNLL shared tasks on universal dependency parsing in 2017 and 2018. The Uppsala Parser is publicly available at https://github.com/UppsalaNLP/uuparser. &lt;br /&gt;
Note that the version installed here may exhibit some slight differences, designed to improve ease of use.&lt;br /&gt;
&lt;br /&gt;
For the full documentation, see the github page.&lt;br /&gt;
&lt;br /&gt;
== Using the Uppsala Parser on Saga ==&lt;br /&gt;
&lt;br /&gt;
* Log into Saga&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /cluster/shared/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load nlpl-uuparser/2.3.0&lt;br /&gt;
&lt;br /&gt;
== Training a parsing model ==&lt;br /&gt;
&lt;br /&gt;
To train a set of parsing models on treebanks from Universal Dependencies (v2.2 or later):&lt;br /&gt;
&lt;br /&gt;
 uuparser --include [languages to include denoted by their treebank id] --outdir my-output-directory  --datadir ud-treebank-dir&lt;br /&gt;
&lt;br /&gt;
For example:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments  --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
will train separate models on UD Swedish-Talbanken, UD English-ParTUT and UD Russian-SynTagRus and store the results in the &amp;lt;code&amp;gt;experiments&amp;lt;/code&amp;gt; folder in your home directory. Model selection is included in the training process by default; that is, at each epoch the current model is evaluated on the UD dev data, and at the end of training the best performing model for each language is selected.&lt;br /&gt;
&lt;br /&gt;
== Predicting with a pre-trained parsing model ==&lt;br /&gt;
&lt;br /&gt;
To predict on UD test data with the models trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
To predict on other texts, prepare the data into the CoNLLU format (see below) and run the following command, assuming the model trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --testfile TESTFILE --outdir ~/experiments --modeldir ~/experiments/sv_talbanken --predict&lt;br /&gt;
&lt;br /&gt;
== Options ==&lt;br /&gt;
&lt;br /&gt;
The parser has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 uuparser --help | less&lt;br /&gt;
&lt;br /&gt;
We recommend you set the &amp;lt;code&amp;gt;--dynet-mem&amp;lt;/code&amp;gt; option to a larger number when running the full training procedure on larger treebanks. Commonly used values are 5000 and 10000 (in MB). Dynet is the neural network library on which the parser is built.&lt;br /&gt;
&lt;br /&gt;
Note that due to random initialization and other non-deterministic elements in the training process, you will not obtain the same results even when training twice under exactly the same circumstances (e.g. languages, number of epochs etc.). To ensure identical results between two runs, we recommend setting the &amp;lt;code&amp;gt;--dynet-seed&amp;lt;/code&amp;gt; option to the same value both times (e.g. &amp;lt;code&amp;gt; --dynet-seed 123456789&amp;lt;/code&amp;gt;) and adding the &amp;lt;code&amp;gt;--use-default-seed&amp;lt;/code&amp;gt; flag. This ensures that Python's random number generator and Dynet both produce the same sequence of random numbers.&lt;br /&gt;
&lt;br /&gt;
== Training a multitreebank parsing model ==&lt;br /&gt;
&lt;br /&gt;
Our parser supports multitreebank parsing, that is a single parsing model for one or more treebanks, which could be from the same or form different languages. To train a ''multitreebank'' model for the three languages in the examples above, we simply add the &amp;lt;code&amp;gt;--multiling&amp;lt;/code&amp;gt; flag when training (note, though, that these models have so far been found to work best for groups of related languages)&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
In this case, instead of creating three separate models in the language-specific subdirectories within ~/experiments, a single model will be created directly in this folder. Predicting on test data is then as easy as:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Note that if you want to have different output directories for training and predicting, the &amp;lt;code&amp;gt;--modeldir&amp;lt;/code&amp;gt; option can be specified when predicting to tell the parser where the pre-trained model can be found.&lt;br /&gt;
&lt;br /&gt;
If you want to use the parser for a language or treebank that was not among the treebanks that the parser was trained for, this can be done in two ways:&lt;br /&gt;
 &lt;br /&gt;
 uuparser --include &amp;quot;sv_pud:sv_talbanken&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_pud&amp;quot; --forced-tbank-emb sv_talbanken --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Where in both cases, sv_pud will be parsed using sv_talbanken as a proxy (i.e. parsing sv_pud as if it is is sv_talbanken). With the first option, multiple treebanks and proxies can be defined separated by space.&lt;br /&gt;
&lt;br /&gt;
== Segmentation ==&lt;br /&gt;
&lt;br /&gt;
In the above examples, we assume pre-segmented input data already in the [http://universaldependencies.org/format.html CONLL-U] format. If your input is raw text, we recommend using UDPipe to segment first. The UDPipe module can be loaded using &amp;lt;code&amp;gt;module load nlpl-udpipe&amp;lt;/code&amp;gt; and then run by typing &amp;lt;code&amp;gt;udpipe&amp;lt;/code&amp;gt; at the command line, see [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe].&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=904</id>
		<title>Parsing/uuparser</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=904"/>
		<updated>2020-01-14T19:46:45Z</updated>

		<summary type="html">&lt;p&gt;Sara: /* Predicting with a pre-trained parsing model */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= The Uppsala Parser =&lt;br /&gt;
&lt;br /&gt;
The Uppsala Parser is a neural transition-based dependency parser based on bist-parser by Eli Kiperwasser and Yoav Goldberg and developed primarily in the context of the CoNLL shared tasks on universal dependency parsing in 2017 and 2018. The Uppsala Parser is publicly available at https://github.com/UppsalaNLP/uuparser. &lt;br /&gt;
Note that the version installed here may exhibit some slight differences, designed to improve ease of use.&lt;br /&gt;
&lt;br /&gt;
For the full documentation, see the github page.&lt;br /&gt;
&lt;br /&gt;
== Using the Uppsala Parser on Saga ==&lt;br /&gt;
&lt;br /&gt;
* Log into Saga&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /cluster/shared/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load nlpl-uuparser/2.3.0&lt;br /&gt;
&lt;br /&gt;
== Training a parsing model ==&lt;br /&gt;
&lt;br /&gt;
To train a set of parsing models on treebanks from Universal Dependencies (v2.2 or later):&lt;br /&gt;
&lt;br /&gt;
 uuparser --include [languages to include denoted by their treebank id] --outdir my-output-directory  --datadir ud-treebank-dir&lt;br /&gt;
&lt;br /&gt;
For example:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments  --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
will train separate models on UD Swedish-Talbanken, UD English-ParTUT and UD Russian-SynTagRus and store the results in the &amp;lt;code&amp;gt;experiments&amp;lt;/code&amp;gt; folder in your home directory. Model selection is included in the training process by default; that is, at each epoch the current model is evaluated on the UD dev data, and at the end of training the best performing model for each language is selected.&lt;br /&gt;
&lt;br /&gt;
== Predicting with a pre-trained parsing model ==&lt;br /&gt;
&lt;br /&gt;
To predict on UD test data with the models trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --predict --datadir UD_DATA_DIR&lt;br /&gt;
&lt;br /&gt;
To predict on other texts, prepare the data into the CoNLLU format (see below) and run the following command, assuming the model trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --testfile TESTFILE --outdir ~/experiments --modeldir ~/experiments/sv_talbanken --predict&lt;br /&gt;
&lt;br /&gt;
== Options ==&lt;br /&gt;
&lt;br /&gt;
The parser has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 uuparser --help | less&lt;br /&gt;
&lt;br /&gt;
We recommend you set the &amp;lt;code&amp;gt;--dynet-mem&amp;lt;/code&amp;gt; option to a larger number when running the full training procedure on larger treebanks. Commonly used values are 5000 and 10000 (in MB). Dynet is the neural network library on which the parser is built.&lt;br /&gt;
&lt;br /&gt;
Note that due to random initialization and other non-deterministic elements in the training process, you will not obtain the same results even when training twice under exactly the same circumstances (e.g. languages, number of epochs etc.). To ensure identical results between two runs, we recommend setting the &amp;lt;code&amp;gt;--dynet-seed&amp;lt;/code&amp;gt; option to the same value both times (e.g. &amp;lt;code&amp;gt; --dynet-seed 123456789&amp;lt;/code&amp;gt;) and adding the &amp;lt;code&amp;gt;--use-default-seed&amp;lt;/code&amp;gt; flag. This ensures that Python's random number generator and Dynet both produce the same sequence of random numbers.&lt;br /&gt;
&lt;br /&gt;
== Training a multitreebank parsing model ==&lt;br /&gt;
&lt;br /&gt;
Our parser supports multitreebank parsing, that is a single parsing model for one or more treebanks, which could be from the same or form different languages. To train a ''multitreebank'' model for the three languages in the examples above, we simply add the &amp;lt;code&amp;gt;--multiling&amp;lt;/code&amp;gt; flag when training (note, though, that these models have so far been found to work best for groups of related languages)&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
In this case, instead of creating three separate models in the language-specific subdirectories within ~/experiments, a single model will be created directly in this folder. Predicting on test data is then as easy as:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Note that if you want to have different output directories for training and predicting, the &amp;lt;code&amp;gt;--modeldir&amp;lt;/code&amp;gt; option can be specified when predicting to tell the parser where the pre-trained model can be found.&lt;br /&gt;
&lt;br /&gt;
If you want to use the parser for a language or treebank that was not among the treebanks that the parser was trained for, this can be done in two ways:&lt;br /&gt;
 &lt;br /&gt;
 uuparser --include &amp;quot;sv_pud:sv_talbanken&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_pud&amp;quot; --forced-tbank-emb sv_talbanken --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Where in both cases, sv_pud will be parsed using sv_talbanken as a proxy (i.e. parsing sv_pud as if it is is sv_talbanken). With the first option, multiple treebanks and proxies can be defined separated by space.&lt;br /&gt;
&lt;br /&gt;
== Segmentation ==&lt;br /&gt;
&lt;br /&gt;
In the above examples, we assume pre-segmented input data already in the [http://universaldependencies.org/format.html CONLL-U] format. If your input is raw text, we recommend using UDPipe to segment first. The UDPipe module can be loaded using &amp;lt;code&amp;gt;module load nlpl-udpipe&amp;lt;/code&amp;gt; and then run by typing &amp;lt;code&amp;gt;udpipe&amp;lt;/code&amp;gt; at the command line, see [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe].&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=903</id>
		<title>Parsing/uuparser</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=903"/>
		<updated>2020-01-14T13:07:04Z</updated>

		<summary type="html">&lt;p&gt;Sara: /* Training a parsing model */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= The Uppsala Parser =&lt;br /&gt;
&lt;br /&gt;
The Uppsala Parser is a neural transition-based dependency parser based on bist-parser by Eli Kiperwasser and Yoav Goldberg and developed primarily in the context of the CoNLL shared tasks on universal dependency parsing in 2017 and 2018. The Uppsala Parser is publicly available at https://github.com/UppsalaNLP/uuparser. &lt;br /&gt;
Note that the version installed here may exhibit some slight differences, designed to improve ease of use.&lt;br /&gt;
&lt;br /&gt;
For the full documentation, see the github page.&lt;br /&gt;
&lt;br /&gt;
== Using the Uppsala Parser on Saga ==&lt;br /&gt;
&lt;br /&gt;
* Log into Saga&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /cluster/shared/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load nlpl-uuparser/2.3.0&lt;br /&gt;
&lt;br /&gt;
== Training a parsing model ==&lt;br /&gt;
&lt;br /&gt;
To train a set of parsing models on treebanks from Universal Dependencies (v2.2 or later):&lt;br /&gt;
&lt;br /&gt;
 uuparser --include [languages to include denoted by their treebank id] --outdir my-output-directory  --datadir ud-treebank-dir&lt;br /&gt;
&lt;br /&gt;
For example:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments  --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
will train separate models on UD Swedish-Talbanken, UD English-ParTUT and UD Russian-SynTagRus and store the results in the &amp;lt;code&amp;gt;experiments&amp;lt;/code&amp;gt; folder in your home directory. Model selection is included in the training process by default; that is, at each epoch the current model is evaluated on the UD dev data, and at the end of training the best performing model for each language is selected.&lt;br /&gt;
&lt;br /&gt;
== Predicting with a pre-trained parsing model ==&lt;br /&gt;
&lt;br /&gt;
To predict on UD test data with the models trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --predict&lt;br /&gt;
&lt;br /&gt;
== Options ==&lt;br /&gt;
&lt;br /&gt;
The parser has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 uuparser --help | less&lt;br /&gt;
&lt;br /&gt;
We recommend you set the &amp;lt;code&amp;gt;--dynet-mem&amp;lt;/code&amp;gt; option to a larger number when running the full training procedure on larger treebanks. Commonly used values are 5000 and 10000 (in MB). Dynet is the neural network library on which the parser is built.&lt;br /&gt;
&lt;br /&gt;
Note that due to random initialization and other non-deterministic elements in the training process, you will not obtain the same results even when training twice under exactly the same circumstances (e.g. languages, number of epochs etc.). To ensure identical results between two runs, we recommend setting the &amp;lt;code&amp;gt;--dynet-seed&amp;lt;/code&amp;gt; option to the same value both times (e.g. &amp;lt;code&amp;gt; --dynet-seed 123456789&amp;lt;/code&amp;gt;) and adding the &amp;lt;code&amp;gt;--use-default-seed&amp;lt;/code&amp;gt; flag. This ensures that Python's random number generator and Dynet both produce the same sequence of random numbers.&lt;br /&gt;
&lt;br /&gt;
== Training a multitreebank parsing model ==&lt;br /&gt;
&lt;br /&gt;
Our parser supports multitreebank parsing, that is a single parsing model for one or more treebanks, which could be from the same or form different languages. To train a ''multitreebank'' model for the three languages in the examples above, we simply add the &amp;lt;code&amp;gt;--multiling&amp;lt;/code&amp;gt; flag when training (note, though, that these models have so far been found to work best for groups of related languages)&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
In this case, instead of creating three separate models in the language-specific subdirectories within ~/experiments, a single model will be created directly in this folder. Predicting on test data is then as easy as:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Note that if you want to have different output directories for training and predicting, the &amp;lt;code&amp;gt;--modeldir&amp;lt;/code&amp;gt; option can be specified when predicting to tell the parser where the pre-trained model can be found.&lt;br /&gt;
&lt;br /&gt;
If you want to use the parser for a language or treebank that was not among the treebanks that the parser was trained for, this can be done in two ways:&lt;br /&gt;
 &lt;br /&gt;
 uuparser --include &amp;quot;sv_pud:sv_talbanken&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_pud&amp;quot; --forced-tbank-emb sv_talbanken --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Where in both cases, sv_pud will be parsed using sv_talbanken as a proxy (i.e. parsing sv_pud as if it is is sv_talbanken). With the first option, multiple treebanks and proxies can be defined separated by space.&lt;br /&gt;
&lt;br /&gt;
== Segmentation ==&lt;br /&gt;
&lt;br /&gt;
In the above examples, we assume pre-segmented input data already in the [http://universaldependencies.org/format.html CONLL-U] format. If your input is raw text, we recommend using UDPipe to segment first. The UDPipe module can be loaded using &amp;lt;code&amp;gt;module load nlpl-udpipe&amp;lt;/code&amp;gt; and then run by typing &amp;lt;code&amp;gt;udpipe&amp;lt;/code&amp;gt; at the command line, see [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe].&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=902</id>
		<title>Parsing/uuparser</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=902"/>
		<updated>2020-01-14T13:06:20Z</updated>

		<summary type="html">&lt;p&gt;Sara: /* Training a multitreebank parsing model */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= The Uppsala Parser =&lt;br /&gt;
&lt;br /&gt;
The Uppsala Parser is a neural transition-based dependency parser based on bist-parser by Eli Kiperwasser and Yoav Goldberg and developed primarily in the context of the CoNLL shared tasks on universal dependency parsing in 2017 and 2018. The Uppsala Parser is publicly available at https://github.com/UppsalaNLP/uuparser. &lt;br /&gt;
Note that the version installed here may exhibit some slight differences, designed to improve ease of use.&lt;br /&gt;
&lt;br /&gt;
For the full documentation, see the github page.&lt;br /&gt;
&lt;br /&gt;
== Using the Uppsala Parser on Saga ==&lt;br /&gt;
&lt;br /&gt;
* Log into Saga&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /cluster/shared/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load nlpl-uuparser/2.3.0&lt;br /&gt;
&lt;br /&gt;
== Training a parsing model ==&lt;br /&gt;
&lt;br /&gt;
To train a set of parsing models on treebanks from Universal Dependencies (v2.2 or later):&lt;br /&gt;
&lt;br /&gt;
 uuparser --include [languages to include denoted by their treebank id] --outdir my-output-directory&lt;br /&gt;
&lt;br /&gt;
For example:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments&lt;br /&gt;
&lt;br /&gt;
will train separate models on UD Swedish-Talbanken, UD English-ParTUT and UD Russian-SynTagRus and store the results in the &amp;lt;code&amp;gt;experiments&amp;lt;/code&amp;gt; folder in your home directory. Model selection is included in the training process by default; that is, at each epoch the current model is evaluated on the UD dev data, and at the end of training the best performing model for each language is selected.&lt;br /&gt;
&lt;br /&gt;
== Predicting with a pre-trained parsing model ==&lt;br /&gt;
&lt;br /&gt;
To predict on UD test data with the models trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --predict&lt;br /&gt;
&lt;br /&gt;
== Options ==&lt;br /&gt;
&lt;br /&gt;
The parser has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 uuparser --help | less&lt;br /&gt;
&lt;br /&gt;
We recommend you set the &amp;lt;code&amp;gt;--dynet-mem&amp;lt;/code&amp;gt; option to a larger number when running the full training procedure on larger treebanks. Commonly used values are 5000 and 10000 (in MB). Dynet is the neural network library on which the parser is built.&lt;br /&gt;
&lt;br /&gt;
Note that due to random initialization and other non-deterministic elements in the training process, you will not obtain the same results even when training twice under exactly the same circumstances (e.g. languages, number of epochs etc.). To ensure identical results between two runs, we recommend setting the &amp;lt;code&amp;gt;--dynet-seed&amp;lt;/code&amp;gt; option to the same value both times (e.g. &amp;lt;code&amp;gt; --dynet-seed 123456789&amp;lt;/code&amp;gt;) and adding the &amp;lt;code&amp;gt;--use-default-seed&amp;lt;/code&amp;gt; flag. This ensures that Python's random number generator and Dynet both produce the same sequence of random numbers.&lt;br /&gt;
&lt;br /&gt;
== Training a multitreebank parsing model ==&lt;br /&gt;
&lt;br /&gt;
Our parser supports multitreebank parsing, that is a single parsing model for one or more treebanks, which could be from the same or form different languages. To train a ''multitreebank'' model for the three languages in the examples above, we simply add the &amp;lt;code&amp;gt;--multiling&amp;lt;/code&amp;gt; flag when training (note, though, that these models have so far been found to work best for groups of related languages)&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
In this case, instead of creating three separate models in the language-specific subdirectories within ~/experiments, a single model will be created directly in this folder. Predicting on test data is then as easy as:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Note that if you want to have different output directories for training and predicting, the &amp;lt;code&amp;gt;--modeldir&amp;lt;/code&amp;gt; option can be specified when predicting to tell the parser where the pre-trained model can be found.&lt;br /&gt;
&lt;br /&gt;
If you want to use the parser for a language or treebank that was not among the treebanks that the parser was trained for, this can be done in two ways:&lt;br /&gt;
 &lt;br /&gt;
 uuparser --include &amp;quot;sv_pud:sv_talbanken&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_pud&amp;quot; --forced-tbank-emb sv_talbanken --outdir ~/experiments --multiling --dynet-mem 5000 --predict --datadir /cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&lt;br /&gt;
&lt;br /&gt;
Where in both cases, sv_pud will be parsed using sv_talbanken as a proxy (i.e. parsing sv_pud as if it is is sv_talbanken). With the first option, multiple treebanks and proxies can be defined separated by space.&lt;br /&gt;
&lt;br /&gt;
== Segmentation ==&lt;br /&gt;
&lt;br /&gt;
In the above examples, we assume pre-segmented input data already in the [http://universaldependencies.org/format.html CONLL-U] format. If your input is raw text, we recommend using UDPipe to segment first. The UDPipe module can be loaded using &amp;lt;code&amp;gt;module load nlpl-udpipe&amp;lt;/code&amp;gt; and then run by typing &amp;lt;code&amp;gt;udpipe&amp;lt;/code&amp;gt; at the command line, see [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe].&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=901</id>
		<title>Parsing/uuparser</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=901"/>
		<updated>2020-01-14T12:51:15Z</updated>

		<summary type="html">&lt;p&gt;Sara: /* Segmentation */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= The Uppsala Parser =&lt;br /&gt;
&lt;br /&gt;
The Uppsala Parser is a neural transition-based dependency parser based on bist-parser by Eli Kiperwasser and Yoav Goldberg and developed primarily in the context of the CoNLL shared tasks on universal dependency parsing in 2017 and 2018. The Uppsala Parser is publicly available at https://github.com/UppsalaNLP/uuparser. &lt;br /&gt;
Note that the version installed here may exhibit some slight differences, designed to improve ease of use.&lt;br /&gt;
&lt;br /&gt;
For the full documentation, see the github page.&lt;br /&gt;
&lt;br /&gt;
== Using the Uppsala Parser on Saga ==&lt;br /&gt;
&lt;br /&gt;
* Log into Saga&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /cluster/shared/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load nlpl-uuparser/2.3.0&lt;br /&gt;
&lt;br /&gt;
== Training a parsing model ==&lt;br /&gt;
&lt;br /&gt;
To train a set of parsing models on treebanks from Universal Dependencies (v2.2 or later):&lt;br /&gt;
&lt;br /&gt;
 uuparser --include [languages to include denoted by their treebank id] --outdir my-output-directory&lt;br /&gt;
&lt;br /&gt;
For example:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments&lt;br /&gt;
&lt;br /&gt;
will train separate models on UD Swedish-Talbanken, UD English-ParTUT and UD Russian-SynTagRus and store the results in the &amp;lt;code&amp;gt;experiments&amp;lt;/code&amp;gt; folder in your home directory. Model selection is included in the training process by default; that is, at each epoch the current model is evaluated on the UD dev data, and at the end of training the best performing model for each language is selected.&lt;br /&gt;
&lt;br /&gt;
== Predicting with a pre-trained parsing model ==&lt;br /&gt;
&lt;br /&gt;
To predict on UD test data with the models trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --predict&lt;br /&gt;
&lt;br /&gt;
== Options ==&lt;br /&gt;
&lt;br /&gt;
The parser has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 uuparser --help | less&lt;br /&gt;
&lt;br /&gt;
We recommend you set the &amp;lt;code&amp;gt;--dynet-mem&amp;lt;/code&amp;gt; option to a larger number when running the full training procedure on larger treebanks. Commonly used values are 5000 and 10000 (in MB). Dynet is the neural network library on which the parser is built.&lt;br /&gt;
&lt;br /&gt;
Note that due to random initialization and other non-deterministic elements in the training process, you will not obtain the same results even when training twice under exactly the same circumstances (e.g. languages, number of epochs etc.). To ensure identical results between two runs, we recommend setting the &amp;lt;code&amp;gt;--dynet-seed&amp;lt;/code&amp;gt; option to the same value both times (e.g. &amp;lt;code&amp;gt; --dynet-seed 123456789&amp;lt;/code&amp;gt;) and adding the &amp;lt;code&amp;gt;--use-default-seed&amp;lt;/code&amp;gt; flag. This ensures that Python's random number generator and Dynet both produce the same sequence of random numbers.&lt;br /&gt;
&lt;br /&gt;
== Training a multitreebank parsing model ==&lt;br /&gt;
&lt;br /&gt;
Our parser supports multitreebank parsing, that is a single parsing model for one or more treebanks, which could be from the same or form different languages. To train a ''multitreebank'' model for the three languages in the examples above, we simply add the &amp;lt;code&amp;gt;--multiling&amp;lt;/code&amp;gt; flag when training (note, though, that these models have so far been found to work best for groups of related languages)&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000&lt;br /&gt;
&lt;br /&gt;
In this case, instead of creating three separate models in the language-specific subdirectories within ~/experiments, a single model will be created directly in this folder. Predicting on test data is then as easy as:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict&lt;br /&gt;
&lt;br /&gt;
Note that if you want to have different output directories for training and predicting, the &amp;lt;code&amp;gt;--modeldir&amp;lt;/code&amp;gt; option can be specified when predicting to tell the parser where the pre-trained model can be found.&lt;br /&gt;
&lt;br /&gt;
If you want to use the parser for a language or treebank that was not among the treebanks that the parser was trained for, this can be done in two ways:&lt;br /&gt;
 &lt;br /&gt;
 uuparser --include &amp;quot;sv_pud:sv_talbanken&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_pud&amp;quot; --forced-tbank-emb sv_talbanken --outdir ~/experiments --multiling --dynet-mem 5000 --predict&lt;br /&gt;
&lt;br /&gt;
Where in both cases, sv_pud will be parsed using sv_talbanken as a proxy (i.e. parsing sv_pud as if it is is sv_talbanken). With the first option, multiple treebanks and proxies can be defined separated by space.&lt;br /&gt;
&lt;br /&gt;
== Segmentation ==&lt;br /&gt;
&lt;br /&gt;
In the above examples, we assume pre-segmented input data already in the [http://universaldependencies.org/format.html CONLL-U] format. If your input is raw text, we recommend using UDPipe to segment first. The UDPipe module can be loaded using &amp;lt;code&amp;gt;module load nlpl-udpipe&amp;lt;/code&amp;gt; and then run by typing &amp;lt;code&amp;gt;udpipe&amp;lt;/code&amp;gt; at the command line, see [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe].&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=900</id>
		<title>Parsing/uuparser</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=900"/>
		<updated>2020-01-14T12:50:28Z</updated>

		<summary type="html">&lt;p&gt;Sara: /* Training a multitreebank parsing model */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= The Uppsala Parser =&lt;br /&gt;
&lt;br /&gt;
The Uppsala Parser is a neural transition-based dependency parser based on bist-parser by Eli Kiperwasser and Yoav Goldberg and developed primarily in the context of the CoNLL shared tasks on universal dependency parsing in 2017 and 2018. The Uppsala Parser is publicly available at https://github.com/UppsalaNLP/uuparser. &lt;br /&gt;
Note that the version installed here may exhibit some slight differences, designed to improve ease of use.&lt;br /&gt;
&lt;br /&gt;
For the full documentation, see the github page.&lt;br /&gt;
&lt;br /&gt;
== Using the Uppsala Parser on Saga ==&lt;br /&gt;
&lt;br /&gt;
* Log into Saga&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /cluster/shared/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load nlpl-uuparser/2.3.0&lt;br /&gt;
&lt;br /&gt;
== Training a parsing model ==&lt;br /&gt;
&lt;br /&gt;
To train a set of parsing models on treebanks from Universal Dependencies (v2.2 or later):&lt;br /&gt;
&lt;br /&gt;
 uuparser --include [languages to include denoted by their treebank id] --outdir my-output-directory&lt;br /&gt;
&lt;br /&gt;
For example:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments&lt;br /&gt;
&lt;br /&gt;
will train separate models on UD Swedish-Talbanken, UD English-ParTUT and UD Russian-SynTagRus and store the results in the &amp;lt;code&amp;gt;experiments&amp;lt;/code&amp;gt; folder in your home directory. Model selection is included in the training process by default; that is, at each epoch the current model is evaluated on the UD dev data, and at the end of training the best performing model for each language is selected.&lt;br /&gt;
&lt;br /&gt;
== Predicting with a pre-trained parsing model ==&lt;br /&gt;
&lt;br /&gt;
To predict on UD test data with the models trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --predict&lt;br /&gt;
&lt;br /&gt;
== Options ==&lt;br /&gt;
&lt;br /&gt;
The parser has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 uuparser --help | less&lt;br /&gt;
&lt;br /&gt;
We recommend you set the &amp;lt;code&amp;gt;--dynet-mem&amp;lt;/code&amp;gt; option to a larger number when running the full training procedure on larger treebanks. Commonly used values are 5000 and 10000 (in MB). Dynet is the neural network library on which the parser is built.&lt;br /&gt;
&lt;br /&gt;
Note that due to random initialization and other non-deterministic elements in the training process, you will not obtain the same results even when training twice under exactly the same circumstances (e.g. languages, number of epochs etc.). To ensure identical results between two runs, we recommend setting the &amp;lt;code&amp;gt;--dynet-seed&amp;lt;/code&amp;gt; option to the same value both times (e.g. &amp;lt;code&amp;gt; --dynet-seed 123456789&amp;lt;/code&amp;gt;) and adding the &amp;lt;code&amp;gt;--use-default-seed&amp;lt;/code&amp;gt; flag. This ensures that Python's random number generator and Dynet both produce the same sequence of random numbers.&lt;br /&gt;
&lt;br /&gt;
== Training a multitreebank parsing model ==&lt;br /&gt;
&lt;br /&gt;
Our parser supports multitreebank parsing, that is a single parsing model for one or more treebanks, which could be from the same or form different languages. To train a ''multitreebank'' model for the three languages in the examples above, we simply add the &amp;lt;code&amp;gt;--multiling&amp;lt;/code&amp;gt; flag when training (note, though, that these models have so far been found to work best for groups of related languages)&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000&lt;br /&gt;
&lt;br /&gt;
In this case, instead of creating three separate models in the language-specific subdirectories within ~/experiments, a single model will be created directly in this folder. Predicting on test data is then as easy as:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict&lt;br /&gt;
&lt;br /&gt;
Note that if you want to have different output directories for training and predicting, the &amp;lt;code&amp;gt;--modeldir&amp;lt;/code&amp;gt; option can be specified when predicting to tell the parser where the pre-trained model can be found.&lt;br /&gt;
&lt;br /&gt;
If you want to use the parser for a language or treebank that was not among the treebanks that the parser was trained for, this can be done in two ways:&lt;br /&gt;
 &lt;br /&gt;
 uuparser --include &amp;quot;sv_pud:sv_talbanken&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_pud&amp;quot; --forced-tbank-emb sv_talbanken --outdir ~/experiments --multiling --dynet-mem 5000 --predict&lt;br /&gt;
&lt;br /&gt;
Where in both cases, sv_pud will be parsed using sv_talbanken as a proxy (i.e. parsing sv_pud as if it is is sv_talbanken). With the first option, multiple treebanks and proxies can be defined separated by space.&lt;br /&gt;
&lt;br /&gt;
== Segmentation ==&lt;br /&gt;
&lt;br /&gt;
In the above examples, we assume pre-segmented input data already in the [http://universaldependencies.org/format.html CONLL-U] format. If your input is raw text, we recommend using UDPipe to segment first. The UDPipe module can be loaded using &amp;lt;code&amp;gt;module load nlpl-udpipe&amp;lt;/code&amp;gt; and then run by typing &amp;lt;code&amp;gt;udpipe&amp;lt;/code&amp;gt; at the command line, see below.&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=899</id>
		<title>Parsing/uuparser</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=899"/>
		<updated>2020-01-14T12:19:17Z</updated>

		<summary type="html">&lt;p&gt;Sara: /* The Uppsala Parser */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= The Uppsala Parser =&lt;br /&gt;
&lt;br /&gt;
The Uppsala Parser is a neural transition-based dependency parser based on bist-parser by Eli Kiperwasser and Yoav Goldberg and developed primarily in the context of the CoNLL shared tasks on universal dependency parsing in 2017 and 2018. The Uppsala Parser is publicly available at https://github.com/UppsalaNLP/uuparser. &lt;br /&gt;
Note that the version installed here may exhibit some slight differences, designed to improve ease of use.&lt;br /&gt;
&lt;br /&gt;
For the full documentation, see the github page.&lt;br /&gt;
&lt;br /&gt;
== Using the Uppsala Parser on Saga ==&lt;br /&gt;
&lt;br /&gt;
* Log into Saga&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /cluster/shared/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load nlpl-uuparser/2.3.0&lt;br /&gt;
&lt;br /&gt;
== Training a parsing model ==&lt;br /&gt;
&lt;br /&gt;
To train a set of parsing models on treebanks from Universal Dependencies (v2.2 or later):&lt;br /&gt;
&lt;br /&gt;
 uuparser --include [languages to include denoted by their treebank id] --outdir my-output-directory&lt;br /&gt;
&lt;br /&gt;
For example:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments&lt;br /&gt;
&lt;br /&gt;
will train separate models on UD Swedish-Talbanken, UD English-ParTUT and UD Russian-SynTagRus and store the results in the &amp;lt;code&amp;gt;experiments&amp;lt;/code&amp;gt; folder in your home directory. Model selection is included in the training process by default; that is, at each epoch the current model is evaluated on the UD dev data, and at the end of training the best performing model for each language is selected.&lt;br /&gt;
&lt;br /&gt;
== Predicting with a pre-trained parsing model ==&lt;br /&gt;
&lt;br /&gt;
To predict on UD test data with the models trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --predict&lt;br /&gt;
&lt;br /&gt;
== Options ==&lt;br /&gt;
&lt;br /&gt;
The parser has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 uuparser --help | less&lt;br /&gt;
&lt;br /&gt;
We recommend you set the &amp;lt;code&amp;gt;--dynet-mem&amp;lt;/code&amp;gt; option to a larger number when running the full training procedure on larger treebanks. Commonly used values are 5000 and 10000 (in MB). Dynet is the neural network library on which the parser is built.&lt;br /&gt;
&lt;br /&gt;
Note that due to random initialization and other non-deterministic elements in the training process, you will not obtain the same results even when training twice under exactly the same circumstances (e.g. languages, number of epochs etc.). To ensure identical results between two runs, we recommend setting the &amp;lt;code&amp;gt;--dynet-seed&amp;lt;/code&amp;gt; option to the same value both times (e.g. &amp;lt;code&amp;gt; --dynet-seed 123456789&amp;lt;/code&amp;gt;) and adding the &amp;lt;code&amp;gt;--use-default-seed&amp;lt;/code&amp;gt; flag. This ensures that Python's random number generator and Dynet both produce the same sequence of random numbers.&lt;br /&gt;
&lt;br /&gt;
== Training a multitreebank parsing model ==&lt;br /&gt;
&lt;br /&gt;
Our parser supports multitreebank parsing, that is a single parsing model for one or more treebanks, which could be from the same or form different languages. To train a ''multitreebank'' model for the three languages in the examples above, we simply add the &amp;lt;code&amp;gt;--multiling&amp;lt;/code&amp;gt; flag when training (note, though, that these models have so far been found to work best for groups of related languages)&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000&lt;br /&gt;
&lt;br /&gt;
In this case, instead of creating three separate models in the language-specific subdirectories within ~/experiments, a single model will be created directly in this folder. Predicting on test data is then as easy as:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict&lt;br /&gt;
&lt;br /&gt;
Note that if you want to have different output directories for training and predicting, the &amp;lt;code&amp;gt;--modeldir&amp;lt;/code&amp;gt; option can be specified when predicting to tell the parser where the pre-trained model can be found.&lt;br /&gt;
&lt;br /&gt;
== Segmentation ==&lt;br /&gt;
&lt;br /&gt;
In the above examples, we assume pre-segmented input data already in the [http://universaldependencies.org/format.html CONLL-U] format. If your input is raw text, we recommend using UDPipe to segment first. The UDPipe module can be loaded using &amp;lt;code&amp;gt;module load nlpl-udpipe&amp;lt;/code&amp;gt; and then run by typing &amp;lt;code&amp;gt;udpipe&amp;lt;/code&amp;gt; at the command line, see below.&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/turboparser&amp;diff=898</id>
		<title>Parsing/turboparser</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/turboparser&amp;diff=898"/>
		<updated>2020-01-14T12:15:57Z</updated>

		<summary type="html">&lt;p&gt;Sara: Created page with &amp;quot;= TurboParser =  TurboParser is a fast and accurate pre-neural dependency parser with linear programming. The package also contains a POS tagger, a semantic role labeler, a en...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= TurboParser =&lt;br /&gt;
&lt;br /&gt;
TurboParser is a fast and accurate pre-neural dependency parser with linear programming. The package also contains a POS tagger, a semantic role labeler, a entity tagger, a coreference resolver, and a constituent (phrase-based) parser. For full documentation, see [http://www.cs.cmu.edu/~ark/TurboParser/ http://www.cs.cmu.edu/~ark/TurboParser/] and [https://github.com/andre-martins/TurboParser https://github.com/andre-martins/TurboParser]. This document will only describe how to use the tagger and parser.&lt;br /&gt;
&lt;br /&gt;
== Using TurboParser on Saga ==&lt;br /&gt;
&lt;br /&gt;
* Log into Saga&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /cluster/shared/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load nlpl-turboparser/2.3.0&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Data formats and conversion ==&lt;br /&gt;
&lt;br /&gt;
TurboParser takes CoNLL-X files as input. Most dcurrent data, including the universal dependencies data available on Saga is in CoNLLU format. If you want to use such data, you first need to convert between these formats, using the provided script (this script comes from Universal dependenices tools by Dan Zeman, and is included in the TurboParser module):&lt;br /&gt;
&lt;br /&gt;
 conllu_to_conllx.pl &amp;lt; INPUT_FILE.conllu &amp;gt; OUTPUT_FILE.conll&lt;br /&gt;
&lt;br /&gt;
TurboTagger accepts data in  native format with one word per line, and two columns, the words and the tags. There is a script for conversion between CoNLL-X and this format:&lt;br /&gt;
&lt;br /&gt;
 create_tagging_corpus.sh INPUT_FILE.conll&lt;br /&gt;
&lt;br /&gt;
which will create the file &amp;quot;INPUT_FILE.conll.tagging&amp;quot;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Using the parser ==&lt;br /&gt;
&lt;br /&gt;
To train a parsing models on a treebank:&lt;br /&gt;
&lt;br /&gt;
 TurboParser --train --file_train=$res_dir/TRAINING_DATA.conll --file_model=MODEL --logtostderr&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
To predict using the model trained above:&lt;br /&gt;
&lt;br /&gt;
 TurboParser --test --evaluate --file_model=MODEL --file_test=TEST_INPUT_FILE --file_prediction=RESULT_FILE --logtostderr &lt;br /&gt;
&lt;br /&gt;
This command writes the aprsed data in RESULT_FILE, and prints the accuracy (since the --evaluate flag is given)&lt;br /&gt;
&lt;br /&gt;
The parser has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 TurboParser --help &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Using the tagger ==&lt;br /&gt;
&lt;br /&gt;
To train a tagging model on a treebank:&lt;br /&gt;
&lt;br /&gt;
 TurboTagger --train --file_train=TRAINING_DATA.conll.tagging --file_model=MODEL --form_cutoff=1 --logtostderr&lt;br /&gt;
&lt;br /&gt;
To predict using the model trained above:&lt;br /&gt;
&lt;br /&gt;
 TurboTagger --test --evaluate --file_model=MODEL --file_test=TEST_INPUT_FILE --file_prediction=RESULT_FILE --logtostderr &lt;br /&gt;
&lt;br /&gt;
This command writes the tagged data in RESULT_FILE, and prints the accuracy (since the --evaluate flag is given)&lt;br /&gt;
&lt;br /&gt;
The tagger has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 TurboTagger --help &lt;br /&gt;
&lt;br /&gt;
== Segmentation ==&lt;br /&gt;
&lt;br /&gt;
In the above examples, we assume pre-segmented input data already in the appropriate format (see above). If your input is raw text, we recommend using UDPipe to segment first. The UDPipe module can be loaded using &amp;lt;code&amp;gt;module load nlpl-udpipe&amp;lt;/code&amp;gt; and then run by typing &amp;lt;code&amp;gt;udpipe&amp;lt;/code&amp;gt; at the command line, see [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe]&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/abel&amp;diff=897</id>
		<title>Parsing/abel</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/abel&amp;diff=897"/>
		<updated>2020-01-14T10:22:32Z</updated>

		<summary type="html">&lt;p&gt;Sara: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Background =&lt;br /&gt;
&lt;br /&gt;
'''This page is outdated and kept for documentation purposes only! It reflects the status of the translation activity mid-2019, before the launch of Puhti and Saga.'''&lt;br /&gt;
&lt;br /&gt;
This page describes resources previously installed on the Abel cluster. For unlinked resources, see pages for currently available software/data on the main parsing page.&lt;br /&gt;
&lt;br /&gt;
= Preprocessing Tools =&lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/repp REPP Tokenizer (English and Norwegian)]&lt;br /&gt;
&lt;br /&gt;
Additionally, a variety of tools for sentence splitting, tokenization, lemmatization, et al.&lt;br /&gt;
are available through the NLPL installations of the&lt;br /&gt;
[http://nltk.org Natural Language Processing Toolkit (NLTK)] and the&lt;br /&gt;
[https://spacy.io spaCy: Natural Language Processing in Python] tools.&lt;br /&gt;
&lt;br /&gt;
= Parsing Systems =&lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/dozat Stanford Graph-Based Parser by Tim Dozat]&lt;br /&gt;
* uuparser&lt;br /&gt;
* UDPipe&lt;br /&gt;
&lt;br /&gt;
= Data =&lt;br /&gt;
&lt;br /&gt;
* Universal Dependencies treebanks&lt;br /&gt;
* Semantic Dependency parsing&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/home&amp;diff=896</id>
		<title>Parsing/home</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/home&amp;diff=896"/>
		<updated>2020-01-14T10:20:42Z</updated>

		<summary type="html">&lt;p&gt;Sara: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Background =&lt;br /&gt;
&lt;br /&gt;
An experimentation environment for data-driven dependency parsing is maintained for NLPL under the coordination of Uppsala University (UU).&lt;br /&gt;
The data is available on the Norwegian Saga cluster and on the Finnish Puhti cluster.&lt;br /&gt;
The software is available on the Norwegian Saga cluster&lt;br /&gt;
&lt;br /&gt;
Initially, software and data were commissioned on the Norwegian Abel supercluster, see [http://wiki.nlpl.eu/index.php/Parsing/abel The Abel page] for legacy information.&lt;br /&gt;
&lt;br /&gt;
= Preprocessing Tools =&lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe]&lt;br /&gt;
&lt;br /&gt;
Additionally, a variety of tools for sentence splitting, tokenization, lemmatization, et al.&lt;br /&gt;
are available through the NLPL installations of the&lt;br /&gt;
[http://nltk.org Natural Language Processing Toolkit (NLTK)] and the&lt;br /&gt;
[https://spacy.io spaCy: Natural Language Processing in Python] tools.&lt;br /&gt;
&lt;br /&gt;
= Parsing Systems =&lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/uuparser The Uppsala Parser]&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe]&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/turboparser TurboParser]&lt;br /&gt;
&lt;br /&gt;
= Training and Evaluation Data = &lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/ud Universal Dependencies v2.0–2.5]&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/sdp Semantic Dependency Parsing]&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/home&amp;diff=895</id>
		<title>Parsing/home</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/home&amp;diff=895"/>
		<updated>2020-01-14T10:18:01Z</updated>

		<summary type="html">&lt;p&gt;Sara: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Background =&lt;br /&gt;
&lt;br /&gt;
An experimentation environment for data-driven dependency parsing is maintained for NLPL under the coordination of Uppsala University (UU).&lt;br /&gt;
Initially, the software and data are commissioned on the Norwegian Abel supercluster.&lt;br /&gt;
&lt;br /&gt;
= Preprocessing Tools =&lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe]&lt;br /&gt;
&lt;br /&gt;
Additionally, a variety of tools for sentence splitting, tokenization, lemmatization, et al.&lt;br /&gt;
are available through the NLPL installations of the&lt;br /&gt;
[http://nltk.org Natural Language Processing Toolkit (NLTK)] and the&lt;br /&gt;
[https://spacy.io spaCy: Natural Language Processing in Python] tools.&lt;br /&gt;
&lt;br /&gt;
= Parsing Systems =&lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/uuparser The Uppsala Parser]&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe]&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/turboparser TurboParser]&lt;br /&gt;
&lt;br /&gt;
= Training and Evaluation Data = &lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/ud Universal Dependencies v2.0–2.5]&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/sdp Semantic Dependency Parsing]&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/abel&amp;diff=894</id>
		<title>Parsing/abel</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/abel&amp;diff=894"/>
		<updated>2020-01-14T10:16:39Z</updated>

		<summary type="html">&lt;p&gt;Sara: Created page with &amp;quot;= Background =  '''This page is outdated and kept for documentation purposes only! It reflects the status of the translation activity mid-2019, before the launch of Puhti and ...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Background =&lt;br /&gt;
&lt;br /&gt;
'''This page is outdated and kept for documentation purposes only! It reflects the status of the translation activity mid-2019, before the launch of Puhti and Saga.'''&lt;br /&gt;
&lt;br /&gt;
This page describes resources previously installed on the Abel cluster.&lt;br /&gt;
&lt;br /&gt;
= Preprocessing Tools =&lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/repp REPP Tokenizer (English and Norwegian)]&lt;br /&gt;
&lt;br /&gt;
Additionally, a variety of tools for sentence splitting, tokenization, lemmatization, et al.&lt;br /&gt;
are available through the NLPL installations of the&lt;br /&gt;
[http://nltk.org Natural Language Processing Toolkit (NLTK)] and the&lt;br /&gt;
[https://spacy.io spaCy: Natural Language Processing in Python] tools.&lt;br /&gt;
&lt;br /&gt;
= Parsing Systems =&lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/dozat Stanford Graph-Based Parser by Tim Dozat]&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=893</id>
		<title>Parsing/uuparser</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/uuparser&amp;diff=893"/>
		<updated>2020-01-13T13:03:00Z</updated>

		<summary type="html">&lt;p&gt;Sara: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= The Uppsala Parser =&lt;br /&gt;
&lt;br /&gt;
The Uppsala Parser is a neural transition-based dependency parser based on bist-parser by Eli Kiperwasser and Yoav Goldberg and developed primarily in the context of the CoNLL shared tasks on universal dependency parsing in 2017 and 2018. The Uppsala Parser is publicly available at https://github.com/UppsalaNLP/uuparser. &lt;br /&gt;
Note that the version installed here may exhibit some slight differences, designed to improve ease of use.&lt;br /&gt;
&lt;br /&gt;
== Using the Uppsala Parser on Saga ==&lt;br /&gt;
&lt;br /&gt;
* Log into Saga&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /cluster/shared/nlpl/software/modules/etc&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load nlpl-uuparser/2.3.0&lt;br /&gt;
&lt;br /&gt;
== Training a parsing model ==&lt;br /&gt;
&lt;br /&gt;
To train a set of parsing models on treebanks from Universal Dependencies (v2.2 or later):&lt;br /&gt;
&lt;br /&gt;
 uuparser --include [languages to include denoted by their treebank id] --outdir my-output-directory&lt;br /&gt;
&lt;br /&gt;
For example:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments&lt;br /&gt;
&lt;br /&gt;
will train separate models on UD Swedish-Talbanken, UD English-ParTUT and UD Russian-SynTagRus and store the results in the &amp;lt;code&amp;gt;experiments&amp;lt;/code&amp;gt; folder in your home directory. Model selection is included in the training process by default; that is, at each epoch the current model is evaluated on the UD dev data, and at the end of training the best performing model for each language is selected.&lt;br /&gt;
&lt;br /&gt;
== Predicting with a pre-trained parsing model ==&lt;br /&gt;
&lt;br /&gt;
To predict on UD test data with the models trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --predict&lt;br /&gt;
&lt;br /&gt;
== Options ==&lt;br /&gt;
&lt;br /&gt;
The parser has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 uuparser --help | less&lt;br /&gt;
&lt;br /&gt;
We recommend you set the &amp;lt;code&amp;gt;--dynet-mem&amp;lt;/code&amp;gt; option to a larger number when running the full training procedure on larger treebanks. Commonly used values are 5000 and 10000 (in MB). Dynet is the neural network library on which the parser is built.&lt;br /&gt;
&lt;br /&gt;
Note that due to random initialization and other non-deterministic elements in the training process, you will not obtain the same results even when training twice under exactly the same circumstances (e.g. languages, number of epochs etc.). To ensure identical results between two runs, we recommend setting the &amp;lt;code&amp;gt;--dynet-seed&amp;lt;/code&amp;gt; option to the same value both times (e.g. &amp;lt;code&amp;gt; --dynet-seed 123456789&amp;lt;/code&amp;gt;) and adding the &amp;lt;code&amp;gt;--use-default-seed&amp;lt;/code&amp;gt; flag. This ensures that Python's random number generator and Dynet both produce the same sequence of random numbers.&lt;br /&gt;
&lt;br /&gt;
== Training a multitreebank parsing model ==&lt;br /&gt;
&lt;br /&gt;
Our parser supports multitreebank parsing, that is a single parsing model for one or more treebanks, which could be from the same or form different languages. To train a ''multitreebank'' model for the three languages in the examples above, we simply add the &amp;lt;code&amp;gt;--multiling&amp;lt;/code&amp;gt; flag when training (note, though, that these models have so far been found to work best for groups of related languages)&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000&lt;br /&gt;
&lt;br /&gt;
In this case, instead of creating three separate models in the language-specific subdirectories within ~/experiments, a single model will be created directly in this folder. Predicting on test data is then as easy as:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv_talbanken en_partut ru_syntagrus&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict&lt;br /&gt;
&lt;br /&gt;
Note that if you want to have different output directories for training and predicting, the &amp;lt;code&amp;gt;--modeldir&amp;lt;/code&amp;gt; option can be specified when predicting to tell the parser where the pre-trained model can be found.&lt;br /&gt;
&lt;br /&gt;
== Segmentation ==&lt;br /&gt;
&lt;br /&gt;
In the above examples, we assume pre-segmented input data already in the [http://universaldependencies.org/format.html CONLL-U] format. If your input is raw text, we recommend using UDPipe to segment first. The UDPipe module can be loaded using &amp;lt;code&amp;gt;module load nlpl-udpipe&amp;lt;/code&amp;gt; and then run by typing &amp;lt;code&amp;gt;udpipe&amp;lt;/code&amp;gt; at the command line, see below.&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/home&amp;diff=892</id>
		<title>Parsing/home</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/home&amp;diff=892"/>
		<updated>2020-01-13T12:54:23Z</updated>

		<summary type="html">&lt;p&gt;Sara: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Background =&lt;br /&gt;
&lt;br /&gt;
An experimentation environment for data-driven dependency parsing is maintained for NLPL under the coordination of Uppsala University (UU).&lt;br /&gt;
Initially, the software and data are commissioned on the Norwegian Abel supercluster.&lt;br /&gt;
&lt;br /&gt;
= Preprocessing Tools =&lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/repp REPP Tokenizer (English and Norwegian)]&lt;br /&gt;
&lt;br /&gt;
Additionally, a variety of tools for sentence splitting, tokenization, lemmatization, et al.&lt;br /&gt;
are available through the NLPL installations of the&lt;br /&gt;
[http://nltk.org Natural Language Processing Toolkit (NLTK)] and the&lt;br /&gt;
[https://spacy.io spaCy: Natural Language Processing in Python] tools.&lt;br /&gt;
&lt;br /&gt;
= Parsing Systems =&lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/uuparser The Uppsala Parser]&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/udpipe UDPipe]&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/dozat Stanford Graph-Based Parser by Tim Dozat]&lt;br /&gt;
&lt;br /&gt;
= Training and Evaluation Data = &lt;br /&gt;
&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/ud Universal Dependencies v2.0–2.5]&lt;br /&gt;
* [http://wiki.nlpl.eu/index.php/Parsing/sdp Semantic Dependency Parsing]&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/ud&amp;diff=839</id>
		<title>Parsing/ud</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/ud&amp;diff=839"/>
		<updated>2019-12-17T10:23:15Z</updated>

		<summary type="html">&lt;p&gt;Sara: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Universal Dependencies =&lt;br /&gt;
&lt;br /&gt;
For syntactic parsing experiments we provide data from the [http://universaldependencies.org/ Universal Dependencies (UD) project] for a high number of languages. The data is provided in v2.0, which was used for the CoNLL shared task 2017, v2.1, v2.2, which was used for the CoNLL shared task 2018, v2.3, v2.4, and v2.5. &lt;br /&gt;
&lt;br /&gt;
All data is available on Saga at &amp;lt;code&amp;gt;/cluster/shared/nlpl/data/parsing/ud&amp;lt;/code&amp;gt; and automatically&lt;br /&gt;
[http://wiki.nlpl.eu/index.php/Infrastructure/replication replicated] to Puhti into &amp;lt;code&amp;gt;/projappl/nlpl/data/parsing/ud&amp;lt;code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== UD version 2.0 ==&lt;br /&gt;
&lt;br /&gt;
folders:&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;code&amp;gt;/cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.0-conll2017&amp;lt;/code&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;code&amp;gt;/cluster/shared/nlpl/data/parsing/ud/ud-test-v2.0-conll2017&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
info:&amp;lt;br&amp;gt;&lt;br /&gt;
Version 2.0 treebanks, archived at http://hdl.handle.net/11234/1-1983. &amp;lt;br&amp;gt;&lt;br /&gt;
70 treebanks, 50 languages, released March 1, 2017.&amp;lt;br&amp;gt;&lt;br /&gt;
Test data 2.0 are archived at http://hdl.handle.net/11234/1-2184. &amp;lt;br&amp;gt;&lt;br /&gt;
81 treebanks, 49 languages, released May 18, 2017.&lt;br /&gt;
&lt;br /&gt;
Release 2.0 has test data released separately from the test data,&lt;br /&gt;
which is reflected in our folder structure. This data was released for&lt;br /&gt;
the CoNLL 2017 shared task.&lt;br /&gt;
&lt;br /&gt;
== UD version 2.1 ==&lt;br /&gt;
&lt;br /&gt;
folders:&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;code&amp;gt;/cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.1&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
info:&amp;lt;br&amp;gt;&lt;br /&gt;
Version 2.1 treebanks are available at http://hdl.handle.net/11234/1-2515. &amp;lt;br&amp;gt;&lt;br /&gt;
102 treebanks, 60 languages, released November 15, 2017.&lt;br /&gt;
&lt;br /&gt;
== UD version 2.2 ==&lt;br /&gt;
&lt;br /&gt;
folders:&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;code&amp;gt;/cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.2&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
info:&amp;lt;br&amp;gt;&lt;br /&gt;
Version 2.2 treebanks are available at http://hdl.handle.net/11234/1-2837. &amp;lt;br&amp;gt;&lt;br /&gt;
122 treebanks, 71 languages, released July 1, 2018.&lt;br /&gt;
&lt;br /&gt;
== UD version 2.3 ==&lt;br /&gt;
&lt;br /&gt;
folders:&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;code&amp;gt;/cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.3&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
info:&amp;lt;br&amp;gt;&lt;br /&gt;
Version 2.2 treebanks are available at http://hdl.handle.net/11234/1-2895. &amp;lt;br&amp;gt;&lt;br /&gt;
129 treebanks, 76 languages, released November 15, 2018.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== UD version 2.4 ==&lt;br /&gt;
&lt;br /&gt;
folders:&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;code&amp;gt;/cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.4&amp;lt;/code&amp;gt;    &lt;br /&gt;
&lt;br /&gt;
info:&amp;lt;br&amp;gt;&lt;br /&gt;
Version 2.4 treebanks are available at http://hdl.handle.net/11234/1-2988. &amp;lt;br&amp;gt;&lt;br /&gt;
146 treebanks, 83 languages, released May 15, 2019.&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== UD version  2.5 ==&lt;br /&gt;
&lt;br /&gt;
folders:&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;code&amp;gt;/cluster/shared/nlpl/data/parsing/ud/ud-treebanks-v2.5&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
info:&amp;lt;br&amp;gt;&lt;br /&gt;
Version 2.5 treebanks are available at http://hdl.handle.net/11234/1-3105. &amp;lt;br&amp;gt;&lt;br /&gt;
157 treebanks, 90 languages, released November 15, 2019.&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Contact =&lt;br /&gt;
Joakim Nivre, Uppsala University&amp;lt;br/&amp;gt;&lt;br /&gt;
Sara Stymne, Uppsala University&amp;lt;br/&amp;gt;&lt;br /&gt;
firstname.lastname@lingfil.uu.se&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/home&amp;diff=122</id>
		<title>Parsing/home</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/home&amp;diff=122"/>
		<updated>2017-12-08T09:10:18Z</updated>

		<summary type="html">&lt;p&gt;Sara: /* Segmentation */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Background =&lt;br /&gt;
&lt;br /&gt;
An experimentation environment for data-driven dependency parsing is maintained for NLPL under the coordination of Uppsala University (UU).&lt;br /&gt;
Initially, the software and data are commissioned on the Norwegian Abel supercluster.&lt;br /&gt;
The Uppsala Parser is publicly available at https://github.com/UppsalaNLP/uuparser. &lt;br /&gt;
Note that the version installed here may exhibit some slight differences, designed to improve ease of use.&lt;br /&gt;
&lt;br /&gt;
= Data =&lt;br /&gt;
&lt;br /&gt;
For parsing experiments data from the [Univeral Dependencies (UD) project](http://universaldependencies.org/), for a high number of languages. We provide the data in version 2.0, which was also used for the CoNLL shared task 2017, and version 2.1. &lt;br /&gt;
&lt;br /&gt;
All data is available on Abel at &amp;lt;code&amp;gt;/projects/nlpl/data/parsing/universal_dependencies&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== UD version 2.0 ==&lt;br /&gt;
&lt;br /&gt;
folders:&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;code&amp;gt;/projects/nlpl/data/parsing/universal_dependencies/ud-treebanks-v2.0-conll2017&amp;lt;/code&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;code&amp;gt;/projects/nlpl/data/parsing/universal_dependencies/ud-test-v2.0-conll2017&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
info:&amp;lt;br&amp;gt;&lt;br /&gt;
Version 2.0 treebanks, archived at http://hdl.handle.net/11234/1-1983. &amp;lt;br&amp;gt;&lt;br /&gt;
70 treebanks, 50 languages, released March 1, 2017.&amp;lt;br&amp;gt;&lt;br /&gt;
Test data 2.0 are archived at http://hdl.handle.net/11234/1-2184. &amp;lt;br&amp;gt;&lt;br /&gt;
81 treebanks, 49 languages, released May 18, 2017.&lt;br /&gt;
&lt;br /&gt;
Release 2.0 has test data released separately from the test data,&lt;br /&gt;
which is reflected in our folder structure. This data was released for&lt;br /&gt;
the CoNLL 2017 shared task.&lt;br /&gt;
&lt;br /&gt;
== UD version 2.1 ==&lt;br /&gt;
&lt;br /&gt;
folders:&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;code&amp;gt;/projects/nlpl/data/parsing/universal_dependencies/ud-treebanks-v2.1&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
info:&amp;lt;br&amp;gt;&lt;br /&gt;
Version 2.1 treebanks are available at http://hdl.handle.net/11234/1-2515. &amp;lt;br&amp;gt;&lt;br /&gt;
102 treebanks, 60 languages, released November 15, 2017.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Using the Uppsala Parser =&lt;br /&gt;
&lt;br /&gt;
* Log into Abel&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /projects/nlpl/software/modulefiles/&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load uuparser&lt;br /&gt;
&lt;br /&gt;
== Training a parsing model ==&lt;br /&gt;
&lt;br /&gt;
To train a set of parsing models on treebanks from Universal Dependencies (v2.1):&lt;br /&gt;
&lt;br /&gt;
 uuparser --include [languages to include denoted by their ISO id] --outdir my-output-directory&lt;br /&gt;
&lt;br /&gt;
If you want to quickly test the parser is correctly loaded and running without waiting for the full training procedure, add the &amp;lt;code&amp;gt;--mini&amp;lt;/code&amp;gt; flag.&lt;br /&gt;
For example:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv en ru&amp;quot; --outdir ~/experiments --mini&lt;br /&gt;
&lt;br /&gt;
will train separate models for UD Swedish, English and Russian and store the results in the &amp;lt;code&amp;gt;experiments&amp;lt;/code&amp;gt; folder in your home directory. The &amp;lt;code&amp;gt;--mini&amp;lt;/code&amp;gt; flag tells the parser to train on just the first 150 sentences of each language, and evaluate on the first 100 sentences of development data. It also tells the parser to train for just 3 epochs, as opposed to the default 30 (see more below under &amp;quot;Options&amp;quot;)&lt;br /&gt;
&lt;br /&gt;
Model selection is included in the training process by default; that is, at each epoch the current model is evaluated on the UD dev data, and at the end of training the best performing model for each language is selected. &lt;br /&gt;
&lt;br /&gt;
== Predicting with a pre-trained parsing model ==&lt;br /&gt;
&lt;br /&gt;
To predict on UD test data with the models trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv en ru&amp;quot; --outdir ~/experiments --predict&lt;br /&gt;
&lt;br /&gt;
You may again include the &amp;lt;code&amp;gt;--mini&amp;lt;/code&amp;gt; flag if you prefer to test on a subset of 50 test sentences.&lt;br /&gt;
&lt;br /&gt;
== Options ==&lt;br /&gt;
&lt;br /&gt;
The parser has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 uuparser --help | less&lt;br /&gt;
&lt;br /&gt;
We recommend you set the &amp;lt;code&amp;gt;--dynet-mem&amp;lt;/code&amp;gt; option to a larger number when running the full training procedure on larger treebanks. Commonly used values are 5000 and 10000 (in MB). Dynet is the neural network library on which the parser is built.&lt;br /&gt;
&lt;br /&gt;
Note that due to random initialization and other non-deterministic elements in the training process, you will not obtain the same results even when training twice under exactly the same circumstances (e.g. languages, number of epochs etc.). To ensure identical results between two runs, we recommend setting the &amp;lt;code&amp;gt;--dynet-seed&amp;lt;/code&amp;gt; option to the same value both times (e.g. &amp;lt;code&amp;gt; --dynet-seed 123456789&amp;lt;/code&amp;gt;) and adding the &amp;lt;code&amp;gt;--use-default-seed&amp;lt;/code&amp;gt; flag. This ensures that Python's random number generator and Dynet both produce the same sequence of random numbers.&lt;br /&gt;
&lt;br /&gt;
== Training a multilingual parsing model ==&lt;br /&gt;
&lt;br /&gt;
Our parser supports multilingual parsing, that is a single parsing model for one or more languages. To train a ''multilingual'' model for the three languages in the examples above, we simply add the &amp;lt;code&amp;gt;--multilingual&amp;lt;/code&amp;gt; flag when training&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv en ru&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000&lt;br /&gt;
&lt;br /&gt;
In this case, instead of creating three separate models in the language-specific subdirectories within ~/experiments, a single model will be created directly in this folder. Predicting on test data is then as easy as:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv en ru&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict&lt;br /&gt;
&lt;br /&gt;
Note that if you want to have different output directories for training and predicting, the &amp;lt;code&amp;gt;--modeldir&amp;lt;/code&amp;gt; option can be specified when predicting to tell the parser where the pre-trained model can be found.&lt;br /&gt;
&lt;br /&gt;
== Segmentation ==&lt;br /&gt;
&lt;br /&gt;
In the above examples, we assume pre-segmented input data already in the [http://universaldependencies.org/format.html CONLL-U] format. If your input is raw text, we recommend using UDPipe to segment first. The UDPipe module can be loaded using &amp;lt;code&amp;gt;module load nlpl-udpipe&amp;lt;/code&amp;gt; and then run by typing &amp;lt;code&amp;gt;udpipe&amp;lt;/code&amp;gt; at the command line, see below.&lt;br /&gt;
&lt;br /&gt;
= Using UDPipe =&lt;br /&gt;
&lt;br /&gt;
UDPipe is available as a module on Abel. It was installed as part of the OPUS activity.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
How to use UDPipe:&lt;br /&gt;
* Log into Abel&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /projects/nlpl/software/modulefiles/&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load nlpl-udpipe&lt;br /&gt;
&lt;br /&gt;
To learn more about using UDPipe, check the official [UDPipe User's Manual](http://ufal.mff.cuni.cz/udpipe/users-manual)&lt;br /&gt;
&lt;br /&gt;
= Contact =&lt;br /&gt;
Aaron Smith, Uppsala University, firstname.lastname@lingfil.uu.se&amp;lt;br&amp;gt;&lt;br /&gt;
Sara Stymne, Uppsala University, firstname.lastname@lingfil.uu.se&amp;lt;br&amp;gt;&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/home&amp;diff=121</id>
		<title>Parsing/home</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/home&amp;diff=121"/>
		<updated>2017-12-08T09:09:48Z</updated>

		<summary type="html">&lt;p&gt;Sara: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Background =&lt;br /&gt;
&lt;br /&gt;
An experimentation environment for data-driven dependency parsing is maintained for NLPL under the coordination of Uppsala University (UU).&lt;br /&gt;
Initially, the software and data are commissioned on the Norwegian Abel supercluster.&lt;br /&gt;
The Uppsala Parser is publicly available at https://github.com/UppsalaNLP/uuparser. &lt;br /&gt;
Note that the version installed here may exhibit some slight differences, designed to improve ease of use.&lt;br /&gt;
&lt;br /&gt;
= Data =&lt;br /&gt;
&lt;br /&gt;
For parsing experiments data from the [Univeral Dependencies (UD) project](http://universaldependencies.org/), for a high number of languages. We provide the data in version 2.0, which was also used for the CoNLL shared task 2017, and version 2.1. &lt;br /&gt;
&lt;br /&gt;
All data is available on Abel at &amp;lt;code&amp;gt;/projects/nlpl/data/parsing/universal_dependencies&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== UD version 2.0 ==&lt;br /&gt;
&lt;br /&gt;
folders:&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;code&amp;gt;/projects/nlpl/data/parsing/universal_dependencies/ud-treebanks-v2.0-conll2017&amp;lt;/code&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;code&amp;gt;/projects/nlpl/data/parsing/universal_dependencies/ud-test-v2.0-conll2017&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
info:&amp;lt;br&amp;gt;&lt;br /&gt;
Version 2.0 treebanks, archived at http://hdl.handle.net/11234/1-1983. &amp;lt;br&amp;gt;&lt;br /&gt;
70 treebanks, 50 languages, released March 1, 2017.&amp;lt;br&amp;gt;&lt;br /&gt;
Test data 2.0 are archived at http://hdl.handle.net/11234/1-2184. &amp;lt;br&amp;gt;&lt;br /&gt;
81 treebanks, 49 languages, released May 18, 2017.&lt;br /&gt;
&lt;br /&gt;
Release 2.0 has test data released separately from the test data,&lt;br /&gt;
which is reflected in our folder structure. This data was released for&lt;br /&gt;
the CoNLL 2017 shared task.&lt;br /&gt;
&lt;br /&gt;
== UD version 2.1 ==&lt;br /&gt;
&lt;br /&gt;
folders:&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;code&amp;gt;/projects/nlpl/data/parsing/universal_dependencies/ud-treebanks-v2.1&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
info:&amp;lt;br&amp;gt;&lt;br /&gt;
Version 2.1 treebanks are available at http://hdl.handle.net/11234/1-2515. &amp;lt;br&amp;gt;&lt;br /&gt;
102 treebanks, 60 languages, released November 15, 2017.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Using the Uppsala Parser =&lt;br /&gt;
&lt;br /&gt;
* Log into Abel&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /projects/nlpl/software/modulefiles/&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load uuparser&lt;br /&gt;
&lt;br /&gt;
== Training a parsing model ==&lt;br /&gt;
&lt;br /&gt;
To train a set of parsing models on treebanks from Universal Dependencies (v2.1):&lt;br /&gt;
&lt;br /&gt;
 uuparser --include [languages to include denoted by their ISO id] --outdir my-output-directory&lt;br /&gt;
&lt;br /&gt;
If you want to quickly test the parser is correctly loaded and running without waiting for the full training procedure, add the &amp;lt;code&amp;gt;--mini&amp;lt;/code&amp;gt; flag.&lt;br /&gt;
For example:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv en ru&amp;quot; --outdir ~/experiments --mini&lt;br /&gt;
&lt;br /&gt;
will train separate models for UD Swedish, English and Russian and store the results in the &amp;lt;code&amp;gt;experiments&amp;lt;/code&amp;gt; folder in your home directory. The &amp;lt;code&amp;gt;--mini&amp;lt;/code&amp;gt; flag tells the parser to train on just the first 150 sentences of each language, and evaluate on the first 100 sentences of development data. It also tells the parser to train for just 3 epochs, as opposed to the default 30 (see more below under &amp;quot;Options&amp;quot;)&lt;br /&gt;
&lt;br /&gt;
Model selection is included in the training process by default; that is, at each epoch the current model is evaluated on the UD dev data, and at the end of training the best performing model for each language is selected. &lt;br /&gt;
&lt;br /&gt;
== Predicting with a pre-trained parsing model ==&lt;br /&gt;
&lt;br /&gt;
To predict on UD test data with the models trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv en ru&amp;quot; --outdir ~/experiments --predict&lt;br /&gt;
&lt;br /&gt;
You may again include the &amp;lt;code&amp;gt;--mini&amp;lt;/code&amp;gt; flag if you prefer to test on a subset of 50 test sentences.&lt;br /&gt;
&lt;br /&gt;
== Options ==&lt;br /&gt;
&lt;br /&gt;
The parser has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 uuparser --help | less&lt;br /&gt;
&lt;br /&gt;
We recommend you set the &amp;lt;code&amp;gt;--dynet-mem&amp;lt;/code&amp;gt; option to a larger number when running the full training procedure on larger treebanks. Commonly used values are 5000 and 10000 (in MB). Dynet is the neural network library on which the parser is built.&lt;br /&gt;
&lt;br /&gt;
Note that due to random initialization and other non-deterministic elements in the training process, you will not obtain the same results even when training twice under exactly the same circumstances (e.g. languages, number of epochs etc.). To ensure identical results between two runs, we recommend setting the &amp;lt;code&amp;gt;--dynet-seed&amp;lt;/code&amp;gt; option to the same value both times (e.g. &amp;lt;code&amp;gt; --dynet-seed 123456789&amp;lt;/code&amp;gt;) and adding the &amp;lt;code&amp;gt;--use-default-seed&amp;lt;/code&amp;gt; flag. This ensures that Python's random number generator and Dynet both produce the same sequence of random numbers.&lt;br /&gt;
&lt;br /&gt;
== Training a multilingual parsing model ==&lt;br /&gt;
&lt;br /&gt;
Our parser supports multilingual parsing, that is a single parsing model for one or more languages. To train a ''multilingual'' model for the three languages in the examples above, we simply add the &amp;lt;code&amp;gt;--multilingual&amp;lt;/code&amp;gt; flag when training&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv en ru&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000&lt;br /&gt;
&lt;br /&gt;
In this case, instead of creating three separate models in the language-specific subdirectories within ~/experiments, a single model will be created directly in this folder. Predicting on test data is then as easy as:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv en ru&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict&lt;br /&gt;
&lt;br /&gt;
Note that if you want to have different output directories for training and predicting, the &amp;lt;code&amp;gt;--modeldir&amp;lt;/code&amp;gt; option can be specified when predicting to tell the parser where the pre-trained model can be found.&lt;br /&gt;
&lt;br /&gt;
== Segmentation ==&lt;br /&gt;
&lt;br /&gt;
In the above examples, we assume pre-segmented input data already in the [http://universaldependencies.org/format.html CONLL-U] format. If your input is raw text, we recommend using UDPipe to segment first. The UDPipe module can be loaded using &amp;lt;code&amp;gt;module load nlpl-udpipe&amp;lt;/code&amp;gt; and then run by typing &amp;lt;code&amp;gt;udpipe&amp;lt;/code&amp;gt; at the command line.&lt;br /&gt;
&lt;br /&gt;
= Using UDPipe =&lt;br /&gt;
&lt;br /&gt;
UDPipe is available as a module on Abel. It was installed as part of the OPUS activity.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
How to use UDPipe:&lt;br /&gt;
* Log into Abel&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /projects/nlpl/software/modulefiles/&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load nlpl-udpipe&lt;br /&gt;
&lt;br /&gt;
To learn more about using UDPipe, check the official [UDPipe User's Manual](http://ufal.mff.cuni.cz/udpipe/users-manual)&lt;br /&gt;
&lt;br /&gt;
= Contact =&lt;br /&gt;
Aaron Smith, Uppsala University, firstname.lastname@lingfil.uu.se&amp;lt;br&amp;gt;&lt;br /&gt;
Sara Stymne, Uppsala University, firstname.lastname@lingfil.uu.se&amp;lt;br&amp;gt;&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Parsing/home&amp;diff=119</id>
		<title>Parsing/home</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Parsing/home&amp;diff=119"/>
		<updated>2017-12-08T08:57:37Z</updated>

		<summary type="html">&lt;p&gt;Sara: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Background =&lt;br /&gt;
&lt;br /&gt;
An experimentation environment for data-driven dependency parsing is maintained for NLPL under the coordination of Uppsala University (UU).&lt;br /&gt;
Initially, the software and data are commissioned on the Norwegian Abel supercluster.&lt;br /&gt;
The Uppsala Parser is publicly available at https://github.com/UppsalaNLP/uuparser. &lt;br /&gt;
Note that the version installed here may exhibit some slight differences, designed to improve ease of use.&lt;br /&gt;
&lt;br /&gt;
= Data =&lt;br /&gt;
&lt;br /&gt;
For parsing experiments data from the Univeral Dependencies (UD) project, for a high number of languages. We provide the data in version 2.0, which was also used for the CoNLL shared task 2017, and version 2.1. &lt;br /&gt;
&lt;br /&gt;
All data is available on Abel at &amp;lt;code&amp;gt;/projects/nlpl/data/parsing/universal_dependencies&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== UD version 2.0 ==&lt;br /&gt;
&lt;br /&gt;
folders:&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;code&amp;gt;/projects/nlpl/data/parsing/universal_dependencies/ud-treebanks-v2.0-conll2017&amp;lt;/code&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;code&amp;gt;/projects/nlpl/data/parsing/universal_dependencies/ud-test-v2.0-conll2017&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
info:&amp;lt;br&amp;gt;&lt;br /&gt;
Version 2.0 treebanks, archived at http://hdl.handle.net/11234/1-1983. &amp;lt;br&amp;gt;&lt;br /&gt;
70 treebanks, 50 languages, released March 1, 2017.&amp;lt;br&amp;gt;&lt;br /&gt;
Test data 2.0 are archived at http://hdl.handle.net/11234/1-2184. &amp;lt;br&amp;gt;&lt;br /&gt;
81 treebanks, 49 languages, released May 18, 2017.&lt;br /&gt;
&lt;br /&gt;
Release 2.0 has test data released separately from the test data,&lt;br /&gt;
which is reflected in our folder structure. This data was released for&lt;br /&gt;
the CoNLL 2017 shared task.&lt;br /&gt;
&lt;br /&gt;
== UD version 2.1 ==&lt;br /&gt;
&lt;br /&gt;
folders:&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;code&amp;gt;/projects/nlpl/data/parsing/universal_dependencies/ud-treebanks-v2.1&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
info:&amp;lt;br&amp;gt;&lt;br /&gt;
Version 2.1 treebanks are available at http://hdl.handle.net/11234/1-2515. &amp;lt;br&amp;gt;&lt;br /&gt;
102 treebanks, 60 languages, released November 15, 2017.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Using the Uppsala Parser =&lt;br /&gt;
&lt;br /&gt;
* Log into Abel&lt;br /&gt;
* Activate the NLPL module repository:&lt;br /&gt;
 module use -a /projects/nlpl/software/modulefiles/&lt;br /&gt;
* Load the most recent version of the uuparser module:&lt;br /&gt;
 module load uuparser&lt;br /&gt;
&lt;br /&gt;
== Training a parsing model ==&lt;br /&gt;
&lt;br /&gt;
To train a set of parsing models on treebanks from Universal Dependencies (v2.1):&lt;br /&gt;
&lt;br /&gt;
 uuparser --include [languages to include denoted by their ISO id] --outdir my-output-directory&lt;br /&gt;
&lt;br /&gt;
If you want to quickly test the parser is correctly loaded and running without waiting for the full training procedure, add the &amp;lt;code&amp;gt;--mini&amp;lt;/code&amp;gt; flag.&lt;br /&gt;
For example:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv en ru&amp;quot; --outdir ~/experiments --mini&lt;br /&gt;
&lt;br /&gt;
will train separate models for UD Swedish, English and Russian and store the results in the &amp;lt;code&amp;gt;experiments&amp;lt;/code&amp;gt; folder in your home directory. The &amp;lt;code&amp;gt;--mini&amp;lt;/code&amp;gt; flag tells the parser to train on just the first 150 sentences of each language, and evaluate on the first 100 sentences of development data. It also tells the parser to train for just 3 epochs, as opposed to the default 30 (see more below under &amp;quot;Options&amp;quot;)&lt;br /&gt;
&lt;br /&gt;
Model selection is included in the training process by default; that is, at each epoch the current model is evaluated on the UD dev data, and at the end of training the best performing model for each language is selected. &lt;br /&gt;
&lt;br /&gt;
== Predicting with a pre-trained parsing model ==&lt;br /&gt;
&lt;br /&gt;
To predict on UD test data with the models trained above:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv en ru&amp;quot; --outdir ~/experiments --predict&lt;br /&gt;
&lt;br /&gt;
You may again include the &amp;lt;code&amp;gt;--mini&amp;lt;/code&amp;gt; flag if you prefer to test on a subset of 50 test sentences.&lt;br /&gt;
&lt;br /&gt;
== Options ==&lt;br /&gt;
&lt;br /&gt;
The parser has numerous options to allow you to fine-control its behaviour. For a full list, type&lt;br /&gt;
&lt;br /&gt;
 uuparser --help | less&lt;br /&gt;
&lt;br /&gt;
We recommend you set the &amp;lt;code&amp;gt;--dynet-mem&amp;lt;/code&amp;gt; option to a larger number when training on larger treebanks. Commonly used values are 5000 and 10000 (in MB). Dynet is the neural network library on which the parser is built.&lt;br /&gt;
&lt;br /&gt;
Note that due to random initialization and other non-deterministic elements in the training process, you will not obtain the same results even when training twice under exactly the same circumstances (e.g. languages, number of epochs etc.). To ensure identical results between two runs, we recommend setting the &amp;lt;code&amp;gt;--dynet-seed&amp;lt;/code&amp;gt; option to the same value both times (e.g. &amp;lt;code&amp;gt; --dynet-seed 123456789&amp;lt;/code&amp;gt;) and adding the &amp;lt;code&amp;gt;--use-default-seed&amp;lt;/code&amp;gt; flag. This ensures that Python's random number generator and Dynet both produce the same sequence of random numbers.&lt;br /&gt;
&lt;br /&gt;
== Training a multilingual parsing model ==&lt;br /&gt;
&lt;br /&gt;
Our parser supports multilingual parsing, that is a single parsing model for one or more languages. To train a ''multilingual'' model for the three languages in the examples above, we simply add the &amp;lt;code&amp;gt;--multilingual&amp;lt;/code&amp;gt; flag when training&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv en ru&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000&lt;br /&gt;
&lt;br /&gt;
In this case, instead of creating three separate models in the language-specific subdirectories created within ~/experiments, a single model will be created directly in this folder. Predicting on test data is then as easy as:&lt;br /&gt;
&lt;br /&gt;
 uuparser --include &amp;quot;sv en ru&amp;quot; --outdir ~/experiments --multiling --dynet-mem 5000 --predict&lt;br /&gt;
&lt;br /&gt;
Note that if you want to have different output directories for training and predicting, the &amp;lt;code&amp;gt;--modeldir&amp;lt;/code&amp;gt; option can be specified when predicting to tell the parser where the pre-trained model can be found.&lt;br /&gt;
&lt;br /&gt;
'''Contact:'''&lt;br /&gt;
Aaron Smith, Uppsala University, firstname.lastname@lingfil.uu.se&lt;/div&gt;</summary>
		<author><name>Sara</name></author>
		
	</entry>
</feed>