<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://wiki.nlpl.eu/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Joerg</id>
	<title>Nordic Language Processing Laboratory - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://wiki.nlpl.eu/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Joerg"/>
	<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/Special:Contributions/Joerg"/>
	<updated>2026-04-25T17:38:39Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.31.10</generator>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Infrastructure/software/eosc&amp;diff=761</id>
		<title>Infrastructure/software/eosc</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Infrastructure/software/eosc&amp;diff=761"/>
		<updated>2019-10-18T07:26:46Z</updated>

		<summary type="html">&lt;p&gt;Joerg: /* Data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Background =&lt;br /&gt;
&lt;br /&gt;
This page provides a working document for requirements in the NLP(L) use case in the EOSC Nordic project.&lt;br /&gt;
&lt;br /&gt;
The NLPL research community (in late 2019) is comprised of many dozens of active users, ranging from&lt;br /&gt;
MSc students to professors; there is much variation in computational experience and ‘Un*x foo’.&lt;br /&gt;
Likewise, computing tasks vary a lot, ranging from maybe a handful of single-cpu jobs to&lt;br /&gt;
thousands of (mildly) parallel or multi-gpu tasks; NLP research quite generally is both data- and compute-intensive.&lt;br /&gt;
&lt;br /&gt;
Typical types of data include potentially large&lt;br /&gt;
[http://wiki.nlpl.eu/index.php/Corpora/home document collections] (for example 130 billion words of&lt;br /&gt;
English extracted from the Common Crawl or vast collections of translated texts in multiple languages),&lt;br /&gt;
pre-computed representations of&lt;br /&gt;
[http://wiki.nlpl.eu/index.php/Vectors/home word or sentence meaning]&lt;br /&gt;
(so-called word embeddings), or more specialized training and evaluation sets&lt;br /&gt;
for supervised machine learning tasks like parsing or machine translation.&lt;br /&gt;
&lt;br /&gt;
After some two years of activity in the NLPL project, its community has collectively&lt;br /&gt;
installed some 80 shared software modules and around six terabytes of primary source data.&lt;br /&gt;
In May 2019, &amp;lt;code&amp;gt;module load&amp;lt;/code&amp;gt; operations for NLPL-maintained software accounted&lt;br /&gt;
for close to five percent of the total on the Norwegian Abel supercluster.&lt;br /&gt;
In sum, preparing the software and data environment for the ‘average’ NLP experiment is no&lt;br /&gt;
small task; duplication of data, software, and effort should be minimized.&lt;br /&gt;
Further, reproducibility and replicability play an increasingly important role in NLP research.&lt;br /&gt;
Other researchers must be enabled to re-run the same experiment (and obtain the same results),&lt;br /&gt;
ideally also several years after the original publication.&lt;br /&gt;
&lt;br /&gt;
= Software =&lt;br /&gt;
&lt;br /&gt;
Relevant software modules comprise general-purpose run-time environments like Java and Python,&lt;br /&gt;
machine learning frameworks like DyNet, PyTorch, SciPy, or TensorFlow, and a myriad of&lt;br /&gt;
discipline-specific tools like CoreNLP, Gensim, Marian, NLTK, Open NMT, spaCy, and others.&lt;br /&gt;
NLPL users typically ‘mix and match’ several of these components, to then build their own&lt;br /&gt;
code on top.&lt;br /&gt;
They will often require specific versions of individual modules, sometimes for good reasons.&lt;br /&gt;
Between 2017 and 2019, the NLPL infrastructure task force has received installation requests&lt;br /&gt;
for individual Python add-ons against language versions 2.7, 3.5, and 3.7, sometimes with&lt;br /&gt;
additional constraints regarding supported versions of, for example, NumPy, PyTorch, or&lt;br /&gt;
TensorFlow.&lt;br /&gt;
&lt;br /&gt;
For compatibility with third-party code and for reproducibility, users should largely&lt;br /&gt;
be free (within reason) to pick the module versions they (believe they) require, modules must&lt;br /&gt;
not change once installed (and announced), and historic or older module versions should&lt;br /&gt;
remain functional over time, ideally many years into the future.&lt;br /&gt;
The NLPL approach to meeting these demands has been to&lt;br /&gt;
[http://wiki.nlpl.eu/index.php/Infrastructure/software/catalogue ‘unbundle’]&lt;br /&gt;
to a high degree, i.e. provision separate add-ons (like Gensim, NumPy, SciPy, TensorFlow, etc.)&lt;br /&gt;
as individual modules and inasmuch as possible provide each module for multiple base language&lt;br /&gt;
versions.&lt;br /&gt;
Abstractly, this design appears adequate and scalable, but module installation needs to be&lt;br /&gt;
automated further, uniformity across different computing environments improved, and users&lt;br /&gt;
better guided in navigating the resulting (large) space of only partially interoperable&lt;br /&gt;
modules.&lt;br /&gt;
&lt;br /&gt;
Uniformity across different computing environments, essentially means that the exact same&lt;br /&gt;
versions of tools (and bundles) are available, and of course that they behave the same on&lt;br /&gt;
all systems.&lt;br /&gt;
To accomplish this goal, it may ultimately be necessary to build the complete software&lt;br /&gt;
stack ‘from the ground up’, i.e. include all dependencies (beyond the core operating system)&lt;br /&gt;
in the NLPL modules collection.&lt;br /&gt;
Otherwise, if one were to build on top of a pre-installed Python, for example, it is likely&lt;br /&gt;
that Python installations will differ (in minor versions, compiler and library versions,&lt;br /&gt;
optional add-on components, and such) across different systems.&lt;br /&gt;
&lt;br /&gt;
= Containerization =&lt;br /&gt;
&lt;br /&gt;
So far, NLPL has shied away from using containers, in part simply because of lacking&lt;br /&gt;
support on some of the target systems (notably Taito), in part because of a concern&lt;br /&gt;
for reduced transparency from the user point of view.&lt;br /&gt;
Also, containerizing individual software modules severely challenges modularization:&lt;br /&gt;
There is no straightforward way to ‘mix and match’ multiple containers into a uniform&lt;br /&gt;
process environment.&lt;br /&gt;
&lt;br /&gt;
However, provisioning the ''full NLPL'' software (and possibly data) environment inside&lt;br /&gt;
a container may offer some benefits, for example compatibility with cloud environments,&lt;br /&gt;
increased uniformity across different systems, and potentially longer-term reproducibility.&lt;br /&gt;
On this view, modularization would obtain within the container, just as it does in the&lt;br /&gt;
current environments on, for example, Abel, Puhti, Saga, and Taito.&lt;br /&gt;
&lt;br /&gt;
= Data =&lt;br /&gt;
&lt;br /&gt;
For the NLPL community it is important to have direct access to essential data sets from the command line. A mounted file system is, therefore, the preferred solution at least at the moment with current workflows. NLPL currently tries to synchronise data sets between the Nordic HPC clusters providing a standardised view on data sets with proper versioning for replication purposes. The structure within the root data folder is roughly like this: 'nlpl-activity/dataset-name/optional-subfolder/release'. The release refers to the version or the date of the release. The path is preferably composed of lower-cased plain ASCII characters only but upper-case letters may appear if necessary.&lt;br /&gt;
&lt;br /&gt;
Mirroring the data is currently done via cron-jobs and the master copy of each data set is on one specific server depending on the main responsible person who maintains the resource.&lt;br /&gt;
&lt;br /&gt;
Some datasets are available for external users without HPC access, for example, OPUS parallel data. This is currently done via ObjectStorage and cPouta at CSC. That collection follows the same release structures as the mounted releases and is also in sync with that data.&lt;br /&gt;
&lt;br /&gt;
Goals for NLPL in EOSC: Better streamline the data maintenance and mirroring procedures. Improve data access libraries and tools to make the work with data sets more transparent. Replicability if results is important. Unnecessary data copying and duplication should be avoided. Documentation is essential.&lt;/div&gt;</summary>
		<author><name>Joerg</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Infrastructure/software/catalogue&amp;diff=596</id>
		<title>Infrastructure/software/catalogue</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Infrastructure/software/catalogue&amp;diff=596"/>
		<updated>2019-02-01T14:07:30Z</updated>

		<summary type="html">&lt;p&gt;Joerg: /* Activity G: OPUS Parallel Corpus */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Background =&lt;br /&gt;
&lt;br /&gt;
This page provides a high-level summary of NLPL-specific software installed on either of our two systems.&lt;br /&gt;
As a rule of thumb, NLPL aims to build on generic software installations provided by the&lt;br /&gt;
system maintainers (e.g. development tools and libraries that are not discipline-specific),&lt;br /&gt;
using the [http://modules.sourceforge.net/ &amp;lt;tt&amp;gt;module&amp;lt;/tt&amp;gt;s infrastructure].&lt;br /&gt;
For example, an environment like OpenNMT is unlikely to be used by other disciplines,&lt;br /&gt;
and NLPL stands to gain from in-house, shared expertise that comes with maintaining&lt;br /&gt;
a project-specific installation.&lt;br /&gt;
On the other hand, the CUDA libraries are general extensions to the operating system&lt;br /&gt;
that most users of deep learning frameworks on gpus will want to use; hence, CUDA is&lt;br /&gt;
most appropriately installed by the core system maintainers.&lt;br /&gt;
Frameworks like PyTorch and TensorFlow, arguably, present a middle ground to this&lt;br /&gt;
rule of thumb:&lt;br /&gt;
In principle, they are not discipline-specific, but in mid-2018 at least the demand for&lt;br /&gt;
installations of these frameworks is strong within NLPL, and the project will likely&lt;br /&gt;
benefit from growing its competencies in this area.&lt;br /&gt;
&lt;br /&gt;
= Module Catalogue =&lt;br /&gt;
&lt;br /&gt;
The discipline-specific modules maintained by NLPL are not activated by default.&lt;br /&gt;
To make available the NLPL directory of module configurations, on top of the&lt;br /&gt;
pre-configured, system-wide modules, one needs to:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
module use -a /proj*/nlpl/software/modulefiles/&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We will at times assume a shell variable &amp;lt;tt&amp;gt;$NLPLROOT&amp;lt;/tt&amp;gt; that points to the&lt;br /&gt;
top-level project directory, i.e. &amp;lt;tt&amp;gt;/projects/nlpl/&amp;lt;/tt&amp;gt; (on Abel) or&lt;br /&gt;
&amp;lt;tt&amp;gt;/proj/nlpl/&amp;lt;/tt&amp;gt; (on Taito).&lt;br /&gt;
For NLPL users, we recommend that one adds the above &amp;lt;tt&amp;gt;module use&amp;lt;/tt&amp;gt; command&lt;br /&gt;
to the shell start-up script, e.g. &amp;lt;tt&amp;gt;.bashrc&amp;lt;/tt&amp;gt; in the user home directory.&lt;br /&gt;
&lt;br /&gt;
= Activity A: Basic Infrastructure =&lt;br /&gt;
&lt;br /&gt;
Interoperability of NLPL installations with each other, as well as with system-wide&lt;br /&gt;
software that is maintained by the core operations teams for Abel and Taito, is no&lt;br /&gt;
small challenge; neither is parallelism across the two systems, for example in&lt;br /&gt;
available software (and versions) and techniques for ‘mixing and matching’.&lt;br /&gt;
These challenges are discussed in some more detail with regard to the&lt;br /&gt;
[http://wiki.nlpl.eu/index.php/Infrastructure/software/python Python programming environment]&lt;br /&gt;
and with regard to&lt;br /&gt;
[http://wiki.nlpl.eu/index.php/Infrastructure/software/frameworks common Deep Learning frameworks].&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Module Name/Version !! Description !! System !! Install Date !! Maintainer&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-cython/0.29.1 || C Extensions for Python || Abel || December 2018 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Infrastructure/software/nltk nlpl-nltk/3.3] || Natural Language Toolkit (NLTK) || Abel, Taito || September 2018 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Infrastructure/software/pytorch nlpl-pytorch/0.4.1] || PyTorch Deep Learning Framework (CPU and GPU) || Abel, Taito || September 2018 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Infrastructure/software/spacy nlpl-spacy/2.0.12] || spaCy: Natural Language Processing in Python || Abel, Taito || October 2018 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Infrastructure/software/tensorflow nlpl-tensorflow/1.11] || TensorFlow Deep Learning Framework (CPU and GPU) || Abel, Taito || September 2018 || Stephan Oepen&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
= Activity B: Statistical and Neural Machine Translation =&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Module Name/Version !! Description !! System !! Install Date !! Maintainer&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Translation/home#Using_the_Moses_module nlpl-moses/mmt-mvp-v0.12.1-2739-gdc42bcb]      || Moses SMT system, including GIZA++, MGIZA, fast_align || Taito || July 2017 || Yves Scherrer&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Translation/home#Using_the_Moses_module nlpl-moses/4.0-65c75ff]      || Moses SMT System Release 4.0, including GIZA++, MGIZA, fast_align, SALM&amp;lt;br/&amp;gt;Some minor fixes added to existing install 2/2018.&amp;lt;br/&amp;gt; Should not break compatibility except when using tokenizer.perl for Finnish or Swedish. || Taito, Abel || November 2017 || Yves Scherrer&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Translation/home#Using_the_Efmaral_module nlpl-efmaral/0.1_2017_07_20] || efmaral and eflomal word alignment tools      || Taito || July 2017 || Yves Scherrer&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Translation/home#Using_the_Efmaral_module nlpl-efmaral/0.1_2017_11_24] || efmaral and eflomal word alignment tools      || Taito, Abel || November 2017 || Yves Scherrer&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Translation/home#Using_the_Efmaral_module nlpl-efmaral/0.1_2018_12_13/17] || efmaral and eflomal word alignment tools      || Taito, Abel || December 2018 || Yves Scherrer&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Translation/home#Using_the_HNMT_module nlpl-hnmt/1.0.1] || HNMT neural machine translation system      || Taito || March 2018 || Yves Scherrer&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Translation/opennmt-py nlpl-opennmt-py/0.2.1] || OpenNMT Python Library || Abel, Taito || September 2018 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Translation/home#Using_the_Marian_module nlpl-marian/1.2.0] || Marian neural machine translation system      || Taito || March 2018 || Yves Scherrer&lt;br /&gt;
|-&lt;br /&gt;
| marian/1.5 || Marian neural machine translation system      || Taito || June 2018 || CSC staff&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Translation/home#Using_the_mttools_module nlpl-mttools/2018_12_23] || A collection of preprocessing and evaluation script for machine translation      || Taito, Abel || December 2018 || Yves Scherrer&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
= Activity C: Data-Driven Parsing =&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Module Name/Version !! Description !! System !! Install Date !! Maintainer&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Parsing/uuparser nlpl-uuparser] || Uppsala Parser || Abel || December 2018 || &lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Parsing/udpipe nlpl-udpipe/1.2.1-devel] || UDPipe 1.2 with Pre-Trained Models      || Taito, Abel || November 2017 || Jörg Tiedemann&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Parsing/dozat nlpl-dozat/201812] || Stanford Graph-Based Parser by Tim Dozat (v3) || Abel || December 2018 || Stephan Oepen&lt;br /&gt;
|-&lt;br /&gt;
| [http://wiki.nlpl.eu/index.php/Parsing/repp nlpl-repp/201812] || REPP Tokenizer (and Sentence Splitter) || Abel || December 2018 || Stephan Oepen&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
= Activity E: Pre-Trained Word Embeddings =&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Module Name/Version !! Description !! System !! Install Date !! Maintainer&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-gensim/3.6.0 || GenSim: Topic Modeling for Humans || Taito, Abel || October 2018 || Stephan Oepen&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
= Activity G: OPUS Parallel Corpus =&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Module Name/Version !! Description !! System !! Install Date !! Maintainer&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-cwb/3.4.12 || Corpus Work Bench (CWB)      || Taito, Abel || November 2017 || Jörg Tiedemann&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-opus/0.1 || Various OPUS Tools      || Taito, Abel || November 2017 || Jörg Tiedemann&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-opus/0.2 || Various OPUS Tools      || Taito, Abel || 2018 || Jörg Tiedemann&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-opus/201901 || Various OPUS Tools      || Taito, Abel || January 2019 || Jörg Tiedemann&lt;br /&gt;
|-&lt;br /&gt;
| nlpl-uplug/0.3.8dev || UPlug Parallel Corpus Tools      || Taito, Abel || November 2017 || Jörg Tiedemann&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Joerg</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Corpora/OPUS&amp;diff=483</id>
		<title>Corpora/OPUS</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Corpora/OPUS&amp;diff=483"/>
		<updated>2018-12-19T18:50:56Z</updated>

		<summary type="html">&lt;p&gt;Joerg: /* http://opus.nlpl.eu */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== http://opus.nlpl.eu ==&lt;br /&gt;
&lt;br /&gt;
OPUS is a collection of open parallel corpora in many languages. It provides bilingually aligned data sets, interfaces, tools and more. The data sets are available in various common formats and are provided for download and for use within the NLPL infrastructure. The service is hosted at CSC in Finland and the core of the data is also available from sigma2 on abel. Tools for processing the data are accessible from taito and more detailed information can be found on the [http://opus.nlpl.eu/trac OPUS Wiki]:&lt;br /&gt;
&lt;br /&gt;
* Information for [http://opus.nlpl.eu/trac/wiki/NLPL NLPL Users]&lt;br /&gt;
* Information about the [http://opus.nlpl.eu/trac#WebAPI OPUS API] for finding resources&lt;br /&gt;
* Information about [http://opus.nlpl.eu/trac/wiki/DataFormats data formats]&lt;br /&gt;
* Information about [http://opus.nlpl.eu/trac/wiki/Tools tools]&lt;br /&gt;
* Information about [http://opus.nlpl.eu/trac/wiki/QueryInterfaces on-line interfaces]&lt;br /&gt;
* Information about [http://opus.nlpl.eu/trac/wiki/WordAlign word alignment] and the [http://opus.nlpl.eu/trac/wiki/WordAlignDB alignment lexicon]&lt;br /&gt;
&lt;br /&gt;
The on-line search interface is available from http://opus.nlpl.eu/bin/opuscqp.pl and the word-alignment-based lexicon is accessible from http://opus.nlpl.eu/lex.php&lt;br /&gt;
&lt;br /&gt;
Contact: [http://blogs.helsinki.fi/tiedeman/ Jörg Tiedemann] via e-mail - firstname.lastname at helsinki.fi (first name without dots)&lt;/div&gt;</summary>
		<author><name>Joerg</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Corpora/home&amp;diff=99</id>
		<title>Corpora/home</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Corpora/home&amp;diff=99"/>
		<updated>2017-11-15T22:30:03Z</updated>

		<summary type="html">&lt;p&gt;Joerg: /* Large Corpora */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;NLPL provides various large data sets. They are available from the connected infrastructure. Please, check the individual pages of each resource.&lt;br /&gt;
&lt;br /&gt;
* [[Corpora/OPUS|OPUS - the collection of open parallel corpora]]&lt;/div&gt;</summary>
		<author><name>Joerg</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Corpora/OPUS&amp;diff=98</id>
		<title>Corpora/OPUS</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Corpora/OPUS&amp;diff=98"/>
		<updated>2017-11-15T22:29:46Z</updated>

		<summary type="html">&lt;p&gt;Joerg: /* OPUS */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== http://opus.nlpl.eu ==&lt;br /&gt;
&lt;br /&gt;
OPUS is a collection of open parallel corpora in many languages. It provides bilingually aligned data sets, interfaces, tools and more. The data sets are available in various common formats and are provided for download and for use within the NLPL infrastructure. The service is hosted at CSC in Finland and the core of the data is also available from sigma2 on abel. Tools for processing the data are accessible from taito and more detailed information can be found on the [http://opus.nlpl.eu/trac OPUS Wiki]:&lt;br /&gt;
&lt;br /&gt;
* Information for [http://opus.nlpl.eu/trac/wiki/NLPL NLPL Users]&lt;br /&gt;
* Information about [http://opus.nlpl.eu/trac/wiki/DataFormats data formats]&lt;br /&gt;
* Information about [http://opus.nlpl.eu/trac/wiki/Tools tools]&lt;br /&gt;
* Information about [http://opus.nlpl.eu/trac/wiki/QueryInterfaces on-line interfaces]&lt;br /&gt;
* Information about [http://opus.nlpl.eu/trac/wiki/WordAlign word alignment] and the [http://opus.nlpl.eu/trac/wiki/WordAlignDB alignment lexicon]&lt;br /&gt;
&lt;br /&gt;
The on-line search interface is available from http://opus.nlpl.eu/bin/opuscqp.pl and the word-alignment-based lexicon is accessible from http://opus.nlpl.eu/lex.php&lt;br /&gt;
&lt;br /&gt;
Contact: [http://blogs.helsinki.fi/tiedeman/ Jörg Tiedemann] via e-mail - firstname.lastname at helsinki.fi (first name without dots)&lt;/div&gt;</summary>
		<author><name>Joerg</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Corpora/OPUS&amp;diff=97</id>
		<title>Corpora/OPUS</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Corpora/OPUS&amp;diff=97"/>
		<updated>2017-11-15T22:26:51Z</updated>

		<summary type="html">&lt;p&gt;Joerg: Created page with &amp;quot;= OPUS =  URL: http://opus.nlpl.eu  OPUS is a collection of open parallel corpora in many languages. It provides bilingually aligned data sets, interfaces, tools and more. The...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= OPUS =&lt;br /&gt;
&lt;br /&gt;
URL: http://opus.nlpl.eu&lt;br /&gt;
&lt;br /&gt;
OPUS is a collection of open parallel corpora in many languages. It provides bilingually aligned data sets, interfaces, tools and more. The data sets are available in various common formats and are provided for download and for use within the NLPL infrastructure. The service is hosted at CSC in Finland and the core of the data is also available from sigma2 on abel. Tools for processing the data are accessible from taito and more detailed information can be found on the [http://opus.nlpl.eu/trac OPUS Wiki]:&lt;br /&gt;
&lt;br /&gt;
* Information for [http://opus.nlpl.eu/trac/wiki/NLPL NLPL Users]&lt;br /&gt;
* Information about [http://opus.nlpl.eu/trac/wiki/DataFormats data formats]&lt;br /&gt;
* Information about [http://opus.nlpl.eu/trac/wiki/Tools tools]&lt;br /&gt;
* Information about [http://opus.nlpl.eu/trac/wiki/QueryInterfaces on-line interfaces]&lt;br /&gt;
* Information about [http://opus.nlpl.eu/trac/wiki/WordAlign word alignment] and the [http://opus.nlpl.eu/trac/wiki/WordAlignDB alignment lexicon]&lt;br /&gt;
&lt;br /&gt;
The on-line search interface is available from http://opus.nlpl.eu/bin/opuscqp.pl and the word-alignment-based lexicon is accessible from http://opus.nlpl.eu/lex.php&lt;br /&gt;
&lt;br /&gt;
Contact: [http://blogs.helsinki.fi/tiedeman/ Jörg Tiedemann] via e-mail - firstname.lastname at helsinki.fi (first name without dots)&lt;/div&gt;</summary>
		<author><name>Joerg</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Corpora/home&amp;diff=96</id>
		<title>Corpora/home</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Corpora/home&amp;diff=96"/>
		<updated>2017-11-15T22:07:03Z</updated>

		<summary type="html">&lt;p&gt;Joerg: Created page with &amp;quot;= Large Corpora =  NLPL provides various large data sets. They are available from the connected infrastructure. Please, check the individual pages of each resource.  * [[Corpo...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Large Corpora =&lt;br /&gt;
&lt;br /&gt;
NLPL provides various large data sets. They are available from the connected infrastructure. Please, check the individual pages of each resource.&lt;br /&gt;
&lt;br /&gt;
* [[Corpora/OPUS|OPUS - the collection of open parallel corpora]]&lt;/div&gt;</summary>
		<author><name>Joerg</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Infrastructure/software/catalogue&amp;diff=64</id>
		<title>Infrastructure/software/catalogue</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Infrastructure/software/catalogue&amp;diff=64"/>
		<updated>2017-11-06T21:22:34Z</updated>

		<summary type="html">&lt;p&gt;Joerg: /* Module catalogue */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Background =&lt;br /&gt;
&lt;br /&gt;
High-level summary of NLPL-specific software installed on either of our two systems&lt;br /&gt;
&lt;br /&gt;
= Module catalogue =&lt;br /&gt;
&lt;br /&gt;
Add the NLPL directory of module configurations:&lt;br /&gt;
* on Taito:&lt;br /&gt;
 module use -a /proj/nlpl/software/modulefiles/&lt;br /&gt;
* on Abel:&lt;br /&gt;
 module use -a /projects/nlpl/software/modulefiles/&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Activity !! Module name/version !! Included software !! Installed on !! Install date&lt;br /&gt;
|-&lt;br /&gt;
| B || moses/mmt-mvp-v0.12.1-2739-gdc42bcb      || Moses SMT system, including GIZA++, MGIZA, fast_align || Taito || 7/2017&lt;br /&gt;
|-&lt;br /&gt;
| B || efmaral/0.1_2017_07_20      || efmaral and eflomal word alignment tools      || Taito || 7/2017&lt;br /&gt;
|-&lt;br /&gt;
| C, G || nlpl-udpipe/1.2.1-devel      || UDPipe with pre-trained models      || Taito, Abel || 11/2017&lt;br /&gt;
|-&lt;br /&gt;
| G || nlpl-cwb/3.4.12      || Corpus Work Bench (CWB)      || Taito, Abel || 11/2017&lt;br /&gt;
|-&lt;br /&gt;
| G || nlpl-opus/0.1      || various OPUS tools      || Taito, Abel || 11/2017&lt;br /&gt;
|-&lt;br /&gt;
| G || nlpl-uplug/0.3.8dev      || Uplug parallel corpus tools      || Taito, Abel || 11/2017&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Joerg</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Infrastructure/wiki&amp;diff=40</id>
		<title>Infrastructure/wiki</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Infrastructure/wiki&amp;diff=40"/>
		<updated>2017-10-16T08:18:56Z</updated>

		<summary type="html">&lt;p&gt;Joerg: Created page with &amp;quot;= Background =  NLPL maintains an installation of the [http://www.mediawiki.org MediaWiki] software to collaboratively maintain information that is relevant to project members...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Background =&lt;br /&gt;
&lt;br /&gt;
NLPL maintains an installation of the [http://www.mediawiki.org MediaWiki] software&lt;br /&gt;
to collaboratively maintain information that is relevant to project members and other&lt;br /&gt;
users of the infrastructure, including technical documentation.&lt;br /&gt;
Bjørn Lindi and Stephan Oepen serve as the wiki administrators, i.e. will try to assist&lt;br /&gt;
in all technical questions regarding access and structure.&lt;br /&gt;
&lt;br /&gt;
= Gaining Access =&lt;br /&gt;
&lt;br /&gt;
To prevent wiki spam, the NLPL wiki is currently not set up for self-help account creation.&lt;br /&gt;
To again editing rights for the wiki, please email you Abel account name to&lt;br /&gt;
Bjørn Lindi or Stephan Oepen.&lt;br /&gt;
&lt;br /&gt;
= Editing Conventions =&lt;br /&gt;
&lt;br /&gt;
For the collaborative authoring process to maintain some semblance of structure,&lt;br /&gt;
please apply the following conventions in editing existing pages and, in particular,&lt;br /&gt;
when creating new pages:&lt;br /&gt;
&lt;br /&gt;
Page names should be comprised of all lower-case english words, organized in a page hierarchy&lt;br /&gt;
that follows the project structure and other organizational principles, e.g. the names&lt;br /&gt;
of individual software packages closer to the leafs of the hierarchy.&lt;/div&gt;</summary>
		<author><name>Joerg</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Translation/documentation&amp;diff=39</id>
		<title>Translation/documentation</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Translation/documentation&amp;diff=39"/>
		<updated>2017-10-16T08:03:03Z</updated>

		<summary type="html">&lt;p&gt;Joerg: Created page with &amp;quot;= Background =  This page provides the user documentation for the software installation of Machine Translation (MT) tools on Taito, particularly the open-source [http://www.st...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Background =&lt;br /&gt;
&lt;br /&gt;
This page provides the user documentation for the software installation of Machine Translation (MT) tools on Taito,&lt;br /&gt;
particularly the open-source [http://www.statmt.org/moses/ Moses] environment for experimentation with statistical MT (SMT).&lt;/div&gt;</summary>
		<author><name>Joerg</name></author>
		
	</entry>
</feed>