<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://wiki.nlpl.eu/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Erikve</id>
	<title>Nordic Language Processing Laboratory - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://wiki.nlpl.eu/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Erikve"/>
	<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/Special:Contributions/Erikve"/>
	<updated>2026-05-16T06:23:25Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.31.10</generator>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Community/training&amp;diff=1657</id>
		<title>Community/training</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Community/training&amp;diff=1657"/>
		<updated>2024-01-30T13:22:25Z</updated>

		<summary type="html">&lt;p&gt;Erikve: panelists&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;'''HPLT &amp;amp; NLPL Winter School on Large Language Models: Creation, Customization, Evaluation, and Use'''&lt;br /&gt;
&lt;br /&gt;
[[File:Skeikampen.2023.jpg|center]]&lt;br /&gt;
&lt;br /&gt;
= Background =&lt;br /&gt;
&lt;br /&gt;
Since 2023, the NLPL network and Horizon Europe&lt;br /&gt;
project ''[https://hplt-project.org High-Performance Language Technologies]'' (HPLT)&lt;br /&gt;
have joined forces to organize the successful winter school series on Web-scale NLP.&lt;br /&gt;
The winter school seeks to stimulate ''community formation'',&lt;br /&gt;
i.e. strengthening interaction and collaboration among&lt;br /&gt;
European research teams in NLP and advancing a shared level of knowledge&lt;br /&gt;
and experience in using high-performance e-infrastructures for large-scale&lt;br /&gt;
NLP research.&lt;br /&gt;
The 2024 edition of the winter school puts special emphasis on&lt;br /&gt;
NLP researchers from countries who participate in the EuroHPC&lt;br /&gt;
[https://www.lumi-supercomputer.eu/lumi-consortium/ LUMI consortium].&lt;br /&gt;
For additional background, please see the archival pages from the&lt;br /&gt;
[http://wiki.nlpl.eu/index.php/Community/training/2018 2018],&lt;br /&gt;
[http://wiki.nlpl.eu/index.php/Community/training/2019 2019],&lt;br /&gt;
[http://wiki.nlpl.eu/index.php/Community/training/2020 2020], and&lt;br /&gt;
[http://wiki.nlpl.eu/index.php/Community/training/2023 2023]&lt;br /&gt;
NLPL Winter Schools.&lt;br /&gt;
&lt;br /&gt;
For early 2024, HPLT will hold its winter school from Sunday, February 4, to&lt;br /&gt;
Tuesday, February 6, 2024, at a&lt;br /&gt;
[https://www.thonhotels.com/our-hotels/norway/skeikampen/ mountain-side hotel]&lt;br /&gt;
(with skiing and walking opportunities) about two hours north of Oslo.&lt;br /&gt;
The project will organize group bus transfer from and to the Oslo&lt;br /&gt;
airport ''Gardermoen'', leaving the airport at 9:45 on Sunday morning&lt;br /&gt;
and returning there around 17:30 on Tuesday afternoon.&lt;br /&gt;
&lt;br /&gt;
The winter school is subsidized by the HPLT project: there is no fee for&lt;br /&gt;
participants and no charge for the bus transfer to and from the&lt;br /&gt;
conference hotel.&lt;br /&gt;
All participants will have to cover their own travel and accomodation&lt;br /&gt;
at Skeikampen, however.&lt;br /&gt;
Two nights at the hotel, including all meals, will come to NOK 3745 (NOK 3345 per person in a shared double room), &lt;br /&gt;
to be paid to the hotel directly.&lt;br /&gt;
&lt;br /&gt;
= Programme =&lt;br /&gt;
&lt;br /&gt;
The 2024 winter school will have a thematic focus on ''Large Language Models: Creation, Customization, Evaluation, and Use''.&lt;br /&gt;
The programme will be comprised of in-depth technical presentations (possibly including some&lt;br /&gt;
hands-on elements) by seasoned experts, with special emphasis on open science and European languages, &lt;br /&gt;
but also include critical reflections on current development trends in LLM-focussed NLP.&lt;br /&gt;
The programme will be complemented with a panel discussion and a ‘walk-through’ of available&lt;br /&gt;
infrastructure on the shared EuroHPC LUMI supercomputer.&lt;br /&gt;
&lt;br /&gt;
Confirmed presenters include:&lt;br /&gt;
&lt;br /&gt;
* [http://afra.alishahi.name Afra Alishahi, Tilburg University, The Netherlands]&lt;br /&gt;
* [https://di.ku.dk/english/staff/vip/?pure=en/persons/631668 Desmond Elliot, University of Copenhagen, Denmark]&lt;br /&gt;
* [https://muennighoff.github.io/ Niklas Muennighoff, Contextual AI]&lt;br /&gt;
* [https://perso.limsi.fr/neveol/bio.html Aurélie Névéol, Interdisciplinary Laboratory of Numerical Sciences, France]&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
!colspan=3|Sunday, February 4, 2024&lt;br /&gt;
|-&lt;br /&gt;
| 13:00 || 14:00 || Lunch&lt;br /&gt;
|-&lt;br /&gt;
| 14:00 || 15:30 || '''Session 1''': Analyzing and Interpreting Deep Neural Models of Language ([http://afra.alishahi.name Afra Alishahi])&lt;br /&gt;
|-&lt;br /&gt;
| 15:30 || 15:50 || Coffee Break&lt;br /&gt;
|-&lt;br /&gt;
| 16:00 || 17:30 || '''Session 2''': Analyzing and Interpreting Deep Neural Models of Language ([http://afra.alishahi.name Afra Alishahi])&lt;br /&gt;
|-&lt;br /&gt;
| 17:30 || 17:50 || Coffee Break&lt;br /&gt;
|-&lt;br /&gt;
| 17:50 || 19:20 || '''Session 3''': Scaling Data-constrained Language Models ([https://muennighoff.github.io/ Niklas Muennighoff])&lt;br /&gt;
|-&lt;br /&gt;
| 19:30 ||  || Dinner&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
!colspan=3|Monday, February 5, 2024&lt;br /&gt;
|-&lt;br /&gt;
|colspan=3 | Breakfast is available from 07:30&lt;br /&gt;
|-&lt;br /&gt;
| 09:00 || 10:30 || '''Session 4''': Bias in Natural Language Processing: focus on large language models ([https://perso.limsi.fr/neveol/bio.html Aurélie Névéol]) &lt;br /&gt;
|-&lt;br /&gt;
|colspan=3| Free time (Lunch is available between 13:00 and 14:30)&lt;br /&gt;
|-&lt;br /&gt;
| 15:00 || 16:30 || '''Session 5''': Multilingual and multimodal language models ([https://di.ku.dk/english/staff/vip/?pure=en/persons/631668 Desmond Elliot]) &lt;br /&gt;
|-&lt;br /&gt;
| 16:30 || 16:50 || Coffee Break&lt;br /&gt;
|-&lt;br /&gt;
| 16:50 || 17:40 || '''Session 6''': Multilingual and multimodal language models ([https://di.ku.dk/english/staff/vip/?pure=en/persons/631668 Desmond Elliot]) &lt;br /&gt;
|-&lt;br /&gt;
| 17:40 || 18:00 || Coffee Break&lt;br /&gt;
|-&lt;br /&gt;
| 18:00 || 19:15 || '''Session 7'''. «Large vs. Small»: panel discussion. Panelists: Per Kummervold (National Library of Norway), Desmond Elliott (University of Copenhagen), Evangelia Gogoulou (RISE, Sweden), Afra Alishahi (Tilburg University), Jan Hajič (Charles University in Prague), and Aurélie Névéol (LISN, France)&lt;br /&gt;
|-&lt;br /&gt;
| 19:30 ||  || Dinner&lt;br /&gt;
|-&lt;br /&gt;
| 21:00 || || '''Evening Session'''. LUMI: BERT in an Hour, GPT in a Week. Speakers: David Samuel (University of Oslo) and Risto Luukkonen (University of Turku, Silo AI)&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
!colspan=3|Tuesday, February 6, 2024&lt;br /&gt;
|-&lt;br /&gt;
|colspan=3| Breakfast is available from 07:30&lt;br /&gt;
|-&lt;br /&gt;
| 08:30 || 10:00 || '''Session 8''': Reproducibility in Natural Language Processing ([https://perso.limsi.fr/neveol/bio.html Aurélie Névéol]) &lt;br /&gt;
|-&lt;br /&gt;
| 10:00 || 10:30 || Coffee Break&lt;br /&gt;
|-&lt;br /&gt;
| 10:30 || 12:00 || '''Session 9''': Understanding and measuring the environmental impact of Natural Language Processing ([https://perso.limsi.fr/neveol/bio.html Aurélie Névéol]) &lt;br /&gt;
|-&lt;br /&gt;
| 12:30 || 13:30 || Lunch&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
= Registration =&lt;br /&gt;
&lt;br /&gt;
In total, we anticipate around 55 participants at the 2024 winter school.&lt;br /&gt;
We have received more requests for participation than we will be able to accommodate,&lt;br /&gt;
and the registration form has now been closed.&lt;br /&gt;
We processed requests for participation on a first-come, first-served basis, with an eye toward regional balance.&lt;br /&gt;
Interested parties who have submitted the registration form were confirmed in three batches, on December 11, on December 15,&lt;br /&gt;
and on December 22, which was also the closing date for winter school registration.&lt;br /&gt;
&lt;br /&gt;
Once confirmed by the organizing team, participant names are published&lt;br /&gt;
on this page, and registration establishes a&lt;br /&gt;
''binding agreement'' with the hotel.&lt;br /&gt;
Therefore, a cancellation fee will be incurred (unless we can find someone else to ‘take over’ last-minute&lt;br /&gt;
spaces), and no-shows will be charged the full price for at least one night&lt;br /&gt;
by the hotel.&lt;br /&gt;
&lt;br /&gt;
= Logistics = &lt;br /&gt;
&lt;br /&gt;
With a few exceptions, winter school participants travel to and from the conference hotel&lt;br /&gt;
jointly on a chartered bus (the HPLT shuttle).&lt;br /&gt;
The bus will leave OSL airport no later than 9:45 CET on Sunday, February 4.&lt;br /&gt;
Thus, please meet up by 9:30 and make your arrival known to your assigned&lt;br /&gt;
‘tour guide’ (who will introduce themselves to you by email beforehand).&lt;br /&gt;
&lt;br /&gt;
The group will gather near the DNB currency exchange booth in the downstairs&lt;br /&gt;
arrivals area, just outside the international arrivals luggage claims and slightly&lt;br /&gt;
to the left as one exits the customs area:&lt;br /&gt;
the yellow dot numbered (18) on the&lt;br /&gt;
[https://avinor.no/globalassets/_oslo-lufthavn/ankomst-arrivals.pdf OSL arrivals map].&lt;br /&gt;
The group will then walk over to the bus terminal, to leave the airport not long after 9:40.&lt;br /&gt;
The drive to the Skeikampen conference hotel will take us about three hours, and the bus&lt;br /&gt;
will make one stop along the way to stretch our legs and fill up on coffee.&lt;br /&gt;
&lt;br /&gt;
The winter school will end with lunch on Tuesday, February 6, before the group returns&lt;br /&gt;
to OSL airport on the HPLT shuttle.&lt;br /&gt;
The bus will leave Skeikampen at 14:00 CET, with an expected arrival time at OSL&lt;br /&gt;
around 17:00 to 17:30 CET. After stopping at the OSL airport, the bus will continue to central Oslo.&lt;br /&gt;
&lt;br /&gt;
= Organization =&lt;br /&gt;
&lt;br /&gt;
The 2024 Winter School is organized by a team of volunteers at the University&lt;br /&gt;
of Oslo, supported by a programme committee from the HPLT and NLPL network and beyond,&lt;br /&gt;
please see below.&lt;br /&gt;
For all inquiries regarding registration, the programme, logistics,&lt;br /&gt;
or such, please contact &amp;lt;code&amp;gt;hplt-training@ifi.uio.no&amp;lt;/code&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
The programme committee is comprised of:&lt;br /&gt;
&lt;br /&gt;
* Isabelle Augenstein (University of Copenhagen, Denmark)&lt;br /&gt;
* Emily M. Bemder (University of Washington, USA)&lt;br /&gt;
* Kenneth Heafield (Edinburgh University, UK)&lt;br /&gt;
* Jindřich Helcl (Charles University, Czech Republic)&lt;br /&gt;
* Marco Kuhlmann (Linköping University, Sweden)&lt;br /&gt;
* Per Egil Kummervold (National Library of Norway)&lt;br /&gt;
* Andrey Kutuzov (University of Oslo, Norway)&lt;br /&gt;
* Joakim Nivre (RISE and Uppsala University, Sweden)&lt;br /&gt;
* Stephan Oepen (University of Oslo, Norway)&lt;br /&gt;
* Sampo Pyysalo (University of Turku, Finland)&lt;br /&gt;
* Gema Ramirez (Prompsit Language Engineering, Spain)&lt;br /&gt;
* Anna Rogers (IT University of Copenhagen, Denmark)&lt;br /&gt;
* Magnus Sahlgreen (AI Sweden)&lt;br /&gt;
* David Samuel (University of Oslo, Norway)&lt;br /&gt;
* Jörg Tiedemann (University of Helsinki, Finland)&lt;br /&gt;
* Erik Velldal (University of Oslo, Norway)&lt;br /&gt;
&lt;br /&gt;
= Participants =&lt;br /&gt;
&lt;br /&gt;
# Afra Alishahi, Tilburg University (The Netherlands)&lt;br /&gt;
# Ali Allaith, University of Copenhagen (Denmark)&lt;br /&gt;
# Nikolay Arefev, University of Oslo (Norway)&lt;br /&gt;
# Joseph Attieh, University of Helsinki (Finland)&lt;br /&gt;
# Christopher Brückner, Charles University in Prague (Czech Republic)&lt;br /&gt;
# Lucas Charpentier, University of Oslo (Norway)&lt;br /&gt;
# Konstantin Dobler, Hasso Plattner Institute (Germany)&lt;br /&gt;
# Aleksei Dorkin, University of Tartu (Estonia)&lt;br /&gt;
# Luise Dürlich, Uppsala University (Sweden)&lt;br /&gt;
# Simen Eide, Schibsted (Norway)&lt;br /&gt;
# Desmond Elliott, University of Copenhagen (Denmark)&lt;br /&gt;
# Kenneth Enevoldsen, Aarhus University (Denmark)&lt;br /&gt;
# Mariia Fedorova, University of Oslo (Norway)&lt;br /&gt;
# Emilie Francis, Gothenburg University (Sweden)&lt;br /&gt;
# Evangelia Gogoulou, RISE (Sweden)&lt;br /&gt;
# Jan Hajič, Charles University in Prague (Czech Republic)&lt;br /&gt;
# Lasse Hansen, Aarhus University Hospital (Denmark)&lt;br /&gt;
# Jindřich Helcl, Charles University in Prague (Czech Republic)&lt;br /&gt;
# Yiping Jin, Pompeu Fabra University (Spain)&lt;br /&gt;
# Amanda Kann, Stockholm University (Sweden)&lt;br /&gt;
# Jan Kostkan, Aarhus University (Denmark)&lt;br /&gt;
# Per Kummervold, National Library og Norway&lt;br /&gt;
# Andrey Kutuzov, University of Oslo (Norway)&lt;br /&gt;
# Tsz Kin Lam, University of Edinburgh (UK)&lt;br /&gt;
# Wenyan Li, University of Copenhagen (Denmark)&lt;br /&gt;
# Pierre Lison, Norsk Regnesentral&lt;br /&gt;
# Jouni Luoma, University of Turku (Finland)&lt;br /&gt;
# Risto Luukkonen, University of Turku (Finland)&lt;br /&gt;
# Arianna Masciolini, Gothenburg University (Sweden)&lt;br /&gt;
# Petter Mæhlum, University of Oslo (Norway)&lt;br /&gt;
# Vladislav Mikhailov, University of Oslo (Norway)&lt;br /&gt;
# Yousuf Ali Mohammed, Gothenburg University (Sweden)&lt;br /&gt;
# Aurélie Névéol, LISN &amp;amp; CNRS (France)&lt;br /&gt;
# Tobias Norlund, AI Sweden (Sweden)&lt;br /&gt;
# Stephan Oepen, University of Oslo (Norway)&lt;br /&gt;
# Lilja Øvrelid, University of Oslo (Norway)&lt;br /&gt;
# Alberto Parola, University of Copenhagen (Denmark)&lt;br /&gt;
# Siddhesh Pawar, University of Copenhagen (Denmark)&lt;br /&gt;
# Erofili Psaltaki, University of Helsinki (Finland)&lt;br /&gt;
# Akseli Reunamo, University of Turku (Finland)&lt;br /&gt;
# David Samuel, University of Oslo (Norway)&lt;br /&gt;
# Ricardo Muñoz Sánchez, Gothenburg University (Sweden)&lt;br /&gt;
# Gautam Kishore Shahi, University of Duisburg-Essen (Germany)&lt;br /&gt;
# Janine Siewert, University of Helsinki (Finland)&lt;br /&gt;
# Étienne Simon, University of Oslo (Norway)&lt;br /&gt;
# Inguna Skadiņa, University of Latvia&lt;br /&gt;
# Ondrej Sotolar, Masaryk University (Czech Republic)&lt;br /&gt;
# Pavel Stranak, Charles University in Prague (Czech Republic)&lt;br /&gt;
# Maria Irena Szawerna, Gothenburg University (Sweden)&lt;br /&gt;
# Jörg Tiedemann, University of Helsinki (Finland)&lt;br /&gt;
# Ekaterina Uetova, Technological University Dublin (Ireland)&lt;br /&gt;
# Erik Velldal, University of Oslo (Norway)&lt;br /&gt;
# Tea Vojtěchová, Charles University in Prague (Czech Republic)&lt;br /&gt;
# Jonas Waldendorf, University of Edinburgh (UK)&lt;br /&gt;
# Jaume Zaragoza-Bernabeu, Prompsit Language Engineering (Spain)&lt;br /&gt;
# Giulio Zhou, University of Edinburgh (UK)&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1377</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1377"/>
		<updated>2021-06-23T11:56:15Z</updated>

		<summary type="html">&lt;p&gt;Erikve: /* NLP tasks */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
The following would be natural places to start. For most of these, while we do have baseline numbers to compare to, we do not have existing set-ups where we could simply plug in a Norwegian BERT and rund, so we may need to identify suitable code for existing BERT-based architectures for e.g. English to re-use. For the first task though (document-level SA on NoReC) Jeremy would have an existing set-up for using mBERT that we could perhaps use. &lt;br /&gt;
&lt;br /&gt;
== NLP tasks ==&lt;br /&gt;
* Structured sentiment analysis: [https://github.com/ltgoslo/norec_fine NoReC_fine]  &lt;br /&gt;
* Sentence-level 2/3-way polarity: [https://github.com/ltgoslo/norec_sentence/ NoReC_sentences]  &lt;br /&gt;
* Negation cues and scopes (evaluation is still being developed): [https://github.com/ltgoslo/norec_neg/ NoReC_neg]&lt;br /&gt;
* PoS tagging: [https://github.com/UniversalDependencies/UD_Norwegian-NynorskLIA ILA] +  NDT [https://github.com/UniversalDependencies/UD_Norwegian-Bokmaal Bokmaal] / [https://github.com/UniversalDependencies/UD_Norwegian-Nynorsk Nynorsk]&lt;br /&gt;
* Dependency parsing: [https://github.com/UniversalDependencies/UD_Norwegian-NynorskLIA ILA] + NDT [https://github.com/UniversalDependencies/UD_Norwegian-Bokmaal Bokmaal] / [https://github.com/UniversalDependencies/UD_Norwegian-Nynorsk Nynorsk] &lt;br /&gt;
* NER: [https://github.com/ltgoslo/norne NorNE] (Bokmål+Nynorsk)&lt;br /&gt;
* Co-reference resolution (annotation ongoing)&lt;br /&gt;
&lt;br /&gt;
== Lexical  ==&lt;br /&gt;
*[https://www.nb.no/sprakbanken/en/resource-catalogue/oai-nb-no-sbr-27/ Word sense disambiguation in context]&lt;br /&gt;
*[https://github.com/ltgoslo/norwegian-synonyms Norwegian synonyms] (for static models)&lt;br /&gt;
*[https://github.com/ltgoslo/norwegian-analogies Norwegian analogies] (for static models)&lt;br /&gt;
*[https://github.com/ltgoslo/norsentlex NorSentLex]: Sentiment lexicon (for static models)&lt;br /&gt;
&lt;br /&gt;
== Text classification ==&lt;br /&gt;
*[https://github.com/ltgoslo/norec NoReC]; document-level ratings.&lt;br /&gt;
*[https://github.com/ltgoslo/talk-of-norway Talk of Norway]&lt;br /&gt;
*[https://github.com/jerbarnes/norwegian_dialect NorDial]&lt;br /&gt;
&lt;br /&gt;
==Other ==&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1376</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1376"/>
		<updated>2021-06-23T11:56:01Z</updated>

		<summary type="html">&lt;p&gt;Erikve: /* NLP tasks */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
The following would be natural places to start. For most of these, while we do have baseline numbers to compare to, we do not have existing set-ups where we could simply plug in a Norwegian BERT and rund, so we may need to identify suitable code for existing BERT-based architectures for e.g. English to re-use. For the first task though (document-level SA on NoReC) Jeremy would have an existing set-up for using mBERT that we could perhaps use. &lt;br /&gt;
&lt;br /&gt;
== NLP tasks ==&lt;br /&gt;
* Structured sentiment analysis: [https://github.com/ltgoslo/norec_fine NoReC_fine]  &lt;br /&gt;
* Sentence-level 2/3-way polarity: [https://github.com/ltgoslo/norec_sentence/ NoReC_sentences]  &lt;br /&gt;
* Negation cues and scopes (evaluation is still being developed): [https://github.com/ltgoslo/norec_neg/ NoReC_neg]&lt;br /&gt;
* PoS tagging: [https://github.com/UniversalDependencies/UD_Norwegian-NynorskLIA ILA] +  NDT [https://github.com/UniversalDependencies/UD_Norwegian-Bokmaal Bokmaal] / [https://github.com/UniversalDependencies/UD_Norwegian-Nynorsk Nynorsk]&lt;br /&gt;
* Dependency parsing: [https://github.com/UniversalDependencies/UD_Norwegian-NynorskLIA ILA] + NDT [https://github.com/UniversalDependencies/UD_Norwegian-Bokmaal Bokmaal] / [https://github.com/UniversalDependencies/UD_Norwegian-Nynorsk Nynorsk] &lt;br /&gt;
* NER: [https://github.com/ltgoslo/norne NorNE] (bokmål+nynorsk)&lt;br /&gt;
* Co-reference resolution (annotation ongoing)&lt;br /&gt;
&lt;br /&gt;
== Lexical  ==&lt;br /&gt;
*[https://www.nb.no/sprakbanken/en/resource-catalogue/oai-nb-no-sbr-27/ Word sense disambiguation in context]&lt;br /&gt;
*[https://github.com/ltgoslo/norwegian-synonyms Norwegian synonyms] (for static models)&lt;br /&gt;
*[https://github.com/ltgoslo/norwegian-analogies Norwegian analogies] (for static models)&lt;br /&gt;
*[https://github.com/ltgoslo/norsentlex NorSentLex]: Sentiment lexicon (for static models)&lt;br /&gt;
&lt;br /&gt;
== Text classification ==&lt;br /&gt;
*[https://github.com/ltgoslo/norec NoReC]; document-level ratings.&lt;br /&gt;
*[https://github.com/ltgoslo/talk-of-norway Talk of Norway]&lt;br /&gt;
*[https://github.com/jerbarnes/norwegian_dialect NorDial]&lt;br /&gt;
&lt;br /&gt;
==Other ==&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1375</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1375"/>
		<updated>2021-06-23T11:55:20Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
The following would be natural places to start. For most of these, while we do have baseline numbers to compare to, we do not have existing set-ups where we could simply plug in a Norwegian BERT and rund, so we may need to identify suitable code for existing BERT-based architectures for e.g. English to re-use. For the first task though (document-level SA on NoReC) Jeremy would have an existing set-up for using mBERT that we could perhaps use. &lt;br /&gt;
&lt;br /&gt;
== NLP tasks ==&lt;br /&gt;
* Structured sentiment analysis: [https://github.com/ltgoslo/norec_fine NoReC_fine]  &lt;br /&gt;
* Sentence-level 2/3-way polarity: [https://github.com/ltgoslo/norec_sentence/ NoReC_sentences]  &lt;br /&gt;
* Negation cues and scopes (evaluation is still being developed): [https://github.com/ltgoslo/norec_neg/ NoReC_neg]&lt;br /&gt;
* PoS tagging: [https://github.com/UniversalDependencies/UD_Norwegian-NynorskLIA ILA] +  NDT [https://github.com/UniversalDependencies/UD_Norwegian-Bokmaal Bokmaal] / [https://github.com/UniversalDependencies/UD_Norwegian-Nynorsk Nynorsk]&lt;br /&gt;
* Dependency parsing: [https://github.com/UniversalDependencies/UD_Norwegian-NynorskLIA ILA] + NDT [https://github.com/UniversalDependencies/UD_Norwegian-Bokmaal Bokmaal] / [https://github.com/UniversalDependencies/UD_Norwegian-Nynorsk Nynorsk] &lt;br /&gt;
* NER: [https://github.com/ltgoslo/norne NorNE]&lt;br /&gt;
* Co-reference resolution (annotation ongoing)&lt;br /&gt;
&lt;br /&gt;
== Lexical  ==&lt;br /&gt;
*[https://www.nb.no/sprakbanken/en/resource-catalogue/oai-nb-no-sbr-27/ Word sense disambiguation in context]&lt;br /&gt;
*[https://github.com/ltgoslo/norwegian-synonyms Norwegian synonyms] (for static models)&lt;br /&gt;
*[https://github.com/ltgoslo/norwegian-analogies Norwegian analogies] (for static models)&lt;br /&gt;
*[https://github.com/ltgoslo/norsentlex NorSentLex]: Sentiment lexicon (for static models)&lt;br /&gt;
&lt;br /&gt;
== Text classification ==&lt;br /&gt;
*[https://github.com/ltgoslo/norec NoReC]; document-level ratings.&lt;br /&gt;
*[https://github.com/ltgoslo/talk-of-norway Talk of Norway]&lt;br /&gt;
*[https://github.com/jerbarnes/norwegian_dialect NorDial]&lt;br /&gt;
&lt;br /&gt;
==Other ==&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1374</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1374"/>
		<updated>2021-06-23T11:53:33Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
The following would be natural places to start. For most of these, while we do have baseline numbers to compare to, we do not have existing set-ups where we could simply plug in a Norwegian BERT and rund, so we may need to identify suitable code for existing BERT-based architectures for e.g. English to re-use. For the first task though (document-level SA on NoReC) Jeremy would have an existing set-up for using mBERT that we could perhaps use. &lt;br /&gt;
&lt;br /&gt;
== NLP tasks ==&lt;br /&gt;
* Structured sentiment analysis: [https://github.com/ltgoslo/norec_fine NoReC_fine]  &lt;br /&gt;
* Sentence-level 2/3-way polarity: [https://github.com/ltgoslo/norec_sentence/ NoReC_sentences]  &lt;br /&gt;
* Negation cues and scopes (evaluation is still being developed): [https://github.com/ltgoslo/norec_neg/ NoReC_neg]&lt;br /&gt;
* PoS tagging: [https://github.com/UniversalDependencies/UD_Norwegian-NynorskLIA ILA] +  NDT [https://github.com/UniversalDependencies/UD_Norwegian-Bokmaal Bokmaal] / [https://github.com/UniversalDependencies/UD_Norwegian-Nynorsk Nynorsk]&lt;br /&gt;
* Dependency parsing: [https://github.com/UniversalDependencies/UD_Norwegian-NynorskLIA ILA] + NDT [https://github.com/UniversalDependencies/UD_Norwegian-Bokmaal Bokmaal] / [https://github.com/UniversalDependencies/UD_Norwegian-Nynorsk Nynorsk] &lt;br /&gt;
* Co-reference resolution (annotation ongoing)&lt;br /&gt;
&lt;br /&gt;
== Lexical  ==&lt;br /&gt;
*[https://www.nb.no/sprakbanken/en/resource-catalogue/oai-nb-no-sbr-27/ Word sense disambiguation in context]&lt;br /&gt;
*[https://github.com/ltgoslo/norwegian-synonyms Norwegian synonyms] (for static models)&lt;br /&gt;
*[https://github.com/ltgoslo/norwegian-analogies Norwegian analogies] (for static models)&lt;br /&gt;
*[https://github.com/ltgoslo/norsentlex NorSentLex]: Sentiment lexicon (for static models)&lt;br /&gt;
&lt;br /&gt;
== Text classification ==&lt;br /&gt;
*[https://github.com/ltgoslo/norec NoReC]; document-level ratings.&lt;br /&gt;
*[https://github.com/ltgoslo/talk-of-norway Talk of Norway]&lt;br /&gt;
*[https://github.com/jerbarnes/norwegian_dialect NorDial]&lt;br /&gt;
&lt;br /&gt;
==Other ==&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1373</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1373"/>
		<updated>2021-06-23T11:52:53Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
The following would be natural places to start. For most of these, while we do have baseline numbers to compare to, we do not have existing set-ups where we could simply plug in a Norwegian BERT and rund, so we may need to identify suitable code for existing BERT-based architectures for e.g. English to re-use. For the first task though (document-level SA on NoReC) Jeremy would have an existing set-up for using mBERT that we could perhaps use. &lt;br /&gt;
&lt;br /&gt;
== NLP tasks ==&lt;br /&gt;
* Structured sentiment analysis: [https://github.com/ltgoslo/norec_fine NoReC_fine]  &lt;br /&gt;
* Sentence-level 2/3-way polarity: [https://github.com/ltgoslo/norec_sentence/ NoReC_sentences]  &lt;br /&gt;
* Negation cues and scopes (evaluation is still being developed): [https://github.com/ltgoslo/norec_neg/ NoReC_neg]&lt;br /&gt;
* PoS tagging: [https://github.com/UniversalDependencies/UD_Norwegian-NynorskLIA ILA] +  NDT [https://github.com/UniversalDependencies/UD_Norwegian-Bokmaal Bokmaal] / [https://github.com/UniversalDependencies/UD_Norwegian-Nynorsk Nynorsk]&lt;br /&gt;
* Dependency parsing: [https://github.com/UniversalDependencies/UD_Norwegian-NynorskLIA ILA] + NDT [https://github.com/UniversalDependencies/UD_Norwegian-Bokmaal Bokmaal] / [https://github.com/UniversalDependencies/UD_Norwegian-Nynorsk Nynorsk] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Lexical  ==&lt;br /&gt;
*[https://www.nb.no/sprakbanken/en/resource-catalogue/oai-nb-no-sbr-27/ Word sense disambiguation in context]&lt;br /&gt;
*[https://github.com/ltgoslo/norwegian-synonyms Norwegian synonyms] (for static models)&lt;br /&gt;
*[https://github.com/ltgoslo/norwegian-analogies Norwegian analogies] (for static models)&lt;br /&gt;
*[https://github.com/ltgoslo/norsentlex NorSentLex]: Sentiment lexicon (for static models)&lt;br /&gt;
&lt;br /&gt;
== Text classification ==&lt;br /&gt;
*[https://github.com/ltgoslo/norec NoReC]; document-level ratings.&lt;br /&gt;
*[https://github.com/ltgoslo/talk-of-norway Talk of Norway]&lt;br /&gt;
*[https://github.com/jerbarnes/norwegian_dialect NorDial]&lt;br /&gt;
&lt;br /&gt;
==Other ==&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1372</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1372"/>
		<updated>2021-06-23T11:51:07Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
The following would be natural places to start. For most of these, while we do have baseline numbers to compare to, we do not have existing set-ups where we could simply plug in a Norwegian BERT and rund, so we may need to identify suitable code for existing BERT-based architectures for e.g. English to re-use. For the first task though (document-level SA on NoReC) Jeremy would have an existing set-up for using mBERT that we could perhaps use. &lt;br /&gt;
&lt;br /&gt;
== NLP tasks ==&lt;br /&gt;
* Structured sentiment analysis: [https://github.com/ltgoslo/norec_fine NoReC_fine]  &lt;br /&gt;
* Sentence-level 2/3-way polarity: [https://github.com/ltgoslo/norec_sentence/ NoReC_sentences]  &lt;br /&gt;
* Negation cues and scopes (evaluation is still being developed): [https://github.com/ltgoslo/norec_neg/ NoReC_neg]&lt;br /&gt;
* PoS tagging: NDT [https://github.com/UniversalDependencies/UD_Norwegian-Bokmaal Bokmaal] / [https://github.com/UniversalDependencies/UD_Norwegian-Nynorsk Nynorsk]&lt;br /&gt;
* Dependency parsing: NDT [https://github.com/UniversalDependencies/UD_Norwegian-Bokmaal Bokmaal] / [https://github.com/UniversalDependencies/UD_Norwegian-Nynorsk Nynorsk] &lt;br /&gt;
* [https://github.com/UniversalDependencies/UD_Norwegian-NynorskLIA Spoken dialects]&lt;br /&gt;
&lt;br /&gt;
== Lexical  ==&lt;br /&gt;
*[https://www.nb.no/sprakbanken/en/resource-catalogue/oai-nb-no-sbr-27/ Word sense disambiguation in context]&lt;br /&gt;
*[https://github.com/ltgoslo/norwegian-synonyms Norwegian synonyms] (for static models)&lt;br /&gt;
*[https://github.com/ltgoslo/norwegian-analogies Norwegian analogies] (for static models)&lt;br /&gt;
*[https://github.com/ltgoslo/norsentlex NorSentLex]: Sentiment lexicon (for static models)&lt;br /&gt;
&lt;br /&gt;
== Text classification ==&lt;br /&gt;
*[https://github.com/ltgoslo/norec NoReC]; document-level ratings.&lt;br /&gt;
*[https://github.com/ltgoslo/talk-of-norway Talk of Norway]&lt;br /&gt;
*[https://github.com/jerbarnes/norwegian_dialect NorDial]&lt;br /&gt;
&lt;br /&gt;
==Other ==&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1370</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1370"/>
		<updated>2021-06-23T11:47:24Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
The following would be natural places to start. For most of these, while we do have baseline numbers to compare to, we do not have existing set-ups where we could simply plug in a Norwegian BERT and rund, so we may need to identify suitable code for existing BERT-based architectures for e.g. English to re-use. For the first task though (document-level SA on NoReC) Jeremy would have an existing set-up for using mBERT that we could perhaps use. &lt;br /&gt;
&lt;br /&gt;
== NLP tasks ==&lt;br /&gt;
* Structured sentiment analysis: [https://github.com/ltgoslo/norec_fine NoReC_fine]  &lt;br /&gt;
* Sentence-level 2/3-way polarity: [https://github.com/ltgoslo/norec_sentence/ NoReC_sentences]  &lt;br /&gt;
* Negation cues and scopes (evaluation is still being developed): [https://github.com/ltgoslo/norec_neg/ NoReC_neg]&lt;br /&gt;
&lt;br /&gt;
=== Linguistic pipeline (dependency parsing or PoS tagging) ===&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-Bokmaal Bokmaal]&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-Nynorsk Nynorsk]&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-NynorskLIA Spoken dialects]&lt;br /&gt;
&lt;br /&gt;
== Lexical  ==&lt;br /&gt;
*[https://www.nb.no/sprakbanken/en/resource-catalogue/oai-nb-no-sbr-27/ Word sense disambiguation in context]&lt;br /&gt;
*[https://github.com/ltgoslo/norwegian-synonyms Norwegian synonyms] (for static models)&lt;br /&gt;
*[https://github.com/ltgoslo/norwegian-analogies Norwegian analogies] (for static models)&lt;br /&gt;
*[https://github.com/ltgoslo/norsentlex NorSentLex]: Sentiment lexicon&lt;br /&gt;
&lt;br /&gt;
== Text classification ==&lt;br /&gt;
*[https://github.com/ltgoslo/norec NoReC]; document-level ratings.&lt;br /&gt;
*[https://github.com/ltgoslo/talk-of-norway Talk of Norway]&lt;br /&gt;
*[https://github.com/jerbarnes/norwegian_dialect NorDial]&lt;br /&gt;
&lt;br /&gt;
==Other ==&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1367</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1367"/>
		<updated>2021-06-23T11:41:09Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
The following would be natural places to start. For most of these, while we do have baseline numbers to compare to, we do not have existing set-ups where we could simply plug in a Norwegian BERT and rund, so we may need to identify suitable code for existing BERT-based architectures for e.g. English to re-use. For the first task though (document-level SA on NoReC) Jeremy would have an existing set-up for using mBERT that we could perhaps use. &lt;br /&gt;
&lt;br /&gt;
== NLP tasks ==&lt;br /&gt;
&lt;br /&gt;
=== NoReC* ===&lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC_fine]: structured sentiment analysis  &lt;br /&gt;
*[https://github.com/ltgoslo/norec_sentence/ NoReC_sentences] sentence-level 2/3-way polarity  &lt;br /&gt;
*[https://github.com/ltgoslo/norec_neg/ NoReC_neg]: negation cues and scopes&lt;br /&gt;
&lt;br /&gt;
=== Linguistic pipeline (dependency parsing or PoS tagging) ===&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-Bokmaal Bokmaal]&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-Nynorsk Nynorsk]&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-NynorskLIA Spoken dialects]&lt;br /&gt;
&lt;br /&gt;
== Lexical  ==&lt;br /&gt;
*[https://www.nb.no/sprakbanken/en/resource-catalogue/oai-nb-no-sbr-27/ Word sense disambiguation in context]&lt;br /&gt;
*[https://github.com/ltgoslo/norwegian-synonyms Norwegian synonyms]&lt;br /&gt;
*[https://github.com/ltgoslo/norwegian-analogies Norwegian analogies]&lt;br /&gt;
*[https://github.com/ltgoslo/norsentlex NorSentLex]: Sentiment lexicon&lt;br /&gt;
&lt;br /&gt;
== Text classification ==&lt;br /&gt;
*[https://github.com/ltgoslo/norec NoReC]; document-level ratings.&lt;br /&gt;
*[https://github.com/ltgoslo/talk-of-norway Talk of Norway]&lt;br /&gt;
*[https://github.com/jerbarnes/norwegian_dialect NorDial]&lt;br /&gt;
&lt;br /&gt;
==Other ==&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1366</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1366"/>
		<updated>2021-06-23T11:37:36Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
The following would be natural places to start. For most of these, while we do have baseline numbers to compare to, we do not have existing set-ups where we could simply plug in a Norwegian BERT and rund, so we may need to identify suitable code for existing BERT-based architectures for e.g. English to re-use. For the first task though (document-level SA on NoReC) Jeremy would have an existing set-up for using mBERT that we could perhaps use. &lt;br /&gt;
&lt;br /&gt;
== NLP tasks ==&lt;br /&gt;
&lt;br /&gt;
=== NoReC* ===&lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC_fine]: structured sentiment analysis  &lt;br /&gt;
*[https://github.com/ltgoslo/norec_sentence/ NoReC_sentences] sentence-level 2/3-way polarity  &lt;br /&gt;
*[https://github.com/ltgoslo/norec_neg/ NoReC_neg]: negation cues and scopes&lt;br /&gt;
&lt;br /&gt;
=== Linguistic pipeline (dependency parsing or PoS tagging) ===&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-Bokmaal Bokmaal]&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-Nynorsk Nynorsk]&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-NynorskLIA Spoken dialects]&lt;br /&gt;
&lt;br /&gt;
== Lexical semantic ==&lt;br /&gt;
*[https://www.nb.no/sprakbanken/en/resource-catalogue/oai-nb-no-sbr-27/ Word sense disambiguation in context]&lt;br /&gt;
&lt;br /&gt;
== Text classification ==&lt;br /&gt;
*[https://github.com/ltgoslo/norec NoReC]; document-level ratings.&lt;br /&gt;
*[https://github.com/ltgoslo/talk-of-norway Talk of Norway]&lt;br /&gt;
*[https://github.com/jerbarnes/norwegian_dialect NorDial]&lt;br /&gt;
&lt;br /&gt;
==Other ==&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1365</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1365"/>
		<updated>2021-06-23T11:37:02Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
The following would be natural places to start. For most of these, while we do have baseline numbers to compare to, we do not have existing set-ups where we could simply plug in a Norwegian BERT and rund, so we may need to identify suitable code for existing BERT-based architectures for e.g. English to re-use. For the first task though (document-level SA on NoReC) Jeremy would have an existing set-up for using mBERT that we could perhaps use. &lt;br /&gt;
&lt;br /&gt;
== NLP tasks ==&lt;br /&gt;
&lt;br /&gt;
=== NoReC* ===&lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC_fine]: structured sentiment analysis  &lt;br /&gt;
*[https://github.com/ltgoslo/norec_sentence/ NoReC_sentences] sentence-level 2/3-way polarity  &lt;br /&gt;
*[https://github.com/ltgoslo/norec_neg/ NoReC_neg]: negation cues and scopes&lt;br /&gt;
&lt;br /&gt;
=== Linguistic pipeline (dependency parsing or PoS tagging) ===&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-Bokmaal Bokmaal]&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-Nynorsk Nynorsk]&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-NynorskLIA Spoken dialects]&lt;br /&gt;
&lt;br /&gt;
== Lexical semantic ==&lt;br /&gt;
*[https://www.nb.no/sprakbanken/en/resource-catalogue/oai-nb-no-sbr-27/ Word sense disambiguation in context]&lt;br /&gt;
&lt;br /&gt;
== Text classification ==&lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC]; for document-level sentiment analysis (i.e. rating prediction). Note that we would want to use another version than the current official release; this has 10k more sentences (and is soon to be officially released).&lt;br /&gt;
*[https://github.com/ltgoslo/talk-of-norway Talk of Norway]&lt;br /&gt;
*[https://github.com/jerbarnes/norwegian_dialect NorDial]&lt;br /&gt;
&lt;br /&gt;
==Other ==&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1363</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1363"/>
		<updated>2021-06-23T11:34:50Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
The following would be natural places to start. For most of these, while we do have baseline numbers to compare to, we do not have existing set-ups where we could simply plug in a Norwegian BERT and rund, so we may need to identify suitable code for existing BERT-based architectures for e.g. English to re-use. For the first task though (document-level SA on NoReC) Jeremy would have an existing set-up for using mBERT that we could perhaps use. &lt;br /&gt;
&lt;br /&gt;
== NLP tasks ==&lt;br /&gt;
&lt;br /&gt;
=== NoReC* ===&lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC_fine]: structured sentiment analysis  &lt;br /&gt;
*[https://github.com/ltgoslo/norec_sentence/ NoReC_sentences] sentence-level 2/3-way polarity  &lt;br /&gt;
*[https://github.com/ltgoslo/norec_neg/ NoReC_neg]: negation cues and scopes&lt;br /&gt;
&lt;br /&gt;
=== Linguistic pipeline (dependency parsing or PoS tagging) ===&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-Bokmaal Bokmaal]&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-Nynorsk Nynorsk]&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-NynorskLIA Spoken dialects]&lt;br /&gt;
&lt;br /&gt;
== Lexical semantic ==&lt;br /&gt;
*[https://www.nb.no/sprakbanken/en/resource-catalogue/oai-nb-no-sbr-27/ Word sense disambiguation in context]&lt;br /&gt;
&lt;br /&gt;
== Text classification ==&lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC]; for document-level sentiment analysis (i.e. rating prediction). Note that we would want to use another version than the current official release; this has 10k more sentences (and is soon to be officially released).&lt;br /&gt;
*[https://github.com/ltgoslo/talk-of-norway Talk of Norway]&lt;br /&gt;
*[https://github.com/jerbarnes/norwegian_dialect NorDial]&lt;br /&gt;
&lt;br /&gt;
==Other ==&lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC_fine]; subset of documents from NoReC annotated with fine-grained sentiment (e.g. for predicting target expression + polarity)&lt;br /&gt;
*[https://github.com/ltgoslo/norne NorNE]; for named entity recognition, extends NDT (also available for the UD version)&lt;br /&gt;
*NoReC_neg; soon to be released; adds negation cues and scopes to the same subset of sentences as in NoReC_fine.&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1361</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1361"/>
		<updated>2021-06-23T11:33:41Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
The following would be natural places to start. For most of these, while we do have baseline numbers to compare to, we do not have existing set-ups where we could simply plug in a Norwegian BERT and rund, so we may need to identify suitable code for existing BERT-based architectures for e.g. English to re-use. For the first task though (document-level SA on NoReC) Jeremy would have an existing set-up for using mBERT that we could perhaps use. &lt;br /&gt;
&lt;br /&gt;
== NLP tasks ==&lt;br /&gt;
&lt;br /&gt;
=== NoReC* ===&lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC_fine]: strucutred sentiment analysis  &lt;br /&gt;
*[https://github.com/ltgoslo/norec_sentence/ NoReC_sentences] sentence-level 2/3-way polarity  &lt;br /&gt;
*[https://github.com/ltgoslo/norec_neg/ NoReC_neg]: negation cues and scopes&lt;br /&gt;
&lt;br /&gt;
=== Linguistic pipeline (dependency parsing or PoS tagging) ===&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-Bokmaal Bokmaal]&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-Nynorsk Nynorsk]&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-NynorskLIA Spoken dialects]&lt;br /&gt;
&lt;br /&gt;
== Lexical semantic ==&lt;br /&gt;
*[https://www.nb.no/sprakbanken/en/resource-catalogue/oai-nb-no-sbr-27/ Word sense disambiguation in context]&lt;br /&gt;
&lt;br /&gt;
== Document classification ==&lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC]; for document-level sentiment analysis (i.e. rating prediction). Note that we would want to use another version than the current official release; this has 10k more sentences (and is soon to be officially released).&lt;br /&gt;
*[https://github.com/ltgoslo/talk-of-norway Talk of Norway]&lt;br /&gt;
*[https://github.com/jerbarnes/norwegian_dialect NorDial]&lt;br /&gt;
&lt;br /&gt;
==Other ==&lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC_fine]; subset of documents from NoReC annotated with fine-grained sentiment (e.g. for predicting target expression + polarity)&lt;br /&gt;
*[https://github.com/ltgoslo/norne NorNE]; for named entity recognition, extends NDT (also available for the UD version)&lt;br /&gt;
*NoReC_neg; soon to be released; adds negation cues and scopes to the same subset of sentences as in NoReC_fine.&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1360</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1360"/>
		<updated>2021-06-23T11:29:43Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
The following would be natural places to start. For most of these, while we do have baseline numbers to compare to, we do not have existing set-ups where we could simply plug in a Norwegian BERT and rund, so we may need to identify suitable code for existing BERT-based architectures for e.g. English to re-use. For the first task though (document-level SA on NoReC) Jeremy would have an existing set-up for using mBERT that we could perhaps use. &lt;br /&gt;
&lt;br /&gt;
== NoReC* == &lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC_fine]: strucutred sentiment analysis  &lt;br /&gt;
*[https://github.com/ltgoslo/norec_sentence/ NoReC_sentences] sentence-level 2/3-way polarity  &lt;br /&gt;
*[https://github.com/ltgoslo/norec_neg/ NoReC_neg]: negation cues and scopes&lt;br /&gt;
&lt;br /&gt;
== Linguistic pipeline (dependency parsing or PoS tagging) ==&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-Bokmaal Bokmaal]&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-Nynorsk Nynorsk]&lt;br /&gt;
*[https://github.com/UniversalDependencies/UD_Norwegian-NynorskLIA Spoken dialects]&lt;br /&gt;
&lt;br /&gt;
== Lexical semantic ==&lt;br /&gt;
*[https://www.nb.no/sprakbanken/en/resource-catalogue/oai-nb-no-sbr-27/ Word sense disambiguation in context]&lt;br /&gt;
&lt;br /&gt;
== Document classification ==&lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC]; for document-level sentiment analysis (i.e. rating prediction). Note that we would want to use another version than the current official release; this has 10k more sentences (and is soon to be officially released).&lt;br /&gt;
*[https://github.com/ltgoslo/talk-of-norway Talk of Norway]&lt;br /&gt;
*[https://github.com/jerbarnes/norwegian_dialect NorDial]&lt;br /&gt;
&lt;br /&gt;
==Other ==&lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC_fine]; subset of documents from NoReC annotated with fine-grained sentiment (e.g. for predicting target expression + polarity)&lt;br /&gt;
*[https://github.com/ltgoslo/norne NorNE]; for named entity recognition, extends NDT (also available for the UD version)&lt;br /&gt;
*NoReC_neg; soon to be released; adds negation cues and scopes to the same subset of sentences as in NoReC_fine.&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Vectors/norlm/norbert&amp;diff=1281</id>
		<title>Vectors/norlm/norbert</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Vectors/norlm/norbert&amp;diff=1281"/>
		<updated>2021-01-13T11:59:41Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;''[http://norlm.nlpl.eu Back to NorLM]''&lt;br /&gt;
&lt;br /&gt;
= NorBERT: Bidirectional Encoder Representations from Transformers =&lt;br /&gt;
&lt;br /&gt;
'''NorBERT''' is a BERT deep learning language model [[https://www.aclweb.org/anthology/N19-1423/ Devlin et al 2019]] trained from scratch for Norwegian. The model can be used to achieve state-of-the-art results for various Norwegian natural language processing tasks.&lt;br /&gt;
These models are part of the ongoing&lt;br /&gt;
[http://norlm.nlpl.eu NorLM initiative] for very large contextualized&lt;br /&gt;
Norwegian language models and associated tools and recipies.&lt;br /&gt;
The NorBERT training setup builds on prior work on&lt;br /&gt;
[https://github.com/TurkuNLP/FinBERT FinBERT] &lt;br /&gt;
by our collaborators at the&lt;br /&gt;
[https://turkunlp.org/ University of Turku].&lt;br /&gt;
&lt;br /&gt;
- '''[http://vectors.nlpl.eu/repository/20/215.zip Download from the NLPL Vector Repository]'''&lt;br /&gt;
&lt;br /&gt;
- '''[https://huggingface.co/ltgoslo/norbert Use with the Huggingface Transformers library]'''&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''NorBERT''' features a custom 30 000 WordPiece vocabulary that has much better coverage of Norwegian words than the multilingual BERT (mBERT) models from Google:&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Vocabulary !! Example of a tokenized sentence&lt;br /&gt;
|-&lt;br /&gt;
| NorBERT || Denne gjengen håper at de sammen skal bidra til å gi kvinne ##fotball ##en i Kristiansand et lenge etterl ##engt ##et løft .&lt;br /&gt;
|-&lt;br /&gt;
| mBERT || Denne g ##jeng ##en h ##å ##per at de sammen skal bid ##ra til å gi k ##vinne ##fo ##t ##ball ##en i Kristiansand et lenge etter ##len ##gte ##t l ##ø ##ft .&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Evaluation ==&lt;br /&gt;
&lt;br /&gt;
We have currently evaluated NorBERT on two standard benchmarks: Part-of-Speech tagging on Bokmål (taken from [https://universaldependencies.org/ the Universal Dependencies project]) and sentence-level binary sentiment classification (created by aggregating the fine-grained annotations in [https://github.com/ltgoslo/norec_fine NoReC_fine] and removing sentences with conflicting or no sentiment). &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Data !! Train !! Dev !! Test&lt;br /&gt;
|-&lt;br /&gt;
| POS || 15,696 || 2,409 || 1939&lt;br /&gt;
|-&lt;br /&gt;
| Sentiment || 2,675|| 516 || 417&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
We fine-tune NorBERT and mBERT for 30 epochs and keep the best model on the dev set. NorBERT outperforms mBERT on both tasks: on POS by 0.6 percentage points, and by 15.6 on sentiment.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Model/task !! mBERT !! NorBERT&lt;br /&gt;
|-&lt;br /&gt;
| Part-of-Speech tagging || ​97.9 || '''98.5'''&lt;br /&gt;
|-&lt;br /&gt;
| Sentence-level binary sentiment classification || ​66.7 || '''82.3'''&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
==Training Corpus==&lt;br /&gt;
We use clean training corpora with ordered sentences:&lt;br /&gt;
&lt;br /&gt;
*[https://www.nb.no/sprakbanken/ressurskatalog/oai-nb-no-sbr-4/ Norsk Aviskorpus] (NAK); 1.7 billion words;&lt;br /&gt;
*[https://dumps.wikimedia.org/nowiki/latest/ Bokmål Wikipedia]; 160 million words;&lt;br /&gt;
*[https://dumps.wikimedia.org/nnwiki/latest/ Nynorsk Wikipedia]; 40 million words;&lt;br /&gt;
&lt;br /&gt;
In total, this comprises about two billion (1 907 072 909) word tokens in 203 million (202 802 665) sentences, both in Bokmål and in Nynorsk; thus, this is a ''joint'' model. In the future, separate Bokmål and Nynorsk models are planned as well.&lt;br /&gt;
&lt;br /&gt;
==Preprocessing ==&lt;br /&gt;
1. Wikipedia texts were extracted using [https://github.com/RaRe-Technologies/gensim/blob/master/gensim/scripts/segment_wiki.py segment_wiki].&lt;br /&gt;
&lt;br /&gt;
2. In NAK, for years up to 2005, the text is in the one-token-per-line format. There are special delimiters signaling the beginning of a new document and providing the URLs. We converted this to running text using a [https://github.com/ltgoslo/NorBERT/blob/main/preprocessing/detokenize.py self-made de-tokenizer].&lt;br /&gt;
&lt;br /&gt;
3. In NAK, everything up to and including 2011 is in the ISO 8859-01 encoding ('Latin-1'). These files were [https://github.com/ltgoslo/NorBERT/blob/main/preprocessing/recode.sh converted] to UTF-8 before any other pre-processing.&lt;br /&gt;
&lt;br /&gt;
4. The resulting corpus was sentence-segmented using [https://stanfordnlp.github.io/stanza/performance.html Stanza]. We left blank lines between documents (and sections in the case of Wikipedia) so that the &amp;quot;next sentence prediction&amp;quot; task doesn't span between documents.&lt;br /&gt;
&lt;br /&gt;
==Vocabulary==&lt;br /&gt;
The vocabulary for the model is of size 30 000 and contains ''cased entries with diacritics''. It is generated from raw text, without, e.g., separating punctuation from word tokens. This means one can feed raw text into NorBERT.&lt;br /&gt;
&lt;br /&gt;
The vocabulary was generated using the SentencePiece algorithm and Tokenizers library ([https://github.com/ltgoslo/NorBERT/blob/main/tokenization/spiece_tokenizer.py code]). The resulting [https://github.com/ltgoslo/NorBERT/blob/main/vocabulary/norwegian_sentencepiece_vocab_30k.json Tokenizers model] was [https://github.com/ltgoslo/NorBERT/blob/main/tokenization/sent2wordpiece.py converted] to the standard [https://github.com/ltgoslo/NorBERT/blob/main/vocabulary/norwegian_wordpiece_vocab_30k.txt BERT WordPiece format].&lt;br /&gt;
&lt;br /&gt;
=NorBERT Model Details=&lt;br /&gt;
&lt;br /&gt;
==Configuration==&lt;br /&gt;
NorBERT corresponds in its configuration to Google's Bert-Base Cased for English, with 12 layers and hidden size 768. [https://github.com/ltgoslo/NorBERT/blob/main/norbert_config.json Configuration file]&lt;br /&gt;
&lt;br /&gt;
==Training Overview==&lt;br /&gt;
NorBERT was trained on the Norwegian academic HPC system called [https://documentation.sigma2.no/hpc_machines/saga.html Saga]. Most of the time the training was distributed across 4 compute nodes and 16 NVIDIA P100 GPUs. Training took approximately 3 weeks. [http://wiki.nlpl.eu/index.php/Eosc/pretraining/nvidia Instructions for reproducing the training setup with EasyBuild]&lt;br /&gt;
&lt;br /&gt;
==Training Code==&lt;br /&gt;
Similar to the creators of [https://github.com/TurkuNLP/FinBERT FinBERT], we employed the [https://github.com/NVIDIA/DeepLearningExamples/tree/master/TensorFlow/LanguageModeling/BERT BERT implementation by NVIDIA] (version 20.06.08) which allows relatively fast multi-node and multi-GPU training.&lt;br /&gt;
&lt;br /&gt;
We made minor changes to this code, mostly to update it to the newer TensorFlow versions ([https://github.com/ltgoslo/NorBERT/tree/main/patches_for_NVIDIA_BERT our patches]).&lt;br /&gt;
&lt;br /&gt;
All the utils we used at the preprocessing and training are published in [https://github.com/ltgoslo/NorBERT our Github repository].&lt;br /&gt;
&lt;br /&gt;
==Training Workflow==&lt;br /&gt;
The Phase 1 (training with maximum sequence length of 128) was being done with batch size 48 and global batch size 48*16=768. Since one global batch contains 768 sentences, approximately 265 000 training steps constitute 1 epoch (one pass over the whole corpus). We have done 3 epochs: 795 000 training steps.&lt;br /&gt;
&lt;br /&gt;
The Phase 2 (training with maximum sequence length of 512) was being done with batch size 8 and global batch size 8*16=128. We aimed at mimicking the original BERT in that at Phase 2 the model should see about 1/9 of the number of sentences seen during Phase 1. Thus, we needed about 68 million sentences, which at the global batch size of 128 boils down to 531 000 training steps more.&lt;br /&gt;
&lt;br /&gt;
Full logs and loss plots can be found [https://github.com/ltgoslo/NorBERT/tree/main/logs here] (the training was on pause on December 25 and 26, since we were solving problems with mixed precision training).&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Vectors/norlm/norelmo&amp;diff=1280</id>
		<title>Vectors/norlm/norelmo</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Vectors/norlm/norelmo&amp;diff=1280"/>
		<updated>2021-01-13T11:59:17Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;''[http://norlm.nlpl.eu Back to NorLM]''&lt;br /&gt;
&lt;br /&gt;
=NorELMo: Embeddings from Language Models for Norwegian=&lt;br /&gt;
&lt;br /&gt;
'''NorELMo''' is a set of bidirectional recurrent ELMo language models trained from scratch on Norwegian Wikipedia, trained as part of the ongoing [http://norlm.nlpl.eu NorLM initiative]. &lt;br /&gt;
ELMo was the first contextualized architecture to become well-known in the NLP community. [[https://www.aclweb.org/anthology/N18-1202/ Peters et al 2018]] describing it got the Best Paper award at the NAACL 2018 conference.  &lt;br /&gt;
&lt;br /&gt;
NorELMO models can be used to achieve state-of-the-art results for various Norwegian natural language processing tasks. In many cases, they may be a viable alternative to [[Vectors/norlm/norbert|NorBERT]], especially if computational resources are scarce.&lt;br /&gt;
&lt;br /&gt;
Download from the [http://vectors.nlpl.eu/repository/ NLPL Vector repository]:&lt;br /&gt;
&lt;br /&gt;
- ID 210: trained on ''lemmatized'' Norwegian Wikipedia Dump of September 2020 ([http://vectors.nlpl.eu/repository/20/210.zip download])&lt;br /&gt;
&lt;br /&gt;
- ID 211: trained on ''tokenized'' Norwegian Wikipedia Dump of September 2020 ([http://vectors.nlpl.eu/repository/20/210.zip download])&lt;br /&gt;
&lt;br /&gt;
==Training Corpus==&lt;br /&gt;
*[https://dumps.wikimedia.org/nowiki/latest/ Norwegian Bokmål Wikipedia] dump from September 2020; 160 million words;&lt;br /&gt;
&lt;br /&gt;
==Preprocessing==&lt;br /&gt;
Both models were trained on the corpus tokenized using [https://ufal.mff.cuni.cz/udpipe UDPipe]. The '''lemmatized''' model is trained on the version of the corpus where raw word forms were replaced with their lemmas (`kontorer' --&amp;gt; `kontor'). For different tasks, different models can be better.&lt;br /&gt;
&lt;br /&gt;
==Vocabulary==&lt;br /&gt;
Both models were trained with vocabularies comprising 100 000 most frequent words in the corresponding training corpus. The vocabularies are published together with the models in the archives linked above.&lt;br /&gt;
&lt;br /&gt;
==Training workflow==&lt;br /&gt;
Each models was trained for 3 epochs with batch size 192. We employed a [https://github.com/ltgoslo/simple_elmo_training version of the original ELMo training code updated to work better with the recent TensorFlow versions].  All the hyperparameters were left at their default values, except LSTM dimensionality reduced to 2048 from the default 4096.&lt;br /&gt;
&lt;br /&gt;
==Usage==&lt;br /&gt;
The NorELMO models are published in two formats: &lt;br /&gt;
&lt;br /&gt;
1. TensorFlow checkpoints&lt;br /&gt;
&lt;br /&gt;
2. HDF5 model files&lt;br /&gt;
&lt;br /&gt;
We recommend to use our [https://pypi.org/project/simple-elmo/ simple-elmo] Python library to do stuff with NorELMo models.&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Vectors/norlm&amp;diff=1263</id>
		<title>Vectors/norlm</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Vectors/norlm&amp;diff=1263"/>
		<updated>2021-01-13T08:05:02Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Norwegian Large-scale Language Models =&lt;br /&gt;
&lt;br /&gt;
[[File:norbert.png|thumb|right|150px]]&lt;br /&gt;
Welcome to the emerging collection of large-scale contextualized&lt;br /&gt;
language models for the Norwegian language.&lt;br /&gt;
NorLM is a joint initiative of the projects &lt;br /&gt;
[https://www.eosc-nordic.eu/ EOSC-Nordic] (European Open Science Cloud) and&lt;br /&gt;
[https://www.mn.uio.no/ifi/english/research/projects/sant/index.html SANT]&lt;br /&gt;
(Sentiment Analysis for Norwegian), &lt;br /&gt;
coordinated by the&lt;br /&gt;
[https://www.mn.uio.no/ifi/english/research/groups/ltg/ Language Technology Group] (LTG)&lt;br /&gt;
at the University of Oslo.&lt;br /&gt;
&lt;br /&gt;
We are working to provide these models and supporting tools for researchers and developers in Natural&lt;br /&gt;
Language Processing (NLP) for the Norwegian language.&lt;br /&gt;
We do so in the hope of facilitating scientific experimentation with and practical applications of state-of-the-art&lt;br /&gt;
NLP architectures, as well as to enable others to develop their own large-scale models, for example for&lt;br /&gt;
domain- or application-specific tasks, language variants, or even other languages than Norwegian.&lt;br /&gt;
&lt;br /&gt;
= Available Models =&lt;br /&gt;
&lt;br /&gt;
At this initial stage of development, Norwegian models for two common architecture variants are available:&lt;br /&gt;
&lt;br /&gt;
* [[Vectors/norlm/elmo|NorELMo: LSTM-Based Architectures]]&lt;br /&gt;
* [[Vectors/norlm/norbert|NorBERT: Transformer-Based Architectures]]&lt;br /&gt;
&lt;br /&gt;
We emphatically welcome all kinds of user feedback, including of course suggestions for improvement&lt;br /&gt;
or suggestions for additional types of Norwegian contextualized language models or associated tools.&lt;br /&gt;
Please contact us via the NorLM technical coordinator,&lt;br /&gt;
[https://www.mn.uio.no/ifi/english/people/aca/andreku/ Andrey Kutuzov].&lt;br /&gt;
= License and Access =&lt;br /&gt;
&lt;br /&gt;
All Norwegian language models from the NorLM initiative are&lt;br /&gt;
publicly available for download from the&lt;br /&gt;
[http://vectors.nlpl.eu/repository NLPL Vectors Repository], with a [https://creativecommons.org/licenses/by/4.0/ CC BY 4.0 license].&lt;br /&gt;
The NorBERT model is also included with the &lt;br /&gt;
[https://huggingface.co/transformers/ Huggingface Transformers Library].&lt;br /&gt;
&lt;br /&gt;
To receive announcements of updates and availability of additional&lt;br /&gt;
models, please self-subscribe to our very low-traffic NorLM&lt;br /&gt;
[http://lists.nlpl.eu/mailman/listinfo/norlm mailing list].&lt;br /&gt;
&lt;br /&gt;
= Acknowledgements =&lt;br /&gt;
&lt;br /&gt;
The NorLM resources are being developed on the Norwegian national supercomputing services operated by&lt;br /&gt;
[https://www.sigma2.no/ UNINETT Sigma2], the National Infrastructure for High Performance Computing and Data Storage in Norway.&lt;br /&gt;
Software provisioning was financially supported through the European&lt;br /&gt;
[https://www.eosc-nordic.eu/ EOSC-Nordic] project; data preparation and evaluation&lt;br /&gt;
were supported by the Norwegian&lt;br /&gt;
[https://www.mn.uio.no/ifi/english/research/projects/sant/index.html SANT] project.&lt;br /&gt;
We are indebted to all funding agencies involved, the University of Oslo, and the&lt;br /&gt;
Norwegian tax payer.&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Vectors/norlm&amp;diff=1254</id>
		<title>Vectors/norlm</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Vectors/norlm&amp;diff=1254"/>
		<updated>2021-01-12T21:08:13Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Norwegian Large-scale Language Models =&lt;br /&gt;
&lt;br /&gt;
[[File:norbert.png|thumb|right|150px]]&lt;br /&gt;
Welcome to the emerging collection of large-scale contextualized&lt;br /&gt;
language models for the Norwegian language.&lt;br /&gt;
NorLM is a joint initiative of the projects &lt;br /&gt;
[https://www.eosc-nordic.eu/ EOSC-Nordic] (European Open Science Cloud) and&lt;br /&gt;
[https://www.mn.uio.no/ifi/english/research/projects/sant/index.html SANT]&lt;br /&gt;
(Sentiment Analysis for Norwegian), &lt;br /&gt;
coordinated by the&lt;br /&gt;
[https://www.mn.uio.no/ifi/english/research/groups/ltg/ Language Technology Group] (LTG)&lt;br /&gt;
at the University of Oslo.&lt;br /&gt;
&lt;br /&gt;
We are working to provide these models and supporting tools for researchers and developers in Natural&lt;br /&gt;
Language Processing (NLP) for the Norwegian language.&lt;br /&gt;
We do so in the hope of facilitating scientific experimentation with and practical applications of state-of-the-art&lt;br /&gt;
NLP architectures, as well as to enable others to develop their own large-scale models, for example for&lt;br /&gt;
domain- or application-specific tasks, language variants, or even other languages than Norwegian.&lt;br /&gt;
&lt;br /&gt;
= Available Models =&lt;br /&gt;
&lt;br /&gt;
At this initial stage of development, Norwegian models for two common architecture variants are available:&lt;br /&gt;
&lt;br /&gt;
* [[Vectors/norlm/elmo|NorELMo: LSTM-Based Architectures]]&lt;br /&gt;
* [[Vectors/norlm/norbert|NorBERT: Transformer-Based Architectures]]&lt;br /&gt;
&lt;br /&gt;
We emphatically welcome all kinds of user feedback, including of course suggestions for improvement&lt;br /&gt;
or suggestions for additional types of Norwegian contextualized language models or associated tools.&lt;br /&gt;
Please contact us via the NorLM technical coordinator,&lt;br /&gt;
[https://www.mn.uio.no/ifi/english/people/aca/andreku/ Andrey Kutuzov].&lt;br /&gt;
= License and Access =&lt;br /&gt;
&lt;br /&gt;
All Norwegian language models from the NorLM initiative are&lt;br /&gt;
publicly available for download from the&lt;br /&gt;
[http://vectors.nlpl.eu/repository NLPL Vectors Repository];&lt;br /&gt;
a subset of the models is also included with the&lt;br /&gt;
[https://huggingface.co/transformers/ Huggingface Transformers Library].&lt;br /&gt;
&lt;br /&gt;
To receive announcements of updates and availability of additional&lt;br /&gt;
models, please self-subscribe to our very low-traffic NorLM&lt;br /&gt;
[http://lists.nlpl.eu/mailman/listinfo/norlm mailing list].&lt;br /&gt;
&lt;br /&gt;
= Acknowledgements =&lt;br /&gt;
&lt;br /&gt;
The NorLM resources are being developed on the Norwegian national supercomputing services operated by&lt;br /&gt;
[https://www.sigma2.no/ UNINETT Sigma2], the National Infrastructure for High Performance Computing and Data Storage in Norway.&lt;br /&gt;
Software provisioning was financially supported through the European&lt;br /&gt;
[https://www.eosc-nordic.eu/ EOSC-Nordic] project; data preparation and evaluation&lt;br /&gt;
were supported by the Norwegian&lt;br /&gt;
[https://www.mn.uio.no/ifi/english/research/projects/sant/index.html SANT] project.&lt;br /&gt;
We are indebted to all funding agencies involved, the University of Oslo, and the&lt;br /&gt;
Norwegian tax payer.&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Vectors/norlm&amp;diff=1253</id>
		<title>Vectors/norlm</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Vectors/norlm&amp;diff=1253"/>
		<updated>2021-01-12T20:56:15Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Norwegian Large-scale Language Models =&lt;br /&gt;
&lt;br /&gt;
[[File:norbert.png|thumb|right|150px]]&lt;br /&gt;
Welcome to the emerging collection of large-scale contextualized&lt;br /&gt;
language models for the Norwegian language.&lt;br /&gt;
NorLM is a joint initiative of the&lt;br /&gt;
[https://www.eosc-nordic.eu/ EOSC-Nordic] (European Open Science Cloud) and&lt;br /&gt;
[https://www.mn.uio.no/ifi/english/research/projects/sant/index.html SANT]&lt;br /&gt;
(Sentiment Analysis for Norwegian) projects,&lt;br /&gt;
coordinated by the&lt;br /&gt;
[https://www.mn.uio.no/ifi/english/research/groups/ltg/ Language Technology Group] (LTG)&lt;br /&gt;
at the University of Oslo.&lt;br /&gt;
&lt;br /&gt;
We are working to provide these models and supporting tools for researchers and developers in Natural&lt;br /&gt;
Language Processing (NLP) for the Norwegian language.&lt;br /&gt;
We do so in the hope of facilitating scientific experimentation with and practical applications of state-of-the-art&lt;br /&gt;
NLP architectures, as well as to enable others to develop their own large-scale models, for example for&lt;br /&gt;
domain- or application-specific tasks, language variants, or even other languages than Norwegian.&lt;br /&gt;
&lt;br /&gt;
= Available Models =&lt;br /&gt;
&lt;br /&gt;
At this initial stage of development, Norwegian models for two common architecture variants are available:&lt;br /&gt;
&lt;br /&gt;
* [[Vectors/norlm/elmo|NorELMo: LSTM-Based Architectures]]&lt;br /&gt;
* [[Vectors/norlm/norbert|NorBERT: Transformer-Based Architectures]]&lt;br /&gt;
&lt;br /&gt;
We emphatically welcome all kinds of user feedback, including of course suggestions for improvement&lt;br /&gt;
or suggestions for additional types of Norwegian contextualized language models or associated tools.&lt;br /&gt;
Please contact us via the NorLM technical coordinator,&lt;br /&gt;
[https://www.mn.uio.no/ifi/english/people/aca/andreku/ Andrey Kutuzov].&lt;br /&gt;
= License and Access =&lt;br /&gt;
&lt;br /&gt;
All Norwegian language models from the NorLM initiative are&lt;br /&gt;
publicly available for download from the&lt;br /&gt;
[http://vectors.nlpl.eu/repository NLPL Vectors Repository];&lt;br /&gt;
a subset of the models is also included with the&lt;br /&gt;
[https://huggingface.co/transformers/ Huggingface Transformers Library].&lt;br /&gt;
&lt;br /&gt;
To receive announcements of updates and availability of additional&lt;br /&gt;
models, please self-subscribe to our very low-traffic NorLM&lt;br /&gt;
[http://lists.nlpl.eu/mailman/listinfo/norlm mailing list].&lt;br /&gt;
&lt;br /&gt;
= Acknowledgements =&lt;br /&gt;
&lt;br /&gt;
The NorLM resources are being developed on the Norwegian national supercomputing services operated by&lt;br /&gt;
[https://www.sigma2.no/ UNINETT Sigma2], the National Infrastructure for High Performance Computing and Data Storage in Norway.&lt;br /&gt;
Software provisioning was financially supported through the European&lt;br /&gt;
[https://www.eosc-nordic.eu/ EOSC-Nordic] project; data preparation and evaluation&lt;br /&gt;
were supported by the Norwegian&lt;br /&gt;
[https://www.mn.uio.no/ifi/english/research/projects/sant/index.html SANT] project.&lt;br /&gt;
We are indebted to all funding agencies involved, the University of Oslo, and the&lt;br /&gt;
Norwegian tax payer.&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1173</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1173"/>
		<updated>2020-12-03T20:50:16Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
The following would be natural places to start. For most of these, while we do have baseline numbers to compare to, we do not have existing set-ups where we could simply plug in a Norwegian BERT and rund, so we may need to identify suitable code for existing BERT-based architectures for e.g. English to re-use. For the first task though (document-level SA on NoReC) Jeremy would have an existing set-up for using mBERT that we could perhaps use. &lt;br /&gt;
&lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC]; for document-level sentiment analysis (i.e. rating prediction). Note that we would want to use another version than the current official release; this has 10k more sentences (and is soon to be officially released).&lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC_fine]; subset of documents from NoReC annotated with fine-grained sentiment (e.g. for predicting target expression + polarity)&lt;br /&gt;
*[https://www.nb.no/sprakbanken/ressurskatalog/oai-nb-no-sbr-10/ NDT]; for dependency parsing or PoS tagging (perhaps best to use the UD version)&lt;br /&gt;
*[https://github.com/ltgoslo/norne NorNE]; for named entity recognition, extends NDT (also available for the UD version)&lt;br /&gt;
*NoReC_neg; soon to be released; adds negation cues and scopes to the same subset of sentences as in NoReC_fine.&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1172</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1172"/>
		<updated>2020-12-03T20:45:19Z</updated>

		<summary type="html">&lt;p&gt;Erikve: /* Emerging Thoughts on Benchmarking */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
The following would be natural places to start. For most of these, while we do have baseline numbers to compare to, we do not have existing set-ups where we could simply plug in a Norwegian BERT and rund, so we may need to identify suitable code for existing BERT-based architectures for e.g. English to re-use. For the first task though (document-level SA on NoReC) Jeremy would have an existing set-up for using mBERT that we could perhaps use. &lt;br /&gt;
&lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC]; for document-level sentiment analysis (i.e. rating prediction). &lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC_fine]; for fine-grained sentiment analysis (e.g. predicting target expression + polarity)&lt;br /&gt;
*[https://www.nb.no/sprakbanken/ressurskatalog/oai-nb-no-sbr-10/ NDT]; for dependency parsing or PoS tagging (perhaps best to use the UD version)&lt;br /&gt;
*[https://github.com/ltgoslo/norne NorNE]; for named entity recognition, extends NDT (also available for the UD version)&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1171</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1171"/>
		<updated>2020-12-03T18:06:15Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
The following would be natural places to start. For most of these one would need to find suitable code for existing BERT-based architectures for e.g. English. For the first though, document-level SA on NoReC, Jeremy would have an existing set-up for using mBERT. &lt;br /&gt;
&lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC]; for document-level sentiment analysis (i.e. rating prediction). &lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC_fine]; for fine-grained sentiment analysis (e.g. predicting target expression + polarity)&lt;br /&gt;
*[https://www.nb.no/sprakbanken/ressurskatalog/oai-nb-no-sbr-10/ NDT]; for dependency parsing or PoS tagging (perhaps best to use the UD version)&lt;br /&gt;
*[https://github.com/ltgoslo/norne NorNE]; for named entity recognition, extends NDT (also available for the UD version)&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1170</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1170"/>
		<updated>2020-12-03T18:03:51Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
This would be natural places to start:&lt;br /&gt;
&lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC]; for document-level sentiment analysis (i.e. rating prediction)&lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC_fine]; for fine-grained sentiment analysis (e.g. predicting target expression + polarity)&lt;br /&gt;
*[https://www.nb.no/sprakbanken/ressurskatalog/oai-nb-no-sbr-10/ NDT]; for dependency parsing or PoS tagging (perhaps best to use the UD version)&lt;br /&gt;
*[https://github.com/ltgoslo/norne NorNE]; for named entity recognition, extends NDT (also available for the UD version)&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1169</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1169"/>
		<updated>2020-12-03T18:03:33Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
This would be natural places to start:&lt;br /&gt;
&lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC]; for document-level sentiment analysis (i.e. rating prediction): &lt;br /&gt;
*[https://github.com/ltgoslo/norec_fine NoReC_fine]; for fine-grained sentiment analysis (e.g. predicting target expression + polarity): &lt;br /&gt;
*[https://www.nb.no/sprakbanken/ressurskatalog/oai-nb-no-sbr-10/ NDT]; for dependency parsing or PoS tagging (perhaps best to use the UD version):  &lt;br /&gt;
*[https://github.com/ltgoslo/norne NorNE]; for named entity recognition, extends NDT (also available for the UD version):&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1168</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1168"/>
		<updated>2020-12-03T18:00:33Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
This would be natural places to start:&lt;br /&gt;
&lt;br /&gt;
NoReC, for document-level sentiment analysis (i.e. rating prediction): &lt;br /&gt;
https://github.com/ltgoslo/norec&lt;br /&gt;
&lt;br /&gt;
NoReC_fine, for fine-grained sentiment analysis (e.g. predicting target expression + polarity): &lt;br /&gt;
https://github.com/ltgoslo/norec_fine&lt;br /&gt;
&lt;br /&gt;
NDT, for dependency parsing or PoS tagging (perhaps best to use the UD version):  &lt;br /&gt;
https://www.nb.no/sprakbanken/ressurskatalog/oai-nb-no-sbr-10/&lt;br /&gt;
&lt;br /&gt;
NorNE, for named entity recognition, extends NDT (also available for the UD version):&lt;br /&gt;
https://github.com/ltgoslo/norne&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
	<entry>
		<id>https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1167</id>
		<title>Eosc/norbert/benchmark</title>
		<link rel="alternate" type="text/html" href="https://wiki.nlpl.eu/index.php?title=Eosc/norbert/benchmark&amp;diff=1167"/>
		<updated>2020-12-03T18:00:05Z</updated>

		<summary type="html">&lt;p&gt;Erikve: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Emerging Thoughts on Benchmarking =&lt;br /&gt;
&lt;br /&gt;
This would be natural places to start:&lt;br /&gt;
&lt;br /&gt;
NoReC, for document-level sentiment analysis (i.e. rating prediction)&lt;br /&gt;
https://github.com/ltgoslo/norec&lt;br /&gt;
&lt;br /&gt;
NoReC_fine, for fine-grained sentiment analysis (e.g. predicting target expression + polarity)&lt;br /&gt;
https://github.com/ltgoslo/norec_fine&lt;br /&gt;
&lt;br /&gt;
NDT, for dependency parsing or PoS tagging (perhaps best to use the UD version) &lt;br /&gt;
https://www.nb.no/sprakbanken/ressurskatalog/oai-nb-no-sbr-10/&lt;br /&gt;
&lt;br /&gt;
NorNE, for named entity recognition, extends NDT (also available for the UD version)&lt;br /&gt;
https://github.com/ltgoslo/norne&lt;/div&gt;</summary>
		<author><name>Erikve</name></author>
		
	</entry>
</feed>