Size: 6491
Comment:
|
Size: 7254
Comment:
|
Deletions are marked like this. | Additions are marked like this. |
Line 22: | Line 22: |
||<style="border:0;padding-left:30px;padding-bottom:15px">Talk summary will be made available shortly.|| | ||<style="border:0;padding-left:30px;padding-bottom:15px">The presentation will cover the task of multi-word lexical simplification, in which a sentence in natural language is made easier to understand by replacing its fragment with a simpler alternative, both of which can consist of many words. In order to explore this new direction, a corpus (MWLS1) including 1462 sentences in English from various sources with 7059 simplifications was prepared through crowdsourcing. Additionally, an automatic solution (Plainifier) for the problem, based on a purpose-trained neural language model, will be discussed along with the evaluation, comparing to human and resource-based baselines. The results of the presented study were also published at the COLING 2020 conference in [[https://coling2020.org/pages/accepted_papers_main_conference|an article of the same title]].|| |
Natural Language Processing Seminar 2020–2021
The NLP Seminar is organised by the Linguistic Engineering Group at the Institute of Computer Science, Polish Academy of Sciences (ICS PAS). It takes place on (some) Mondays, normally at 10:15 am, in the seminar room of the ICS PAS (ul. Jana Kazimierza 5, Warszawa). All recorded talks are available on YouTube. |
NOTE: Due to restriction of admission to the Institute building, only staff and speakers (including external ones) may currently take part in the seminar on site. For all other participants the seminar will be broadcast – please use the link next to the presentation title. |
5 October 2020 |
Piotr Rybak, Robert Mroczkowski, Janusz Tracz (ML Research at Allegro.pl), Ireneusz Gawlik (ML Research at Allegro.pl & AGH University of Science and Technology) |
In recent years, a series of BERT-based models improved the performance of many natural language processing systems. During this talk, we will briefly introduce the BERT model as well as some of its variants. Next, we will focus on the available BERT-based models for Polish language and their results on the KLEJ benchmark. Finally, we will dive into the details of the new model developed in cooperation between ICS PAS and Allegro. |
14 December 2020 |
Piotr Przybyła (Linguistic Engineering Group, Institute of Computer Science, Polish Academy of Sciences) |
Multi-Word Lexical Simplification |
The presentation will cover the task of multi-word lexical simplification, in which a sentence in natural language is made easier to understand by replacing its fragment with a simpler alternative, both of which can consist of many words. In order to explore this new direction, a corpus (MWLS1) including 1462 sentences in English from various sources with 7059 simplifications was prepared through crowdsourcing. Additionally, an automatic solution (Plainifier) for the problem, based on a purpose-trained neural language model, will be discussed along with the evaluation, comparing to human and resource-based baselines. The results of the presented study were also published at the COLING 2020 conference in an article of the same title. |
Please see also the talks given in 2000–2015 and 2015–2020. |