+ All Categories
Home > Documents > DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following...

DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following...

Date post: 04-Jul-2020
Category:
Upload: others
View: 3 times
Download: 0 times
Share this document with a friend
23
DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR 2018 demonstration track aims at bringing together Academia and Industry and creating a forum where more than written or spoken descriptions of research are available. Thus, demos allow attendees to try and test them during their presentation in a dedicated session adopting a more informal setting. Products, systems or tools are examples of accepted demos, where both early – research prototypes and mature systems were considered. July 24-26, 2018, Canela - Brazil Valeria de Paiva (Nuance, USA) Rodrigo Wilkens (Université Catholique de Louvain, Belgium) Fernando Batista (INESC-ID & ISCTE-IUL, Portugal) Demos Chairs
Transcript
Page 1: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

DEMONSTRATION SESSION

Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR2018 demonstration track aims at bringing together Academia and Industry and creating a forumwhere more than written or spoken descriptions of research are available. Thus, demos allowattendees to try and test them during their presentation in a dedicated session adopting a moreinformal setting. Products, systems or tools are examples of accepted demos, where both early –research prototypes and mature systems were considered.

July 24-26, 2018, Canela - Brazil

Valeria de Paiva (Nuance, USA)Rodrigo Wilkens (Université Catholique de Louvain, Belgium)Fernando Batista (INESC-ID & ISCTE-IUL, Portugal)Demos Chairs

Page 2: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

ACCEPTED DEMOS

• A computational grammar for PortugueseBruno Cuconato and Alexandre Rademaker

• LX-SemanticSimilarityJoão Silva, Marcos Garcia, João Rodrigues, and António Branco

• TEITOK: TEI for corpus linguistics Maarten Janssen

• DeepBonDD: a Deep neural approach to Boundary and Disfluency DetectionMarcos Treviso, Anderson Smidarle, Lilian Hubner, and Sandra Aluisio

• CL-CONLLU Universal Dependencies in Common Lisp Alexandre Rademaker, Fabricio Chalub, Bruno Cuconato, Henrique Muniz, GuilhermePassos

• Automatically Generating Temporal Proto-Narratives From Portuguese HeadlinesArian Pasquali, Vítor Mangaravite, Ricardo Campos, Alípio Mário Jorge, and Adam Jatowt

• SMILLE: Supporting Portuguese as Second LanguageLeonardo Zilio, Rodrigo Wilkens, Maria José Finatto, and Cédrick Fairon

Page 3: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

A computational grammar for Portuguese

Bruno Cuconato1 and Alexandre Rademaker1,2

1 FGV/EMAp2 IBM Research

Abstract. This work presents an ongoing effort towards a Portuguesegrammar under the Grammatical Framework (GF) formalism. GF andthe new grammar are briefly introduced, and then we employ the gram-mar to parse HPSG’s Matrix MRS test suite. We will demonstrate theuse of the grammar in the parsing of text and in natural language appli-cations.

Keywords: grammatical framework · computational grammar · typetheory · functional programming

1 Introduction

Grammatical Framework (GF) is a programming language for grammar writing.It is a functional programming language, with syntax inspired by the Haskellprogramming language [2]; it draws from intuitionistic type theory for its typesystem [3].

GF’s forte lies at multilingual processing. It applies to natural languagesthe distinction made for programming languages: that of abstract and concretesyntaxes. Separating them allows GF to specify a single abstract grammar forseveral concrete languages. Translation between two natural languages there-fore becomes parsing of concrete syntax to its abstract representation, and thenfurther linearization to the target language.

Writing a grammar for even a fragment of a natural language is a complextask. GF boasts a module system, so GF grammars can import other grammarsfor code reusing. GF grammars can thus be divided in resource and applicationgrammars: while the former intend to provide useful linguistic constructs fordownstream grammars in a suitable and stable application programming inter-face (API) (like software libraries do to programs [4]), the latter aim to applythese and other definitions to domain-specific applications.

The GF Resource Grammar Library (RGL) declares a common abstract syn-tax for resource grammars, with a number of grammatical categories, construc-tion functions, and a small test lexicon. Each resource grammar then defines thisstructure in parallel, and is also free to add language-specific extensions.

Listing 1.1. RGL API, resource grammar, and application grammar output examples

> import -retain present/TryEng.gfo

Page 4: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

> cc -one mkS (mkCl (mkNP this_Det (mkN "candy ")) (mkA "good "))

this candy is good

> import present/LangEng.gfo

> p -lang=Eng "these fish are rotten"

PhrUtt NoPConj (UttS (UseCl (TTAnt TPres ASimul) PPos

(PredVP (DetCN (DetQuant this_Quant NumPl) (UseN fish_N ))

(UseComp (CompAP (PositA rotten_A )))))) NoVoc

> import FoodsEng.gf FoodsPor.gf

> p -lang=Eng -tr "that pizza is delicious" | l -lang=Por

Pred (That Pizza) Delicious

essa pizza e deliciosa

In listing 1.1, we can see (in order) the user importing and using the Englishresource grammar API to build a simple sentence; the user importing the Englishresource grammar interface and parsing a sentence with it (notice the detailedoutput of the syntactic structure); finally, the user imports a domain-specificapplication grammar, parses a sentence with it, and linearizes the obtained treein Portuguese. Because the application grammar is specialized to a domain, itcan produce smaller and more semantic trees. 3

2 The Portuguese resource grammar

The current GF RGL supports more than thirty languages, with varying degreesof completeness. This work presents current work on the addition of a Portugueseresource grammar (henceforth PRG) to the RGL.

As an example of the utility of the PRG, a programmer wanting to createa multilingual application grammar involving a Portuguese lexicon would haveto hard code the lexicon’s inflection tables in the application. With the PRG,she can import the resource grammar, which includes a concrete syntax and acomplete set of paradigms for building words. She can then use an overloadedconstructor mkC (for any given class C) which accepts a variable number of argu-ments dependent on the word’s irregularity. For most words, simply providingtheir uninflected form is sufficient to obtain the correct inflection table [1].

3 Experiments and Discussion

In order to test the PRG, we used HPSG’s Matrix MRS test suite of 107 sentencesin English. 4 Our experiment was as follows: we parsed the English sentencesinto trees, removing spuriously ambiguous ones, and linearized the resultingtrees into Portuguese. The Portuguese linearizations were then compared totheir corresponding sentences in the test suite, and analyzed with respect to

3 Generally, application grammars also produce less trees than resource grammars.4 http://moin.delph-in.net/MatrixMrsTestSuiteEn

2 Cuconato and Rademaker

Page 5: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

grammatical correctness. We do not test the translated sentences for equivalencebecause translation equivalence is not a goal of the RGL [4].

Even when parsing the simple sentences of the test suite, the issue of ambi-guity arises. Consider the sentence [Some bark ]. Considering the context of theother sentences it is clear that ‘bark’ here is meant as a verb. But our grammarcan not know such a thing, and then outputs three possible trees, one for barkas a noun, and two for bark as a verb.

Another example of ambiguity is in [the dog could bark ]. The RGL distin-guishes between ‘can’ in the sense of ‘know’ and in the sense of ‘being capable’.These have the same linearization in English, but the Portuguese grammar canthen offer two possible translations [o cachorro sabia ladrar ] and [o cachorropodia ladrar ].

The test suite allowed us to find several mistakes in our implementation.For instance, the handling of compound nouns is wrong, translating [the tobaccogarden dog barked ] to *[o tabaco o jardim o cachorro ladrava].

Besides correcting the mistakes found in the Portuguese linearizations, thereare missing constructors that prevented the linearization of some trees, and a fewphenomena that are still to be implemented, such as the contraction in *[haviagatos em o jardim].

4 Conclusion

When complete, the Portuguese resource grammar will be one of few freely-available computational grammars for Portuguese. In addition to also being opensource, GF offers a whole ecosystem of tools for the use of GF grammars in NLPapplications: compilation of grammars to several formats (such as a portablebinary format and formats for speech recognition grammars), the possibility ofembedding grammars in Haskell, Java, Python, and C# programs, and of course,the use of the RGL for multilingual applications.

In the demo, we will give examples of morphology paradigms of GF and theiruse in Portuguese, as well as offer examples of application grammars using thePRG, such as a logic to natural language translator following [5].

References

1. Detrez, G., Ranta, A.: Smart paradigms and the predictability and complexity ofinflectional morphology. In: Proceedings of the 13th Conference of the EACL. pp.645–653. ACL, Stroudsburg, PA, USA (2012)

2. Marlow, S., et al.: Haskell 2010 language report (2010)3. Ranta, A.: Grammatical Framework: a type-theoretical grammar formalism. Journal

of Functional Programming 14(2), 145189 (2004)4. Ranta, A.: The GF resource grammar library. Linguistic Issues in Language Tech-

nology 2(2), 1–63 (2009)5. Ranta, A.: Translating between language and logic: What is easy and what is diffi-

cult. In: CADE. pp. 5–25. Springer (2011)

A computational grammar for Portuguese 3

Page 6: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

LX-SemanticSimilarity?

Joao Silva1, Marcos Garcia2, Joao Rodrigues1, and Antonio Branco1

1 University of LisbonNLX—Natural Language and Speech Group, Department of Informatics

Faculdade de Ciencias, Campo Grande, 1749-016 Lisboa, Portugal{jsilva,joao.rodrigues,antonio.branco}@di.fc.ul.pt

2 University of Coruna, Faculty of [email protected]

Abstract. We present the LX-SemanticSimilarity web service and therespective demo, offered as an online service for human users. The webservice provides an API to common operations over the LX-DSemVectorsword embeddings for Portuguese without requiring the embeddings to beloaded locally.

Keywords: Web service · Online service · Distributional semantics ·Word embeddings · Portuguese.

1 Introduction

Distributional semantic models, also known as word embeddings, represent themeaning of an expression as a high-dimension vector of real numbers. This vecto-rial representation of meaning allows, among other possibilities, to reify semanticsimilarity in terms of distance in a vector space. Having a way to quantitativelymeasure semantic similarity has opened up many avenues of research that ex-plore how the integration of distributional features can improve a variety ofnatural language processing tasks, such as determining similarity between words[4], formal semantics [1], sentiment analysis [2], etc.

High-quality embeddings are hard to obtain due to the amount of data andcomputational effort required. LX-DSemVectors [7] are publicly available wordembeddings for Portuguese and their existence helps in this regard, though theystill require a great deal of RAM to operate and some technical skills, whichmay pose problems for some researchers, including from the Digital Humanities.In this paper, we present the LX-SemanticSimilarity web service, which pro-vides access to the LX-DSemVectors through an API with several operationscommonly used on such semantic representations.

? The research presented here was partly supported by the ANI/3279/2016 grant,by the Infrastructure for the Science and Technology of the Portuguese Language(PORTULAN / CLARIN), and by a Juan de la Cierva grant (IJCI-2016-29598).

Page 7: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

2 LX-DSemVectors embbedings and LX-LR4DistSemEvalevaluation datasets

LX-DSemVectors [7] are the first publicly available word embeddings for Por-tuguese. Trained over a corpus of 1.7 billion words, these embeddings were evalu-ated over the LX-4WAnalogies dataset [7], a translation of the de facto standardEnglish dataset for analogies [4], and were found to have a performance at thelevel of the state-of-the-art.

LX-LR4DistSemEval [5] is a collection of datasets adapted via translationfrom various English gold standard datasets for different mainstream evaluationtasks for embeddings, namely the analogy task, the conceptual categorizationtask and the semantic similarity task. These datasets provide a standard wayto intrinsically evaluate and compare distributional semantic models for Por-tuguese.

3 LX-SemanticSimilarity

The embeddings in LX-DSemVectors require nearly 6 GB of RAM when loaded,making them unfeasible to use on many desktop computers. We have found thatloading them on a server and accessing them through a web service help to neatlycircumvent this issue.

3.1 Web service

The LX-SemanticSimilarity web service exposes an API with operations com-monly used on word embeddings, namely getting the (cosine) similarity betweentwo words, and also between two sets of words; finding the top-n most simi-lar words, allowing to specify words that contribute positively and words thatcontribute negatively; and getting the n words closest to a given word.

The server works as a XML-RPC wrapper around the gensim [6] library.Having a standard protocol like XML-RPC makes it easy to use any of a varietyof programming languages on the client side, as seen in the following example inPython that queries the service to get the similarity between the words “batata”(potato) and “banana”:

import xmlrpc.client

lxsemsim = xmlrpc.client.ServerProxy(url)

result = lxsemsim.similarity("batata", "banana")

print(result)

3.2 Online service and demo

The LX-SemanticSimilarity online service/demo (http://lxsemsimil.di.fc.ul.pt/)is built on top of the web service and showcases some simple examples of pos-sible applications of embeddings. The users are presented with two modes of

5 Silva et al.

Page 8: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

Fig. 1. Examples of output by LX-SemanticSimilarity online service and demo

operation: They can either (i) provide two words to see their distance and aninteractive visualization of their surrounding vector-space; or (ii) provide a singleword to see a list of the most similar words to it, in a tabular format and as aword cloud. The outputs of these two modes are exemplified in Figure 1.

The first mode is supported by the t-SNEJS JavaScript library,3 an imple-mentation of the t-SNE [3] dimensionality reduction technique; while the word-cloud image is generated by resorting to the wordcloud4 Python package.

References

1. Baroni, M., Bernardi, R., Zamparelli, R.: Frege in space: A program for composi-tional distributional semantics. Linguistic Issues in Language Technology 9, 241–346(2014)

2. Li, J., Jurafsky, D.: Do multi-sense embeddings improve natural language under-standing? arXiv preprint arXiv:1506.01070 (2015)

3. van der Maaten, L., Hinton, G.: Visualizing high-dimensional data using t-SNE.Journal of Machine Learning Research 9, 2579–2605 (2008)

4. Mikolov, T., Chen, K., Corrado, G., Dean, J.: Efficient estimation of word represen-tations in vector space. arXiv preprint arXiv:1301.3781 (2013)

5. Querido, A., de Carvalho, R., Rodrigues, J., Garcia, M., Correia, C., Rendeiro, N.,Pereira, R.V., Campos, M., Silva, J., Branco, A.: LX-LR4DistSemEval: A collec-tion of language resources for the evaluation of distributional semantic models ofPortuguese. Revista da Associacao Portuguesa de Linguıstica 3 (2017)

6. Rehurek, R., Sojka, P.: Software framework for topic modelling with large corpora.In: Proceedings of the LREC 2010 Workshop on New Challenges for NLP Frame-works. pp. 45–50 (2010)

7. Rodrigues, J., Branco, A., Neale, S., Silva, J.: LX-DSemVectors: Distributional se-mantics models for the Portuguese language. In: Proceedings of the 12th Interna-tional Conference on the Computational Processing of Portuguese (PROPOR’16).pp. 259–270 (2016)

3 https://github.com/karpathy/tsnejs4 https://github.com/amueller/word cloud

LX-SemanticSimilarity 6

Page 9: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

TEITOK: TEI for corpus linguistics

Maarten Janssen

CELGA-ILTEC

Abstract. TEITOK is an online environment for building, maintain-ing and searching TEI-based corpora. It provides a wide range of toolsto work with different corpora, for instance for visualizing manuscripttranscriptions along their facsimile images, searching the corpus usingCWB, visualizing documents on the world map, searching dependencyrelations, visualizing time-aligned spoken corpora, etc. This demo willdisplay a number of key uses of the TEITOK framework.

Keywords: Dependency grammar, TEI/XML, Corpus annotation

1 Introduction

TEITOK [1] is a online platform for visualizing, searching, and editing corpora inwhich the corpus texts are kept in the TEI/XML format, a rich and widely usedXML standard for digital texts (http://www.tei-c.org). Other than text-basedcorpora, TEI-based corpora can keep all the typographic information the originalcontains, with a very rich set of annotations defined to account for the demandsof very different types of corpora. This makes TEI the preferred framework incorpora where detailed annotation is crucial, for instance for historical corporawhere detailed paleographic annotation is essential, spoken corpora where pausesand truncations are highly relevant, and learner corpora where annotation ofcorrections and errors by the student are of primary importance.

TEITOK provides an online platform that makes the notoriously difficult taskof using TEI for linguistic annotation easier, by providing a visualization for TEIdocuments, and use generic scripts to interact with a number of computationaltools behind the screens. These scripts allow you to tokenize, annotate, and parsethe TEI documents in the browser, all with simple buttons. It also can exportthe collection of TEI files to a CWB corpus [2], making the corpus searchable.And it provides the option to edit the metadata in the XML file as well as theannotations on a given token using a simple HTML form. This allows researchersfrom a wide range of linguistic areas to build annotated corpora without the needto have detailed information about the processes behind the screens.

TEITOK in principle assumes you created a TEI document with any of themany TEI writing tools out there, and takes over from there, allowing you toadd linguistic annotation into the XML to enrich it and turn it into a searchablecorpus. It does however also have the option to create specific types of TEIdocuments directly from within the interface. And it comes with a set of scriptsto convert various formats into a TEITOK corpus.

Page 10: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

TEITOK can be used with virtually any type of TEI document, indepen-dently of which purpose the document was created for. It can hence make awide variety of TEI documents tokenized, editable, and searchable. To makeTEITOK even more usable for specific linguistic areas, it has various dedicatedvisualization modules to turn it into a dedicated tool for historical corpora,learner corpora, spoken corpora, and LRL corpora.

During this demo we will show how useful TEITOK is for each of theseareas, showing examples from the growing number of corpora using the system.TEITOK is freely available software, that can be installed on a local server. Moreinformation about how to obtain the tool, as well as links to a growing number ofprojects using the system, can be found at the website: (http://www.teitok.org).

2 Dedicated modules

TEITOK has a modular design in which the same XML file can be displayed andedited in a number of different ways, depending on its content. Below is a selec-tion of tools for two specific types of documents: transcriptions of manuscripts,and transcriptions of sound. The modules listed are not mutually exclusive:learner corpora often combine transcriptions of written and oral exams, whereTEITOK can combine both types into a single corpus while keeping their verydifferent nature. They could even be used on the very same XML file if we havea manuscript that was read out aloud. And these modules can then be combinedwith any number of additional modules to add dependency relations, interlinearglossed texts, visualizing documents on the world map, any many more.

2.1 Manuscript transcription

For corpora that are transcriptions of manuscripts, typically either historicalcorpora or learner corpora, TEITOK provides a number of facsimile-orientedoptions. Firstly, TEITOK can store multiple orthographic realizations on eachtoken, for instance the original spelling and the modernized orthography. In thetext visualization, you can select which of those version to display. This makesit possible to switch between, say, a paleographic, a critical, and a normalizedversion of the same text with a simple click, all built from a single XML source.

TEITOK can also keep track of bounding boxes for each XML node, ie. whichpart of the facsimile image corresponds to the paragraph, manuscript line, orword. And when lines (line breaks in TEI) contain bounding boxes, it can presentthe document in a interleaved version, showing the transcription of each linebelow a cut-out of the manuscript (see http://alfclul.clul.ul.pt/teitok/junius11).

It also provides an option to split a facsimile image into lines, and then tran-scribe that page line-by-line, keeping track of the progress of the transcription.The fact that in this way, TEITOK displays the manuscript line directly abovethe transcription makes the transcription process not only quicker but also oftenmore accurate since you get direct visual verification.

8 Janssen

Page 11: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

When each token is provided with bounding boxes, it can also provide avisualization similar to a searchable PDF: an image with a hidden text layer,allowing you to search directly in the image, select text from the image, and getinformation about each token when moving the mouse over a word.

2.2 Spoken corpora

For spoken corpora, TEITOK provides a number of options to create and vi-sualize time-aligned transcriptions. When for a transcription, the correspondingsound file is provided in the metadata, TEITOK can display the sound file abovethe text transcription, and if the utterances are time-aligned, it can produce aplay button in front of each utterance. To create a time-aligned TEI document,TEITOK provides a script to convert EXMARALDA [3] files into the TEI for-mat, turning segments on the tiers into utterances, ordered in an interview-stylemanner.

When utterances are time-aligned, searching through the CWB corpus willnot only provide a list of resulting utterances, but also allow you to directlylisten to the corresponding sound fragment. This makes it easy to find spokenexamples in the corpus based on orthographic clues.

To get a more dedicated speech-driven interface, TEITOK also provides aninterface similar to that found in speech software like ELAN [4] or Praat [5]:a waveform image with the transcription below. The utterances are orderedvertically, and will scroll along as the sound is playing. And you can click on anyutterance to listen to the corresponding sound.

The system also allows you to create a time-aligned transcription directlyin the TEITOK interface: you can select a segment on the timeline, create anutterance for it, and type in the transcription in an HTML form. This way,TEITOK provides a quick interface to create time-aligned spoken corpora inTEI.

References

1. Janssen, M.: TEITOK: Text-faithful annotated corpora. Proceedings of the 10thInternational Conference on Language Resources and Evaluation, LREC 2016 (2016)4037–4043

2. Evert, S., Hardie, A.: Twenty-first century corpus workbench: Updating a queryarchitecture for the new millennium. In: Corpus Linguistics 2011. (2011)

3. Schmidt, T.: Exmaralda - ein modellierungs- und visualisierungsverfahren fr diecomputergesttzte transkription gesprochener sprache. In Buchberger, E., ed.: Pro-ceedings of Konvens 2004. Volume 5. (2004) DE.

4. Wittenburg, P., Brugman, H., Russel, A., Klassmann, A., Sloetjes, H.: ELAN: aprofessional framework for multimodality research. In: Proceedings of LREC 2006.(2006)

5. Boersma, P., Weenink, D.: Praat, a system for doing phonetics by computer. GlotInternational 5(9/10) (2001) 341–345

TEITOK: TEI for corpus linguistics 9

Page 12: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

DeepBonDD: a Deep neural approach toBoundary and Disfluency Detection

Marcos Treviso1, Anderson Smidarle2, Lilian Hubner2, and Sandra Aluisio1

1 Institute of Mathematics and Computer Science, University of Sao Paulo, Brazil2 Pontifical Catholic University of Rio Grande do Sul,Brazil

[email protected], [email protected], [email protected],

[email protected]

Abstract. In this paper, we present DeepBonDD, a web applicationresponsible for segmenting transcripts which also detects disfluenciespresent in the tests. The use of DeepBonDD in transcripts allows fur-ther application of natural language processing tools that depend onwell-formed texts, such as taggers and parsers.

Keywords: Sentence Segmentation · Disfluency Detection · ImpairedSpeech

1 Introduction

In recent years, mild cognitive impairment (MCI) has received great attentionbecause it may represent a preclinical stage of Alzheimer’s Disease (AD). Sev-eral studies have shown that speech production is a sensitive task to detect agingeffects and to differentiate individuals with MCI from healthy ones. Automaticlinguistic analysis tools have been applied to transcripts of narratives in En-glish [2] and also in Brazilian Portuguese [1]. However, the absence of sentenceboundary information and the presence of disfluencies in transcripts prevent thedirect application of Natural Language Processing (NLP) methods that dependon well-formed texts, such as taggers and parsers. Fig. 1 shows the result of atranscript from a neuropsychological retelling task, that does not include capi-talization nor sentence segmentation, and presents disfluencies.

The Sentence Segmentation or sentence boundary detection task can be seenas a specific case of the punctuation recovery task, which attempts not onlyto detect sentence boundaries but also the types of punctuation that occur inthese places. The Disfluency Detection task is concerned with finding regions ofdisfluencies and categorizing them into their types, such as: (i) fillers, which areusually used by the interlocutor to indicate hesitation or to keep control of aconversation, e.g. “ah, hm, bom, entao, digo”; (ii) and edit disfluencies, whichoccur when the interlocutor makes a statement that is not complete or correctand therefore he himself corrects or changes his statement, e.g. “pro castelo naverdade ela vai trabalhar no castelo ne” in Fig. 1.

Here, we present DeepBonDD, a web application responsible for segmentingtranscripts [3] which also detects disfluencies present in the tests [4].

Page 13: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

cinderela a história da cinderela... ela:: encontra um cavaleiro com com um cavalo dai ela fica amiga desse cavalo tudo isso é próximo de um castelo e ela vai pro castelo pro castelo na verdade ela vai trabalhar no castelo né e ela começa a fazer lá...

Fig. 1: Narrative excerpt transcribed using the NURC annotation manual3

2 DeepBonDD

The full pipeline of DeepBonDD can be seen in Fig. 2. After receiving a cleantranscript, i.e. with no punctuation marks and with lower case letters, the pro-cessing steps of DeepBonDD are: (1) Call the sentence boundary detector andthe filler remover (filled pauses and discursive markers only) in parallel; (2) Com-bine the output of these two processes, that is, the sentence boundary signals areentered in their proper places and the filled pauses that were classified as fillersare removed; (3) Apply the edit disfluencies remover (repetitions and revisions),generating a final segmented and free of disfluencies transcription.

fillerremover

transcription

sentence boundarydetector

segmented transcription

list of filledpauses

segmented and without disfluencies

transcription

repetitions and revisions

remover

transcription without fillers

Fig. 2: Processing steps in DeepBonDD

The DeepBonDD’s web interface for the full pipeline can be seen in Fig. 3abefore the processing steps, and in Fig. 3b after applying the pipeline to thetranscript. Furthermore, the user can detect sentence boundaries, fillers or editdisfluencies independently. The user can also select configurations for each de-tection, choosing whether or not to use part-of-speech tags and handcrafted fea-tures, and choosing available trained models (such as MLP, RCNN and RCNNwith CRF) to deal with lexical and audio information. To detect filled pauses,an optional list of filled pauses to be removed can be uploaded by the user.

3 http://www.letras.ufrj.br/nurc-rj/

11 Treviso et al.

Page 14: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

(a) Before (b) After

Fig. 3: DeepBonDD’s web interface

3 Evaluation and Demonstration

In the demonstration session to be held in PROPOR 2018, attendees will beable to use DeepBonDD’s web interface to segment narratives into sentencesand to remove disfluencies of them. We will make available 3 datasets withnarratives in Brazilian Portuguese: (i) The Cinderella Narrative Dataset - 60narrative samples (20 subjects with MCI, 20 with DA and 20 normal elderlycontrol subjects (CTL)) from a narrative production test based on 22 pictures;(ii) The Dog Story Dataset - 10 narratives transcripts (6 CTL and 4 MCI)from a narrative production test based on seven pictures; (iii) Lucia’s Story - 10narrative transcripts (5 CTL, 2 MCI and 3 DA) from the retelling test of BALE(Bateria de Avaliacao da Linguagem no Envelhecimento).

References

1. Aluısio, S.M., da Cunha, A.L., Scarton, C.: Evaluating progression of alzheimer’sdisease by regression and classification methods in a narrative language test inportuguese. In: Silva, J., Ribeiro, R., Quaresma, P., Adami, A., Branco, A. (eds.)International Conference on Computational Processing of the Portuguese Language.pp. 109–114. Springer (2016)

2. Lehr, M., Prud’hommeaux, E.T., Shafran, I., Roark, B.: Fully automated neuropsy-chological assessment for detecting mild cognitive impairment. In: INTERSPEECH.pp. 1039–1042 (2012)

3. Treviso, M., Shulby, C., Aluısio, S.: Evaluating word embeddings for sentence bound-ary detection in speech transcripts. In: Proceedings of the 11th Brazilian Symposiumin Information and Human Language Technology. pp. 151–160 (2017)

4. Treviso, M.V.: Segmentacao de sentencas e deteccao de disfluencias em nar-rativas transcritas de testes neuropsicologicos. Master’s thesis, Universidadede Sao Paulo (2017), http://www.teses.usp.br/teses/disponiveis/55/55134/

tde-05022018-090740/

DeepBonDD: a Deep neural approach to Boundary and Disfluency Detection 12

Page 15: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

CL-CONLLUUniversal Dependencies in Common Lisp

Alexandre Rademaker1,2, Fabricio Chalub1,Bruno Cuconato2, Henrique Muniz1,2, and Guilherme Paulino-Passos1,3

1 IBM Research2 FGV/EMAp

3 PESC/COPPE/UFRJ

Abstract. The growing interest in the Universal Dependencies projectfor creating corpora in different languages, using a common morphologi-cal and syntactic tags, motivate different research groups involved in thecreation and maintenance of corpora the demand for tools for editing,correction and display of syntactic trees. Here we present cl-conllu, aCommon Lisp library for manipulating CoNLL-U files, the file formatused by the Universal Dependencies project.

1 Introduction

The use of different tags for morphological and syntactic annotations, as wellas different annotation conventions, makes it difficult to develop multi-languagesyntax analysis tools and to study common linguistic phenomena between dif-ferent languages [3]. To solve this problem, the Universal Dependencies (UD)project aims to create consistent linguistic annotations between different lan-guages. Recently, the UD project launched version 2.0 of its treebanks [4], alreadyused in the shared task of the Conference on Computational Natural LanguageLearning (CoNLL 2017).

The advancement of the UD project demands tools for helping on corporamaintenance. In particular, as part of the UD-Portuguese-Bosque [6] corpusmaintenance effort, we developed a library called cl-conllu for manipulatingthe CoNLL-U files in the Common Lisp (CL) language. The library provides fea-tures such as reading and writing CoNNL-U files, annotation validation, batchtransformations, queries, the production of different views of syntax trees, eval-uating annotations and comparing different annotations of the same sentence.

2 The CoNLL-U format

Following a syntactic model of dependencies, UD considers that each word isdependent on some other (except for the root of the phrase), called its head,through a specific dependency relation. Besides, by its adoption of lexicalism, inUD the basic units of annotation are syntactic words (not spelling or phonological

Page 16: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

words) [3]. Hence, contractions and clitics are divided, for example, do is brokeninto two tokens de+o.

For the representation of annotations following these principles, the CoNLL-U format was developed. Each file can contain multiple sentences, separatedby a blank line. Sentences start with metadata which is then followed by line-separated words, each of which comprising 10 tab-separated fields, such as se-quential numbering (ID), original form in the text (FORM), lemma (LEMMA),the UD PoS tag (UPOS), morphological attributes (FEATS), index of its headtoken (HEAD), and the universal dependency relation (DEPREL). Multi-wordtokens (orthographic tokens that have been broken into more than one word)also receive a line of their own.

3 The cl-conllu library

The primary data structures of the library are the sentence, token, and multiwordtoken classes. A sentence has as its attributes (‘slot’ in CL) the main list of itstokens and multiword tokens. We chose to keep the tokens and multi-word tokensin separate lists to facilitate the use of these structures by various other libraryfunctions.

The functions read-conllu and write-conllu are the functions for readingand writing CoNNL-U files, respectively. The first one, receives a ‘string’ or anobject of class pathname.4 It returns a linked list of objects of the class sentence.The write-conllu function receives a linked list of sentence objects and a filename and writes the sentences to the file. Among format conversions, the librarycurrently supports the conversion of CoNLL-U files to Prolog and RDF.

Three significant recent additions to the library are: (1) a rule language tofacilitate batch transformations; (2) visualization of the syntactic trees; and (3)a standard query language in syntax trees. These recent additions are the focusof this article.

Starting with the visualization, the function tree-sentence receives a sen-tence and a ‘stream’ and produces a nice vertical tree showing the tokens connec-tions. This function has been inspired by similar function in the UDAPI library[5].

To allow batch transformations, the apply-rules-from-file function hasbeen implemented. This functionality was inspired by the program ‘Corte e Cos-tura’ [2] and it was built for batch correction of annotations (syntactic andmorphological). The function receives a list of rules, a CoNLL-U file to be readand a CoNLL-U to be generated. The function also produces a log file of therule applications. Listing 1.1 presents a rule with more than one pattern, withvariables, in the left-hand side followed by a list of conditions. The variables areidentifiers CL beginning with the character ”?”. The conditions are formed byan operator, a token field that we are interested in testing, and a string that canbe a regular expression.

4 In CL, the ‘pathname’ class represents a path in the operating system’s file system[7].

14 Rademaker et al.

Page 17: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

Listing 1.1. Example of rule

(=> ( ( ? a ( match lemma ” [ aA ] te ” ) )(?b (= lemma ” entao ” ) ) )

( ( ? a ( set upostag ”ADV” ) )(?b ( set upostag ”ADV” ) ) ) )

The query function operates over the trees. It was created to facilitate thelocalization of sentences given a pattern in the corpus. The query language wasinspired by [1].

Listing 1.2. query example

CL> ( query ’ ( nsubj( advc l

(and ( upostag ”VERB” ) ( lemma ” c o r r e r ” ) )( upostag ”VERB” ) )

( upostag ”PROP” ) ) ∗ s en t ence s ∗)

4 Conclusion

We intend to continue adding features to the library, such as: (1) better supportfor sentence validation; (2) expansion of the rule language with support forvariables over expressions and not just variables for tokens, possibly combiningthe query language with the rules language; and (3) support for sentence editinginteractively and other forms of syntactic tree visualization. Finally, we intendto add even more test cases to increase the robustness of the library. The libraryand its source code can be downloaded from the http://github.com/own-pt/

cl-conllu repository and an initial documentation is in the same repository.

References

1. Luotolahti, J., Kanerva, J., Pyysalo, S., Ginter, F.: Sets: Scalable and efficient treesearch in dependency graphs. Proc. of the 2015 Conference of the North AmericanChapter of the Association for Computational Linguistics: Demonstrations. (2015)

2. Mota, C., Santos, D.: Corte e costura no AC/DC: auxiliando a melhoria da anotacaonos corpos (Sep 2009), http://www.linguateca.pt/acesso/corte-e-costura.pdf

3. Nivre, J., de Marneffe, M., Ginter, F., Goldberg, Y., Hajic, J., Manning, C.D., Mc-Donald, R.T., Petrov, S., Pyysalo, S., Silveira, N., Tsarfaty, R., Zeman, D.: Universaldependencies v1: A multilingual treebank collection. (2016)

4. Nivre, J.e.a.: Universal dependencies 2.0 (2017), http://hdl.handle.net/11234/1-1983, LINDAT/CLARIN digital library at the Institute of Formal and AppliedLinguistics, Charles University

5. Popel, M., Zabokrtsky, Z., Vojtek, M.: Udapi: Universal api for universal depen-dencies. Proc. of the NoDaLiDa 2017 Workshop on Universal Dependencies. (May2017)

6. Rademaker, A., Chalub, F., Real, L., Freitas, C., Bick, E., de Paiva Universal De-pendencies for Portuguese, V.: Universal dependencies for portuguese. Proc. of theFourth International Conference on Dependency Linguistics (Depling). (Sep 2017)

7. Steele Jr, G.L.: Common Lisp: the Language. Digital Press 20, 124 (1984)

CL-CONLLU Universal Dependencies in Common Lisp 15

Page 18: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

Automatically Generating TemporalProto-Narratives From Portuguese Headlines

Arian Pasquali1,2, Vítor Mangaravite1, Ricardo Campos1,3, Alípio MárioJorge1,2, and Adam Jatowt4

1 LIAAD – [email protected]

2 FCUP, University of Porto, Portugal{arian.pasquali,amjorge}@fc.up.pt

3 Polytechnic Institute of Tomar - Smart Cities Research Center, [email protected]

4 Kyoto University, [email protected]

Abstract. Conta-me Histórias (Tell me stories) is an online tool thatallows users to automatically generate proto-narrative timelines based onsearch queries on the Portuguese Web Archive. Our approach is basedon a Keyphrase Extraction algorithm and a peak detection method toselect relevant stories over time. It offers a friendly user interface thatenables users to study and revisit topics in the past thus providing adifferent perspective on historical narratives.

Keywords: Information Retrieval · Temporal Summarization

1 Introduction

Despite the latest advances in the field of natural language processing, gener-ating consistent narratives is still an open problem which has been attractingincreasing attention from the research community [4], Portuguese included, overthe last few years. In this work, we aim to give a step ahead in this promisingresearch area by proposing an unsupervised method to automatically generate aproto-narrative timeline based on a (temporal) collection of news headlines. Inour work, a proto-narrative is a rudimentary narrative made of weakly connectedand temporally ordered sentences.

Conta-me Histórias2 is an online application that allows users to exploretopics and relevant events throughout time, without having to read an entirecollection of news articles. This may be very useful, not only for journalistslooking for historical information but also for anyone interested in researchingforgotten stories. One such tool may work as an innovative way to explore pastevents, contributing to a better-informed society.

2 http://bit.ly/ContameHistorias

Page 19: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

2 Proposed Solution

In Conta-me Histórias, users interact with the system through a friendly inter-face where is possible to specify a query (a free text field) and a time interval(last five, ten or twenty years) to build an interactive timeline. In order to guar-antee the plurality and the diversity of the information, we consider news from24 selected Portuguese news providers.

We make use of the Portuguese Web Archive API to obtain a set of newsheadlines related to the user’s query. The Portuguese Web Archive is a third-party initiative that aims to preserve web pages collected from the web since1996. Periodically it collects and stores entire websites, processing the data tomake it searchable and finally providing a full-text search service that enablesthe retrieval of the past versions of the site. In order to build the temporal nar-ratives, we propose a framework to identify relevant headlines and importantdates. This process can be described in 3 steps: (1) Time intervals: to deter-mine the time-intervals we begin by dividing the timespan into 60 equi-widthintervals (partitions). These partitions are used to find peaks of occurrences.The interval boundaries are then given by the fewer partition (smallest peak)among each pair of peaks; (2) Keyphrase candidates detection: For every timeinterval, we then select keyphrase candidates based on the relevance of eachterm that is part of the headline. To this regard, we use an adapted version ofthe keyword extractor YAKE! [1, 2] which, unlike the original version, consid-ers multiple (news) documents; (3) Finally, we eliminate similar keyphrases ina deduplication analysis applying sequence-matcher similarity. As an add-on wealso provide a sentiment analysis tool to classify each headline as for its sen-timent and also a named entity detection method to select related entities inorder to present a world cloud with the most relevant names in the interface.To this regard, we apply a lexicon-based sentiment analysis method that usesSentiLex-PT01 [3], a specialized lexicon for the Portuguese language. Finally,to perform named entity detection we apply PAMPO (PAttern Matching andPOs tagging based algorithm for NER)[5], the method relies on flexible patternmatching, part-of-speech tagging and lexical-based rules and it was developedto process texts written in Portuguese.

A snapshot of the result can be seen in Figure 1 for the query "Dilma Rouss-eff". The result here presented shows the most relevant extracted keyphrasesselected by YAKE! algorithm [2] for each time interval. Each keyphrase may beconstituted by 3-grams up to 100. In addition, by hovering over the titles, we maynot only access the publication date of the document, but also the archived webpage, thus giving the user the possibility to enhance his/her knowledge about agiven topic.

3 Conclusion

In the era of post-truth and fake news, web archive initiatives are importantcontributions to preserve history. In this context, our demo may be considered

17 Pasquali et al.

Page 20: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

Fig. 1. Example of result narrative timeline for the query Dilma Rousseff

an additional solution that allows users to better explore this kind of repository.We believe that making this demo publicly available and accessible for everyoneis an important contribution to foster not only related research, but also theuser’s search experience when looking for past events and summarizing complexinformation. Despite using a web archive as data source, this tool can be adaptedto support different kinds of data sources. Therefore, in the future, we aim totest this framework on top of alternative collections.

4 Acknowledgements

This work is partially funded by the ERDF through the COMPETE 2020 Pro-gramme within project POCI-01-0145-FEDER-006961, and by National Fundsthrough the FCT as part of project UID/EEA/50014/2013.

References

1. Campos, R. and Mangaravite, V. and Pasquali, A. and Jorge, A.M. and Nunes, C.and Jatowt, A.: YAKE! collection-independent automatic keyword extractor. ECIR,806–810 (2018)

2. Campos, R. and Mangaravite, V. and Pasquali, A. and Jorge, A.M. and Nunes, C.and Jatowt, A.: A Text Feature Based Automatic Keyword Extraction Method forSingle Documents. ECIR, 684–691 (2018)

3. Silva, M. J. and Carvalho, P. and Costa, C and Sarmento L.: Automatic Expansionof a Social Judgment Lexicon for Sentiment Analysis. Technical Report TR 10-08.University of Lisbon, Faculty of Sciences, LASIGE, (2010)

4. Jorge, A. and Campos, R. and Jatowt, A. and Nunes, S. First International Work-shop on Narrative Extraction from Text (2016)

5. Rocha, C. and Jorge, A. and Sionara, R. and Brito, P. and Pimenta, C. and Rezende,S. PAMPO: using pattern matching and pos-tagging for effective Named Entitiesrecognition in Portuguese, (2016)

6. Gomes D. and Cruz D. and Miranda J. and Costa M. and Fontes S.: Acquiringand providing access to historical web collections, 10th International Conference onPreservation of Digital Objects, (2013)

Automatically Generating Temporal Proto-Narratives From Portuguese Headlines 18

Page 21: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

SMILLE: Supporting Portuguese as SecondLanguage?

Leonardo Zilio1, Rodrigo Wilkens1, Maria Jose Finatto2, and Cedrick Fairon1

1 Universite catholique de Louvain, Belgium,{leonardo.zilio,rodrigo.wilkens,cedrick.fairon}@uclouvain.be

2 Universidade Federal do Rio Grande do Sul, Brazil, [email protected]

Abstract. This demo presents SMILLE for Portuguese, a system forenhancing pedagogically relevant grammatical structures that can helplanguage learners to read Web-based texts and, at the same time, focuson grammatical structures that are important for their learning process.

Keywords: Second language acquisition · Text enhancements · Gram-matical structures · Reading Assistant.

1 Introduction

Research on the field of Second Language Acquisition (SLA) has already shownthat the mere presentation of raw input to a language learner is not enoughfor ensuring that learning will take place [3]. One way of solving the lack ofsalience in raw input, as suggested by Smith and Truscott [6], is the use of “inputenhancements”, so that the relevant linguistic information is highlighted. This“focus-on-form strategy” [1] has provided a new way to assist language learners,and some studies have shown that input enhancements represent a positive stepin transforming input into intake (e.g. [4, 5]).

In this demo, we present the Smart and Immersive Language Learning Envi-ronment (SMILLE) in its version for Portuguese. SMILLE is a system that canautomatically analyze and enhance written texts by employing Natural LanguageProcessing (NLP) techniques for retrieving pedagogically relevant grammaticalstructures. These structures are directly linked to the different language lev-els described by the Common European Framework of Reference for Languages(CEFR) [7]. SMILLE was developed based on a scenario in which the users arealready taking second language classes and wish to continue the language learn-ing activity by reading Web-based material that corresponds to their interests.In this case, SMILLE can help not only with the text-understanding process, forit has built-in access to dictionaries and meaning-related information, but alsowith improving the users’ awareness of the grammatical structures that corre-spond to their language level. As such, SMILLE can be seen as a complementaryapplication to a language course, where the grammatical structures associated

? Supported by the Walloon Region (Projects BEWARE 1510637 and 1610378) andAltissia International.

Page 22: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

to the user’s level will be on focus (by means of text enhancements), with thebonus of having a plethora of new vocabulary available, since it is designed toprocess any user-chosen, Web-based text.

2 System Description

SMILLE was originally developed for English [8], and then was further extendedto Portuguese, both understood as a foreign language for the learner. Followingthe idea of WERTi [3], it was designed for the users to have independence forchoosing on-line reading material in the foreign language. Using the selectedtext, SMILLE provides a reading assistant module that helps the users to noticelinguistic content of the target language by enhancing (i.e. highlighting) languagestructures in context, while also offering the possibility of looking up meaningand word class information.

SMILLE links the detected information to the guidelines of the CEFR [7], sothat the grammatical enhancements are not limited to isolated linguistic struc-tures, but covers the needs for a given language level. By applying rules on topof the parser annotation, SMILLE also detects grammatical structures that arenot directly retrieved from parsing.

For retrieving the relevant content in the chosen Web page, SMILLE crawlsover the HTML structure and extracts its text content. This text content is thenparsed with PassPort [9], a dependency parsing model for Portuguese trainedwith the Stanford parser [2]. The parsed text content is then analyzed with rulesfor creating new tags for each relevant grammatical structure. After this process,a new Web page is constructed, showing the same information extracted fromthe original one, but with new encoded information that allows for on-the-flymodifications of the text.

Much of the grammatical information that is detected by SMILLE for Por-tuguese requires only that PassPort correctly analyze the word or structure inquestion. However, some structures are retrieved based on rules specifically writ-ten for them. As such, SMILLE combines the analysis done by the parser withhand-written rules to extract text information that would not be easily iden-tified, and would not be salient, in a raw input. The grammatical annotationprovided by SMILLE for Portuguese was tested in different genres in Zilio et al.[10] and achieved an overall mean precision of 84.07%.

SMILLE is divided in two modules: a reading assistant and a teaching assis-tant. The reading assistant is responsible for enhancing grammatical structures,providing access to language dictionaries and also to grammatical explanations ofeach of the structures being displayed. The enhancements are done in real-time,so that the user can change the highlighted structures on the fly. They modifythe text in terms of color coding (font and background colors) and by changingit to boldface (Fig 1). The option for these three modifications is based on theresults of [5], which have shown that the use of three modifications was amongthe best ways of enhancing a text structure without saturation. The teachingassistant suggests what information should be displayed based on a user profile.

20 Zilio et al.

Page 23: DEMONSTRATION SESSION Following the spirit of the Demos …€¦ · DEMONSTRATION SESSION Following the spirit of the Demos Session of PROPOR’2010, 2012, 2014, and 2016, the PROPOR

Fig. 1. Enhanced sentence highlighting the use of a clitic pronoun

SMILLE for Portuguese recognizes up to 71 pedagogically relevant grammat-ical structures in written texts. The rules for these structures encompass boththe Brazilian and the European variants and are based on the CEFR levels fromA1 to B2, so that each rule is linked to a specific level.

In this demonstration, the users will be able to see the system working onthe Web pages that they choose. The interface will be presented via SMILLE’sChrome extension, and users will be able to chose different structures to highlighton the fly by interacting with SMILLE’s menu. The access to online dictionarieswill be available, as will some features of automatic exercise generation. SMILLEallows for two types of interfaces: one that preserves as much as possible of theoriginal Website, and another that extracts the raw text and presents it in aseparate window; and users will be able to interact with both of them.

References

1. Doughty, C.: Second language instruction does make a difference. Studies in secondlanguage acquisition 13(04), 431–469 (1991)

2. Manning, C.D., Surdeanu, M., Bauer, J., Finkel, J., Bethard, S.J., McClosky,D.: The Stanford CoreNLP natural language processing toolkit. In: Associationfor Computational Linguistics (ACL) System Demonstrations. pp. 55–60 (2014),http://www.aclweb.org/anthology/P/P14/P14-5010

3. Meurers, D., Ziai, R., Amaral, L., Boyd, A., Dimitrov, A., Metcalf, V., Ott, N.: En-hancing authentic web pages for language learners. In: Proceedings of the NAACLHLT 2010 Fifth Workshop on Innovative Use of NLP for Building EducationalApplications. pp. 10–18. Association for Computational Linguistics (2010)

4. Plonsky, L., Ziegler, N.: The call-sla interface: Insights from a second-order syn-thesis (2016)

5. Simard, D.: Differential effects of textual enhancement formats on intake. System37(1), 124–135 (2009)

6. Smith, M.S., Truscott, J.: Explaining input enhancement: A mogul perspective.International Review of Applied Linguistics in Language Teaching 52(3), 253–281(2014)

7. Verhelst, N., Van Avermaet, P., Takala, S., Figueras, N., North, B.: Common Euro-pean Framework of Reference for Languages: learning, teaching, assessment. Cam-bridge University Press (2009)

8. Zilio, L., Wilkens, R., Fairon, C.: Using nlp for enhancing second language acqui-sition. In: Proceedings of Recent Advances in Natural Language Processing. pp.839–846 (2017)

9. Zilio, L., Wilkens, R., Fairon, C.: Passport: A dependency parsing model for por-tuguese. In: Proceedings of the International Conference on the ComputationalProcessing of Portuguese (2018)

10. Zilio, L., Wilkens, R., Fairon, C.: Smille for portuguese: Annotation and analysisof grammatical structures in a pedagogical context. In: Proceedings of the Inter-national Conference on the Computational Processing of Portuguese (2018)

SMILLE: Supporting Portuguese as Second Language 21


Recommended