Search Results for author: Luciana Benotti

Found 18 papers, 2 papers with code

The Impact of Answers in Referential Visual Dialog

no code implementations ReInAct 2021 Mauricio Mazuecos, Patrick Blackburn, Luciana Benotti

Here we present work in progress in the study of the impact of different answering models in human generated questions in GuessWhat?!.

Question Generation Question-Generation +1

Visually Grounded Follow-up Questions: a Dataset of Spatial Questions Which Require Dialogue History

1 code implementation ACL (splurobonlp) 2021 Tianai Dong, Alberto Testoni, Luciana Benotti, Raffaella Bernardi

We call the question that restricts the context: trigger, and we call the spatial question that requires the trigger question to be answered: zoomer.

They Are Not All Alike: Answering Different Spatial Questions Requires Different Grounding Strategies

no code implementations EMNLP (SpLU) 2020 Alberto Testoni, Claudio Greco, Tobias Bianchi, Mauricio Mazuecos, Agata Marcante, Luciana Benotti, Raffaella Bernardi

By analyzing LXMERT errors and its attention mechanisms, we find that our classification helps to gain a better understanding of the skills required to answer different spatial questions.

What kinds of errors do reference resolution models make and what can we learn from them?

no code implementations Findings (NAACL) 2022 Jorge Sánchez, Mauricio Mazuecos, Hernán Maina, Luciana Benotti

Referring resolution is the task of identifying the referent of a natural language expression, for example “the woman behind the other woman getting a massage”.

CVQA: Culturally-diverse Multilingual Visual Question Answering Benchmark

no code implementations10 Jun 2024 David Romero, Chenyang Lyu, Haryo Akbarianto Wibowo, Teresa Lynn, Injy Hamed, Aditya Nanda Kishore, Aishik Mandal, Alina Dragonetti, Artem Abzaliev, Atnafu Lambebo Tonja, Bontu Fufa Balcha, Chenxi Whitehouse, Christian Salamea, Dan John Velasco, David Ifeoluwa Adelani, David Le Meur, Emilio Villa-Cueva, Fajri Koto, Fauzan Farooqui, Frederico Belcavello, Ganzorig Batnasan, Gisela Vallejo, Grainne Caulfield, Guido Ivetta, Haiyue Song, Henok Biadglign Ademtew, Hernán Maina, Holy Lovenia, Israel Abebe Azime, Jan Christian Blaise Cruz, Jay Gala, Jiahui Geng, Jesus-German Ortiz-Barajas, Jinheon Baek, Jocelyn Dunstan, Laura Alonso Alemany, Kumaranage Ravindu Yasas Nagasinghe, Luciana Benotti, Luis Fernando D'Haro, Marcelo Viridiano, Marcos Estecha-Garitagoitia, Maria Camila Buitrago Cabrera, Mario Rodríguez-Cantelar, Mélanie Jouitteau, Mihail Mihaylov, Mohamed Fazli Mohamed Imam, Muhammad Farid Adilazuarda, Munkhjargal Gochoo, Munkh-Erdene Otgonbold, Naome Etori, Olivier Niyomugisha, Paula Mónica Silva, Pranjal Chitale, Raj Dabre, Rendi Chevi, Ruochen Zhang, Ryandito Diandaru, Samuel Cahyawijaya, Santiago Góngora, Soyeong Jeong, Sukannya Purkayastha, Tatsuki Kuribayashi, Thanmay Jayakumar, Tiago Timponi Torrent, Toqeer Ehsan, Vladimir Araujo, Yova Kementchedjhieva, Zara Burzo, Zheng Wei Lim, Zheng Xin Yong, Oana Ignat, Joan Nwatu, Rada Mihalcea, Thamar Solorio, Alham Fikri Aji

Visual Question Answering (VQA) is an important task in multimodal AI, and it is often used to test the ability of vision-language models to understand and reason on knowledge present in both visual and textual data.

Question Answering Visual Question Answering

Selectively Answering Visual Questions

no code implementations3 Jun 2024 Julian Martin Eisenschlos, Hernán Maina, Guido Ivetta, Luciana Benotti

We perform the first in-depth analysis of calibration methods and metrics for VQA with in-context learning LMMs.

Avg In-Context Learning +2

A methodology to characterize bias and harmful stereotypes in natural language processing in Latin America

1 code implementation14 Jul 2022 Laura Alonso Alemany, Luciana Benotti, Hernán Maina, Lucía González, Mariela Rajngewerc, Lautaro Martínez, Jorge Sánchez, Mauro Schilman, Guido Ivetta, Alexia Halvorsen, Amanda Mata Rojo, Matías Bordone, Beatriz Busaniche

Our methodology is based on the following principles: * focus on the linguistic manifestations of discrimination on word embeddings and language models, not on the mathematical properties of the models * reduce the technical barrier for discrimination experts%, be it social scientists, domain experts or other * characterize through a qualitative exploratory process in addition to a metric-based approach * address mitigation as part of the training process, not as an afterthought

Bias Detection Decision Making +1

A recipe for annotating grounded clarifications

no code implementations NAACL 2021 Luciana Benotti, Patrick Blackburn

In this paper, we argue that dialogue clarification mechanisms make explicit the process of interpreting the communicative intents of the speaker's utterances by grounding them in the various modalities in which the dialogue is situated.

Natural Language Understanding

Grounding as a Collaborative Process

no code implementations EACL 2021 Luciana Benotti, Patrick Blackburn

Collaborative grounding is a fundamental aspect of human-human dialog which allows people to negotiate meaning.

Language Acquisition

On the role of effective and referring questions in GuessWhat?!

no code implementations WS 2020 Mauricio Mazuecos, Alberto Testoni, Raffaella Bernardi, Luciana Benotti

Regarding our first metric, we find that successful dialogues do not have a higher percentage of effective questions for most models.

Modeling Student Response Times: Towards Efficient One-on-one Tutoring Dialogues

no code implementations WS 2018 Luciana Benotti, Jayadev Bhaskaran, Sigtryggur Kjartansson, David Lang

Knowing how long it would normally take a student to respond to different types of questions could help tutors optimize their own time while answering multiple dialogues concurrently, as well as deciding when to prompt a student again.

Math

Cannot find the paper you are looking for? You can Submit a new open access paper.