Andrea Zugarini

ORCID: 0000-0003-0344-1656
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Natural Language Processing Techniques
  • Topic Modeling
  • Speech Recognition and Synthesis
  • Multimodal Machine Learning Applications
  • Artificial Intelligence in Games
  • Speech and dialogue systems
  • Linguistic Variation and Morphology
  • Authorship Attribution and Profiling
  • Reinforcement Learning in Robotics
  • Diet and metabolism studies
  • Metabolism and Genetic Disorders
  • Semantic Web and Ontologies
  • Text and Document Classification Technologies
  • Renal function and acid-base balance
  • Software Engineering Research
  • Explainable Artificial Intelligence (XAI)
  • Language, Metaphor, and Cognition
  • Retinal Imaging and Analysis
  • Domain Adaptation and Few-Shot Learning
  • Model Reduction and Neural Networks
  • Influenza Virus Research Studies
  • Imbalanced Data Classification Techniques
  • Mental Health Research Topics
  • Advanced Text Analysis Techniques
  • Music and Audio Processing

Expert System (Italy)
2024

University of Sussex
2023

University of Siena
2016-2022

University of Florence
2018-2021

Florence (Netherlands)
2020-2021

Vision-language models (VLMs), which process image and text inputs, are increasingly integrated into chat assistants other consumer AI applications. Without proper safeguards, however, VLMs may give harmful advice (e.g. how to self-harm) or encourage unsafe behaviours consume drugs). Despite these clear hazards, little work so far has evaluated VLM safety the novel risks created by multimodal inputs. To address this gap, we introduce MSTS, a Multimodal Safety Test Suite for VLMs. MSTS...

10.48550/arxiv.2501.10057 preprint EN arXiv (Cornell University) 2025-01-17

The common practice in Machine Learning research is to evaluate the top-performing models based on their performance. However, this often leads overlooking other crucial aspects that should be given careful consideration. In some cases, performance differences between various approaches may insignificant, whereas factors like production costs, energy consumption, and carbon footprint taken into account. Large Language Models (LLMs) are widely used academia industry address NLP problems....

10.5121/ijnlc.2024.13102 article EN International Journal on Natural Language Computing 2024-02-28

We consider a scenario where an artificial agent is reading stream of text composed set narrations, and it informed about the identity some individuals that are mentioned in portion currently being read. The expected to learn follow thus disambiguating mentions discovering new individuals. focus on case which entities relations propose end-to-end trainable memory network learns discover disambiguate them online manner, performing one-shot learning dealing with small number sparse...

10.1109/tnnls.2019.2955597 article EN IEEE Transactions on Neural Networks and Learning Systems 2019-12-25

Real-world business applications require a trade-off between language model performance and size. We propose new method for compression that relies on vocabulary transfer. evaluate the various vertical domains downstream tasks. Our results indicate transfer can be effectively used in combination with other techniques, yielding significant reduction size inference time while marginally compromising performance.

10.18653/v1/2022.emnlp-industry.41 preprint EN cc-by 2022-01-01

Most Machine Learning research evaluates the best solutions in terms of performance. However, race for performing model, many important aspects are often overlooked when, on contrary, they should be carefully considered. In fact, sometimes gaps performance between different approaches neglectable, whereas factors such as production costs, energy consumption, and carbon footprint must take into consideration. Large Language Models (LLMs) extensively adopted to address NLP problems academia...

10.5121/csit.2024.140203 article EN 2024-01-27

This report provides an introduction to some Machine Learning tools within the most common development environments. It mainly focuses on practical problems, skipping any theoretical introduction. is oriented both students trying approach and experts looking for new frameworks.

10.48550/arxiv.1703.05298 preprint EN other-oa arXiv (Cornell University) 2017-01-01

MOTIVATION Alkaptonuria (AKU) is a rare and genetic disease which causes discoloration of bone (a process called ‘ochronosis’) induces early-onset osteoarthritis. AKU data have not been organized yet the has no approved biomarkers. The ability to collect, integrate analyze relevant streams core for developing “Precision Medicine Ecosystem” AKU-dedicated in biological resources are shared between researchers, clinicians patients. Computational modeling can be useful guide generate an...

10.7287/peerj.preprints.2174v1 preprint EN 2016-06-28

MOTIVATION Alkaptonuria (AKU) is a rare and genetic disease which causes discoloration of bone (a process called ‘ochronosis’) induces early-onset osteoarthritis. AKU data have not been organized yet the has no approved biomarkers. The ability to collect, integrate analyze relevant streams core for developing “Precision Medicine Ecosystem” AKU-dedicated in biological resources are shared between researchers, clinicians patients. Computational modeling can be useful guide generate an...

10.7287/peerj.preprints.2174 preprint EN 2016-06-28

In the last decades, deep learning approaches achieved impressive results in many research fields, such as Computer Vision and Natural Language Processing (NLP). NLP particular has greatly benefit from unsupervised methods that allow to learn distributed representation of language. On race for better performances Models have reached hundred billions parameters nowadays. Despite remarkable results, models are still far being fully exploited real world applications. Indeed, these black-boxes,...

10.1109/ijcnn54540.2023.10191364 article EN 2022 International Joint Conference on Neural Networks (IJCNN) 2023-06-18

Large Language Models have proven highly successful at modelling a variety of tasks. However, this comes steep computational cost that hinders wider industrial uptake. In paper, we present MWT: Multi-Word Tokenizer goes beyond word boundaries by representing frequent multi-word expressions as single tokens. MWTs produce more compact and efficient tokenization yields two benefits: (1) Increase in performance due to greater coverage input data given fixed sequence length budget; (2) Faster...

10.18653/v1/2023.emnlp-industry.58 preprint EN cc-by 2023-01-01

Crossword puzzles are popular linguistic games often used as tools to engage students in learning. Educational crosswords characterized by less cryptic and more factual clues that distinguish them from traditional crossword puzzles. Despite there exist several publicly available clue-answer pair databases for crosswords, educational pairs datasets missing. In this article, we propose a methodology build clue generation can be instruct Large Language Models (LLMs). By gathering Wikipedia...

10.48550/arxiv.2404.06186 preprint EN arXiv (Cornell University) 2024-04-09

Understanding human attention is crucial for vision science and AI. While many models exist free-viewing, less known about task-driven image exploration. To address this, we introduce CapMIT1003, a dataset with captions click-contingent explorations, to study during the captioning task. We also present NevaClip, zero-shot method predicting visual scanpaths by combining CLIP NeVA algorithms. NevaClip generates fixations align representations of foveated stimuli captions. The simulated...

10.48550/arxiv.2408.09948 preprint EN arXiv (Cornell University) 2024-08-19

Traditional approaches to Named Entity Recognition (NER) frame the task into a BIO sequence labeling problem. Although these systems often excel in downstream at hand, they require extensive annotated data and struggle generalize out-of-distribution input domains unseen entity types. On contrary, Large Language Models (LLMs) have demonstrated strong zero-shot capabilities. While several works address Zero-Shot NER English, little has been done other languages. In this paper, we define an...

10.48550/arxiv.2409.15933 preprint EN arXiv (Cornell University) 2024-09-24

Large language models present opportunities for innovative Question Answering over Knowledge Graphs (KGQA). However, they are not inherently designed query generation. To bridge this gap, solutions have been proposed that rely on fine-tuning or ad-hoc architectures, achieving good results but limited out-of-domain distribution generalization. In study, we introduce a novel approach called Dynamic Few-Shot Learning (DFSL). DFSL integrates the efficiency of in-context learning and semantic...

10.48550/arxiv.2407.01409 preprint EN arXiv (Cornell University) 2024-07-01

Recently, several specialized instruction-tuned Large Language Models (LLMs) for Named Entity Recognition (NER) have emerged. Compared to traditional NER approaches, these models strong generalization capabilities. Existing LLMs mainly focus on zero-shot in out-of-domain distributions, being fine-tuned an extensive number of entity classes that often highly or completely overlap with test sets. In this work instead, we propose SLIMER, approach designed tackle never-seen-before named tags by...

10.48550/arxiv.2407.01272 preprint EN arXiv (Cornell University) 2024-07-01

Paraphrasing is the task of re-writing an input text using other words, without altering meaning original content. Conversational systems can exploit automatic paraphrasing to make conversation more natural, e.g., talking about a certain topic different paraphrases in time instants. Recently, automatically generating has been approached context Natural Language Generation (NLG). While many existing simply consist rule-based models, recent success Deep Neural Networks several NLG tasks...

10.1109/snams.2019.8931824 preprint EN 2019-10-01

Understanding the mechanisms underlying human attention is a fundamental challenge for both vision science and artificial intelligence. While numerous computational models of free-viewing have been proposed, less known about task-driven image exploration. To address this gap, we present CapMIT1003, database captions click-contingent explorations collected during captioning tasks. CapMIT1003 based on same stimuli from well-known MIT1003 benchmark, which eye-tracking data under conditions...

10.48550/arxiv.2305.12380 preprint EN cc-by arXiv (Cornell University) 2023-01-01

Transformers have been established as one of the most effective neural approach in performing various Natural Language Processing tasks. However, following common trend modern deep architectures, their scale has quickly grown to an extent that reduces concrete possibility for several enterprises train such models from scratch. Indeed, despite high-level performances, general drawback requiring a huge amount training data, computational resources and energy consumption be successfully...

10.1109/ijcnn54540.2023.10191680 article EN 2022 International Joint Conference on Neural Networks (IJCNN) 2023-06-18

Most Machine Learning research evaluates the best solutions in terms of performance. However, race for performing model, many important aspects are often overlooked when, on contrary, they should be carefully considered. In fact, sometimes gaps performance between different approaches neglectable, whereas factors such as production costs, energy consumption, and carbon footprint must take into consideration. Large Language Models (LLMs) extensively adopted to address NLP problems academia...

10.48550/arxiv.2311.01256 preprint EN other-oa arXiv (Cornell University) 2023-01-01
Coming Soon ...