Anil Ramakrishna

ORCID: 0000-0002-7999-0531
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Topic Modeling
  • Natural Language Processing Techniques
  • Anomaly Detection Techniques and Applications
  • Sentiment Analysis and Opinion Mining
  • Privacy-Preserving Technologies in Data
  • Data Visualization and Analytics
  • Semantic Web and Ontologies
  • Domain Adaptation and Few-Shot Learning
  • Machine Learning and Data Classification
  • Speech and dialogue systems
  • Autism Spectrum Disorder Research
  • Media Influence and Health
  • Music and Audio Processing
  • Neural Networks and Applications
  • Data Stream Mining Techniques
  • Speech Recognition and Synthesis
  • Advanced Text Analysis Techniques
  • Simulation Techniques and Applications
  • Hate Speech and Cyberbullying Detection
  • Mobile Crowdsensing and Crowdsourcing
  • Humor Studies and Applications
  • Customer churn and segmentation
  • Text Readability and Simplification
  • Gender Studies in Language
  • Machine Learning in Healthcare

Amazon (United States)
2021-2023

University of Southern California
2013-2021

Southern California University for Professional Studies
2019

University College of Commerce & Business Management
2017

Osmania University
2017

Viterbo University
2015

Over the past few years, Federated Learning (FL) has become an emerging machine learning technique to tackle data privacy challenges through collaborative training. In algorithm, clients submit a locally trained model, and server aggregates these parameters until convergence. Despite significant efforts that have been made FL in fields like computer vision, audio, natural language processing, applications utilizing multimodal streams remain largely unexplored. It is known broad real-world...

10.1145/3580305.3599825 article EN public-domain Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining 2023-08-04

As the world changes, we need to be able update our models and correct false information without costly retraining. Knowledge-based model editing enables precise modifications weights of large language in order modify encoded within. Recent approaches have seen success enabling recall edited for thousands edits at once. However, these fail produce that account associated contextual information. We present K-Edit, an effective approach generating contextually consistent knowledge edits. By...

10.48550/arxiv.2502.10626 preprint EN arXiv (Cornell University) 2025-02-14

Children with Autism Spectrum Disorder (ASD) are known to have difficulty in producing and perceiving emotional facial expressions. Their expressions often perceived as atypical by adult observers. This paper focuses on data driven ways analyze quantify atypicality of children ASD. Our objective is uncover those characteristics gestures that induce the sense Using a carefully collected motion capture database, without ASD compared within six basic emotion categories employing methods from...

10.1109/icassp.2015.7178080 article EN 2015-04-01

Anil Ramakrishna, Victor R. Martínez, Nikolaos Malandrakis, Karan Singla, Shrikanth Narayanan. Proceedings of the 55th Annual Meeting Association for Computational Linguistics (Volume 1: Long Papers). 2017.

10.18653/v1/p17-1153 article EN cc-by Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2017-01-01

Atypical speech prosody is a primary characteristic of autism spectrum disorders (ASD), yet it often excluded from diagnostic instrument algorithms due to poor subjective reliability. Robust, objective prosodic cues can enhance our understanding those aspects which are atypical in autism. In this work, we connect signal-derived descriptors perceptions awkwardness. Subjectively, more awkward less expressive (more monotone) and has perceived rate/rhythm, volume, intonation. We also find...

10.21437/interspeech.2015-374 article EN Interspeech 2022 2015-09-06

Violent content in movies can influence viewers’ perception of the society. For example, frequent depictions certain demographics as perpetrators or victims abuse shape stereotyped attitudes. In this work, we propose to characterize aspects violent solely from language used scripts. This makes our method applicable a movie earlier stages creation even before it is produced. complementary previous works which rely on audio video post production. Our approach based broad range features...

10.1609/aaai.v33i01.3301671 article EN Proceedings of the AAAI Conference on Artificial Intelligence 2019-07-17

We describe a unified active transfer learning framework called Hierarchical Active Transfer Learning (HATL). HATL exploits cluster structure shared between different data domains to perform by imputing labels for unlabeled target and generate effective label queries during learning. The resulting is flexible enough not only adaptive accelerated but also unsupervised semi-supervised derive an intuitive useful upper bound on HATL's error when used infer points. present results synthetic that...

10.1137/1.9781611974010.58 article EN 2015-06-30

Direct content analysis reveals important details about movies including those of gender representations and potential biases.We investigate the differences between male female character depictions in movies, based on patterns language used.Specifically, we use an automatically generated lexicon linguistic norms characterizing ladenness.We multivariate to correlate them with elements movie production.The proposed metric differentiates utterances exhibits some interesting interactions genres...

10.18653/v1/d15-1234 article EN cc-by Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2015-01-01

Humor is an important social construct that serves several roles in human communication. Though subjective, it culturally ubiquitous and often used to diffuse tension, specially intense conversations such as those psychotherapy sessions. Automatic recognition of humor has been considerable interest the natural language processing community thanks its relevance conversational agents. In this work, we present a model for Motivational Interviewing based We use Long Short Term Memory (LSTM)...

10.21437/interspeech.2018-1583 article EN Interspeech 2022 2018-08-28

Rahul Sharma, Anil Ramakrishna, Ansel MacLaughlin, Anna Rumshisky, Jimit Majmudar, Clement Chung, Salman Avestimehr, Gupta. Proceedings of the 2022 Conference North American Chapter Association for Computational Linguistics: Human Language Technologies. 2022.

10.18653/v1/2022.naacl-main.196 article EN cc-by Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies 2022-01-01

Federated Learning (FL) is a popular algorithm to train machine learning models on user data constrained edge devices (for example, mobile phones) due privacy concerns. Typically, FL trained with the assumption that no part of can be egressed from edge. However, in many production settings, specific data-modalities/meta-data are limited device while others not. For commercial SLU systems, it typically desired prevent transmission biometric signals (such as audio recordings input prompt)...

10.48550/arxiv.2403.01615 preprint EN arXiv (Cornell University) 2024-03-03

Decoding methods for large language models (LLMs) usually struggle with the tradeoff between ensuring factuality and maintaining diversity. For example, a higher p threshold in nucleus (top-p) sampling increases diversity but decreases factuality, vice versa. In this paper, we propose REAL (Residual Entropy from Asymptotic Line) sampling, decoding method that achieves improved over by predicting an adaptive of $p$. Specifically, predicts step-wise likelihood LLM to hallucinate, lowers when...

10.48550/arxiv.2406.07735 preprint EN arXiv (Cornell University) 2024-06-11

In this work, we introduce the Learnable Response Scoring Function (LARS) for Uncertainty Estimation (UE) in generative Large Language Models (LLMs). Current scoring functions probability-based UE, such as length-normalized and semantic contribution-based weighting, are designed to solve specific aspects of problem but exhibit limitations, including inability handle biased probabilities under-performance low-resource languages like Turkish. To address these issues, propose LARS, a function...

10.48550/arxiv.2406.11278 preprint EN arXiv (Cornell University) 2024-06-17

Knowledge graphs (KGs) complement Large Language Models (LLMs) by providing reliable, structured, domain-specific, and up-to-date external knowledge. However, KGs LLMs are often developed separately must be integrated after training. We introduce Tree-of-Traversals, a novel zero-shot reasoning algorithm that enables augmentation of black-box with one or more KGs. The equips LLM actions for interfacing KG the to perform tree search over possible thoughts find high confidence paths. evaluate...

10.48550/arxiv.2407.21358 preprint EN arXiv (Cornell University) 2024-07-31

We propose a constraint learning schema for fine-tuning Large Language Models (LLMs) with attribute control. Given training corpus and control criteria formulated as sequence-level on model outputs, our method fine-tunes the LLM while enhancing satisfaction minimal impact its utility generation quality. Specifically, approach regularizes by penalizing KL divergence between desired output distribution, which satisfies constraints, LLM's posterior. This regularization term can be approximated...

10.48550/arxiv.2410.05559 preprint EN arXiv (Cornell University) 2024-10-07

Machine unlearning has been used to remove unwanted knowledge acquired by large language models (LLMs). In this paper, we examine machine from an optimization perspective, framing it as a regularized multi-task problem, where one task optimizes forgetting objective and another the model performance. particular, introduce normalized gradient difference (NGDiff) algorithm, enabling us have better control over trade-off between objectives, while integrating new, automatic learning rate...

10.48550/arxiv.2410.22086 preprint EN arXiv (Cornell University) 2024-10-29

Contrastive decoding (CD) (Li et al., 2023) improves the next-token distribution of a large expert language model (LM) using small amateur LM. Although CD is applied to various LMs and domains enhance open-ended text generation, it still unclear why often works well, when could fail, how we can make better. To deepen our understanding CD, first theoretically prove that be viewed as linearly extrapolating logits from huge hypothetical We also highlight linear extrapolation unable output most...

10.48550/arxiv.2411.01610 preprint EN arXiv (Cornell University) 2024-11-03

10.18653/v1/2024.emnlp-main.484 article EN Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2024-01-01
Coming Soon ...