Lavender Yao Jiang

ORCID: 0000-0003-2464-3281
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Artificial Intelligence in Healthcare and Education
  • Machine Learning in Healthcare
  • Advanced Graph Neural Networks
  • Topic Modeling
  • Biomedical Text Mining and Ontologies
  • Radiomics and Machine Learning in Medical Imaging
  • Brain Tumor Detection and Classification
  • Brain Metastases and Treatment
  • Complex Network Analysis Techniques
  • Medical Coding and Health Information
  • Stochastic Gradient Optimization Techniques
  • Privacy-Preserving Technologies in Data
  • Medical Imaging and Analysis
  • EEG and Brain-Computer Interfaces
  • Speech and dialogue systems
  • Functional Brain Connectivity Studies
  • Age of Information Optimization
  • Medical Imaging Techniques and Applications
  • Neural dynamics and brain function
  • Semantic Web and Ontologies
  • Intracerebral and Subarachnoid Hemorrhage Research
  • Machine Learning and ELM
  • Frailty in Older Adults
  • Chronic Disease Management Strategies
  • Natural Language Processing Techniques

NYU Langone Health
2022-2025

New York University
2022-2024

Hinge Health
2023

Carnegie Mellon University
2019-2020

Abstract Physicians make critical time-constrained decisions every day. Clinical predictive models can help physicians and administrators by forecasting clinical operational events. Existing structured data-based have limited use in everyday practice owing to complexity data processing, as well model development deployment 1–3 . Here we show that unstructured notes from the electronic health record enable training of language models, which be used all-purpose engines with low-resistance...

10.1038/s41586-023-06160-y article EN cc-by Nature 2023-06-07

The adoption of large language models (LLMs) in healthcare demands a careful analysis their potential to spread false medical knowledge. Because LLMs ingest massive volumes data from the open Internet during training, they are potentially exposed unverified knowledge that may include deliberately planted misinformation. Here, we perform threat assessment simulates data-poisoning attack against Pile, popular dataset used for LLM development. We find replacement just 0.001% training tokens...

10.1038/s41591-024-03445-1 article EN cc-by-nc-nd Nature Medicine 2025-01-08

The detection and tracking of metastatic cancer over the lifetime a patient remains major challenge in clinical trials real-world care. Advances deep learning combined with massive datasets may enable development tools that can address this challenge. We present NYUMets-Brain, world's largest, longitudinal, dataset consisting imaging, follow-up, medical management 1,429 patients. Using we developed Segmentation-Through-Time, neural network which explicitly utilizes longitudinal structure...

10.1038/s41467-024-52414-2 article EN cc-by-nc-nd Nature Communications 2024-09-17

BACKGROUND: The development of accurate machine learning algorithms requires sufficient quantities diverse data. This poses a challenge in health care because the sensitive and siloed nature biomedical information. Decentralized through federated (FL) avoid data aggregation by instead distributing to before centrally updating one global model. OBJECTIVE: To establish multicenter collaboration assess feasibility using FL train models for intracranial hemorrhage (ICH) detection without sharing...

10.1227/neu.0000000000002198 article EN Neurosurgery 2022-11-08

Graph convolutional neural networks (GCNNs) are a powerful extension of deep learning techniques to graph-structured data problems. We empirically evaluate several pooling methods for GCNNs, and combinations those graph with three different architectures: GCN, TAGCN, GraphSAGE. confirm that pooling, especially DiffPool, improves classification accuracy on popular datasets find that, average, TAGCN achieves comparable or better than GCN GraphSAGE, particularly larger sparser structures.

10.1109/ieeeconf44664.2019.9048796 article EN 2019-11-01

<title>Abstract</title> The detection and tracking of metastatic cancer over the lifetime a patient remains major challenge in clinical trials real-world care. <sup>1–3</sup> Recent advances deep learning combined with massive, datasets may enable development tools that can address this challenge. We present our work NYUMets Project to develop NYUMets-Brain novel longitudinal neural network (DNN), segmentation-through-time (STT). is world's largest, longitudinal, dataset consisting imaging,...

10.21203/rs.3.rs-2444113/v1 preprint EN cc-by Research Square (Research Square) 2023-01-11

Advances in large language models (LLMs) provide new opportunities healthcare for improved patient care, clinical decision-making, and enhancement of physician administrator workflows. However, the potential these importantly depends on their ability to generalize effectively across environments populations, a challenge often underestimated early development. To better understand reasons challenges inform mitigation approaches, we evaluated ClinicLLM, an LLM trained [HOSPITAL]'s notes,...

10.48550/arxiv.2402.10965 preprint EN arXiv (Cornell University) 2024-02-14

We present a non-invasive deep learning approach for tracking cortical spreading depressions (CSDs) in scalp electroencephalography (EEG) signals. Our method, which we refer to as CSD spatially aware convolutional network or CSD-SpArC, combines neural network, extracts temporal features from the EEG signal of each electrode, with graph exploits spatial structure signals on scalp. Using high-density EEG, this combination networks misses no CSDs, even narrowest ones (informed by widths...

10.1109/ner49283.2021.9441333 article EN 2021-05-04

Hongyi Zheng, Yixin Zhu, Lavender Jiang, Kyunghyun Cho, Eric Oermann. Proceedings of the 61st Annual Meeting Association for Computational Linguistics (Volume 4: Student Research Workshop). 2023.

10.18653/v1/2023.acl-srw.18 article EN cc-by 2023-01-01

Traditional evaluation metrics for classification in natural language processing such as accuracy and area under the curve fail to differentiate between models with different predictive behaviors despite their similar performance metrics. We introduce sensitivity score, a metric that scrutinizes models' at vocabulary level provide insights into disparities decision-making logic. assess score on set of representative words test using two classifiers trained hospital readmission statistics....

10.48550/arxiv.2211.07047 preprint EN cc-by-nc-sa arXiv (Cornell University) 2022-01-01

Recent advances in large language models have led to renewed interest natural processing healthcare using the free text of clinical notes. One distinguishing characteristic notes is their long time span over multiple documents. The unique structure creates a new design choice: when context length for model predictor limited, which part should we choose as input? Existing studies either inputs with domain knowledge or simply truncate them. We propose framework analyze sections high predictive...

10.48550/arxiv.2307.07051 preprint EN public-domain arXiv (Cornell University) 2023-01-01

INTRODUCTION: Existing clinical prediction algorithms mostly leverage small cohorts of structured data (e.g., medical imaging or laboratory data). However, large language models have demonstrated the ability to utilize unstructured outperform other machine learning approaches given sufficient data. Training on notes offers a possible alternative algorithm development in tasks. METHODS: An unlabeled dataset over seven million radiology reports and patient histories) was collected from four...

10.1227/neu.0000000000002809_754 article EN Neurosurgery 2024-03-15

INTRODUCTION: Surgical research demands the development of clinical registries, often through time-intensive manual chart review. Natural language processing (NLP) may accelerate registry development, and an ideal automatic (autoregistry) algorithm would be highly accurate while requiring minimal data annotation. NLP approaches including bespoke Regular Expression (RegEx) classifiers Large Language Models (LLM) possess distinct strengths weaknesses have not been compared in setting...

10.1227/neu.0000000000002809_227 article EN Neurosurgery 2024-03-15

Packing and shuffling tokens is a common practice in training auto-regressive language models (LMs) to prevent overfitting improve efficiency. Typically documents are concatenated chunks of maximum sequence length (MSL) then shuffled. However setting the atom size, for each data chunk accompanied by random shuffling, MSL may lead contextual incoherence due from different being packed into same chunk. An alternative approach utilize padding, another packing strategy, avoid only including one...

10.48550/arxiv.2408.09621 preprint EN arXiv (Cornell University) 2024-08-18

Large language models (LLMs) have recently emerged as powerful tools, finding many medical applications. LLMs' ability to coalesce vast amounts of information from sources generate a response-a process similar that human expert-has led see potential in deploying LLMs for clinical use. However, medicine is setting where accurate reasoning paramount. Many researchers are questioning the effectiveness multiple choice question answering (MCQA) benchmarks, frequently used test LLMs. Researchers...

10.48550/arxiv.2412.10982 preprint EN arXiv (Cornell University) 2024-12-14

INTRODUCTION: The development of accurate and generalizable machine learning algorithms requires sufficient quantities diverse data. This poses a challenge in healthcare due to the sensitive siloed nature biomedical information. Decentralized through federated (FL) paradigm avoid need for data aggregation by instead distributing itself before centrally updating one global model. METHODS: Five academic neurosurgery departments across US collaborated establish network using computed tomography...

10.1227/neu.0000000000002375_316 article EN Neurosurgery 2023-03-16

Zihao Yang, Chenkang Zhang, Muru Wu, Xujin Liu, Lavender Jiang, Kyunghyun Cho, Eric Oermann. Proceedings of the 61st Annual Meeting Association for Computational Linguistics (Volume 4: Student Research Workshop). 2023.

10.18653/v1/2023.acl-srw.19 article EN cc-by 2023-01-01

Graph neural networks (GNNs) extend convolutional (CNNs) to graph based data. A question that arises is how much performance improvement does the underlying structure in GNN provide over CNN (that ignores this structure). To address question, we introduce edge entropy and evaluate good an indicator it for possible of GNNs CNNs. Our results on node classification with synthetic real datasets show lower values predict larger expected gains CNNs, and, conversely, higher leads smaller gains.

10.1109/ieeeconf51394.2020.9443451 article EN 2014 48th Asilomar Conference on Signals, Systems and Computers 2020-11-01

Graph neural networks (GNNs) extend convolutional (CNNs) to graph-based data. A question that arises is how much performance improvement does the underlying graph structure in GNN provide over CNN (that ignores this structure). To address question, we introduce edge entropy and evaluate good an indicator it for possible of GNNs CNNs. Our results on node classification with synthetic real datasets show lower values predict larger expected gains CNNs, and, conversely, higher leads smaller gains.

10.48550/arxiv.2012.08698 preprint EN cc-by-sa arXiv (Cornell University) 2020-01-01

Graph convolutional neural networks (GCNNs) are a powerful extension of deep learning techniques to graph-structured data problems. We empirically evaluate several pooling methods for GCNNs, and combinations those graph with three different architectures: GCN, TAGCN, GraphSAGE. confirm that pooling, especially DiffPool, improves classification accuracy on popular datasets find that, average, TAGCN achieves comparable or better than GCN GraphSAGE, particularly larger sparser structures.

10.48550/arxiv.2004.03519 preprint EN other-oa arXiv (Cornell University) 2020-01-01
Coming Soon ...