Nan Rosemary Ke

ORCID: 0009-0003-7647-8449
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Bayesian Modeling and Causal Inference
  • Topic Modeling
  • Domain Adaptation and Few-Shot Learning
  • Cancer-related Molecular Pathways
  • Speech Recognition and Synthesis
  • Reinforcement Learning in Robotics
  • Advanced Graph Neural Networks
  • Explainable Artificial Intelligence (XAI)
  • Advanced Breast Cancer Therapies
  • Machine Learning and Algorithms
  • Natural Language Processing Techniques
  • Generative Adversarial Networks and Image Synthesis
  • Gaussian Processes and Bayesian Inference
  • Machine Learning in Healthcare
  • Neural Networks and Applications
  • Multimodal Machine Learning Applications
  • Protein Degradation and Inhibitors
  • Lung Cancer Research Studies
  • Advanced Bandit Algorithms Research
  • Data Stream Mining Techniques
  • Machine Learning and Data Classification
  • Advanced Image and Video Retrieval Techniques
  • Adversarial Robustness in Machine Learning
  • Cancer Genomics and Diagnostics
  • Ovarian cancer diagnosis and treatment

DeepMind (United Kingdom)
2023

Northeastern University
2023

Peking University
2021-2022

Chongqing University
2022

Polytechnique Montréal
2017-2021

Université de Montréal
2015-2021

China University of Geosciences
2021

Mila - Quebec Artificial Intelligence Institute
2021

Harvard University
2011-2018

Dana-Farber Cancer Institute
2011-2018

The two fields of machine learning and graphical causality arose are developed separately. However, there is, now, cross-pollination increasing interest in both to benefit from the advances other. In this article, we review fundamental concepts causal inference relate them crucial open problems learning, including transfer generalization, thereby assaying how can contribute modern research. This also applies opposite direction: note that most work starts premise variables given. A central...

10.1109/jproc.2021.3058954 article EN cc-by Proceedings of the IEEE 2021-02-26

We present MILABOT: a deep reinforcement learning chatbot developed by the Montreal Institute for Learning Algorithms (MILA) Amazon Alexa Prize competition. MILABOT is capable of conversing with humans on popular small talk topics through both speech and text. The system consists an ensemble natural language generation retrieval models, including template-based bag-of-words sequence-to-sequence neural network latent variable models. By applying to crowdsourced data real-world user...

10.48550/arxiv.1709.02349 preprint EN other-oa arXiv (Cornell University) 2017-01-01

The use of dialogue systems as a medium for human-machine interaction is an increasingly prevalent paradigm. A growing number conversation strategies that are learned from large datasets. There well documented instances where interactions with these system have resulted in biased or even offensive conversations due to the data-driven training process. Here, we highlight potential ethical issues arise research, including: implicit biases systems, rise adversarial examples, sources privacy...

10.1145/3278721.3278777 article EN 2018-12-27

Abstract Recent studies suggest that targeting transcriptional machinery can lead to potent and selective anticancer effects in cancers dependent on high constant expression of certain transcription factors for growth survival. Cyclin-dependent kinase 7 (CDK7) is the catalytic subunit CDK-activating complex. Its function required both cell-cycle regulation control gene expression. CDK7 has recently emerged as an attractive cancer target because its inhibition leads decreased transcript...

10.1158/0008-5472.can-19-0119 article EN Cancer Research 2019-05-07

We propose zoneout, a novel method for regularizing RNNs. At each timestep, zoneout stochastically forces some hidden units to maintain their previous values. Like dropout, uses random noise train pseudo-ensemble, improving generalization. But by preserving instead of dropping units, gradient information and state are more readily propagated through time, as in feedforward stochastic depth networks. perform an empirical investigation various RNN regularizers, find that gives significant...

10.48550/arxiv.1606.01305 preprint EN other-oa arXiv (Cornell University) 2016-01-01

The two fields of machine learning and graphical causality arose developed separately. However, there is now cross-pollination increasing interest in both to benefit from the advances other. In present paper, we review fundamental concepts causal inference relate them crucial open problems learning, including transfer generalization, thereby assaying how can contribute modern research. This also applies opposite direction: note that most work starts premise variables are given. A central...

10.48550/arxiv.2102.11107 preprint EN other-oa arXiv (Cornell University) 2021-01-01

Many efforts have been devoted to training generative latent variable models with autoregressive decoders, such as recurrent neural networks (RNN). Stochastic successful in capturing the variability observed natural sequential data speech. We unify ideas from recently proposed architectures into a stochastic model: each step sequence is associated that used condition dynamics for future steps. Training performed amortized variational inference where approximate posterior augmented RNN runs...

10.48550/arxiv.1711.05411 preprint EN other-oa arXiv (Cornell University) 2017-01-01

CDK7 has emerged as an exciting target in oncology due to its roles two important processes that are misregulated cancer cells: cell cycle and transcription. This report describes the discovery of

10.1021/acs.jmedchem.1c01171 article EN cc-by Journal of Medicinal Chemistry 2021-11-02

Deep Neural Network (DNN) acoustic models have yielded many state-of-the-art results in Automatic Speech Recognition (ASR) tasks. More recently, Recurrent (RNN) been shown to outperform DNNs counterparts. However, DNN and RNN tend be impractical deploy on embedded systems with limited computational capacity. Traditionally, the approach for platforms is either train a small directly, or that learns output distribution of large DNN. In this paper, we utilize transfer knowledge We use model...

10.48550/arxiv.1504.01483 preprint EN other-oa arXiv (Cornell University) 2015-01-01

E-type cyclins (cyclins E1 and E2) are components of the core cell cycle machinery overexpressed in many human tumor types. E thought to drive proliferation by activating cyclin-dependent kinase 2 (CDK2). The cyclin gene represents site recurrent integration hepatitis B virus pathogenesis hepatocellular carcinoma, this event is associated with strong up-regulation expression. Regardless underlying mechanism tumorigenesis, majority liver cancers overexpress cyclins. Here we used conditional...

10.1073/pnas.1711477115 article EN Proceedings of the National Academy of Sciences 2018-01-16

Most recommender systems recommend a list of items. The user examines the list, from first item to last, and often chooses attractive does not examine rest. This type behavior can be modeled by cascade model. In this work, we study cascading bandits, an online learning variant model where goal is $K$ most items large set $L$ candidate We propose two algorithms for solving problem, which are based on idea linear generalization. key in our solutions that learn predictor attraction...

10.48550/arxiv.1603.05359 preprint EN other-oa arXiv (Cornell University) 2016-01-01

The fundamental challenge in causal induction is to infer the underlying graph structure given observational and/or interventional data. Most existing algorithms operate by generating candidate graphs and evaluating them using either score-based methods (including continuous optimization) or independence tests. In our work, we instead treat inference process as a black box design neural network architecture that learns mapping from both data structures via supervised training on synthetic...

10.48550/arxiv.2204.04875 preprint EN cc-by arXiv (Cornell University) 2022-01-01

We propose a simple technique for encouraging generative RNNs to plan ahead. train "backward" recurrent network generate given sequence in reverse order, and we encourage states of the forward model predict cotemporal backward model. The is used only during training, plays no role sampling or inference. hypothesize that our approach eases modeling long-term dependencies by implicitly forcing hold information about longer-term future (as contained states). show empirically achieves 9%...

10.48550/arxiv.1708.06742 preprint EN other-oa arXiv (Cornell University) 2017-01-01

Often, the performance on a supervised machine learning task is evaluated with emph{task loss} function that cannot be optimized directly. Examples of such loss functions include classification error, edit distance and BLEU score. A common workaround for this problem to instead optimize emph{surrogate function, as instance cross-entropy or hinge loss. In order remedy effective, it important ensure minimization surrogate results in loss, condition we call emph{consistency loss}. work, propose...

10.48550/arxiv.1511.06456 preprint EN other-oa arXiv (Cornell University) 2015-01-01

Learning long-term dependencies in extended temporal sequences requires credit assignment to events far back the past. The most common method for training recurrent neural networks, back-propagation through time (BPTT), information be propagated backwards every single step of forward computation, potentially over thousands or millions steps. This becomes computationally expensive even infeasible when used with long sequences. Importantly, biological brains are unlikely perform such detailed...

10.48550/arxiv.1809.03702 preprint EN other-oa arXiv (Cornell University) 2018-01-01
Coming Soon ...