Alex Ayoub

ORCID: 0000-0003-0341-8747
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Reinforcement Learning in Robotics
  • Advanced Bandit Algorithms Research
  • Supply Chain and Inventory Management
  • Experimental Behavioral Economics Studies
  • Model Reduction and Neural Networks
  • Computability, Logic, AI Algorithms
  • Protein Structure and Dynamics
  • Enzyme Structure and Function
  • Machine Learning in Bioinformatics
  • Explainable Artificial Intelligence (XAI)
  • Smart Grid Energy Management
  • Misinformation and Its Impacts

Intarcia Therapeutics (United States)
2023

University of Alberta
2020-2021

Abstract Three billion years of evolution has produced a tremendous diversity protein molecules 1 , but the full potential proteins is likely to be much greater. Accessing this been challenging for both computation and experiments because space possible larger than those have functions. Here we introduce Chroma, generative model complexes that can directly sample novel structures sequences, conditioned steer process towards desired properties To enable this, diffusion respects conformational...

10.1038/s41586-023-06728-8 article EN cc-by Nature 2023-11-15

This paper studies model-based reinforcement learning (RL) for regret minimization. We focus on finite-horizon episodic RL where the transition model $P$ belongs to a known family of models $\mathcal{P}$, special case which is when in $\mathcal{P}$ take form linear mixtures: $P_θ = \sum_{i=1}^{d} θ_{i}P_{i}$. propose based algorithm that optimism principle: In each episode, set are `consistent' with data collected constructed. The criterion consistency total squared error incurs task...

10.48550/arxiv.2006.01107 preprint EN other-oa arXiv (Cornell University) 2020-01-01

We propose training fitted Q-iteration with log-loss (FQI-LOG) for batch reinforcement learning (RL). show that the number of samples needed to learn a near-optimal policy FQI-LOG scales accumulated cost optimal policy, which is zero in problems where acting optimally achieves goal and incurs no cost. In doing so, we provide general framework proving $\textit{small-cost}$ bounds, i.e. bounds scale achievable cost, RL. Moreover, empirically verify uses fewer than FQI trained squared loss on...

10.48550/arxiv.2403.05385 preprint EN arXiv (Cornell University) 2024-03-08

We prove that single-parameter natural exponential families with subexponential tails are self-concordant polynomial-sized parameters. For subgaussian we establish an exact characterization of the growth rate self-concordance parameter. Applying these findings to bandits allows us fill gaps in literature: show optimistic algorithms for generalized linear enjoy regret bounds both second-order (scale variance optimal arm's reward distribution) and free dependence on bound problem parameter...

10.48550/arxiv.2410.01112 preprint EN arXiv (Cornell University) 2024-10-01
Coming Soon ...