Conferences
-
Agnostic Interactive Imitation Learning: New Theory and Practical Algorithms.
ICML 2024 (To appear).
[arXiv]
-
Efficient Low-Rank Matrix Estimation, Experimental Design, and Arm-Set-Dependent Low-Rank Bandits.
ICML 2024 (To appear).
[arXiv]
-
The Human-AI Substitution game: active learning from a strategic labeler.
ICLR 2024.
[OpenReview]
-
Efficient Active Learning Halfspaces with Tsybakov Noise: A Non-convex Optimization Approach.
AISTATS 2024.
[arXiv]
-
Kullback-Leibler Maillard Sampling for Multi-armed Bandits with Bounded Rewards.
NeurIPS 2023.
[arXiv]
-
Fair coexistence of heterogeneous networks: A novel probabilistic multi-armed bandit approach.
WiOpt 2023.
[link]
-
Hierarchical Unimodal Bandits.
ECML-PKDD 2022.
[link]
-
On Efficient Online Imitation Learning via Classification.
NeurIPS 2022.
[arXiv] [Talk Slides]
-
PopArt: Efficient Sparse Regression and Experimental Design for Optimal Sparse Linear Bandits.
NeurIPS 2022.
[arXiv] [Talk Slides]
-
Thompson Sampling for Robust Transfer in Multi-Task Bandits.
ICML 2022.
[arXiv] [Poster] [Related Talk Slides]
-
Margin-distancing for safe model explanation.
AISTATS 2022.
[arXiv]
-
Provably Efficient Multi-Task Reinforcement Learning with Model Transfer.
NeurIPS 2021.
[arXiv] [Poster] [slides]
-
Improved Algorithms for Efficient Active Learning Halfspaces with Massart and Tsybakov noise.
COLT 2021.
[arXiv] [slides] [poster]
-
Multitask Bandit Learning through Heterogeneous Feedback Aggregation.
AISTATS 2021.
[arXiv] [Spotlight talk (by Zhi)]
-
Active Online Learning with Hidden Shifting Domains.
AISTATS 2021.
[arXiv] [Spotlight talk (by Yining)]
-
Attribute-Efficient Learning of Halfspaces with Malicious Noise: Near-Optimal Label Complexity and Noise Tolerance.
ALT 2021.
[arXiv]
-
Crush Optimism with Pessimism: Structured Bandits Beyond Asymptotic Optimality.
NeurIPS 2020.
[arXiv]
-
Efficient Contextual Bandits with Continuous Actions.
NeurIPS 2020.
[arXiv] [poster] [Long talk]
-
Efficient Active Learning of Sparse Halfspaces with Arbitrary Bounded Noise.
NeurIPS 2020 (oral presentation).
[arXiv] [talk]
-
Deep Batch Active Learning by Diverse, Uncertain Gradient Lower Bounds.
ICLR 2020 (talk).
[arXiv] [code]
-
Contextual Bandits with Continuous Actions: Smoothing, Zooming, and Adapting.
COLT 2019.
[arXiv]
-
Warm-starting Contextual Bandits: Robustly Combining Supervised and Bandit Feedback.
ICML 2019.
[arXiv] [code]
-
Bandit Multiclass Linear Classification: Efficient Algorithms for the Separable Case.
ICML 2019.
[arXiv]
-
Efficient Active Learning of Sparse Halfspaces.
COLT 2018.
[arXiv]
-
Revisiting Perceptron: Efficient and Label-Optimal Learning of Halfspaces.
NeurIPS 2017.
[arXiv] [poster]
-
Efficient Online Bandit Multiclass Learning with \( \tilde O(\sqrt{T}) \) Regret.
ICML 2017.
[arXiv]
-
Search Improves Label for Active Learning.
NeurIPS 2016.
[arXiv]
-
The Extended Littlestone's Dimension for Learning with Mistakes and Abstentions.
COLT 2016.
[arXiv]
-
Active Learning from Weak and Strong Labelers.
NeurIPS 2015.
[arXiv]
-
Spectral Learning of Large Structured HMMs for Comparative Epigenomics.
NeurIPS 2015.
[arXiv] [code]
-
Beyond Disagreement-based Agnostic Active Learning.
NeurIPS 2014 (spotlight presentation).
[pdf] [arXiv]
Journals
-
Fair Probabilistic Multi-Armed Bandit with Applications to Network Optimization.
IEEE Transactions on Machine Learning in Communications and Networking (TMLCN) 2024.
[IEEE Xplore]
-
Contextual Bandits with Continuous Actions: Smoothing, Zooming, and Adapting.
JMLR 2019.
[JMLR]
Workshop Contributions
-
A Potential-based Framework for Online Learning with Mistakes and Abstentions.
NeurIPS 2016 workshop on Reliable Machine Learning in the Wild.
[pdf]
-
Search Improves Label for Active Learning.
ICML Workshop on Data Efficient Machine Learning 2016.
[pdf]
-
Active Learning from Weak and Strong Labelers.
ICML Active Learning Workshop 2015.
[pdf]
-
Improved Algorithms for Confidence-Rated Prediction with Error Guarantees.
NeurIPS 2013 Workshop on Learning Faster from Easy Data.
[pdf]
Thesis
-
Active Learning and Confidence-rated Prediction.
PhD Thesis, UCSD, 2017.
[Local version (with typos fixed)] [UC eScholarship]