research interests

I’m broadly interested in statistics, machine learning, and optimization. I’m also interested in lots of applications: to finance, operations research, public policy, social good, sustainability, epidemiology, healthcare, autonomous vehicles, analytics, …. A lot of my recent work has focused on building reliable and trustworthy machine learning systems, by taking a close look at the (many) statistical and computational issues that arise after a machine learning model has been deployed into real-world systems and scientific applications.

On a more technical level, I’m interested in – and have recently worked on projects related to – the following areas:

  • distribution shift, robust optimization, subpopulation-level performance, responsible AI
  • conformal inference, distribution-free uncertainty quantification
  • weak supervision
  • tuning parameter-free stochastic optimization
  • implicit regularization
  • sparse regression
  • large-scale multiple testing
  • risk estimation, model selection

representative publications

Here are a few (representative) publications related to the above topics:



all publications

  1. The Lifecycle of a Statistical Model: Model Failure Detection, Identification, and Refitting
    Alnur Ali, Maxime Cauchois, and John Duchi
    Journal of Machine Learning Research (under review), 2023
    #subpopulation-level performance #responsible AI #conformal inference #weak supervision #large-scale multiple testing
  2. Predictive Inference with Weak Supervision
    Maxime Cauchois, Suyash Gupta, Alnur Ali, and John Duchi
    Journal of Machine Learning Research (under review), 2022
    #conformal inference #weak supervision #responsible AI
  3. A Comment and Erratum on "Excess Optimism: How Biased is the Apparent Error of an Estimator Tuned by SURE?"
    Maxime Cauchois, Alnur Ali, and John Duchi
    Journal of the American Statistical Association (under review), 2022
    #risk estimation #model selection
  4. Minimum-Distortion Embedding
    Akshay Agrawal, Alnur Ali, and Stephen Boyd
    Foundations and Trends in Machine Learning, 2021
    #representation learning #convex optimization
  5. Minimizing Oracle-Structured Composite Functions
    Xinyue Shen, Alnur Ali, and Stephen Boyd
    Optimization and Engineering, 2021
    #stochastic optimization
  6. Accelerated Gradient Flow: Risk, Stability, and Implicit Regularization
    Yue Sheng and Alnur Ali
    2021
    #implicit regularization #convex optimization
  7. Computationally Efficient Posterior Inference With Langevin Monte Carlo and Early Stopping
    Dushyant Sahoo, Alnur Ali, and Edgar Dobriban
    2021
    #implicit regularization #Markov chain Monte Carlo
  8. Robust Validation: Confident Predictions Even When Distributions Shift
    Maxime Cauchois, Suyash Gupta, Alnur Ali, and John Duchi
    Journal of the American Statistical Association, 2021
    #distribution shift #robust optimization #conformal inference #responsible AI
  9. The Implicit Regularization of Stochastic Gradient Flow for Least Squares
    International Conference on Machine Learning (ICML), 2020
    #implicit regularization #stochastic optimization
  10. Confidence Bands for a Log-Concave Density
    Guenther Walther, Alnur Ali, Xinyue Shen, and Stephen Boyd
    Journal of Computational and Graphical Statistics, 2020
    #uncertainty quantification #convex optimization
  11. A Continuous-Time View of Early Stopping for Least Squares
    Alnur Ali, Zico Kolter, and Ryan Tibshirani
    International Conference on Artificial Intelligence and Statistics (AISTATS), 2019
    #implicit regularization #convex optimization
  12. The Generalized Lasso Problem and Uniqueness
    Alnur Ali and Ryan Tibshirani
    Electronic Journal of Statistics, 2019
    #sparse regression #convex optimization
  13. Communication-Avoiding Optimization Methods for Distributed Massive-Scale Sparse Inverse Covariance Estimation
    International Conference on Artificial Intelligence and Statistics (AISTATS), 2018
    #sparse inverse covariance estimation #convex optimization #sparse regression
  14. A Semismooth Newton Method for Fast, Generic Convex Programming
    Alnur Ali, Eric Wong, and Zico Kolter
    International Conference on Machine Learning (ICML), 2017
    #convex optimization
  15. Generalized Pseudolikelihood Methods for Inverse Covariance Estimation
    International Conference on Artificial Intelligence and Statistics (AISTATS), 2017
    #sparse inverse covariance estimation #convex optimization #sparse regression
  16. The Multiple Quantile Graphical Model
    Alnur Ali, Zico Kolter, and Ryan Tibshirani
    Advances in Neural Information Processing Systems 29 (NeurIPS), 2016
    #sparse inverse covariance estimation #convex optimization #sparse regression
  17. Disciplined Convex Stochastic Programming: A New Framework for Stochastic Optimization
    Proceedings of the 31st Conference on Uncertainty in Artificial Intelligence (UAI), 2015
    #stochastic optimization
  18. Active Learning With Model Selection
    Alnur Ali, Rich Caruana, and Ashish Kapoor
    AAAI Conference on Artificial Intelligence (AAAI), 2014
    #model selection
  19. Experiments With Kemeny Ranking: What Works When?
    Alnur Ali and Marina Meila
    Mathematical Social Sciences, 2012
    #ranking
  20. Learning Lexicon Models from Search Logs for Query Expansion
    Jianfeng Gao, Xiaodong He, Shasha Xie, and Alnur Ali
    Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning (EMNLP-CoNLL), 2012
    #ranking
  21. Preferences in College Applications: A Nonparametric Bayesian Analysis of Top-10 Rankings
    Neural Information Processing Systems (NeurIPS) Workshop on Computational Social Science, 2010
    #ranking #Markov chain Monte Carlo