Elad Hazan

Summary

Elad Hazan is an Israeli-American computer scientist, academic, author and researcher. He is a Professor of Computer Science at Princeton University, and the co-founder and director of Google AI Princeton.[1][2]

Elad Hazan
NationalityIsraeli-American
Occupation(s)Computer scientist, academic, author and researcher
AwardsBell Labs Prize
Marie Curie Fellow, European Research Council
Google Research Award
Amazon Research Award
Academic background
EducationB.Sc., Computer Science, Tel Aviv University
M.Sc., Computer Science, Tel Aviv University
Ph.D., Computer Science, Princeton University
Academic work
InstitutionsPrinceton University
Websitehttps://www.ehazan.com/

Hazan co-invented adaptive gradient methods and the AdaGrad algorithm. He has published over 150 articles and has several patents awarded. He has worked machine learning and mathematical optimization, and more recently on control theory and reinforcement learning.[3] He has authored a book, entitled Introduction to Online Convex Optimization. Hazan is the co-founder of In8 Inc., which was acquired by Google in 2018.[4]

Education edit

Hazan studied at Tel Aviv University and received his bachelor's and master's degrees in Computer Science in 2001 and 2002, respectively. He then moved to the United States, earning his Doctoral Degree in Computer Science from Princeton University in 2006 under Sanjeev Arora.[1]

Career edit

Upon receiving his doctoral degree, Hazan held an appointment as a Research Staff Member in the Theory Group at IBM Almaden Research Center in 2006. Following this appointment, he joined Technion - Israel Institute of Technology as an assistant professor in 2010 and was tenured and promoted to Associate Professor in 2013.[1] In 2015, he joined Princeton University as an Assistant Professor of Computer Science, and later became Professor of Computer Science in 2016. Since 2018, he has been serving as a Director of Google AI Princeton.[5]

Research edit

Hazan's research primarily focuses on machine learning, mathematical optimization, control theory and reinforcement learning. He is the co-inventor of five US patents.

Machine learning and mathematical optimization edit

Hazan co-introduced adaptive subgradient methods to dynamically incorporate knowledge of the geometry of the data observed in earlier iterations, and to perform more informative gradient-based learning. The AdaGrad algorithm changed optimization for deep learning and serves as the basis for today's fastest algorithms. In his study, he also made substantial contributions to the theory of online convex optimization, including the Online Newton Step and Online Frank Wolfe algorithm, projection free methods, and adaptive-regret algorithms.[6]

In the area of mathematical optimization, Hazan proposed the first sublinear-time algorithms for linear classification as well as semi-definite programming. He also gave the first linearly converging Frank-Wolfe-type algorithm.[7]

More recently, Hazan and his group proposed a new paradigm for differentiable reinforcement learning called non-stochastic control, which applies online convex optimization to control.[8]

Awards and honors edit

  • 2002–2006 - Gordon Wu fellowship, Princeton University
  • 2008 - Machine Learning Journal Award best student paper, COLT
  • 2009, 2012 - IBM Pat Goldberg Memorial Best Paper Award[9]
  • 2011–2015 - European Research Council Marie Curie Fellow
  • 2011, 2015 - Google Research Award[10][11]
  • 2012 - ICML Best Student Paper Runner Up
  • 2013 - European Research Council (ERC) Starting Grant Recipient[12]
  • 2017 - Bell Labs Prize[13]
  • 2018 - Amazon Research Award[14]

Bibliography edit

Books edit

  • Introduction to Online Convex Optimization (2016) ISBN 9781521003442

Selected articles edit

  • Hazan, E., Agarwal, A., & Kale, S. (2007). Logarithmic regret algorithms for online convex optimization. Machine Learning, 69(2-3), 169–192.
  • Duchi, J., Hazan, E., & Singer, Y. (2011). Adaptive subgradient methods for online learning and stochastic optimization. Journal of machine learning research, 12(7).
  • Arora, S., Hazan, E., & Kale, S. (2012). The multiplicative weights update method: a meta-algorithm and applications. Theory of Computing, 8(1), 121–164.
  • Hazan, E. (2019). Introduction to online convex optimization. arXiv preprint arXiv:1909.05207.
  • Clarkson, K. L., Hazan, E., & Woodruff, D. P. (2012). Sublinear optimization for machine learning. Journal of the ACM (JACM), 59(5), 1-49.
  • Garber, D., & Hazan, E. (2013). A linearly convergent conditional gradient algorithm with applications to online and stochastic optimization. arXiv preprint arXiv:1301.4666.
  • Agarwal, N., Bullins, B., Hazan, E., Kakade, S., & Singh, K. (2019, May). Online control with adversarial disturbances. In International Conference on Machine Learning (pp. 111–119). PMLR.

References edit

  1. ^ a b c "Elad Hazan".
  2. ^ NJ.com, Rob Jennings | NJ Advance Media for (2018-12-19). "Self-driving cars, speech recognition. See Google's new AI lab at Princeton". nj. Retrieved 2021-07-31.
  3. ^ "Elad Hazan - Google Scholar".
  4. ^ "Teaching Machines to Learn by Themselves".
  5. ^ "Google AI lab formally opens in downtown Princeton, bolstering innovation and invention".
  6. ^ "Adaptive Subgradient Methods for Online Learning and Stochastic Optimization" (PDF).
  7. ^ Allen-Zhu, Zeyuan; Hazan, Elad; Hu, Wei; Li, Yuanzhi (2017). "Linear Convergence of a Frank-Wolfe Type Algorithm over Trace-Norm Balls". arXiv:1708.02105 [cs.LG].
  8. ^ Hazan, Elad (2019). "Introduction to Online Convex Optimization". arXiv:1909.05207 [cs.LG].
  9. ^ "Pat Goldberg Memorial Best Paper Awards". 25 July 2016.
  10. ^ "Google Research Awards – December 2011" (PDF).
  11. ^ "Google Faculty Research Awards ¬ February 2015" (PDF).
  12. ^ "ERC Starting Grants 2013 Results" (PDF).
  13. ^ "Nokia announces its 2016 Bell Labs Prize winners".
  14. ^ "Amazon Research Award - Recipients".