Algorithms for Convex Optimization - Nisheeth K. Vishnoi

Algorithms for Convex Optimization

Buch | Hardcover
200 Seiten
2021
Cambridge University Press (Verlag)
978-1-108-48202-8 (ISBN)
99,75 inkl. MwSt
Algorithms for Convex Optimization are the workhorses of data-driven, technological advancements in machine learning and artificial intelligence. This concise, modern guide to deriving these algorithms is self-contained and accessible to advanced students, practitioners, and researchers in computer science, operations research, and data science.
In the last few years, Algorithms for Convex Optimization have revolutionized algorithm design, both for discrete and continuous optimization problems. For problems like maximum flow, maximum matching, and submodular function minimization, the fastest algorithms involve essential methods such as gradient descent, mirror descent, interior point methods, and ellipsoid methods. The goal of this self-contained book is to enable researchers and professionals in computer science, data science, and machine learning to gain an in-depth understanding of these algorithms. The text emphasizes how to derive key algorithms for convex optimization from first principles and how to establish precise running time bounds. This modern text explains the success of these algorithms in problems of discrete optimization, as well as how these methods have significantly pushed the state of the art of convex optimization itself.

Nisheeth K. Vishnoi is a Professor of Computer Science at Yale University. His research areas include theoretical computer science, optimization, and machine learning. He is a recipient of the Best Paper Award at IEEE FOCS in 2005, the IBM Research Pat Goldberg Memorial Award in 2006, the Indian National Science Academy Young Scientist Award in 2011, and the Best Paper award at ACM FAccT in 2019. He was elected an ACM Fellow in 2019. He obtained a bachelor degree in Computer Science and Engineering from IIT Bombay and a Ph.D. in Algorithms, Combinatorics and Optimization from Georgia Institute of Technology.

1. Bridging continuous and discrete optimization; 2. Preliminaries; 3. Convexity; 4. Convex optimization and efficiency; 5. Duality and optimality; 6. Gradient descent; 7. Mirror descent and multiplicative weights update; 8. Accelerated gradient descent; 9. Newton's method; 10. An interior point method for linear programming; 11. Variants of the interior point method and self-concordance; 12. Ellipsoid method for linear programming; 13. Ellipsoid method for convex optimization.

Erscheinungsdatum
Zusatzinfo Worked examples or Exercises
Verlagsort Cambridge
Sprache englisch
Maße 156 x 235 mm
Gewicht 660 g
Themenwelt Informatik Theorie / Studium Algorithmen
Mathematik / Informatik Mathematik Angewandte Mathematik
Mathematik / Informatik Mathematik Finanz- / Wirtschaftsmathematik
ISBN-10 1-108-48202-3 / 1108482023
ISBN-13 978-1-108-48202-8 / 9781108482028
Zustand Neuware
Haben Sie eine Frage zum Produkt?
Mehr entdecken
aus dem Bereich