An Introduction to Optimization
Home > Mathematics and Science Textbooks > Mathematics > An Introduction to Optimization: With Applications to Machine Learning
An Introduction to Optimization: With Applications to Machine Learning

An Introduction to Optimization: With Applications to Machine Learning

|
     0     
5
4
3
2
1




International Edition


About the Book

An Introduction to Optimization Accessible introductory textbook on optimization theory and methods, with an emphasis on engineering design, featuring MATLAB® exercises and worked examples Fully updated to reflect modern developments in the field, the Fifth Edition of An Introduction to Optimization fills the need for an accessible, yet rigorous, introduction to optimization theory and methods, featuring innovative coverage and a straightforward approach. The book begins with a review of basic definitions and notations while also providing the related fundamental background of linear algebra, geometry, and calculus. With this foundation, the authors explore the essential topics of unconstrained optimization problems, linear programming problems, and nonlinear constrained optimization. In addition, the book includes an introduction to artificial neural networks, convex optimization, multi-objective optimization, and applications of optimization in machine learning. Numerous diagrams and figures found throughout the book complement the written presentation of key concepts, and each chapter is followed by MATLAB® exercises and practice problems that reinforce the discussed theory and algorithms. The Fifth Edition features a new chapter on Lagrangian (nonlinear) duality, expanded coverage on matrix games, projected gradient algorithms, machine learning, and numerous new exercises at the end of each chapter. An Introduction to Optimization includes information on: The mathematical definitions, notations, and relations from linear algebra, geometry, and calculus used in optimization Optimization algorithms, covering one-dimensional search, randomized search, and gradient, Newton, conjugate direction, and quasi-Newton methods Linear programming methods, covering the simplex algorithm, interior point methods, and duality Nonlinear constrained optimization, covering theory and algorithms, convex optimization, and Lagrangian duality Applications of optimization in machine learning, including neural network training, classification, stochastic gradient descent, linear regression, logistic regression, support vector machines, and clustering. An Introduction to Optimization is an ideal textbook for a one- or two-semester senior undergraduate or beginning graduate course in optimization theory and methods. The text is also of value for researchers and professionals in mathematics, operations research, electrical engineering, economics, statistics, and business.

Table of Contents:
Preface xv About the Companion Website xviii Part I Mathematical Review 1 1 Methods of Proof and Some Notation 3 1.1 Methods of Proof 3 1.2 Notation 5 Exercises 5 2 Vector Spaces and Matrices 7 2.1 Vector and Matrix 7 2.2 Rank of a Matrix 11 2.3 Linear Equations 16 2.4 Inner Products and Norms 18 Exercises 20 3 Transformations 23 3.1 Linear Transformations 23 3.2 Eigenvalues and Eigenvectors 24 3.3 Orthogonal Projections 26 3.4 Quadratic Forms 27 3.5 Matrix Norms 32 Exercises 35 4 Concepts from Geometry 39 4.1 Line Segments 39 4.2 Hyperplanes and Linear Varieties 39 4.3 Convex Sets 41 4.4 Neighborhoods 43 4.5 Polytopes and Polyhedra 44 Exercises 45 5 Elements of Calculus 47 5.1 Sequences and Limits 47 5.2 Differentiability 52 5.3 The Derivative Matrix 54 5.4 Differentiation Rules 57 5.5 Level Sets and Gradients 58 5.6 Taylor Series 61 Exercises 65 Part II Unconstrained Optimization 67 6 Basics of Set-Constrained and Unconstrained Optimization 69 6.1 Introduction 69 6.2 Conditions for Local Minimizers 70 Exercises 78 7 One-Dimensional Search Methods 87 7.1 Introduction 87 7.2 Golden Section Search 87 7.3 Fibonacci Method 91 7.4 Bisection Method 97 7.5 Newton’s Method 98 7.6 Secant Method 101 7.7 Bracketing 103 7.8 Line Search in Multidimensional Optimization 103 Exercises 105 8 Gradient Methods 109 8.1 Introduction 109 8.2 Steepest Descent Method 110 8.3 Analysis of Gradient Methods 117 Exercises 126 9 Newton’s Method 133 9.1 Introduction 133 9.2 Analysis of Newton’s Method 135 9.3 Levenberg–Marquardt Modification 138 9.4 Newton’s Method for Nonlinear Least Squares 139 Exercises 142 10 Conjugate Direction Methods 145 10.1 Introduction 145 10.2 Conjugate Direction Algorithm 146 10.2.1 Basic Conjugate Direction Algorithm 146 10.3 Conjugate Gradient Algorithm 151 10.4 Conjugate Gradient Algorithm for Nonquadratic Problems 154 Exercises 156 11 Quasi-Newton Methods 159 11.1 Introduction 159 11.2 Approximating the Inverse Hessian 160 11.3 Rank One Correction Formula 162 11.4 DFP Algorithm 166 11.5 BFGS Algorithm 170 Exercises 173 12 Solving Linear Equations 179 12.1 Least-Squares Analysis 179 12.2 Recursive Least-Squares Algorithm 187 12.3 Solution to a Linear Equation with Minimum Norm 190 12.4 Kaczmarz’s Algorithm 191 12.5 Solving Linear Equations in General 194 Exercises 201 13 Unconstrained Optimization and Neural Networks 209 13.1 Introduction 209 13.2 Single-Neuron Training 211 13.3 Backpropagation Algorithm 213 Exercises 222 14 Global Search Algorithms 225 14.1 Introduction 225 14.2 Nelder–Mead Simplex Algorithm 225 14.3 Simulated Annealing 229 14.3.1 Randomized Search 229 14.3.2 Simulated Annealing Algorithm 229 14.4 Particle Swarm Optimization 231 14.4.1 Basic PSO Algorithm 232 14.4.2 Variations 233 14.5 Genetic Algorithms 233 14.5.1 Basic Description 233 14.5.1.1 Chromosomes and Representation Schemes 234 14.5.1.2 Selection and Evolution 234 14.5.2 Analysis of Genetic Algorithms 238 14.5.3 Real-Number Genetic Algorithms 243 Exercises 244 Part III Linear Programming 247 15 Introduction to Linear Programming 249 15.1 Brief History of Linear Programming 249 15.2 Simple Examples of Linear Programs 250 15.3 Two-Dimensional Linear Programs 256 15.4 Convex Polyhedra and Linear Programming 258 15.5 Standard Form Linear Programs 260 15.6 Basic Solutions 264 15.7 Properties of Basic Solutions 267 15.8 Geometric View of Linear Programs 269 Exercises 273 16 Simplex Method 277 16.1 Solving Linear Equations Using Row Operations 277 16.2 The Canonical Augmented Matrix 283 16.3 Updating the Augmented Matrix 284 16.4 The Simplex Algorithm 285 16.5 Matrix Form of the Simplex Method 291 16.6 Two-Phase Simplex Method 294 16.7 Revised Simplex Method 297 Exercises 301 17 Duality 309 17.1 Dual Linear Programs 309 17.2 Properties of Dual Problems 316 17.3 Matrix Games 321 Exercises 324 18 Nonsimplex Methods 331 18.1 Introduction 331 18.2 Khachiyan’s Method 332 18.3 Affine Scaling Method 334 18.3.1 Basic Algorithm 334 18.3.2 Two-Phase Method 337 18.4 Karmarkar’s Method 339 18.4.1 Basic Ideas 339 18.4.2 Karmarkar’s Canonical Form 339 18.4.3 Karmarkar’s Restricted Problem 341 18.4.4 From General Form to Karmarkar’s Canonical Form 342 18.4.5 The Algorithm 345 Exercises 349 19 Integer Linear Programming 351 19.1 Introduction 351 19.2 Unimodular Matrices 351 19.3 The Gomory Cutting-Plane Method 358 Exercises 366 Part IV Nonlinear Constrained Optimization 369 20 Problems with Equality Constraints 371 20.1 Introduction 371 20.2 Problem Formulation 373 20.3 Tangent and Normal Spaces 374 20.4 Lagrange Condition 379 20.5 Second-Order Conditions 387 20.6 Minimizing Quadratics Subject to Linear Constraints 390 Exercises 394 21 Problems with Inequality Constraints 399 21.1 Karush–Kuhn–Tucker Condition 399 21.2 Second-Order Conditions 406 Exercises 410 22 Convex Optimization Problems 417 22.1 Introduction 417 22.2 Convex Functions 419 22.3 Convex Optimization Problems 426 22.4 Semidefinite Programming 431 22.4.1 Linear Matrix Inequalities and Their Properties 431 22.4.2 LMI Solvers 435 22.4.2.1 Finding a Feasible Solution Under LMI Constraints 436 22.4.2.2 Minimizing a Linear Objective Under LMI Constraints 438 22.4.2.3 Minimizing a Generalized Eigenvalue Under LMI Constraints 440 Exercises 442 23 Lagrangian Duality 449 23.1 Overview 449 23.2 Notation 449 23.3 Primal–Dual Pair 450 23.4 General Duality Properties 451 23.4.1 Convexity of Dual Problem 451 23.4.2 Primal Objective in Terms of Lagrangian 451 23.4.3 Minimax Inequality Chain 452 23.4.4 Optimality of Saddle Point 452 23.4.5 Weak Duality 453 23.4.6 Duality Gap 453 23.5 Strong Duality 454 23.5.1 Strong Duality ⇔ Minimax Equals Maximin 454 23.5.2 Strong Duality ⇒ Primal Unconstrained Minimization 455 23.5.3 Strong Duality ⇒ Optimality 455 23.5.4 Strong Duality ⇒ KKT (Including Complementary Slackness) 455 23.5.5 Strong Duality ⇒ Saddle Point 456 23.6 Convex Case 456 23.6.1 Convex Case: KKT ⇒ Strong Duality 456 23.6.2 Convex Case: Regular Optimal Primal ⇒ Strong Duality 457 23.6.3 Convex Case: Slater’s Condition ⇒ Strong Duality 457 23.7 Summary of Key Results 457 Exercises 458 24 Algorithms for Constrained Optimization 459 24.1 Introduction 459 24.2 Projections 459 24.3 Projected Gradient Methods with Linear Constraints 462 24.4 Convergence of Projected Gradient Algorithms 465 24.4.1 Fixed Points and First-Order Necessary Conditions 466 24.4.2 Convergence with Fixed Step Size 468 24.4.3 Some Properties of Projections 469 24.4.4 Armijo Condition 470 24.4.5 Accumulation Points 471 24.4.6 Projections in the Convex Case 472 24.4.7 Armijo Condition in the Convex Case 474 24.4.8 Convergence in the Convex Case 480 24.4.9 Convergence Rate with Line-Search Step Size 481 24.5 Lagrangian Algorithms 483 24.5.1 Lagrangian Algorithm for Equality Constraints 484 24.5.2 Lagrangian Algorithm for Inequality Constraints 486 24.6 Penalty Methods 489 Exercises 495 25 Multiobjective Optimization 499 25.1 Introduction 499 25.2 Pareto Solutions 499 25.3 Computing the Pareto Front 501 25.4 From Multiobjective to Single-Objective Optimization 505 25.5 Uncertain Linear Programming Problems 508 25.5.1 Uncertain Constraints 508 25.5.2 Uncertain Objective Function Coefficients 511 25.5.3 Uncertain Constraint Coefficients 513 25.5.4 General Uncertainties 513 Exercises 513 Part V Optimization in Machine Learning 517 26 Machine Learning Problems and Feature Engineering 519 26.1 Machine Learning Problems 519 26.1.1 Data with Labels and Supervised Learning 519 26.1.2 Data Without Labels and Unsupervised Learning 521 26.2 Data Normalization 522 26.3 Histogram of Oriented Gradients 524 26.4 Principal Component Analysis and Linear Autoencoder 526 26.4.1 Singular Value Decomposition 526 26.4.2 Principal Axes and Principal Components of a Data Set 527 26.4.3 Linear Autoencoder 529 Exercises 530 27 Stochastic Gradient Descent Algorithms 537 27.1 Stochastic Gradient Descent Algorithm 537 27.2 Stochastic Variance Reduced Gradient Algorithm 540 27.3 Distributed Stochastic Variance Reduced Gradient 542 27.3.1 Distributed Learning Environment 542 27.3.2 SVRG in Distributed Optimization 543 27.3.3 Communication Versus Computation 545 27.3.4 Data Security 545 Exercises 546 28 Linear Regression and Its Variants 553 28.1 Least-Squares Linear Regression 553 28.1.1 A Linear Model for Prediction 553 28.1.2 Training the Model 554 28.1.3 Computing Optimal ̂w 554 28.1.4 Optimal Predictor and Performance Evaluation 555 28.1.5 Least-Squares Linear Regression for Data Sets with Vector Labels 556 28.2 Model Selection by Cross-Validation 559 28.3 Model Selection by Regularization 562 Exercises 564 29 Logistic Regression for Classification 569 29.1 Logistic Regression for Binary Classification 569 29.1.1 Least-Squares Linear Regression for Binary Classification 569 29.1.2 Logistic Regression for Binary Classification 570 29.1.3 Interpreting Logistic Regression by Log Error 572 29.1.4 Confusion Matrix for Binary Classification 573 29.2 Nonlinear Decision Boundary via Linear Regression 575 29.2.1 Least-Squares Linear Regression with Nonlinear Transformation 576 29.2.2 Logistic Regression with Nonlinear Transformation 578 29.3 Multicategory Classification 580 29.3.1 One-Versus-All Multicategory Classification 580 29.3.2 Softmax Regression for Multicategory Classification 581 Exercises 584 30 Support Vector Machines 589 30.1 Hinge-Loss Functions 589 30.1.1 Geometric Interpretation of the Linear Model 589 30.1.2 Hinge Loss for Binary Data Sets 590 30.1.3 Hinge Loss for Multicategory Data Sets 592 30.2 Classification by Minimizing Hinge Loss 593 30.2.1 Binary Classification by Minimizing Average Hinge Loss 593 30.2.2 Multicategory Classification by Minimizing E hww or E hcs 594 30.3 Support Vector Machines for Binary Classification 596 30.3.1 Hard-Margin Support Vector Machines 596 30.3.2 Support Vectors 598 30.3.3 Soft-Margin Support Vector Machines 599 30.3.4 Connection to Hinge-Loss Minimization 602 30.4 Support Vector Machines for Multicategory Classification 602 30.5 Kernel Trick 603 30.5.1 Kernels 603 30.5.2 Kernel Trick 604 30.5.3 Learning with Kernels 605 30.5.3.1 Regularized Logistic Regression with Nonlinear Transformation for Binary Classification 605 30.5.3.2 Regularized Hinge-Loss Minimization for Binary Classification 606 Exercises 607 31 K-Means Clustering 611 31.1 K-Means Clustering 611 31.2 K-Means++ forCenterInitialization 615 31.3 Variants of K-Means Clustering 617 31.3.1 K-Means Clustering Based on 1-Norm Regularization 617 31.3.2 PCA-Guided K-Means Clustering 619 31.4 Image Compression by Vector Quantization and K-Means Clustering 622 Exercises 623 References 627 Index 635


Best Sellers


Product Details
  • ISBN-13: 9781119877639
  • Publisher: John Wiley & Sons Inc
  • Publisher Imprint: John Wiley & Sons Inc
  • Height: 259 mm
  • No of Pages: 672
  • Returnable: N
  • Sub Title: With Applications to Machine Learning
  • Width: 185 mm
  • ISBN-10: 1119877636
  • Publisher Date: 18 Sep 2023
  • Binding: Hardback
  • Language: English
  • Returnable: N
  • Spine Width: 41 mm
  • Weight: 1412 gr


Similar Products

Add Photo
Add Photo

Customer Reviews

REVIEWS      0     
Click Here To Be The First to Review this Product
An Introduction to Optimization: With Applications to Machine Learning
John Wiley & Sons Inc -
An Introduction to Optimization: With Applications to Machine Learning
Writing guidlines
We want to publish your review, so please:
  • keep your review on the product. Review's that defame author's character will be rejected.
  • Keep your review focused on the product.
  • Avoid writing about customer service. contact us instead if you have issue requiring immediate attention.
  • Refrain from mentioning competitors or the specific price you paid for the product.
  • Do not include any personally identifiable information, such as full names.

An Introduction to Optimization: With Applications to Machine Learning

Required fields are marked with *

Review Title*
Review
    Add Photo Add up to 6 photos
    Would you recommend this product to a friend?
    Tag this Book Read more
    Does your review contain spoilers?
    What type of reader best describes you?
    I agree to the terms & conditions
    You may receive emails regarding this submission. Any emails will include the ability to opt-out of future communications.

    CUSTOMER RATINGS AND REVIEWS AND QUESTIONS AND ANSWERS TERMS OF USE

    These Terms of Use govern your conduct associated with the Customer Ratings and Reviews and/or Questions and Answers service offered by Bookswagon (the "CRR Service").


    By submitting any content to Bookswagon, you guarantee that:
    • You are the sole author and owner of the intellectual property rights in the content;
    • All "moral rights" that you may have in such content have been voluntarily waived by you;
    • All content that you post is accurate;
    • You are at least 13 years old;
    • Use of the content you supply does not violate these Terms of Use and will not cause injury to any person or entity.
    You further agree that you may not submit any content:
    • That is known by you to be false, inaccurate or misleading;
    • That infringes any third party's copyright, patent, trademark, trade secret or other proprietary rights or rights of publicity or privacy;
    • That violates any law, statute, ordinance or regulation (including, but not limited to, those governing, consumer protection, unfair competition, anti-discrimination or false advertising);
    • That is, or may reasonably be considered to be, defamatory, libelous, hateful, racially or religiously biased or offensive, unlawfully threatening or unlawfully harassing to any individual, partnership or corporation;
    • For which you were compensated or granted any consideration by any unapproved third party;
    • That includes any information that references other websites, addresses, email addresses, contact information or phone numbers;
    • That contains any computer viruses, worms or other potentially damaging computer programs or files.
    You agree to indemnify and hold Bookswagon (and its officers, directors, agents, subsidiaries, joint ventures, employees and third-party service providers, including but not limited to Bazaarvoice, Inc.), harmless from all claims, demands, and damages (actual and consequential) of every kind and nature, known and unknown including reasonable attorneys' fees, arising out of a breach of your representations and warranties set forth above, or your violation of any law or the rights of a third party.


    For any content that you submit, you grant Bookswagon a perpetual, irrevocable, royalty-free, transferable right and license to use, copy, modify, delete in its entirety, adapt, publish, translate, create derivative works from and/or sell, transfer, and/or distribute such content and/or incorporate such content into any form, medium or technology throughout the world without compensation to you. Additionally,  Bookswagon may transfer or share any personal information that you submit with its third-party service providers, including but not limited to Bazaarvoice, Inc. in accordance with  Privacy Policy


    All content that you submit may be used at Bookswagon's sole discretion. Bookswagon reserves the right to change, condense, withhold publication, remove or delete any content on Bookswagon's website that Bookswagon deems, in its sole discretion, to violate the content guidelines or any other provision of these Terms of Use.  Bookswagon does not guarantee that you will have any recourse through Bookswagon to edit or delete any content you have submitted. Ratings and written comments are generally posted within two to four business days. However, Bookswagon reserves the right to remove or to refuse to post any submission to the extent authorized by law. You acknowledge that you, not Bookswagon, are responsible for the contents of your submission. None of the content that you submit shall be subject to any obligation of confidence on the part of Bookswagon, its agents, subsidiaries, affiliates, partners or third party service providers (including but not limited to Bazaarvoice, Inc.)and their respective directors, officers and employees.

    Accept

    New Arrivals

    Inspired by your browsing history


    Your review has been submitted!

    You've already reviewed this product!