Stochastic Optimization for Large-scale Machine Learning
Home > Computing and Information Technology > Computer programming / software engineering > Algorithms and data structures > Stochastic Optimization for Large-scale Machine Learning
Stochastic Optimization for Large-scale Machine Learning

Stochastic Optimization for Large-scale Machine Learning

|
     0     
5
4
3
2
1




International Edition


About the Book

Advancements in the technology and availability of data sources have led to the `Big Data' era. Working with large data offers the potential to uncover more fine-grained patterns and take timely and accurate decisions, but it also creates a lot of challenges such as slow training and scalability of machine learning models. One of the major challenges in machine learning is to develop efficient and scalable learning algorithms, i.e., optimization techniques to solve large scale learning problems. Stochastic Optimization for Large-scale Machine Learning identifies different areas of improvement and recent research directions to tackle the challenge. Developed optimisation techniques are also explored to improve machine learning algorithms based on data access and on first and second order optimisation methods. Key Features: Bridges machine learning and Optimisation. Bridges theory and practice in machine learning. Identifies key research areas and recent research directions to solve large-scale machine learning problems. Develops optimisation techniques to improve machine learning algorithms for big data problems. The book will be a valuable reference to practitioners and researchers as well as students in the field of machine learning.

Table of Contents:
List of Figures List of Tables Preface Section I BACKGROUND Introduction 1.1 LARGE-SCALE MACHINE LEARNING 1.2 OPTIMIZATION PROBLEMS 1.3 LINEAR CLASSIFICATION 1.3.1 Support Vector Machine (SVM) 1.3.2 Logistic Regression 1.3.3 First and Second Order Methods 1.3.3.1 First Order Methods 1.3.3.2 Second Order Methods 1.4 STOCHASTIC APPROXIMATION APPROACH 1.5 COORDINATE DESCENT APPROACH 1.6 DATASETS 1.7 ORGANIZATION OF BOOK Optimisation Problem, Solvers, Challenges and Research Directions 2.1 INTRODUCTION 2.1.1 Contributions 2.2 LITERATURE 2.3 PROBLEM FORMULATIONS 2.3.1 Hard Margin SVM (1992) 2.3.2 Soft Margin SVM (1995) 2.3.3 One-versus-Rest (1998) 2.3.4 One-versus-One (1999) 2.3.5 Least Squares SVM (1999) 2.3.6 v-SVM (2000) 2.3.7 Smooth SVM (2001) 2.3.8 Proximal SVM (2001) 2.3.9 Crammer Singer SVM (2002) 2.3.10 Ev-SVM (2003) 2.3.11 Twin SVM (2007) 2.3.12 Capped lp-norm SVM (2017) 2.4 PROBLEM SOLVERS 2.4.1 Exact Line Search Method 2.4.2 Backtracking Line Search 2.4.3 Constant Step Size 2.4.4 Lipschitz & Strong Convexity Constants 2.4.5 Trust Region Method 2.4.6 Gradient Descent Method 2.4.7 Newton Method 2.4.8 Gauss-Newton Method 2.4.9 Levenberg-Marquardt Method 2.4.10 Quasi-Newton Method 2.4.11 Subgradient Method 2.4.12 Conjugate Gradient Method 2.4.13 Truncated Newton Method 2.4.14 Proximal Gradient Method 2.4.15 Recent Algorithms 2.5 COMPARATIVE STUDY 2.5.1 Results from Literature 2.5.2 Results from Experimental Study 2.5.2.1 Experimental Setup and Implementation Details 2.5.2.2 Results and Discussions 2.6 CURRENT CHALLENGES AND RESEARCH DIRECTIONS 2.6.1 Big Data Challenge 2.6.2 Areas of Improvement 2.6.2.1 Problem Formulations 2.6.2.2 Problem Solvers 2.6.2.3 Problem Solving Strategies/Approaches 2.6.2.4 Platforms/Frameworks 2.6.3 Research Directions 2.6.3.1 Stochastic Approximation Algorithms 2.6.3.2 Coordinate Descent Algorithms 2.6.3.3 Proximal Algorithms 2.6.3.4 Parallel/Distributed Algorithms 2.6.3.5 Hybrid Algorithms 2.7 CONCLUSION Section II FIRST ORDER METHODS Mini-batch and Block-coordinate Approach 3.1 INTRODUCTION 3.1.1 Motivation 3.1.2 Batch Block Optimization Framework (BBOF) 3.1.3 Brief Literature Review 3.1.4 Contributions 3.2 STOCHASTIC AVERAGE ADJUSTED GRADIENT (SAAG) METHODS 3.3 ANALYSIS 3.4 NUMERICAL EXPERIMENTS 3.4.1 Experimental setup 3.4.2 Convergence against epochs 3.4.3 Convergence against Time 3.5 CONCLUSION AND FUTURE SCOPE Variance Reduction Methods 4.1 INTRODUCTION 4.1.1 Optimization Problem 4.1.2 Solution Techniques for Optimization Problem 4.1.3 Contributions 4.2 NOTATIONS AND RELATED WORK 4.2.1 Notations 4.2.2 Related Work 4.3 SAAG-I, II AND PROXIMAL EXTENSIONS 4.4 SAAG-III AND IV ALGORITHMS 4.5 ANALYSIS 4.6 EXPERIMENTAL RESULTS 4.6.1 Experimental Setup 4.6.2 Results with Smooth Problem 4.6.3 Results with non-smooth Problem 4.6.4 Mini-batch Block-coordinate versus mini-batch setting 4.6.5 Results with SVM 4.7 CONCLUSION Learning and Data Access 5.1 INTRODUCTION 5.1.1 Optimization Problem 5.1.2 Literature Review 5.1.3 Contributions 5.2 SYSTEMATIC SAMPLING 5.2.1 Definitions 5.2.2 Learning using Systematic Sampling 5.3 ANALYSIS 5.4 EXPERIMENTS 5.4.1 Experimental Setup 5.4.2 Implementation Details 5.4.3 Results 5.5 CONCLUSION Section III SECOND ORDER METHODS Mini-batch Block-coordinate Newton Method 6.1 INTRODUCTION 6.1.1 Contributions 6.2 MBN 6.3 EXPERIMENTS 6.3.1 Experimental Setup 6.3.2 Comparative Study 6.4 CONCLUSION Stochastic Trust Region Inexact Newton Method 7.1 INTRODUCTION 7.1.1 Optimization Problem 7.1.2 Solution Techniques 7.1.3 Contributions 7.2 LITERATURE REVIEW 7.3 TRUST REGION INEXACT NEWTON METHOD 7.3.1 Inexact Newton Method 7.3.2 Trust Region Inexact Newton Method 7.4 STRON 7.4.1 Complexity 7.4.2 Analysis 7.5 EXPERIMENTAL RESULTS 7.5.1 Experimental Setup 7.5.2 Comparative Study 7.5.3 Results with SVM 7.6 EXTENSIONS 7.6.1 PCG Subproblem Solver 1 7.6.2 Stochastic Variance Reduced Trust Region Inexact Newton Method 7.7 CONCLUSION Section IV CONCLUSION Conclusion and Future Scope 8.1 FUTURE SCOPE 142 Bibliography Index


Best Sellers


Product Details
  • ISBN-13: 9781032131757
  • Publisher: Taylor & Francis Ltd
  • Publisher Imprint: CRC Press
  • Height: 254 mm
  • No of Pages: 158
  • Width: 178 mm
  • ISBN-10: 1032131756
  • Publisher Date: 19 Nov 2021
  • Binding: Hardback
  • Language: English
  • Weight: 582 gr


Similar Products

Add Photo
Add Photo

Customer Reviews

REVIEWS      0     
Click Here To Be The First to Review this Product
Stochastic Optimization for Large-scale Machine Learning
Taylor & Francis Ltd -
Stochastic Optimization for Large-scale Machine Learning
Writing guidlines
We want to publish your review, so please:
  • keep your review on the product. Review's that defame author's character will be rejected.
  • Keep your review focused on the product.
  • Avoid writing about customer service. contact us instead if you have issue requiring immediate attention.
  • Refrain from mentioning competitors or the specific price you paid for the product.
  • Do not include any personally identifiable information, such as full names.

Stochastic Optimization for Large-scale Machine Learning

Required fields are marked with *

Review Title*
Review
    Add Photo Add up to 6 photos
    Would you recommend this product to a friend?
    Tag this Book Read more
    Does your review contain spoilers?
    What type of reader best describes you?
    I agree to the terms & conditions
    You may receive emails regarding this submission. Any emails will include the ability to opt-out of future communications.

    CUSTOMER RATINGS AND REVIEWS AND QUESTIONS AND ANSWERS TERMS OF USE

    These Terms of Use govern your conduct associated with the Customer Ratings and Reviews and/or Questions and Answers service offered by Bookswagon (the "CRR Service").


    By submitting any content to Bookswagon, you guarantee that:
    • You are the sole author and owner of the intellectual property rights in the content;
    • All "moral rights" that you may have in such content have been voluntarily waived by you;
    • All content that you post is accurate;
    • You are at least 13 years old;
    • Use of the content you supply does not violate these Terms of Use and will not cause injury to any person or entity.
    You further agree that you may not submit any content:
    • That is known by you to be false, inaccurate or misleading;
    • That infringes any third party's copyright, patent, trademark, trade secret or other proprietary rights or rights of publicity or privacy;
    • That violates any law, statute, ordinance or regulation (including, but not limited to, those governing, consumer protection, unfair competition, anti-discrimination or false advertising);
    • That is, or may reasonably be considered to be, defamatory, libelous, hateful, racially or religiously biased or offensive, unlawfully threatening or unlawfully harassing to any individual, partnership or corporation;
    • For which you were compensated or granted any consideration by any unapproved third party;
    • That includes any information that references other websites, addresses, email addresses, contact information or phone numbers;
    • That contains any computer viruses, worms or other potentially damaging computer programs or files.
    You agree to indemnify and hold Bookswagon (and its officers, directors, agents, subsidiaries, joint ventures, employees and third-party service providers, including but not limited to Bazaarvoice, Inc.), harmless from all claims, demands, and damages (actual and consequential) of every kind and nature, known and unknown including reasonable attorneys' fees, arising out of a breach of your representations and warranties set forth above, or your violation of any law or the rights of a third party.


    For any content that you submit, you grant Bookswagon a perpetual, irrevocable, royalty-free, transferable right and license to use, copy, modify, delete in its entirety, adapt, publish, translate, create derivative works from and/or sell, transfer, and/or distribute such content and/or incorporate such content into any form, medium or technology throughout the world without compensation to you. Additionally,  Bookswagon may transfer or share any personal information that you submit with its third-party service providers, including but not limited to Bazaarvoice, Inc. in accordance with  Privacy Policy


    All content that you submit may be used at Bookswagon's sole discretion. Bookswagon reserves the right to change, condense, withhold publication, remove or delete any content on Bookswagon's website that Bookswagon deems, in its sole discretion, to violate the content guidelines or any other provision of these Terms of Use.  Bookswagon does not guarantee that you will have any recourse through Bookswagon to edit or delete any content you have submitted. Ratings and written comments are generally posted within two to four business days. However, Bookswagon reserves the right to remove or to refuse to post any submission to the extent authorized by law. You acknowledge that you, not Bookswagon, are responsible for the contents of your submission. None of the content that you submit shall be subject to any obligation of confidence on the part of Bookswagon, its agents, subsidiaries, affiliates, partners or third party service providers (including but not limited to Bazaarvoice, Inc.)and their respective directors, officers and employees.

    Accept

    New Arrivals

    Inspired by your browsing history


    Your review has been submitted!

    You've already reviewed this product!