Buy Enhancing LLM Performance by Andy Way at Bookstore UAE
Book 1
Book 2
Book 3
Book 1
Book 2
Book 3
Book 1
Book 2
Book 3
Book 1
Book 2
Book 3
Home > Computing and Information Technology > Computer science > Artificial intelligence > Machine learning > Enhancing LLM Performance: Efficacy, Fine-Tuning, and Inference Techniques(7 Machine Translation: Technologies and Applications)
37%
Enhancing LLM Performance: Efficacy, Fine-Tuning, and Inference Techniques(7 Machine Translation: Technologies and Applications)

Enhancing LLM Performance: Efficacy, Fine-Tuning, and Inference Techniques(7 Machine Translation: Technologies and Applications)


     0     
5
4
3
2
1



Available


X
About the Book

This book is a pioneering exploration of the state-of-the-art techniques that drive large language models (LLMs) toward greater efficiency and scalability. Edited by three distinguished experts--Peyman Passban, Mehdi Rezagholizadeh, and Andy Way--this book presents practical solutions to the growing challenges of training and deploying these massive models. With their combined experience across academia, research, and industry, the authors provide insights into the tools and strategies required to improve LLM performance while reducing computational demands. This book is more than just a technical guide; it bridges the gap between research and real-world applications. Each chapter presents cutting-edge advancements in inference optimization, model architecture, and fine-tuning techniques, all designed to enhance the usability of LLMs in diverse sectors. Readers will find extensive discussions on the practical aspects of implementing and deploying LLMs in real-world scenarios. The book serves as a comprehensive resource for researchers and industry professionals, offering a balanced blend of in-depth technical insights and practical, hands-on guidance. It is a go-to reference book for students, researchers in computer science and relevant sub-branches, including machine learning, computational linguistics, and more.

Table of Contents:
Introduction and Fundamentals.- SPEED: Speculative Pipelined Execution for Efficient Decoding.- Efficient LLM Inference on CPUs.- KronA: Parameter-Efficient Tuning with Kronecker Adapter.- LoDA: Low-Dimensional Adaptation of Large Language Models.- Sparse Fine-Tuning for Inference Acceleration of Large Language Models.- TCNCA: Temporal CNN with Chunked Attention for Efficient Training on Long Sequences.- Class-Based Feature Knowledge Distillation.- On the Use of Cross-Attentive Fusion Techniques for Audio-Visual Speaker Verification.- An Efficient Clustering Algorithm for Self-Supervised Speaker Recognition.- Remaining Issues for AI.

About the Author :
Peyman Passban obtained his Ph.D. in 2018 from Dublin City University, Dublin, Ireland, where he also spent one year as a postdoctoral researcher and worked on neural machine translation models for low-resource and morphologically complex languages. Following that, Peyman has transitioned into full-time research roles within the industry, in which he has held a variety of positions such as Sr Researcher, Research Lead, Director of Engineering in different companies including Huawei Technologies, Amazon, and a few others. He published more than 20 papers in different venues (ACL, AAAI, NAACL, EMNLP, COLING, TALLIP, etc.) and is an active member of the research community (Organizer at LoResMT, Organizer and Area Chair at ENLSP, Committee Member at ACL, EMNLP, Reviewer at Springer Journal of Machine Translation, and many others). The influence of his work extends beyond research facilities. Peyman has actively participated in various sectors, including life sciences, conversational AI agents, and smart glasses. Andy Way obtained a B.Sc. (Hons) in 1986, an M.Sc. in 1989, and his Ph.D. in 2001 from the University of Essex, Colchester, U.K. From 1988 to 91, he worked at the University of Essex, U.K., on the Eurotra MT project. He joined DCU in 1991 and is employed as Full Professor. He was Recipient of the 2015 DCU President's Research Award for Science and Engineering, and in 2019, he received the extremely prestigious Award of Honour from the International Association for Machine Translation for his services to the community. Prof. Way co-founded the SFI-funded Centre CNGL in 2007 and the ADAPT Centre in 2015. He took a career break from 2011 to 2013 to work in the translation industry in the UK. On his return to DCU in January 2014, Prof. Way acted as Deputy Director of CNGL and subsequently Deputy Director and Co-Applicant of ADAPT. Mehdi Rezagholizadeh obtained a B.Sc. in 2009, an M.Sc. in 2011 from the University of Tehran, and Ph.D. in 2016 from McGill University in Electrical and Computer Engineering (Centre of Intelligent Machines). He joined Huawei in January 2017 and his research focus has been on different deep learning and NLP projects such as generative adversarial networks, neural machine translation, adversarial neural text generation, and efficient NLP for pre-trained language models. He is now Principal Research Scientist, and he has been leading the NLP team of Huawei Noah's Ark Lab in Canada since 2018. He has more than 8 years' industrial experience in broad spectrum roles such as Software Developer, Research Engineer, Research Scientist, and Team Leader. Mehdi has contributed to more than 15 patents and 50 published papers in top journals and conferences such as TACL, NeurIPS, AAAI, ACL, NAACL, EMNLP, EACL, Interspeech, and ICASSP.


Best Sellers


Product Details
  • ISBN-13: 9783031857461
  • Publisher: Springer International Publishing AG
  • Publisher Imprint: Springer International Publishing AG
  • Height: 235 mm
  • No of Pages: 183
  • Series Title: 7 Machine Translation: Technologies and Applications
  • Width: 155 mm
  • ISBN-10: 3031857461
  • Publisher Date: 05 Jul 2025
  • Binding: Hardback
  • Language: English
  • Returnable: N
  • Sub Title: Efficacy, Fine-Tuning, and Inference Techniques


Similar Products

Add Photo
Add Photo

Customer Reviews

REVIEWS      0     
Click Here To Be The First to Review this Product
Enhancing LLM Performance: Efficacy, Fine-Tuning, and Inference Techniques(7 Machine Translation: Technologies and Applications)
Springer International Publishing AG -
Enhancing LLM Performance: Efficacy, Fine-Tuning, and Inference Techniques(7 Machine Translation: Technologies and Applications)
Writing guidlines
We want to publish your review, so please:
  • keep your review on the product. Review's that defame author's character will be rejected.
  • Keep your review focused on the product.
  • Avoid writing about customer service. contact us instead if you have issue requiring immediate attention.
  • Refrain from mentioning competitors or the specific price you paid for the product.
  • Do not include any personally identifiable information, such as full names.

Enhancing LLM Performance: Efficacy, Fine-Tuning, and Inference Techniques(7 Machine Translation: Technologies and Applications)

Required fields are marked with *

Review Title*
Review
    Add Photo Add up to 6 photos
    Would you recommend this product to a friend?
    Tag this Book Read more
    Does your review contain spoilers?
    What type of reader best describes you?
    I agree to the terms & conditions
    You may receive emails regarding this submission. Any emails will include the ability to opt-out of future communications.

    CUSTOMER RATINGS AND REVIEWS AND QUESTIONS AND ANSWERS TERMS OF USE

    These Terms of Use govern your conduct associated with the Customer Ratings and Reviews and/or Questions and Answers service offered by Bookswagon (the "CRR Service").


    By submitting any content to Bookswagon, you guarantee that:
    • You are the sole author and owner of the intellectual property rights in the content;
    • All "moral rights" that you may have in such content have been voluntarily waived by you;
    • All content that you post is accurate;
    • You are at least 13 years old;
    • Use of the content you supply does not violate these Terms of Use and will not cause injury to any person or entity.
    You further agree that you may not submit any content:
    • That is known by you to be false, inaccurate or misleading;
    • That infringes any third party's copyright, patent, trademark, trade secret or other proprietary rights or rights of publicity or privacy;
    • That violates any law, statute, ordinance or regulation (including, but not limited to, those governing, consumer protection, unfair competition, anti-discrimination or false advertising);
    • That is, or may reasonably be considered to be, defamatory, libelous, hateful, racially or religiously biased or offensive, unlawfully threatening or unlawfully harassing to any individual, partnership or corporation;
    • For which you were compensated or granted any consideration by any unapproved third party;
    • That includes any information that references other websites, addresses, email addresses, contact information or phone numbers;
    • That contains any computer viruses, worms or other potentially damaging computer programs or files.
    You agree to indemnify and hold Bookswagon (and its officers, directors, agents, subsidiaries, joint ventures, employees and third-party service providers, including but not limited to Bazaarvoice, Inc.), harmless from all claims, demands, and damages (actual and consequential) of every kind and nature, known and unknown including reasonable attorneys' fees, arising out of a breach of your representations and warranties set forth above, or your violation of any law or the rights of a third party.


    For any content that you submit, you grant Bookswagon a perpetual, irrevocable, royalty-free, transferable right and license to use, copy, modify, delete in its entirety, adapt, publish, translate, create derivative works from and/or sell, transfer, and/or distribute such content and/or incorporate such content into any form, medium or technology throughout the world without compensation to you. Additionally,  Bookswagon may transfer or share any personal information that you submit with its third-party service providers, including but not limited to Bazaarvoice, Inc. in accordance with  Privacy Policy


    All content that you submit may be used at Bookswagon's sole discretion. Bookswagon reserves the right to change, condense, withhold publication, remove or delete any content on Bookswagon's website that Bookswagon deems, in its sole discretion, to violate the content guidelines or any other provision of these Terms of Use.  Bookswagon does not guarantee that you will have any recourse through Bookswagon to edit or delete any content you have submitted. Ratings and written comments are generally posted within two to four business days. However, Bookswagon reserves the right to remove or to refuse to post any submission to the extent authorized by law. You acknowledge that you, not Bookswagon, are responsible for the contents of your submission. None of the content that you submit shall be subject to any obligation of confidence on the part of Bookswagon, its agents, subsidiaries, affiliates, partners or third party service providers (including but not limited to Bazaarvoice, Inc.)and their respective directors, officers and employees.

    Accept

    Fresh on the Shelf


    Inspired by your browsing history


    Your review has been submitted!

    You've already reviewed this product!