Zum Hauptinhalt springen Zur Suche springen Zur Hauptnavigation springen
Herzlich Willkommen!
This book is a pioneering exploration of the state-of-the-art techniques that drive large language models (LLMs) toward greater efficiency and scalability. Edited by three distinguished experts—Peyman Passban, Mehdi Rezagholizadeh, and Andy Way—this book presents practical solutions to the growing challenges of training and deploying these massive models. With their combined experience across academia, research, and industry, the authors provide insights into the tools and strategies required to improve LLM performance while reducing computational demands.This book is more than just a technical guide; it bridges the gap between research and real-world applications. Each chapter presents cutting-edge advancements in inference optimization, model architecture, and fine-tuning techniques, all designed to enhance the usability of LLMs in diverse sectors. Readers will find extensive discussions on the practical aspects of implementing and deploying LLMs in real-world scenarios. The book serves as a comprehensive resource for researchers and industry professionals, offering a balanced blend of in-depth technical insights and practical, hands-on guidance. It is a go-to reference book for students, researchers in computer science and relevant sub-branches, including machine learning, computational linguistics, and more.
Peyman Passban obtained his Ph.D. in 2018 from Dublin City University, Dublin, Ireland, where he also spent one year as a postdoctoral researcher and worked on neural machine translation models for low-resource and morphologically complex languages. Following that, Peyman has transitioned into full-time research roles within the industry, in which he has held a variety of positions such as Sr Researcher, Research Lead, Director of Engineering in different companies including Huawei Technologies, Amazon, and a few others. He published more than 20 papers in different venues (ACL, AAAI, NAACL, EMNLP, COLING, TALLIP, etc.) and is an active member of the research community (Organizer at LoResMT, Organizer and Area Chair at ENLSP, Committee Member at ACL, EMNLP, Reviewer at Springer Journal of Machine Translation, and many others). The influence of his work extends beyond research facilities. Peyman has actively participated in various sectors, including life sciences, conversational AI agents, and smart glasses.Andy Way obtained a B.Sc. (Hons) in 1986, an M.Sc. in 1989, and his Ph.D. in 2001 from the University of Essex, Colchester, U.K. From 1988 to 91, he worked at the University of Essex, U.K., on the Eurotra MT project. He joined DCU in 1991 and is employed as Full Professor. He was Recipient of the 2015 DCU President’s Research Award for Science and Engineering, and in 2019, he received the extremely prestigious Award of Honour from the International Association for Machine Translation for his services to the community. Prof. Way co-founded the SFI-funded Centre CNGL in 2007 and the ADAPT Centre in 2015. He took a career break from 2011 to 2013 to work in the translation industry in the UK. On his return to DCU in January 2014, Prof. Way acted as Deputy Director of CNGL and subsequently Deputy Director and Co-Applicant of ADAPT.Mehdi Rezagholizadeh obtained a B.Sc. in 2009, an M.Sc. in 2011 from the University of Tehran, and Ph.D. in 2016 from McGill University in Electrical and Computer Engineering (Centre of Intelligent Machines). He joined Huawei in January 2017 and his research focus has been on different deep learning and NLP projects such as generative adversarial networks, neural machine translation, adversarial neural text generation, and efficient NLP for pre-trained language models. He is now Principal Research Scientist, and he has been leading the NLP team of Huawei Noah's Ark Lab in Canada since 2018. He has more than 8 years’ industrial experience in broad spectrum roles such as Software Developer, Research Engineer, Research Scientist, and Team Leader. Mehdi has contributed to more than 15 patents and 50 published papers in top journals and conferences such as TACL, NeurIPS, AAAI, ACL, NAACL, EMNLP, EACL, Interspeech, and ICASSP.