24-25 May 2024, Edinburgh, UK
Contact Us: Luo Mai | Edoardo Ponti
Please contact the organizers for more details.
The 2024 International Workshop on Efficient GenAI aims to bring together researchers and practitioners in the field of generative AI, focusing on Large Language Models (LLMs) and Large Multi-Modal Models (LMMs). The objective is to foster collaboration and share insights to improve the training and deployment efficiency of these models. This initiative is expected to drive innovation across the entire system stack, encompassing models, algorithms, software, and hardware. This workshop is funded by the the recently launched Edinburgh Generative AI Lab (GAIL).
We are pleased to announce that we have confirmed invited speakers from University of Edinburgh, UC Berkeley, MIT, Imperial College London, Mila - Quebec AI Institute, GraphCore, Cohere AI, Tenstorrent, Amazon AI Lab, Malted AI, and EPCC (UK’s Supercomputing Centre). We have released a tentative schedule.
The workshop will be held at the Informatics Forum, University of Edinburgh, in room G.07.
The workshop will include keynote speeches, panel discussions, invited talks, and networking sessions.
Please reserve your place for the workshop by clicking the “Register for Workshop” button at the top of the page.
PhD students and researchers are invited to give lightning talks and present posters. To apply, click the “Give a talk and/or present poster of your work” button at the top of the page.
The workshop covers the following topics:
Exploring innovations in language and multi-modal models, such as nanoT5, retrieval-based LLMs, and multi-modal integration techniques exemplified by LLaVA and GPT4V.
Delving into efficient strategies for LLM/LMM training, including parameter-efficient fine-tuning and resource-efficient distributed training approaches.
Addressing efficient inference methodologies, such as speculative decoding, along with compression and quantization techniques.
Focusing on the challenges of efficiently managing datasets for LLMs and LMMs and enhancing the effectiveness of large-scale, intelligent retrieval.
Evaluating software solutions pivotal in training and deploying LLMs, including vLLM, PageAttention, DeepSpeed, Megatron-LM, and AI compilers.
Discussing the deployment of LLMs on emerging hardware platforms such as NVIDIA Graph-Hopper, GraphCore IPU, Cerebras Systems, and Tenstorrent Chips.
Encouraging dialogues at the intersection of LLM efficiency and sectors such as energy, environment, and social studies, highlighting shared challenges and solutions.
Luo Mai
University of Edinburgh
Edoardo Ponti
University of Edinburgh
Pasquale Minervini
University of Edinburgh
Antreas Antoniou
University of Edinburgh
Xuan Sun
University of Edinburgh
Shangmin Guo
University of Edinburgh
Yao Fu
University of Edinburgh
Yeqi Huang
University of Edinburgh
Mirella Lapata
University of Edinburgh