Deep networks with billions of parameters trained on large datasets have achieved unprecedented success in various applications, ranging from medical diagnostics to urban planning and autonomous driving, to name a few. However, training large models is contingent on exceptionally large and expensive computational resources. Such infrastructures consume substantial energy, produce a massive amount of carbon footprint, and often soon become obsolete and turn into e-waste. While there has been a persistent effort to improve the performance of machine learning models, their sustainability is often neglected. This realization has motivated the community to look closer at the sustainability and efficiency of machine learning, by identifying the most relevant model parameters or model structures. In this workshop, we examine the community’s progress toward these goals and aim to identify areas that call for additional research efforts. In particular, by bringing researchers with diverse backgrounds, we will focus on the limitations of existing methods for model compression and discuss the tradeoffs among model size and performance. The main goal of the workshop is to bring together researchers from academia, and industry with diverse expertise and points of view on network compression, to discuss how to effectively evaluate and enforce machine learning pipelines to better comply with sustainability and efficiency constraints. Our workshop will consist of a diverse set of speakers (ranging from researchers with hardware background to researchers in neurobiology, and algorithmic ML community) to discuss sparse training algorithms and hardware limitations in various machine learning domains, ranging from robotics and task automation, to vision, natural language processing, and reinforcement learning. The workshop aims to further develop these research directions for the machine learning community.
Fri 12:00 a.m. - 12:05 a.m.
|
Opening Remarks by Organizers
SlidesLive Video » |
🔗 |
Fri 12:05 a.m. - 12:35 a.m.
|
Dan Alistarh: Sparsity is Coming of Age
(
Keynote Talk
)
|
Dan Alistarh 🔗 |
Fri 12:35 a.m. - 12:55 a.m.
|
Gintare Karolina Dziugaite
(
Invited talk
)
SlidesLive Video » |
🔗 |
Fri 12:55 a.m. - 1:15 a.m.
|
Martha White
(
Invited talk
)
SlidesLive Video » |
🔗 |
Fri 1:15 a.m. - 1:35 a.m.
|
Yani Ioannou
(
Invited talk
)
SlidesLive Video » |
🔗 |
Fri 1:35 a.m. - 2:05 a.m.
|
Panel: Understanding Sparsity
(
Panel
)
SlidesLive Video » Panelists: Dan Alistarh, Gintare Karolina Dziugaite, Martha White, Yani Ioannou Moderators: Elena Mocanu and Aleksandra Nowak |
🔗 |
Fri 2:05 a.m. - 2:15 a.m.
|
Coffee Break
|
🔗 |
Fri 2:15 a.m. - 2:25 a.m.
|
Spotlight: PopSparse: Accelerated block sparse matrix multiplication on IPU
(
Spotlight presentation
)
SlidesLive Video » |
Zhiyi Li · Douglas Orr · Valeriu Ohan · Godfrey Da Costa · Tom Murray · Adam Sanders · Deniz Beker · Dominic Masters 🔗 |
Fri 2:25 a.m. - 2:35 a.m.
|
Spotlight: Ten Lessons We Have Learned in the New ''Sparseland'': A Short Handbook for Sparse Neural Network Researchers
(
Spotlight presentation
)
SlidesLive Video » |
Shiwei Liu · Zhangyang Wang 🔗 |
Fri 2:35 a.m. - 2:45 a.m.
|
Spotlight: Dynamic Sparsity Is Channel-Level Sparsity Learner
(
Spotlight presentation
)
SlidesLive Video » |
Lu Yin · Gen Li · Meng Fang · Li Shen · Tianjin Huang · Zhangyang Wang · Xiaolong Ma · Mykola Pechenizkiy · Shiwei Liu 🔗 |
Fri 2:45 a.m. - 3:30 a.m.
|
Poster session
|
🔗 |
Fri 3:30 a.m. - 4:30 a.m.
|
Lunch break
|
🔗 |
Fri 4:30 a.m. - 5:00 a.m.
|
Aakanksha Chowdhery
(
Invited talk
)
SlidesLive Video » |
Aakanksha Chowdhery 🔗 |
Fri 5:00 a.m. - 5:10 a.m.
|
Spotlight: Massive Language Models Can be Accurately Pruned in One-Shot
(
Spotlight presentation
)
SlidesLive Video » |
Elias Frantar · Dan Alistarh 🔗 |
Fri 5:10 a.m. - 5:20 a.m.
|
Spotlight: Efficient Backpropagation for Sparse Training with Speedup
(
Spotlight presentation
)
|
Mahdi Nikdan · Tommaso Pegolotti · Eugenia Iofinova · Eldar Kurtic · Dan Alistarh 🔗 |
Fri 5:20 a.m. - 5:30 a.m.
|
Spotlight: Automatic Noise Filtering with Dynamic Sparse Training in Deep Reinforcement Learning
(
Spotlight presentation
)
SlidesLive Video » |
Bram Grooten · Ghada Sokar · Shibhansh Dohare · Elena Mocanu · Matthew E. Taylor · Mykola Pechenizkiy · Decebal Mocanu 🔗 |
Fri 5:30 a.m. - 6:15 a.m.
|
Breakout session
link »
Topics for brainstorming in small groups: 1) Sparsity for real 2) Sparsity & Generalisation 3) Sparsity & Reinforcement Learning 4) Sparsity & Continual Learning 5) Sparsity Terminology & History 6) Model, activation, and data sparsity 7) Sparsity and Green AI You can find the topics and more details on the attached link. |
🔗 |
Fri 6:15 a.m. - 6:30 a.m.
|
Coffee Break
|
🔗 |
Fri 6:30 a.m. - 6:50 a.m.
|
Ce Zhang
(
Invited talk
)
SlidesLive Video » |
Ce Zhang 🔗 |
Fri 6:50 a.m. - 7:10 a.m.
|
Pavlo Molchanov
(
Invited talk
)
SlidesLive Video » |
Pavlo Molchanov 🔗 |
Fri 7:10 a.m. - 7:30 a.m.
|
Jeff Dean
(
Invited talk
)
SlidesLive Video » |
Jeff Dean 🔗 |
Fri 7:30 a.m. - 8:00 a.m.
|
Panel: Role of Sparsity on Scaling of Neural Networks
(
Panel
)
SlidesLive Video » Panelists: Jeff Dean, Aakanksha Chowdhery, Ce Zhang, Nir Shavit, Pavlo Molchanov Moderator: Utku Evci |
🔗 |
Fri 8:00 a.m. - 8:05 a.m.
|
Closing remarks
SlidesLive Video » |
🔗 |
Fri 8:00 a.m. - 8:45 a.m.
|
Virtual Poster session ( Poster session ) link » | 🔗 |
Fri 9:00 a.m. - 10:00 a.m.
|
Sparsity social
|
🔗 |