Workshop
ICLR 2023 Workshop on Sparsity in Neural Networks: On practical limitations and tradeoffs between sustainability and efficiency
Baharan Mirzasoleiman · Zhangyang Wang · Decebal Constantin Mocanu · Elena Mocanu · Utku Evci · Trevor Gale · Aleksandra Nowak · Ghada Sokar · Zahra Atashgahi
AD12
Fri 5 May, midnight PDT
Deep networks with billions of parameters trained on large datasets have achieved unprecedented success in various applications, ranging from medical diagnostics to urban planning and autonomous driving, to name a few. However, training large models is contingent on exceptionally large and expensive computational resources. Such infrastructures consume substantial energy, produce a massive amount of carbon footprint, and often soon become obsolete and turn into e-waste. While there has been a persistent effort to improve the performance of machine learning models, their sustainability is often neglected. This realization has motivated the community to look closer at the sustainability and efficiency of machine learning, by identifying the most relevant model parameters or model structures. In this workshop, we examine the community’s progress toward these goals and aim to identify areas that call for additional research efforts. In particular, by bringing researchers with diverse backgrounds, we will focus on the limitations of existing methods for model compression and discuss the tradeoffs among model size and performance. The main goal of the workshop is to bring together researchers from academia, and industry with diverse expertise and points of view on network compression, to discuss how to effectively evaluate and enforce machine learning pipelines to better comply with sustainability and efficiency constraints. Our workshop will consist of a diverse set of speakers (ranging from researchers with hardware background to researchers in neurobiology, and algorithmic ML community) to discuss sparse training algorithms and hardware limitations in various machine learning domains, ranging from robotics and task automation, to vision, natural language processing, and reinforcement learning. The workshop aims to further develop these research directions for the machine learning community.
Schedule
Fri 12:00 a.m. - 12:05 a.m.
|
Opening Remarks by Organizers
(
Opening Remarks by Organizers
)
>
SlidesLive Video |
🔗 |
Fri 12:05 a.m. - 12:35 a.m.
|
Dan Alistarh: Sparsity is Coming of Age
(
Keynote Talk
)
>
|
Dan Alistarh 🔗 |
Fri 12:35 a.m. - 12:55 a.m.
|
Gintare Karolina Dziugaite
(
Invited talk
)
>
SlidesLive Video |
🔗 |
Fri 12:55 a.m. - 1:15 a.m.
|
Martha White
(
Invited talk
)
>
SlidesLive Video |
🔗 |
Fri 1:15 a.m. - 1:35 a.m.
|
Yani Ioannou
(
Invited talk
)
>
SlidesLive Video |
🔗 |
Fri 1:35 a.m. - 2:05 a.m.
|
Panel: Understanding Sparsity
(
Panel
)
>
SlidesLive Video |
🔗 |
Fri 2:05 a.m. - 2:15 a.m.
|
Coffee Break
|
🔗 |
Fri 2:15 a.m. - 2:25 a.m.
|
Spotlight: PopSparse: Accelerated block sparse matrix multiplication on IPU
(
Spotlight presentation
)
>
SlidesLive Video |
Zhiyi Li · Douglas Orr · Valeriu Ohan · Godfrey Da Costa · Tom Murray · Adam Sanders · Deniz Beker · Dominic Masters 🔗 |
Fri 2:25 a.m. - 2:35 a.m.
|
Spotlight: Ten Lessons We Have Learned in the New ''Sparseland'': A Short Handbook for Sparse Neural Network Researchers
(
Spotlight presentation
)
>
SlidesLive Video |
Shiwei Liu · Zhangyang Wang 🔗 |
Fri 2:35 a.m. - 2:45 a.m.
|
Spotlight: Dynamic Sparsity Is Channel-Level Sparsity Learner
(
Spotlight presentation
)
>
SlidesLive Video |
Lu Yin · Gen Li · Meng Fang · Li Shen · Tianjin Huang · Zhangyang Wang · Xiaolong Ma · Mykola Pechenizkiy · Shiwei Liu 🔗 |
Fri 2:45 a.m. - 3:30 a.m.
|
Poster session
(
Poster session
)
>
|
🔗 |
Fri 3:30 a.m. - 4:30 a.m.
|
Lunch break
(
Lunch break
)
>
|
🔗 |
Fri 4:30 a.m. - 5:00 a.m.
|
Aakanksha Chowdhery
(
Invited talk
)
>
SlidesLive Video |
Aakanksha Chowdhery 🔗 |
Fri 5:00 a.m. - 5:10 a.m.
|
Spotlight: Massive Language Models Can be Accurately Pruned in One-Shot
(
Spotlight presentation
)
>
SlidesLive Video |
Elias Frantar · Dan Alistarh 🔗 |
Fri 5:10 a.m. - 5:20 a.m.
|
Spotlight: Efficient Backpropagation for Sparse Training with Speedup
(
Spotlight presentation
)
>
|
Mahdi Nikdan · Tommaso Pegolotti · Eugenia Iofinova · Eldar Kurtic · Dan Alistarh 🔗 |
Fri 5:20 a.m. - 5:30 a.m.
|
Spotlight: Automatic Noise Filtering with Dynamic Sparse Training in Deep Reinforcement Learning
(
Spotlight presentation
)
>
SlidesLive Video |
Bram Grooten · Ghada Sokar · Shibhansh Dohare · Elena Mocanu · Matthew E. Taylor · Mykola Pechenizkiy · Decebal Mocanu 🔗 |
Fri 5:30 a.m. - 6:15 a.m.
|
Breakout session ( Breakout session ) > link | 🔗 |
Fri 6:15 a.m. - 6:30 a.m.
|
Coffee Break
|
🔗 |
Fri 6:30 a.m. - 6:50 a.m.
|
Ce Zhang
(
Invited talk
)
>
SlidesLive Video |
Ce Zhang 🔗 |
Fri 6:50 a.m. - 7:10 a.m.
|
Pavlo Molchanov
(
Invited talk
)
>
SlidesLive Video |
Pavlo Molchanov 🔗 |
Fri 7:10 a.m. - 7:30 a.m.
|
Jeff Dean
(
Invited talk
)
>
SlidesLive Video |
Jeff Dean 🔗 |
Fri 7:30 a.m. - 8:00 a.m.
|
Panel: Role of Sparsity on Scaling of Neural Networks
(
Panel
)
>
SlidesLive Video |
🔗 |
Fri 8:00 a.m. - 8:05 a.m.
|
Closing remarks
(
Closing remarks
)
>
SlidesLive Video |
🔗 |
Fri 8:00 a.m. - 8:45 a.m.
|
Virtual Poster session ( Poster session ) > link | 🔗 |
Fri 9:00 a.m. - 10:00 a.m.
|
Sparsity social
(
Sparsity social
)
>
|
🔗 |