Call for Papers

ScaDL 2023: Fifth IPDPS Workshop on Scalable Deep Learning over Parallel and Distributed Infrastructure 

Scope of the Workshop

Recently, Deep Learning (DL) has received tremendous attention in the research community because of the impressive results obtained for a large number of machine learning problems. The success of state-of-the-art deep learning systems relies on training deep neural networks over a massive amount of training data, which typically requires a large-scale distributed computing infrastructure to run. In order to run these jobs in a scalable and efficient manner, on cloud infrastructure or dedicated HPC systems, several interesting research topics have emerged which are specific to DL. The sheer size and complexity of deep learning models when trained over a large amount of data makes them harder to converge in a reasonable amount of time. It demands advancement along multiple research directions such as, model/data parallelism, model/data compression, distributed optimization algorithms for DL convergence, synchronization strategies, efficient communication, federated learning and specific hardware acceleration. Distributed DL becomes even more challenging when one considers additional desiderata of trustworthiness such as privacy, adversarial robustness, and fairness.

Areas of Interest

In this workshop, we solicit research papers focused on distributed deep learning aiming to achieve efficiency and scalability for deep learning jobs over distributed and parallel systems. Papers focusing both on algorithms as well as systems are welcome. We invite authors to submit papers on topics including but not limited to:

ScaDL Research Directions

ScaDL seeks to advance the following research directions:

This intersection of distributed/parallel computing and deep learning is becoming critical and demands specific attention to address the above topics which some of the broader forums may not be able to provide. The aim of this workshop is to foster collaboration among researchers from distributed/parallel computing and deep learning communities to share the relevant topics as well as results of the current approaches lying at the intersection of these areas.

Submission Link

Please log in to Linklings here (create an account if necessary). Once you login, you will find a link to submissions for the ScaDL workshop.

Key Dates

Paper Submission Deadline: February 5, 2023 February 14, 2023

Acceptance Notification: February 26, 2023

Camera ready papers due: March 7, 2023


Author Instructions

ScaDL 2023 accepts submissions in two categories:

Regular papers: 8-10 pages

Short papers/Work in progress: 4 pages

The aforementioned lengths include all technical content, references and appendices.

We encourage submissions that are original research work, work in progress, case studies, vision papers, and industrial experience papers.

Papers should be formatted using IEEE conference style, including figures, tables, and references. The IEEE conference style templates for MS Word and LaTeX provided by IEEE eXpress Conference Publishing are available for download. See the latest versions at


General Chairs

Kaoutar El Maghraoui, IBM Research AI, USA

Daniele Lezzi, Barcelona Supercomputing Center, Spain

Program Committee Chairs

Alex Gittens, Rensselaer Polytechnic Institute (RPI), USA

Misbah Mubarak, NVIDIA, USA

Steering Committee

Parijat Dube, IBM Research AI, USA

Danilo Ardagna, Politecnico di Milano, Italy

Vinod Muthusamy, IBM Research AI

Ashish Verma, Amazon

Jayaram Kallapalayam Radhakrishnan, IBM Research AI, USA

Yogish Sabharwal, IBM Research AI, India