Program

DIMES'24 took place in Room 615B of Hilton Austin.

The workshop proceedings are available in the ACM Digital Library. The individual papers and presentation slides are linked below.

Schedule

09:00 am — Welcome

09:05 am — Keynote: Data, Storage, Memory, and Overall System Efficiency
Gustavo Alonso (ETH Zurich)

09:50 am — Session 1: Let the Numbers Do the Talking – Benchmarks and Models (Chair: Timo Hönig)

10:30 am — Discussion with Speakers from Session 1

10:45 am — Coffee Break

11:00 am — Session 2: There Must Be Order – Memory and Caches (Chair: Frank Bellosa)

12:00 pm — Discussion with Speakers from Session 2

12:15 pm — Lunch

01:00 pm — Keynote: Realistic Expectations for CXL Memory Pools
Daniel S. Berger (Microsoft Azure & University of Washington)

01:45 pm — Session 3: A Silver Lining? – CXL Systems (Chair: Andreas Polze)

02:25 pm — Discussion with Speakers from Session 3

02:40 pm — Short Break

02:50 pm — Panel Discussion
Gustavo Alonso (ETH Zurich), Daniel S. Berger (Microsoft Azure & University of Washington), Philip Levis (Stanford), Andreas Polze (HPI, University of Potsdam), Emmett Witchel (The University of Texas at Austin)

03:35 pm — Coffee Break

04:00 pm — Session 4: Cards on the Table — Demos

04:20 pm — Best Paper Award & Closing Remarks

Keynotes

Realistic Expectations for CXL Memory Pools

Daniel S. Berger (Microsoft Azure and University of Washington)

CXL promises increased memory bandwidth/capacity and cross-node pooling and sharing. While the CLX standard enables many possibilities and inviting configurations, systems research directions depend on what is realistic and can be deployed in datacenters. Our work in Azure Research has contributed to the standardization, development, and deployment of CXL over the past five years. In this talk, we share lessons learned and current expectations for realistic CXL deployments from the perspective of a major cloud platform. We review deployment steps such as CXL 1.1 local memory expansion, CXL 2.0/3.0 memory pools, and CXL 3.0/3.1+ memory sharing with back invalidate, and contrast them with timelines and hardware availability. We also comment on architectural and cost considerations that shed light on specific design questions such as pool sizes and sharing implementations.

Daniel S. Berger
Daniel S. Berger

Daniel is a Principal Researcher at Azure Systems Research and an Affiliate Assistant Professor in Computer Science at the University of Washington. His research focuses on improving the efficiency, sustainability, and reliability of cloud platforms. Before joining Microsoft Azure, he was a member of the Systems Research group at Microsoft Research.

He is the recipient of an ACM ASPLOS 2023 distinguished paper award, USENIX OSDI 2023 best paper award, 2021 ACM SOSP Best Paper Award, the 2018 Mark Stehlik Fellowship at Carnegie Mellon University, and best paper awards at IFIP Performance and ACM WiSec.

Data, Storage, Memory, and Overall System Efficiency

Gustavo Alonso (ETH Zurich)

Many applications dominating the computing landscape are data intensive: data analytics, machine learning, large language models, recommendation systems, etc. The amount of data processed by these systems is staggering and continues to grow at an exponential rate. While the use of more and more data has led to impressive progress in many areas, including storage and memory systems, it has an often-ignored side effect: data movement is expensive, requires many resources, and it is often inefficiently managed. Any serious attempt at improving the sustainability and overall efficiency of data centers must necessarily include improvements in the way we handle and process data. In this talk I will first show comment on the environmental impact of current IT. Then I will discuss why existing systems are inherently inefficient in data movement, resource utilization, and processing requirements. I will then present potential solutions that take advantage of the trends towards specialization and the large economies of scale of the cloud, suggesting along the way how to design data centric architectures at the storage and memory level that are more energy and resource efficient than what we have today.

Gustavo Alonso
Gustavo Alonso

Gustavo Alonso is a professor in the Department of Computer Science of ETH Zurich where he is a member of the Systems Group and the head of the Institute of Computing Platforms. He leads the AMD HACC (Heterogeneous Accelerated Compute Cluster) deployment at ETH, with several hundred users worldwide, a research facility that supports exploring data center hardware-software co-design. His research interests include data management, cloud computing architecture, and building systems on modern hardware. Gustavo holds degrees in telecommunication from the Madrid Technical University and a MS and PhD in Computer Science from UC Santa Barbara. Previous to joining ETH, he was a research scientist at IBM Almaden in San Jose, California.

Gustavo has received 4 Test-of-Time Awards for his research in databases, software runtimes, middleware, and mobile computing. He is an ACM Fellow, an IEEE Fellow, a Distinguished Alumnus of the Department of Computer Science of UC Santa Barbara, and has received the Lifetime Achievements Award from the European Chapter of ACM SIGOPS (EuroSys).