Academia's premiere storage systems research center.
Leading research in storage systems, databases, ML systems, cloud computing, data lakes, etc.
Talks by Recent PDL Personnel
Benjamin Berg, Assistant Professor, University of North Carolina
A New Methodology for Parallel Job Scheduling
Huaicheng Li, Assistant Professor, Virginia Tech
Towards Predictable and Efficient Datacenter Storage
Lin Ma, Assistant Professor, University of Michigan
Putting Your Database on Autopilot:
Self-driving Database Management Systems
Recent PDL Publications
The PDL Packet - Summer 2022 Newsletter
TMO: Transparent Memory Offloading in Datacenters
BEST PAPER AT ASPLOS '22!
Johannes Weiner, Niket Agarwal, Dan Schatzberg, Leon Yang, Hao Wang, Blaise Sanouillet, Bikash Sharma, Tejun Heo, Mayank Jain, Chunqiang Tang, Dimitrios Skarlatos. ASPLOS ’22, February 28 – March 4, 2022, Lausanne, Switzerland.
The unrelenting growth of the memory needs of emerging datacenter applications, along with ever increasing cost and volatility of DRAM prices, has led to DRAM being a major infrastructure expense. Alternative technologies, such as NVMe SSDs and upcoming NVM devices, offer higher capacity than DRAM at a fraction of the cost and power. One promising approach is to transparently offload colder memory to cheaper memory technologies via kernel or hypervisor techniques. [...more]
Kangaroo: Caching Billions of Tiny Objects on Flash
BEST PAPER AT SOSP'21!
Sara McAllister, Benjamin Berg, Julian Tutuncu-Macias, Juncheng Yang, Sathya Gunasekar, Jimmy Lu, Daniel Berger, Nathan Beckmann, Gregory R. Ganger
Proceedings of the 28th ACM Symposium on Operating Systems Principles (SOSP '21) October 25-28, 2021. Virtual Event.
Many social-media and IoT services have very large working sets consisting of billions of tiny (~100 B) objects. Large, flash-based caches are important to serving these working sets at acceptable monetary cost. However, caching tiny objects on flash is challenging for two reasons: (i) SSDs can read/write data only in multi-KB pages that are much larger than a single object, stressing the limited number of times flash can be written; and (ii) very few bits per cached object can be kept in DRAM without losing flash's cost advantage. [...more]