Loading…
Attending this event?
4 - 5 December 2024 | Geneva, Switzerland
View More Details & Registration

The Sched app allows you to build your schedule but is separate from your event registration. You must be registered for Cephalocon 2024 to participate in the sessions. If you have not registered but would like to join us, please go to the event registration page to purchase a registration.

This schedule is automatically displayed in Central European Time. To see the schedule in your preferred timezone, select from the drop-down menu located at the bottom of the menu to the right.
Wednesday, December 4
 

08:00 CET

Registration + Badge Pick-up
Wednesday December 4, 2024 08:00 - 19:30 CET
Wednesday December 4, 2024 08:00 - 19:30 CET
Lower Foyer

09:00 CET

Keynote: Welcome & Opening Remarks - Enrica Porcari, Head of Information Technology Department, CERN
Wednesday December 4, 2024 09:00 - 09:10 CET
Speakers
avatar for Enrica Porcari

Enrica Porcari

Head Of Information Technology Department, CERN
Enrica Maria Porcari is Head of the IT Department at CERN.Previously Enrica was the UN World Food Programme’s Chief Information Officer and Director of Technology and Chair of the UN’s Emergency Telecommunications Cluster. In this role Enrica drove WFP to be the leading edge of... Read More →
Wednesday December 4, 2024 09:00 - 09:10 CET
SG Auditorium A

09:10 CET

Keynote Sessions: To Be Announced
Wednesday December 4, 2024 09:10 - 09:40 CET
Wednesday December 4, 2024 09:10 - 09:40 CET
SG Auditorium A

09:40 CET

Keynote: Powering Modern Enterprise Data Storage with Ceph - Vincent Hsu, VP & CTO of IBM Storage & Neha Ojha, Senior Development Manager, IBM
Wednesday December 4, 2024 09:40 - 09:55 CET
The evolution of data centers for the last several decades is evolving from siloed enterprise data centers to cloud data centers then hybrid cloud architecture. More than ever, customers are asking for a consistent architecture for the on-prem data centers behave like cloud – Cloud native architecture on prem. This is perfect for Ceph since we have seen evidence of Ceph adoption in enterprises across different industries.
 
Ceph’s vision is to serve wide variety of workloads in data centers and Ceph has the best technology to address today’s storage business needs. This talk will describe the topics and focus areas that IBM is proposing to the Ceph open-source community to accelerate Ceph adoption for enterprise workloads. We would like to see the larger community invest and innovate to accelerate these capabilities. 
Speakers
avatar for Vincent Hsu

Vincent Hsu

IBM Fellow, VP and CTO of IBM Storage, IBM
Vincent Hsu is an IBM Fellow, VP and CTO of IBM storage. He oversees the technical strategy across IBM storage portfolio. He represents IBM in Ceph governance board and actively participating in Ceph community activities. Vincent Hsu has been in storage industry over 30 years across... Read More →
avatar for Neha Ojha

Neha Ojha

Senior Development Manager, IBM
Neha Ojha is the Senior Development Manager of the Ceph Core Team at IBM and an elected member of the Ceph Executive Council, where she leads the project's open-source community initiatives. Before her current role, Neha served as the Project Technical Lead for Ceph's RADOS component... Read More →
Wednesday December 4, 2024 09:40 - 09:55 CET
SG Auditorium A

10:00 CET

Keynote: State of the Cephalopod - Speakers To Be Announced
Wednesday December 4, 2024 10:00 - 10:40 CET
Wednesday December 4, 2024 10:00 - 10:40 CET
SG Auditorium A

10:40 CET

Break
Wednesday December 4, 2024 10:40 - 11:00 CET
Wednesday December 4, 2024 10:40 - 11:00 CET
Lower Foyer

10:40 CET

Solutions Showcase
Wednesday December 4, 2024 10:40 - 19:30 CET
Wednesday December 4, 2024 10:40 - 19:30 CET
Lower Foyer

11:00 CET

Beyond Particle Physics: The Impact of Ceph and OpenStack on CERN's Multi-Datacenter Cloud Strategy - Enrico Bocchi, CERN, European Organization for Nuclear Research & Jose Castro Leon, CERN
Wednesday December 4, 2024 11:00 - 11:35 CET
CERN IT operates a large-scale storage and computing infrastructure at the service of scientific research and its user community: Ceph provides block, object, and file storage at a scale of 100 PBs, while OpenStack provisions bare-metal nodes, VMs, and virtual networking managing more than 450k CPUs. With the advent of a new computing center a few kilometers away from the main campus, compute and storage resources have been re-imagined to extend the capabilities offered by the infrastructure, while putting upfront clear design choices to favor availability and ease of operations. . In this presentation we report on how the new computing center was designed to host cohesively compute and storage resources, how integration with the existing computing center was achieved, and which new capabilities have been unlocked thanks to the newly-built DC. For Ceph in particular, we share insights on achieving data locality with compute resources, deploying a multi-site object storage service, and running a CephFS service that spans across both data centers.
Speakers
avatar for Jose Castro Leon

Jose Castro Leon

Cloud Technical Leader, CERN, European Organization for Nuclear Research
Jose is the Technical Leader for the CERN Cloud Infrastructure Service. He holds a Msc in Computer Science from Universidad de Oviedo. He joined CERN in 2010 and since then he was been working in virtualisation first and then he become part of the cloud team who build the CERN's OpenStack-based... Read More →
avatar for Enrica Porcari

Enrica Porcari

Head Of Information Technology Department, CERN
Enrica Maria Porcari is Head of the IT Department at CERN.Previously Enrica was the UN World Food Programme’s Chief Information Officer and Director of Technology and Chair of the UN’s Emergency Telecommunications Cluster. In this role Enrica drove WFP to be the leading edge of... Read More →
Wednesday December 4, 2024 11:00 - 11:35 CET
SG Auditorium C
  Session Presentation
  • Audience Level Any

11:00 CET

Ceph NVMe-of Road Map - Orit Wasserman & Mike Burkhart, IBM
Wednesday December 4, 2024 11:00 - 11:35 CET
Discover the current status and our future plans for the Ceph NVMe-oF gateway. This session will cover the latest developments, highlighting new features focusing on security and performance enhancements. Join us to see what’s next for Ceph NVMe-oF!
Speakers
avatar for Mike Burkhart

Mike Burkhart

Technical Product Manager - IBM Storage Ceph, NVMe/TCP and VMware Integration, IBM
Mike is a 25 year veteran of the IT data center space, spanning software development and testing, data center and hybrid cloud architecture, as well as now product management. Currently he collaborates with the brilliant engineers who develop Ceph to bring new features to opensource... Read More →
avatar for Orit Wasserman

Orit Wasserman

Distingiushed Engineer, IBM
Orit is a Distinguished Engineer at IBM, specializing in Software Defined Storage (Ceph) and storage for containerized apps (OpenShift Data Foundation) as well as hybrid/multi-cloud. With a strong background as a software engineer and architect, Orit's passion lies in open-source... Read More →
Wednesday December 4, 2024 11:00 - 11:35 CET
SG Auditorium B
  Session Presentation
  • Audience Level Any

11:00 CET

Enhancing Observability and Monitoring for Large Ceph Clusters at Scale - Filipp Akinfiev, Clyso GmbH
Wednesday December 4, 2024 11:00 - 11:35 CET
Maintaining performance and reliability in large Ceph clusters, especially with Rados Gateway (RadosGW), is challenging. Traditional observability approaches often generate excessive data without providing actionable insights. This speech introduces an advanced observability architecture that combines basic monitoring with on-demand detailed and event-triggered monitoring, ensuring continuous visibility and dynamic responsiveness. We'll explore: - The four-layer architecture, detailing producers, the NATS messaging backbone, and consumers. - Monitoring Techniques, including basic, on-demand, and event-triggered monitoring. - A Case Study, sharing proof of concept insights and lessons learned. - Future Directions, discussing potential advancements. This presentation is designed for cloud infrastructure engineers, SREs, and DevOps professionals looking to implement a scalable observability framework that improves system health and performance.
Speakers
avatar for Filipp Akinfiev

Filipp Akinfiev

Senior System Architect, Clyso GmbH
With over 30 years of experience in software development and system administration, I possess a broad technical knowledge in software and system architecture, as well as integration architectures. I specialize in designing and implementing forward-thinking, high-performance software... Read More →
Wednesday December 4, 2024 11:00 - 11:35 CET
SG Auditorium A
  Session Presentation
  • Audience Level Any

11:40 CET

Keeping Ceph RGW Object Storage Consistent - Jane Zhu, Bloomberg
Wednesday December 4, 2024 11:40 - 12:15 CET
Data powers Bloomberg’s financial products. Ceph clusters are the backbone of Bloomberg’s internal S3 cloud storage systems, which host this data and serve billions of requests a day.During the intensive usage of the Ceph RGW object storage with multi-site settings, we encountered different types of data inconsistencies, such as bucket-index and RADOS object inconsistency, unfinished transactions, and multi-site replication inconsistency, etc. These inconsistencies may potentially be caused by software bugs, race conditions, system timeout, and other reasons. Since we cannot guarantee the system is always bug-free and operating smoothly, it’s crucial that we can identify the inconsistency – should it happen – and fix or report it. While there are existing tools and code in place to help address some of these issues, their usage has limitations. As such, we are proposing a scalable and extensible bucket scrubbing approach to systematically check and identify and fix any inconsistency in the RGW object storage system at the bucket-level, if possible. This talk will discuss the design of this bucket scrubbing system and a prototype of it that we are implementing at Bloomberg.
Speakers
avatar for Jane Zhu

Jane Zhu

Senior Software Engineer, Bloomberg
Dr. Jane Zhu is a Senior Software Engineer in the Storage Engineering team at Bloomberg. Jane and her team designed and built a highly available, scalable, and durable software-defined cloud storage platform inside the Bloomberg ecosystem. Jane worked in the industry for more than... Read More →
Wednesday December 4, 2024 11:40 - 12:15 CET
SG Auditorium C
  Session Presentation
  • Audience Level Any

11:40 CET

Pull Requests and Reviews for Good - Gregory Farnum & Sam Just, IBM
Wednesday December 4, 2024 11:40 - 12:15 CET
You’ve built an amazing new feature for your Ceph use case and want to share it with the world. Now what? You make a pull request! Learn how to prepare and present your new work effectively and successfully within the Ceph community, and understand what you can expect from other contributors who are reviewing it. 
Conversely, you are asked to review a PR, and maybe it needs work. How do you share feedback in a way that it’s heard and handled? Which feedback is worth sharing at which stage of the process? And what are you promising to everybody else when you provide that “Reviewed-by” tag? This talk is for anybody who writes or reviews code (or wants to!) in the Ceph project. Prepared and presented by two of Ceph’s original four tech leads, hear about pitfalls and tips developed submitting PRs, mentoring new developers, and reviewing code from drive-by contributors, Ceph startups, consultancies, and future maintainers!
Speakers
avatar for Sam Just

Sam Just

Engineer, IBM
Sam began working on the Ceph project in 2011. Most of his time currently is spent working on crimson, the next generation ceph-osd implementation.
avatar for Gregory Farnum

Gregory Farnum

CephFS Engineering Manager, IBM
Greg Farnum has been in the core Ceph development group since 2009. Greg has contributed major work to CephFS and RADOS, contributed foundational work in the early days of RBD and RGW, previously served as the CephFS tech lead, and now manages IBM’s CephFS development team while... Read More →
Wednesday December 4, 2024 11:40 - 12:15 CET
SG Auditorium B

11:40 CET

The SMB Report Card - John Mulligan, IBM
Wednesday December 4, 2024 11:40 - 12:15 CET
Is 2024 the year of SMB on Ceph? It was for me. In his talk I will discuss the progress of our effort to add managed SMB suport to Ceph. Focusing primarily on the Orchestration aspects of this work, I will talk about some of the major steps it took to integrate Samba with Ceph, some of the projects outside Ceph that help make it happen, and some of our future plans. We will look into the commands needed to set up an SMB Cluster and Shares and demonstrate the workflow involved in connecting some Windows and Linux clients to an SMB Share using Active Directory authentication.
Speakers
avatar for John Mulligan

John Mulligan

Software Developer, IBM
John Mulligan is a developer at IBM working on the Ceph team. John's current focus is adding SMB support to Ceph. In addition to SMB/Samba John is interested in topics including Containers, Python, and Orchestration.
Wednesday December 4, 2024 11:40 - 12:15 CET
SG Auditorium A

12:15 CET

Lunch
Wednesday December 4, 2024 12:15 - 13:45 CET
Wednesday December 4, 2024 12:15 - 13:45 CET
Globe of Science and Innovation

13:45 CET

Session To Be Announced - Nathan Goulding, Vultr
Wednesday December 4, 2024 13:45 - 14:00 CET
Speakers
avatar for Nathan Goulding

Nathan Goulding

Senior Vice President, Engineering, Vultr
Nathan Goulding is an entrepreneurial-minded, product-focused technical leader with over 20 years of infrastructure, platform, and software as-a-service experience. As SVP, Engineering at Vultr, Nathan leads the engineering and technical product management teams. Prior to Vultr, Nathan... Read More →
Wednesday December 4, 2024 13:45 - 14:00 CET
SG Auditorium C

13:45 CET

Bringing a Ceph Based Enterprise Email System Into the Field - Danny Al-Gaaf, Deutsche Telekom AG
Wednesday December 4, 2024 13:45 - 14:20 CET
Deutsche Telekom operates a growing email system with several million accounts and billions of emails stored on traditional NFS. Six years ago we introduced librmb (librados mailbox) to the community, a universal open source library to store emails in a Ceph cluster. Librmb uses RADOS to store email directly in Ceph to achieve maximum performance through parallel access from many email gateways simultaneously, for millions of active customers. Email systems are much too complex to be simulated in a way which would allow to verify if the switch to librmb will work for a large number of users. Therefore a field test with active customers was necessary to provide an educated guess on the behavior of the final setup. This presentation will cover the results from artificial and real field tests with more than 1 million accounts/users. The results include the experience and learnings of migrating from the existing email system into Ceph, an extended time of running the system and from migrating the accounts out of the test system. We will provide an insight into our learnings, found issues, potential solutions and an outlook into our next steps towards a Ceph based email system.
Speakers
avatar for Danny Al-Gaaf

Danny Al-Gaaf

Senior Cloud Technologist, Deutsche Telekom AG
Danny Al-Gaaf is a Senior Cloud Technologist working for Deutsche Telekom. As a Ceph upstream developer he is a driver for using Ceph at Deutsche Telekom. For the last 15 years his professional focus has been on Linux and open source. He works actively in several upstream communities... Read More →
Wednesday December 4, 2024 13:45 - 14:20 CET
SG Auditorium A

13:45 CET

Crimson Project Update - Matan Breizman & Aishwarya Mathuria, IBM
Wednesday December 4, 2024 13:45 - 14:20 CET
The Crimson project is an effort to build a replacement ceph-osd well suited to the new reality of low latency, high throughput, persistent memory and NVMe technologies. Built on the Seastar C++ framework, crimson-osd aims to be able to fully exploit modern devices by minimizing latency, CPU overhead, and cross-core communication. This talk will discuss the current state of Crimson going into the Tentacle release.
Speakers
avatar for Aishwarya Mathuria

Aishwarya Mathuria

Senior Software Engineer, IBM
avatar for Matan Breizman

Matan Breizman

Crimson's Tech Lead, IBM
Matan is Crimson's tech lead, which is the next generation Ceph OSD. Part of the Core Ceph development group since 2021.
Wednesday December 4, 2024 13:45 - 14:20 CET
SG Auditorium B

14:25 CET

Follow the Advice, but Not Too Closely - Building the Cluster YOU Need - Gregory Orange, Pawsey Supercomputing Centre
Wednesday December 4, 2024 14:25 - 14:35 CET
Pawsey receives infrequent blocks of government funding to provide supercomputing, cloud and data services typically at no cost to scientific researchers in Australia. This has a big impact on both the way requirements are determined and on the design and procurement model. Our new 20PB cluster pushes the bounds of what is recommended, for performance and for node density, but we are confident it will serve researchers' needs over the lifespan of the equipment and beyond. This talk will look at the details of this design, and how we determined them, tested them, and increased our confidence of success. Many of these factors are useful in deploying any new cluster, so you will take away some tips on how to get started, or use the experience you already have to build the next one.
Speakers
avatar for Gregory Orange

Gregory Orange

Gregory Orange, Pawsey Supercomputing Centre
Greg deploys and runs Ceph and OpenStack for scientists in Australia, and is involved in the floor tiles, the RGW nodes, the teams who look after it all, and nearly everything in between. Automation and efficiency are of great interest to him, and so is the well-being of all the people... Read More →
Wednesday December 4, 2024 14:25 - 14:35 CET
SG Auditorium A

14:25 CET

Conditional End2end Tracing - Yuval Lifshitz, IBM & Deepika Upadhyay, Clyso GmBH
Wednesday December 4, 2024 14:25 - 15:00 CET
End to end tracing can help debug latency issues between the RGW and the OSD, giving a complete picture of request flow, but tracing itself has a performance impact on the system. When your production system runs into issues, the last thing you want is to put it under more pressure! In this session we will demonstrate how to use Lua scripting on the RGW to turn opentelemetry based tracing only for some of the incoming requests. Allowing us to focus on the problem without slowing down the entire system.
Speakers
avatar for Yuval Lifshitz

Yuval Lifshitz

Senior Technical Staff Member, IBM
Yuval Lifshitz works as a Senior Technical Staff Member at IBM. His current focus is enriching the Ceph ecosystem by adding connectivity between the Rados Object Gateway and external message brokers (Kafka, Knative, RabbitMQ, etc.). He also added Lua scripting into the Rados Object... Read More →
avatar for Deepika Upadhyay

Deepika Upadhyay

Ceph Engineer, Clyso GmBH
Deepika is currently working as a Ceph Engineer at Clyso GmBH and is a contributor to Ceph and Rook project, she has worked as an Outreachy intern for Ceph with focus on adding tracing to Ceph OSD. She has also worked as a RADOS and RBD(Block based storage) teams and now working with... Read More →
Wednesday December 4, 2024 14:25 - 15:00 CET
SG Auditorium B

14:25 CET

The Art of Teuthology - Patrick Donnelly, IBM, Inc.
Wednesday December 4, 2024 14:25 - 15:00 CET
The Ceph project has used the Teuthology testing framework for much of its history. The custom framework is used to schedule batch jobs that perform e2e testing of Ceph. This is orchestrated using a suite of YAML fragments to alternate test modes, configurations, workloads, and other parameters. Teuthology assembles these fragments into a static matrix with potentially dozens of dimensions ultimately producing a combinatoric explosion of jobs which are evaluated, in practice, as smaller subsets for scheduling. We will explore an alternative directed graph model for constructing jobs from a suite of YAML fragments using path walks. Code adapted to this model has been constructed to produce subsets in linear time and provide Lua scriptable control of YAML fragment generation. The latter new feature empowers us to test Ceph with more rigor and completeness. For example, upgrade suites can be constructed using all possible versions of Ceph that are valid upgrade paths to a target release. We will explore this and other enhancements in depth. The audience can expect to leave with a firm and visual understanding of how QA is performed on Ceph and a vision for the future testing.
Speakers
avatar for Patrick Donnelly

Patrick Donnelly

Software Architect, IBM, Inc.
Patrick Donnelly is a Software Architect at IBM, Inc. working as part of the global development team on the open source Ceph distributed storage system. Patrick has principally worked on the Ceph file system (CephFS) since 2016. He has been working on Open Source projects for the... Read More →
Wednesday December 4, 2024 14:25 - 15:00 CET
SG Auditorium C

14:40 CET

Erasure Coding: 5 Ways to Split a Squid - Jamie Pryde, IBM
Wednesday December 4, 2024 14:40 - 14:50 CET
Ceph supports a wide range of Erasure Coding plugins and techniques. This talk will give a brief overview of the different erasure coding plugins supported by Ceph along with some interesting detail about how the various techniques for each plugin works. After giving an overview of the different plugins and techniques, we will discuss our motivations for changing the default plugin from Jerasure to ISA-L, including highlighting some of the more recent optimisations supported by ISA-L. We will also show some preliminary performance results supporting this decision.
Speakers
avatar for Jamie Pryde

Jamie Pryde

Storage Software Engineer, IBM
I have been working Ceph performance and erasure coding since the beginning of 2024, after almost 10 years of developing encryption and security solutions for the IBM FlashSystems line.
Wednesday December 4, 2024 14:40 - 14:50 CET
SG Auditorium A

14:50 CET

CephBox - a Home Cloud Personal Storage Solution - Karun Josy, IBM
Wednesday December 4, 2024 14:50 - 15:00 CET
// What if... there is a 24*7 running Ceph cluster at every home? // What is CephBox project? A DIY opensource Home cloud NAS solution with Ceph as the backbone. Why CephBox? A product must be a solution to real world problem: Cloud storage for personal use is 'expensive' and 'superfluous', whereas the external Hard drives and USB stick which we use as alternative have a short 'shell life'; waiting for the memories to be lost forever. Although personal data doesn't have high monetary value associated to it like enterprise business data, it has tremendous emotional value. There should be a de-centralized hybrid solution for personal data storage without cutting off the benefits of cloud storage. I would like to present my research and experience of building a small but efficient Ceph cluster on Raspberry pi servers to manage the storage needs of my family and myself. Agenda : Building a Ceph cluster on Raspberry pi servers. Why is CephBox better than other NAS alternatives. How to use S3 apps to store and automatically backup photos from the phone in the home network. How to create a 3-2-1 storage solution with CephBox which guarantee lifetime longevity of your personal data.
Speakers
avatar for Karun Josy

Karun Josy

IBM L3 Ceph Engineer, IBM
I'm Karun Josy, I started my Ceph journey around 2016 at the time of release of Jewel version, when I designed and deployed Ceph clusters as the storage backend for a ServerHosting company. Then I joined Red Hat as a technical support Engineer for Ceph and ODF before moving to Ceph... Read More →
Wednesday December 4, 2024 14:50 - 15:00 CET
SG Auditorium A
  Lightning Talk

15:05 CET

Remote Replication in MicroCeph: RBD and Beyond - Utkarsh Bhatt, Canonical
Wednesday December 4, 2024 15:05 - 15:15 CET
Remote replication (for block, file, and object workload) is a highly desirable feature for backup, migration, and disaster recovery. Ceph offers a highly capable but non-homogenous user experience for remote replication across different workloads (RBD mirroring, CephFS mirroring, and RGW Multisite). The Squid release of MicroCeph introduces a new set of APIs that expose standardized procedures for remote cluster awareness and remote replication for the mentioned Ceph workloads. This lightning talk will highlight implementation details while demoing RBD remote replication in MicroCeph and plans for CephFs and RGW remote replication roadmap.
Speakers
UB

Utkarsh Bhatt

Software Engineer, Canonical
Utkarsh Bhatt is a Software Engineer in the Ceph Engineering team. His team is responsible for producing the packages, charms, snaps, rocks and everything in between for the Canonical Ceph storage solutions. He graduated in 2020, and joined Canonical in May, 2022 after working for... Read More →
Wednesday December 4, 2024 15:05 - 15:15 CET
SG Auditorium B

15:05 CET

Exploring RocksDB in RGW: How We Manage Tombstones - Sungjoon Koh, LINE Plus
Wednesday December 4, 2024 15:05 - 15:40 CET
LINE, a global mobile messenger, has adopted Ceph as its main object storage. It is used to store different kinds of data, such as log files and application data. Thanks to its scalability, billions of objects are stored in our clusters. However, over time, object deletions lead to the accumulation of tombstones in RocksDB, resulting in delays during iteration. Slow iteration not only impacts LIST operation but also stalls subsequent requests. To address this issue, we first collected RocksDB metric called "skip count", which indicates the total number of tombstones detected during iterations. We then deployed a new job which compacts OSDs with high skip counts to prevent stalls. Additionally, we analyzed the pattern of tombstones and found out that a few prefixes account for over 80% tombstones, throughout the entire OSD. Based on this observation, we propose range-based compaction. In this presentation, we will first explain the basics of RocksDB and its role in Ceph Object Storage. Then, we will share our experience how we handled the RocksDB issue. Lastly, we will discuss our proposal for range-based compaction, which could further optimize overall system performance.
Speakers
avatar for Sungjoon Koh

Sungjoon Koh

Cloud Storage Engineer, LINE Plus
Sungjoon Koh is a cloud storage engineer at LINE Plus Corporation, focusing on object storage and NVMe-oF-based block storage services. His current interests include enhancing Ceph's compatibility with the S3 standard and developing object migration features. Before joining LINE Plus... Read More →
Wednesday December 4, 2024 15:05 - 15:40 CET
SG Auditorium C

15:05 CET

RBD in Squid and Beyond - Ramana Krisna Venkatesh Raja, IBM Canada Ltd & Prasanna Kumar Kalever, IBM
Wednesday December 4, 2024 15:05 - 15:40 CET
This talk will provide an overview of the new features and notable improvements in Ceph's block device component, RBD, in the Squid release. We will discuss topics such as the new feature to mirror RBD groups, improvements in live-migrating RBD images, various performance optimizations in RBD, and improved support for Windows. The session will also cover what's next for RBD in the Tentacle release. The goal is to keep new and experienced RBD users up-to-date with the latest that RBD has to offer and future plans for RBD.
Speakers
avatar for Prasanna Kumar Kalever

Prasanna Kumar Kalever

IBM, Software Architect, IBM
Prasanna Kumar Kalever works as a Software Architect at IBM and is a member of Ceph RBD team. Ex-RedHatter, Author of block storage support on Gluster which kick started RedHat's Openshift Data Foundation, also instrumental to its integration with Kubernetes. His contributions include... Read More →
avatar for Ramana Raja

Ramana Raja

Senior Software Engineer, IBM Canada Ltd
I am a developer working on Ceph's RBD component with a focus on RBD mirroring. I have made numerous code contributions to RBD and CephFS components of the Ceph project. I was previously the maintainer of the CephFS's driver for the Open Stack Manila project. I've also contributed... Read More →
Wednesday December 4, 2024 15:05 - 15:40 CET
SG Auditorium A

15:40 CET

Break
Wednesday December 4, 2024 15:40 - 16:00 CET
Wednesday December 4, 2024 15:40 - 16:00 CET
Lower Foyer

16:00 CET

Ceph at 20 Years! Still the Best for Modern Storage - Dan van der Ster, CLYSO
Wednesday December 4, 2024 16:00 - 16:35 CET
This talk explores why Ceph is the best software-defined storage solution available, highlighting its evolution since 2004 and its leadership over today's alternatives. Ceph uses innovative technologies to stay relevant. It offers block, object, and file storage through a unified system, reducing complexity and management overhead. CRUSH and Placement Groups provide scalability and resilience, allowing Ceph clusters to span hardware generations without disruptive migrations. BlueStore enhances performance with flexible replication and erasure coding. Stretch clusters and mirroring enable robust disaster recovery. Scale-out metadata in RGW and CephFS support performance for AI workloads. Ceph remains trendy with top integrations for on-premises cloud platforms, thanks to its pluggable architecture and community contributions. It's free, open-source, vendor-free, and easily installable with orchestration tools. With competition closing in, Ceph's community must innovate to stay ahead. We'll offer insights from today's toughest storage requirements, suggesting technical evolutions for the OSD, RGW, and MDS to keep Ceph at the forefront of software-defined storage solutions.
Speakers
avatar for Dan van der Ster

Dan van der Ster

CTO, CLYSO
Dan is CTO for CLYSO, developing and supporting solutions with Ceph, open infrastructure, and cloud native products and services. Dan contributes to the open source Ceph Foundation and community as Executive Council Member since 2021 and Board Member since 2015. Previously Dan was... Read More →
Wednesday December 4, 2024 16:00 - 16:35 CET
SG Auditorium A
  Session Presentation
  • Audience Level Any

16:00 CET

Revisiting Ceph's Performance After 4 Years - Wido den Hollander, Your.Online
Wednesday December 4, 2024 16:00 - 16:35 CET
As new generations of hardware become available and Ceph is improved, how does it's performance change? If we look back 4 years, how did Ceph's performance improve (or not)?
Speakers
avatar for Wido den Hollander

Wido den Hollander

CTO, Your.Online
Wido has been a part of the Ceph community for over 10 years. Long time user, developer and advocate of the future of storage. He has worked as Ceph consultant and trainer and is now CTO of Your.Online, a European-based hosting group with companies throughout Europe and a large Ceph... Read More →
Wednesday December 4, 2024 16:00 - 16:35 CET
SG Auditorium C

16:40 CET

Ceph Manager Module Design and Operation, an in-Depth Review - Brad Hubbard, Redhat & Prashant Dhange, IBM Canada Ltd.
Wednesday December 4, 2024 16:40 - 17:15 CET
This session will cover overall ceph manager design and operational aspects of the ceph MGR daemon. We will begin by giving an introduction to the MGR architecture, move on to discussing functionality of the mgr DaemonServer, mgr client, python module registry, base mgr module, and loading and unloading of the mgr modules. We will then move on to discuss module debugging, an example of GIL deadlock debugging, and how to troubleshoot MGR bugs and plugin issues. Finally, we discuss new features including tracking mgr ops and further improvements planned for future releases.
Speakers
avatar for Prashant Dhange

Prashant Dhange

Ceph rados core engineer, IBM Canada Ltd.
With 15+ years of experience in storage and cloud computing, Prashant is a experienced professional with a strong background in system programming. Prashant's focus lies in developing and optimizing storage solutions, particularly through his in-depth work with Ceph RADOS, a pivotal... Read More →
avatar for Brad Hubbard

Brad Hubbard

Principal Software Engineer, Redhat
Involved in supporting and contributing to the ceph project for well over ten years. Most recently as a RADOS core engineer working on features and bugs, both upstream and down, as well as advocating for the customer and expediting their issues internally. I have a passion for complex... Read More →
Wednesday December 4, 2024 16:40 - 17:15 CET
SG Auditorium C

16:40 CET

Maximizing the Value of Your Rados Gateway with Ingress Strategies - Michaela Lang, Red Hat & Daniel Parkes, IBM
Wednesday December 4, 2024 16:40 - 17:15 CET
Drawing from the insights gained from attending Ceph Days 2022, Customers previous talks, and my personal experience with ServiceMesh deployments, I have observed that many organizations need help implementing a cluster-wide rate limiting and metrics visibility across their RGW and Buckets. We discover a range of exciting use cases as we explore Envoy's capabilities in rate limiting, filtering, and collecting metrics on S3 activities executed against RGW. These include header-based filtering for multi-region deployments, additional OAuth token enforcement, and the ability to monitor user-to-bucket metrics without post-processing command outputs at ease. To illustrate these capabilities, I will lead a hands-on lab demonstration, showcasing Envoy's frontend role for RGW on various use cases. Rate limiting S3 requests per client Rate limiting S3 requests per region Rate limiting S3 requests per user/address/bucket (more granular level of control) Utilize Prometheus for metrics collection and monitoring Examine the use of geo-regional traffic flow scenarios in RGW Examine the use of traffic stream replication for disaster recovery scenarios in RGW
Speakers
avatar for Michaela Lang

Michaela Lang

Ms, Red Hat
Born in Vienna in 1977 started with Red Hat Linux 6 in 99 and managed to put my fingers on nearly all technoligies now landed at Red Hat where I even get paid for doing things I love to do.
avatar for Daniel Parkes

Daniel Parkes

IBM Storage Ceph Technical Product Manager, IBM
Daniel Parkes has been a die-hard Infrastructure enthusiast for many years with a massive passion for open-source technologies and a keen eye for innovation. Daniel is working in the IBM Storage Ceph Product Management team, focusing on the IBM Storage Ceph Object Storage offering... Read More →
Wednesday December 4, 2024 16:40 - 17:15 CET
SG Auditorium A
  Session Presentation
  • Audience Level Any

16:40 CET

The Challenge of Storing Small Objects on a Large Scale - Luis Domingues & Ján Senko, Proton AG
Wednesday December 4, 2024 16:40 - 17:15 CET
As an online sevices provider, storage is a critical part of Proton. With customers all around the world, exchanging e-mails, and backing up their data, the storage stack needs to be accessible 24/7. In this talk we will share the challeges of managing some Ceph clusters to serve those customers. How we manage 100PiB of small objects accross 6'000+ OSDs. Some experiances we tried with OMAP. And what we do to always be online.
Speakers
avatar for Luis Domingues

Luis Domingues

Storage Engineer, Proton AG
Luis Domingues graduated from HES-SO on distributed IT systems. After a few years at Kudelski group, he joint proton where he works now as a storage engineer.
avatar for Ján Senko

Ján Senko

Head of Storage, Proton AG
Ján has founded the Storage department at Proton, pioneered Ceph and is responsible for several types of data Storage encompassing more than 100PB of data. Luis is a Ceph Engineer responsible for keeping our production Ceph clusters running smoothly.
Wednesday December 4, 2024 16:40 - 17:15 CET
SG Auditorium B

17:20 CET

User Feedback: Shaping the Future of Ceph - Dan van der Ster, CLYSO
Wednesday December 4, 2024 17:20 - 17:55 CET
Join us for an interactive Birds of a Feather session where Ceph users come together with developers to share their experiences. This is your chance to provide direct feedback on what’s working well and what’s not in your Ceph deployments. Whether you’ve had successes or faced challenges, your insights are crucial. Together, we’ll discuss concrete technical ideas for improvements and collaborate on shaping the future of Ceph. Bring your experiences, pain points, and suggestions to help make Ceph even better!
Speakers
avatar for Dan van der Ster

Dan van der Ster

CTO, CLYSO
Dan is CTO for CLYSO, developing and supporting solutions with Ceph, open infrastructure, and cloud native products and services. Dan contributes to the open source Ceph Foundation and community as Executive Council Member since 2021 and Board Member since 2015. Previously Dan was... Read More →
Wednesday December 4, 2024 17:20 - 17:55 CET
SG Auditorium A
  Birds of a Feather
  • Audience Level Any

17:20 CET

Improving Ceph Economics with QAT Hardware Offload - Philip Williams, Canonical
Wednesday December 4, 2024 17:20 - 17:55 CET
Ceph, the words most popular open source software defined storage system, has offered storage efficiency features such as block device compression, object compression and server-side object encryption for a number of releases. However, enabling these features has always come as a trade-off between the additional performance required (in terms of cores/GHz) vs the raw storage cost, ultimately driving users away from these features. In this talk we will walk through several different scenarios where Intel's QAT offload is used to enable these features without significant overhead to primary processing, and still yields greater performance without causing increased cost per GB.
Speakers
avatar for Philip Williams

Philip Williams

Product Manager, Canonical
Philip is a Product Manager at Canonical responsible for Ceph and other storage solutions. He has over 18 years experience in the storage industry, having previously been responsible for storage infrastructure and products at a number of leading technology companies.
Wednesday December 4, 2024 17:20 - 17:55 CET
SG Auditorium B

17:20 CET

Supporting 3 Availability Zones Stretch Cluster - Kamoltat (Junior) Sirivadhna, IBM
Wednesday December 4, 2024 17:20 - 17:55 CET
A Ceph cluster stretched across 3 zones faces a potential scenario where data loss can occur due to unforeseeable circumstances. An example of such a scenario is when we have 6 replicas spread across 3 datacenters with a min_size of 3 and the setup is intended to prevent I/O from happening when there is only 1 datacenter available, however, there is an edge case where a placement group (PG) becomes available due to a lack of safeguarding during the process of temporary PG mappings in order ensure data availability. This scenario poses a risk when the sole surviving data center accepts writes, and then suddenly the 2 unavailable data centers come back up. At the same time, the surviving data center suddenly goes down, which means we would have a data loss situation. To prevent such a scenario from happening, we created a solution that utilizes an existing feature in stretch mode that would restrict how we choose the OSDs that would go into the acting set of a PG. This talk will take a deep dive into how this feature is implemented in the latest Ceph upstream as well as other features that improve the user experience with stretch cluster in the latest Ceph upstream release.
Speakers
avatar for Kamoltat (Junior) Sirivadhna

Kamoltat (Junior) Sirivadhna

Software Engineer RADOS, IBM
Junior has been a Ceph contributor for 4 years, some of his work includes enhancing Stretch Mode/ Stretch Cluster features in Ceph and improving the PG auto scaler module. Furthermore, he also contributes to the enhancement of Teuthology, a Ceph Integration testing framework that... Read More →
Wednesday December 4, 2024 17:20 - 17:55 CET
SG Auditorium C
  Session Presentation
  • Audience Level Any

18:00 CET

Attendee Reception
Wednesday December 4, 2024 18:00 - 19:30 CET
Wednesday December 4, 2024 18:00 - 19:30 CET
Solutions Showcase
 
Thursday, December 5
 

08:00 CET

Registration + Badge Pick-up
Thursday December 5, 2024 08:00 - 18:00 CET
Thursday December 5, 2024 08:00 - 18:00 CET
Lower Foyer

09:00 CET

Keynote Sessions: To Be Announced
Thursday December 5, 2024 09:00 - 09:55 CET
Thursday December 5, 2024 09:00 - 09:55 CET
SG Auditorium A

10:00 CET

Keynote: Cephalocon Developers AMA: Meet the Minds Behind Open Source Ceph
Thursday December 5, 2024 10:00 - 10:30 CET
Thursday December 5, 2024 10:00 - 10:30 CET
SG Auditorium A

10:30 CET

Break
Thursday December 5, 2024 10:30 - 11:00 CET
Thursday December 5, 2024 10:30 - 11:00 CET
Lower Foyer

10:30 CET

Solutions Showcase
Thursday December 5, 2024 10:30 - 16:00 CET
Thursday December 5, 2024 10:30 - 16:00 CET
Lower Foyer

11:00 CET

Ceph Telemetry - the Why, What, and How - Yaarit Hatuka, IBM
Thursday December 5, 2024 11:00 - 11:35 CET
Whether you are a Ceph user or a developer, you have probably wondered at some point: How many Ceph clusters are out there? What Ceph versions are they running? What does their storage capacity distribution look like? Answers to these questions and more are available thanks to Ceph’s telemetry module. In this session, we will deep dive into this module and explore the value it brings to users and developers alike.
Speakers
avatar for Yaarit Hatuka

Yaarit Hatuka

Senior software engineer, IBM
Senior software engineer at IBM working on Ceph telemetry and observability
Thursday December 5, 2024 11:00 - 11:35 CET
SG Auditorium A
  Session Presentation
  • Audience Level Any

11:00 CET

Migrating AI/ML Data Pipelines for Real Time Streaming Applications from HDFS to Ceph - Vahideh Alinouri, Etraveligroup
Thursday December 5, 2024 11:00 - 11:35 CET
Etraveligroup’s AI/ML team has migrated data pipelines from HDFS to Ceph to ensure efficient management and performance. To achieve this, they deployed Ceph RGW with a replicated data pool on NVMe disks to meet the technical demands and automation needs. The request flow involves migrating real time streaming applications and training datasets for machine learning that assist with real time sales analytics and automated pricing. The crucial migration steps include transferring production pipelines from HDFS to in-house object storage. Precise tuning is necessary to manage c-states, power saving, and IOMMU on physical servers and preventing write locking during the dynamic resharding process. Key factors such as failure domain, co-locating RGWs with OSDs or MONs, beast thread pool size, memory profiling, mclock profile, and the number of RGW services are essential for achieving the target metrics and ensuring a successful migration.
Speakers
avatar for Vahideh Alinouri

Vahideh Alinouri

Cloud engineer, Etraveli Group AB
Vahideh Alinouri is a well-rounded Cloud Engineer with a strong background in cloud computing and academic experience. At EtraveliGroup, a leading company for tech solutions and fulfillment capabilities for online sales of flight, she designed and supported their private cloud infrastructure... Read More →
Thursday December 5, 2024 11:00 - 11:35 CET
SG Auditorium B

11:00 CET

SWITCH: Operations, Data Management and Automation - Theofilos Mouratidis, SWITCH
Thursday December 5, 2024 11:00 - 11:35 CET
SWITCH is the national research and education network (NREN) of Switzerland, a non-profit organisation that provides services to the universities and schools of the country. In the storage circle of the cloud team at SWITCH, we maintain and procure Ceph clusters mainly for S3. We have 3 iterations that differ in terms of automation and features namely OSv1/2/3. We currently develop the latest iteration using Ansible in a GitOps way, where the code is the source of truth and changes to the code automatically deploy configuration changes to various clusters. In this session, we will talk about the OSv3 Ansible collection and configuration management repos, where from defining an inventory that looks like the `ceph orch host ls` output and a short yaml file, we can immediately bootstrap clusters that connect together and provide multisite S3, without any manual steps. Now that we deploy our new clusters using the new technologies, we are in the migration phase where we try to maintain the old dying clusters (OSv1/2) and slowly migrate S3 data to the new ones with minimal or no user intervention.
Speakers
avatar for Theofilos Mouratidis

Theofilos Mouratidis

Cloud Engineer, SWITCH
My name is Theofilos Mouratidis and I am from Greece. I am currently a cloud engineer at SWITCH. I have a strong theoretical background and research interest in distributed systems. In the past I have worked for CERN and Proton in similar positions. I enjoy the sunny weather and go... Read More →
Thursday December 5, 2024 11:00 - 11:35 CET
SG Auditorium C

11:40 CET

Advancing BlueStore with Real-World Insights - Adam Kupczyk, IBM
Thursday December 5, 2024 11:40 - 12:15 CET
In past years we have invested significant effort to improve BlueStore's IO latency and throughput. Testing, including aging, have always done using artificial workloads. Obviously we optimized for those scenarios. Now we want to open new chapter in BlueStore maturity. Whenever possible we will use real-life workloads, provided by Ceph users. We will test new components and new proposed settings against those workloads. Testing aging will be augmented by shortcuts that will get the aging process complete faster. The ultimate goal is to preserve high performance new deployments enjoy as long as possible. We want to share this plan with community, get developers involved and convince users to share their workloads.
Speakers
avatar for Adam Kupczyk

Adam Kupczyk

Mr., IBM
Mathematician by education. Engineer and programmer by job. Tester by necessity. Graduated Adam Mickiewicz University, Poznan. 25 years in software development.
Thursday December 5, 2024 11:40 - 12:15 CET
SG Auditorium C

11:40 CET

Ceph Operations Retrospective: Lessons from Scaling and Managing a Multi-PB Cluster - Mattias Skohg, Engin IT Nordic AB
Thursday December 5, 2024 11:40 - 12:15 CET
This talk will cover our experiences managing and scaling a Ceph cluster (pacific) from 4.5PB to 9PB) (raw) while in production, primarliy using CephFS and RGW. We will go through the cluster design and hardware, operational challenges, bugs hit, decisions made and how we resolved them.
Speakers
avatar for Mattias Skohg

Mattias Skohg

CTO, Engin IT Nordic AB
Mattias is the co-founder of Engin IT Nordic AB and a senior storage consultant. He has designed, installed and in many cases managed clusters with a combined capacity of over over 50PB.
Thursday December 5, 2024 11:40 - 12:15 CET
SG Auditorium A
  Session Presentation
  • Audience Level Any

11:40 CET

D4N in RGW - Pritha Srivastava, IBM
Thursday December 5, 2024 11:40 - 12:15 CET
D4N is an intelligent caching framework within Ceph RGW. This project is a collaboration between the MOC team at Boston University and Ceph RGW team. This talk outlines the brief history of the D4N Research Project and its use cases. Then it will discuss the current ongoing effort to upstream D4N into Ceph which will include its architecture based on the Zipper API in RGW and its use-cases for s3, analytics workload acceleration and use cases in the AI/ML domain.
Speakers
avatar for Pritha Srivastava

Pritha Srivastava

Software Architect, IBM
I am a Software Developer with IBM and have been working on Ceph RGW for the last 8 and a half years. Prior to that I worked with HP for StoreVirtual. I have completed my MS from University of Colorado at Boulder.
Thursday December 5, 2024 11:40 - 12:15 CET
SG Auditorium B

12:15 CET

Lunch
Thursday December 5, 2024 12:15 - 13:45 CET
Thursday December 5, 2024 12:15 - 13:45 CET
Globe of Science and Innovation

13:45 CET

Archive Zone: Lessons Learned - Ismael Puerto, INDITEX
Thursday December 5, 2024 13:45 - 14:20 CET
In this session, we will delve into the history and evolution of our Ceph clusters dedicated to the archive zone in production. We'll cover the entire journey, from the initial hardware selection to the deployment, and share the critical lessons we've learned along the way. Key topics include: Hardware Selection: How we chose the right hardware for our archive zone, including considerations and trade-offs. Common Mistakes: The pitfalls and mistakes we encountered during the deployment process, and how we overcame them. Best Practices: Steps and strategies to ensure a successful deployment, focusing on reliability, scalability, and performance. Optimization Tips: Techniques to optimize your Ceph cluster for archival purposes, ensuring efficient storage and retrieval of data. By the end of this talk, you will have a comprehensive understanding of the challenges and solutions involved in deploying a Ceph archive zone, enabling you to avoid common pitfalls and achieve a successful implementation in your environment.
Speakers
avatar for Ismael

Ismael

Solution Architect, INDITEX
I am a Solution Architect at Inditex, responsible for operating and evolving services based on Ceph and Kubernetes. I have been working with Ceph for six years, handling all types of storage: Object, Block, and FileSystem. My top priorities are maintaining resilience, performance... Read More →
Thursday December 5, 2024 13:45 - 14:20 CET
SG Auditorium C

13:45 CET

Understanding Ceph One Performance Counter at a Time - Marcel Lauhoff, Clyso GmbH
Thursday December 5, 2024 13:45 - 14:20 CET
Let’s demystify Ceph performance counters by following the journey of data through the system, from the initial client I/O request to the moment it reaches the disk. We'll introduce performance counters, explaining what they are, how they work, and how to access them effectively. Using a simple workload, we'll analyze key Ceph performance counters to illustrate how data flows through various subsystems and what these metrics reveal about system performance. By understanding these metrics, attendees will gain insights into diagnosing issues and optimizing their Ceph deployments.
Speakers
avatar for Marcel Lauhoff

Marcel Lauhoff

Staff Software Engineer, Clyso GmbH
Marcel Lauhoff is a seasoned software engineer with nearly a decade of experience in software-defined storage. His Ceph journey began with his Master's thesis, which focused on adapting Ceph for use in archives. He previously worked on the S3GW project, a standalone cloud S3 service... Read More →
Thursday December 5, 2024 13:45 - 14:20 CET
SG Auditorium B

13:45 CET

Volume Shadow Copies with CephFS: A Samba-Centric Approach - Shwetha Acharya, IBM
Thursday December 5, 2024 13:45 - 14:20 CET
In this talk, we delve into the integration of snapshots as shadow copies within CephFS, emphasizing their utilization and management from a Windows client perspective through Samba. The focus will be on the vfs_ceph_snapshots module, which facilitates the seamless presentation of CephFS snapshots as volume shadow copies to Windows clients. This capability is effective both in environments using kernel mounts and those employing the ceph VFS module, offering a flexible approach to snapshot management. We will conduct practical demonstrations to showcase the process of creating, accessing, and restoring these snapshots, while also evaluating the observations made. Furthermore, the presentation will outline future plans aimed at enhancing the reliability and versatility of this functionality, ensuring it meets the evolving needs of users.
Speakers
avatar for Shwetha Acharya

Shwetha Acharya

Miss, IBM
I work for the CephFS Engineering team at IBM, where I focus on SMB integration with CephFS. I also maintain the Geo-Replication component of GlusterFS.
Thursday December 5, 2024 13:45 - 14:20 CET
SG Auditorium A

14:25 CET

Ceph Notification at Scale: Notification V2, an Enterprise-Level Feature - Krunal Chheda, Bloomberg
Thursday December 5, 2024 14:25 - 15:00 CET
Data powers Bloomberg’s financial products. Ceph clusters are the backbone of Bloomberg’s internal S3 cloud storage systems, which host this data and serve billions of requests a day. Given the importance of these cloud storage systems to the company’s broader infrastructure, it is crucial that any S3 feature we enable must be an enterprise-level feature that scales, provides telemetry, is secure, and has fault-tolerant capabilities. Working together with the community, Bloomberg has enhanced the existing notification feature built into Ceph enabling it to be used by Bloomberg and other enterprise users of Ceph. This talk will discuss all the work that went into building “notification_v2,” the new enterprise-class S3 notifications platform and all of its new features.
Speakers
avatar for Krunal Chheda

Krunal Chheda

Senior Software Engineer, Bloomberg
Krunal Chheda is a Senior Software Engineer at Bloomberg and is part of the Storage Engineering team in the company's Technology Infrastructure department. He has worked in the industry for more than 12 years and holds a Masters degree in Software Engineering from BITS Pilani India... Read More →
Thursday December 5, 2024 14:25 - 15:00 CET
SG Auditorium B

14:25 CET

State of CephFS: Three Easy Pieces - Venky Shankar, IBM & Patrick Donnelly, IBM, Inc.
Thursday December 5, 2024 14:25 - 15:00 CET
This talk focusses on the current (and near future) state of the three pieces that make up a Ceph File System - Ceph Metadata Sever (MDS), Clients and a set of Ceph Manager Plugins. Much advancements have been made to the Ceph File System recently, opening up gateways for wider adoption. Some features are already available in recent releases and some are under development. We detail these enhancements by breaking up nicely into each of the three pieces. Ceph File System specific manager plugins have come a long way to now becoming the de-facto for subvolume/crash-consistent snapshot management and mirroring. We discuss about those. And finally, we peek into what is upcoming in CephFS for Tentacle ("T") release. Existing and new CephFS users would find it helpful to assess and plan ahead for its adoption.
Speakers
avatar for Patrick Donnelly

Patrick Donnelly

Software Architect, IBM, Inc.
Patrick Donnelly is a Software Architect at IBM, Inc. working as part of the global development team on the open source Ceph distributed storage system. Patrick has principally worked on the Ceph file system (CephFS) since 2016. He has been working on Open Source projects for the... Read More →
avatar for Venky Shankar

Venky Shankar

CephFS PTL, IBM
I have worked in distributed file systems for over a decade. Currently leading the Ceph File Systems team and part of the Ceph Leadership Team.
Thursday December 5, 2024 14:25 - 15:00 CET
SG Auditorium C
  Session Presentation
  • Audience Level Any

15:05 CET

Benchmarking: Repeatable & Comparable - Trent Lloyd, Canonical (Ubuntu)
Thursday December 5, 2024 15:05 - 15:40 CET
Your goal when benchmarking should be to ensure that the results are both continuously repeatable and fairly comparable to previously attempts. This is all too easy to get wrong. Benchmarking of any kind often presents tricky business, but storage has always presented particularly difficult challenges as even the simple Hard Drive has interesting performance characteristics that vary greatly depending on the workload or even chance. You might hope that was solved by SSDs, and that is true to an extent for real workloads, but they tend to give even more misleading results during synthetic benchmarks. I'll work through many different causes of inconsistent results in benchmarking both individual components and over-all performance of a Ceph cluster, with specific examples and graphs of real attempts. Items covered include - Working set size - Bi-modal SSD performance due to flash block management - Thin provisioning - Bandwidth limitations of SSDs, Backplanes, PCIe buses, CPUs, Memory and Networks - Filesystems - Caches of all kinds - Inconsistencies from benchmarking freshly deployed Ceph clusters - Benchmarking tools (Don't use anything other than fio, aws-cli is slow) - And more
Speakers
avatar for Trent Lloyd

Trent Lloyd

Mr, Canonical (Ubuntu)
Trent Lloyd is a long time passionate speaker and member of the Linux & Open Source community, having first presented at the age of 15 to an audience at linux.conf.au 2003. He has spent the last 9 years in the Ubuntu Support Team at Canonical as a Sustaining Software Engineer specialising... Read More →
Thursday December 5, 2024 15:05 - 15:40 CET
SG Auditorium C

15:05 CET

From ISCSI to NVMe-of: A Comparative Look on Storage Protocol Advantages and Challenges - Andy Muthmann, croit GmbH
Thursday December 5, 2024 15:05 - 15:40 CET
We’d like to share our experience with NVMe-oF: * What deployment options are available for NVMe-oF, how they work and what advantages/disadvantages they have * Performance of NVMe-oF compared to iSCSI and raw kernel rbd * Issues with iSCSI and how NVMe-oF fits here * NVMe-of Multipathing
Speakers
avatar for Andy Muthmann

Andy Muthmann

CEO, croit GmbH
Thursday December 5, 2024 15:05 - 15:40 CET
SG Auditorium B

15:40 CET

Break
Thursday December 5, 2024 15:40 - 16:00 CET
Thursday December 5, 2024 15:40 - 16:00 CET
Lower Foyer

16:00 CET

Arming Ceph: CI Enhancement and UADK Integration - Rongqi Sun, HUAWEI TECHNOLOGIES CO., LTD.
Thursday December 5, 2024 16:00 - 16:10 CET
The ARM ecosystem is becoming increasingly robust, and ensuring the stability and high performance of Ceph on the ARM platform is crucial. Due to some reasons, the ARM CI has been almost abandoned in recent years. Over the past year, Rongqi's team has been actively fixing environment and software issues to maintain Ceph's operation on ARM. The team also continuously monitors Ceph's performance on ARM. In this Lightning Talk, he will showcase the achievements on the ARM CI up to now. Additionally, in the performance section, he will discuss how Ceph's integration with UADK (User Space Accelerator Development Kit, a general-purpose user space accelerator framework that uses shared virtual address (SVA) technology to provide a unified programming interface for hardware acceleration computing cryptography and compression algorithms) enables compression under low CPU load. The T version will include this by default.
Speakers
avatar for Rongqi Sun

Rongqi Sun

Software Engineer, HUAWEI TECHNOLOGIES CO., LTD.
Working at Huawei, Ceph Arch Maintainer, dedicated to adapting and optimizing Ceph for ARM.
Thursday December 5, 2024 16:00 - 16:10 CET
SG Auditorium B
  Lightning Talk
  • Audience Level Any

16:00 CET

Get That Cluster Back Online, but Hurry Slowly - Gregory Orange, Pawsey Supercomputing Centre
Thursday December 5, 2024 16:00 - 16:30 CET
Pawsey offers supercomputing, cloud and data services to Australian researchers, and our Petabyte-scale Ceph RGW clusters have become the central storage solution no matter which combination of services is required. Ceph is both complicated and resilient, meaning there are many things which can go wrong. However, with a suitable design and in the right hands the data is protected and the service can be brought back online. I would like to tell you about the months-long situation leading to a significant outage of our largest cluster, problems we encountered, how we discovered and fixed them, who helped along the way, and the lessons we have recorded for the future. These lessons relate to chasing masked errors, technology choices, team resources and community engagement.
Speakers
avatar for Gregory Orange

Gregory Orange

Gregory Orange, Pawsey Supercomputing Centre
Greg deploys and runs Ceph and OpenStack for scientists in Australia, and is involved in the floor tiles, the RGW nodes, the teams who look after it all, and nearly everything in between. Automation and efficiency are of great interest to him, and so is the well-being of all the people... Read More →
Thursday December 5, 2024 16:00 - 16:30 CET
SG Auditorium C

16:00 CET

Ceph Made Easy: One Dashboard for Multiple Ceph Clusters - Nizamudeen A, IBM India Private Ltd
Thursday December 5, 2024 16:00 - 16:35 CET
The presentation is about a solution that we have created in the Ceph Dashboard for managing and monitoring multiple ceph clusters from a single cluster that we call a hub cluster. This approach simplifies the complexities of managing multiple clusters by providing a more streamlined and efficient user experience. I will describe the architecture of our implementation and how this would help admins to manage many clusters, ensuring optimal performance, reliability and ease of use. I will also demo various features which can leverage the multi-cluster setup like setting up replication between multiple clusters. Also with multiple clusters connected on a single cluster, this will also provide an overview dashboard where important information on the other clusters can be monitored including the real-time alerts that goes on in the other clusters. I’ll also share how we are planning to improve the feature and our testing strategies around it.
Speakers
avatar for Nizamudeen A

Nizamudeen A

Software Engineer, IBM India Private Ltd
Software Engineer and component lead of Ceph Dashboard. I started 5 years ago as an intern at Red Hat contributing to Rook Operator. Eventually moved into Ceph Dashboard and started looking into the usability improvements and implementing workflows in the UI. Later picked up the lead... Read More →
Thursday December 5, 2024 16:00 - 16:35 CET
SG Auditorium A

16:10 CET

Want to Become a Ceph Ambassador? - Gaurav Sitlani, IBM
Thursday December 5, 2024 16:10 - 16:20 CET
The Ceph Ambassador program recognizes individuals in the community who are participating and helping with community activities and engaging with other community members. They are an amazing group of Ceph experts and community leaders in various regions across the world who are available to support the Ceph community within their region. This talk will cover an Overview of the Ceph Ambassador program and its goals. I’ll be sharing about the various events and initiatives involving the ambassadors to inspire and drive future activities, community growth, engagement, and development collecting feedback from the community. It has been a nice year in the Ceph community with Ceph Days being successfully organized across the world. The Ceph Ambassadors look forward to inspiring the community and contributing towards the Ceph project's success.
Speakers
avatar for Gaurav Sitlani

Gaurav Sitlani

Ceph Community Ambassadors Lead, IBM
Gaurav Sitlani is originally from Jaipur which is also known as the "Pink City of India". Joined Red Hat as an Intern where he started working on Ceph in 2018 supporting Enterprise customers. He graduated from the University of Pune with a Bachelor's degree in Computer Engineering... Read More →
Thursday December 5, 2024 16:10 - 16:20 CET
SG Auditorium B
  Lightning Talk
  • Audience Level Any

16:20 CET

High Performance Erasure Coding - the Story so Far - Connor Fawcett, IBM
Thursday December 5, 2024 16:20 - 16:30 CET
We have been working to improve erasure coding performance, such that it performs almost as well as replica 3 configurations for all configurations, including block and file. Our aim here is to expose the wider Ceph community to the changes we're making to erasure coding and pave the way for wider adoption as we increase visibility of performance improvements over time. In this talk we will share the latest news on erasure coding improvements, highlight both merged and in progress PRs that are of interest, and include performance benchmarks on the latest experimental code.
Speakers
avatar for Connor Fawcett

Connor Fawcett

Storage Software Engineer, IBM
After almost a decade working as a storage developer on the IBM FlashSystems line of products, I have moved on to pastures new and transitioned to working on Ceph performance as of the beginning of 2024.
Thursday December 5, 2024 16:20 - 16:30 CET
SG Auditorium B
  Lightning Talk
  • Audience Level Any

16:40 CET

Pulpito 2.0: A Tour of Pulpito Next-Generation - Vallari Agrawal, IBM
Thursday December 5, 2024 16:40 - 16:50 CET
Introducing the new Pulpito interface with unique features, including job scheduling from web, one-click run termination, rerunning and sharing saved command configurations, and more! This talk will walk you through all the new capabilities and tricks of Pulpito-NG, which streamline the entire Teuthology experience.
Speakers
avatar for Vallari Agrawal

Vallari Agrawal

Software Engineer, IBM
Vallari is a Software Engineer at IBM Ceph. She is a member of Ceph NVMe-oF and Teuthology team, and is additionally a Google Summer of Code Mentor and Org Admin.
Thursday December 5, 2024 16:40 - 16:50 CET
SG Auditorium A
  Lightning Talk

16:40 CET

Cost-Effective, Dense, and Performant Prometheus Storage via QLC - Anthony D'Atri, Dreamsnake Productions
Thursday December 5, 2024 16:40 - 17:10 CET
Prometheus is the metrics ecosystem of choice for modern computing, with exporters for Ceph, RAID HBAs, Redfish, time synchronization, and the panoply provided by node_exporter. Exporters are scraped multiple times per minute for effective queries, each ingesting as many as several housand metrics per system. Data may be kept locally or in external solutions including Ceph RGW. Retention of a year or more is valuable for trending and comparisons. A moderate-size deployment can easily fill tens or hundreds of terabytes. As retention and cardinality grow, so does processing. Prometheus will GC and flush its WAL every two hours, which can manifest visible yet spurious artifacts in visualization tools like Grafana and false alarms from alertmanager rules. Rotational media just don't cut it. While HDDs of capacities as large as 30TB are available, rotational + seek latencies, SATA stenosis, interminable resilvering, and SMR severely limit their viability. SSDs are increasingly viable as HDD replacements. We can improve cost and density by tailoring to the workload: intermittent sequential writes, and frequent random reads. This is a classic workload for modern QLC SSDs.
Speakers
avatar for Anthony D'Atri

Anthony D'Atri

Principled Engineer, Dreamsnake Productions
Anthony has run Ceph at scale for over eleven years and one can say that it is literally a part of him - ask him in person and he'll show you why. He is also an observability advocate and contributes daily to Ceph documentation and may be open to new opportunities.
Thursday December 5, 2024 16:40 - 17:10 CET
SG Auditorium C
  Session Presentation

16:40 CET

Encrypting Data in Ceph: An Overview and Opportunities - David Mohren, 45Drives
Thursday December 5, 2024 16:40 - 17:15 CET
Encrypting data in large storage platforms like Ceph has become a standard practice in the industry. However, when attempting to utilize these features in Ceph, a typical user may find the process overwhelming. One reason may be the lack of a consolidated Ceph Docs document that explains and compares the various encryption solutions available in Ceph. This could assist Ceph users in identifying the optimal solution for their specific use case. My objective is to provide a foundational overview of data encryption in Ceph and clarify the various encryption solutions currently available within Ceph. At the end I will talk about that none of the Ceph encryption solutions give a storage end-user the ability to figure out if someone other than authorized parties have tampered with the data. Therefore I will present a method on how to solve the problem. Our solution is to integrate AEAD ciphers into Ceph (as done in Azure,Google Cloud,AWS) and use OMAP to store the resulting authentication tags. My goal is to initiate a dialogue within the community to ascertain whether the solution I have developed is applicable within the context of Ceph and to get guidance on how to proceed from here
Speakers
avatar for David Mohren

David Mohren

Research Assistant, 45Drives
I am currently a graduate student at the University of New Brunswick (Canada) and the University of Applied Sciences Bonn-Rhein-Sieg (Germany) pursuing a dual Master Computer Science Degree under the supervision of Dr. Kenneth Kent. I am working on a Ceph data security research project... Read More →
Thursday December 5, 2024 16:40 - 17:15 CET
SG Auditorium B

16:50 CET

Characterizing Crc Operation on Ceph - Rixin Luo, Huawei Technologies Co., Ltd
Thursday December 5, 2024 16:50 - 17:00 CET
A cyclic redundancy check (CRC) is an error-detecting code widely used in digital networks and storage devices to identify accidental changes in digital data. Ceph employs CRC checksums to ensure data correctness during network transmission and to maintain data integrity in Bluestore. For RBD 4K random write and random read workloads, we observed that the CRC32C operation size primarily ranges from 4 to 292 bytes. Specifically, for 4K random writes, 80% of the operations are smaller than 292 bytes, while for 4K random reads, 83% are smaller than 254 bytes. The most common sizes are 28 bytes and 41 bytes, accounting for nearly 40%; these correspond to 'preamble_block_t' and 'ceph_msg_header2', respectively. Given the characteristics of the CRC32C operation, we optimized it for smaller sizes by using inline assembly code to unfold the CRC32C function 'ceph_crc32c_aarch64'. Additionally, we optimized the encoding procedure to minimize the occurrence of small-sized data segments. We tested these optimizations on the RBD 4K random write and random read workloads, resulting in a 2% improvement in throughput.
Speakers
avatar for Rixin Luo

Rixin Luo

Senior software engineer, Huawei Technologies Co., Ltd
I am a senior software engineer from Huawei. I have worked for years on enabling and optimizing Ceph performance on aarch64.
Thursday December 5, 2024 16:50 - 17:00 CET
SG Auditorium A

17:00 CET

Async I/O with Libcephfs - Dhairya Parmar, IBM
Thursday December 5, 2024 17:00 - 17:10 CET
File system gateways like NFS-Ganesha are capable of immense throughput but can be hindered by thread-centric synchronicity. The default I/O interfaces built in libcephfs rely on threading this way which makes it difficult to achieve high performance throughput. Enter async I/O in libcephfs – a transformative approach that changes how file system gateways interact with CephFS. By leveraging async I/O, protocols can maximize resource utilization, enhance system responsiveness especially to low latency operations. In this talk, we will explore how NFS-Ganesha led to the development of async I/O in CephFS, detailing its journey from inception through various phases of improvement and stabilization. We will discuss the current state of async I/O in CephFS, including how NFS-Ganesha makes the most out of it, how it boosts performance with CephFS, and future applications of the API on other gateways. Think I/O in CephFS? Think Async I/O!
Speakers
avatar for Dhairya Parmar

Dhairya Parmar

Software Engineer, IBM
Dhairya is an open-source enthusiast currently part of the CephFS team. He primarily works on the CephFS client, MDS (Metadata Server), and the manager NFS module. Before joining Ceph, he interned at Red Hat, where he worked on converting legacy sudo-SSSD bash scripts into efficient... Read More →
Thursday December 5, 2024 17:00 - 17:10 CET
SG Auditorium A

17:20 CET

Backing up CephFS - Abhishek Lekshmanan, CERN, European Organization for Nuclear Research
Thursday December 5, 2024 17:20 - 17:55 CET
This Birds of a Feather session aims to explore best practices in the backup of CephFS deployments, encourage technical exchange and highlight where the system can evolve to improve support for this crucial use-case. We encourage participants to share their experience in this area. Discussion points - Applications: i) Which backup applications have been tried? ii) Tricks to avoid iterating on every inode, e.g., recursive mtime. - Performance and configuration: Handling/controlling impact of backups on the live system. - Snapshots: i) Impact of snapshots on the live system; ii) Snapdiff efficiency; iii) Ability to export snapshots as blobs to arbitraty backends, e.g., tape; iv) Experiences with snapshot mirroring. - Integrations: OpenStack Manila.
Speakers
avatar for Abhishek Lekshmanan

Abhishek Lekshmanan

Mr, CERN, European Organization for Nuclear Research
Abhishek is a Computing Engineer at CERN, where he has worked for the past couple of years in various aspects of development and operations on distributed storage systems for physics and non-physics use cases. Abhishek was previously involved in the Ceph project as a maintainer for... Read More →
Thursday December 5, 2024 17:20 - 17:55 CET
SG Auditorium B
  Birds of a Feather
  • Audience Level Any

17:20 CET

The ‘Scrub-Type to Limitations’ Matrix - Ronen Friedman, IBM
Thursday December 5, 2024 17:20 - 17:55 CET
The scrub ‘restrictions overrides’ matrix: Scrubs can be triggered by multiple conditions, with each trigger resulting in a specific set of scrub-session behaviors and a specific set of limitations / restrictions that apply or are overridden (operator-initiated scrub, for example, are allowed to run on whatever day of the week or hour, regardless of configuration). The matrix of ‘scrub-type to restrictions’ was never fully nor consistently documented. Starting with the ‘Reef’, through ‘Squid’, and - hopefully - finalized in ‘Tentacles’ - we are working on clarifying, documenting and implementing the desired behaviors - the desired matrix. I will present, with the goal of receiving feedback from the Ceph community, what was already released with Squid, and - more important - what changes to this matrix are planned for Tentacles. For the community, this would be a great opportunity to influence the fine details of what will be part of the next Ceph release.
Speakers
avatar for Ronen Friedman

Ronen Friedman

Software Architect, IBM
Ronen has been developing software for more than thirty years. He is a member of the Rados core team at Red Hat and now IBM for the last 5 years. Currently is the maintainer for the Ceph OSD Scrub.
Thursday December 5, 2024 17:20 - 17:55 CET
SG Auditorium C

17:20 CET

Deploy Ceph in Kubernetes: Rook Deep Dive - Travis Nielsen & Subham Rai, IBM
Thursday December 5, 2024 17:20 - 17:55 CET
Do you run Kubernetes? Rook is the answer to natively integrate Ceph with Kubernetes, with K8s workloads running in production for over six years. Rook configures Ceph to provide stable block (RWO), shared file system (RWX), and object storage (S3). We will give an overview of Rook as well as deep dive into the core configuration features including encryption, recent feature additions, and future plans for the project. Rook was accepted as a graduated project by the Cloud Native Computing Foundation in October 2020.
Speakers
avatar for Travis Nielsen

Travis Nielsen

Rook Lead Maintainer, IBM
Travis Nielsen is a Senior Technical Staff Member for IBM where he is a maintainer on Rook and member of the ODF and Ceph engineering team. Prior to IBM and Red Hat, Travis worked in storage at Quantum and Symform, a P2P storage startup, and was an engineering lead for the Windows... Read More →
avatar for Subham Rai

Subham Rai

Software Engineer, IBM
I'm Subham Rai from India. I have more than 3 years of experience in software or more specifically in the storage industry working mainly in rook-ceph. I hold a B.tech degree and I'm also RHCSA certified. I have spoken at FOSDEM 2022, Cephalocon 2023(Amsterdam).  I'm in the top 4... Read More →
Thursday December 5, 2024 17:20 - 17:55 CET
SG Auditorium A
 
  • Filter By Date
  • Filter By Venue
  • Filter By Type
  • Audience Level
  • Session Slides Attached
  • Timezone

Share Modal

Share this link via

Or copy link

Filter sessions
Apply filters to sessions.