Skip to main content

DataStax launches Kubernetes operator for open source Cassandra database

Today, DataStax, the commercial company behind the open source Apache Cassandra project, announced an open source Kubernetes operator developed by the company to run a cloud native version of the database.

When Sam Ramji, chief strategy officer at DataStax, came over from Google last year, the first thing he did was take the pulse of customers, partners and community members around Kubernetes and Cassandra, and they found there was surprisingly limited support.

While some companies had built Kubernetes support themselves, DataStax lacked one to call its own. Given that Kubernetes was born inside Google, and the company has widely embraced the notion of containerization in general, Ramji wanted there to be an operator specifically designed by the company to give customers a general starting point with Kubernetes.

“What’s special about the Kube operator that we’re offering to the community as an opinion — one of many — is that we have done the work to generalize the operator to Cassandra wherever it might be implemented,” Ramji told TechCrunch.

Ramji says that most companies that have created their own Kubernetes operators tend to specialize for their own particular requirements, which is fine, but as the company built on top of Cassandra, they wanted to come up with a general version that could appeal broader range of use cases.

In Kubernetes, the operator is how the DevOps team packages, manages and deploys an application, giving it the instructions it needs to run correctly. DataStax has created this operator specifically to run Cassandra with a broad set of assumptions.

Cassandra is a powerful database because it stays running when many others fall down. As such it is used by companies as varied as Apple, eBay and Netflix to run their key services. This new Kubernetes implementation will enable anyone who wishes to run Cassandra as a containerized application, helping push it into a modern development realm.

The company also announced a free help service for engineers trying to cope with increased usage on their databases due to COVID-19. They are calling the program, “Keep calm and Cassandra on.” The engineers charged with keeping systems like Cassandra running are called Site Reliability Engineers or SREs.

“The new service is completely free SRE-to-SRE support calls. So our SREs are taking calls from Apache Cassandra users anywhere in the world, no matter what version they’re using if they’re trying to figure out how to keep it up to stand up to the increased demand,” Ramji explained.

DataStax was founded in 2010 and has raised over $190 million, according to PitchBook data.



from TechCrunch https://ift.tt/3awwBXt
via IFTTT

Comments

Popular posts from this blog

Apple’s AI Push: Everything We Know About Apple Intelligence So Far

Apple’s WWDC 2025 confirmed what many suspected: Apple is finally making a serious leap into artificial intelligence. Dubbed “Apple Intelligence,” the suite of AI-powered tools, enhancements, and integrations marks the company’s biggest software evolution in a decade. But unlike competitors racing to plug AI into everything, Apple is taking a slower, more deliberate approach — one rooted in privacy, on-device processing, and ecosystem synergy. If you’re wondering what Apple Intelligence actually is, how it works, and what it means for your iPhone, iPad, or Mac, you’re in the right place. This article breaks it all down.   What Is Apple Intelligence? Let’s get the terminology clear first. Apple Intelligence isn’t a product — it’s a platform. It’s not just a chatbot. It’s a system-wide integration of generative AI, machine learning, and personal context awareness, embedded across Apple’s OS platforms. Think of it as a foundational AI layer stitched into iOS 18, iPadOS 18, and m...

The Silent Revolution of On-Device AI: Why the Cloud Is No Longer King

Introduction For years, artificial intelligence has meant one thing: the cloud. Whether you’re asking ChatGPT a question, editing a photo with AI tools, or getting recommendations on Netflix — those decisions happen on distant servers, not your device. But that’s changing. Thanks to major advances in silicon, model compression, and memory architecture, AI is quietly migrating from giant data centres to the palm of your hand. Your phone, your laptop, your smartwatch — all are becoming AI engines in their own right. It’s a shift that redefines not just how AI works, but who controls it, how private it is, and what it can do for you. This article explores the rise of on-device AI — how it works, why it matters, and why the cloud’s days as the centre of the AI universe might be numbered. What Is On-Device AI? On-device AI refers to machine learning models that run locally on your smartphone, tablet, laptop, or edge device — without needing constant access to the cloud. In practi...

Max Q: Psyche(d)

In this issue: SpaceX launches NASA asteroid mission, news from Relativity Space and more. © 2023 TechCrunch. All rights reserved. For personal use only. from TechCrunch https://ift.tt/h6Kjrde via IFTTT