Video Series: Lowering Costs with the Data Optimization Cycle

It’s no secret that data costs have become unruly in a cloud native world. When cardinality goes up and costs explode, teams can end up feeling like they’re flying blind, while companies feel the effects of squeezed budgets.

In this first episode of our new series, Chronologues, copywriter at Chronosphere Sophie Kohler is joined by observability experts at Chronosphere to discuss each step of the data optimization cycle, and how to transform your observability from a cost center, to a competitive advantage.

Featuring

A woman in a black dress sitting on a stone bench, starring in a video series on lowering costs through data optimization cycle.
Sophie Kohler

Copywriter
Chronosphere

Sophie Kohler is part of the content team at Chronosphere where she writes blogs as well as creates videos and other educational content for a business-to-business audience.

A man in a white shirt is taking a selfie during a webinar.
Scott Kelly

Senior Product Marketing Manager
Chronosphere

Scott is a Senior Product Marketing Manager at Chronosphere. Previously, Scott was in product marketing at VMware (via the Pivotal acquisition) where he worked on the Tanzu Observability (Wavefront) team and partner go-to-market efforts for VMware’s Tanzu portfolio with AWS and Microsoft Azure. Before VMware, Scott spent three years in product marketing at Dynatrace. In his free time, you’ll find Scott at his local CrossFit “box”, doing home improvement projects, and spending time with his family.

A smiling woman in a blue shirt at a Cloud Native Observability webinar.
Julia Blase

Product Manager
Chronosphere

Kim Oehmichen

Technical Account Manager
Chronosphere

Lowering costs with the data Optimization Cycle 

Sophie Kohler: I think it’s time to talk about costs.  I know it’s on everyone’s minds right now, and in economic downtime, it’s impossible not to feel the effects of high costs and squeezed budgets. For organizations looking to successfully compete, all eyes are on what’s racking up the big bills: data. What’s changed that’s creating this need to control data?

The move from VMs to containers has created an explosion of growth. Cloud native environments actually emit between 10 times  and 100 times more observability data than traditional, VM-based environments. This brings up two challenges. One,  observability data costs skyrocket and become a big budgeting item.  Two, engineers are drowning in a flood of noise, when trying to perform simple tasks, like troubleshooting.

Which means  longer incidents, which means impacted customers, which means  higher costs with worse outcomes. It even ends up affecting your return on investment. “Thanks for telling us what we’re dealing with here, Sophie. Do you have any solutions?”  Okay, okay. I know a few folks who do.  And get this, it follows a three step process. But before we dive into what those three steps are, let’s talk about what it all starts with. Step zero, if you will: centralized governance.

Setting a foundation with centralized governance

Since observability data costs are becoming both costly and unpredictable, observability teams need to lay one crucial foundation: the ability to understand their organization’s usage – by team or service or whatever dimension is important. This allows for chargebacks or showbacks.  And once observability teams have visibility into consumption, they can take steps to make it more predictable. Like allocating portions of the observability system capacity with guardrails, like quotas.

And now that we’ve actually cultivated an environment for allocating priorities and driving accountability, let’s start with step one of the process: Analyze. Actually, let’s go talk to Scott Kelly to hear what he has to say about it. Scott, are you there?

Analyzing incoming data in real time

Scott Kelly:  Yeah, happy to help.  So before taking action on their data, engineering teams must be able to analyze the incoming data in real time to address issues such as cardinality spikes that could currently be impacting the observability system. And they also need the ability to understand both the  cost and the utility of the data. Just knowing if the data is used or not really isn’t enough.

To understand the cost, Chronosphere analyzes all of the incoming data with our  traffic analyzer, to show teams information about the cardinality and volume of the various data sets in real time, and historically.  To understand the value, we provide a   metrics usage analyzer  that allows teams to see the cost  and value  of their metrics side by side. The metrics usage analyzer lets teams slice and dice usage by different dimensions – such as least valuable, most valuable, most used, et cetera, for fine grained cost usage analysis.  It provides detailed information about where their metric is used, how the metric is used, how often the metric is used, and who’s using the metric. We even generate a utility score from this analysis for each metric to make it easy to understand how useful it is.

Sophie Kohler:  Interesting, interesting. So, once you’ve identified the data, what can you do about it? You need strong refining capabilities. Julia, can you help us out?

Refining data with various tools on the write and read path

Julia Blase:  Hi, Sophie. Sure. In order to refine your data,  you might need a few different tools. You’ll want a few tools on both the write path  as data comes into your system, and also on the read path  as data comes out of your storage system. On the write path, I think some of those classic tools are tools that let you drop metrics that are not being used at all today, things that you just don’t need and don’t want to pay for.

Then, on the read  side, you might have complex ways to query data that give you different relevant insights into your system. And you may want to store  those under something like a new metric name. Something that in Chronosphere, we called a   derived metric name. [ It] makes it really easy for people to find and use even if they’re brand new to your system.

Sophie Kohler: Because of the dynamic nature of [the] cloud and its agility, the value of that data may change. Kim, are there  rules  that we  still need to put in place?

Adjust for efficiency and maximize your return on investment

Kim Oehmichen:  That’s a great question, Sophie. So for your large scale cloud native environment, you and your teams need insights into how effective your refining policies are so that you can make adjustments as necessary. Without this level of transparency, your teams cannot ensure that they’re using their assigned capacity or quota efficiently – which results in your organization not being able to maximize the return of investment of your observability practices. As TAMs (Technical Account Managers),  we also help you to identify and put the right policies and configurations in place that work best for your organization and scale.

In other words, we help you to continuously adjust for efficiency, troubleshooting faster by ensuring that the data your engineers need to manipulate during your debugging process are quick and responsive.

Let’s have a closer look at the query accelerator. It continuously scans for slower queries and automatically creates a refined alternative.  What I think is  really cool  is that they can simply create a query that returns the data they need, and our query accelerator will ensure that it performs optimally wherever it is used.

Sophie Kohler: That is  very cool indeed. Now, with observability costs and data growth finally under control, your organization can actually do more with the time and savings, and make their observability a  competitive advantage rather than a cost center.

Analyze, refine, operate. Thanks for listening and catch you next time, where we can talk more about how you can set your team up for success in a cloud native world.