How Labra engineers stopped getting logged out mid-debug

... and unified fragmented multi-cloud observability with Oodle

How Labra engineers stopped getting logged out mid-debug
Seat sharing woes: when a teammate logs in, you get logged out.

Summary

Labra runs a multi-cloud SaaS platform across AWS, Azure, and GCP. Their observability was split between New Relic and cloud-provider specific tools: twice the cost, four times the number of tools to monitor.

Challenges:

  • Observability access limited by pricing tiers and product complexity
    • Expensive per-user pricing. Account-sharing leads to engineers
      getting logged-out mid-debug.
    • "New Relic has a ton of features we can't use: they're not simple"
  • Double-paying: CloudWatch and New Relic ingesting the same data
  • Alerting was noisy, delayed, or absent entirely
  • Azure logs took five minutes to appear
  • Distributed tracing was manual and incomplete

With Oodle:

  • Unified observability for all signals: logs, metrics, traces.
  • Ask AI Assistant in plain English to debug across all signals.
  • Powerful features, simple UX.
  • AI assistant built a 50-panel dashboard in 1.5 hours (vs 1.5 days manually)
  • MCP server plugged into Labra's custom LangChain debugging agent.

We've been struggling a lot with monitoring, alerting. We have the tools. It's just not much time or expertise put towards that part.

Nestor

Principal Engineer, Labra

Who Is Labra

Labra is a SaaS go-to-market platform that helps independent software vendors list, sell, and operate through cloud marketplaces on AWS, Azure, and GCP. The platform orchestrates everything from CRM integration to marketplace transactions for enterprise Independent Software Vendors.

Over 40 developers run roughly 20 services across four environments (QA, staging, production, DR). AWS usage has ECS Fargate, Lambda, RDS, DynamoDB. Container Apps, Azure Functions, Cosmos DB are used on Azure, and GCP uses GKE and Google Cloud Functions. The stack is primarily Python, with some Node.js and a critical Golang service.

For a team processing marketplace transactions across three clouds, observability isn't optional. When a listing sync fails or a transaction gets stuck, engineers need to trace it through Lambda functions, ECS services, and third-party APIs. Fast.

Paying Twice, Debugging Twice

Labra's observability had grown organically. CloudWatch was the default for AWS: ECS logs, Lambda logs, infrastructure metrics. But CloudWatch alone wasn't enough for a team that needed APM, error tracking, and distributed tracing. So they added New Relic.

The intent was to consolidate. The opposite happened. New Relic had reliability problems, and the team couldn't afford gaps in log availability. So they kept both.

We wanted to have high availability of logs. We were worried about downtime, so we stored logs in both places, CloudWatch and New Relic. But now, going forward, we'd love to see it in one place only.

Priyank Raj

Senior Software Engineer, Labra

ECS services shipped logs to both CloudWatch and New Relic. Lambda logs went to CloudWatch first, then got forwarded to New Relic through a separate pipeline. Two tools ingesting the same data. Two bills every month.

New Relic had the capabilities. The team just couldn't use them. Alert configuration was complex enough that monitors were either misconfigured, noisy, or not set up at all. Dead letter queue notifications fired for internal test accounts. Critical alerts never got created because nobody had the time.

Since alert creation in Oodle was much easier, the team had created all the alerts they need right away, including the ones they always wanted on New Relic but were putting it off.

We're not utilizing New Relic to its full capability. The main point is we're not getting alerts properly.

Nishant Kumar

Senior Software Engineer, Labra

The cost kept climbing. Per-user pricing meant one admin account with three concurrent logins for the entire company. A fourth engineer logging in would kick someone out mid-debug. Everyone else relied on CloudWatch or secondhand Slack summaries.

New Relic is very expensive, so we have just a few users. Only three can log in at a time. If a fourth person logs in while you are debugging, you get logged out in the middle.

Priyank Raj

Senior Software Engineer, Labra

On Azure, logs took five minutes to surface. For a team debugging production issues across clouds, five minutes is an eternity.

The logs on Azure take like five minutes to show up, and that is painful

Nestor

Principal Engineer, Labra

Distributed tracing existed, but barely. The team had instrumented some services with New Relic's proprietary SDK. Gaps were common. Tracing across clouds didn't work. And because the instrumentation was tied to New Relic's SDK rather than OpenTelemetry, any future vendor change would mean re-instrumenting everything.

The wish list was clear: one platform for logs, metrics, and traces. Alerting that works without a week of configuration. No per-user fees. And support that responds immediately.

Why Oodle

Four things stood out during the initial evaluation.

The UI made sense immediately. Priyank, who leads Labra's platform engineering, had spent enough time in New Relic had many complaints. Oodle was the opposite.

The UI is simpler here and easier to debug. In New Relic, it's clumsy. At least for me, it's very difficult to navigate. But here, everything is straightforward. You don't need to drill down deep.

Priyank Raj

Senior Software Engineer, Labra

AI: built-in, not bolted-on. Oodle's AI assistant creates alerts, writes PromQL, builds entire dashboards, and investigates firing alerts by correlating logs and metrics.

For a team that had given up on configuring New Relic's alerting, this tipped the evaluation.

We use the Oodle AI assistant like this: this is my service, this is my request ID, get me the RCA. It comes with artifacts and saves a lot of time.

Priyank Raj

Senior Software Engineer, Labra

No per-user fees. Oodle charges by ingestion volume and retention. Every developer gets access. For a team where 40+ engineers shared three New Relic logins, this was liberating.

Rapid-response support. Responsiveness is a trust signal when you're evaluating a newer vendor. The Oodle team answered questions quickly, shipped features during the PoC, and worked alongside Labra's engineers on integration. With Oodle University tutorials and Slack support, the Labra DevOps team handled integrations with ease.

Support on Slack is very good. We get replies in minutes.

Priyank Raj

Senior Software Engineer, Labra

From PoC to Production in Days

The PoC started small: three QA ECS services, connected via an OpenTelemetry collector running as a sidecar container. Logs and traces started flowing immediately.

Log ingestion was seamless. Just connecting with the OTel collector and things are flowing good. The best part I liked about the platform is tracing, that heatmap. It really helps.

Shashwat

Lead Software Engineer, Labra

Traces in Oodle

Looking at how quick and easy the hyper-focused POC process was, the team expanded to production within the first week. The integration was, in Priyank's words, "straightforward". They added the OTel collector sidecar to existing ECS task definitions. Production telemetry started flowing without application code changes.

Alert creation followed. The DevOps team set up monitors for ALB error rates using Oodle's AI assistant. No PromQL expertise required.

I did not face any challenges on the alert creation part. It's good!

Swapna Mannem

Senior DevOps Engineer, Labra

Oodle's AWS integration pulled in CloudWatch metrics for RDS, ALB, ECS, ElastiCache, and SQS. Infrastructure metrics from EC2 instances came through a Datadog agent forwarding to Oodle. Within two weeks, the team had logs, metrics, traces, and alerts running in both QA and production.

Support showed up when it mattered. When a developer hit a Python logging compatibility issue (structlog didn't work with OTel's auto-instrumentation), an Oodle engineer sent back a verified, working code example within minutes.

The New World with Oodle

One Platform, Every Developer

No more choosing between CloudWatch and New Relic depending on which service you're debugging or whether you have a seat. Every developer at Labra now has access to the same logs, metrics, and traces. No per-user fees. No seat restrictions. No secondhand Slack summaries.

Within a week, 10-15 developers were using Oodle for real debugging on migrated services: searching logs by request ID, tracing requests across services, building dashboards for marketplace transaction metrics. After one production deployment, a developer found the root cause in Oodle's AI summary at the top of the log view instead of scrolling through hundreds of log lines. Alerts were firing to Slack. This wasn't a sandbox evaluation. It was production work.

AI That Debugs, Not Just Alerts

Oodle's AI assistant replaced the manual alerting workflow. Instead of constructing PromQL queries and hoping the threshold was right, the team describes what they want monitored. The assistant creates the alert, catches syntax errors, and suggests improvements.

Labra took it further. The team had already built a custom LangChain agent for internal debugging. They connected Oodle's MCP server to this agent.

When alerts are triggered, the agent uses the MCP and gets artifacts out of it. The MCP had minimum delays.

Priyank Raj

Senior Software Engineer, Labra

They had tried the same approach with New Relic's MCP and AWS CloudWatch's MCP. Neither performed well. Oodle's was faster and returned useful results consistently.

Dashboard building got faster too. Priyank built a 20-panel monitoring dashboard in an hour and a half with the AI assistant. The same work done manually — find each metric, write the PromQL, verify the graph, repeat 20 times — would have taken a day and a half.

Oodle AI Assistant creating dashboards

Endpoint Monitoring Without the Price Tag

Labra was paying for synthetic canary jobs through CloudWatch Synthetics. It worked, but the cost was hard to justify for basic availability checks.

Labra mentioned the need for synthetic monitoring during a Thursday PoC call. By the next meeting, Oodle had it working in dev. The day after that, the feature was live in Labra's instance. It covers HTTP, ping, DNS, TCP, traceroute, and SSL certificate checks. For Labra's public endpoints, it replaced the canary setup at a fraction of the cost.

Synthetic Monitors in Oodle

Traces That Show the Shape of Your Latency

With the OTel collector already in place for logs, traces came along for free. Shashwat called out the trace heatmap as his favorite feature: spot latency outliers at a glance, without writing queries or sifting through raw spans.

Trace details: reveal latency distribution of specific spans

If your team is paying for many observability tools and fully using none, juggling alerts that fire for the wrong reasons while critical ones arrive hours late — that's exactly where Labra was.

See why teams switch to Oodle →