[HN Gopher] Show HN: Wild Moose - Autonomous agent for productio...
___________________________________________________________________
Show HN: Wild Moose - Autonomous agent for production debugging
Hi Hacker News! We launched an autonomous agent that helps debug
production issues, and we're curious to get your feedback. Today's
GenAI devtools, such as Copilot, are limited: they are great for
writing code, but we all know that programming is only 20% coding,
and 80% debugging. So how can GenAI be used for debugging? As
opposed to code completion or test automation, production debugging
is not about generating text. Debugging is mostly about root-cause
analysis. We realized two things: 1) Generative AI is drastically
changing the way we work with data, thanks to its ability to not
only generate queries, but also run code and analyze unstructured
data. This enables building better data-exploration experiences
with far more intuitive interfaces. 2) RCA is all about exploring
different types of data. When you don't know why a transaction was
dropped or which customers are affected - you explore metrics,
logs, your code, other people's code, old slack messages, and
whatnot, to figure out what's broken. Putting those two together,
we built an autonomous agent that helps debug production issues.
Our LLM "moose" (ok, it's corny but we like it) connects to your
logs, metrics, and other observability data, and lets you extract
and analyze them by conversing with it. Once it gets a task, it
will start reasoning, invoking APIs, and running code, until it
comes back with an answer. For example, when requested to "show me
IDs of transactions that took over 1 minute today", it will fetch
those transactions from Datadog for you. You might then ask it if
long-running transactions correlate with a metric such as DB CPU
load. It will fetch the metric values, visualize them on a graph
alongside the long transaction frequency, and give you the answer.
Our software both runs code and invokes API calls; the interplay
between these is nontrivial to design and a fertile ground for
innovation. There are "textbook" solutions to let agents write and
run code (open sourced by, for example, Open Interpreter), and also
to invoke tools/APIs (for example, Gorilla). But doing both
together is harder, and yet required. We welcome your thoughts on
this! Try our tool with your Datadog's logs and metrics >>
https://app.wildmoose.ai/slack/install Setup demo >>
https://tinyurl.com/mvaj8emf If you want to see other
integrations, or have ideas for features, or you've spotted
behaviors that seem off - we'd love to hear. Hit us up in the
comments!
Author : yasmind
Score : 25 points
Date : 2023-10-03 16:08 UTC (6 hours ago)
(HTM) web link (www.wildmoose.ai)
(TXT) w3m dump (www.wildmoose.ai)
| pico-games wrote:
| Being able to go through massive amounts of production logs
| quickly is super interesting - how can we pinpoint and trace
| requests quickly going through multiple microservices?
| roeischuster wrote:
| Roei the CTO here :) During onboarding we preliminarily index
| some data about your microservice architecture so we have some
| "common language" with your Datadog instance. Couple that with
| run-of-the-mill distributed tracing deployed at the user's end,
| and our moose can start querying/reasoning about multi-service
| transactions, and/or at the finer granularity of individual
| requests and log lines.
|
| Does this answer your question?
| infixed wrote:
| One of the issues with AI products is that they often require a
| good amount of input to use. It seems to me that a great AI
| product would save me time and do things for me without being
| asked to.
|
| For example, at our company we have a quite a few of alerts set
| up. Datadog also automatically detects anomalies. It would be
| neat if this (or something else) could automatically do an
| initial triage without being prompted and give me a free
| headstart on issues that come in.
|
| Otherwise, it feels like it's "work" to learn how to use the
| product, which seems to miss the promise of AI (doing things for
| us!).
| yasmind wrote:
| 100%. We already have some of those more-automated features,
| e.g. giving you context about recent changes that might be
| related to an alert, like deployments that might be related,
| feature flags, etc. But def interesting to do more around
| triage. Are you using anything today for triage or is it all
| manual?
| morsela wrote:
| Pretty wild. I can see these kinds of ai-based tools become used
| more often as they connect to more systems.
|
| With that said, in my experience, usually, coming up with a good
| AI question can sometimes be harder than the act of looking for
| the data itself, i.e. asking "what were the last 3 errors
| yesterday" is not something I have ever done.
| jadbox wrote:
| It sounds neat. I'd need to see how productive it is for a real
| world production environment to figure out the value. The splash
| page design though is an absolute delight for me. Scrolling feels
| slow though.
| yasmind wrote:
| Thanks, we'll fix the scrolling issue!
|
| Do you have any examples for recent prod issues you faced?
| Would be interesting to see how we could help speed up the
| investigation there.
___________________________________________________________________
(page generated 2023-10-03 23:00 UTC)