TNS
VOXPOP
As a JavaScript developer, what non-React tools do you use most often?
Angular
0%
Astro
0%
Svelte
0%
Vue.js
0%
Other
0%
I only use React
0%
I don't use JavaScript
0%
Large Language Models / Observability

LLo11yPop: Nvidia, Grafana Working on LLM for Observability

The two companies are creating a number of projects aimed at giving engineers a better sense of how AI model training is progressing and solve other observability pain points.
Oct 23rd, 2024 1:00pm by
Featued image for: LLo11yPop: Nvidia, Grafana Working on LLM for Observability
Photo of Aaron Erickson, senior engineering manager at Nvidia, at ObservabilityCON.

NEW YORK — It wasn’t an announcement per se, but Nvidia Senior Engineering Manager Aaron Erickson described during a keynote here at ObservabilityCON several AI initiatives Nvidia is working with Grafana on.

The projects include the two organizations developing AI training to better understand model performance and consistency. Another initiative leverages telemetry to create an interface for observability into large language models and AI applications.

As Erickson described in the Sept. 24 keynote, Nvidia relies on Grafana Cloud for its observability support.

One of the more interesting projects — called LLo11yPop — is an LLM for observability. Nvidia is developing an LLM designed with Grafana’s support to ask questions like, “Show me a graph of job failures.” Or, “What are the five best possible causes of last night’s issues?” Or, “Which clusters need maintenance over time?”

Erickson told conference attendees during the keynote address, “It’s still early days, but we’re starting to get answers from the data center using multiple LLMs trained in different areas of expertise. It’s an incredible concept, and we’re excited about the possibilities. We believe this, and more agents added over time, will allow us to solve a wide range of problems.”

Telemetry to Check LLM Training Status

Nvidia is also working with Grafana on a Grafana-led application for training observability; Nvidia is a design partner with Grafana Labs on this project.

Nvidia’s status as a Grafana customer, Erickson said, is “critical for what we’re trying to accomplish: actually being able to understand the telemetry of how training is going

“Imagine you’re trying to build a foundation model or doing a massive tune of a model. You need to understand: Is the model converging? Is the training progressing? And, on a lower level, is the GPU performance stable? Is the heat OK? All of this is crucial to evaluate if you’re making a good investment in these training runs.”

Moreover, Erickson said during his keynote: “One of our key convictions is that this system won’t work unless it’s grounded in truth. The query results from the LLM must be backed by actual data, like what’s shown on a Grafana dashboard. We know LLMs aren’t perfect — they do hallucinate — so we want to use them to build dashboards dynamically.

“You should be able to ask a question, get a dashboard link, and then dive deeper into the data center’s status across the world, across cloud providers, and across continents where we have GPUs.”

Grafana is working with several AI companies, in addition to Nvidia, “to push the boundaries on what we can monitor, learn and generate using both technologies,” Tom Wilkie, CTO of Grafana Labs, told me.

“Our current efforts with Nvidia will help drive more useful observability to both those running and monitoring GPU infrastructure as well as those building and training models to run on GPUs. This foundational work can lead to a number of more reliable outputs from today’s models, and is something that should only continue to improve in accuracy and value to users as model performance improves.”

What’s Under the Hood?

NIM retrieval agents based on Nvidia NIM (Nvidia Inference Microservices) technology, serves as the building blocks for the system Nvidia is building. NIM can also be described as offering optimized inference microservices for large-scale LLM deployments.

These agents gather data from sources like Grafana to answer questions. The questions come from analyst agents trained on how different applications run in data centers. With an observability agent framework for GPU fleet management using a multi-LLM compound model in the architecture design, agents manage orchestration and task execution for observability frameworks. These are orchestrated by what’s called an OODA loop — Observe, Orient, Decide, Act.

“The loop drives the agents to ask questions, identify what went wrong, and take action, like opening Jira tickets or calling PagerDuty,” Erickson told The New Stack.

Debugging, of course, is paramount as without proper observability tools, making application and network fixes can take weeks or even months, he said.

“If they fail, that’s a lot of wasted resources for no good reason,” Erickson said. “That’s why it’s really important to have this insight, and we’re thrilled to be design partners, helping build out this capability. This collaboration allows us to achieve our goals and helps other customers running training jobs on GPUs too.”

‘Who Among Us Hasn’t Hallucinated a Little?’

During his talk, Erickson waxed not so nostalgic during his days as vice president of engineering for an unnamed “observability company.”

“Imagine waking up every day at 6 a.m., responsible for helping your CEO understand, by 8 a.m., what happened the night before,” he said. “My routine was to read the backscroll in Slack, call directors and [independent contributors], and track down the people handling incidents.

“I had to gather information like which customers were impacted, which regions were affected and possible root causes. We needed all this to write the incident report and inform customers. This happened every day for six months while we transitioned to the cloud — it was not a fun time.”

LLMs are not perfect, or at least not yet. Frequent hallucinations across all LLMs are one of the most cited failures. But at the same token, humans do have hallucinations as well.

“You’ve had an incident, and the next morning, a senior leader asks what happened,” Erickson said. “Who among us hasn’t hallucinated a little to fill in the gaps? If you’re skeptical, just ask your developers. It happens all the time.

“This inspired me to think: with GPT-4, we can take a human question and translate it into a query language like PromQL, or anything we use with Grafana, to figure out what happened last night. As an engineer, this allows me to start with a rough idea, ask follow-up questions, and get closer to plausible causes — and while it may not be perfect, neither are humans.”

Created with Sketch.
TNS DAILY NEWSLETTER Receive a free roundup of the most recent TNS articles in your inbox each day.