Instana AI Strategy

A unifying interaction model for AI in observability — built before agentic patterns had a name.

Client
IBM
Role
Senior Design Lead
Year
2023-2024

Thesis

At Instana, a new focus on AI was starting to surface across the product in disconnected pockets: incident summarization in one squad, predictive alerting in another, chatbot research in a third. The work was siloed and narrow in focus.

I saw the trajectory of the market, the AI direction coming out of IBM, and the Instana need for a bigger strategy and a unifying interaction model. The framework I built defined which AI capabilities the product should reach for and how they should connect, with a hub-and-spoke interaction model at its center: the chatbot as hub, bidirectionally connected to the rest of the product.

This was early. Agentic AI did not yet have a settled name in the industry, and the interaction patterns had not yet crystallized into conventions.

I was already leading three concurrent design efforts at Instana. I started this initiative as a fourth, pulled in collaborators across PM, engineering, and architecture, and shared the framework as scaffolding for other teams exploring AI in their own corners. This set the stage for the AI work that shows up in the product today.

The small ask

The assignment that landed on one of my teams was small. Design a Generate-Summary button on the incident page. The user clicks, the system writes a paragraph of natural language describing what was wrong with the incident, the user moves on. The capability already existed in the product in a guardrailed form, surfaced through a single button on a single page. My job was to make it more useful.

The brief made sense. Incidents in the product are dense by design: grouped alerts, root-cause analysis, related events, automation policies, ongoing collaboration notes. Reading one quickly is hard. A summary button is a reasonable response to that, and shipping it as a self-contained, guardrailed feature was a reasonable way to handle a still-uncertain technology.

On its own, it would have shipped fine.

The existing Generate-Summary buttonNotes and Activity panel on an incident page, with the “Preview” tag visible.

The realization

Once I started looking past the button, the picture across the product was harder to ignore.

Smart alerts had been doing pattern detection for years without ever being framed as AI. Predictive alerting was being sketched out on another team. Chatbot research was running on my own team alongside the summary button. Other AI possibilities were being explored elsewhere in the product. Each effort was reasonable in isolation, but there was no one whose job it was to see how the pieces fit together.

The questions that needed asking were obvious once you stepped back. How do all these capabilities relate to each other? What is the user’s mental model of AI in the product? Where does AI live, and where does it not? Without answers, the product was going to ship a half-dozen disconnected AI features, each making its own assumptions about how the user should encounter them.

It wasn’t anyone’s job, so I made it mine.

The framework

My response was a framework. A top-down articulation of what AI in the product should be, how its pieces should connect, and how users should encounter it.

The framework had four parts that fit together: a capabilities taxonomy, an interaction model, a set of UI integration patterns, and a technical architecture beneath them. Each part answered a question that was not being asked anywhere else in the product.

Capabilities taxonomy

The starting move was naming what AI actually was inside the product. Five capability types, each playing a different role.

Naming the types gave every team a vocabulary to locate their work in. The summary button sat in Generative. Smart alerts sat in Predictive. Both were AI in a meaningful sense, and both belonged in the same picture.

01
Generative
Interprets, summarizes, suggests. The chatbot lives here.
02
Predictive
Detects anomalies, surfaces what’s coming. Smart alerts and predictive alerting.
03
RAG
Pulls in documentation, runbooks, and historical context to ground responses in source material.
04
Agentic
Performs actions and remediations on the user’s behalf.
05
Assistive
Sharpens search, recommendations, and small enhancements throughout the product.
Interaction model — the chatbot as hub

The capabilities had to connect, not just coexist. I put the chatbot at the center of the interaction model and made the rest of the product the spokes around it.

The chatbot pointed outward to do work in the product. It navigated the user to relevant views. It built things (dashboards, runbooks, summaries) on demand. It served as the entry point for AI features embedded elsewhere in the UI.

The product pointed back to do work in the chatbot. A probable root cause analysis or an insight on a page could launch a chat thread for deeper investigation. AI outputs landed in the chat panel for follow-up.

The chatbot acted as connective tissue between AI capabilities and the rest of the product.

Hub-and-spoke interaction modelChatbot at center; bidirectional spokes — outward to navigate, build, and trigger embedded AI; inward from page elements that launch chat threads.
UI integration models

Knowing what AI is and how it connects was not enough. The framework also had to specify where AI shows up in the interface and what kind of behavior lives at each surface. Four patterns covered the territory.

Each pattern had a different design challenge. Knowing the framework allowed new features to be built systematically and efficiently within it.

01
Self-Contained Chat
The dedicated chat interface, available across the product. Open-ended Q&A, debugging, and the bidirectional interactions all happen here.
02
Chat-Driven Navigation
AI woven into how the user moves through the product. The chat surfaces the relevant page; search gets generative reformulation; the product becomes navigable in plain language.
03
Artifact Creation
Generative AI embedded in authoring surfaces. Build a dashboard, draft a query, scaffold a runbook inside the tools the user is already working in.
04
Contextual Launch
AI features attached to specific objects in the product. The Summarize button on an incident; the same pattern on a Suggest button for a query, an Explain button for a metric.
Working with architecture

I worked with the architects to understand the systems beneath the framework. My role wasn't to design the technical architecture — it was to communicate the mission to engineering, weigh tradeoffs with the architects building it, and translate the resulting limits and opportunities back to the design teams.

Technical flowInput → LLM → branch (NL→API for REST/GraphQL · NL→Widget for UI artifacts) → Response.

The concepts

I designed use cases to test the framework and to evangelize it. Each one exercised a different part of the system, often more than one at a time.

Incident summarization

The Summarize button itself shipped in Q1 2025. This was the original feature behind the strategy, and it landed in the framework as a Contextual Launch: AI attached to a specific object in the product, surfaced at the moment of need, generating output that lives in context.

Treating the button as one instance of a pattern, instead of a one-off feature, meant designing it to coexist gracefully with everything else AI was going to do across the product.

Shipped Summarize feature in contextChat home with the Summarize affordance · streaming summary generation · final summary output in context.
Audience-aware sharing

When something breaks in production, an incident gets shared many times to many different people. Executives need to know impact and duration. On-call SREs need probable cause and potential fixes. Downstream engineers need the full technical picture.

I designed a sharing flow where the user picks who the share is for, and the AI generates a summary tailored to that audience:

  • Executive overview. Top-line impact: who’s affected, how long, severity.
  • Triage report. Probable cause and possible solutions for whoever’s getting paged.
  • Detailed share. Full technical detail for the engineer who will do the work.

The flow sat inside the existing share modal — a third option alongside invite a collaborator and share a link, with the audience picker driving the template. This is Artifact Creation, not a chat conversation. The user clicks share, picks the audience type, and the system generates a bespoke summary on demand.

Audience-aware shareSharing modal with audience options · executive summary view · triage report view.
Generating a dashboard from chat

A user asks the chatbot something specific: “Show me latency for our checkout service over the last week.” The chatbot’s answer is the dashboard itself. It builds it on the fly, then navigates the user to it. The user lands on a new persistent dashboard view, ready to be saved, shared, or extended.

This single flow exercises three patterns at once. It starts in Self-Contained Chat (the user asks). It produces an Artifact (the dashboard). It ends in Chat-Driven Navigation (the user lands on the new view). The chatbot moves from answerer to builder to guide in one user action.

Chat-to-dashboard concept flowChat input · chatbot generating a dashboard inline · user landed on the new persistent dashboard view.
Runbooks with a chat refinement loop

When an incident hits, the response often involves running remediation steps that may or may not have been documented before. I designed a flow where the chatbot, anchored to the incident at hand, generates a candidate runbook, pulling from documentation, prior incident histories, and the engineer’s intent in the moment.

The runbook lands in the product as an artifact. The user can run it as-is, or click into it to open a chat thread for refinement: “Make this safer,” “Skip step three,” “Add a rollback at the end.” The chat takes the artifact, modifies it, and returns it to the product.

This is Artifact Creation feeding into the bidirectional product-to-chat loop from the interaction model, and pointing toward Agentic AI, where the next step is the chatbot offering to execute the runbook itself.

AI-generated runbook with chat-refinement flowChatbot generating a candidate runbook · runbook in the product with a “Refine in chat” affordance · chat thread refining the runbook.

Each use case touched a different part of the framework. The Summarize button proved Contextual Launch could ship. Audience-aware sharing was Artifact Creation, tuned to the recipient. The dashboard flow chained three patterns into one user action. The runbook loop pointed at the Agentic edge. The framework was producing things that fit together as a system.

The pitch & where it went

I took the use cases and paired them with the system’s IA and taxonomy. The synthesis ran end-to-end: vision, capabilities taxonomy, four UI integration models, hub-and-spoke interaction model, technical flow, phased roadmap, risks and ethics. The roadmap itself moved through four phases (foundations, expanded AI features, automated actions, adaptive AI), the last two of which are the territory the industry now calls agentic.

I captured the synthesis in a Mural and built pitch decks, flows, and other artifacts from it for different audiences. I started sharing the work out, working with the PMs and architects I had sold on the vision. The aim was to turn the synthesis into a funded work stream with portfolio-level sponsorship behind it.

After presenting and documenting the work, I separated from the company. The framework didn’t ship under my name, but the direction it was pointing in continued to shape the product.

Since I left, Instana has shipped Intelligent Incident Investigation (preview), using agentic AI for faster root cause identification — in the same chatbot-as-investigation-tool direction the framework was pushing toward. They also released GenAI Observability for monitoring AI applications running on the platform.

The strategic direction the framework was articulating is where the product went.

Leader
Gartner Magic Quadrant for Observability Platforms — July 2025
IBM Instana Observability

The work was rewarding for what it required: looking past the assignment in front of me to the system underneath, designing for a technology before its conventions had settled, and trying to bring others along to a vision that hadn’t yet been blessed from above. That kind of work doesn’t always fit within your tenure. Products at this scale are built through a ton of collaboration, and it’s fun getting to shape the path of something that continues to live on.

Note: Due to NDAs, some assets have been redrawn or generalized.