Ultimate UX Toolkit: Data, Context, and Evals for Better Design

Learn UX, Product, AI on Coursera

Stay relevant. Upskill now—before someone else does.

AI is changing the product landscape, it's not going to take your job, but the person who knows how to use it properly will. Get up to speed, fast, with certified online courses from Google, Microsoft, IBM and leading Universities.

  • ✔  Free courses and unlimited access
  • ✔  Learn from industry leaders
  • ✔  Courses from Stanford, Google, Microsoft

Spots fill fast - enrol now!

Search 100+ Courses

Introduction: Navigating the New Frontier of User Experience in AI

In today’s rapidly evolving digital landscape, user experience (UX) design is undergoing a profound transformation driven by advances in artificial intelligence (AI). Traditional UX practices centered around predictable, static interfaces are giving way to dynamic, real-time behaviors generated by probabilistic systems. As AI models become integral to product functionality, designers must develop new skills and mental models to craft experiences that are responsive, ethical, and aligned with user intent. This article explores the emerging toolkit—centered on data, context, and evaluations—that empowers UX professionals to shape AI-driven experiences effectively.

The Shift from Predictable Screens to Probabilistic Systems

Historically, UX design focused on creating interfaces that render predetermined outputs—screens that users could predict and navigate confidently. However, with the advent of AI, especially large language models and multimodal systems, experiences are no longer fixed. Instead, they unfold dynamically based on probabilistic algorithms making decisions in real time. Users now interact with systems that generate responses, recommendations, or actions influenced by vast amounts of training data and contextual cues. This shift challenges designers to think beyond static layouts and consider how models behave under varying circumstances.

Designing for Behavior, Not Just Screens

The core challenge lies in designing how AI models behave rather than merely crafting visual components. This requires a paradigm shift where UX practitioners collaborate closely with engineering and product teams to influence system prompts, fine-tuning strategies, and evaluation frameworks. For example, instead of designing a static FAQ page, a designer might help define how an AI assistant interprets ambiguous questions or how it manages sensitive topics. This behavioral approach ensures the experience aligns with user expectations and ethical standards.

The New Role of UX in Shaping AI Experience

In many organizations, core AI experience design is already happening upstream within engineering and product teams through prompt engineering, data curation, and system orchestration. UX often enters later with efforts focused on making outputs more aesthetically pleasing or on-brand. Yet, the most impactful contributions occur when UX takes a proactive role in defining system behavior from the outset.

As Andrew Ng highlighted, the bottleneck in AI development has shifted from engineering speed to product strategy—deciding what to build and why. For UX professionals, this means expanding influence beyond traditional design artifacts into areas like model behavior specification and evaluation criteria. By developing empathy for users and understanding their mental models, designers can guide system interactions to be more relevant, safe, and trustworthy.

Embracing a Systemic Mental Model for AI Design

This transition demands a systemic thinking approach—viewing AI ecosystems as interconnected agents working collaboratively over time. Agentic design involves orchestrating multiple specialized AI agents that communicate, delegate tasks, and adapt based on context. Designing such ecosystems requires an understanding of how these agents coordinate, how failures propagate through the system, and how users can control or override workflows.

For example, in a customer support AI ecosystem, different agents might handle inquiries related to billing, technical issues, or account management. The UX designer’s role includes ensuring seamless handoffs between agents while maintaining transparency about system capabilities and limitations.

The Essential Toolkit: Data, Context, and Evals

Data: The Foundation of Model Behavior

Data shapes what an AI model knows and predicts. Pretraining datasets influence broad capabilities; fine-tuning data refines specific behaviors; interaction-generated data adapts the system over time. Designers should scrutinize the sources of training data—considering its diversity, biases, and ethical implications—and ask:

  • What behaviors does this data encourage?
  • What perspectives might it exclude?
  • How was this data acquired?

For instance, models trained on internet-scraped data often contain biases or harmful patterns—issues that designers must actively address through dataset curation or model fine-tuning.

Context: Shaping Real-Time Interactions

The context window encompasses user inputs, environmental signals, retrieved knowledge bases, prompts, safeguards, and instructions that assemble the experience at runtime. Writing effective system prompts or constraints is a design act—using plain language or markup languages like XLM—to steer model responses appropriately.

Consider safety concerns such as models generating inappropriate content or suggesting harmful actions—especially when trained on vast unfiltered datasets. Designers must define boundaries—what systems should not do—and embed safeguards directly into context definitions.

Evaluations: Measuring Success & Safety

Evaluation strategies determine what success looks like—what responses are acceptable versus unsafe or irrelevant. Without rigorous evals—including relevance metrics, safety checks, fairness assessments—they cannot identify failure modes or guide improvements.

Frameworks like RISK (Relevant, Inclusive, Safe, Kind) help structure these evaluations for specific use cases. Combining human review with automated tools—such as code-based evaluators or language-model judges—ensures comprehensive oversight.

Continual assessment enables iterative refinement: updating datasets for bias mitigation; revising prompts; connecting responses to knowledge bases; or even deploying antagonistic models that test system boundaries.

From Individual Models to Ecosystems: Agentic Design & Systemic Thinking

The future of UX in AI involves designing ecosystems of specialized agents working collaboratively—they share information, delegate tasks internally—and adapt over time based on feedback loops. This agentic approach requires understanding how agents coordinate and how users can control complex workflows.

This systemic perspective ensures robustness against failures and transparency for users—key factors in building trust in AI products. For example, in a multimodal interface combining text and images generated by different agents, UX must facilitate understanding of each component’s role while maintaining seamless interaction.

The Ethical Dimension: Responsible Design in AI

Designers bear responsibility for shaping safe and ethical AI experiences. This entails defining what behaviors are permissible within the system through careful prompt design and evaluation criteria. Neglecting safeguards can lead to harmful outputs—such as inappropriate content generation or misinformation—which erodes user trust and exposes organizations to risk.

Implementing strong safeguards isn’t just technical but deeply rooted in intentional design decisions—choosing not to implement certain features or redirect conversations proactively reflects values embedded within the system’s behavior.

Practical Steps for Incorporating AI Behavior into UX Practice

  • Educate yourself about model behavior: Understand training data origins and limitations.
  • Collaborate upstream: Work with product managers and engineers during prompt formulation and system architecture planning.
  • Define clear evaluation metrics: Use frameworks like RISK to measure relevance and safety systematically.
  • Create explicit context: Write thoughtful prompts and constraints that guide model responses ethically.
  • Implement continuous monitoring: Regularly assess model outputs post-deployment to catch drift or harmful behaviors early.

In Closing: Redefining UX in an AI-Driven World

The integration of AI into products is fundamentally redefining what it means to craft user experiences. No longer can UX practitioners focus solely on static screens; instead, they must become architects of behavior—designing how models learn from data, interpret context, and are evaluated for safety and relevance.

This requires adopting systemic thinking complemented by a new toolkit containing data stewardship skills, context management techniques, and rigorous evaluation strategies. By moving upstream alongside engineering and product teams—and embracing their role as behavioral designers—UX professionals can shape responsible AI experiences that truly serve users’ needs while mitigating risks.

The future belongs to those who see AI not just as technology but as material—a malleable substance they can craft with precision and care. Invest in learning these tools today to lead your organization into this new era of user-centered AI design.

Oops. Something went wrong. Please try again.
Please check your inbox

Want Better Results?

Start With Better Ideas

Subscribe to the productic newsletter for AI-forward insights, resources, and strategies

Meet Maia - Designflowww's AI Assistant
Maia is productic's AI agent. She generates articles based on trends to try and identify what product teams want to talk about. Her output informs topic planning but never appear as reader-facing content (though it is available for indexing on search engines).