Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Google integrates Gemini CLI into Zed code editor

      August 28, 2025

      10 Benefits of Integrating React.js Vibe Coding into Your Agile DevOps Pipeline

      August 28, 2025

      Designing For TV: The Evergreen Pattern That Shapes TV Experiences

      August 27, 2025

      Amplitude launches new self-service capabilities for marketing initiatives

      August 27, 2025

      This Vizio soundbar has impressive surround sound, and it’s on sale

      August 29, 2025

      DJI’s ultralight wireless Mic 3 captures great audio – even in tricky situations

      August 29, 2025

      OpenAI gives its voice agent superpowers to developers – look for more apps soon

      August 29, 2025

      T-Mobile will give you 4 free iPhone 16 phones right now – here’s how to get yours

      August 29, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      Optimizing Laravel Livewire Performance with Computed Properties

      August 29, 2025
      Recent

      Optimizing Laravel Livewire Performance with Computed Properties

      August 29, 2025

      Smart Cache Package for Laravel

      August 29, 2025

      This Week in Laravel: Filament 4 Videos and Pest 4 Browser Testing

      August 29, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Containers in 2025: Docker vs. Podman for Modern Developers

      August 29, 2025
      Recent

      Containers in 2025: Docker vs. Podman for Modern Developers

      August 29, 2025

      FOSS Weekly #25.35: New Gerhwin DE, grep Command, Nitro init system, KDE Customization and More Linux Stuff

      August 29, 2025

      19 Beautiful Themes to Get a Better Visual Experience With VS Code

      August 29, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Tech & Work»Design Patterns For AI Interfaces

    Design Patterns For AI Interfaces

    July 14, 2025

    So you need to design a new AI feature for your product. How would you start? How do you design flows and interactions? And how do you ensure that that new feature doesn’t get abandoned by users after a few runs?

    In this article, I’d love to share a very simple but systematic approach to how I think about designing AI experiences. Hopefully, it will help you get a bit more clarity about how to get started.

    This article is part of our ongoing series on UX. You can find more details on design patterns and UX strategy in Smart Interface Design Patterns 🍣 — with live UX training coming up soon. Jump to table of contents.

    The Receding Role of AI Chat

    One of the key recent shifts is a slow move away from traditional “chat-alike” AI interfaces. As Luke Wroblewski wrote, when agents can use multiple tools, call other agents and run in the background, users orchestrate AI work more — there’s a lot less chatting back and forth.

    In fact, chatbots are rarely a great experience paradigm — mostly because the burden of articulating intent efficiently lies on the user. But in practice, it’s remarkably difficult to do well and very time-consuming.

    Chat doesn’t go away, of course, but it’s being complemented with task-oriented UIs — temperature controls, knobs, sliders, buttons, semantic spreadsheets, infinite canvases — with AI providing predefined options, presets, and templates.

    There, AI emphasizes the work, the plan, the tasks — the outcome, instead of the chat input. The results are experiences that truly amplify value for users by sprinkling a bit of AI in places where it delivers real value to real users.

    To design better AI experiences, we need to study 5 key areas that we need to shape.

    Input UX: Expressing Intent

    Conversational AI is a very slow way of helping users express and articulate their intent. Usability tests show that users often get lost in editing, reviewing, typing, and re-typing. It’s painfully slow, often taking 30-60 seconds for input.

    As it turns out, people have a hard time expressing their intent well. In fact, instead of writing prompts manually, it’s a good idea to ask AI to write a prompt to feed itself.

    With Flora AI, users can still write prompts, but they visualize their intent with nodes by connecting various sources visually. Instead of elaborately explaining to AI how we need the pipeline to work, we attach nodes and commands on a canvas.

    With input for AI, being precise is slow and challenging. Instead, we can abstract away the object we want to manipulate, and give AI precise input by moving that abstracted object on a canvas. That’s what Krea.ai does.

    In summary, we can minimize the burden of typing prompts manually — with AI-generated pre-prompts, prompt extensions, query builders, and also voice input.

    Output UX: Displaying Outcomes

    AI output doesn’t have to be merely plain text or a list of bullet points. It must be helpful to drive people to insights, faster. For example, we could visualize output by creating additional explanations based on the user’s goal and motivations.

    For example, Amelia Wattenberger visualized AI output for her text editor PenPal by adding style lenses to explore the content from. The output could be visualized in sentence lengths and scales Sad — Happy, Concrete — Abstract, and so on.

    The outcome could also be visualized on a map, which, of course, is expected for an AI GIS analyst. Also, users can access individual data layers, turn them on and off, and hence explore the data on the map.

    We can also use forced ranking and prioritizations to suggest best options and avoid choice paralysis — even if a user asks for top 10 recommendations. We can think about ways to present results as a data table, or a dashboard, or a visualization on a map, or as a structured JSON file, for example.

    Refinement UX: Tweaking Output

    Users often need to cherry-pick some bits from the AI output and bring them together in a new place — and often they need to expand on one section, synthesize bits from another section, or just refine the outcome to meet their needs.

    Refinement is usually the most painful part of the experience, with many fine details being left to users to explain elaborately. But we can use good old-fashioned UI controls like knobs, sliders, buttons, and so on to improve that experience, similar to how Adobe Firefly does it (image above).

    We can also use presets, bookmarks, and allow users to highlight specific parts of the outcome that they’d like to change — with contextual prompts acting on highlighted parts of the output, rather than global prompts.

    AI Actions: Tasks To Complete

    With AI agents, we can now also allow users to initiate tasks that AI can perform on their behalf, such as scheduling events, planning, and deep research. We could also ask to sort results or filter them in a specific way.

    But we can also add features to help users use AI output better — e.g., by visualizing it, making it shareable, allowing transformations between formats, or also posting to Slack, Jira, and so on.

    AI Integration: Where Work Happens

    Many AI interactions are locked within a specific product, but good AI experiences happen where the actual work happens. It would be quite unusual to expect a dedicated section for Autocomplete, for example, but we do so for AI features.

    The actual boost in productivity comes when users rely on AI as a co-pilot or little helper in the tools they use daily for work. It’s seamless integrations into Slack, Teams, Jira, GitHub, and so on — the tools that people use anyway. Dia Browser and Dovetail are great examples of it in action.

    Wrapping Up

    Along these five areas, we can explore ways to minimize the cost of interaction with a textbox, and allow users to interact with the points of interest directly, by tapping, clicking, selecting, highlighting, and bookmarking.

    Many products are obsessed with being AI-first. But you might be way better off by being AI-second instead. The difference is that we focus on user needs and sprinkle a bit of AI across customer journeys where it actually adds value.

    And AI products don’t have to be AI-only. There is a lot of value in mapping into the mental models that people have adopted over the years, and enhance them with AI, similar to how we do it with browsers’ autofill, rather than leaving users in front of a frightening and omnipresent text box.

    Useful Resources

    • Where Should AI Sit In Your UI?, by Sharang Sharma
    • Shape of AI: Design Patterns, by Emily Campbell
    • AI UX Patterns, by Luke Bennis
    • Design Patterns For Trust With AI, via Sarah Gold
    • AI Guidebook Design Patterns, by Google
    • Usable Chat Interfaces to AI Models, by Luke Wroblewski
    • The Receding Role of AI Chat, by Luke Wroblewski
    • Agent Management Interface Patterns, by Luke Wroblewski
    • Designing for AI Engineers, by Eve Weinberg

    Meet “Smart Interface Design Patterns”

    You can find more details on design patterns and UX in Smart Interface Design Patterns, our 15h-video course with 100s of practical examples from real-life projects — with a live UX training later this year. Everything from mega-dropdowns to complex enterprise tables — with 5 new segments added every year. Jump to a free preview. Use code BIRDIE to save 15% off.

    Meet Smart Interface Design Patterns, our video course on interface design & UX.


    • Video + UX Training
    • Video only

    Video + UX Training

    $ 495.00 $ 699.00

    Get Video + UX Training

    25 video lessons (15h) + Live UX Training.
    100 days money-back-guarantee.

    Video only

    $ 300.00$ 395.00


    Get the video course

    40 video lessons (15h). Updated yearly.
    Also available as a UX Bundle with 2 video courses.

    Source: Read More 

    news
    Facebook Twitter Reddit Email Copy Link
    Previous ArticleAmazon launches spec-driven AI IDE, Kiro
    Next Article Akka introduces platform for distributed agentic AI

    Related Posts

    Tech & Work

    Google integrates Gemini CLI into Zed code editor

    August 28, 2025
    Tech & Work

    10 Benefits of Integrating React.js Vibe Coding into Your Agile DevOps Pipeline

    August 28, 2025
    Leave A Reply Cancel Reply

    For security, use of Google's reCAPTCHA service is required which is subject to the Google Privacy Policy and Terms of Use.

    Continue Reading

    CVE-2025-24340 – CtrlX OS Password Disclosure Vulnerability

    Common Vulnerabilities and Exposures (CVEs)

    Microsoft Forms Was Down for Some Users; But Now Fixed

    Operating Systems

    Better CSS Shapes Using shape() — Part 3: Curves

    News & Updates

    How Red Hat just quietly, radically transformed enterprise server Linux

    News & Updates

    Highlights

    Linux

    VirtualBox 7.1.8 Adds Support for Linux Kernel 6.14

    April 15, 2025

    Ubuntu 25.04 is out this week and many will be turning to virtual machine to…

    Is ChatGPT down for you? You’re not alone – here’s the latest

    June 10, 2025

    DragonForce Ransomware Cartel Hits UK Retailers with Custom Payloads and Global Extortion Campaign

    May 5, 2025

    This mirror wraps your reflection inside Microsoft Paint — but you only have two days to order your own

    April 17, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.