how do we design UX for AI?
ambient intelligence, inferring intent
- we need to move beyond the constraints of the chat box
- “ambient intelligence” is key - designers must consider where “intelligence” should show up to feel natural
- system should infer intent to augment the human, like auto-complete
- there should be a collaborative experience
- the system should react to signals, like a thermostat reacts to temperature
- example: granola.ai - records conversations ambiently for summarization and recap…operates in the background
- of course, there is tension between ambience and privacy
- one must ask: “if this thing i’m building succeeds, what are the downstream consequences?”
- of course, there is tension between ambience and privacy
- we are again in “early days of UX” when it comes to AI experiences - the rules for what is good or bad have not been fleshed out yet
- we should be thinking, “what does the next generation of this technology look like?”
- there is a shift to designing reactive probabilistic experiences
hyper-personalization. dynamic, on-the-fly UI
- a vision of the future of HCI :
- AI is involved in every interface interaction.
- AI is in every widget
- there will be intelligent ‘mode’ switching, where one ‘gesture’ could have infinite meanings, implicitly determined by AI through context
- any pixel can be UI - no more CSS, every single aspect of UI can be dynamically altered, including content structure, mental models, etc.
- proposed scenarios
- write without ever typing - interactions like ‘rotating’ a paragraph to automatically reorganize sentence structure
- everything is dynamic - properties are dynamically determined, text boxes have generative autocomplete
- See: AI-Instruments Paper
- drawing canvases that autocomplete
- defining new art brushes with text
- one gesture, infinite meanings
- given some lineart, a scribble over the sky could mean color-fill the sky, while a scribble over a lamp could mean to remove that lamp from the drawing
- AI predicts your intention
- Sequence of interactions and system state inform context
- questions:
- how do we embrace ambiguity for delivering magical experiences?
- what are novel feedforward and negotiation mechanisms?
- what are robust affordances and semantic anchors?
- e.g. a book page intuitively indicates a swiping interaction
- “Your interface, your dance partner”
- co-adaptive, hyper-personalized experiences
- “the perfect app for you”
- ^ a point brought up: how is collaboration changed or inhibited when everybody sees a view tailored only for them?
leveraging AI for business
- it is most important to understand the business goals, in order to express opinions and have real influence
- digital spaces should have goal-driven, dynamic navigation - high priority tasks should be dynamically surfaced and easily navigable
- system should understand what task the user is doing, based on where they click, etc.
- products have been designed around personas, structured around “jobs to be done”, defining typical tasks and pain points
- AI is changing what jobs should be done - personas may no longer be future-proof
- shift away from persona to “goals to be done”
- classify tasks as tedious, fulfilling, etc. and delegate agents
AI for creativity
-
to use AI as a creative, thinking partner, it needs enough context on you - it needs human-scale, long memory → “what deserves to be remembered?”
-
one experimental method: routine dual journaling, where user and AI both journal their reflections on their collaborations - builds long-term memory - an imitation of consciousness, cognitive roleplay, simulation → simulacra
- reflection instead of summary
-
LyricStudio - a tool assisting musicians with writing lyrics - next-line generative complete - but not intended to grow a dependency or be addictive
- users report having lasting creativity boosts even after ceasing to use the app
-
this should be a goal of AI systems: amplify human creativity instead of replacing it - non-addictive AI systems
-
There is no objective truth, and LLMs should not try to become a “100% accurate”, infallible source - we should not deprive people of the search for truth
-
Hallucinations = willingness to be wrong, which is arguably something that should be encouraged in creative contexts - we should be taking advantage of LLMs in exploring the whole probabilistic space, rather than focusing on ‘convergence’ and steering it down the most probabilistic path
AI for ‘companionship’ - supporting a journey, not one task
- human journeys are continuous but,
- “today, we design AI moments in product silos”
- but users care about navigating towards goals
- they are seeking (exploring options, information gathering) → striving (goal-setting, prioritizing) → shifting (reflecting, reframing)
- AI should be able to proactively help with, and adjust to, the context of the person’s journey
- Scene recognition
- Sentiment awareness
- Signals should be ‘linked across these silos’ - frame AI interactions by journey instead of by app
- again, all about inferring from context
considering consequences, trust and mistakes
-
LLMs must own mistakes and understand the root cause
-
Important to understand:
- what types of mistakes does AI make?
- how easy is it to spot those mistakes?
- it’s hard to fact-check LLM outputs!
- what are the consequences of those mistakes?
-
See: When combinations of humans and AI are useful - A systematic review and meta-analysis
- Interestingly, studies find that human+AI combos can be less productive than human or AI acting alone
-
We should be designing for error - mitigate overreliance, help people recognize errors
- example: one study highlighted tokens with low generation probability
- see: HAX toolkit design library
- need visibility into workings of AI agents
-
when LLMs provide ‘reasons’ for their answers this can actually increase overreliance because humans are more inclined to trust it when it could still be wrong
-
need multimodal validation
-
it all goes back to the question: “is the user able to accomplish their goal?”
-
“pathetic fallacy” - projecting human-ness onto inanimate objects - humans will form emotional bonds with AI regardless of their perceived sentience
-
“AI personification” = giving AI human attributes - this should not be pursued unless truly necessary for a specific intent, clearly differentiable from real humans, and bias resistant
-
currently, AI agents can be difficult to differentiate from people when inserted into workflows, which causes problems or controversy