Добавить новость
smi24.net
PYMNTS.com
Январь
2026
1 2 3
4
5 6 7 8 9 10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31

Commerce Finds Its Voice

0

Voice will finally pull agentic commerce onto the mobile phone by turning complex, desktop‑only “go do this for me” prompts into natural, spoken conversations that consumers can have anywhere. The platforms that win 2026 will be those that embed capable voice agents deeply into devices, apps, and operating systems, not just those that bolt AI onto legacy assistants.

Try typing a thoughtful, 40‑word prompt into your favorite LLM on a mobile phone. Not a two‑ or three‑word Google‑style query, but the kind of detailed prompt agents require to be most efficient, with constraints, preferences, trade‑offs, timing and explicit instructions. One that is eight to twenty times longer than a typical keyword search.

Even with predictive text and autocorrect, it’s a pain. Typing on a phone is slow and error‑prone. Research on mobile text entry shows error rates in the mid‑single to low‑double digits, and those rates rise as task complexity increases. At 40 words, that can mean as many as eight corrections before a prompt is even submitted, each one adding friction and cognitive load.​

That friction isn’t just annoying. It explains a great deal about where the Prompt Economy actually lives today.

Despite years and billions spent perfecting mobile‑first design, most prompt‑heavy and agent‑driven interactions still happen on desktops and laptops. Not because consumers prefer them, but because keyboards and large screens make it easier to express complex intentions. When people need to be precise, conditional, and explicit, they gravitate toward devices that make that easier.​

That’s not the mobile phone today.

For agentic commerce to migrate meaningfully to mobile, that constraint has to be addressed.

Agentic commerce needs to find its voice.

Commerce Finds Its Voice

For more than a decade, digital commerce has been organized around mobile interfaces optimized for tapping, scrolling and keyword‑based search. What’s changing now isn’t the interface itself, but how the commerce journey begins. And what consumers will expect of those experiences. Agents move shopping on mobile from “search, tap, scroll, click, compare, repeat” to “go do this for me.” Consumers will replace discovery‑driven navigation with intent‑driven execution by agents.

Consider what that actually looks like in practice. Instead of hours searching, clicking and comparing five different sites and multiple options within them, they say to their favorite agent:

“Plan and book a week-long family ski trip for two adults and two children under 14, for around $3,500, with direct flights, ski-in/ski-out lodging, transportation to and from the resort included, ski lessons and daytime activities for the kids, and charge it to the card that maximizes rewards, and if a better deal, also leverages my loyalty memberships at Delta and Marriott, and allows cancellation up to a week out.”

That’s not a query. It’s a complex chain of instructions that a consumer would ask a personal assistant to help with. A request that requires memory, judgment, optimization across merchants and services, and execution across payment rails.

The Voice Assistant That Wasn’t

When Amazon launched Alexa in 2014, the big idea was simple. People would talk to machines the way they talk to each other. From the start, my expectation was that voice would become a starting point for commerce and payments, not just a novelty for setting timers and playing music. For years, that made mine a somewhat lonely position, continually making the case that the most natural interface of all would ultimately become the front door to digital shopping, bill payment and everyday financial decisions, even as early usage data suggested otherwise.

Over the years, I was also clear about why that vision stalled. First‑generation voice assistants were good at handling commands but poor at managing conversation. Like a desktop or laptop, they required being at a specific device and using a “wake word” to initiate what was, in truth, not much of a conversation. They struggled with nuance, trade‑offs and the iterative back‑and‑forth that complicated commerce decisions require.​

Voice without context and reason wasn’t really that smart.

From Command Devices to Embedded Agents

That constraint is now fading. Large language models and agentic systems fundamentally change what voice can do. Spoken prompts can be long, conditional and iterative. Context can persist across turns. Agents can evaluate options, make decisions, and take action. Voice stops being a keyword command and becomes a conversational interface.

Just as important is where those voice agents now live. They are no longer tied to a single smart speaker on a kitchen counter. Voice is becoming embedded. In earbuds, in cars, in mobile banking and retail apps, in operating systems, and soon in dedicated audio‑first devices that pair tightly with frontier models. Instead of shouting across the room to a box on the counter, consumers will speak quietly into the microphone that is already next to their ear, in their dashboard or in their hand.​

On mobile, that shift is decisive. Voice turns the phone from a scrolling device into an instruction device, removing the single biggest barrier to complex mobile commerce: typing.

Today’s Mobile Assistants as On-ramps

As agentic capabilities mature, two distinct strategies for voice are emerging.

One approach embeds voice into systems that are already smart. In this model, the system already understands users, transactions, preferences, constraints and consequences. It knows how money moves, how risk is managed, and how exceptions are handled. Voice becomes the mechanism through which intent triggers an action, not the place where intelligence resides.

The other approach attempts to make voice itself the intelligence layer. Here, speech interfaces are treated as voice-operating systems, with memory, reasoning and orchestration layered on top of conversation. The ambition is for voice to become the place where decisions are formed rather than expressed.

We see these strategies playing out in real life.

The installed base of mobile voice assistants — Siri, Google Assistant, Bixby and Alexa on phones — has already trained consumers to talk to their devices, even if only for simple tasks like asking for the weather, setting reminders or placing a quick reorder. What changes in 2026 is not consumer willingness to speak, but the intelligence consumers find on the other end of the microphone. Amazon’s Alexa+ is attempting to reinvent itself for an agentic era after years as a command‑and‑control assistant, while Apple’s Siri, once the pioneer of voice, has visibly lagged in generative AI and remains an open question in agent‑driven commerce.​

In parallel, AI‑native platforms are racing to define the new default mobile agent. OpenAI is investing heavily in voice, consolidating teams around a new audio model and an audio‑first device effort led by Jony Ive’s hardware studio. Voice-native startups are creating voice operating systems that support already-smart AI platforms.

The opportunity for retail is well within reach. And massive. The 2025 Visa Global Digital Shopping Index, produced with PYMNTS Intelligence, documents the rise of the “mobile window shopper,” a consumer who browses on mobile multiple times per week and converts those sessions into purchases at roughly three times the rate of the typical mobile shopper, especially in higher‑margin retail categories. These consumers skew toward higher‑income households and parents.​

In other words, retail’s most valuable customers already live on mobile, with their shopping preferences only a few spoken words away.

What these consumers lack isn’t intent, since they are using AI-native platforms today to search for what they want to buy. It’s a low‑friction way to activate agentic commerce on the device they use for just about every aspect of their daily lives. While commuting, at lunch, walking the dog, running errands, at the kid’s hockey game, standing in a store aisle or talking through options while watching TV on the couch.

Agentic commerce won’t be unlocked by better buttons or more elegant screens. It will be unlocked when consumers can simply say what they want and trust that it will be done. Voice is the interface that finally makes that possible at scale, in the one place commerce actually happens: in the palm of the consumer’s hand.

The question 2026 will answer is whether consumers respond more to voice built natively into intelligent AI platforms or to attempts to make legacy voice assistants smarter.

The post Commerce Finds Its Voice appeared first on PYMNTS.com.















Музыкальные новости






















СМИ24.net — правдивые новости, непрерывно 24/7 на русском языке с ежеминутным обновлением *