May 22, 2025
Schaun Wheeler

Why Agentic Systems Don't Operate Over a Huge State Space

May 22, 2025
Schaun Wheeler

Why Agentic Systems Don't Operate Over a Huge State Space

May 22, 2025
Schaun Wheeler

Why Agentic Systems Don't Operate Over a Huge State Space

May 22, 2025
Schaun Wheeler

Why Agentic Systems Don't Operate Over a Huge State Space

Agentic systems don't operate over a huge, chaotic message space. They operate over structured action sets — defined semantic categories that make up a treatment policy.

At Aampe, a typical treatment policy might be composed from action sets: day of week, time of day, channel, value proposition, product/offering, tone of voice, incentive level/type, and so on.

Each set has maybe 3 to 20 possible actions. So yes, the space is combinatorially large—but it's semantically tractable. The agent's task isn't to predict the perfect message. It's to select the best combination of these features for a given user at a given moment. That policy then determines what content gets sent.

The main complexity of user engagement is in the user space - all of the different events that can happen, either individually or in sequence, in their journey though the app. Agents avoid getting bogged down in that complexity by modeling the action space instead of the user space. The user space gets handled via policy selection rather than direct modeling.

That's not just a design choice. It's the reality of online, real-time, organic human behavior. Every customer's behavior is messy, dynamic, and only partially observable. Some users are effectively cold starts. Others have interacted enough that the agent has built confidence in their preferences.

Agentic learners track this directly: every action is represented as a beta distribution, continuously updated per user. So for any feature — say, "push" vs. "email," or "high discount" vs. "low urgency" — the agent knows

  1. what it expects the reward to be, and

  2. how confident it is in that estimate

When the agent selects a policy, it's balancing expected value and *uncertainty. Some actions are taken because they're strong bets. Others are taken because uncertainty is high and requires exploration.

Because treatments are structured, and every send is tied to a selected policy, we can analyze outcomes in detail:

  • Which features are being chosen confidently vs. tentatively?

  • How often exploratory actions outperform exploitative ones?

  • How does agent learning translate into measurable gains over time?

This is the core advantage of agentic personalization: it doesn't just deliver content. It runs structured experiments, learns revealed preferences, and adapts policy decisions at the feature level for every individual user.

The table below shows the percent difference between explore-mode messages and exploit-mode messages for one of our customers. The columns of the table represent eight different use cases agents were assigned to orchestrate, and each row represents a different app event. The higher the value (and the more blue the cell), the more exploit-mode messages outperformed explore-mode messages.

0

Related

Shaping the future of marketing with Aampe through innovation, data.

Jun 16, 2025

Schaun Wheeler

Discover how Aampe's agents employ causal analysis to accurately measure user engagement outcomes, even when influenced by external messages. By isolating the effects of their own actions from other variables, Aampe ensures precise attribution and effective decision-making in a complex messaging environment.

Jun 16, 2025

Schaun Wheeler

Discover how Aampe's agents employ causal analysis to accurately measure user engagement outcomes, even when influenced by external messages. By isolating the effects of their own actions from other variables, Aampe ensures precise attribution and effective decision-making in a complex messaging environment.

Jun 16, 2025

Schaun Wheeler

Discover how Aampe's agents employ causal analysis to accurately measure user engagement outcomes, even when influenced by external messages. By isolating the effects of their own actions from other variables, Aampe ensures precise attribution and effective decision-making in a complex messaging environment.

Jun 16, 2025

Schaun Wheeler

Discover how Aampe's agents employ causal analysis to accurately measure user engagement outcomes, even when influenced by external messages. By isolating the effects of their own actions from other variables, Aampe ensures precise attribution and effective decision-making in a complex messaging environment.

Jun 10, 2025

Schaun Wheeler

Explore how agentic systems define and execute decisions. This article delves into the five key decision types—Go/No-Go, Context, Creative Policy, Item Recommendation, and Freshness—that guide autonomous agents in delivering personalized user experiences. Learn how these systems prioritize meaningful choices to enhance engagement and effectiveness.

Jun 10, 2025

Schaun Wheeler

Explore how agentic systems define and execute decisions. This article delves into the five key decision types—Go/No-Go, Context, Creative Policy, Item Recommendation, and Freshness—that guide autonomous agents in delivering personalized user experiences. Learn how these systems prioritize meaningful choices to enhance engagement and effectiveness.

Jun 10, 2025

Schaun Wheeler

Explore how agentic systems define and execute decisions. This article delves into the five key decision types—Go/No-Go, Context, Creative Policy, Item Recommendation, and Freshness—that guide autonomous agents in delivering personalized user experiences. Learn how these systems prioritize meaningful choices to enhance engagement and effectiveness.

Jun 10, 2025

Schaun Wheeler

Explore how agentic systems define and execute decisions. This article delves into the five key decision types—Go/No-Go, Context, Creative Policy, Item Recommendation, and Freshness—that guide autonomous agents in delivering personalized user experiences. Learn how these systems prioritize meaningful choices to enhance engagement and effectiveness.

Jun 5, 2025

Schaun Wheeler

Explore Aampe's innovative approach to personalization, combining classical recommender systems for item selection with real-time reinforcement learning agents for dynamic message composition. Learn how this dual-path strategy enhances user engagement and content relevance.

Jun 5, 2025

Schaun Wheeler

Explore Aampe's innovative approach to personalization, combining classical recommender systems for item selection with real-time reinforcement learning agents for dynamic message composition. Learn how this dual-path strategy enhances user engagement and content relevance.

Jun 5, 2025

Schaun Wheeler

Explore Aampe's innovative approach to personalization, combining classical recommender systems for item selection with real-time reinforcement learning agents for dynamic message composition. Learn how this dual-path strategy enhances user engagement and content relevance.

Jun 5, 2025

Schaun Wheeler

Explore Aampe's innovative approach to personalization, combining classical recommender systems for item selection with real-time reinforcement learning agents for dynamic message composition. Learn how this dual-path strategy enhances user engagement and content relevance.

Jun 3, 2025

Schaun Wheeler

Explore why focusing solely on short-term metrics like lift can be misleading when assessing adaptive systems, and discover alternative approaches for meaningful evaluation.

Jun 3, 2025

Schaun Wheeler

Explore why focusing solely on short-term metrics like lift can be misleading when assessing adaptive systems, and discover alternative approaches for meaningful evaluation.

Jun 3, 2025

Schaun Wheeler

Explore why focusing solely on short-term metrics like lift can be misleading when assessing adaptive systems, and discover alternative approaches for meaningful evaluation.

Jun 3, 2025

Schaun Wheeler

Explore why focusing solely on short-term metrics like lift can be misleading when assessing adaptive systems, and discover alternative approaches for meaningful evaluation.

Load More

Load More

Load More

Load More