Skip to content

Commit 57fc553

Browse files
espressif-botameyinamdarkedarsKedar SovaniFBEZ
authored
Sync Merge: blog/annoucing_esp_private_agents_platform (#615)
* First draft of ESP Private Agents Platform Intro blog post Signed-off-by: Amey Inamdar <[email protected]> * Change ESP Private Agents Platform to Private AI Agents Platform Signed-off-by: Amey Inamdar <[email protected]> * Minor updates with video links * Include multiple firmware variants, update YouTube links * Fix video link * Update to webp * Apply 3 suggestion(s) to 1 file(s) Co-authored-by: Francesco Bez <[email protected]> * Added featured image and removed the PNG files --------- Signed-off-by: Amey Inamdar <[email protected]> Co-authored-by: Amey Inamdar <[email protected]> Co-authored-by: Kedar Sovani <[email protected]> Co-authored-by: Kedar Sovani <[email protected]> Co-authored-by: Francesco Bez <[email protected]> Co-authored-by: pedrominatel <[email protected]>
1 parent d92778b commit 57fc553

File tree

8 files changed

+115
-0
lines changed

8 files changed

+115
-0
lines changed
-81 KB
Loading
3.37 KB
Loading
14.4 KB
Loading
12.4 KB
Loading
66.6 KB
Loading
18.4 KB
Loading
718 KB
Loading
Lines changed: 115 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,115 @@
1+
---
2+
title: "Introducing Private AI Agents Platform"
3+
date: 2025-12-09
4+
showAuthor: false
5+
authors:
6+
- "amey-inamdar"
7+
- "chirag-atal"
8+
tags:
9+
- AI
10+
- Agents
11+
- IoT
12+
- LLM
13+
- ESP32
14+
summary: "This article announces the Private AI Agent Platform"
15+
---
16+
17+
{{< figure
18+
src="esp-private-agents-logo.webp"
19+
alt="Private AI Agents Platform"
20+
>}}
21+
22+
We are proud to announce the "Private AI Agents Platform" – a self-hosted, multi-modal AI agent platform that you can deploy in your AWS account. This platform is optimized and tuned for building and managing AI Agents that can be integrated in connected devices and companion phone apps and web dashboards. In this blog post we provide an introduction to this platform.
23+
24+
{{< youtube eYz9nHDg9Ao >}}
25+
26+
## AI Agents and IoT
27+
28+
The first question we need to answer is what an AI agent is, and why it has become so relevant for IoT. An AI agent is the workflow implementation that combines large language models (LLMs) with a tool calling layer that allows to read and change device state, invoke APIs, access knowledge base and make decisions based on the context. For device makers, this unlocks a new class of capabilities such as intelligent assistants in companion apps, natural multi-lingual voice interfaces for device control, intelligent customer support that works with the combination of actual device state and authoritative product documentation, and ultimately, ambient intelligence driven by events from sensors, cloud rules and other services.
29+
30+
## High Level Architecture
31+
32+
Engineering such an agent implementation for IoT systems requires a careful consideration to modality (text, speech), scalability, security, cost and safety. Keeping all these considerations in mind, the Private AI Agents Platform is built on AWS cloud services using AWS Fargate as a primary application platform and Amazon Bedrock Foundation Models as backend LLM system. Espressif has made an instance of this available for a [public trial](https://agents.espressif.com). For the production cases, you can deploy the same in your own AWS account. Here is a high-level architecture of the system.
33+
{{< figure
34+
src="esp-private-agents-block-dia.webp"
35+
alt="Private Agents Platform Architecture"
36+
>}}
37+
38+
Now let's have a look at how we can create and use agents for various use-cases mentioned above.
39+
40+
## Defining the Agent
41+
42+
An **Agent** consists of an LLM selection, a system prompt—written in plain English—that defines its behavior, along with a set of tools that give it the ability to perform actions. The **AI Agents admin dashboard**, available at <https://agents.espressif.com>, allows you to create and configure agents using pre-built templates or by defining them entirely from scratch.
43+
44+
{{< youtube HhcYJGAx0kU >}}
45+
46+
### LLM Selection
47+
48+
The Agent definition process allows developers to choose from a range of supported **AWS Bedrock Foundation Models**, giving full control over performance, cost, and behavior. Each interaction modality—**voice** and **text**—can be assigned its own LLM, enabling optimized configurations such as lightweight models for conversational voice interfaces and more capable models for complex text reasoning. This flexibility ensures that every agent can be tuned precisely to its use-case while maintaining consistent deployment and governance within the customer's own cloud environment.
49+
50+
### System Prompt
51+
52+
Every agent—whether created from a template or built from scratch—includes a **system prompt**. The system prompt defines the agent's behavior and establishes its persona, such as a voice controller, storyteller, or customer support assistant. When you name an agent, the dashboard can automatically generate an initial system prompt using the LLM, which you may keep, modify, or replace entirely. During interactions, the user's query is combined with this system prompt, guiding the LLM to reason and respond in alignment with the agent's intended role.
53+
54+
### Tools
55+
56+
**Tools** are pluggable actions that an agent can invoke to perform specific tasks. The dashboard includes several pre-defined tools—such as ESP RainMaker control, Volume Control, and Emotion Detection—and allows you to create custom tools tailored to your application. The platform supports two categories of tools.
57+
58+
#### Remote Tools
59+
60+
The Private AI Agents Platform supports the [Model Context Protocol (MCP)](https://en.wikipedia.org/wiki/Model_Context_Protocol), enabling device makers to plug their agents into a much broader ecosystem of third-party MCP-compatible services and tools. With MCP servers, agents can seamlessly interact with external systems—such as querying a knowledge base, communicating to the device cloud, creating a customer ticket, or communicating with other services—without building custom integrations. This pluggability dramatically accelerates development and ensures agents can evolve as the ecosystem grows. The platform also supports OAuth authentication to the MCP servers, allowing securely logging in and getting the user context wherever required.
61+
62+
#### Local Tools
63+
64+
In addition to remote tools powered by MCP servers, the Private AI Agents Platform also supports **Local Tools**, which allow tool calls to be executed directly on the client—such as the IoT device itself or a companion mobile application. Local Tools enable agents to trigger real-time, device-level actions without requiring any backend integration. For example, if a smart device needs to respond to a voice command like *"Turn off the bedroom fan"*, the developers can simply expose a Local Tool such as `toggleFan` or `setSpeed`, and the agent will deliver the instruction directly to the device or app for immediate execution. This makes Local Tools ideal for tight device control loops, offline-capable interactions, and low-latency experiences, ensuring that agents can act as naturally and responsively as users expect.
65+
66+
Once an agent is defined, you can test it directly from the web dashboard to observe how it responds to queries and invokes tools in real time. The dashboard provides an interactive interface for trying out both **text** and **voice** modalities, making it easy to validate behavior, fine-tune prompts, and verify tool integrations before deploying the agent into your application.
67+
68+
## Getting Started on a Real Device: Step-by-Step
69+
70+
Once your agent is configured, you can run it on real hardware using any of the three supported development kits: **EchoEar**, **ESP32-S3-Box**, and **M5Stack Core S3**. The **agent client** is fully compatible with all three platforms, enabling you to test voice interactions, tool execution, and end-to-end behavior exactly as users would experience it.
71+
72+
<div style="display: flex; gap: 1rem; justify-content: center; margin: 2rem 0;">
73+
<figure style="flex: 1; margin: 0;">
74+
<img src="2.webp" alt="Development Kit 2" style="width: 100%;" />
75+
</figure>
76+
<figure style="flex: 1; margin: 0;">
77+
<img src="1.webp" alt="Development Kit 1" style="width: 100%;" />
78+
</figure>
79+
<figure style="flex: 1; margin: 0;">
80+
<img src="3.webp" alt="Development Kit 3" style="width: 100%;" />
81+
</figure>
82+
</div>
83+
84+
### 1. Program the firmware
85+
86+
You can flash pre-compiled firmware to these boards directly from your web browser without requiring any toolchain installation. This makes it incredibly easy to get started and validate your agent in minutes. The full source code for the firmware will also be made available soon, allowing developers to explore, customize, and extend the client implementation as needed. For instructions on flashing and using the device please refer to: <https://github.com/espressif/esp-agents-firmware>.
87+
88+
Currently, 2 kinds of firmware are supported:
89+
90+
* Generic Assistant/Friend
91+
* Voice-assissted Matter Controller with Thread support
92+
93+
### 2. Provision the device using the ESP RainMaker Home app
94+
95+
Next, open the **ESP RainMaker Home** app to configure the device's Wi-Fi credentials. The app guides you through the provisioning process, ensuring the device connects securely to your network and becomes ready to communicate with your agent.
96+
97+
### 3. Configure a new Agent into the Device
98+
99+
This is an optional step, where you can change the default Agent running on the device. Once the device is online, use the same mobile app to link it to your agent. Each agent created in the dashboard has a unique **Agent URL** represented as a QR code. Simply scan this QR code with your phone's camera, and the app will automatically write the correct **Agent ID** into the device. Note that, the device firmware has a default agent configured in it which makes it act as a "Friend" of the user.
100+
101+
### 4. Interact with the device using voice
102+
103+
With provisioning and configuration complete, your device is now fully connected to the agent. Speak naturally to the device to issue commands, ask questions, or trigger actions. The agent will process your input, reason about it, and call the appropriate tools in real time.
104+
105+
## Trying out Phone App Chat Agent
106+
107+
A completely different application of the Private Agent Platform is to directly integrate AI Agent into your connected device's companion app. The **ESP RainMaker Home** app also includes a built-in **Chat Agent** feature that demonstrates this. Any agent you create in the dashboard can be configured as the chat agent within the app, allowing you to test conversational interactions immediately with no additional development required.
108+
109+
{{< youtube gEo4JWVOQrw >}}
110+
111+
Because the chat agent can invoke **Local Tools**, it can retrieve real-time device status directly from the user's hardware, allowing support scenarios such as *"Why is my purifier showing a red light?"* or *"Check if my fan is connected."* This combination of conversational understanding and live device context enables a much more effective and personalized support experience within the companion app.
112+
113+
114+
## Conclusion
115+
In this article we introduced the Private AI Agents Platform and showed how it enabled developers to build and deploy AI agents within their own AWS environments. We outlined how agents were configured using LLMs, system prompts, and tools, and how they were tested across both device hardware and companion apps. Together, these components demonstrated how the platform supported secure and scalable AI-driven interactions for connected products.

0 commit comments

Comments
 (0)