Version: 3.x

Getting started with Analytics

Visualise and process Rasa assistant metrics in the tooling of choice.

Rasa Pro Only
Rasa Pro License

You'll need a license to get started with Rasa Pro. Talk with Sales

Analytics Data Pipeline helps visualize and process Rasa assistant metrics in the tooling (BI tools, data warehouses) of your choice. Visualizations and analysis of the production assistant and its conversations allow you to assess ROI and improve the performance of the assistant over time.

Rasa Pro Analytics - Sessions per channelRasa Pro Analytics - Escalation rateRasa Pro Analytics - Intents distribution over timeRasa Pro Analytics - Top intents

Measuring a Rasa assistant's success and making strategic investments will lead to better business outcomes. It will also help you understand the needs of your end users and better serve them.

An overview of the components of Rasa Pro.

Rasa Pro will connect to your production assistant using Kafka. The pipeline will compute analytics as soon as the docker container is successfully deployed and connected to your data warehouse and Kafka instances.

Types of metrics

Metrics collected from your assistant can broadly be categorized as

  • User Analytics: Who are the users of the assistant, and how do they feel about it? Examples: demographics, channels, sentiment analysis
  • Usage Analytics: How is the assistant’s overall health and what kind of traffic is coming to it? Examples: total number of sessions, time per session, errors and error rates
  • Conversation Analytics: What happened during the conversation? Examples: number of messages sent, abandonment depth, number of topics introduced by user, top N intents
  • Business Analytics: How is the assistant performing with regard to business goals? Examples: ROI of assistant per LoB, time comparison of assistant vs agent, containment rate

In this version of the Analytics pipeline, measurement of the following metrics is possible

Number of conversationsUsage AnalyticsTotal number of conversations
Number of usersUsage AnalyticsTotal number of users
Number of sessionsUsage AnalyticsGross traffic to assistant
Channels usedUsage AnalyticsSessions by channel
User session countUser AnalyticsTotal number of user sessions or average sessions per user
Top N intentsConversation AnalyticsTop intents across all users
Avg response timeConversation AnalyticsAverage response time for assistant
Containment rateBusiness Analytics% of conversations handled purely by assistant (not handed to human agent
Abandonment rateBusiness Analytics% of abandoned conversations
Escalation rateBusiness Analytics% of conversations escalated to human agent

For examples of how you can extract these metrics, see Example queries.


  • A production deployment of Kafka is required to set up Rasa Pro. We recommend using Amazon Managed Streaming for Apache Kafka.

  • A production deployment of a data lake needs to be connected to the data pipeline. Rasa Pro directly supports the following data lakes:

    Virtually any other data lakes can be configured to sync with your deployment of PostgreSQL. You can find additional instructions on how to connect your PostgreSQL deployment to either BigQuery or Snowflake in the Connect a data warehouse step.

    We recommend managed deployments of your data lake to minimize maintenance efforts.

1. Connect an assistant

To connect an assistant to Rasa Pro Services, you need to connect the assistant to an event broker. The assistant will stream all events to the event broker, which will then be consumed by Rasa Pro Services.

The configuration of the assistant is the first step of Installation and Configuration. No additional configuration is required to connect the assistant to the Analytics pipeline. After the assistant is deployed, the Analytics pipeline will receive the data from the assistant and persist it to your data warehouse which will be configured in the next step.

2. Connect a data warehouse

You can choose to configure the analytics pipeline to stream the transformed conversational data to two different data warehouse types in AWS:

Additionally, any data warehouse is supported if it is configured in sync with your deployment of PostgreSQL, as recommended for Redshift.

We have included brief instructions for syncing your PostgreSQL deployment with:


You can use Amazon Relational Database Service (RDS) to create a PostgreSQL DB instance which is the environment that will run your PostgreSQL database.

First, you must set up Amazon RDS by completing the instructions listed here. Next, create the PostgreSQL DB instance. You can follow one of the following instruction sets:

To build the DB URL in the DBAPI format specified in the Connect an assistance section you must enter the database credentials. You must obtain the database username and password after you select a database authentication option during the process of the PostgreSQL DB instance creation:

Finally, when running the Analytics pipeline Docker container, set the environment variable RASA_ANALYTICS_DB_URL to the PostgreSQL Amazon RDS DB instance URL.


If you prefer instead to set up Amazon Redshift as your choice of data lake, you can choose to either stream the data from the PostgreSQL source database within the Amazon RDS DB instance created at the PostgreSQL step to the Redshift target or connect directly to Amazon Redshift.

Streaming from PostgreSQL to Redshift

If you meet the prerequisites for using an Amazon Redshift database as a target, you will need to implement two steps:

  1. configure the PostgreSQL source for AWS Database Migration Service (DMS) by following these instructions
  2. configure the Redshift target for AWS DMS following the instructions here.

Direct connection

You can get started on enabling the direct connection with Redshift by following these resources on creating an Amazon Redshift cluster. Before you do so, you should take into account that streaming historical data directly to Redshift could take much longer than streaming directly to or via the PostgreSQL RDS instance.

You must update the RASA_ANALYTICS_DB_URL to the Redshift cluster DB URL which must follow the following format:

redshift://<USER>:<PASSWORD>@<AWS URL>:5439/<DB NAME>

For example:

Redshift write performance

As a result of performance considerations, we strongly advise against choosing a direct connection to Redshift. Redshift is a great data lake for analytics but lacks the necessary write performance to directly stream data to it.


To stream data from PostgreSQL to BigQuery, you can use Datastream for BigQuery. Datastream for BigQuery supports several PostgreSQL deployment types, including CloudSQL.

Before you begin, make sure to check the Datastream prerequisites, as well as additional Datastream networking connectivity requirements.

You can closely follow this quickstart guide on replicating data from PostgreSQL CloudSQL to BigQuery with Datastream.

Alternatively, you can deep dive into the following Datastream set-up guides:


You can sync your PostgreSQL deployment manually or via an automated partner solution.

The instructions for manual sync include the following steps:

  • Extract data from PostgreSQL to file using COPY INTO command. You should also explore the Snowflake data loading best practices before extraction.
  • Stage the extracted data files to either internal or external locations such as AWS S3, Google Cloud Storage or Microsoft Azure.
  • Copy staged files to Snowflake tables using COPY INTO command. You can decide to use bulk data loading into Snowflake or to load continuously using Snowpipe. Alternatively you can also benefit from this plugin to load data to an existing Snowflake table.

3. Ingest past conversations (optional)

When Analytics is connected to your Kafka instance, it will consume all prior events on the Kafka topic and ingest them into the database. Kafka has a retention policy for events on a topic which defaults to 7 days.

If you want to process events from conversations that are older than the retention policy configured for the Rasa topic, you can manually ingest events from past conversations.

Manually ingesting data from past conversations requires a connection to the tracker store. The tracker store contains past conversations and a connection to the Kafka cluster. Use the rasa export command to export the events stored in the tracker store to Kafka:

rasa export --endpoints endpoints.yml

Configure the export to read from your production tracker store and write to Kafka as an event broker, e.g.

type: SQL
dialect: "postgresql"
url: "localhost"
db: "tracker"
username: postgres
password: password
type: kafka
topic: rasa-events
url: localhost:29092
partition_by_sender: true

Running manual ingestion of past events multiple times will result in duplicated events. There is currently no deduplication implemented in Analytics. Every ingested event will be stored in the database, even if it was processed previously.

4. Connect a BI Solution

Connecting a business intelligence platform to the data warehouse varies for each platform. We provide example instructions for Metabase and Tableau but you can use any BI platform which supports AWS Redshift or PostgreSQL.

Example: Metabase

Metabase is a free and open-source business intelligence platform. It provides a simple interface to query and visualize data. Metabase can be connected to PostgreSQL or Redshift databases.

Example: Tableau

Tableau is a business intelligence platform. It provides a flexible interface to build business intelligence dashboards. Tableau can be connected to PostgreSQL or Redshift databases.