Skip to content

Conversation

@rainbowFi
Copy link
Contributor

Description

Provide information and an example to explain how customers can predict their costs when using AIT

Checklist

@coderabbitai
Copy link

coderabbitai bot commented Jan 15, 2026

Important

Review skipped

Auto reviews are disabled on this repository.

Please check the settings in the CodeRabbit UI or the .coderabbit.yaml file in this repository. To trigger a single review, invoke the @coderabbitai review command.

You can disable this status message by setting the reviews.review_status to false in the CodeRabbit configuration file.


Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share

Comment @coderabbitai help to get the list of available commands and usage tips.

@rainbowFi rainbowFi marked this pull request as draft January 15, 2026 17:07
@rainbowFi rainbowFi marked this pull request as ready for review January 15, 2026 22:19
@rainbowFi rainbowFi added the review-app Create a Heroku review app label Jan 15, 2026
@ably-ci ably-ci temporarily deployed to ably-docs-ait-149-prici-dkbttu January 15, 2026 22:20 Inactive
@ably-ci ably-ci temporarily deployed to ably-docs-ait-149-prici-dkbttu January 16, 2026 10:02 Inactive
- the number of subscribers receiving the response.
- the [token streaming pattern](/docs/ai-transport/features/token-streaming#token-streaming-patterns) you choose.

For example, an AI support chatbot sending a response of 250 tokens at 70 tokens/s to a single client using the [message-per-response](/docs/ai-transport/features/token-streaming/message-per-response) pattern would consume 90 inbound messages, 90 outbound messages and 90 persisted messages. See the [AI support chatbot pricing example](/docs/platform/pricing/examples/ai-chatbot) for a full breakdown of the costs in this scenario.
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
For example, an AI support chatbot sending a response of 250 tokens at 70 tokens/s to a single client using the [message-per-response](/docs/ai-transport/features/token-streaming/message-per-response) pattern would consume 90 inbound messages, 90 outbound messages and 90 persisted messages. See the [AI support chatbot pricing example](/docs/platform/pricing/examples/ai-chatbot) for a full breakdown of the costs in this scenario.
For example, suppose an AI support chatbot sends a response of 250 tokens, each as a discrete update, using the [message-per-response](/docs/ai-transport/features/token-streaming/message-per-response) pattern, and that there is a single client subscribed to that channel. This will result in usage of 250 inbound messages, 250 outbound messages and 250 persisted messages. See the [AI support chatbot pricing example](/docs/platform/pricing/examples/ai-chatbot) for a full breakdown of the costs in this scenario.


### Effect of append rollup

The calculation above uses the default append rollup window of 40ms, chosen to control costs with minimum impact on responsiveness. For a text chatbot use case, you could increase the window to 200ms without noticably impacting the user experience.
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The calculation above uses the default append rollup window of 40ms

Does it? Where is that in the calculation?
70 updates/s = ~14ms per token, so rollup will be happening if there's a 40ms interval

@ably-ci ably-ci temporarily deployed to ably-docs-ait-149-prici-dkbttu January 16, 2026 12:28 Inactive
@mschristensen mschristensen force-pushed the ait-251-overview branch 2 times, most recently from 13a9c10 to c3d90cc Compare January 16, 2026 13:20
@ably-ci ably-ci temporarily deployed to ably-docs-ait-149-prici-dkbttu January 16, 2026 13:43 Inactive
|---------------|---------------------------|----------------|------|
| 40ms | 360 per chat | 1092M | $2730.00/M chats |
| 100ms | 144 per chat | 444M | $1110.00/M chats |
| 200ms | 72 per chat | 228M | $570.00/M chats |
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Why are we asking a customer to decide what token rate they want?

We are the experts on AI UX, we are building a product to solve this problem, this should not be something we push down to developers to decide on. There is no benefit to this approach whatsoever.

Base automatically changed from ait-251-overview to AIT-129-AIT-Docs-release-branch January 16, 2026 14:11
@ably-ci ably-ci temporarily deployed to ably-docs-ait-149-prici-dkbttu January 16, 2026 15:28 Inactive
@ably-ci ably-ci temporarily deployed to ably-docs-ait-149-prici-dkbttu January 16, 2026 15:44 Inactive
Copy link
Member

@paddybyers paddybyers left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

lgtm

- the number of subscribers receiving the response.
- the [token streaming pattern](/docs/ai-transport/token-streaming#token-streaming-patterns) you choose.

For example, suppose an AI support chatbot sends a response of 300 tokens, each as a discrete update, using the [message-per-response](/docs/ai-transport/token-streaming/message-per-response) pattern, and with a single client subscribed to the channel. With AI Transport's [append rollup](/docs/ai-transport/messaging/token-rate-limits#per-response), this will result in usage of 100 inbound messages, 100 outbound messages and 100 persisted messages. See the [AI support chatbot pricing example](/docs/platform/pricing/examples/ai-chatbot) for a full breakdown of the costs in this scenario.
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
For example, suppose an AI support chatbot sends a response of 300 tokens, each as a discrete update, using the [message-per-response](/docs/ai-transport/token-streaming/message-per-response) pattern, and with a single client subscribed to the channel. With AI Transport's [append rollup](/docs/ai-transport/messaging/token-rate-limits#per-response), this will result in usage of 100 inbound messages, 100 outbound messages and 100 persisted messages. See the [AI support chatbot pricing example](/docs/platform/pricing/examples/ai-chatbot) for a full breakdown of the costs in this scenario.
For example, suppose an AI support chatbot sends a response of 300 tokens, each as a discrete update, using the [message-per-response](/docs/ai-transport/token-streaming/message-per-response) pattern, and with a single client subscribed to the channel. With AI Transport's [append rollup](/docs/ai-transport/messaging/token-rate-limits#per-response), those 300 input tokens will be conflated giving rise to 100 discrete inbound messages, resulting in 100 outbound messages and 100 persisted messages. See the [AI support chatbot pricing example](/docs/platform/pricing/examples/ai-chatbot) for a full breakdown of the costs in this scenario.

@ably-ci ably-ci temporarily deployed to ably-docs-ait-149-prici-dkbttu January 16, 2026 17:03 Inactive
@rainbowFi rainbowFi merged commit 42916b2 into AIT-129-AIT-Docs-release-branch Jan 16, 2026
1 of 2 checks passed
@rainbowFi rainbowFi deleted the ait-149-pricing branch January 16, 2026 17:04
@rainbowFi rainbowFi removed the review-app Create a Heroku review app label Jan 16, 2026
mschristensen pushed a commit that referenced this pull request Jan 16, 2026
Add information about pricing to the overview page and include a worked example for illustration
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Development

Successfully merging this pull request may close these issues.

5 participants