-
Notifications
You must be signed in to change notification settings - Fork 46
AIT-149: Add pricing information to docs #3106
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
Important Review skippedAuto reviews are disabled on this repository. Please check the settings in the CodeRabbit UI or the You can disable this status message by setting the Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out. Comment |
0be0654 to
56a0dfc
Compare
| - the number of subscribers receiving the response. | ||
| - the [token streaming pattern](/docs/ai-transport/features/token-streaming#token-streaming-patterns) you choose. | ||
|
|
||
| For example, an AI support chatbot sending a response of 250 tokens at 70 tokens/s to a single client using the [message-per-response](/docs/ai-transport/features/token-streaming/message-per-response) pattern would consume 90 inbound messages, 90 outbound messages and 90 persisted messages. See the [AI support chatbot pricing example](/docs/platform/pricing/examples/ai-chatbot) for a full breakdown of the costs in this scenario. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
| For example, an AI support chatbot sending a response of 250 tokens at 70 tokens/s to a single client using the [message-per-response](/docs/ai-transport/features/token-streaming/message-per-response) pattern would consume 90 inbound messages, 90 outbound messages and 90 persisted messages. See the [AI support chatbot pricing example](/docs/platform/pricing/examples/ai-chatbot) for a full breakdown of the costs in this scenario. | |
| For example, suppose an AI support chatbot sends a response of 250 tokens, each as a discrete update, using the [message-per-response](/docs/ai-transport/features/token-streaming/message-per-response) pattern, and that there is a single client subscribed to that channel. This will result in usage of 250 inbound messages, 250 outbound messages and 250 persisted messages. See the [AI support chatbot pricing example](/docs/platform/pricing/examples/ai-chatbot) for a full breakdown of the costs in this scenario. |
|
|
||
| ### Effect of append rollup | ||
|
|
||
| The calculation above uses the default append rollup window of 40ms, chosen to control costs with minimum impact on responsiveness. For a text chatbot use case, you could increase the window to 200ms without noticably impacting the user experience. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The calculation above uses the default append rollup window of 40ms
Does it? Where is that in the calculation?
70 updates/s = ~14ms per token, so rollup will be happening if there's a 40ms interval
13a9c10 to
c3d90cc
Compare
| |---------------|---------------------------|----------------|------| | ||
| | 40ms | 360 per chat | 1092M | $2730.00/M chats | | ||
| | 100ms | 144 per chat | 444M | $1110.00/M chats | | ||
| | 200ms | 72 per chat | 228M | $570.00/M chats | |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Why are we asking a customer to decide what token rate they want?
We are the experts on AI UX, we are building a product to solve this problem, this should not be something we push down to developers to decide on. There is no benefit to this approach whatsoever.
f9b4362 to
c46b859
Compare
paddybyers
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
lgtm
| - the number of subscribers receiving the response. | ||
| - the [token streaming pattern](/docs/ai-transport/token-streaming#token-streaming-patterns) you choose. | ||
|
|
||
| For example, suppose an AI support chatbot sends a response of 300 tokens, each as a discrete update, using the [message-per-response](/docs/ai-transport/token-streaming/message-per-response) pattern, and with a single client subscribed to the channel. With AI Transport's [append rollup](/docs/ai-transport/messaging/token-rate-limits#per-response), this will result in usage of 100 inbound messages, 100 outbound messages and 100 persisted messages. See the [AI support chatbot pricing example](/docs/platform/pricing/examples/ai-chatbot) for a full breakdown of the costs in this scenario. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
| For example, suppose an AI support chatbot sends a response of 300 tokens, each as a discrete update, using the [message-per-response](/docs/ai-transport/token-streaming/message-per-response) pattern, and with a single client subscribed to the channel. With AI Transport's [append rollup](/docs/ai-transport/messaging/token-rate-limits#per-response), this will result in usage of 100 inbound messages, 100 outbound messages and 100 persisted messages. See the [AI support chatbot pricing example](/docs/platform/pricing/examples/ai-chatbot) for a full breakdown of the costs in this scenario. | |
| For example, suppose an AI support chatbot sends a response of 300 tokens, each as a discrete update, using the [message-per-response](/docs/ai-transport/token-streaming/message-per-response) pattern, and with a single client subscribed to the channel. With AI Transport's [append rollup](/docs/ai-transport/messaging/token-rate-limits#per-response), those 300 input tokens will be conflated giving rise to 100 discrete inbound messages, resulting in 100 outbound messages and 100 persisted messages. See the [AI support chatbot pricing example](/docs/platform/pricing/examples/ai-chatbot) for a full breakdown of the costs in this scenario. |
20df5cb to
01ab0f8
Compare
Co-authored-by: Paddy Byers <paddy.byers@gmail.com>
20f161a to
91ef836
Compare
42916b2
into
AIT-129-AIT-Docs-release-branch
Add information about pricing to the overview page and include a worked example for illustration
Description
Provide information and an example to explain how customers can predict their costs when using AIT
Checklist