βPremium Features
Description of the different premium features
Get a Taste Tier
Skip The Waiting Line
Our waiting room allows limiting how many people can use the platform at the same time and allows us to make sure enough hardware is provisioned to support our active users. We can offer our service for free thanks to premium subscribers that are willing to support financially the platform. This benefit let you skip the waiting line every time you want to chat with our bots.
Memory Manager
While you don't get access to Semantic Memory 2.0, you will soon get access to the Memory Manager that let you add memories that you believe are important to your conversation. These memories must be added manually and are not created automatically.
True Supporter Tier
4K Context (Memory)
When you chat with our bots, we generate a text prompt that is composed of the bot personality and example dialogues and a portion of your conversation history. The longer the prompt, the more server resources it uses. Our current model has a limit of 4096 tokens (words) but for our Free Tier, we limit that to 2048 tokens.
When calculating total tokens (words) we need to account for both input and output, so assuming an allowed output of 180 tokens and a bot definition of 900 tokens, in that example, our free tier users would have 1000 tokens (words) to fit a few of the last turns of their conversation.
In that same example, premium subscribers would have 3048 tokens (words) to fit the conversation, so that means 3x more of the conversation history can be included in the prompt.
That means the response can better leverage what was discussed before.
Longer Responses
Thanks to the 4096 tokens, True Supporters, and I'm All In tiers benefit from a max response of up to 300 tokens instead of the default 180. This means less often will the response appear to be incomplete by having been truncated.
Semantic Memory
Even with 4K Context, because we need to fit our prompt within the 4096 tokens that we have available, we can only include a portion of your conversation history. With semantic memory, we try to find a semantically (based on meaning) relevant portion of your previous conversation and include these tidbits into the prompt, even if they are not the most recent things discussed.
An example of this would be if you were discussing the details of a particular book and then shifted the conversation to a different topic, like music. If later in the conversation you refer back to the book, even if several turns have occurred since the book was last mentioned, the messages the most relevant about the book would get added to the prompt.
The advantage of semantic memory is that it doesnβt solely rely on the recentness of the conversation but also its relevance.
Semantic Memory 2.0
Semantic Memory 2.0 is now available to True Supporters. Please read about it here
Conversation Images
This new feature allows generating images within your conversation. Our AI will use the bot image, bot definition and the last turn of your conversation to generate images on request. You can read more about this feature.
As a True Supporter, you'll have the capability to generate images on all trained chatbots; however, the ability to train new chatbots to create images in conversation requires the I'm All In tier.
I'm All In Tier
Priority Generation Queue
This benefit ensures that your requests are given priority over other users. Whenever you initiate a conversation or send a message, your request is placed at the front of the queue for immediate processing. This results in faster response times and a smoother chatting experience, especially during peak usage hours or when server demand is high.
Conversation Images on Private Chatbots
You can train the AI on your private chatbots images so that they can generate conversation images.
Generation Settings
You can experiment by controlling inference temperature, top_p, and top_k. This is for the most advanced users that like to experiment.
Access to Advanced Models
You get access to over 10 different models, each designed for specific needs and preferences. These models range from 8 billion to 141 billion parameters. For instance, the WizardLM-141B model is 17 times more powerful than our default model, offering unmatched performance and capabilities.
These models are optimized for various purposes, allowing you to choose the one that best suits your role-play style and taste.
For a limited time, you'll get to experiment with our very powerful model, WizardLM-141B.
Why is WizardLM-2 special?
Multilingual Capabilities: Fluent in English, French, Italian, German, and Spanish. It'll beat all the models that you have tried so far.
Optimized for Reasoning: WizardLM-2 is optimized for reasoning, which makes it perfect for engaging and deep role-playing.
8k Context (Memory)
You get access to 8192 tokens, doubling your bot memory over what can fit in 4K context!
Last updated