Unlimited
No rate-limits, no censorship, and unlimited token generation.
Zero-log
No-logging policy to keep your data safe.
Money-back guarantee
Flat monthly pricing with money-back guarantee if you are not satisfied.
The most unrestricted LLM Platform.
Assistant
Chat with an LLM to use as a personal assistant.
AI Agents
Create agentic workflows using LLMs without worrying about token use.
Roleplay
Chat with your AI companions without being censored or counting your tokens. Sillytavern, RisuAI, and other RP frontends supported!
Data Processing
Run through your data with any LLM without any limits or censorship.
Code Completion
Connect to code completion plugins and write code faster.
Applications
Create AI-powered applications cost-effectively.
Frequently asked questions
How can there be no limits?
With our pricing strategy, we charge based on parallel requests which allows us to easily calculate and scale how many GPUs we need.
How do I use this?
Arli AI is an LLM inference platform that provides API and chat access to interact with the latest models. Check out our quick-start page in our docs!
How do I contact Arli AI support?
You can contact us at contact@arliai.comm or through our contact us form.
Do you keep logs of prompts and generation?
We strictly do not keep any logs of user requests or generations.
Why is Arli AI better than other LLM providers?
We provide the most unrestricted LLM platform with no rate-limits to tokens or requests on top of no censorship.
Is there a hidden limit imposed?
We only have a limit of how many parallel requests a user can make as explained in the pricing page.
Why use Arli AI API instead of self-hosting LLMs?
Using Arli AI will cost you significantly less than other inference platforms, not to mention renting GPUs in the cloud or paying electricity to run your own GPUs.
What if I want to use a model that's not here?
If a model you want to use is not in our Models page, you can contact us to request to add it.