Unlimited
No rate-limits, no censorship, and unlimited token generations.
Zero-log
Absolutely no logs are kept of requests or generations.
Money-back guarantee
Flat monthly pricing with money-back guarantee if you are not satisfied.
The most unrestricted LLM Platform.
Frequently asked questions
What is Arli AI?
Arli AI is a cost-effective unlimited generations LLM Inference API platform with a zero-log policy.
How can there be no limits?
We do not have limits to the number of tokens or requests processed for any of our plan. However, we do limit the number of parallel requests an account can make and response speeds are dependant on current request load on our API.
How are the response speeds?
Our regular plans (Starter, Core, Advanced, Professional, Ultimate) are subject to load balancing adjustments in order to provide acceptable response speeds for all users. If you require faster and more consistent generations, contact us for a custom plan tailored to your needs.
Do you keep logs of prompts and generation?
We strictly do not keep any logs of user requests or generations. User requests and the responses never touch storage media.
How do you have so many models?
We extract high-rank LoRA for the finetuned models. This is allows us to hotswap LoRAs on the fly as needed while maintaining near-lossless performance.
What quantization do you use for the models?
All of our base models (no - suffix) are using Compressed-Tensors INT8 W8A8 quantization with the finetuned models (with - suffix) being high-rank FP16 LoRAs applied on top.
Can I use it with x frontend?
As long as the frontend supports using the standard OpenAI API endpoint, then it will work with Arli AI. As a bonus, we also have per-API-key parameter overrides that lets you set inference parameters even if the frontend does not support it.
What is Midtrans? Is there another way to pay?
We are based in Indonesia, and therefore use Midtrans which is an Indonesian payment processor. If you have issues paying via Midtrans, we can help you with different methods of payment.
Why use Arli AI API instead of self-hosting LLMs?
We have the most models on offer compared to other providers, and using Arli AI will cost you significantly less than paying for rented GPUs or paying for electricity to run your own GPUs.
Where do I find the latest updates?
Join our subreddit and our discord server where we regularly post updates and discuss the models!
What if I want to use a model that's not here?
If a model you want to use is not in our Models page, as long as it is based on one of the base models, you can contact us to request to add it.