这是一个关于免费的大模型api的合集,并精选了一部分模型
This is a collection of free LLM apis, and selected some models
我会尽可能更新维护这个项目(目前只有我一个人)
I will keep maintaining and updating this project to the best of my ability
入选原则是:限制请求速率而不是token > 尽可能多的来源 > 尽可能新且好的模型 > 足够用的请求速率
The selection criteria are: limit request rate over token count > more sources > newer and better models > sufficient rate limits
主要是有一定热度的文本模型
Primarily text models that have gained some popularity
目前只接受提供了OpenAI格式的API
At present, only accepted OpenAI-formated API
欢迎大家分享更多api
Welcome to share more apis
这个表格是由Gemini 3 Flash生成的,由Taple渲染
This table was generated by Gemini 3 Flash, Rendered by Taple
- API: https://api.chatanywhere.tech
- Rate Limits: Not Limited
- Tip: -
- Models:
gpt-4o-mini
- API: https://api.siliconflow.cn/v1
- Rate Limits: 1000 RPM (each model)
- Tip: -
- Models:
deepseek-ai/DeepSeek-R1-0528-Qwen3-8BQwen/Qwen3-8BTHUDM/glm-4-9b-chatTHUDM/GLM-4-9B-0414THUDM/GLM-Z1-9B-0414THUDM/GLM-4.1V-9B-Thinking
- API: https://openrouter.ai/api/v1
- Rate Limits: 20 RPM / 200 RPD (each model)
- Tip: -
- Models:
deepseek/deepseek-r1-0528:freemoonshotai/kimi-k2:freez-ai/glm-4.5-air:freeqwen/qwen3-coder:freeopenai/gpt-oss-120b:freeopenai/gpt-oss-20b:freearcee-ai/trinity-mini:freenvidia/nemotron-3-nano-30b-a3b:freenvidia/nemotron-nano-12b-v2-vl:freenvidia/nemotron-nano-9b-v2:freearcee-ai/trinity-large-preview:freeliquid/lfm-2.5-1.2b-thinking:freeliquid/lfm-2.5-1.2b-instruct:freeallenai/molmo-2-8b:free
- API: https://chat.intern-ai.org.cn/api/v1
- Rate Limits: 10 RPM
- Tip: 密钥有效期6个月 / The key is valid for 6 months
- Models:
intern-latestintern-s1-miniinternvl3-78b
- API: https://api.suanli.cn/v1
- Rate Limits: Unknown
- Tip: 算力由他人设备共享提供 / Shared computing by other people's devices
- Models:
free:QwQ-32B
-
API: https://generativelanguage.googleapis.com/v1beta/openai
-
Rate Limits: 30 RPM / 1440 RPD
-
Tip: GFW
-
Models:
gemma-3-27bgemma-3-12bgemma-3-4b
-
Rate Limits: 5 RPM / 20 RPD
-
Models:
gemini-2.5-flashgemini-3-flash
-
Rate Limits: 10 RPM / 20 RPD
-
Models:
gemini-2.5-flash-lite
- API: https://api.cohere.ai/compatibility/v1
- Rate Limits: 20 RPM
- Tip:
- 绑定支付方式可以使用速率限制更宽松的 Production Key / Binding payment methods can use rate limiting and relaxed Production Key
- GFW
- Models:
command-a-03-2025command-a-vision-07-2025
- API: https://open.bigmodel.cn/api/paas/v4/
- Rate Limits: 只有并发数限制(均为30) / Only the number of concurrent transactions is limited (both 30).
- Tip: -
- Models:
GLM-4-Flash-250414GLM-Z1-FlashGLM-4.5-FlashGLM-4.1V-Thinking-FlashGLM-4.6V-Flash
- API: https://models.github.ai/inference
- Rate Limits: 15 RPM / 150 RPD
- Tip:
- 如果使用 Azure API,可以使用更多模型 / If used Azure API, more models available
- Models:
openai/gpt-4.1-miniopenai/gpt-4.1openai/gpt-4o
- API: https://api.gemai.cc/v1
- Rate Limits: Unknown
- Tip:
- 赠送¥100额度 / Gift ¥100 Credit
- Models:
gpt-5.1gpt-5.1-high[官逆]gemini-2.5-pro[满血A]gemini-3-pro-preview[满血A]gemini-3-pro-preview-thinking[满血B]gemini-2.5-flash[满血C]gemini-2.5-flash-thinking[满血C]gemini-2.5-pro-thinking[特价C]128k-claude-sonnet-4-5-20250929deepseek-v3.2-expdeepseek-v3.2-exp-thinkinggrok-4grok-4.1
- API: https://api520.pro/v1
- Rate Limits: Unknown
- Tip:
- 赠送¥100额度 / Gift ¥100 Credit
- Models:
熊猫-按量-特供顶级-官方正向满血-claude-haiku-4.5熊猫-按量-特供顶级-官方正向满血-claude-opus-4熊猫-按量-特供顶级-官方正向满血-claude-opus-4.1熊猫-按量-特供顶级-官方正向满血-claude-opus-4.1-thinking熊猫-按量-特供顶级-官方正向满血-claude-opus-4.5熊猫-按量-特供顶级-官方正向满血-claude-opus-4.5-thinking熊猫-按量-特供顶级-官方正向满血-claude-sonnet-4.5熊猫-按量-特供顶级-官方正向满血-claude-sonnet-4.5-thinking熊猫-A-4-deepseek-v3.2-满血熊猫-A-5-deepseek-v3.1-terminus-thinking-满血熊猫-A-5-deepseek-v3.2-thinking-满血熊猫-A-1-gemini-2.5-flash-lite-0925熊猫-A-2-gemini-2.5-flash-0925熊猫特供-按量-SS-gemini-2.5-pro熊猫特供-按量-SS-gemini-2.5-pro-thinking熊猫特供-按量-SSS-gemini-3-pro-preview熊猫-A-5-kimi-k2-thinking-满血熊猫-A-3-速度极快-gpt-oss-120b熊猫-A-5-Grok-4-fast-200w上下文已知最大熊猫-A-5-grok-4.1熊猫-A-4-豆包-doubao-seed-1.6熊猫-A-5-豆包-doubao-seed-1.6-thinking熊猫-A-5-智谱glm-4.6-满血熊猫-A-3-qwen3-max-满血熊猫-A-4-minimax-m2
-
自行收集 / Self-collected
-
投稿(B站等) / Contributed by others
- llm_benchmark:个人评测榜单,可信度高,而且收录更全 / A personal review list, it is highly credible, and it is more comprehensive
- Artifical Analysis
- LMArena
