Unleash your creativity
Check our our current supported models below. We support the best, highest-rated models available, and we update the selection of models we support frequently. Sign up to receive a monthly update about our latest models and other news, and participate in our Discord Server to ask questions or request additional models.
TOP LLMS
Sao10K/
L3-70B-Euryale-v2.2
Coherent, emotional and very creative.
- RP, Storywriting.
- FP8 Dynamic
- L3.1-70B-Euryale-v2.2-FP8-Dynamic
- Context: 16K
- RP Instruct: https://files.catbox.moe/1c9sp0.json
- RP Context: https://files.catbox.moe/5wwpin.json
Settings provided by: ShotMisser64
rAIfle/
SorcererLM-8x22b-bf16
- RP
- BF16
- rAIfle-SorcererLM-8x22b-bf16
- Context: 16K
- Recommended Settings: https://files.catbox.moe/9tj7m0.json
Anthracite-org/
Magnum-72b-v4
This model has spatial awareness, memory and detailed descriptions to keep the generation entertaining. Very good creativity and NSFW.
- RP, Storywriting.
- FP8 Dynamic
- anthracite-org-magnum-v4-72b-FP8-Dynamic
- Context: 32k
- Preset: https://files.catbox.moe/rqei05.json
- RP Instruct: https://files.catbox.moe/btnhau.json
- RP Context: https://files.catbox.moe/7kct3f.json
Settings provided by: GERGE
ALL MODELS
meta-llama/
Llama-3.2-11B-Vision-Instruct
- BF16
- GP
- Llama-3.2-11B-Vision-Instruct-Turbo
- Context: 128K
Mistralai/
Mixtral-8x7B-Instruct-v0.1
- BF16
- GP
- Mixtral-8x7B-Instruct-v0.1
- Context: 32K
Qwen/
Qwen2-72B-Instruct
- BF16
- GP
- Qwen2-72B-Instruct
- Context: 32K
TheDrummer/
Rocinante-12B-v1.1
- RP
- BF16
- TheDrummer-Rocinante-12B-v1.1
- Context: 32K
Sao10K/
L3.1 70B Hanami x1
- GP, RP
- FP16
- Sao10K-L3.1-70B-Hanami-x1
- Context: 32K
Envoid/
Llama3 TenyxChat DaybreakStorywriter 70B
- FP8 Dynamic
- RP
- Llama-3-TenyxChat-DaybreakStorywriter-70B-fp8-dynamic
- Context: 16K
- RP Instruct: https://files.catbox.moe/i3z4wv.json
- RP Context: https://files.catbox.moe/1k8p5b.json
Settings provided by: ShotMisser64
Alpindale/
WizardLM 2 8x22B
- RP
- Storywriting
- General purpose
- WizardLM-2-8x22B
- Context: 16K
- Context template: Alpaca
- RP Instruct: https://files.catbox.moe/q0a07u.json
- RP Context: https://files.catbox.moe/68194o.json
Settings provided by: GERGE
Sophosympatheia/
Midnight Miqu 70B v1.5
- RP
- FP16
- Midnight-Miqu-70B-v1.5
- Context: 18K
- Preset: https://files.catbox.moe/l8e5zt.json
- RP Instruct: https://files.catbox.moe/eaj6gy.json
- RP Context: https://files.catbox.moe/mvn3jo.json
Settings provided by: ShadingCrawler
nvidia/
Llama-3.1-Nemotron-70B-Instruct-HF
- GP
- BF16
- nvidia-Llama-3.1-Nemotron-70B-Instruct-HF
- Context: 32K
- Recommended Settings: https://files.catbox.moe/7e6zjo.json
Qwen/
Qwen2.5-72B-Instruct
- GP
- FP8
- Qwen2.5-72B-Instruct-Turbo
- Context: 32K
TheDrummer/
UnslopNemo-12B-v4.1
- RP
- BF16
- Context: 32K
- Recommended Settings: https://files.catbox.moe/7e6zjo.json
Infermatic/
MN-12B-Inferor-v0.0
All the qualities of the best models merged into one.
- RP
- More info: https://huggingface.co/Infermatic/MN-12B-Inferor-v0.0
- Context: 32K
- Settings & Review: https://infermatic.ai/infermatic-mn-12b-inferor-v0-0/
Qwen/
QwQ-32B-Preview
All the qualities of the best models merged into one.
- RP
- More info: https://huggingface.co/Qwen/QwQ-32B-Preview
- Context: 32K
NousResearch/
Hermes-3-Llama-3.1-70B
All the qualities of the best models merged into one.
- RP
- More info: https://huggingface.co/NousResearch/Hermes-3-Llama-3.1-70B-FP8
- Context: 64K
Anthracite-org/
magnum-v2-72b
This model is fine-tuned on top of Qwen-2 72B Instruct.
- More info: https://huggingface.co/anthracite-org/magnum-v2-72b
- FP8: https://huggingface.co/Infermatic/magnum-v2-72b-FP8-Dynamic
- Context: 32K
meta-llama/
Llama-3.3-70B-Instruct
The Meta Llama 3.3 multilingual large language model (LLM) is a pretrained and instruction tuned generative model in 70B (text in/text out).
Sao10K/
L3.3-70B-Euryale-v2.3
A direct replacement / successor to Euryale v2.2
Sao10K/
72B-Qwen2.5-Kunou-v1
Another version of Euryale with with a Qwen base model.
New!
inflatebot/
MN-12B-Mag-Mell-R1
This is a merge of pre-trained language Mistral Nemo models.
- More info: https://huggingface.co/inflatebot/MN-12B-Mag-Mell-R1
- Context: 32K
New!
EVA-UNIT-01/
EVA-Qwen2.5-72B-v0.1
A RP/storywriting specialist model, full-parameter finetune of Qwen2.5-72B on mixture of synthetic and natural data.
- Context: 32K
- Settings: [todo: - need to support nested list]
- Temperature: 1
- Min-P: 0.05
- Top-A: 0.2
- Repetition Penalty: 1.03
Guides & SETTINGS
Models
L3-70B-Euryale-v2.1
Meet L3 70B Euryale v2.1: Your New Creative Companion What is L3 70B Euryale v2.1 [...]
20
Aug
Aug
Guides
Using Infermatic.ai API with SillyTavern
SillyTavern is one of the most popular interfaces to interact with LLMs. We have been [...]
21
Jun
Jun
Docs
API Docs
Frequently Asked Questions from Geek to Geek
- What is prompt engineering, and why is it critical in working with LLMs?
- How can I design effective prompts for LLMs?
- What are some standard techniques used in prompt engineering?
- How does prompt length impact the output of an LLM?
- How do LLMs understand and generate human-like text?
- What is the difference between Llama, Mixtral, and Qwen?
- What are some examples of advanced use cases of prompt engineering with LLMs?
- How do I choose the best LLM model for my project?
- What are large language models, and how do they differ from traditional NLP models?
- Can LLMs write code well?