Available Models

Chat AI provides a large assortment of state-of-the-art open-weight Large Language Models (LLMs) which are hosted on our platform with the highest standards of data protection. The data sent to these models, including the prompts and message contents, are never stored at any location on our systems. Additionally, Chat AI offers models hosted externally such as Anthropic Claude and OpenAI GPT-5.

Available models are regularly upgraded as newer, more capable ones are released. We select models to include in our services based on user demand, cost, and performance across various benchmarks, such as HumanEval, MATH, HellaSwag, MMLU, etc. Certain models are more capable at specific tasks and with specific settings, which are described below to the best of our knowledge.


List of open-weight models, hosted by GWDG

OrganizationModelOpenRelease dateContext window in tokensAdvantagesLimitationsRecommended settings
πŸ‡¨πŸ‡­ Swiss AIApertus 70B Instruct 2509yesSep 202565kFully open-source, Multilingual-temp=0.8
top_p=0.9
πŸ‡¨πŸ‡³ DeepSeekDeepSeek R1 Distill Llama 70ByesJan 202532kGood overall performance,
faster than R1
Censorship, political biasdefault
temp=0.7, top_p=0.8
πŸ‡«πŸ‡· MistralDevstral 2 123B Instruct 2512yesDec 2025256KCoding, agentic tasks-default
πŸ‡ΊπŸ‡Έ GoogleGemma 3 27B InstructyesMar 2025128kVision, great overall performance-default
πŸ‡ΈπŸ‡¬ Z.aiGLM-4.7yesDec 2025200kGreat performance-temp=1.0
top_p=0.95
πŸ‡¨πŸ‡³ OpenGVLabInternVL 3.5 30B A3ByesAug 202540kVision, lightweight and fast-default
πŸ‡ΊπŸ‡Έ GoogleMedGemma 27B InstructyesJul 2025128kVision, medical knowledge-default
πŸ‡ΊπŸ‡Έ MetaLlama 3.1 8B InstructyesJul 2024128kFast overall performance-default
πŸ‡ΊπŸ‡Έ MetaLlama 3.3 70B InstructyesJul 2024128kGood overall performance,
reasoning and creative writing
-default
temp=0.7, top_p=0.8
πŸ‡«πŸ‡· MistralMistral Large Instruct 3 675B Instruct 2512yesDec 2025256KGreat overall performance, vision-temp < 0.1
higher temp for creative use cases
πŸ‡ΊπŸ‡Έ OpenAIGPT OSS 120ByesAug 2025128kGreat overall performance, fast-default
πŸ‡¨πŸ‡³ Alibaba CloudQwen 3 235B A22B Thinking 2507yesJul 2025222k-outdatedtemp=0.6, top_p=0.95
πŸ‡¨πŸ‡³ Alibaba CloudQwen 3 30B A3B Instruct 2507yesJul 2025256kGood performance, fast-temp=0.6, top_p=0.95
πŸ‡¨πŸ‡³ Alibaba CloudQwen 3 30B A3B Thinking 2507yesJul 2025256k-outdatedtemp=0.6, top_p=0.95
πŸ‡¨πŸ‡³ Alibaba CloudQwen 3 32ByesApr 202532k-outdateddefault
πŸ‡¨πŸ‡³ Alibaba CloudQwen 3.5 122B A10ByesFeb 2026256KVision, great overall performance-temp=0.6, top_p=0.95
πŸ‡¨πŸ‡³ Alibaba CloudQwen 3.5 27ByesFeb 2026256KVision, good overall performance-temp=0.6, top_p=0.95
πŸ‡¨πŸ‡³ Alibaba CloudQwen 3.5 35B A3ByesFeb 2026256KVision, good overall performance-temp=0.6, top_p=0.95
πŸ‡¨πŸ‡³ Alibaba CloudQwen 3.5 397B A17ByesFeb 2026256KVision, great overall performance-temp=0.6, top_p=0.95
πŸ‡¨πŸ‡³ Alibaba CloudQwen 3 Coder 30B A3B InstructyesJul 2025256kCoding-temp=0.7, top_p=0.8
πŸ‡¨πŸ‡³ Alibaba CloudQwen 3 Omni 30B A3B InstructyesSep 2025256kMultimodal-default
πŸ‡¨πŸ‡³ Alibaba CloudQwen 3 VL 30B A3B InstructyesOct 2025262k-outdateddefault
πŸ‡©πŸ‡ͺ OpenGPT-XTeuken 7B Instruct ResearchyesNov 2024128kEuropean languages-default
πŸ‡ΊπŸ‡Έ intfloat x MistralE5 Mistral 7B InstructyesJan 20244096EmbeddingsAPI Only-

List of external models, hosted by external providers

OrganizationModelOpenRelease dateContext window in tokensAdvantagesLimitationsRecommended settings
πŸ‡ΊπŸ‡Έ AnthropicClaude Opus 4.6noFeb 20261MState-of-the-art reasoning, coding, complex analysis-default
πŸ‡ΊπŸ‡Έ AnthropicClaude Sonnet 4.6noFeb 20261MBalanced performance, fast responses-default
πŸ‡ΊπŸ‡Έ OpenAIGPT-5.4noMar 2026272KProfessional knowledge work, coding, data analysis, agentic workflows-default
πŸ‡ΊπŸ‡Έ OpenAIGPT-5.4 MininoMar 2026272KFast overall performance-default
πŸ‡ΊπŸ‡Έ OpenAIGPT-5.4 NanonoMar 2026272KFastest overall performance-default
πŸ‡ΊπŸ‡Έ OpenAIGPT-5.3 ChatnoMar 2026128KMultimodal chat, adaptive chain-of-thought, safety guardrails-default
πŸ‡ΊπŸ‡Έ OpenAIGPT-5.2 ChatnoDec 2025400kGreat overall performance-default
πŸ‡ΊπŸ‡Έ OpenAIGPT-5.2noDec 2025400kGreat overall performance-default
πŸ‡ΊπŸ‡Έ OpenAIGPT-5.1 ChatnoNov 2025400kGreat overall performance-default
πŸ‡ΊπŸ‡Έ OpenAIGPT-5.1noNov 2025400kGreat overall performance-default
πŸ‡ΊπŸ‡Έ OpenAIGPT-5 ChatnoAug 2025400kGood overall performance-default
πŸ‡ΊπŸ‡Έ OpenAIGPT-5noAug 2025400kGood overall performance, reasoning-default
πŸ‡ΊπŸ‡Έ OpenAIGPT-5 MininoAug 2025400kFast overall performance-default
πŸ‡ΊπŸ‡Έ OpenAIGPT-5 NanonoAug 2025400kFastest overall performance-default
πŸ‡ΊπŸ‡Έ OpenAIo3noApr 2025200k-outdateddefault
πŸ‡ΊπŸ‡Έ OpenAIo3-mininoJan 2025200k-outdateddefault
πŸ‡ΊπŸ‡Έ OpenAIGPT-4.1noApr 20251M-outdateddefault
πŸ‡ΊπŸ‡Έ OpenAIGPT-4.1 MininoJun 20241M-outdateddefault

Open-weight models, hosted by GWDG

The models listed in this section are hosted on our platform with the highest standards of data protection. The data sent to these models, including the prompts and message contents, are never stored at any location on our systems.

Apertus 70B Instruct

Apertus is a fully open language model designed to push the boundaries of transparent and compliant AI. It supports over 1,800 languages and a context window size of up to 65,536 tokens, using only fully compliant and open training data. The model achieves comparable performance to closed-source models while respecting opt-out consent of data owners. It was pretrained on 15T tokens with a staged curriculum of web, code, and math data.

DeepSeek R1 Distill Llama 70B

Developed by the Chinese company DeepSeek (深度求紒), DeepSeek R1 Distill Llama 70B is a dense model distilled from DeepSeek-R1 but based on LLama 3.3 70B, in order to fit the capabilities and performance of R1 into a 70B parameter-size model.

Warning

DeepSeek models have been reported to produce politically biased responses, and censor certain topics that are sensitive for the Chinese government.

Devstral 2 123B Instruct 2512

Developed by mistralai, Devstral 2 is an agentic LLM designed for software engineering and coding tasks. It is capable of exploring codebases, working with multiple files, and powering software engineering agents.

Google Gemma 3 27B Instruct

Gemma is Google’s family of light, open-weights models developed with the same research used in the development of its commercial Gemini model series. Gemma 3 27B Instruct is quite fast and thanks to its support for vision (image input), it is a great choice for all sorts of conversations.

GLM-4.7

GLM-4.7 is a coding-focused model that delivers significant improvements over its predecessor in multilingual agentic coding and terminal-based tasks. It achieves strong performance on SWE-bench, SWE-bench Multilingual, and Terminal Bench 2.0. GLM-4.7 also excels at tool use, web browsing, and mathematical reasoning, with notable gains on benchmarks like HLE and τ²-Bench.

InternVL 3.5 30B-A3B

InternVL 3.5 30B-A3B is a lightweight, fast and powerful multimodal model developed by OpenGVLab. It significantly advances versatility, reasoning capability, and efficiency, by featuring a Visual Resolution Router (ViR) for dynamic visual token adjustment and Decoupled Vision-Language Deployment (DvD) for efficient GPU load balancing, achieving up to 4Γ— inference speedup compared to its predecessor. The model excels at multimodal reasoning, OCR, document understanding, multi-image comprehension, video understanding, GUI tasks, and embodied agency.

Google MedGemma 27B Instruct

MedGemma 27B Instruct is a variant of Gemma 3 suitable for medical text and image comprehension. It has been trained on a variety of medical image data, including chest X-rays, dermatology images, ophthalmology images, and histopathology slides, as well as medical text, such as medical question-answer pairs, and FHIR-based electronic health record data. MedGemma variants have been evaluated on a range of clinically relevant benchmarks to illustrate their baseline performance.

Meta Llama 3.1 8B Instruct

The standard model we recommend. It is the most lightweight with the fastest performance and good results across all benchmarks. It is sufficient for general conversations and assistance.

Meta Llama 3.3 70B Instruct

Achieves good overall performance, on par with GPT-4, but with a much larger context window and more recent knowledge cutoff. Best in English comprehension and further linguistic reasoning, such as translations, understanding dialects, slang, colloquialism and creative writing.

Mistral Large 3 Instruct 675B Instruct 2512

Developed by Mistral AI, Mistral Large 3 is a general-purpose multimodal MoE model with 675B total and 41B active parameters. This model is fine-tuned for instruction tasks, ideal for chat, agentic and instruction based use cases. It supports dozens of languages, including English, French, Spanish, German, Italian, Portuguese, Dutch, Chinese, Japanese, Korean, Arabic, and has a large 256K context window size.

OpenAI GPT OSS 120B

In August 2025, OpenAI released the gpt-oss model series, consisting of two open-weight LLMs that are optimized for faster inference with state-of-the-art performance across many domains, including reasoning and tool use. According to OpenAI, the gpt-oss-120b model achieves near-parity with OpenAI o4-mini on core reasoning benchmarks.

Qwen 3 235B A22B Thinking 2507

Expanding on Qwen 3 235B A22B, one of the best-performing models of the Qwen 3 series, Qwen 3 235B A22B Thinking 2507 has a significantly improved performance on reasoning tasks, including logical reasoning, mathematics, science, coding, and academic benchmarks. It is an MoE model with 235B total parameters and 22B activated parameters, and achieves state-of-the-art results among open-weights thinking models. This model is outdated and not recommended anymore.

Qwen 3 30B A3B Instruct 2507

This MoE model features 30.5B total parameters with 3.3B activated parameters for efficient inference. It delivers significant improvements in instruction following, logical reasoning, text comprehension, mathematics, science, coding, and tool usage, with better alignment for subjective and open-ended tasks. The model supports a 256K native context length and operates in non-thinking mode, achieving strong performance across knowledge, reasoning, coding, and multilingual benchmarks.

Qwen 3 30B A3B Thinking 2507

The Thinking variant of Qwen 3 30B A3B is optimized for complex reasoning tasks. It excels at mathematical problem-solving (AIME25, HMMT25), logical reasoning (ZebraLogic), and coding challenges, while maintaining strong performance on knowledge benchmarks like MMLU-Pro and GPQA. This model also demonstrates strong alignment capabilities, scoring high on IFEval, Arena-Hard v2, and creative writing benchmarks. This model is outdated and not recommended anymore.

Qwen 3 32B

Qwen 3 32B is a large dense model developed by Alibaba Cloud released in April 2025. It supports reasoning and outperforms or is at least on par with other strong reasoning models such as OpenAI o1 and DeepSeek R1. This model is outdated and not recommended anymore.

Qwen 3 Coder 30B A3B Instruct

Qwen 3 Coder 30B A3B Instruct is a specialized coding model that achieves strong performance on agentic coding, browser-use, and other foundational coding tasks among open models.

Qwen 3 Omni 30B A3B Instruct

Qwen3 Omni is a natively multilingual omni-modal foundation model that processes text, images, audio, and video. It achieves state-of-the-art performance on many audio/video benchmarks with ASR, audio understanding, and voice conversation performance comparable to Gemini 2.5 Pro. The model features a novel MoE-based Thinker–Talker architecture with AuT pretraining, supports 119 text languages, 19 speech input languages, and enables low-latency interaction with flexible control via system prompts.

Qwen 3 VL 30B A3B Instruct

Qwen3 VL is a powerful vision-language model in the Qwen series, featuring comprehensive upgrades across visual perception, reasoning, and agent capabilities. This MoE model (30B total, 3B active) excels as a visual agent that can operate PC/mobile GUIs, generate code from images/videos (Draw.io/HTML/CSS/JS), and perform advanced spatial reasoning with 2D and 3D grounding. It supports native 256K context for long-form video understanding, recognizes a wide range of entities including celebrities, anime, products, and landmarks, and offers robust OCR across 32 languages. Text understanding is on par with pure LLMs, enabling seamless text-vision fusion. This model is outdated and not recommended anymore.

Qwen 3.5 122B A10B

Qwen 3.5 122B A10B is a powerful language model developed by Alibaba Cloud. With 122 billion parameters it delivers strong performance across reasoning, coding, and general tasks. The model supports vision capabilities for multimodal applications.

Qwen 3.5 27B

Qwen 3.5 27B is a large language model developed by Alibaba Cloud. With 27 billion parameters it provides good overall performance across various tasks while being more memory-efficient. The model supports vision capabilities for understanding and processing images alongside text.

Qwen 3.5 35B A3B

Qwen 3.5 35B A3B is an MoE model with 35 billion total parameters and 3 billion activated parameters for efficient inference. It delivers good overall performance across reasoning, coding, and general tasks. The model supports vision capabilities for multimodal applications and operates efficiently.

Qwen 3.5 397B A17B

Qwen 3.5 397B A17B is a MoE model with 397 billion total parameters and 17 billion activated parameters. It represents one of the most powerful open-weight models available, delivering exceptional performance across reasoning, coding, mathematics, and general tasks. The model supports vision capabilities, and provides state-of-the-art performance among open models.

OpenGPT-X Teuken 7B Instruct Research

OpenGPT-X is a research project funded by the German Federal Ministry of Economics and Climate Protection (BMWK) and led by Fraunhofer, Forschungszentrum JΓΌlich, TU Dresden, and DFKI. Teuken 7B Instruct Research v0.4 is an instruction-tuned 7B parameter multilingual LLM pre-trained with 4T tokens, focusing on covering all 24 EU languages and reflecting European values.


External models, hosted by external providers

Warning

These OpenAI and Anthropic models are hosted on external providers, and Chat AI only relays the contents of your messages to their servers. We therefore recommend the open-weight models, hosted by us, to ensure the highest security and data privacy.

Anthropic Claude Opus 4.6

Claude Opus 4.6 is Anthropic’s currently most capable model, designed for complex reasoning, advanced coding, and nuanced analysis. It excels at multi-step problem solving, scientific reasoning, and creative tasks. Opus 4.6 offers state-of-the-art performance across a wide range of benchmarks.

Anthropic Claude Sonnet 4.6

Claude Sonnet 4.6 provides an excellent balance between capability and speed. It is optimized for production workloads and delivers fast, reliable responses for coding, analysis, and conversational tasks.

OpenAI GPT-5, 5.1, 5.2, 5.3, and 5.4 Series

OpenAI’s GPT-5 series models achieve state-of-the-art performance across various benchmarks. The series consists of the following models along with their intended use cases:

  • OpenAI GPT-5.4: Built for professional knowledge work, including document and spreadsheet tasks, coding, data analysis, agentic workflows, and software automation.
  • OpenAI GPT-5.4 Mini: A lightweight variant of GPT-5.4 for cost-sensitive applications.
  • OpenAI GPT-5.4 Nano: A highly optimized variant of GPT-5.4. Ideal for applications requiring low latency.
  • OpenAI GPT-5.3 Chat: Fast, context-aware chat experiences with adaptive chain-of-thought reasoning, tuned safety and instruction-following, multimodal chat, and agent development support. Ideal for interactive assistants, customer care, IT helpdesk, HR, and sales enablement.
  • OpenAI GPT-5/5.1/5.2 Chat: Designed for advanced, natural, multimodal, and context-aware conversations.
  • OpenAI GPT-5/5.1/5.2: Designed for logic-heavy and multi-step tasks.
  • OpenAI GPT-5 Mini: A lightweight variant of GPT-5 for cost-sensitive applications.
  • OpenAI GPT-5 Nano: A highly optimized variant of GPT-5.

OpenAI GPT-4.1

OpenAI’s GPT-4.1-class models improve on the older GPT-4 series. These models also outperform GPT-4o and GPT-4o Mini, especially in coding and instruction following. They have a large context window size of 1M tokens, with improved long-context comprehension, and an updated knowledge cutoff of June 2024.

OpenAI GPT-4.1 Mini

This was developed as a more cost-effective and faster alternative to GPT-4.1.

OpenAI o1 and o1 Mini

OpenAI’s o1-class models were developed to perform complex reasoning tasks. These models have now been superceded by the o3-series, and are therefore no longer recommended.

OpenAI o3

Released in April 2025, OpenAI’s o3-class models were developed to perform complex reasoning tasks across the domains of coding, math, science, visual perception, and more. These models have an iterative thought process, and therefore take their time to process internally before responding to the user. The thought process for o3 models are not shown to the user.

OpenAI o3 Mini

This was developed as a more cost-effective and faster alternative to o3.