Alpha
amallo chat Icon
model

Qwen 3 Next 80B Thinking

Qwen 3 Next 80B Thinking is a specialized reasoning large language model developed by Alibaba Cloud, featuring a high-sparsity Mixture of Experts architecture and an internal chain-of-thought mechanism. It is designed for complex tasks in mathematics, logic, and software engineering, operating with 80 billion total parameters but only 3 billion active per token.

3h ago2
model

Mistral Small 3.2 24B Instruct

Mistral Small 3.2 24B Instruct is an open-weight multimodal large language model featuring 23.6 billion parameters and a 128,000-token context window. Released by Mistral AI in June 2025, it is optimized for high-speed instruction following, coding proficiency, and visual document analysis.

3h ago2
model

Qwen 3 14B

Qwen 3 14B is a dense 14.8 billion parameter large language model developed by Alibaba Cloud, featuring a hybrid reasoning engine that allows toggling between thinking and non-thinking modes. Released in April 2025 under an Apache 2.0 license, it achieves performance parity with much larger previous-generation models in STEM, coding, and logical reasoning tasks.

3h ago2
model

Mistral Small 3.1 24B Instruct

Mistral Small 3.1 24B Instruct is a 24-billion parameter multimodal model by Mistral AI that balances computational efficiency with advanced reasoning and vision capabilities. It features a 128,000-token context window and is optimized for local deployment on consumer-grade hardware and enterprise-scale agentic workflows.

3h ago2
model

Qwen 3 235B

Qwen 3 235B is a flagship large language model developed by Alibaba Cloud, featuring a Sparse Mixture-of-Experts (MoE) architecture with 235 billion total parameters. It is designed for high-capacity reasoning, multimodal tasks, and extensive context processing, serving as a key open-weights model in the Qwen 3 series.

3h ago2
model

Llama 4 Maverick

Llama 4 Maverick is a specialized iteration of Meta's Llama 4 large language model series, optimized for autonomous agentic workflows, multi-step reasoning, and precise tool-use. It utilizes a sparse Mixture-of-Experts (MoE) architecture to balance high-level intelligence with inference efficiency.

3h ago2
model

Claude Opus 4.0

Claude Opus 4.0 is Anthropic's flagship large language model featuring a hybrid reasoning architecture and an extended thinking mode for complex analytical tasks. It is optimized for autonomous software engineering and long-horizon research, and is the first model to implement AI Safety Level 3 protocols.

3h ago2
model

Claude Sonnet 4.5

Claude Sonnet 4.5 is a mid-tier large language model in the Claude 4.5 family developed by Anthropic, optimized for complex software engineering and autonomous agentic workflows. It features a 200,000-token context window and advanced computer control capabilities.

3h ago2
model

Claude Sonnet 4.0

Claude Sonnet 4.0 is a frontier large language model developed by Anthropic that balances high-level intelligence with operational speed. It is optimized for software engineering and autonomous agentic workflows through a unique hybrid reasoning architecture.

3h ago2
model

Claude Sonnet 3.7

Claude 3.7 Sonnet is a multimodal large language model developed by Anthropic, released in February 2025. It is the first "hybrid reasoning model" featuring an "extended thinking" mode designed for complex software engineering and mathematical tasks.

3h ago1
model

Claude Sonnet 3.5

Claude 3.5 Sonnet is a mid-tier large language model developed by Anthropic that balances high-level reasoning with operational speed. Released in June 2024 and upgraded in October 2024, it introduced advanced capabilities in coding, visual processing, and an experimental 'computer use' feature for interacting with standard desktop interfaces.

3h ago2
model

Claude Sonnet 4.6

Claude Sonnet 4.6 is a mid-tier large language model in Anthropic's Claude 4.6 family, designed to balance high-speed processing with advanced reasoning and agentic capabilities. Released in February 2026, it features a 1-million-token context window and native multimodality, supporting complex workflows in software development and data analysis.

3h ago3
model

Claude Opus 4.6

Claude Opus 4.6 is Anthropic's frontier large language model designed for high-complexity cognitive tasks, featuring a sparse Mixture-of-Experts architecture and a 1-million-token context window. It is optimized for enterprise-level reasoning, autonomous agentic workflows, and sophisticated multimodal analysis of text, images, and video.

3h ago2
model

Grok 3 Fast

Grok 3 Fast is a high-speed, low-latency large language model developed by xAI, optimized for rapid inference and real-time data synthesis using a sparse Mixture-of-Experts architecture. It is designed to balance intelligence with operational efficiency, featuring deep integration with the X platform for up-to-the-minute information processing.

3h ago2
model

Claude Opus 3

Claude 3 Opus is the flagship large language model of Anthropic's Claude 3 family, designed for high-level reasoning, complex analysis, and multimodal input processing. Released in March 2024, it was the first model to surpass GPT-4 on several key industry benchmarks before being succeeded by the Claude 3.5 and 4 series.

3h ago2
model

QwQ 32B

QwQ 32B is a 32-billion-parameter reasoning large language model developed by Alibaba Cloud’s Qwen team, utilizing test-time compute and reinforcement learning to excel in complex mathematical and programming tasks.

3h ago3
model

GPT-4o mini

GPT-4o mini is a small-scale multimodal large language model developed by OpenAI, released in July 2024 as a highly efficient and cost-effective successor to GPT-3.5 Turbo. It features a 128K token context window and is optimized for low-latency tasks such as customer support, real-time responses, and high-volume data processing.

3h ago2
model

Gemini 3.1 Pro

Gemini 3.1 Pro is a multimodal large language model developed by Google DeepMind, designed for advanced reasoning, scientific knowledge, and autonomous agentic tasks. It features a unique "dynamic thinking" mechanism and a 1-million-token input context window with significantly expanded output limits.

3h ago2