Alpha
Wiki Icon
Grok Code Fast 1

Grok Code Fast 1

1m

Grok Code Fast 1 is a specialized 314B parameter Mixture-of-Experts model by xAI designed for high-speed software engineering and agentic coding workflows.

Head Latent Attention

Head Latent Attention

1m

Multi-Head Latent Attention (MLA) is a low-rank attention mechanism that reduces the memory footprint of the Key-Value (KV) cache in large language models by compressing information into a latent space.

Gemini 3.1 Pro

Gemini 3.1 Pro

2m

Gemini 3.1 Pro is a multimodal large language model developed by Google DeepMind, designed for advanced reasoning, scientific knowledge, and autonomous agentic tasks. It features a unique "dynamic thinking" mechanism and a 1-million-token input context window with significantly expanded output limits.

Sonar Reasoning

Sonar Reasoning

3m

Sonar Reasoning is a specialized large language model developed by Perplexity AI that integrates deep chain-of-thought reasoning with real-time web search for analytical and fact-grounded responses. Built on the Llama 3.1 70B architecture, the model utilizes inference-time compute scaling to decompose complex queries and verify information through an iterative processing loop.

GPT-4o

GPT-4o

3m

GPT-4o is a multimodal large language model developed by OpenAI that natively processes and generates text, audio, and visual data within a single integrated neural network. It features significantly reduced latency compared to previous iterations, enabling real-time human-computer interactions such as live translation and interactive tutoring.

Grok 4.20

Grok 4.20

4m

Grok 4.20 is a large language model developed by xAI, featuring a unique "4 Agents" multi-agent collaboration system and a 2-million-token context window. Released in March 2026, it is designed for high-reasoning tasks and real-time information synthesis using data from the X platform.

Gemini 3 Flash

Gemini 3 Flash

4m

Gemini 3 Flash is a high-speed, efficiency-oriented multimodal large language model developed by Google DeepMind, featuring a 1-million-token context window and optimized for low-latency enterprise applications.

QwQ 32B

QwQ 32B

6m

QwQ 32B is a 32-billion-parameter reasoning large language model developed by Alibaba Cloud’s Qwen team, utilizing test-time compute and reinforcement learning to excel in complex mathematical and programming tasks.

V3.2

V3.2

6m

DeepSeek V3.2 is an open-weight large language model that unifies general-purpose instruction following and specialized reasoning within a Mixture-of-Experts architecture. It introduces DeepSeek Sparse Attention (DSA) to maintain linear computational efficiency across its 128,000-token context window.

Meta

Meta

7m

Meta Platforms, Inc. is a global technology conglomerate and leader in social media and artificial intelligence, known for its Family of Apps and a strategic commitment to the metaverse and open-source AI development.

Black Forest Labs

Black Forest Labs

7m

Black Forest Labs is a German artificial intelligence research organization specializing in generative image and video models, founded in 2024 by the creators of Stable Diffusion.

Kimi K2.5

Kimi K2.5

9m

Kimi K2.5 is a multimodal large language model developed by Moonshot AI that utilizes a Mixture-of-Experts architecture and a context window of up to 2 million tokens. It is designed for advanced logical reasoning, coding, and complex agentic workflows, positioning it as a significant competitor in the Chinese AI market.

Gemini 2.0 Flash

Gemini 2.0 Flash

10m

Gemini 2.0 Flash is a natively multimodal large language model developed by Google DeepMind, optimized for low-latency 'agentic' workflows and high-speed processing. It features a 1-million-token context window and supports real-time streaming for interactive applications through its Multimodal Live API.

Model Autonomy

Model Autonomy

12m

Model autonomy refers to the ability of an artificial intelligence system to independently set goals, make decisions, and execute actions through continuous sensing, reasoning, and tool interaction. It represents a progression from static rule-based automation to dynamic agentic workflows that require minimal human intervention.

Qwen 3 Next 80B Instruct

Qwen 3 Next 80B Instruct

13m

Qwen 3 Next 80B Instruct is a high-efficiency large language model featuring an ultra-sparse Mixture-of-Experts (MoE) architecture and hybrid attention. It activates only 3 billion of its 80 billion parameters per inference step, enabling significant throughput advantages for long-context tasks up to 256,000 tokens.

Qwen 3 32B

Qwen 3 32B

15m

Qwen 3 32B is a mid-sized, open-weights large language model developed by Alibaba Cloud, designed to balance high reasoning capabilities in coding and mathematics with hardware efficiency. It features a dense Transformer architecture with 32.5 billion parameters and a 128,000-token context window.

Moonshot AI

Moonshot AI

16m

Moonshot AI is a prominent Beijing-based artificial intelligence startup specializing in large language models (LLMs) and multimodal systems, known for its flagship Kimi chatbot and long-context window technology. Established in 2023, it is recognized as one of China's 'new four AI tigers' and has achieved significant market valuation through rapid technological scaling and strategic investment.

Gemini 2.5 Pro

Gemini 2.5 Pro

16m

Gemini 2.5 Pro is a natively multimodal reasoning model developed by Google DeepMind that utilizes a Mixture-of-Experts architecture and supports a massive context window of up to 2 million tokens.