Key Features
• High-Speed Performance
-1. Runs at 1,200 tokens per second via Cerebras inference, 10x faster than Gemini 2.0 Flash
-2. Near-instant answer generation for quick queries
-3. Optimized for low-latency search and reasoning tasks
• Enhanced Reasoning
-1. Built on Llama 3.3 70B with custom training for factuality and clarity
-2. Outperforms GPT-4o mini and Claude 3.5 Haiku in user satisfaction
-3. Matches or exceeds frontier models like GPT-4o on key benchmarks
• Search-Optimized Design
-1. Integrates real-time web search with citations for transparency
-2. 32,000 token context window for handling complex inputs
-3. Excels in resolving conflicting or missing information
• User Experience Focus
-1. Improved readability with concise, markdown-formatted responses
-2. Tops IFEval (instruction-following) and MMLU (world knowledge) benchmarks
-3. Customizable via Perplexity Pro settings or Sonar API
• Accessibility
-1. Available to Perplexity Pro users as a default model option
-2. Open to developers via the Sonar API for integration
• Versatile Applications
-1. Powers fast Q&A, research, and content generation
-2. Ideal for enterprises needing accurate, real-time insights