Smaug-72B-v0.1 - AI Language Models Tool
Overview
Smaug-72B-v0.1 is an open-source large language model for text generation developed by Abacus.AI. It is based on Qwen-72B and finetuned with the DPO-Positive (DPOP) technique, achieving strong results on benchmarks such as MT-Bench and surpassing an average score of 80% on the Open LLM Leaderboard.
Key Features
- Open-source large language model optimized for text generation
- 72B-parameter model family (Smaug-72B)
- Built on the Qwen-72B foundation model
- Finetuned using the DPO-Positive (DPOP) technique
- High benchmark performance, including MT-Bench
- First open model to exceed 80% average on Open LLM Leaderboard
Ideal Use Cases
- Research and academic evaluation of LLM capabilities
- Benchmarking against other open models
- Generating high-quality natural language text
- Building open-source chat assistants and agents
- Fine-tuning for downstream NLP tasks and experiments
Getting Started
- Visit the Hugging Face model page for Smaug-72B-v0.1
- Review the model card, README, and license details
- Download checkpoints or use Hugging Face-hosted endpoints
- Follow repository instructions to load model and tokenizer
- Run included example inference scripts or minimal pipeline
- Evaluate outputs on your validation dataset before production
Pricing
No pricing information provided. Model is open-source; hosting and inference costs depend on your infrastructure.
Key Information
- Category: Language Models
- Type: AI Language Models Tool