Dolly v2 (7B): Open-source, instruction-following LLM for commercial use.
Overview:Dolly v2 (7B) is an instruction-following large language model trained on the Databricks machine learning platform, licensed for commercial use. It is based on the Pythia-6.9b model and fine-tuned on a dataset of approximately 15,000 instruction/response pairs.Key Features:
Intended Use:Dolly v2 (7B) is designed for various natural language processing tasks, including:
Language Support:The model primarily supports English language tasks.
Architecture:Dolly v2 (7B) is based on the Pythia-6.9b architecture, which is a transformer-based model.Training Data:
Performance Metrics:While Dolly v2 (7B) is not state-of-the-art, it demonstrates surprisingly high-quality instruction-following behavior. Some benchmark results include:
BenchmarkScoreARC (25-shot) 0.392 HellaSwag (10-shot) 0.633838 MMLU (5-shot) 0.406997 TruthfulQA (0-shot)0.444444
Comparison to Other Models:Dolly v2 (7B) underperforms compared to larger models like GPT-3 (175B parameters) but offers a good balance between performance and resource requirements.
Code Sample:
Dolly v2 (7B) was developed with ethical considerations in mind. The training data does not contain obscenity, intellectual property, or personally identifying information about non-public figures. However, it may reflect biases present in the data generated by Databricks employees.
Dolly v2 (7B) is released under the Apache 2.0 license, which allows for both research and commercial use.