Instruction-following model by Databricks, fine-tuned for diverse language tasks.
Overview:
Dolly v2 (3B) is an instruction-following large language model created by Databricks, designed to follow instructions and perform various language tasks. Based on the Pythia-2.8b model, Dolly v2 (3B) has been fine-tuned on a dataset of approximately 15k instruction/response pairs to enhance its ability to generate high-quality responses to prompts.
Dolly v2 (3B) is designed for various natural language processing tasks including brainstorming, classification, closed and open question answering, generation, information extraction, and summarization. It is suitable for applications requiring high-quality instruction following, though it is not state-of-the-art.
Supports English. Other languages might be supported but with potentially less accuracy due to the training data being primarily in English.
Dolly v2 (3B) is based on the Pythia-2.8b model, a Transformer-based architecture.
The model was trained on a dataset of approximately 15,000 instruction/response pairs generated by Databricks employees. This dataset, named databricks-dolly-15k, covers various domains mentioned in the InstructGPT paper, including brainstorming, classification, QA, and summarization.
Databricks is committed to developing AI technologies that are helpful, honest, and harmless. The model has limitations and may produce biased or harmful outputs, reflecting the biases present in the training data.
Dolly v2 (3B) is released under a permissive license (CC-BY-SA), allowing for commercial use.