

Vicuna FastChat T5 (3B) open-source chatbot model by LM-SYS, fine-tuned for diverse conversations.
Model Name: FastChat-T5 (3B)
Developer/Creator: LM-SYS (primarily Dacheng Li, Lianmin Zheng, and Hao Zhang)
Release Date: April 2023
Version: Current Version
Model Type: Text (Chatbot)
FastChat-T5 is designed for commercial chatbot applications and research in natural language processing. It can be used for generating responses in conversational agents and other NLP tasks.
Supports English. Other languages may be supported but with reduced accuracy due to the training data being primarily in English.
FastChat-T5 is based on an encoder-decoder transformer architecture. The encoder processes the input text bidirectionally, creating hidden representations. The decoder then uses cross-attention to focus on these representations while generating the response autoregressively from a starting token.
The training data reflects the conversations shared by users on ShareGPT, which may introduce certain biases. The diversity is limited to the topics and styles of interactions present in these conversations.
FastChat-T5 may inherit biases from the ShareGPT dataset. Users should be cautious of potential ethical issues, including biased or harmful outputs, and use the model responsibly.
FastChat-T5 is licensed under the Apache License 2.0, which allows for commercial and non-commercial use.
Model Name: FastChat-T5 (3B)
Developer/Creator: LM-SYS (primarily Dacheng Li, Lianmin Zheng, and Hao Zhang)
Release Date: April 2023
Version: Current Version
Model Type: Text (Chatbot)
FastChat-T5 is designed for commercial chatbot applications and research in natural language processing. It can be used for generating responses in conversational agents and other NLP tasks.
Supports English. Other languages may be supported but with reduced accuracy due to the training data being primarily in English.
FastChat-T5 is based on an encoder-decoder transformer architecture. The encoder processes the input text bidirectionally, creating hidden representations. The decoder then uses cross-attention to focus on these representations while generating the response autoregressively from a starting token.
The training data reflects the conversations shared by users on ShareGPT, which may introduce certain biases. The diversity is limited to the topics and styles of interactions present in these conversations.
FastChat-T5 may inherit biases from the ShareGPT dataset. Users should be cautious of potential ethical issues, including biased or harmful outputs, and use the model responsibly.
FastChat-T5 is licensed under the Apache License 2.0, which allows for commercial and non-commercial use.