Taibao-7B: An Intellectual Property Domain Large Language Model

Time: Dec. 2023

The current Taibao model is obtained by performing supervised, efficient fine-tuning using LoRA on the foundation model Qwen-7B for instruction tasks. The 7-billion-parameter Qwen-7B foundation model underwent multiple rounds of pretraining on high-quality datasets comprising over 2.5 trillion tokens across Chinese, English, code, mathematics, and specialized domains. This pretraining has endowed the model with a foundational capability for instruction understanding. Its natural language processing and comprehension abilities rank among the best in domestic open-source large models, significantly surpassing other models of similar scale while closely trailing a series of the most advanced closed-source models.

The model focuses on the field of intelligent intellectual property (IP) services, with research and development centered around key aspects of IP creation, management, protection, and utilization. It aims to provide users with comprehensive, efficient, and effective IP services.

We released the LoRA weigth of Taibao-7B, you need combine it with Qwen-7B

More details please see our github repository.

Cite

@misc{Taibao,
  author = {Qiyao Wang, Yuan Lin},
  title = {Taibao-7B: An Intellectual Property Domain Large Language Model},
  year = {2023},
  publisher = {GitHub},
  journal = {GitHub repository},
  howpublished = {\url{https://github.com/QiYao-Wang/Taibao}},
}
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference API
Unable to determine this model's library. Check the docs .

Collection including QiYao-Wang/Taibao-7B