rwkv-7b-jp-lora / README.md
OpenMOSE's picture
Update README.md
7c063ea

license: apache-2.0

a LoRA fine-tuned model with Japanese dataset

LoRA Experiment

rwkv-7b-jp-280.pth is merged model with base

Base Model

https://huggingface.co./BlinkDL/rwkv-4-raven/tree/main

RWKV-4-Raven-7B-v10-Eng89%25-Jpn10%25-Other1%25-20230420-ctx4096.pth

Parameters: Lora Rank 1024 Lora Alpha 2048 ctx length 1024 Lora Size: almost 1.6GB

Dataset

https://huggingface.co./datasets/kunishou/hh-rlhf-49k-ja

https://huggingface.co./datasets/kunishou/databricks-dolly-15k-ja