YuJuLin
nps798
AI & ML interests
None yet
Organizations
None yet
nps798's activity
is it available to change output text format to Markdown?
8
#21 opened about 2 months ago
by
MhmdSyd
It also works with (Taiwan's) Traditinal Chainese.
#3 opened about 2 months ago
by
nps798
Brilliant !
#3 opened 5 months ago
by
nps798
Looking into training detail
#1 opened 8 months ago
by
nps798
Why do you add a normalize layer in the end of the model? Does it affect how we fine tune results?
3
#13 opened 8 months ago
by
nps798
How do you build your own MoE model?
4
#1 opened 9 months ago
by
nps798
The same audio. Why are the results on Hugging Face different from the results of the model on GitHub?
1
#31 opened 11 months ago
by
deepsadness
厲害!請教預訓練所需硬體資源 以及訓練代碼原理
3
#2 opened 10 months ago
by
nps798
Code to continue pretrain
#1 opened 11 months ago
by
nps798
Has a massive repetition problem
14
#29 opened about 1 year ago
by
Delcos
finetuning
3
#1 opened about 1 year ago
by
almugabo
Hi what did you train this model with, and what were hyperparams?
7
#1 opened about 1 year ago
by
teknium
Excellent model ! Asking about training details
#3 opened about 1 year ago
by
nps798
How is this model different from Llama 2-7B?
7
#8 opened about 1 year ago
by
dheerajpai
Curious about how the model was trained to support Taiwan Chinese so well
1
#1 opened about 1 year ago
by
nps798
using oobaboga to load model fail for 70b chat ggml Q2_k anad Q3_K_S
2
#2 opened over 1 year ago
by
nps798