File size: 3,402 Bytes
edcecaf
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
74a15e1
 
 
 
 
 
 
5e705c2
74a15e1
d991380
 
 
b2a0c84
d991380
 
 
 
 
 
 
3d1dd7a
d991380
 
 
 
 
 
3d1dd7a
d991380
 
 
 
 
 
 
 
5e705c2
 
 
 
cd32f9a
5e705c2
63e407b
 
 
 
 
ebe22e5
63e407b
 
 
 
cd32f9a
63e407b
f0b8cc2
63e407b
3d1dd7a
569d420
 
63e407b
 
af403d0
63e407b
 
569d420
f0b8cc2
5e705c2
63e407b
ebe22e5
63e407b
ebe22e5
63e407b
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
---
base_model: unsloth/llama-3-8b-Instruct-bnb-4bit
language:
- en
license: apache-2.0
tags:
- text-generation-inference
- transformers
- unsloth
- llama
- gguf
---

# Uploaded  model

- **Developed by:** Deeokay
- **License:** apache-2.0
- **Finetuned from model :** unsloth/llama-3-8b-Instruct-bnb-4bit

This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.

[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)


# README

This is a test model on a the following 
- a private dataset
- slight customization on llama3 template (no new tokens | no new configs)
- Works with Ollama create with just "FROM path/to/model" as Modelfile (llama3 template works no issues)

# HOW TO USE

The whole point of conversion for me was I wanted to be able to to use it through Ollama or (other local options)
For Ollama, it required to be a GGUF file. Once you have this it is pretty straight forward (if it is in llama3 which this model is)

Quick Start: 
- You must already have Ollama running in your setting
- Download the unsloth.Q4_K_M.gguf model from Files
- In the same directory create a file call "Modelfile"
- Inside the "Modelfile" type

```python
FROM ./unsloth.Q4_K_M.gguf

```
- Save a go back to the folder (folder where model + Modelfile exisit)
- Now in terminal make sure you are in the same location of the folder and type in the following command

```python
ollama create mycustomai  # "mycustomai" <- you can name it anything u want
```

This GGUF is based on llama3-3-8B-Instruct thus ollama doesn't need anything else to auto configure this model 

After than you should be able to use this model to chat! 


# NOTE: DISCLAIMER

Please note this is not for the purpose of production, but result of Fine Tuning through self learning

The llama3 Special Tokens where kept the same, however the format was slight customized using the available tokens

I have foregone the {{.System}} part as this would be updated when converting the llama3. 

I wanted to test if the model would understand additional headers that I created such as what my datasets has 
- Analaysis, Classification, Sentiment

Mulitple pass through my ~70K personalized customized dataset. 

If would like to know how I started creating my dataset, you can check this link 
[Crafting GPT2 for Personalized AI-Preparing Data the Long Way (Part1)](https://medium.com/@deeokay/the-soul-in-the-machine-crafting-gpt2-for-personalized-ai-9d38be3f635f)

As the data was getting created with custom GPT2 special tokens, I had to convert that to the llama3 Template. 

However I got creative again.. the training data has the following Template:

```
<|begin_of_text|>
<|start_header_id|>user<|end_header_id|>
{{.Prompt}}<|eot_id|><|start_header_id|>analysis<|end_header_id|>
{{.Analysis}}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
{{.Response}}<|eot_id|><|start_header_id|>classification<|end_header_id|>
{{.Classification}}<|eot_id|><|start_header_id|>sentiment<|end_header_id|>
{{.Sentiment}}<|eot_id|> <|start_header_id|>user<|end_header_id|>
<|end_of_text|> 

```

The llama3 standard template holds, and can be created in Ollama through normal llama3 template

Will be updating this periodically.. as I have limited colab resources..