File size: 1,078 Bytes
6438b36
 
09245b4
 
 
6438b36
 
 
 
 
 
 
 
 
09245b4
6438b36
60cc967
6438b36
60cc967
 
 
6438b36
60cc967
6438b36
60cc967
 
 
6438b36
d10e001
6438b36
09245b4
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
---
library_name: transformers
license: apache-2.0
language:
- ko
---

# Model Card for Model ID


## Model Details

### Model Description

The Gemma Self-Attention Merged model is a large language model created by merging the self-attention layers of an [English-based Gemma 7B model](https://huggingface.co./google/gemma-1.1-7b-it) and a [Korean-based Gemma 7B model](beomi/gemma-ko-7b). This merger allows the model to leverage the capabilities of both the English and Korean models, resulting in a more versatile and capable language model that can perform well on tasks involving both English and Korean text.

The key features of this merged model include:

- Increased self-attention capacity with doubled number of attention heads
- Ability to handle both English and Korean language input
- Potential for improved performance on a wide range of natural language processing tasks

#### Chat template

**system:** system message...  
**B:** user message...  
**A:** assistant message...  

### Model Sources

- **Repository:** https://github.com/lcw99/merge-gemma-attn.git