File size: 1,303 Bytes
516a027
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
python main. py 
    Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
model. safetensors.index.json: 100%|
| 13.5k/13.5k [00:00‹?, PB/s]
model-00001-of-00002. safetensors: 100%
| 4.95G/4.95G [07:27<00:00, 11. 1MB/s]
model-00002-of-00002. safetensors: 100%
67. 1M/67.1M [00:05<00:00, 11.5MB/s]
Downloading shards: 100% ||
| 2/2 [07:3500:00, 227.61s/it]
Gemma's activation function should be approximate GeLU and not exact GeLU. Changing the activation function to 'gelu_pytorch_tanh.if you want to use the legacy "gelu', edit the "model.config to
set hidden_activation=gelu*
instead of todden act
instead of hidden_act. See https://github.com/huggingface/transformers/pull/29402 for
more details.
Loading checkpoint shards: 100%|
| 2/2 [00:03<00:00, 1.87s/itl
generation_config json: 100%||
137/13700:00<?」3B/s]
nexa model result:
a pouto using the specified caea and resolutiou stones iption: rame rs a photo (cama a):)
Captures
- camera (str): Specifies the camera
to use. Can be \'front\' or \'back\'. The default is \'back\'. \n\n
Returns: \n
- str: The string contains the file
2624 t 12 4a. 
Photo if nees at ay 96 83662387968t, ample: /storage/emulated/o/Pictures/NAPP/3N
123456.Jpg\'\n latency: 367.85967230796814