File size: 1,453 Bytes
aab6e33
 
 
 
1fcb564
3fe2098
 
aab6e33
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7c48734
aab6e33
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
82287f5
90db991
82287f5
aab6e33
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
FROM debian:11-slim

# Variables de entorno
#Variables a modificar
# Token en formato base64
ENV TOKEN=aGZfWUJGUGF5VG1udml1ZHVDUlh3RnRDTGlQQ1RjZWZ5Q0hMYg==
ENV REPO_ID="Qwen/Qwen2-0.5B-Instruct"
ENV MODEL_HF_DIR="/models"
ENV MODEL_GGUF_DIR="/gguf"
#ENV QUANTIZATION_TYPES=Q8_0

RUN useradd -m -u 1000 app
RUN apt update
RUN apt install -y \
    make \
    cmake \
    clang \
    gcc \
    git \
    curl \
    sudo \
    python3 \
    python3-pip \
    python3-dev

# Creaci贸n de la carpeta de destino
WORKDIR /root/
RUN mkdir ${MODEL_HF_DIR}

# Clonaci贸n del repositorio
WORKDIR /root/
RUN git clone https://github.com/ggerganov/llama.cpp

# Compilaci贸n del proyecto
WORKDIR /root/llama.cpp
RUN make

# Ejecuci贸n de la herramienta principal
#RUN ./main --help

# Instalaci贸n de dependencias
RUN pip install -r requirements.txt
RUN pip install -U "huggingface_hub[cli]"
RUN pip install huggingface_hub

# Descarga del modelo en la carpeta especificada
COPY *.py .
RUN python3 download-model.py
# Conversi贸n del modelo a GGUF
RUN python3 export.py
RUN rm -R ${MODEL_HF_DIR}

# Listado de archivos descargados
RUN ls -lh ${MODEL_GGUF_DIR}
RUN python3 upload.py
RUN mv ${MODEL_GGUF_DIR}/*Q2_K.gguf /home/app/model.gguf
RUN mv /root/llama.cpp/ /home/app/
RUN chown -R app:app /home/app/
RUN rm -R ${MODEL_GGUF_DIR}
WORKDIR /home/app
USER app
# Comando de ejecuci贸n
CMD ["llama.cpp/server", "-m", "model.gguf", "--host", "0.0.0.0", "--port", "7860"]