Spaces:
Running
on
T4
Running
on
T4
Update Dockerfile
Browse files- Dockerfile +7 -1
Dockerfile
CHANGED
@@ -1,9 +1,10 @@
|
|
1 |
FROM nvidia/cuda:12.2.2-cudnn8-runtime-ubuntu22.04 AS llama_builder
|
2 |
|
|
|
|
|
3 |
RUN apt-get update && apt-get install --no-install-recommends -y \
|
4 |
build-essential \
|
5 |
python3-pip \
|
6 |
-
python3-dev \
|
7 |
&& apt-get clean && rm -rf /var/lib/apt/lists/*
|
8 |
|
9 |
RUN python -m venv /opt/venv
|
@@ -15,6 +16,10 @@ RUN CMAKE_ARGS="-DLLAMA_CUDA=on" pip3 install --no-cache-dir llama-cpp-python
|
|
15 |
|
16 |
FROM nvcr.io/nvidia/pytorch:23.10-py3 AS dep_builder
|
17 |
|
|
|
|
|
|
|
|
|
18 |
RUN python -m venv /opt/venv
|
19 |
ENV PATH=/opt/venv/bin:$PATH \
|
20 |
NVIDIA_VISIBLE_DEVICES=all \
|
@@ -30,6 +35,7 @@ RUN apt-get update && apt-get install --no-install-recommends -y \
|
|
30 |
ffmpeg \
|
31 |
libsm6 \
|
32 |
libxext6 \
|
|
|
33 |
python3-pip \
|
34 |
python3-dev \
|
35 |
ninja-build \
|
|
|
1 |
FROM nvidia/cuda:12.2.2-cudnn8-runtime-ubuntu22.04 AS llama_builder
|
2 |
|
3 |
+
ARG DEBIAN_FRONTEND=noninteractive
|
4 |
+
|
5 |
RUN apt-get update && apt-get install --no-install-recommends -y \
|
6 |
build-essential \
|
7 |
python3-pip \
|
|
|
8 |
&& apt-get clean && rm -rf /var/lib/apt/lists/*
|
9 |
|
10 |
RUN python -m venv /opt/venv
|
|
|
16 |
|
17 |
FROM nvcr.io/nvidia/pytorch:23.10-py3 AS dep_builder
|
18 |
|
19 |
+
RUN apt-get update && apt-get install --no-install-recommends -y \
|
20 |
+
python3-pip \
|
21 |
+
&& apt-get clean && rm -rf /var/lib/apt/lists/*
|
22 |
+
|
23 |
RUN python -m venv /opt/venv
|
24 |
ENV PATH=/opt/venv/bin:$PATH \
|
25 |
NVIDIA_VISIBLE_DEVICES=all \
|
|
|
35 |
ffmpeg \
|
36 |
libsm6 \
|
37 |
libxext6 \
|
38 |
+
python3.10
|
39 |
python3-pip \
|
40 |
python3-dev \
|
41 |
ninja-build \
|