pdr209 commited on
Commit
f20a742
·
1 Parent(s): 542a8df

Update lcm/README

Browse files
Files changed (1) hide show
  1. lcm/README.md +7 -5
lcm/README.md CHANGED
@@ -6,17 +6,18 @@ language:
6
  tags:
7
  - stable-diffusion
8
  - stable-diffusion-xl
 
9
  - tensorrt
10
  - text-to-image
11
  ---
12
 
13
- # Stable Diffusion XL 1.0 TensorRT
14
 
15
  ## Introduction
16
 
17
  This repository hosts the Latent Consistency Model(LCM) TensorRT versions of **Stable Diffusion XL 1.0** created in collaboration with [NVIDIA](https://huggingface.co/nvidia). The optimized versions give substantial improvements in speed and efficiency.
18
 
19
- See the [usage instructions](#usage-example) for how to run the SDXL pipeline with the ONNX files hosted in this repository. The first invocation produces plan files in `engine_xl_base` and `engine_xl_refiner` specific to the accelerator being run on and are reused for later invocations.
20
 
21
  ## Model Description
22
 
@@ -26,7 +27,7 @@ See the [usage instructions](#usage-example) for how to run the SDXL pipeline wi
26
  - **Model Description:** This is a Latent Consistency Model (LCM) version of the [SDXL base 1.0](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0) and [SDXL refiner 1.0](https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-1.0) models for [NVIDIA TensorRT](https://developer.nvidia.com/tensorrt) optimized inference
27
 
28
 
29
- ## Performance Comparison
30
 
31
  #### Timings for 4 steps at 1024x1024
32
 
@@ -38,7 +39,7 @@ See the [usage instructions](#usage-example) for how to run the SDXL pipeline wi
38
 
39
  ## Usage Example
40
 
41
- 1. Following the [setup instructions](https://github.com/rajeevsrao/TensorRT/blob/release/8.6/demo/Diffusion/README.md) on launching a TensorRT NGC container.
42
  ```shell
43
  git clone https://github.com/rajeevsrao/TensorRT.git
44
  cd TensorRT
@@ -50,7 +51,7 @@ docker run --rm -it --gpus all -v $PWD:/workspace nvcr.io/nvidia/pytorch:23.11-p
50
  ```shell
51
  git lfs install
52
  git clone https://huggingface.co/stabilityai/stable-diffusion-xl-1.0-tensorrt
53
- cd stable-diffusion-xl-1.0-tensorrt
54
  git lfs pull
55
  cd ..
56
  ```
@@ -64,6 +65,7 @@ python3 -m pip install --pre --upgrade --extra-index-url https://pypi.nvidia.com
64
  ```
65
 
66
  4. Perform TensorRT optimized inference
 
67
  ```
68
  python3 demo_txt2img_xl.py \
69
  ""Astronaut in a jungle, cold color palette, muted colors, detailed, 8k"" \
 
6
  tags:
7
  - stable-diffusion
8
  - stable-diffusion-xl
9
+ - stable-diffusion-xl-lcm
10
  - tensorrt
11
  - text-to-image
12
  ---
13
 
14
+ # Stable Diffusion XL 1.0 LCM TensorRT
15
 
16
  ## Introduction
17
 
18
  This repository hosts the Latent Consistency Model(LCM) TensorRT versions of **Stable Diffusion XL 1.0** created in collaboration with [NVIDIA](https://huggingface.co/nvidia). The optimized versions give substantial improvements in speed and efficiency.
19
 
20
+ See the [usage instructions](#usage-example) for how to run the SDXL pipeline with the ONNX files hosted in this repository.
21
 
22
  ## Model Description
23
 
 
27
  - **Model Description:** This is a Latent Consistency Model (LCM) version of the [SDXL base 1.0](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0) and [SDXL refiner 1.0](https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-1.0) models for [NVIDIA TensorRT](https://developer.nvidia.com/tensorrt) optimized inference
28
 
29
 
30
+ ## Performance
31
 
32
  #### Timings for 4 steps at 1024x1024
33
 
 
39
 
40
  ## Usage Example
41
 
42
+ 1. Following the [setup instructions](https://github.com/rajeevsrao/TensorRT/blob/release/9.2/demo/Diffusion/README.md) on launching a TensorRT NGC container.
43
  ```shell
44
  git clone https://github.com/rajeevsrao/TensorRT.git
45
  cd TensorRT
 
51
  ```shell
52
  git lfs install
53
  git clone https://huggingface.co/stabilityai/stable-diffusion-xl-1.0-tensorrt
54
+ cd stable-diffusion-xl-1.0-tensorrt/lcm
55
  git lfs pull
56
  cd ..
57
  ```
 
65
  ```
66
 
67
  4. Perform TensorRT optimized inference
68
+ * The first invocation produces plan files in --engine-dir specific to the accelerator being run on and are reused for later invocations.
69
  ```
70
  python3 demo_txt2img_xl.py \
71
  ""Astronaut in a jungle, cold color palette, muted colors, detailed, 8k"" \