Spaces:
Running
on
Zero
Apply for community grant: Personal project (gpu)
Hello, I’m a member of the VITA team. We’re excited to announce the release of our VITA 1.5 model, which has gained significant attention and popularity within the open-source community. To make it more accessible, we’re currently working on deploying it on HuggingFace Spaces, allowing more users to explore its capabilities.
I’m reaching out to apply for a community grant to support this project. We would greatly appreciate access to either 1 GPU with 80GB RAM or 2 GPUs with 40GB RAM each. Your support would be invaluable in helping us achieve this goal. Thank you for considering our request!
Hi, @hysts . Our basic demo requires only a single GPU machine with 40GB of VRAM. Is this setup available? We would greatly appreciate your support in this regard.
Thanks for your support!
Hello @hysts ,
I have a question regarding the demo deployment on ZeroGPU. Specifically, our demo requires modifications to model config files as well as files located in the .../lib/python3.10/site-packages/vllm/ directory.
You can refer to the following link for further details: Run VITA Demo.
I would like to know how I can achieve this without a PRO account.
Looking forward to your advice!
@lxysl
Oh, so, your demo depends on vLLM? Unfortunately, I don't think ZeroGPU is compatible with it. My understanding is that vLLM requires a dedicated GPU, but ZeroGPU only uses GPU during inference.
That said, I don't know much about vllm, so I might be misunderstanding something.
As for the original question, I think you can add subprocess.run
in your app.py
to run shell commands that modify those files. Also, I just enabled dev mode for this Space so that you can test it more easily. But please note that changes made from the terminal in dev mode will be refreshed when the Space is restarted, so ultimately, you need to make the changes using subprocess.run
.