Spaces:
Sleeping
Sleeping
modified README.md
Browse files
README.md
CHANGED
@@ -10,4 +10,25 @@ pinned: false
|
|
10 |
license: mit
|
11 |
---
|
12 |
|
13 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
10 |
license: mit
|
11 |
---
|
12 |
|
13 |
+
A simple interface for testing out NA-LLM-qa model based on Llama-3-8B checkpoint.
|
14 |
+
|
15 |
+
# Use
|
16 |
+
|
17 |
+
```sh
|
18 |
+
python -m venv .venv
|
19 |
+
source .venv/bin/activate
|
20 |
+
pip install -r requirements.txt
|
21 |
+
gradio app.py
|
22 |
+
```
|
23 |
+
|
24 |
+
# Notes
|
25 |
+
|
26 |
+
- The model is hosted with HuggingFace Inference Endpoint.
|
27 |
+
- The Endpoint may be paused due to inactivity. In that case, calling a signal will "wake up" the endpoint, but it would take around several minutes.
|
28 |
+
- For now, the endpoint is gated. Set appropriate `hf_token` with READ permission to the organization.
|
29 |
+
- Input filtering
|
30 |
+
- The model performs unexpectedly for non-questions
|
31 |
+
- For this reason, a simple SVM-based filter is applied
|
32 |
+
- The filter is a `OneClassSVM` trained with question sections of na-llm
|
33 |
+
- The model, along with corresponding vectorizer, is saved in `question_undetector.pkl` as `(vectorizer, model)` object.
|
34 |
+
- The hosting machine should be powerful enough at least to run a simple SVM model
|