gyulukeyi commited on
Commit
79e94d6
1 Parent(s): d56f0c8

modified README.md

Browse files
Files changed (1) hide show
  1. README.md +22 -1
README.md CHANGED
@@ -10,4 +10,25 @@ pinned: false
10
  license: mit
11
  ---
12
 
13
- An example chatbot using [Gradio](https://gradio.app), [`huggingface_hub`](https://huggingface.co/docs/huggingface_hub/v0.22.2/en/index), and the [Hugging Face Inference API](https://huggingface.co/docs/api-inference/index).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  license: mit
11
  ---
12
 
13
+ A simple interface for testing out NA-LLM-qa model based on Llama-3-8B checkpoint.
14
+
15
+ # Use
16
+
17
+ ```sh
18
+ python -m venv .venv
19
+ source .venv/bin/activate
20
+ pip install -r requirements.txt
21
+ gradio app.py
22
+ ```
23
+
24
+ # Notes
25
+
26
+ - The model is hosted with HuggingFace Inference Endpoint.
27
+ - The Endpoint may be paused due to inactivity. In that case, calling a signal will "wake up" the endpoint, but it would take around several minutes.
28
+ - For now, the endpoint is gated. Set appropriate `hf_token` with READ permission to the organization.
29
+ - Input filtering
30
+ - The model performs unexpectedly for non-questions
31
+ - For this reason, a simple SVM-based filter is applied
32
+ - The filter is a `OneClassSVM` trained with question sections of na-llm
33
+ - The model, along with corresponding vectorizer, is saved in `question_undetector.pkl` as `(vectorizer, model)` object.
34
+ - The hosting machine should be powerful enough at least to run a simple SVM model