File size: 1,881 Bytes
83b66b6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
# PruneSLU-15M: Efficient Model for On-Device Spoken Language Understanding

**PruneSLU-15M** is an optimized version of the [openai/whisper-tiny.en](https://huggingface.co./openai/whisper-tiny.en) model, specifically tailored for Spoken Language Understanding (SLU) tasks. This model has been pruned and further fine-tuned to achieve a lightweight yet powerful solution, ideal for deployment in resource-constrained environments.

### Model Overview

- **Base Model:** [openai/whisper-tiny.en](https://huggingface.co./openai/whisper-tiny.en)
- **Task:** Spoken Language Understanding (SLU)
- **Dataset:** Fine-tuned on the [STOP dataset](https://github.com/facebookresearch/fairseq/tree/main/examples/audio_nlp/nlu)
- **Pruning Techniques:** Vocabulary pruning and layer-wise structural pruning have been applied to reduce the model size while maintaining performance. Post-pruning, the model was retrained to ensure robustness.

### Key Features

- **Compact Size:** With only 15 million parameters, PruneSLU-15M is highly efficient, making it suitable for on-device applications.
- **High Performance:** Despite significant pruning, the model retains strong performance on SLU tasks, as demonstrated by evaluations on the STOP dataset.
- **Easy Integration:** The model can be effortlessly loaded and used via the Hugging Face Transformers library.

### Usage

To load the PruneSLU-15M model in Hugging Face, use the following code:

```python
from transformers import WhisperForConditionalGeneration

model = WhisperForConditionalGeneration.from_pretrained("kodiak619/PruneSLU-15M")
```

### Applications
PruneSLU-15M is ideal for scenarios where computational resources are limited, such as mobile devices or embedded systems. It is particularly well-suited for tasks like voice command recognition, intent detection, and other SLU-related applications in low-resource settings.