wesley7137's picture
Create README.md
a049b1c verified
|
raw
history blame
3.92 kB
# Question Complexity Rating Dataset
## Dataset Description
This dataset contains pairs of questions and their respective complexity ratings. The complexity ratings are provided on a scale from 0.0 to 1.0, where 0.0 indicates the simplest questions and 1.0 indicates the most complex questions. The dataset is intended to be used for training and evaluating models that classify or rank questions based on their complexity.
## Dataset Contents
The dataset is provided as a CSV file with the following columns:
- `instruction`: The question or prompt.
- `label`: The complexity rating of the question, a float between 0.0 and 1.0.
### Example
| instruction | label |
| ---------------------------------------------- | ----- |
| When did Virgin Australia start operating? | 0.2 |
| Which is a species of fish? Tope or Rope | 0.8 |
| Why can camels survive for long without water? | 0.5 |
## Usage
The dataset can be used for various natural language processing (NLP) tasks such as:
- Complexity prediction
- Question ranking
- Model evaluation
- Transfer learning for related tasks
## Use Cases
### 1. Complexity Prediction
Train a model to predict the complexity of new questions based on the provided ratings. This can be useful in educational technology, customer support, and other domains where question complexity is a relevant factor.
### 2. Question Ranking
Rank a set of questions by their complexity to prioritize easier questions for beginners and more complex questions for advanced users.
### 3. Model Evaluation
Use the dataset to evaluate the performance of models designed to assess or rank question complexity.
## How to Use the Dataset
To use the dataset, load the CSV file using a library like pandas and preprocess it as needed for your specific use case.
### Loading the Dataset
```python
import pandas as pd
# Load the dataset
df = pd.read_csv('path_to_your_dataset.csv')
# Display the first few rows of the dataset
print(df.head())
```
## Training a Model
Here is a simple example of how to train a model using the dataset:
```python
import pandas as pd
from sklearn.model_selection import train_test_split
from sklearn.linear_model import LinearRegression
# Load the dataset
df = pd.read_csv('path_to_your_dataset.csv')
# Split the dataset into training and test sets
train_df, test_df = train_test_split(df, test_size=0.2, random_state=42)
# Extract features and labels
X_train = train_df['instruction']
y_train = train_df['label']
X_test = test_df['instruction']
y_test = test_df['label']
# Convert text data to numerical features (example using TF-IDF)
from sklearn.feature_extraction.text import TfidfVectorizer
vectorizer = TfidfVectorizer()
X_train_tfidf = vectorizer.fit_transform(X_train)
X_test_tfidf = vectorizer.transform(X_test)
# Train a simple model (example using Linear Regression)
model = LinearRegression()
model.fit(X_train_tfidf, y_train)
# Evaluate the model
predictions = model.predict(X_test_tfidf)
print(f"Predictions: {predictions}")
```
Citation
If you use this dataset in your research, please cite it as follows:
```sql
@dataset{question_complexity_rating_2024,
title = {Question Complexity Rating Dataset},
author = {Wes Lagarde},
year = {2024},
url = {https://huggingface.co./wesley7137/question_complexity_classification},
description = {A dataset of question and complexity rating pairs for natural language processing tasks.}
}
```
License
This dataset is licensed under the Apache 2.0 License.
Acknowledgements
We would like to thank all contributors and the community for their valuable feedback and support.
css
Copy code
This README file provides a comprehensive overview of your dataset, including its contents, usage, and example code for loading and using the dataset. Adjust the content as necessary to fit your specific dataset details and preferences.