A newer version of this model is available: prithivMLmods/QwQ-LCoT2-7B-Instruct

QwQ-LCoT-7B-Instruct Model File

The QwQ-LCoT-7B-Instruct is a fine-tuned language model designed for advanced reasoning and instruction-following tasks. It leverages the Qwen2.5-7B base model and has been fine-tuned on the amphora/QwQ-LongCoT-130K dataset, focusing on chain-of-thought (CoT) reasoning. This model is optimized for tasks requiring logical reasoning, detailed explanations, and multi-step problem-solving, making it ideal for applications such as instruction-following, text generation, and complex reasoning tasks.

Quickstart with Transformers

Here provides a code snippet with apply_chat_template to show you how to load the tokenizer and model and how to generate contents.

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "prithivMLmods/QwQ-LCoT-7B-Instruct"

model = AutoModelForCausalLM.from_pretrained(
    model_name,
    torch_dtype="auto",
    device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained(model_name)

prompt = "How many r in strawberry."
messages = [
    {"role": "system", "content": "You are a helpful and harmless assistant. You are Qwen developed by Alibaba. You should think step-by-step."},
    {"role": "user", "content": prompt}
]
text = tokenizer.apply_chat_template(
    messages,
    tokenize=False,
    add_generation_prompt=True
)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)

generated_ids = model.generate(
    **model_inputs,
    max_new_tokens=512
)
generated_ids = [
    output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
]

response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]

Sample Long CoT:

Screenshot 2024-12-13 211732.png


Key Features:

  1. Model Size:

    • 7.62B parameters (FP16 precision).
  2. Model Sharding:

    • The model weights are split into 4 shards (safetensors) for efficient storage and download:
      • model-00001-of-00004.safetensors (4.88 GB)
      • model-00002-of-00004.safetensors (4.93 GB)
      • model-00003-of-00004.safetensors (4.33 GB)
      • model-00004-of-00004.safetensors (1.09 GB)
  3. Tokenizer:

    • Byte-pair encoding (BPE) based.
    • Files included:
      • vocab.json (2.78 MB)
      • merges.txt (1.82 MB)
      • tokenizer.json (11.4 MB)
    • Special tokens mapped in special_tokens_map.json (e.g., <pad>, <eos>).
  4. Configuration Files:

    • config.json: Defines model architecture and hyperparameters.
    • generation_config.json: Settings for inference and text generation tasks.

Training Dataset:


Use Cases:

  1. Instruction Following:
    Handle user instructions effectively, even for multi-step tasks.

  2. Reasoning Tasks:
    Perform logical reasoning and generate detailed step-by-step solutions.

  3. Text Generation:
    Generate coherent, context-aware responses.


Downloads last month
433
Safetensors
Model size
7.62B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for prithivMLmods/QwQ-LCoT-7B-Instruct

Base model

Qwen/Qwen2.5-7B
Finetuned
(212)
this model
Merges
13 models
Quantizations
13 models

Dataset used to train prithivMLmods/QwQ-LCoT-7B-Instruct

Spaces using prithivMLmods/QwQ-LCoT-7B-Instruct 2

Collections including prithivMLmods/QwQ-LCoT-7B-Instruct