YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co./docs/hub/model-cards#model-card-metadata)

Llama-3-3B CodeSearchNet Fine-tuned

This repository hosts a Llama 3 (3B) model fine-tuned on the CodeSearchNet dataset, which contains code in six programming languages.

πŸ“ Model Details

  • Base Model: Llama 3 (3B)
  • Fine-tuning Dataset: CodeSearchNet
  • Languages Covered: Python, Java, JavaScript, PHP, Ruby, Go
  • Training Method: Supervised fine-tuning (SFT) with a contrastive loss objective for code search tasks
  • Tokenization: Llama 3 tokenizer with additional tokens for code-specific keywords
  • Frameworks Used: Hugging Face transformers, PyTorch, PEFT (for LoRA-based tuning)

πŸ“š Dataset

The model is trained on the CodeSearchNet dataset, which contains:

  • Function-level code snippets
  • Paired natural language descriptions
  • Multiple programming languages for multi-language search support

Dataset Sources

πŸš€ Training Setup

  • Hardware: NVIDIA A100 GPUs
  • Batch Size: 16
  • Learning Rate: 2e-5 with cosine annealing
  • Max Sequence Length: 512
  • Fine-tuning Duration: 3 epochs

πŸ” Intended Use

  • Code Search: Retrieve relevant code snippets given a natural language query
  • Code Completion: Provide context-aware code suggestions
  • Code-to-Text Generation: Explain code functionality in natural language
  • Multi-language Code Retrieval: Search across different programming languages
Downloads last month
14
Safetensors
Model size
3.21B params
Tensor type
F32
Β·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.