Thanks for being patient! πŸ’œπŸ’œ

Model Card for Model ID

A fine-tuned version of Gemma2 9B for the Telugu Language Instruct on eswardivi/Telugu_InstructData dataset using Unsloth.

Model Details

Model Description

This model has been finetuned with eswardivi/Telugu_InstructData on unsloth/gemma-2-9b. This is to ensure there is a raise in Telugu Language Model usage in daily applications. This model will respond in english text where as the content is in telugu language.

For example instead of ఎలా ఉన్నాడ్? it will respond with ela unnav?. This is to ensure the model is not using more tokens for unicode characters.

  • Developed by: pavankumarbalijepalli
  • Model type: CASUAL_LM
  • Language(s) (NLP): English, Telugu
  • License: MIT
  • Finetuned from model: unsloth/gemma-2-9b

Model Sources

Uses

Model is supposed to be used for the cases where you have a natural language question in telugu language to get a response in return in telugu language. The context should be below 8192 tokens. The output will be generated in telugu.

Downstream Use

# WILL ADD CODE HERE SOON.

Out-of-Scope Use

Generating Unintended Code:

While the model can converse in telugu language, it may not be robust enough to handle complex logic or edge cases. Using it to generate critical production code could lead to errors or unexpected behavior in databases.

Security Risks:

will_add_soon.

Beyond its Training Scope:

The model is trained on telugu Language. Using it for a different languages could lead to inaccurate or nonsensical responses.

Bias, Risks, and Limitations

Bias and Fairness:

The model's training data may contain biases that are reflected in the responses. This could lead to unfair or discriminatory outcomes, especially if the data is not carefully curated.

Interpretability and Explainability:

Language models are often "black boxes" where it's difficult to understand how they learn a specific language, in this case Telugu. This lack of interpretability makes it challenging to debug errors or ensure the generated queries are safe and efficient.

Replacing Human Expertise:

will_add_soon

Recommendations

Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model.

Training Details

Training Data

@misc {divi_eswar_chowdary_2025,
author = { {Divi Eswar Chowdary} },
title = { Telugu_InstructData (Revision b734b7d) },
year = 2025,
url = { https://huggingface.co./datasets/eswardivi/Telugu_InstructData },
doi = { 10.57967/hf/4666 },
publisher = { Hugging Face }
}

Evaluation

Testing Data, Factors & Metrics

Testing Data

Used eswardivi/Telugu_InstructData and split the data into training and testing datasets. The holdout dataset is used for testing the model.

############# WILL UPDATE SOON #############
#### Factors

<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
The complexity of the questions are calculated using the number of tables per question, number of joins, group by, and sub queries per answer. This complexity is used to prepare the test data by stratifying the split around the complexity.

#### Metrics

<!-- These are the evaluation metrics being used, ideally with a description of why. -->
* __Execution Success:__ This metric is used to find out if the generated query is executable without arising any errors. For this, a sqllite3 connection is made to the memory, and using context the dummy tables are created. Then the predicted SQL is executed. This checks out if the generated query is in proper syntax, and if the model is hallucinating any new columns.
* __Inference Time:__ This metric is used to find out which model is providing results in less amount of time. This combined with the execution success, gives the efficiency of the model.
- 
### Results

* __Execution Success:__ Finetuned Phi-2 has 29% more success rate than the SQLCoder-7b-2
* __Inference Time:__ Finetuned Phi-2 has 41% increased inference speed than SQLCoder-7b-2  

#### Summary
* __Reduced Inference Time and Memory Footprint:__ The fine-tuned Phi-2 model
demonstrated a reduction in inference time and memory usage compared to the DeFog
SQLCoder. This is attributed to Phi-2's smaller size and the efficiency of quantization
techniques employed during fine-tuning. This finding implies that NL2SQL models can
be deployed on lower-powered devices like laptops or even mobile phones, potentially
democratizing access to this technology for a wider range of users.

* __Competitive Performance on Easy and Medium Queries:__ The fine-tuned Phi-2
achieved comparable performance to the DeFog SQLCoder in terms of accuracy on easy,
medium, and hard difficulty queries. This indicates that Phi-2, despite its smaller size,
can effectively handle a significant portion of real-world NL2SQL tasks, especially for
simpler queries.

* __Challenges with Complex Queries:__ While Phi-2 performed well on easier queries, it
encountered challenges with complex queries, exhibiting a drop in execution success
compared to the DeFog SQLCoder. This highlights the trade-off between model size and
complexity, suggesting that larger models might still be necessary for tackling highly
intricate tasks.

* __Potential for Further Improvement:__ The fine-tuning process employed in this study
can be further optimized by exploring different hyperparameter configurations and
potentially investigating alternative fine-tuning techniques like adapter-based methods.
This optimization has the potential to improve the model's performance on complex
queries while maintaining its efficiency.

Environmental Impact

Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).

  • Hardware Type: RTX 4070 PCIE 12GB X 1
  • Hours used: 14 Hours
  • Cloud Provider: Private Local Machine
  • Compute Region: Asia-East-1
  • Carbon Emitted: 1.81 kg eq. CO2

Citation

BibTeX:

@misc {pavan_kumar_balijepalli_2025,
    author       = { {Pavan Kumar Balijepalli} },
    title        = { telLM-gemma2-9b (Revision e6f5f18) },
    year         = 2025,
    url          = { https://huggingface.co./pavankumarbalijepalli/telLM-gemma2-9b },
    doi          = { 10.57967/hf/4668 },
    publisher    = { Hugging Face }
}
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no pipeline_tag.

Model tree for pavankumarbalijepalli/telLM-gemma2-9b

Base model

unsloth/gemma-2-9b
Finetuned
(39)
this model

Dataset used to train pavankumarbalijepalli/telLM-gemma2-9b

Collection including pavankumarbalijepalli/telLM-gemma2-9b