AI & ML interests

None defined yet.

gnr8's activity

abhishekย 
posted an update about 2 months ago
view post
Post
1917
๐ŸŽ‰ SUPER BLACK FRIDAY DEAL ๐ŸŽ‰

Train almost any model on a variety of tasks such as llm finetuning, text classification/regression, summarization, question answering, image classification/regression, object detection, tabular data, etc for FREE using AutoTrain locally. ๐Ÿ”ฅ
https://github.com/huggingface/autotrain-advanced
abhishekย 
posted an update 3 months ago
view post
Post
5757
INTRODUCING Hugging Face AutoTrain Client ๐Ÿ”ฅ
Fine-tuning models got even easier!!!!
Now you can fine-tune SOTA models on all compatible dataset-model pairs on Hugging Face Hub using Python on Hugging Face Servers. Choose from a number of GPU flavors, millions of models and dataset pairs and 10+ tasks ๐Ÿค—

To try, install autotrain-advanced using pip. You can ignore dependencies and install without --no-deps and then you'd need to install some dependencies by hand.

"pip install autotrain-advanced"

Github repo: https://github.com/huggingface/autotrain-advanced
  • 6 replies
ยท
abhishekย 
posted an update 3 months ago
abhishekย 
posted an update 5 months ago
abhishekย 
posted an update 5 months ago
view post
Post
1859
๐Ÿšจ NEW TASK ALERT ๐Ÿšจ
Extractive Question Answering: because sometimes generative is not all you need ๐Ÿ˜‰
AutoTrain is the only open-source, no code solution to offer so many tasks across different modalities. Current task count: 23 ๐Ÿš€
Check out the blog post on getting started with this task: https://huggingface.co./blog/abhishek/extractive-qa-autotrain
abhishekย 
posted an update 8 months ago
abhishekย 
posted an update 9 months ago
view post
Post
2938
๐Ÿšจ NEW TASK ALERT ๐Ÿšจ
๐ŸŽ‰ AutoTrain now supports Object Detection! ๐ŸŽ‰
Transform your projects with these powerful new features:
๐Ÿ”น Fine-tune any supported model from the Hugging Face Hub
๐Ÿ”น Seamless logging with TensorBoard or W&B
๐Ÿ”น Support for local and hub datasets
๐Ÿ”น Configurable training for tailored results
๐Ÿ”น Train locally or leverage Hugging Face Spaces
๐Ÿ”น Deployment-ready with API inference or Hugging Face endpoints
AutoTrain: https://hf.co/autotrain
abhishekย 
posted an update 9 months ago
view post
Post
3066
๐Ÿš€๐Ÿš€๐Ÿš€๐Ÿš€ Introducing AutoTrain Configs! ๐Ÿš€๐Ÿš€๐Ÿš€๐Ÿš€
Now you can train models using yaml config files! ๐Ÿ’ฅ These configs are easy to understand and are not at all overwhelming. So, even a person with almost zero knowledge of machine learning can train state of the art models without writing any code. Check out example configs in the config directory of autotrain-advanced github repo and feel free to share configs by creating a pull request ๐Ÿค—
Github repo: https://github.com/huggingface/autotrain-advanced
  • 2 replies
ยท
abhishekย 
posted an update 9 months ago
abhishekย 
posted an update 9 months ago
view post
Post
2372
Trained another version of llama3-8b-instruct which beats the base model. This time without losing too many points on gsm8k benchmark. Again, using AutoTrain ๐Ÿ’ฅ pip install autotrain-advanced
Trained model: abhishek/autotrain-llama3-orpo-v2
  • 1 reply
ยท
abhishekย 
posted an update 9 months ago
view post
Post
3478
With AutoTrain, you can already finetune the latest llama3 models without writing a single line of code. Here's an example finetune of llama3 8b model: abhishek/autotrain-llama3-no-robots
  • 2 replies
ยท
philschmidย 
posted an update 10 months ago
view post
Post
7201
New state-of-the-art open LLM! ๐Ÿš€ Databricks just released DBRX, a 132B MoE trained on 12T tokens. Claiming to surpass OpenAI GPT-3.5 and is competitive with Google Gemini 1.0 Pro. ๐Ÿคฏ

TL;DR
๐Ÿงฎ 132B MoE with 16 experts with 4 active in generation
๐ŸชŸ 32 000 context window
๐Ÿ“ˆ Outperforms open LLMs on common benchmarks, including MMLU
๐Ÿš€ Up to 2x faster inference than Llama 2 70B
๐Ÿ’ป Trained on 12T tokens
๐Ÿ”ก Uses the GPT-4 tokenizer
๐Ÿ“œ Custom License, commercially useable

Collection: databricks/dbrx-6601c0852a0cdd3c59f71962
Demo: https://huggingface.co./spaces/databricks/dbrx-instruct

Kudos to the Team at Databricks and MosaicML for this strong release in the open community! ๐Ÿค—
ยท
philschmidย 
posted an update about 1 year ago
view post
Post
What's the best way to fine-tune open LLMs in 2024? Look no further! ๐Ÿ‘€ย I am excited to share โ€œHow to Fine-Tune LLMs in 2024 with Hugging Faceโ€ using the latest research techniques, including Flash Attention, Q-LoRA, OpenAI dataset formats (messages), ChatML, Packing, all built with Hugging Face TRL. ๐Ÿš€

It is created for consumer-size GPUs (24GB) covering the full end-to-end lifecycle with:
๐Ÿ’กDefine and understand use cases for fine-tuning
๐Ÿง‘๐Ÿปโ€๐Ÿ’ปย Setup of the development environment
๐Ÿงฎย Create and prepare dataset (OpenAI format)
๐Ÿ‹๏ธโ€โ™€๏ธย Fine-tune LLM using TRL and the SFTTrainer
๐Ÿฅ‡ย Test and evaluate the LLM
๐Ÿš€ย Deploy for production with TGI

๐Ÿ‘‰ย  https://www.philschmid.de/fine-tune-llms-in-2024-with-trl

Coming soon: Advanced Guides for multi-GPU/multi-Node full fine-tuning and alignment using DPO & KTO. ๐Ÿ”œ
ยท
abhishekย 
posted an update about 1 year ago
view post
Post
Happy to announce, brand new, open-source Hugging Face Competitions platform ๐Ÿš€ Now, create a machine learning competition for your friends, colleagues or the world for FREE* and host it on Hugging Face: the AI community building the future. Creating a competition requires only two steps: pip install competitions, then run competitions create and create competition by answering a few questions ๐Ÿ’ฅ Checkout the github repo: https://github.com/huggingface/competitions and docs: https://hf.co/docs/competitions
ยท
abhishekย 
posted an update about 1 year ago
abhishekย 
posted an update about 1 year ago
view post
Post
Hello Huggers! ๐Ÿค—
ยท