Clelia Astra Bertelli's picture

Clelia Astra Bertelli

as-cle-bert

AI & ML interests

Biology + Artificial Intelligence = โค๏ธ | AI for sustainable development, sustainable development for AI | Researching on Machine Learning Enhancement | I love automation for everyday things | Blogger | Open Source

Recent Activity

posted an update about 12 hours ago
I just released a fully automated evaluation framework for your RAG applications!๐Ÿ“ˆ GitHub ๐Ÿ‘‰ https://github.com/AstraBert/diRAGnosis PyPi ๐Ÿ‘‰ https://pypi.org/project/diragnosis/ It's called ๐๐ข๐‘๐€๐†๐ง๐จ๐ฌ๐ข๐ฌ and is a lightweight framework that helps you ๐—ฑ๐—ถ๐—ฎ๐—ด๐—ป๐—ผ๐˜€๐—ฒ ๐˜๐—ต๐—ฒ ๐—ฝ๐—ฒ๐—ฟ๐—ณ๐—ผ๐—ฟ๐—บ๐—ฎ๐—ป๐—ฐ๐—ฒ ๐—ผ๐—ณ ๐—Ÿ๐—Ÿ๐— ๐˜€ ๐—ฎ๐—ป๐—ฑ ๐—ฟ๐—ฒ๐˜๐—ฟ๐—ถ๐—ฒ๐˜ƒ๐—ฎ๐—น ๐—บ๐—ผ๐—ฑ๐—ฒ๐—น๐˜€ ๐—ถ๐—ป ๐—ฅ๐—”๐—š ๐—ฎ๐—ฝ๐—ฝ๐—น๐—ถ๐—ฐ๐—ฎ๐˜๐—ถ๐—ผ๐—ป๐˜€. You can launch it as an application locally (it's Docker-ready!๐Ÿ‹) or, if you want more flexibility, you can integrate it in your code as a python package๐Ÿ“ฆ The workflow is simple: ๐Ÿง  You choose your favorite LLM provider and model (supported, for now, are Mistral AI, Groq, Anthropic, OpenAI and Cohere) ๐Ÿง  You pick the embedding models provider and the embedding model you prefer (supported, for now, are Mistral AI, Hugging Face, Cohere and OpenAI) ๐Ÿ“„ You prepare and provide your documents โš™๏ธ Documents are ingested into a Qdrant vector database and transformed into a synthetic question dataset with the help of LlamaIndex ๐Ÿ“Š The LLM is evaluated for the faithfulness and relevancy of its retrieval-augmented answer to the questions ๐Ÿ“Š The embedding model is evaluated for hit rate and mean reciprocal ranking (MRR) of the retrieved documents And the cool thing is that all of this is ๐—ถ๐—ป๐˜๐˜‚๐—ถ๐˜๐—ถ๐˜ƒ๐—ฒ ๐—ฎ๐—ป๐—ฑ ๐—ฐ๐—ผ๐—บ๐—ฝ๐—น๐—ฒ๐˜๐—ฒ๐—น๐˜† ๐—ฎ๐˜‚๐˜๐—ผ๐—บ๐—ฎ๐˜๐—ฒ๐—ฑ: you plug it in, and it works!๐Ÿ”Œโšก Even cooler? This is all built on top of LlamaIndex and its integrations: no need for tons of dependencies or fancy workarounds๐Ÿฆ™ And if you're a UI lover, Gradio and FastAPI are there to provide you a seamless backend-to-frontend experience๐Ÿ•ถ๏ธ So now it's your turn: you can either get diRAGnosis from GitHub ๐Ÿ‘‰ https://github.com/AstraBert/diRAGnosis or just run a quick and painless: ```bash uv pip install diragnosis ``` To get the package installed (lightning-fast) in your environment๐Ÿƒโ€โ™€๏ธ Have fun and feel free to leave feedback and feature/integrations requests on GitHub issuesโœจ
commented on their article 1 day ago
streamlit_supabase_auth_ui
commented on their article 14 days ago
streamlit_supabase_auth_ui
View all activity

Organizations

Social Post Explorers's profile picture Hugging Face Discord Community's profile picture GreenFit AI's profile picture

Posts 38

view post
Post
607
I just released a fully automated evaluation framework for your RAG applications!๐Ÿ“ˆ

GitHub ๐Ÿ‘‰ https://github.com/AstraBert/diRAGnosis
PyPi ๐Ÿ‘‰ https://pypi.org/project/diragnosis/

It's called ๐๐ข๐‘๐€๐†๐ง๐จ๐ฌ๐ข๐ฌ and is a lightweight framework that helps you ๐—ฑ๐—ถ๐—ฎ๐—ด๐—ป๐—ผ๐˜€๐—ฒ ๐˜๐—ต๐—ฒ ๐—ฝ๐—ฒ๐—ฟ๐—ณ๐—ผ๐—ฟ๐—บ๐—ฎ๐—ป๐—ฐ๐—ฒ ๐—ผ๐—ณ ๐—Ÿ๐—Ÿ๐— ๐˜€ ๐—ฎ๐—ป๐—ฑ ๐—ฟ๐—ฒ๐˜๐—ฟ๐—ถ๐—ฒ๐˜ƒ๐—ฎ๐—น ๐—บ๐—ผ๐—ฑ๐—ฒ๐—น๐˜€ ๐—ถ๐—ป ๐—ฅ๐—”๐—š ๐—ฎ๐—ฝ๐—ฝ๐—น๐—ถ๐—ฐ๐—ฎ๐˜๐—ถ๐—ผ๐—ป๐˜€.

You can launch it as an application locally (it's Docker-ready!๐Ÿ‹) or, if you want more flexibility, you can integrate it in your code as a python package๐Ÿ“ฆ

The workflow is simple:
๐Ÿง  You choose your favorite LLM provider and model (supported, for now, are Mistral AI, Groq, Anthropic, OpenAI and Cohere)
๐Ÿง  You pick the embedding models provider and the embedding model you prefer (supported, for now, are Mistral AI, Hugging Face, Cohere and OpenAI)
๐Ÿ“„ You prepare and provide your documents
โš™๏ธ Documents are ingested into a Qdrant vector database and transformed into a synthetic question dataset with the help of LlamaIndex
๐Ÿ“Š The LLM is evaluated for the faithfulness and relevancy of its retrieval-augmented answer to the questions
๐Ÿ“Š The embedding model is evaluated for hit rate and mean reciprocal ranking (MRR) of the retrieved documents

And the cool thing is that all of this is ๐—ถ๐—ป๐˜๐˜‚๐—ถ๐˜๐—ถ๐˜ƒ๐—ฒ ๐—ฎ๐—ป๐—ฑ ๐—ฐ๐—ผ๐—บ๐—ฝ๐—น๐—ฒ๐˜๐—ฒ๐—น๐˜† ๐—ฎ๐˜‚๐˜๐—ผ๐—บ๐—ฎ๐˜๐—ฒ๐—ฑ: you plug it in, and it works!๐Ÿ”Œโšก

Even cooler? This is all built on top of LlamaIndex and its integrations: no need for tons of dependencies or fancy workarounds๐Ÿฆ™
And if you're a UI lover, Gradio and FastAPI are there to provide you a seamless backend-to-frontend experience๐Ÿ•ถ๏ธ

So now it's your turn: you can either get diRAGnosis from GitHub ๐Ÿ‘‰ https://github.com/AstraBert/diRAGnosis
or just run a quick and painless:

uv pip install diragnosis


To get the package installed (lightning-fast) in your environment๐Ÿƒโ€โ™€๏ธ

Have fun and feel free to leave feedback and feature/integrations requests on GitHub issuesโœจ

Articles 10

Article
8

Why we (don't) need export control