README / README.md
sumuks's picture
Update README.md
2187e14 verified
|
raw
history blame
2.55 kB
metadata
title: README
emoji: ⚙️
colorFrom: blue
colorTo: purple
sdk: static
pinned: false
short_description: Unsupervised Human Preference Learning

LINK TO PAPER

⚙️ Preference Agents

Preference Agents is an organization focused on developing and releasing small language models ("preference agents") that enable efficient personalization of larger language models (LLMs). Our agents learn user preferences and generate natural language rules that guide LLMs to produce tailored content, without requiring extensive fine-tuning of the larger models.

🎯 Our Approach

We train small, locally deployable language models to act as "steering wheels" for larger, pre-trained LLMs. These agents learn user preferences from small, personalized datasets and encode these preferences into concise natural language rules. These rules are then provided as context to the larger LLM, guiding its output towards the desired personalized style and content.

📦 Resources

Datasets

We release three datasets for research on personalized language modeling:

  • Enron-42k: A curated subset of the Enron email corpus, focused on original content creation. It contains approximately 40,240 emails from 191 unique senders.
  • The New Yorker: A curated subset of the All The News 2.0 corpus, containing 4000 articles from the New Yorker.
  • LAMP 3U Subset: A subset of the LAMP 3U Amazon product reviews dataset, containing 22,500 reviews from 15 users.

Both datasets are licensed under the Creative Commons Attribution-NonCommercial 4.0 International (CC BY-NC 4.0) license.

🚀 How to Use Our Resources

Datasets

from datasets import load_dataset

enron_dataset = load_dataset("preference-agents/Enron-42k", split="train")
lamp_dataset = load_dataset("preference-agents/LAMP-3U-Subset", split="train")

📊 Evaluation

Our approach was evaluated using GPT-4o and human evaluations, demonstrating significant improvements over baselines like zero-shot generation, few-shot learning, and naive fine-tuning.

📜 Citation

If you use our resources in your research or applications, please cite our paper:

pending

🙏 Acknowledgements

We thank Meta AI for providing the Llama-3 models, Google AI for access to Gemini 1.5 Pro, and Anthropic for access to Claude 3.5 Sonnet. We also acknowledge the creators of the Enron email corpus and the LAMP 3U dataset for making their valuable resources available to the research community.