Awesome RLHF Collection A curated collection of datasets, models, Spaces, and papers on Reinforcement Learning from Human Feedback (RLHF). β’ 11 items β’ Updated Oct 2, 2023 β’ 7
Handbook v0.1 models and datasets Collection Models and datasets for v0.1 of the alignment handbook β’ 6 items β’ Updated Nov 10, 2023 β’ 24
INaturalist-2021 Fine-tunes Collection Fine-tune experiments for various `timm` models on the INaturalist 2021 Challenge dataset (https://github.com/visipedia/inat_comp/tree/master/2021) β’ 5 items β’ Updated Oct 25, 2023 β’ 6
Latent Consistency Models LoRAs Collection Latent Consistency Models for Stable Diffusion - LoRAs and full fine-tuned weights β’ 4 items β’ Updated Nov 10, 2023 β’ 102
AudioPaLM: A Large Language Model That Can Speak and Listen Paper β’ 2306.12925 β’ Published Jun 22, 2023 β’ 53
GLIGEN: Open-Set Grounded Text-to-Image Generation Paper β’ 2301.07093 β’ Published Jan 17, 2023 β’ 3