[Part 2] General discussion.

#2
by Lewdiculous - opened

[Part 2]

@Nitral-AI @ABX-AI

Continue here, the other discussion got too long to scroll comfortably.

Lewdiculous pinned discussion

0_2.webp

You're getting closer. Tho I think LocalBasedMan really prefers regular anime Sumika haha.

Hahah, yeah this was just for you to thank for new thread :3
I would never force 3D over 2D anime enjoyers!! I'm super happy just using the style for my merges.
https://huggingface.co./ABX-AI/Laymonade-7B - I think this one came out alright! now to go for quants, then 9b ^^

BTW @Lewdiculous I did do a lot of 2D anime at the start of my ai imagery journey but quickly got out of it as it was too good to even build on

https://www.deviantart.com/abroxis/gallery/87178213/crimson-petals-of-the-azalea-dynasty
https://www.deviantart.com/abroxis/gallery/87165704/historical-figures-of-the-future
https://www.deviantart.com/abroxis/gallery/87176370/the-celestial-sisterhood-mothers-of-all-life

It's just automatically perfect, pretty much. MJ's niji algorithm is insane. I went into 3D because I felt it had far more space for stylizing and actually coming up with new styles so to speak

Wait shouldnt the config be like this? @ABX-AI

{
  "_name_or_path": "example/example",
  "architectures": [
    "MistralForCausalLM"
  ],
  "attention_dropout": 0.0,
  "bos_token_id": 1,
  "eos_token_id": 2,
  "hidden_act": "silu",
  "hidden_size": 4096,
  "initializer_range": 0.02,
  "intermediate_size": 14336,
  "max_position_embeddings": 32768,
  "model_type": "mistral",
  "num_attention_heads": 32,
  "num_hidden_layers": 32,
  "num_key_value_heads": 8,
  "rms_norm_eps": 1e-05,
  "rope_theta": 1000000.0,
  "sliding_window": null,
  "tie_word_embeddings": false,
  "torch_dtype": "bfloat16",
  "transformers_version": "4.38.2",
  "use_cache": true,
  "vocab_size": 32000
}

im seeing this in laymonade:

{
  "_name_or_path": "KatyTheCutie/LemonadeRP-4.5.3",
  "architectures": [
    "MistralForCausalLM"
  ],
  "attention_dropout": 0.0,
  "bos_token_id": 1,
  "eos_token_id": 2,
  "hidden_act": "silu",
  "hidden_size": 4096,
  "initializer_range": 0.02,
  "intermediate_size": 14336,
  "max_position_embeddings": 32768,
  "model_type": "mistral",
  "num_attention_heads": 32,
  "num_hidden_layers": 32,
  "num_key_value_heads": 8,
  "rms_norm_eps": 1e-05,
  "rope_theta": 10000.0,
  "sliding_window": 4096,
  "tie_word_embeddings": false,
  "torch_dtype": "bfloat16",
  "transformers_version": "4.39.1",
  "use_cache": true,
  "vocab_size": 32000
}

Is that for mistral-v0.2- layla?

I gave up on trying to use it as base.

Just copy the config from mistralai/Mistral-7B-Instruct-v0.2 or alpindale/Mistral-7B-v0.2-hf

@Virt-io no thats the official config from instruct 0.2, but it should still work on layla 0.2 i would assume.

This is the config for layla 0.2 that they provide which has the added chatml tokens into the vocab.

{
  "_name_or_path": "/home/layla/src/text-generation-webui/models/mistral-7b-v0.2",
  "architectures": [
    "MistralForCausalLM"
  ],
  "attention_dropout": 0.0,
  "bos_token_id": 1,
  "eos_token_id": 2,
  "hidden_act": "silu",
  "hidden_size": 4096,
  "initializer_range": 0.02,
  "intermediate_size": 14336,
  "max_position_embeddings": 32768,
  "model_type": "mistral",
  "num_attention_heads": 32,
  "num_hidden_layers": 32,
  "num_key_value_heads": 8,
  "pad_token_id": 0,
  "rms_norm_eps": 1e-05,
  "rope_theta": 1000000.0,
  "sliding_window": 32768,
  "tie_word_embeddings": false,
  "torch_dtype": "bfloat16",
  "transformers_version": "4.38.2",
  "use_cache": false,
  "vocab_size": 32002
}

This is also part of the problem:

image.png

From what I imagined, the "double thumbs up" (normal) is human rating? Then you have the same for reward repo, so if you arrange them by just Double Thumbs up, you will see a different list.
There is also this "by repo" double likes, which I don't understand, but a couple of my merges scored well there: (is it counting the actual likes on HF for the whole repository?)

image.png

Then, this is just "double tumbs up" ratio:

image.png

Overall, for role-play specifically, I think synthetic scoring is not accurate, the reward model doesn't know what good kink is, unless it's claude-3 or smth like that... which it isn't ^^

Oh btw, it re-slipped back into first. I was going to work on merges today, but now i may just play games instead.

Overall, for role-play specifically, I think synthetic scoring is not accurate, the reward model doesn't know what good kink is, unless it's claude-3 or smth like that... which it isn't ^^

Which is why i appreciate the battles here are determined by the actual end users (elo) and not by the reward model.

Overall, for role-play specifically, I think synthetic scoring is not accurate, the reward model doesn't know what good kink is, unless it's claude-3 or smth like that... which it isn't ^^

Which is why i appreciate the battles here are determined by the actual end users (elo) and not by the reward model.

Ooh ok, so the ELO is mosly user-based? That's pretty good, then, I thought it weighs in the bot battles a lot more. See how well-spread out the battles are? I'm curious how that can be achieved, those users must be wanking off furiously 24/7 hahaah

I actually hadn't looked into how massive that platform is... yeah, it makes sense now ^^ They could easily do this with users all the way

@ABX-AI

those users must be wanking off furiously 24/7 hahaah

KEKW

I'll say the Chai users are special

Do you know how chai does there elo?

They mentioned a few days ago in the announcement for the new elo system that there will be a write-up for it coming soon.
So i guess we're stuck waiting for now

I was contacted by Chai-Will, and did ask about it, and the answer was Elo is a human score. I also gave feedback that it would be nice to have an info panel with more explanations on what each ranking column does, and I understood the feedback will be taken in consideration, so that was definitely great to hear :)

Sign up or log in to comment