v000000's picture
Update README.md
18b685c verified
|
raw
history blame
777 Bytes
---
base_model: Monero/Manticore-13b-Chat-Pyg-Guanaco
tags:
- manticore
- llama-cpp
- llama
---
<!DOCTYPE html>
<style>
h1 {
color: #FF0000;
text-decoration: none;
}
</style>
<html lang="en">
<head>
</head>
<body>
<h1>!!! Archive of LLaMa-1-13B Model !!!</h1>
</body>
</html>
# May 27, 2023 - Monero/Manticore-13b-Chat-Pyg-Guanaco
v000000
This model was converted to GGUF format from [`Monero/Manticore-13b-Chat-Pyg-Guanaco`](https://huggingface.co./Monero/Manticore-13b-Chat-Pyg-Guanaco) using llama.cpp
Refer to the [original model card](https://huggingface.co./Monero/Manticore-13b-Chat-Pyg-Guanaco) for more details on the model.
* [Quants in repo:] static Q5_K_M, static Q6_K, static Q8_0
Manticore-13b-Chat-Pyg with the Guanaco 13b qLoRa from TimDettmers applied