Vocalis-Q8_0.gguf
8
8.0B
Q8
base_model:meta-llama/Meta-Llama-3-8B-Instruct
by
lex-au
Other
OTHER
8B params
New
8 downloads
Early-stage
Edge AI:
Mobile
Laptop
Server
18GB+ RAM
Mobile
Laptop
Server
Quick Summary
🧠Model Card: LLaMA 3 8B Instruct – Conversational Roleplay Enhanced Model Name: `lex-au/vocalis-Q80.
Device Compatibility
Mobile
4-6GB RAM
Laptop
16GB RAM
Server
GPU
Minimum Recommended
8GB+ RAM
Code Examples
text
@misc{llama3-8b-2024,
author = {Meta AI},
title = {LLaMA 3 8B Instruct},
year = {2024},
publisher = {Meta},
howpublished = {\url{https://ai.meta.com/llama/}}
}
@misc{vocalis-roleplay-2025,
author = {Lex-au},
title = {Vocalis-Q8_0: Fine-Tuned LLaMA 3 for Conversational Roleplay},
note = {Low-latency roleplay model optimised for use in the Vocalis webapp},
year = {2025},
howpublished = {\url{https://github.com/Lex-au/Vocalis}}
}text
@misc{llama3-8b-2024,
author = {Meta AI},
title = {LLaMA 3 8B Instruct},
year = {2024},
publisher = {Meta},
howpublished = {\url{https://ai.meta.com/llama/}}
}
@misc{vocalis-roleplay-2025,
author = {Lex-au},
title = {Vocalis-Q8_0: Fine-Tuned LLaMA 3 for Conversational Roleplay},
note = {Low-latency roleplay model optimised for use in the Vocalis webapp},
year = {2025},
howpublished = {\url{https://github.com/Lex-au/Vocalis}}
}text
@misc{llama3-8b-2024,
author = {Meta AI},
title = {LLaMA 3 8B Instruct},
year = {2024},
publisher = {Meta},
howpublished = {\url{https://ai.meta.com/llama/}}
}
@misc{vocalis-roleplay-2025,
author = {Lex-au},
title = {Vocalis-Q8_0: Fine-Tuned LLaMA 3 for Conversational Roleplay},
note = {Low-latency roleplay model optimised for use in the Vocalis webapp},
year = {2025},
howpublished = {\url{https://github.com/Lex-au/Vocalis}}
}text
@misc{llama3-8b-2024,
author = {Meta AI},
title = {LLaMA 3 8B Instruct},
year = {2024},
publisher = {Meta},
howpublished = {\url{https://ai.meta.com/llama/}}
}
@misc{vocalis-roleplay-2025,
author = {Lex-au},
title = {Vocalis-Q8_0: Fine-Tuned LLaMA 3 for Conversational Roleplay},
note = {Low-latency roleplay model optimised for use in the Vocalis webapp},
year = {2025},
howpublished = {\url{https://github.com/Lex-au/Vocalis}}
}text
@misc{llama3-8b-2024,
author = {Meta AI},
title = {LLaMA 3 8B Instruct},
year = {2024},
publisher = {Meta},
howpublished = {\url{https://ai.meta.com/llama/}}
}
@misc{vocalis-roleplay-2025,
author = {Lex-au},
title = {Vocalis-Q8_0: Fine-Tuned LLaMA 3 for Conversational Roleplay},
note = {Low-latency roleplay model optimised for use in the Vocalis webapp},
year = {2025},
howpublished = {\url{https://github.com/Lex-au/Vocalis}}
}text
@misc{llama3-8b-2024,
author = {Meta AI},
title = {LLaMA 3 8B Instruct},
year = {2024},
publisher = {Meta},
howpublished = {\url{https://ai.meta.com/llama/}}
}
@misc{vocalis-roleplay-2025,
author = {Lex-au},
title = {Vocalis-Q8_0: Fine-Tuned LLaMA 3 for Conversational Roleplay},
note = {Low-latency roleplay model optimised for use in the Vocalis webapp},
year = {2025},
howpublished = {\url{https://github.com/Lex-au/Vocalis}}
}text
@misc{llama3-8b-2024,
author = {Meta AI},
title = {LLaMA 3 8B Instruct},
year = {2024},
publisher = {Meta},
howpublished = {\url{https://ai.meta.com/llama/}}
}
@misc{vocalis-roleplay-2025,
author = {Lex-au},
title = {Vocalis-Q8_0: Fine-Tuned LLaMA 3 for Conversational Roleplay},
note = {Low-latency roleplay model optimised for use in the Vocalis webapp},
year = {2025},
howpublished = {\url{https://github.com/Lex-au/Vocalis}}
}text
@misc{llama3-8b-2024,
author = {Meta AI},
title = {LLaMA 3 8B Instruct},
year = {2024},
publisher = {Meta},
howpublished = {\url{https://ai.meta.com/llama/}}
}
@misc{vocalis-roleplay-2025,
author = {Lex-au},
title = {Vocalis-Q8_0: Fine-Tuned LLaMA 3 for Conversational Roleplay},
note = {Low-latency roleplay model optimised for use in the Vocalis webapp},
year = {2025},
howpublished = {\url{https://github.com/Lex-au/Vocalis}}
}text
@misc{llama3-8b-2024,
author = {Meta AI},
title = {LLaMA 3 8B Instruct},
year = {2024},
publisher = {Meta},
howpublished = {\url{https://ai.meta.com/llama/}}
}
@misc{vocalis-roleplay-2025,
author = {Lex-au},
title = {Vocalis-Q8_0: Fine-Tuned LLaMA 3 for Conversational Roleplay},
note = {Low-latency roleplay model optimised for use in the Vocalis webapp},
year = {2025},
howpublished = {\url{https://github.com/Lex-au/Vocalis}}
}text
@misc{llama3-8b-2024,
author = {Meta AI},
title = {LLaMA 3 8B Instruct},
year = {2024},
publisher = {Meta},
howpublished = {\url{https://ai.meta.com/llama/}}
}
@misc{vocalis-roleplay-2025,
author = {Lex-au},
title = {Vocalis-Q8_0: Fine-Tuned LLaMA 3 for Conversational Roleplay},
note = {Low-latency roleplay model optimised for use in the Vocalis webapp},
year = {2025},
howpublished = {\url{https://github.com/Lex-au/Vocalis}}
}Deploy This Model
Production-ready deployment in minutes
Together.ai
Instant API access to this model
Production-ready inference API. Start free, scale to millions.
Try Free APIReplicate
One-click model deployment
Run models in the cloud with simple API. No DevOps required.
Deploy NowDisclosure: We may earn a commission from these partners. This helps keep LLMYourWay free.