chatglm2-6b-32k

191
293
6.0B
3 languages
by
zai-org
Code Model
OTHER
6B params
New
191 downloads
Early-stage
Edge AI:
Mobile
Laptop
Server
14GB+ RAM
Mobile
Laptop
Server
Quick Summary

💻 Github Repo • 🐦 Twitter • 📃 [GLM@ACL 22] [GitHub] • 📃 [GLM-130B@ICLR 23] [GitHub] - 我们优化了KV Cache的存储方式,减少了显存碎片的产生。基于优化后的代码,模型可以在约20G显存的情况下处理32K长度的上下文(FP/...

Device Compatibility

Mobile
4-6GB RAM
Laptop
16GB RAM
Server
GPU
Minimum Recommended
6GB+ RAM

Deploy This Model

Production-ready deployment in minutes

Together.ai

Instant API access to this model

Fastest API

Production-ready inference API. Start free, scale to millions.

Try Free API

Replicate

One-click model deployment

Easiest Setup

Run models in the cloud with simple API. No DevOps required.

Deploy Now

Disclosure: We may earn a commission from these partners. This helps keep LLMYourWay free.