Junrui2021
6 models • 1 total models in database
Sort by:
Chatglm3 6b Q4 K M GGUF
Junrui2021/chatglm3-6b-Q4KM-GGUF This model was converted to GGUF format from `THUDM/chatglm3-6b` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
NaNK
llama-cpp
34
1
Qwen2.5-7B-Instruct-Q4_K_M-GGUF
NaNK
llama-cpp
14
0
Qwen2-7B-Instruct-Q4_K_M-GGUF
NaNK
llama-cpp
6
0
bert-based-chinese-openvino
This model was converted to OpenVINO from `Azion/bert-based-chinese` using optimum-intel via the export space.
NaNK
—
3
0
Baichuan2-7B-Chat-Q4_K_M-GGUF
NaNK
llama-cpp
2
0
YOLO_SleepPose
—
0
1