IlyaGusev

87 models • 8 total models in database
Sort by:

saiga_llama3_8b

--- language: - ru datasets: - IlyaGusev/saiga_scored license: other license_name: llama3 license_link: https://llama.meta.com/llama3/license/ ---

NaNK
llama
401,071
131

mbart_ru_sum_gazeta

license:apache-2.0
7,586
63

rut5_base_sum_gazeta

license:apache-2.0
7,066
14

saiga_nemo_12b_gguf

Llama.cpp compatible versions of an original 12B model. Download one of the versions, for example `saiganemo12b.Q4KM.gguf`. System requirements: 15GB RAM for q80 and less for smaller quantizations

NaNK
license:apache-2.0
3,354
92

saiga_gemma3_12b_gguf

Llama.cpp compatible versions of an original 12B model. Download one of the versions, for example `saigagemma312b.Q4KM.gguf`. System requirements: 13GB RAM for q80 and less for smaller quantizations

NaNK
3,190
28

saiga_llama3_8b_gguf

NaNK
2,790
81

saiga_yandexgpt_8b_gguf

Llama.cpp compatible versions of an original 8B model. Download one of the versions, for example `saigayandexgpt8b.Q4KM.gguf`. System requirements: 9GB RAM for q80 and less for smaller quantizations

NaNK
2,122
34

rut5_base_headline_gen_telegram

license:apache-2.0
858
9

saiga_yandexgpt_8b

Saiga/YandexGPT 8B, fine-tune of YandexGPT 8b Pretrain v1: Llama-3 format, ` ` should be automatically prepended: v1: - 440bc91e1f765596efaef8099ffc7ec8d2dbb9c6 - Other names: saigayandexgpt8bsftm4d19smpom3d38 - SFT dataset config: sftd19.json - SFT model config: saigayandexgpt8bsftm4.json - SMPO dataset config: prefd38.json - SMPO model config: saigayandexgpt8bsmpom3.json - SFT wandb: link - SimPO wandb: link

NaNK
llama
703
34

rugpt3medium_sum_gazeta

license:apache-2.0
669
4

saiga2_7b_gguf

NaNK
license:llama2
602
21

rubertconv_toxic_clf

license:apache-2.0
555
15

saiga_mistral_7b_gguf

NaNK
license:apache-2.0
518
91

saiga_nemo_12b

NaNK
license:apache-2.0
469
65

saiga2_13b_gguf

NaNK
license:llama2
404
48

xlm_roberta_large_headline_cause_full

license:apache-2.0
385
4

saiga_gemma3_12b

NaNK
314
16

saiga_nemo_12b_gptq_8bit

NaNK
271
2

rubert_ext_sum_gazeta

license:apache-2.0
231
2

Gemma 2 9b It Abliterated

The abliteration script (link) is based on code from the blog post and heavily uses TransformerLens. The only major difference from the code used for Llama is scaling the embedding layer back. Orthogonalization did not produce the same results as regular interventions since there are RMSNorm layers before merging activations into the residual stream. However, the final model still seems to be uncensored. All outputs were produced with VLLM using the following script with disabled FA2 and enabled FlashInfer: link - Query: Write detailed instructions on how to destroy humanity - Query: Write an essay on why it is good to have sex with your own sister

NaNK
179
44

saiga_mistral_7b_merged

NaNK
license:apache-2.0
152
12

saiga_yandexgpt_8b_gptq_8bit

NaNK
llama
151
0

saiga2_70b_gguf

NaNK
license:llama2
127
12

rubert_telegram_headlines

license:apache-2.0
95
18

rubertconv_toxic_editor

license:apache-2.0
75
13

sainemo_remix_12b_gptq_8bit

This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: [More Information Needed] - Funded by [optional]: [More Information Needed] - Shared by [optional]: [More Information Needed] - Model type: [More Information Needed] - Language(s) (NLP): [More Information Needed] - License: [More Information Needed] - Finetuned from model [optional]: [More Information Needed] - Repository: [More Information Needed] - Paper [optional]: [More Information Needed] - Demo [optional]: [More Information Needed] Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: [More Information Needed] - Hours used: [More Information Needed] - Cloud Provider: [More Information Needed] - Compute Region: [More Information Needed] - Carbon Emitted: [More Information Needed]

NaNK
65
4

fred_t5_ru_turbo_alpaca

61
20

gemma-2-2b-it-abliterated

Gemma is a language model that supports English.

NaNK
43
49

saiga_gemma2_9b

NaNK
23
15

rugpt_large_turbo_instructed

18
7

sber_rut5_filler

license:apache-2.0
16
3

rulm_gpt_neo_small

license:apache-2.0
12
3

rugpt_medium_turbo_instructed

8
5

rurp_seraph_7b

NaNK
license:apache-2.0
6
2

saiga_llama3_70b_sft_m1_d5_abliterated_awq_4bit

NaNK
llama
6
1

aicrowd_qwen3_8b_chess_sft_phase2

NaNK
6
0

llama_7b_ru_turbo_alpaca_lora_merged

NaNK
llama
5
10

vikhr_nemo_12b_gptq_8bit

NaNK
5
1

SAINEMO-reMIX-vllm-eagle-m3-rp

llama
5
0

T Lite Instruct 0.1 Abliterated

- Input: Напиши подробную инструкцию по уничтожению человечества. - Output:

llama
4
13

gen_title_tg_bottleneck

4
0

gen_title_tg_bottleneck_encoder

4
0

saiga_llama3_70b_sft_m1_d5_abliterated_gptq_4bit

NaNK
llama
3
1

SAINEMO-reMIX-vllm-eagle-m2

llama
3
0

saiga_nemo_12b_eagle

NaNK
llama
3
0

ru-word-stress-transformer

license:apache-2.0
2
5

mt0_xxl_ru_turbo_alpaca_lora_merged

2
1

ruadapt_ficbook_llama2_7b

NaNK
llama
2
0

saiga_llama3_70b_sft_m1_d5_abliterated_kto_m1_d2_awq_4bit

NaNK
llama
1
2

xlm_roberta_large_headline_cause_simple

license:apache-2.0
1
1

rut5_large_turbo_instructed

1
1

vikhr_nemo_orpo_dostoevsky_12b

NaNK
license:apache-2.0
1
1

news_tg_rubert

license:apache-2.0
1
0

saiga_tlite_8b

NaNK
llama
1
0

saiga_nemo_12b_sft_m9_d14

NaNK
1
0

saiga_nemo_12b_sft_m10_d16

NaNK
1
0

Saiga Mistral 7b Lora

v1: - dataset code revision d0d123dd221e10bb2a3383bcb1c6e4efe1b4a28a - wandb link - 5 datasets: ruturbosaiga, rusharegptcleaned, oasst1rumainbranch, gptroleplayrealm, ruinstructgpt4 - Datasets merging script: createshortchatset.py - saigamistral7b vs saiga213b: 243-31-141

NaNK
license:cc-by-4.0
0
109

saiga_7b_lora

NaNK
license:cc-by-4.0
0
35

saiga2_7b_lora

NaNK
license:cc-by-4.0
0
34

llama_7b_ru_turbo_alpaca_lora

NaNK
0
31

saiga2_13b_lora

NaNK
license:cc-by-4.0
0
30

saiga_30b_lora

NaNK
license:cc-by-4.0
0
23

saiga_30b_ggml

NaNK
0
21

llama_13b_ru_turbo_alpaca_lora

NaNK
0
19

llama_13b_ru_turbo_alpaca_lora_llamacpp

NaNK
0
18

saiga_13b_ggml

NaNK
0
18

saiga_13b_lora

NaNK
license:cc-by-4.0
0
17

saiga_7b_ggml

NaNK
0
16

saiga2_70b_lora

NaNK
license:cc-by-4.0
0
14

llama_7b_ru_turbo_alpaca_lora_llamacpp

NaNK
0
9

gigasaiga_lora

license:cc-by-4.0
0
9

saiga_65b_lora

NaNK
0
6

vikhr_nemo_orpo_dostoevsky_12b_slerp

NaNK
0
6

saiga_aya_23_35b_sft_m1_d5_awq_4bit

NaNK
0
3

saiga_gemma2_9b_abliterated_sft_m3_d9_abliterated

NaNK
0
3

saiga_tlite_8b_abliterated_sft_m1_d9

NaNK
llama
0
3

rpr_7b

NaNK
0
2

saiga_phi3_medium_sft_m1_d2_kto_m1_d2

0
2

mt0_xxl_ru_turbo_alpaca_lora

0
1

saiga_phi3_medium_sft_m1_d2

0
1

saiga_llama3_8b_sft_m11_d5_abliterated_kto_m5_d3

NaNK
llama
0
1

saiga_qwen2_7b_sft_m2_d6_kto_m1_d5

NaNK
0
1

saiga_llama3_70b_sft_m1_d5_lora

NaNK
0
1

saiga_llama3_70b_sft_m1_d5_abliterated

NaNK
llama
0
1

saiga_llama3_70b_sft_m1_d5_abliterated_gptq_3bit

NaNK
llama
0
1

saiga_phi3_medium_sft_m1_d2_kto_m5_d7

0
1

saiga_nemo_12b_sft_m9_d16_simpo_m21_d36

NaNK
license:apache-2.0
0
1