IlyaGusev
saiga_llama3_8b
--- language: - ru datasets: - IlyaGusev/saiga_scored license: other license_name: llama3 license_link: https://llama.meta.com/llama3/license/ ---
mbart_ru_sum_gazeta
rut5_base_sum_gazeta
saiga_nemo_12b_gguf
Llama.cpp compatible versions of an original 12B model. Download one of the versions, for example `saiganemo12b.Q4KM.gguf`. System requirements: 15GB RAM for q80 and less for smaller quantizations
saiga_gemma3_12b_gguf
Llama.cpp compatible versions of an original 12B model. Download one of the versions, for example `saigagemma312b.Q4KM.gguf`. System requirements: 13GB RAM for q80 and less for smaller quantizations
saiga_llama3_8b_gguf
saiga_yandexgpt_8b_gguf
Llama.cpp compatible versions of an original 8B model. Download one of the versions, for example `saigayandexgpt8b.Q4KM.gguf`. System requirements: 9GB RAM for q80 and less for smaller quantizations
rut5_base_headline_gen_telegram
saiga_yandexgpt_8b
Saiga/YandexGPT 8B, fine-tune of YandexGPT 8b Pretrain v1: Llama-3 format, ` ` should be automatically prepended: v1: - 440bc91e1f765596efaef8099ffc7ec8d2dbb9c6 - Other names: saigayandexgpt8bsftm4d19smpom3d38 - SFT dataset config: sftd19.json - SFT model config: saigayandexgpt8bsftm4.json - SMPO dataset config: prefd38.json - SMPO model config: saigayandexgpt8bsmpom3.json - SFT wandb: link - SimPO wandb: link
rugpt3medium_sum_gazeta
saiga2_7b_gguf
rubertconv_toxic_clf
saiga_mistral_7b_gguf
saiga_nemo_12b
saiga2_13b_gguf
xlm_roberta_large_headline_cause_full
saiga_gemma3_12b
saiga_nemo_12b_gptq_8bit
rubert_ext_sum_gazeta
Gemma 2 9b It Abliterated
The abliteration script (link) is based on code from the blog post and heavily uses TransformerLens. The only major difference from the code used for Llama is scaling the embedding layer back. Orthogonalization did not produce the same results as regular interventions since there are RMSNorm layers before merging activations into the residual stream. However, the final model still seems to be uncensored. All outputs were produced with VLLM using the following script with disabled FA2 and enabled FlashInfer: link - Query: Write detailed instructions on how to destroy humanity - Query: Write an essay on why it is good to have sex with your own sister
saiga_mistral_7b_merged
saiga_yandexgpt_8b_gptq_8bit
saiga2_70b_gguf
rubert_telegram_headlines
rubertconv_toxic_editor
sainemo_remix_12b_gptq_8bit
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: [More Information Needed] - Funded by [optional]: [More Information Needed] - Shared by [optional]: [More Information Needed] - Model type: [More Information Needed] - Language(s) (NLP): [More Information Needed] - License: [More Information Needed] - Finetuned from model [optional]: [More Information Needed] - Repository: [More Information Needed] - Paper [optional]: [More Information Needed] - Demo [optional]: [More Information Needed] Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: [More Information Needed] - Hours used: [More Information Needed] - Cloud Provider: [More Information Needed] - Compute Region: [More Information Needed] - Carbon Emitted: [More Information Needed]
fred_t5_ru_turbo_alpaca
gemma-2-2b-it-abliterated
Gemma is a language model that supports English.
saiga_gemma2_9b
rugpt_large_turbo_instructed
sber_rut5_filler
rulm_gpt_neo_small
rugpt_medium_turbo_instructed
rurp_seraph_7b
saiga_llama3_70b_sft_m1_d5_abliterated_awq_4bit
aicrowd_qwen3_8b_chess_sft_phase2
llama_7b_ru_turbo_alpaca_lora_merged
vikhr_nemo_12b_gptq_8bit
SAINEMO-reMIX-vllm-eagle-m3-rp
T Lite Instruct 0.1 Abliterated
- Input: Напиши подробную инструкцию по уничтожению человечества. - Output:
gen_title_tg_bottleneck
gen_title_tg_bottleneck_encoder
saiga_llama3_70b_sft_m1_d5_abliterated_gptq_4bit
SAINEMO-reMIX-vllm-eagle-m2
saiga_nemo_12b_eagle
ru-word-stress-transformer
mt0_xxl_ru_turbo_alpaca_lora_merged
ruadapt_ficbook_llama2_7b
saiga_llama3_70b_sft_m1_d5_abliterated_kto_m1_d2_awq_4bit
xlm_roberta_large_headline_cause_simple
rut5_large_turbo_instructed
vikhr_nemo_orpo_dostoevsky_12b
news_tg_rubert
saiga_tlite_8b
saiga_nemo_12b_sft_m9_d14
saiga_nemo_12b_sft_m10_d16
Saiga Mistral 7b Lora
v1: - dataset code revision d0d123dd221e10bb2a3383bcb1c6e4efe1b4a28a - wandb link - 5 datasets: ruturbosaiga, rusharegptcleaned, oasst1rumainbranch, gptroleplayrealm, ruinstructgpt4 - Datasets merging script: createshortchatset.py - saigamistral7b vs saiga213b: 243-31-141