DAMO-NLP-SG

53 models • 4 total models in database
Sort by:

VideoLLaMA3-7B

NaNK
videollama3_qwen2
90,200
70

SigLIP-NaViT

videollama3_vision_encoder
33,323
1

VL3-SigLIP-NaViT

videollama3_vision_encoder
28,591
9

VideoLLaMA3-2B

NaNK
videollama3_qwen2
8,149
14

VideoLLaMA2.1-7B-AV

NaNK
videollama2_qwen2
3,955
14

VideoLLaMA2-7B

NaNK
videollama2_mistral
1,354
42

VideoLLaMA2.1-7B-16F

NaNK
videollama2_qwen2
1,268
10

VideoLLaMA3-7B-Image

NaNK
videollama3_qwen2
380
10

VideoLLaMA3-2B-Image

NaNK
videollama3_qwen2
280
8

Zero Shot Classify SSTuning XLM R

Zero-shot text classification (multilingual version) trained with self-supervised tuning Zero-shot text classification model trained with self-supervised tuning (SSTuning). It was introduced in the paper Zero-Shot Text Classification via Self-Supervised Tuning by Chaoqun Liu, Wenxuan Zhang, Guizhen Chen, Xiaobao Wu, Anh Tuan Luu, Chip Hong Chang, Lidong Bing and first released in this repository. The model is tuned with unlabeled data using a first sentence prediction (FSP) learning objective. The FSP task is designed by considering both the nature of the unlabeled corpus and the input/output format of classification tasks. The training and validation sets are constructed from the unlabeled corpus using FSP. During tuning, BERT-like pre-trained masked language models such as RoBERTa and ALBERT are employed as the backbone, and an output layer for classification is added. The learning objective for FSP is to predict the index of the correct label. A cross-entropy loss is used for tuning the model. Model variations There are four versions of models released. The details are: | Model | Backbone | #params | lang | acc | Speed | #Train |------------|-----------|----------|-------|-------|----|-------------| | zero-shot-classify-SSTuning-base | roberta-base | 125M | En | Low | High | 20.48M | | zero-shot-classify-SSTuning-large | roberta-large | 355M | En | Medium | Medium | 5.12M | | zero-shot-classify-SSTuning-ALBERT | albert-xxlarge-v2 | 235M | En | High | Low| 5.12M | | zero-shot-classify-SSTuning-XLM-R | xlm-roberta-base | 278M | Multi | - | - | 20.48M | Please note that zero-shot-classify-SSTuning-XLM-R is trained with 20.48M English samples only. However, it can also be used in other languages as long as xlm-roberta supports. Please check this repository for the performance of each model. Intended uses & limitations The model can be used for zero-shot text classification such as sentiment analysis and topic classification. No further finetuning is needed. How to use You can try the model with the Colab Notebook.

license:mit
141
9

VideoLLaMA2-7B-Base

NaNK
videollama2_mistral
86
6

VideoRefer-VideoLLaMA3-7B

NaNK
videollama3_qwen2
68
11

Qwen2.5-7B-LongPO-128K

NaNK
license:apache-2.0
49
4

VideoRefer-VideoLLaMA3-2B

NaNK
videollama3_qwen2
41
7

VideoLLaMA2-7B-16F

NaNK
videollama2_mistral
35
15

Zero Shot Classify SSTuning Base

Zero-shot text classification (base-sized model) trained with self-supervised tuning Zero-shot text classification model trained with self-supervised tuning (SSTuning). It was introduced in the paper Zero-Shot Text Classification via Self-Supervised Tuning by Chaoqun Liu, Wenxuan Zhang, Guizhen Chen, Xiaobao Wu, Anh Tuan Luu, Chip Hong Chang, Lidong Bing and first released in this repository. The model is tuned with unlabeled data using a learning objective called first sentence prediction (F...

license:mit
33
8

Zero Shot Classify SSTuning ALBERT

Zero-shot text classification (model based on albert-xxlarge-v2) trained with self-supervised tuning Zero-shot text classification model trained with self-supervised tuning (SSTuning). It was introduced in the paper Zero-Shot Text Classification via Self-Supervised Tuning by Chaoqun Liu, Wenxuan Zhang, Guizhen Chen, Xiaobao Wu, Anh Tuan Luu, Chip Hong Chang, Lidong Bing and first released in this repository. Model description The model is tuned with unlabeled data using a learning objective c...

license:mit
25
5

CLEX-Phi-2-32K

NaNK
license:mit
24
10

Zero Shot Classify SSTuning Large

Zero-shot text classification (large-sized model) trained with self-supervised tuning Zero-shot text classification model trained with self-supervised tuning (SSTuning). It was introduced in the paper Zero-Shot Text Classification via Self-Supervised Tuning by Chaoqun Liu, Wenxuan Zhang, Guizhen Chen, Xiaobao Wu, Anh Tuan Luu, Chip Hong Chang, Lidong Bing and first released in this repository. Model description The model is tuned with unlabeled data using a learning objective called first sen...

license:mit
21
2

VideoLLaMA2.1-7B-16F-Base

NaNK
videollama2_qwen2
21
1

VideoRefer-7B

NaNK
license:apache-2.0
20
5

VideoLLaMA2-72B

NaNK
videollama2_qwen2
15
10

Mistral-7B-LongPO-256K-EXP

NaNK
license:apache-2.0
11
0

Mistral-7B-LongPO-128K

NaNK
license:apache-2.0
9
2

VideoLLaMA2-7B-16F-Base

NaNK
videollama2_mistral
8
2

VideoRefer-7B-stage2.5

NaNK
license:apache-2.0
4
2

CLEX-7B-Chat-16K

NaNK
llama
3
3

mt-llama-7b-delta

NaNK
llama
3
2

PMR-xxlarge

license:mit
3
2

NER-PMR-large

license:mit
3
2

CLEX-7B-16K

NaNK
llama
2
3

PMR-base

license:mit
2
2

PMR-large

license:mit
2
2

mPMR-large

license:mit
2
2

EQA-PMR-large

license:mit
2
2

VideoLLaMA2-72B-Base

NaNK
license:apache-2.0
2
1

CLEX-Mixtral-8x7B-32K

NaNK
license:mit
1
3

mPMR-base

license:mit
1
2

roberta-time_identification

1
2

VideoLLaMA2-8x7B-Base

NaNK
videollama2_mixtral
1
2

VideoRefer-7B-stage2

NaNK
license:apache-2.0
1
1

siglip2-so400m-patch14-384-navit

videollama3_vision_encoder
1
0

Video-LLaMA-Series

license:bsd-3-clause
0
47

Video-LLaMA-2-7B-Finetuned

NaNK
0
13

Video-LLaMA-2-13B-Finetuned

NaNK
0
13

Video-LLaMA-2-7B-Pretrained

NaNK
0
8

DiGIT

license:mit
0
4

CLEX-LLaMA-2-7B-64K

NaNK
llama
0
3

VideoLLaMA2-8x7B

NaNK
videollama2_mixtral
0
3

Video-LLaMA-2-13B-Pretrained

NaNK
0
1

rememo-large

license:apache-2.0
0
1

CLEX-Mixtral-8x7B-Chat-32K

NaNK
license:mit
0
1

LiT-B-32_CC12M

license:apache-2.0
0
1