clip-vit-base-patch32_lego-minifigure
35
license:mit
by
Armaggheddon
Image Model
OTHER
32B params
New
35 downloads
Early-stage
Edge AI:
Mobile
Laptop
Server
72GB+ RAM
Mobile
Laptop
Server
Quick Summary
Model Card for clip-vit-base-patch32lego-minifigure This model is a finetuned version of the `openai/clip-vit-base-patch32` CLIP (Contrastive Language-Image Pr...
Device Compatibility
Mobile
4-6GB RAM
Laptop
16GB RAM
Server
GPU
Minimum Recommended
30GB+ RAM
Code Examples
pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import AutoModelForZeroShotImageClassification, AutoProcessor
model = AutoModelForZeroShotImageClassification.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
processor = AutoProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
from transformers import pipeline
model = "Armaggheddon/clip-vit-base-patch32_lego-minifigure"
clip_classifier = pipeline("zero-shot-image-classification", model=model)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)pythontransformers
import torch
from transformers import CLIPModel
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
model_fp16 = model.to(torch.float16)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)Use casespythontransformers
import torch
from transformers import CLIPTokenizerFast, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
tokenizer = CLIPTokenizerFast.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure")
text = ["a photo of a lego minifigure"]
tokens = tokenizer(text, return_tensors="pt", padding=True).to(device)
outputs = model.get_text_features(**tokens)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)pythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
image = Image.open("path_to_image.jpg")
inputs = processor(images=image, return_tensors="pt").to(device)
outputs = model.get_image_features(**inputs)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Zero-shot image classificationpythontransformers
import torch
from PIL import Image
from transformers import CLIPProcessor, CLIPModel
from datasets import load_dataset
device = "cuda" if torch.cuda.is_available() else "cpu"
model = CLIPModel.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
processor = CLIPProcessor.from_pretrained("Armaggheddon/clip-vit-base-patch32_lego-minifigure", device_map="auto").to(device)
dataset = load_dataset("Armaggheddon/lego_minifigure_captions", split="test")
captions = [
"a photo of a lego minifigure with a t-shirt with a pen holder",
"a photo of a lego minifigure with green pants",
"a photo of a lego minifigure with a red cap",
]
image = dataset[0]["image"]
inputs = processor(text=captions, images=image, return_tensors="pt", padding=True).to(device)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probabilities = logits_per_image.softmax(dim=1)
max_prob_idx = torch.argmax(logits_per_image, dim=1)Deploy This Model
Production-ready deployment in minutes
Together.ai
Instant API access to this model
Production-ready inference API. Start free, scale to millions.
Try Free APIReplicate
One-click model deployment
Run models in the cloud with simple API. No DevOps required.
Deploy NowDisclosure: We may earn a commission from these partners. This helps keep LLMYourWay free.