opus-mt-synthetic-en-so

177
1
3 languages
license:cc-by-4.0
by
Helsinki-NLP
Other
OTHER
New
177 downloads
Early-stage
Edge AI:
Mobile
Laptop
Server
Unknown
Mobile
Laptop
Server
Quick Summary

AI model with specialized capabilities.

Code Examples

Usagepythontransformers
from transformers import MarianMTModel, MarianTokenizer

# Load the pre-trained model and tokenizer
model_name = "Helsinki-NLP/opus-mt-synthetic-en-so"
model = MarianMTModel.from_pretrained(model_name)
tokenizer = MarianTokenizer.from_pretrained(model_name)

# Example source text (English)
source_texts = ["Hello, how are you?", "Good morning!", "What is your name?"]

# Tokenize the input texts
inputs = tokenizer(source_texts, return_tensors="pt", padding=True, truncation=True)

# Generate translations
translated_ids = model.generate(inputs["input_ids"])

# Decode the generated tokens to get the translated text
translated_texts = tokenizer.batch_decode(translated_ids, skip_special_tokens=True)

# Print the translations
for src, tgt in zip(source_texts, translated_texts):
    print(f"Source: {src} => Translated: {tgt}")
Usagepythontransformers
from transformers import MarianMTModel, MarianTokenizer

# Load the pre-trained model and tokenizer
model_name = "Helsinki-NLP/opus-mt-synthetic-en-so"
model = MarianMTModel.from_pretrained(model_name)
tokenizer = MarianTokenizer.from_pretrained(model_name)

# Example source text (English)
source_texts = ["Hello, how are you?", "Good morning!", "What is your name?"]

# Tokenize the input texts
inputs = tokenizer(source_texts, return_tensors="pt", padding=True, truncation=True)

# Generate translations
translated_ids = model.generate(inputs["input_ids"])

# Decode the generated tokens to get the translated text
translated_texts = tokenizer.batch_decode(translated_ids, skip_special_tokens=True)

# Print the translations
for src, tgt in zip(source_texts, translated_texts):
    print(f"Source: {src} => Translated: {tgt}")
Citation Informationbibtex
@article{degibert2025scaling,
  title={Scaling Low-Resource MT via Synthetic Data Generation with LLMs},
  author={de Gibert, Ona and Attieh, Joseph and Vahtola, Teemu and Aulamo, Mikko and Li, Zihao and V{\'a}zquez, Ra{\'u}l and Hu, Tiancheng and Tiedemann, J{\"o}rg},
  journal={arXiv preprint arXiv:2505.14423},
  year={2025}
}
Citation Informationbibtex
@article{degibert2025scaling,
  title={Scaling Low-Resource MT via Synthetic Data Generation with LLMs},
  author={de Gibert, Ona and Attieh, Joseph and Vahtola, Teemu and Aulamo, Mikko and Li, Zihao and V{\'a}zquez, Ra{\'u}l and Hu, Tiancheng and Tiedemann, J{\"o}rg},
  journal={arXiv preprint arXiv:2505.14423},
  year={2025}
}

Deploy This Model

Production-ready deployment in minutes

Together.ai

Instant API access to this model

Fastest API

Production-ready inference API. Start free, scale to millions.

Try Free API

Replicate

One-click model deployment

Easiest Setup

Run models in the cloud with simple API. No DevOps required.

Deploy Now

Disclosure: We may earn a commission from these partners. This helps keep LLMYourWay free.