Model Garden
Featured Models
Dolly 2.0 Inference
Dolly 2.0 – The World’s First, Truly Open Instruction-Tuned LLM on IPUs – Inference
OpenAssistant Pythia 12B Inference
OpenAssistant Pythia 12B is an open-source and commercially usable chat-based assistant model trained on the OpenAssistant Conversations Dataset (OASST1)
Whisper Inference
Speech Transcription on IPUs using OpenAI's Whisper - Inference
GPT-J 6B Fine-tuning
Text entailment on IPU using GPT-J 6B on PyTorch using fine-tuning.
Flan-T5-Large/XL Inference
Flan-T5-Large/XL inference on IPUs with Hugging Face
Stable Diffusion Text-to-Image Inference
The popular latent diffusion model for generative AI with support for text-to-image on IPUs using Hugging Face Optimum.
YOLOv4 Inference
YOLOv4 - You Only Look Once - a convolutional neural network model that performs object detection tasks on IPUs using PyTorch.
BERT-Large Fine-tuning
HuggingFace Optimum implementation for fine-tuning a BERT-Large transformer model.
Library
Dolly 2.0 Inference
Dolly 2.0 – The World’s First, Truly Open Instruction-Tuned LLM on IPUs – Inference
OpenAssistant Pythia 12B Inference
OpenAssistant Pythia 12B is an open-source and commercially usable chat-based assistant model trained on the OpenAssistant Conversations Dataset (OASST1)
Whisper Inference
Speech Transcription on IPUs using OpenAI's Whisper - Inference
Llama 2 Inference
Run Meta’s latest Open Source Large Language Model Inference on IPUs
Stable Diffusion 2 Text-to-Image Inference
The popular latent diffusion model for generative AI with support for text-to-image on IPUs using Hugging Face Optimum.
Stable Diffusion Text-to-Image Inference
The popular latent diffusion model for generative AI with support for text-to-image on IPUs using Hugging Face Optimum.
Stable Diffusion Image-to-Image Inference
The popular latent diffusion model for generative AI with support for image-to-image on IPUs using Hugging Face Optimum.
Stable Diffusion Inpainting Inference
The popular latent diffusion model for generative AI with support for inpainting on IPUs using Hugging Face Optimum.
GPT-J 6B Fine-tuning
Text entailment on IPU using GPT-J 6B on PyTorch using fine-tuning.
GPT-J 6B Inference
Text generation on IPU using GPT-J 6B on PyTorch for inference.
RGCN Training
Training a GNN to do Fraud Detection using Relational Graph Convolution Network (RGCN) on IPUs with PyG (PyTorch Geometric)
GPT-3 Fine-tuning
GPT-3 (Generative Pretrained Transformer 3) is a state-of-the-art language processing AI model developed by OpenAI.
GPT-3 Inference
GPT-3 (Generative Pretrained Transformer 3) is a state-of-the-art language processing AI model developed by OpenAI.
GPT2-Large Training
GPT2-L training in PyTorch leveraging the Hugging Face Transformers library.
GPT2-Large Inference
GPT2-L inference in PyTorch leveraging the Hugging Face Transformers library.
GPT2-Medium Training
GPT2-M training in PyTorch leveraging the Hugging Face Transformers library.
GPT2-Medium Fine-tuning
HuggingFace Optimum implementation for fine-tuning a GPT2-Medium transformer model.
GPT2-Medium Inference
GPT2-M inference in PyTorch leveraging the Hugging Face Transformers library.
GPT2-Small Training
GPT2-S training in PyTorch leveraging the Hugging Face Transformers library.
GPT2-Small Fine-tuning
HuggingFace Optimum implementation for fine-tuning a GPT2-Small transformer model.
GPT2-Small Inference
GPT2-S inference in PyTorch leveraging the Hugging Face Transformers library.
Flan-T5-Large/XL Inference
Flan-T5-Large/XL inference on IPUs with Hugging Face
T5-Small Fine-Tuning
Summarization on IPU using T5 Small with Hugging Face Optimum - Fine-Tuning
MT5-Small Fine-Tuning
Machine Translation on IPUs using MT5-Small with Hugging Face - Fine-tuning
MT5-Large Inference
Zero-Shot Text Classification on IPUs using MT5-Large with Hugging Face - Inference
GPS++ Training
A hybrid GNN/Transformer for training Molecular Property Prediction using IPUs on the PCQM4Mv2 dataset. Winner of the Open Graph Benchmark Large-Scale Challenge.
GPS++ Inference
A hybrid GNN/Transformer for Molecular Property Prediction inference using IPUs trained on the PCQM4Mv2 dataset. Winner of the Open Graph Benchmark Large-Scale Challenge.
Distributed KGE - TransE (256) Training
Knowledge graph embedding (KGE) for link-prediction training on IPUs using Poplar with the WikiKG90Mv2 dataset. Winner of the Open Graph Benchmark Large-Scale Challenge.
Distributed KGE - TransE (256) Inference
Knowledge graph embedding (KGE) for link-prediction inference on IPUs using Poplar with the WikiKG90Mv2 dataset. Winner of the Open Graph Benchmark Large-Scale Challenge.
Distributed KGE - TransE (256) Training
Knowledge graph embedding (KGE) for link-prediction training on IPUs using PyTorch with the WikiKG90Mv2 dataset. Winner of the Open Graph Benchmark Large-Scale Challenge.
BERT-Large Training
BERT-Large (Bidirectional Encoder Representations from Transformers) using PyTorch for NLP training on IPUs.
BERT-Large Training
BERT-Large (Bidirectional Encoder Representations from Transformers) using TensorFlow 1 for NLP training on IPUs.
BERT-Large Inference
BERT-Large (Bidirectional Encoder Representations from Transformers) for NLP inference on IPUs with TensorFlow 1.
BERT-Large Training
BERT-Large (Bidirectional Encoder Representations from Transformers) using TensorFlow 2 for NLP training on IPUs.
BERT-Large Training
BERT-Large (Bidirectional Encoder Representations from Transformers) using PopART for NLP training on IPUs.
BERT-Large Inference
BERT-Large (Bidirectional Encoder Representations from Transformers) using PopART for NLP inference on IPUs.
BERT-Large Fine-tuning
HuggingFace Optimum implementation for fine-tuning a BERT-Large transformer model.
BERT-Large Pretraining
HuggingFace Optimum implementation for pre-training a BERT-Large transformer model.
DistilBERT Training
DistilBERT is a small, fast, cheap and light Transformer model trained by distilling BERT base using Hugging Face Optimum on IPUs.
BERT-Base Training
BERT-Base (Bidirectional Encoder Representations from Transformers) using PyTorch for NLP training on IPUs.
BERT-Base Training
BERT-Base (Bidirectional Encoder Representations from Transformers) using TensorFlow 2 for NLP training on IPUs.
BERT-Base Training
BERT-Base (Bidirectional Encoder Representations from Transformers) using TensorFlow 1 for NLP training on IPUs.
BERT-Base Training
BERT-Base (Bidirectional Encoder Representations from Transformers) using PopART for NLP training on IPUs.
BERT-Base Inference
BERT-Base (Bidirectional Encoder Representations from Transformers) using PopART for NLP inference on IPUs.
BERT-Base Training
BERT-Base pre-training and SQuAD fine-tuning using Baidu's PaddlePaddle framework on IPUs.
BERT-Base Pretraining
HuggingFace Optimum implementation for pretraining a BERT-Base transformer model using bert-based-uncased datasets.
BERT-Base Fine-tuning
HuggingFace Optimum implementation for fine-tuning a BERT-Base transformer model using bert-base-uncased on the squad dataset.
RoBERTa-Large Training
HuggingFace Optimum implementation for training RoBERTa-Large - a transformer model for sequence classification, token classification or question answering.
RoBERTa-Base Fine-tuning
HuggingFace Optimum implementation for fine-tuning RoBERTa-Base on the squad dataset for text generation and comprehension tasks
RoBERTa-Base Fine-tuning
HuggingFace Optimum implementation for fine-tuning RoBERTa-Base on the squad_v2 dataset for text generation and comprehension tasks
LXMERT Fine-tuning
HuggingFace Optimum implementation for fine-tuning LXMERT on the gqa-lxmert dataset for learning vision-and-language cross-modality representations.
DeBERTa Training
HuggingFace Optimum implementation for training DeBERTa - a transformer models that improves BERT and RoBERTa models using disentangled attention and enhanced mask decoder.
LXMERT Fine-tuning
HuggingFace Optimum implementation for fine-tuning LXMERT on the vqa-lxmert dataset for learning vision-and-language cross-modality representations.
DeBERTa Inference
SQuAD and MNLI on IPUs using DeBERTa with Hugging Face - Inference
HuBERT Training
HuggingFace Optimum implementation for training HuBERT (Hidden-Unit BERT) for self-supervised speech representation learning approach.
BART Training
HuggingFace Optimum implementation for training BART - a transformer model for text generation and comprehension tasks
GroupBERT Training
GroupBERT - an enhanced transformer architecture with efficient grouped structures in TensorFlow 1.
PackedBERT Training
New BERT packing algorithm that removes padding for more efficient training in PyTorch.
PackedBERT Training
New BERT packing algorithm that removes padding for more efficient training in PopART.
PackedBERT Fine-tuning
New BERT packing algorithm that removes padding for more efficient fine-tuning in Hugging Face.
PackedBERT Inference
New BERT packing algorithm that removes padding for more efficient inference in Hugging Face.
Conformer-Medium Training
A variant of the conformer model based on WeNet (not ESPnet) using PyTorch which uses a hybrid CTC/attention architecture with transformer or conformer as an encoder.
CLIP Training
CLIP (Contrastive Language-Image Pre-Training) - a neural network trained on a variety of (image, text) pairs using PyTorch.
ViT (Vision Transformer) Fine-tuning
ViT (Vision Transformer) fine-tuning in PyTorch using Hugging Face transformers.
ViT (Vision Transformer) Pretraining
ViT (Vision Transformer) pretraining in PyTorch using Hugging Face transformers.
ViT (Vision Transformer) Fine-tuning
HuggingFace Optimum implementation for fine-tuning a ViT (vision transformer) model.
DINO Training
Self-supervised Vision Transformer model for training in PyTorch.
YOLOv3 Training
YOLOv3 - You Only Look Once - a convolutional neural network model that performs object detection tasks on IPUs using TensorFlow 1.