E5
|
Language
|
Deploy E5, a text embedding model series.
|
Colab
Model card
|
Instant ID
|
Language, Vision
|
Deploy Instant ID, an identity preserving, text-to-image generation model.
|
Colab
Model card
|
Llama 3
|
Language
|
Explore and build with Meta's Llama 3 models (8B, 70B) on Vertex AI.
|
Model Card
|
Gemma
|
Language
|
Open weight models (2B, 7B) that are built from the same research and technology used to create Google's Gemini models.
|
Model Card
|
CodeGemma
|
Language
|
Open weight models (2B, 7B) designed for code generation and code completion that are built from the same research and technology used to create Google's Gemini models.
|
Model Card
|
PaliGemma
|
Language
|
Open weight 3B model designed for image captioning tasks and visual question and answering tasks that's built from the same research and technology used to create Google's Gemini models.
|
Model Card
|
Vicuna v1.5
|
Language
|
Deploy Vicuna v1.5 series models, which are foundation models fine-tuned from LLama2 for text generation.
|
Model Card
|
NLLB
|
Language
|
Deploy nllb series models for multi-language translation.
|
Model Card
Colab
|
Mistral-7B
|
Language
|
Deploy Mistral-7B, a foundational model for text generation.
|
Model Card
Colab
|
BioGPT
|
Language
|
Deploy BioGPT, a text generative model for the biomedical domain.
|
Model Card
Colab
|
BiomedCLIP
|
Language, Vision
|
Deploy BiomedCLIP, a multimodal foundation model for the biomedical domain.
|
Model Card
Colab
|
ImageBind
|
Language, Vision,
Audio
|
Deploy ImageBind, a foundational model for multimodal embedding.
|
Model Card
Colab
|
DITO
|
Language, Vision
|
Finetune and deploy DITO, a multimodal foundation model for open vocabulary object detection tasks.
|
Model Card
Colab
|
OWL-ViT v2
|
Language, Vision
|
Deploy OWL-ViT v2, a multimodal foundation model for open vocabulary object detection tasks.
|
Model Card
Colab
|
FaceStylizer (Mediapipe)
|
Vision
|
A generative pipeline to transform human face images to a new style.
|
Model Card
Colab
|
Llama 2
|
Language
|
Finetune and deploy Meta's Llama 2 foundation models (7B, 13B, 70B) on Vertex AI.
|
Model Card
|
Code Llama
|
Language
|
Deploy Meta's Code Llama foundation models (7B, 13B, 34B) on Vertex AI.
|
Model Card
|
Falcon-instruct
|
Language
|
Finetune and deploy Falcon-instruct models (7B, 40B) by using PEFT.
|
Colab
Model Card
|
OpenLLaMA
|
Language
|
Finetune and deploy OpenLLaMA models (3B, 7B, 13B) by using PEFT.
|
Colab
Model Card
|
T5-FLAN
|
Language
|
Finetune and deploy T5-FLAN (base, small, large).
|
Model Card
(fine-tuning pipeline included)
|
BERT
|
Language
|
Finetune and deploy BERT by using PEFT.
|
Colab
Model Card
|
BART-large-cnn
|
Language
|
Deploy BART, a transformer encoder-encoder (seq2seq) model with a bidirectional (BERT-like) encoder and an autoregressive (GPT-like) decoder.
|
Colab
Model Card
|
RoBERTa-large
|
Language
|
Finetune and deploy RoBERTa-large by using PEFT.
|
Colab
Model Card
|
XLM-RoBERTa-large
|
Language
|
Finetune and deploy XLM-RoBERTa-large (a multilingual version of RoBERTa) by using PEFT.
|
Colab
Model Card
|
Dolly-v2-7b
|
Language
|
Deploy Dolly-v2-7b, an instruction-following large language model with 6.9 billion parameters.
|
Colab
Model Card
|
Stable Diffusion XL v1.0
|
Language, Vision
|
Deploy Stable Diffusion XL v1.0, which supports text-to-image generation.
|
Colab
Model card
|
Stable Diffusion XL Lightning
|
Language, Vision
|
Deploy Stable Diffusion XL Lightning, a text-to-image generation model.
|
Colab
Model card
|
Stable Diffusion v2.1
|
Language, Vision
|
Finetune and deploy Stable Diffusion v2.1 (supports text-to-image generation) by using Dreambooth.
|
Colab
Model card
|
Stable Diffusion 4x upscaler
|
Language, Vision
|
Deploy Stable Diffusion 4x upscaler, which supports text conditioned image superresolution.
|
Colab
Model card
|
InstructPix2Pix
|
Language, Vision
|
Deploy InstructPix2Pix, which supports image editing by using a text prompt.
|
Colab
Model Card
|
Stable Diffusion Inpainting
|
Language, Vision
|
Finetune and deploy Stable Diffusion Inpainting, which supports inpainting a masked image by using a text prompt.
|
Colab
Model Card
|
SAM
|
Language, Vision
|
Deploy Segment Anything, which supports zero-shot image segmentation.
|
Colab
Model Card
|
Text-to-video (ModelScope)
|
Language, Vision
|
Deploy ModelScope text-to-video, which supports text-to-video generation.
|
Colab
Model Card
|
Text-to-video zero-shot
|
Language, Vision
|
Deploy Stable Diffusion text-to-video generators, which support zero-shot text-to-video generation.
|
Colab
Model Card
|
Pic2Word Composed Image Retrieval
|
Language, Vision
|
Deploy Pic2Word, which supports multi-modal composed image retrieval.
|
Colab
Model Card
|
BLIP2
|
Language, Vision
|
Deploy BLIP2, which supports image captioning and visual-question-answering.
|
Colab
Model Card
|
Open-CLIP
|
Language, Vision
|
Finetune and deploy the Open-CLIP, which supports zero-shot classification.
|
Colab
Model card
|
F-VLM
|
Language, Vision
|
Deploy F-VLM, which supports open vocabulary image object detection.
|
Colab
Model Card
|
tfhub/EfficientNetV2
|
Vision
|
Finetune and deploy the Tensorflow Vision implementation of the EfficientNetV2 image classification model.
|
Colab
Model Card
|
EfficientNetV2 (TIMM)
|
Vision
|
Finetune and deploy the PyTorch implementation of the EfficientNetV2 image classification model.
|
Colab
Model Card
|
Proprietary/EfficientNetV2
|
Vision
|
Finetune and deploy the Google proprietary checkpoint of the EfficientNetV2 image classification model.
|
Colab
Model Card
|
EfficientNetLite (MediaPipe)
|
Vision
|
Finetune EfficientNetLite image classification model through MediaPipe model maker.
|
Colab
Model Card
|
tfvision/vit
|
Vision
|
Finetune and deploy the TensorFlow Vision implementation of the ViT image classification model.
|
Colab
Model Card
|
ViT (TIMM)
|
Vision
|
Finetune and deploy the PyTorch implementation of the ViT image classification model.
|
Colab
Model Card
|
Proprietary/ViT
|
Vision
|
Finetune and deploy the Google proprietary checkpoint of the ViT image classification model.
|
Colab
Model Card
|
Proprietary/MaxViT
|
Vision
|
Finetune and deploy the Google proprietary checkpoint of the MaxViT hybrid (CNN + ViT) image classification model.
|
Colab
Model Card
|
ViT (JAX)
|
Vision
|
Finetune and deploy the JAX implementation of the ViT image classification model.
|
Colab
Model Card
|
tfvision/SpineNet
|
Vision
|
Finetune and deploy the Tensorflow Vision implementation of the SpineNet object detection model.
|
Colab
Model Card
|
Proprietary/Spinenet
|
Vision
|
Finetune and deploy the Google proprietary checkpoint of the SpineNet object detection model.
|
Colab
Model Card
|
tfvision/YOLO
|
Vision
|
Finetune and deploy the TensorFlow Vision implementation of the YOLO one-stage object detection model.
|
Colab
Model Card
|
Proprietary/YOLO
|
Vision
|
Finetune and deploy the Google proprietary checkpoint of the YOLO one-stage object detection model.
|
Colab
Model Card
|
YOLOv8 (Keras)
|
Vision
|
Finetune and deploy the Keras implementation of the YOLOv8 model for object detection.
|
Colab
Model Card
|
tfvision/YOLOv7
|
Vision
|
Finetune and deploy YOLOv7 model for object detection.
|
Colab
Model Card
|
ByteTrack Video Object Tracking
|
Vision
|
Run batch prediction for video object tracking by using ByteTrack tracker.
|
Colab
Model Card
|
ResNeSt (TIMM)
|
Vision
|
Finetune and deploy the PyTorch implementation of the ResNeSt image classification model.
|
Colab
Model Card
|
ConvNeXt (TIMM)
|
Vision
|
Finetune and deploy ConvNeXt, a pure convolutional model for image classification inspired by the design of Vision Transformers.
|
Colab
Model Card
|
CspNet (TIMM)
|
Vision
|
Finetune and deploy the CSPNet (Cross Stage Partial Network) image classification model.
|
Colab
Model Card
|
Inception (TIMM)
|
Vision
|
Finetune and deploy the Inception image classification model.
|
Colab
Model Card
|
DeepLabv3+ (with checkpoint)
|
Vision
|
Finetune and deploy the DeepLab-v3 Plus model for semantic image segmentation.
|
Colab
Model Card
|
Faster R-CNN (Detectron2)
|
Vision
|
Finetune and deploy the Detectron2 implementation of the Faster R-CNN model for image object detection.
|
Colab
Model Card
|
RetinaNet (Detectron2)
|
Vision
|
Finetune and deploy the Detectron2 implementation of the RetinaNet model for image object detection.
|
Colab
Model Card
|
Mask R-CNN (Detectron2)
|
Vision
|
Finetune and deploy the Detectron2 implementation of the Mask R-CNN model for image object detection and segmentation.
|
Colab
Model Card
|
ControlNet
|
Vision
|
Finetune and deploy the ControlNet text-to-image generation model.
|
Colab
Model Card
|
MobileNet (TIMM)
|
Vision
|
Finetune and deploy the PyTorch implementation of the MobileNet image classification model.
|
Colab
Model Card
|
MobileNetV2 (MediaPipe) Image Classification
|
Vision
|
Finetune the MobileNetV2 image classification model by using MediaPipe model maker.
|
Colab
Model Card
|
MobileNetV2 (MediaPipe) Object Detection
|
Vision
|
Finetune the MobileNetV2 object detection model by using MediaPipe model maker.
|
Colab
Model Card
|
MobileNet-MultiHW-AVG (MediaPipe)
|
Vision
|
Finetune the MobileNet-MultiHW-AVG object detection model by using MediaPipe model maker.
|
Colab
Model Card
|
DeiT
|
Vision
|
Finetune and deploy the DeiT (Data-efficient Image Transformers) model for image classification.
|
Colab
Model Card
|
BEiT
|
Vision
|
Finetune and deploy the BEiT (Bidirectional Encoder representation from Image Transformers) model for image classification.
|
Colab
Model Card
|
Hand Gesture Recognition (MediaPipe)
|
Vision
|
Finetune and deploy on-device the Hand Gesture Recognition models by using MediaPipe.
|
Colab
Model Card
|
Average Word Embedding Classifier (MediaPipe)
|
Vision
|
Finetune and deploy on-device the Average Word Embedding Classifier models by using MediaPipe.
|
Colab
Model Card
|
MobileBERT Classifier (MediaPipe)
|
Vision
|
Finetune and deploy on-device the MobileBERT Classifier models by using MediaPipe.
|
Colab
Model Card
|
MoViNet Video Clip Classification
|
Video
|
Finetune and deploy MoViNet video clip classification models.
|
Colab
Model Card
|
MoViNet Video Action Recognition
|
Video
|
Finetune and deploy MoViNet models for action recognition inference.
|
Colab
Model Card
|
Stable Diffusion XL LCM
|
Vision
|
Deploy this model which uses the Latent Consistency Model (LCM)
to enhance text-to-image
generation in Latent Diffusion Models by enabling faster and
high-quality image creation with fewer steps.
|
Colab
Model Card
|
LLaVA 1.5
|
Vision, Language
|
Deploy LLaVA 1.5 models.
|
Colab
Model Card
|
Pytorch-ZipNeRF
|
Vision, Video
|
Train the Pytorch-ZipNeRF model which is a state-of-the-art
implementation of the ZipNeRF algorithm in the Pytorch framework,
designed for efficient and accurate 3D reconstruction
from 2D images.
|
Colab
Model Card
|
WizardLM
|
Language
|
Deploy WizardLM which is a large language model (LLM)
developed by Microsoft, fine-tuned on complex instructions by
adapting the Evol-Instruct method.
|
Colab
Model Card
|
WizardCoder
|
Language
|
Deploy WizardCoder which is a large language model (LLM)
developed by Microsoft, fine-tuned on complex instructions
by adapting the Evol-Instruct method to the domain of code.
|
Colab
Model Card
|
Mixtral 8x7B
|
Language
|
Deploy the Mixtral 8x7B model which is a Mixture of Experts (MoE)
large language model (LLM) developed by Mistral AI. It is a
decoder-only model with 46.7B parameters and was reported to match
or outperform LLaMA 2 70B and GPT 3.5 on many benchmarks.
|
Colab
Model Card
|
Llama 2 (Quantized)
|
Language
|
Fine-tune & deploy a quantized version of Meta's Llama 2
models.
|
Colab
Model Card
|
LaMa (Large Mask Inpainting)
|
Vision
|
Deploy LaMa which uses fast Fourier convolutions (FFCs),
a high receptive field perceptual loss and large training masks
allows for resolution-robust image inpainting.
|
Colab
Model Card
|
AutoGluon
|
Tabular
|
With AutoGluon you can train and deploy high-accuracy machine
learning and deep learning models for tabular data.
|
Colab
Model Card
|