LLM Models
Llama 3.2
LLAMA3_2_3B
Llama 3.2 3B model (BF16 precision).LLAMA3_2_3B_QLORA
Llama 3.2 3B model with QLoRA quantization.LLAMA3_2_3B_SPINQUANT
Llama 3.2 3B model with SpinQuant quantization.LLAMA3_2_1B
Llama 3.2 1B model (BF16 precision).LLAMA3_2_1B_QLORA
Llama 3.2 1B model with QLoRA quantization.LLAMA3_2_1B_SPINQUANT
Llama 3.2 1B model with SpinQuant quantization.Qwen 3
QWEN3_0_6B
Qwen 3 0.6B model (BF16 precision).QWEN3_0_6B_QUANTIZED
Qwen 3 0.6B model with 8-bit quantization.QWEN3_1_7B
Qwen 3 1.7B model (BF16 precision).QWEN3_1_7B_QUANTIZED
Qwen 3 1.7B model with 8-bit quantization.QWEN3_4B
Qwen 3 4B model (BF16 precision).QWEN3_4B_QUANTIZED
Qwen 3 4B model with 8-bit quantization.Hammer 2.1
HAMMER2_1_0_5B
Hammer 2.1 0.5B model (BF16 precision).HAMMER2_1_0_5B_QUANTIZED
Hammer 2.1 0.5B model with 8-bit quantization.HAMMER2_1_1_5B
Hammer 2.1 1.5B model (BF16 precision).HAMMER2_1_1_5B_QUANTIZED
Hammer 2.1 1.5B model with 8-bit quantization.HAMMER2_1_3B
Hammer 2.1 3B model (BF16 precision).HAMMER2_1_3B_QUANTIZED
Hammer 2.1 3B model with 8-bit quantization.SmolLM 2
SMOLLM2_1_135M
SmolLM 2 135M model (BF16 precision).SMOLLM2_1_135M_QUANTIZED
SmolLM 2 135M model with 8-bit quantization.SMOLLM2_1_360M
SmolLM 2 360M model (BF16 precision).SMOLLM2_1_360M_QUANTIZED
SmolLM 2 360M model with 8-bit quantization.SMOLLM2_1_1_7B
SmolLM 2 1.7B model (BF16 precision).SMOLLM2_1_1_7B_QUANTIZED
SmolLM 2 1.7B model with 8-bit quantization.Qwen 2.5
QWEN2_5_0_5B
Qwen 2.5 0.5B model (BF16 precision).QWEN2_5_0_5B_QUANTIZED
Qwen 2.5 0.5B model with 8-bit quantization.QWEN2_5_1_5B
Qwen 2.5 1.5B model (BF16 precision).QWEN2_5_1_5B_QUANTIZED
Qwen 2.5 1.5B model with 8-bit quantization.QWEN2_5_3B
Qwen 2.5 3B model (BF16 precision).QWEN2_5_3B_QUANTIZED
Qwen 2.5 3B model with 8-bit quantization.Phi 4
PHI_4_MINI_4B
Phi 4 Mini 4B model (BF16 precision).PHI_4_MINI_4B_QUANTIZED
Phi 4 Mini 4B model with 8-bit quantization.LFM 2.5
LFM2_5_1_2B_INSTRUCT
LFM 2.5 1.2B Instruct model (FP16 precision).LFM2_5_1_2B_INSTRUCT_QUANTIZED
LFM 2.5 1.2B Instruct model with 8-bit quantization.Classification Models
EFFICIENTNET_V2_S
EfficientNet V2 Small classification model. Platform-optimized (CoreML for iOS, XNNPACK for Android).Object Detection Models
SSDLITE_320_MOBILENET_V3_LARGE
SSDLite 320 MobileNet V3 Large object detection model.RF_DETR_NANO
RF-DETR Nano object detection model.Style Transfer Models
STYLE_TRANSFER_CANDY
Candy style transfer model. Platform-optimized (CoreML for iOS, XNNPACK for Android).STYLE_TRANSFER_MOSAIC
Mosaic style transfer model. Platform-optimized (CoreML for iOS, XNNPACK for Android).STYLE_TRANSFER_RAIN_PRINCESS
Rain Princess style transfer model. Platform-optimized (CoreML for iOS, XNNPACK for Android).STYLE_TRANSFER_UDNIE
Udnie style transfer model. Platform-optimized (CoreML for iOS, XNNPACK for Android).Speech to Text Models
WHISPER_TINY_EN
Whisper Tiny English-only model.WHISPER_TINY_EN_QUANTIZED
Whisper Tiny English-only model with quantization.WHISPER_BASE_EN
Whisper Base English-only model.WHISPER_SMALL_EN
Whisper Small English-only model.WHISPER_TINY
Whisper Tiny multilingual model.WHISPER_BASE
Whisper Base multilingual model.WHISPER_SMALL
Whisper Small multilingual model.Semantic Segmentation Models
DEEPLAB_V3_RESNET50
DeepLab V3 with ResNet50 backbone.DEEPLAB_V3_RESNET101
DeepLab V3 with ResNet101 backbone.DEEPLAB_V3_MOBILENET_V3_LARGE
DeepLab V3 with MobileNet V3 Large backbone.LRASPP_MOBILENET_V3_LARGE
LRASPP with MobileNet V3 Large backbone.FCN_RESNET50
FCN with ResNet50 backbone.FCN_RESNET101
FCN with ResNet101 backbone.Quantized Variants
Quantized versions are also available:DEEPLAB_V3_RESNET50_QUANTIZEDDEEPLAB_V3_RESNET101_QUANTIZEDDEEPLAB_V3_MOBILENET_V3_LARGE_QUANTIZEDLRASPP_MOBILENET_V3_LARGE_QUANTIZEDFCN_RESNET50_QUANTIZEDFCN_RESNET101_QUANTIZED