# DeepSparkInference **Repository Path**: guo-dida/deepsparkinference ## Basic Information - **Project Name**: DeepSparkInference - **Description**: DeepSparkInference推理模型示例库甄选了216个推理大小模型示例,小模型涵盖计算机视觉,自然语言处理,语音识别等领域;大模型涉及vLLM,TGI等多种框架,并将持续扩展更多、更新的模型示例。 - **Primary Language**: Python - **License**: Apache-2.0 - **Default Branch**: master - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 33 - **Created**: 2025-08-08 - **Last Updated**: 2025-08-08 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README [](README_en.md) [](README.md) # DeepSparkInference
Homepage

DeepSparkInference ModelZoo, as a core project of the DeepSpark open-source community, was officially open-sourced in March 2024. The first release selected 48 inference model examples, covering fields such as computer vision, natural language processing, and speech recognition. More AI domains will be gradually expanded in the future. The models in DeepSparkInference provide inference examples and guidance documents for running on inference engines IGIE or ixRT self-developed by Iluvatar CoreX. Some models provide evaluation results based on the self-developed GPGPU Zhikai 100. IGIE (Iluvatar GPU Inference Engine) is a high-performance, highly gene, and end-to-end AI inference engine developed based on the TVM framework. It supports multi-framework model, quantization, graph optimization, multi-operator library support, multi-backend support, and automatic operator tuning, providing an easy-to-deploy, high-throughput, and low-latency complete solution for inference scenarios. ixRT (Iluvatar CoreX RunTime) is a high-performance inference engine independently developed by Iluvatar CoreX, focusing on maximizing the performance of Iluvatar CoreX's GPGPU and achieving high-performance inference for models in various fields. ixRT supports features such as dynamic shape inference, plugins, and INT8/FP16 inference. DeepSparkInference will be updated quarterly, and model categories will be gradually enriched, with large model inference to be expanded in the future. ## ModelZoo ### LLM (Large Language Model) | Model | vLLM | TRT-LLM | TGI | IXUCA SDK | |-------------------------------|--------------------------------------------------------|---------------------------------------|------------------------------------|-----------| | Baichuan2-7B | [✅](models/nlp/llm/baichuan2-7b/vllm) | | | 4.3.0 | | ChatGLM-3-6B | [✅](models/nlp/llm/chatglm3-6b/vllm) | | | 4.3.0 | | ChatGLM-3-6B-32K | [✅](models/nlp/llm/chatglm3-6b-32k/vllm) | | | 4.3.0 | | DeepSeek-R1-Distill-Llama-8B | [✅](models/nlp/llm/deepseek-r1-distill-llama-8b/vllm) | | | 4.3.0 | | DeepSeek-R1-Distill-Llama-70B | [✅](models/nlp/llm/deepseek-r1-distill-llama-70b/vllm) | | | 4.3.0 | | DeepSeek-R1-Distill-Qwen-1.5B | [✅](models/nlp/llm/deepseek-r1-distill-qwen-1.5b/vllm) | | | 4.3.0 | | DeepSeek-R1-Distill-Qwen-7B | [✅](models/nlp/llm/deepseek-r1-distill-qwen-7b/vllm) | | | 4.3.0 | | DeepSeek-R1-Distill-Qwen-14B | [✅](models/nlp/llm/deepseek-r1-distill-qwen-14b/vllm) | | | 4.3.0 | | DeepSeek-R1-Distill-Qwen-32B | [✅](models/nlp/llm/deepseek-r1-distill-qwen-32b/vllm) | | | 4.3.0 | | Llama2-7B | [✅](models/nlp/llm/llama2-7b/vllm) | [✅](models/nlp/llm/llama2-7b/trtllm) | | 4.3.0 | | Llama2-13B | | [✅](models/nlp/llm/llama2-13b/trtllm) | | 4.3.0 | | Llama2-70B | | [✅](models/nlp/llm/llama2-70b/trtllm) | | 4.3.0 | | Llama3-70B | [✅](models/nlp/llm/llama3-70b/vllm) | | | 4.3.0 | | Qwen-7B | [✅](models/nlp/llm/qwen-7b/vllm) | | | 4.3.0 | | Qwen1.5-7B | [✅](models/nlp/llm/qwen1.5-7b/vllm) | | [✅](models/nlp/llm/qwen1.5-7b/tgi) | 4.3.0 | | Qwen1.5-14B | [✅](models/nlp/llm/qwen1.5-14b/vllm) | | | 4.3.0 | | Qwen1.5-32B Chat | [✅](models/nlp/llm/qwen1.5-32b/vllm) | | | 4.3.0 | | Qwen1.5-72B | [✅](models/nlp/llm/qwen1.5-72b/vllm) | | | 4.3.0 | | Qwen2-7B Instruct | [✅](models/nlp/llm/qwen2-7b/vllm) | | | 4.3.0 | | Qwen2-72B Instruct | [✅](models/nlp/llm/qwen2-72b/vllm) | | | 4.3.0 | | StableLM2-1.6B | [✅](models/nlp/llm/stablelm/vllm) | | | 4.3.0 | ### Computer Vision #### Classification | Model | Prec. | IGIE | ixRT | IXUCA SDK | |------------------------|-------|--------------------------------------------------------|-----------------------------------------------------------|-----------| | AlexNet | FP16 | [✅](models/cv/classification/alexnet/igie) | [✅](models/cv/classification/alexnet/ixrt) | 4.3.0 | | | INT8 | [✅](models/cv/classification/alexnet/igie) | [✅](models/cv/classification/alexnet/ixrt) | 4.3.0 | | CLIP | FP16 | [✅](models/cv/classification/clip/igie) | [✅](models/cv/classification/clip/ixrt) | 4.3.0 | | Conformer-B | FP16 | [✅](models/cv/classification/conformer_base/igie) | | 4.3.0 | | ConvNeXt-Base | FP16 | [✅](models/cv/classification/convnext_base/igie) | [✅](models/cv/classification/convnext_base/ixrt) | 4.3.0 | | ConvNext-S | FP16 | [✅](models/cv/classification/convnext_s/igie) | | 4.3.0 | | ConvNeXt-Small | FP16 | [✅](models/cv/classification/convnext_small/igie) | [✅](models/cv/classification/convnext_small/ixrt) | 4.3.0 | | ConvNeXt-Tiny | FP16 | [✅](models/cv/classification/convnext_tiny/igie) | | 4.3.0 | | CSPDarkNet53 | FP16 | [✅](models/cv/classification/cspdarknet53/igie) | [✅](models/cv/classification/cspdarknet53/ixrt) | 4.3.0 | | | INT8 | | [✅](models/cv/classification/cspdarknet53/ixrt) | 4.3.0 | | CSPResNet50 | FP16 | [✅](models/cv/classification/cspresnet50/igie) | [✅](models/cv/classification/cspresnet50/ixrt) | 4.3.0 | | | INT8 | | [✅](models/cv/classification/cspresnet50/ixrt) | 4.3.0 | | CSPResNeXt50 | FP16 | [✅](models/cv/classification/cspresnext50/igie) | [✅](models/cv/classification/cspresnext50/ixrt) | 4.3.0 | | DeiT-tiny | FP16 | [✅](models/cv/classification/deit_tiny/igie) | [✅](models/cv/classification/deit_tiny/ixrt) | 4.3.0 | | DenseNet121 | FP16 | [✅](models/cv/classification/densenet121/igie) | [✅](models/cv/classification/densenet121/ixrt) | 4.3.0 | | DenseNet161 | FP16 | [✅](models/cv/classification/densenet161/igie) | [✅](models/cv/classification/densenet161/ixrt) | 4.3.0 | | DenseNet169 | FP16 | [✅](models/cv/classification/densenet169/igie) | [✅](models/cv/classification/densenet169/ixrt) | 4.3.0 | | DenseNet201 | FP16 | [✅](models/cv/classification/densenet201/igie) | [✅](models/cv/classification/densenet201/ixrt) | 4.3.0 | | EfficientNet-B0 | FP16 | [✅](models/cv/classification/efficientnet_b0/igie) | [✅](models/cv/classification/efficientnet_b0/ixrt) | 4.3.0 | | | INT8 | | [✅](models/cv/classification/efficientnet_b0/ixrt) | 4.3.0 | | EfficientNet-B1 | FP16 | [✅](models/cv/classification/efficientnet_b1/igie) | [✅](models/cv/classification/efficientnet_b1/ixrt) | 4.3.0 | | | INT8 | | [✅](models/cv/classification/efficientnet_b1/ixrt) | 4.3.0 | | EfficientNet-B2 | FP16 | [✅](models/cv/classification/efficientnet_b2/igie) | [✅](models/cv/classification/efficientnet_b2/ixrt) | 4.3.0 | | EfficientNet-B3 | FP16 | [✅](models/cv/classification/efficientnet_b3/igie) | [✅](models/cv/classification/efficientnet_b3/ixrt) | 4.3.0 | | EfficientNet-B4 | FP16 | [✅](models/cv/classification/efficientnet_b4/igie) | [✅](models/cv/classification/efficientnet_b4/ixrt) | 4.3.0 | | EfficientNet-B5 | FP16 | [✅](models/cv/classification/efficientnet_b5/igie) | [✅](models/cv/classification/efficientnet_b5/ixrt) | 4.3.0 | | EfficientNet-B6 | FP16 | [✅](models/cv/classification/efficientnet_b6/igie) | | 4.3.0 | | EfficientNetV2 | FP16 | [✅](models/cv/classification/efficientnet_v2/igie) | [✅](models/cv/classification/efficientnet_v2/ixrt) | 4.3.0 | | | INT8 | | [✅](models/cv/classification/efficientnet_v2/ixrt) | 4.3.0 | | EfficientNetv2_rw_t | FP16 | [✅](models/cv/classification/efficientnetv2_rw_t/igie) | [✅](models/cv/classification/efficientnetv2_rw_t/ixrt) | 4.3.0 | | EfficientNetv2_s | FP16 | [✅](models/cv/classification/efficientnet_v2_s/igie) | [✅](models/cv/classification/efficientnet_v2_s/ixrt) | 4.3.0 | | GoogLeNet | FP16 | [✅](models/cv/classification/googlenet/igie) | [✅](models/cv/classification/googlenet/ixrt) | 4.3.0 | | | INT8 | [✅](models/cv/classification/googlenet/igie) | [✅](models/cv/classification/googlenet/ixrt) | 4.3.0 | | HRNet-W18 | FP16 | [✅](models/cv/classification/hrnet_w18/igie) | [✅](models/cv/classification/hrnet_w18/ixrt) | 4.3.0 | | | INT8 | | [✅](models/cv/classification/hrnet_w18/ixrt) | 4.3.0 | | InceptionV3 | FP16 | [✅](models/cv/classification/inception_v3/igie) | [✅](models/cv/classification/inception_v3/ixrt) | 4.3.0 | | | INT8 | [✅](models/cv/classification/inception_v3/igie) | [✅](models/cv/classification/inception_v3/ixrt) | 4.3.0 | | Inception-ResNet-V2 | FP16 | | [✅](models/cv/classification/inception_resnet_v2/ixrt) | 4.3.0 | | | INT8 | | [✅](models/cv/classification/inception_resnet_v2/ixrt) | 4.3.0 | | Mixer_B | FP16 | [✅](models/cv/classification/mlp_mixer_base/igie) | | 4.3.0 | | MNASNet0_5 | FP16 | [✅](models/cv/classification/mnasnet0_5/igie) | | 4.3.0 | | MNASNet0_75 | FP16 | [✅](models/cv/classification/mnasnet0_75/igie) | | 4.3.0 | | MNASNet1_0 | FP16 | [✅](models/cv/classification/mnasnet1_0/igie) | | 4.3.0 | | MNASNet1_3 | FP16 | [✅](models/cv/classification/mnasnet1_3/igie) | | 4.3.0 | | MobileNetV2 | FP16 | [✅](models/cv/classification/mobilenet_v2/igie) | [✅](models/cv/classification/mobilenet_v2/ixrt) | 4.3.0 | | | INT8 | [✅](models/cv/classification/mobilenet_v2/igie) | [✅](models/cv/classification/mobilenet_v2/ixrt) | 4.3.0 | | MobileNetV3_Large | FP16 | [✅](models/cv/classification/mobilenet_v3_large/igie) | | 4.3.0 | | MobileNetV3_Small | FP16 | [✅](models/cv/classification/mobilenet_v3/igie) | [✅](models/cv/classification/mobilenet_v3/ixrt) | 4.3.0 | | MViTv2_base | FP16 | [✅](models/cv/classification/mvitv2_base/igie) | | 4.2.0 | | RegNet_x_16gf | FP16 | [✅](models/cv/classification/regnet_x_16gf/igie) | | 4.3.0 | | RegNet_x_1_6gf | FP16 | [✅](models/cv/classification/regnet_x_1_6gf/igie) | | 4.3.0 | | RegNet_x_3_2gf | FP16 | [✅](models/cv/classification/regnet_x_3_2gf/igie) | | 4.3.0 | | RegNet_x_32gf | FP16 | [✅](models/cv/classification/regnet_x_32gf/igie) | | 4.3.0 | | RegNet_x_400mf | FP16 | [✅](models/cv/classification/regnet_x_400mf/igie) | | 4.3.0 | | RegNet_y_1_6gf | FP16 | [✅](models/cv/classification/regnet_y_1_6gf/igie) | | 4.3.0 | | RegNet_y_16gf | FP16 | [✅](models/cv/classification/regnet_y_16gf/igie) | | 4.3.0 | | RegNet_y_3_2gf | FP16 | [✅](models/cv/classification/regnet_y_3_2gf/igie) | | 4.3.0 | | RegNet_y_32gf | FP16 | [✅](models/cv/classification/regnet_y_32gf/igie) | | 4.3.0 | | RegNet_y_400mf | FP16 | [✅](models/cv/classification/regnet_y_400mf/igie) | | 4.3.0 | | RepVGG | FP16 | [✅](models/cv/classification/repvgg/igie) | [✅](models/cv/classification/repvgg/ixrt) | 4.3.0 | | Res2Net50 | FP16 | [✅](models/cv/classification/res2net50/igie) | [✅](models/cv/classification/res2net50/ixrt) | 4.3.0 | | | INT8 | | [✅](models/cv/classification/res2net50/ixrt) | 4.3.0 | | ResNeSt50 | FP16 | [✅](models/cv/classification/resnest50/igie) | | 4.3.0 | | ResNet101 | FP16 | [✅](models/cv/classification/resnet101/igie) | [✅](models/cv/classification/resnet101/ixrt) | 4.3.0 | | | INT8 | [✅](models/cv/classification/resnet101/igie) | [✅](models/cv/classification/resnet101/ixrt) | 4.3.0 | | ResNet152 | FP16 | [✅](models/cv/classification/resnet152/igie) | | 4.3.0 | | | INT8 | [✅](models/cv/classification/resnet152/igie) | | 4.3.0 | | ResNet18 | FP16 | [✅](models/cv/classification/resnet18/igie) | [✅](models/cv/classification/resnet18/ixrt) | 4.3.0 | | | INT8 | [✅](models/cv/classification/resnet18/igie) | [✅](models/cv/classification/resnet18/ixrt) | 4.3.0 | | ResNet34 | FP16 | | [✅](models/cv/classification/resnet34/ixrt) | 4.3.0 | | | INT8 | | [✅](models/cv/classification/resnet34/ixrt) | 4.3.0 | | ResNet50 | FP16 | [✅](models/cv/classification/resnet50/igie) | [✅](models/cv/classification/resnet50/ixrt) | 4.3.0 | | | INT8 | [✅](models/cv/classification/resnet50/igie) | | 4.3.0 | | ResNetV1D50 | FP16 | [✅](models/cv/classification/resnetv1d50/igie) | [✅](models/cv/classification/resnetv1d50/ixrt) | 4.3.0 | | | INT8 | | [✅](models/cv/classification/resnetv1d50/ixrt) | 4.3.0 | | ResNeXt50_32x4d | FP16 | [✅](models/cv/classification/resnext50_32x4d/igie) | [✅](models/cv/classification/resnext50_32x4d/ixrt) | 4.3.0 | | ResNeXt101_64x4d | FP16 | [✅](models/cv/classification/resnext101_64x4d/igie) | [✅](models/cv/classification/resnext101_64x4d/ixrt) | 4.3.0 | | ResNeXt101_32x8d | FP16 | [✅](models/cv/classification/resnext101_32x8d/igie) | [✅](models/cv/classification/resnext101_32x8d/ixrt) | 4.3.0 | | SEResNet50 | FP16 | [✅](models/cv/classification/se_resnet50/igie) | | 4.3.0 | | ShuffleNetV1 | FP16 | | [✅](models/cv/classification/shufflenet_v1/ixrt) | 4.3.0 | | ShuffleNetV2_x0_5 | FP16 | [✅](models/cv/classification/shufflenetv2_x0_5/igie) | [✅](models/cv/classification/shufflenetv2_x0_5/ixrt) | 4.3.0 | | ShuffleNetV2_x1_0 | FP16 | [✅](models/cv/classification/shufflenetv2_x1_0/igie) | [✅](models/cv/classification/shufflenetv2_x1_0/ixrt) | 4.3.0 | | ShuffleNetV2_x1_5 | FP16 | [✅](models/cv/classification/shufflenetv2_x1_5/igie) | [✅](models/cv/classification/shufflenetv2_x1_5/ixrt) | 4.3.0 | | ShuffleNetV2_x2_0 | FP16 | [✅](models/cv/classification/shufflenetv2_x2_0/igie) | [✅](models/cv/classification/shufflenetv2_x2_0/ixrt) | 4.3.0 | | SqueezeNet 1.0 | FP16 | [✅](models/cv/classification/squeezenet_v1_0/igie) | [✅](models/cv/classification/squeezenet_v1_0/ixrt) | 4.3.0 | | | INT8 | | [✅](models/cv/classification/squeezenet_v1_0/ixrt) | 4.3.0 | | SqueezeNet 1.1 | FP16 | [✅](models/cv/classification/squeezenet_v1_1/igie) | [✅](models/cv/classification/squeezenet_v1_1/ixrt) | 4.3.0 | | | INT8 | | [✅](models/cv/classification/squeezenet_v1_1/ixrt) | 4.3.0 | | SVT Base | FP16 | [✅](models/cv/classification/svt_base/igie) | | 4.3.0 | | Swin Transformer | FP16 | [✅](models/cv/classification/swin_transformer/igie) | | 4.3.0 | | Swin Transformer Large | FP16 | | [✅](models/cv/classification/swin_transformer_large/ixrt) | 4.3.0 | | Twins_PCPVT | FP16 | [✅](models/cv/classification/twins_pcpvt/igie) | | 4.3.0 | | VAN_B0 | FP16 | [✅](models/cv/classification/van_b0/igie) | | 4.3.0 | | VGG11 | FP16 | [✅](models/cv/classification/vgg11/igie) | | 4.3.0 | | VGG13 | FP16 | [✅](models/cv/classification/vgg13/igie) | | 4.3.0 | | VGG13_BN | FP16 | [✅](models/cv/classification/vgg13_bn/igie) | | 4.3.0 | | VGG16 | FP16 | [✅](models/cv/classification/vgg16/igie) | [✅](models/cv/classification/vgg16/ixrt) | 4.3.0 | | | INT8 | [✅](models/cv/classification/vgg16/igie) | | 4.3.0 | | VGG19 | FP16 | [✅](models/cv/classification/vgg19/igie) | | 4.3.0 | | VGG19_BN | FP16 | [✅](models/cv/classification/vgg19_bn/igie) | | 4.3.0 | | ViT | FP16 | [✅](models/cv/classification/vit/igie) | | 4.3.0 | | Wide ResNet50 | FP16 | [✅](models/cv/classification/wide_resnet50/igie) | [✅](models/cv/classification/wide_resnet50/ixrt) | 4.3.0 | | | INT8 | [✅](models/cv/classification/wide_resnet50/igie) | [✅](models/cv/classification/wide_resnet50/ixrt) | 4.3.0 | | Wide ResNet101 | FP16 | [✅](models/cv/classification/wide_resnet101/igie) | | 4.3.0 | #### Object Detection | Model | Prec. | IGIE | ixRT | IXUCA SDK | |------------|-------|-------------------------------------------------|-------------------------------------------------|-----------| | ATSS | FP16 | [✅](models/cv/object_detection/atss/igie) | [✅](models/cv/object_detection/atss/ixrt) | 4.3.0 | | CenterNet | FP16 | [✅](models/cv/object_detection/centernet/igie) | [✅](models/cv/object_detection/centernet/ixrt) | 4.3.0 | | DETR | FP16 | | [✅](models/cv/object_detection/detr/ixrt) | 4.3.0 | | FCOS | FP16 | [✅](models/cv/object_detection/fcos/igie) | [✅](models/cv/object_detection/fcos/ixrt) | 4.3.0 | | FoveaBox | FP16 | [✅](models/cv/object_detection/foveabox/igie) | [✅](models/cv/object_detection/foveabox/ixrt) | 4.3.0 | | FSAF | FP16 | [✅](models/cv/object_detection/fsaf/igie) | [✅](models/cv/object_detection/fsaf/ixrt) | 4.3.0 | | GFL | FP16 | [✅](models/cv/object_detection/gfl/igie) | | 4.3.0 | | HRNet | FP16 | [✅](models/cv/object_detection/hrnet/igie) | [✅](models/cv/object_detection/hrnet/ixrt) | 4.3.0 | | PAA | FP16 | [✅](models/cv/object_detection/paa/igie) | [✅](models/cv/object_detection/paa/ixrt) | 4.3.0 | | RetinaFace | FP16 | [✅](models/cv/object_detection/retinaface/igie) | [✅](models/cv/object_detection/retinaface/ixrt) | 4.3.0 | | RetinaNet | FP16 | [✅](models/cv/object_detection/retinanet/igie) | [✅](models/cv/object_detection/retinanet/ixrt) | 4.3.0 | | RTMDet | FP16 | [✅](models/cv/object_detection/rtmdet/igie) | | 4.3.0 | | SABL | FP16 | [✅](models/cv/object_detection/sabl/igie) | | 4.3.0 | | SSD | FP16 | [✅](models/cv/object_detection/ssd/igie) | | 4.3.0 | | YOLOF | FP16 | [✅](models/cv/object_detection/yolof/igie) | | 4.3.0 | | YOLOv3 | FP16 | [✅](models/cv/object_detection/yolov3/igie) | [✅](models/cv/object_detection/yolov3/ixrt) | 4.3.0 | | | INT8 | [✅](models/cv/object_detection/yolov3/igie) | [✅](models/cv/object_detection/yolov3/ixrt) | 4.3.0 | | YOLOv4 | FP16 | [✅](models/cv/object_detection/yolov4/igie) | [✅](models/cv/object_detection/yolov4/ixrt) | 4.3.0 | | | INT8 | [✅](models/cv/object_detection/yolov4/igie16) | [✅](models/cv/object_detection/yolov4/ixrt16) | 4.3.0 | | YOLOv5 | FP16 | [✅](models/cv/object_detection/yolov5/igie) | [✅](models/cv/object_detection/yolov5/ixrt) | 4.3.0 | | | INT8 | [✅](models/cv/object_detection/yolov5/igie) | [✅](models/cv/object_detection/yolov5/ixrt) | 4.3.0 | | YOLOv5s | FP16 | | [✅](models/cv/object_detection/yolov5s/ixrt) | 4.3.0 | | | INT8 | | [✅](models/cv/object_detection/yolov5s/ixrt) | 4.3.0 | | YOLOv6 | FP16 | [✅](models/cv/object_detection/yolov6/igie) | [✅](models/cv/object_detection/yolov6/ixrt) | 4.3.0 | | | INT8 | | [✅](models/cv/object_detection/yolov6/ixrt) | 4.3.0 | | YOLOv7 | FP16 | [✅](models/cv/object_detection/yolov7/igie) | [✅](models/cv/object_detection/yolov7/ixrt) | 4.3.0 | | | INT8 | [✅](models/cv/object_detection/yolov7/igie) | [✅](models/cv/object_detection/yolov7/ixrt) | 4.3.0 | | YOLOv8 | FP16 | [✅](models/cv/object_detection/yolov8/igie) | [✅](models/cv/object_detection/yolov8/ixrt) | 4.3.0 | | | INT8 | [✅](models/cv/object_detection/yolov8/igie) | [✅](models/cv/object_detection/yolov8/ixrt) | 4.3.0 | | YOLOv9 | FP16 | [✅](models/cv/object_detection/yolov9/igie) | [✅](models/cv/object_detection/yolov9/ixrt) | 4.3.0 | | YOLOv10 | FP16 | [✅](models/cv/object_detection/yolov10/igie) | [✅](models/cv/object_detection/yolov10/ixrt) | 4.3.0 | | YOLOv11 | FP16 | [✅](models/cv/object_detection/yolov11/igie) | [✅](models/cv/object_detection/yolov11/ixrt) | 4.3.0 | | YOLOv12 | FP16 | [✅](models/cv/object_detection/yolov12/igie) | | 4.3.0 | | YOLOv13 | FP16 | [✅](models/cv/object_detection/yolov13/igie) | | 4.3.0 | | YOLOX | FP16 | [✅](models/cv/object_detection/yolox/igie) | [✅](models/cv/object_detection/yolox/ixrt) | 4.3.0 | | | INT8 | [✅](models/cv/object_detection/yolox/igie) | [✅](models/cv/object_detection/yolox/ixrt) | 4.3.0 | #### Face Recognition | Model | Prec. | IGIE | ixRT | IXUCA SDK | |---------|-------|------|----------------------------------------------|-----------| | FaceNet | FP16 | | [✅](models/cv/face_recognition/facenet/ixrt) | 4.3.0 | | | INT8 | | [✅](models/cv/face_recognition/facenet/ixrt) | 4.3.0 | #### OCR (Optical Character Recognition) | Model | Prec. | IGIE | IXUCA SDK | |---------------|-------|---------------------------------------|-----------| | Kie_layoutXLM | FP16 | [✅](models/cv/ocr/kie_layoutxlm/igie) | 4.3.0 | | SVTR | FP16 | [✅](models/cv/ocr/svtr/igie) | 4.3.0 | #### Pose Estimation | Model | Prec. | IGIE | ixRT | IXUCA SDK | |----------------------|-------|-----------------------------------------------|----------------------------------------------------------|-----------| | HRNetPose | FP16 | [✅](models/cv/pose_estimation/hrnetpose/igie) | | 4.3.0 | | Lightweight OpenPose | FP16 | | [✅](models/cv/pose_estimation/lightweight_openpose/ixrt) | 4.3.0 | | RTMPose | FP16 | [✅](models/cv/pose_estimation/rtmpose/igie) | [✅](models/cv/pose_estimation/rtmpose/ixrt) | 4.3.0 | #### Instance Segmentation | Model | Prec. | IGIE | ixRT | IXUCA SDK | |------------|-------|------|-----------------------------------------------------|-----------| | Mask R-CNN | FP16 | | [✅](models/cv/instance_segmentation/mask_rcnn/ixrt) | 4.2.0 | | SOLOv1 | FP16 | | [✅](models/cv/instance_segmentation/solov1/ixrt) | 4.3.0 | #### Semantic Segmentation | Model | Prec. | IGIE | ixRT | IXUCA SDK | |-------|-------|------------------------------------------------|------|-----------| | UNet | FP16 | [✅](models/cv/semantic_segmentation/unet/igie) | | 4.3.0 | #### Multi-Object Tracking | Model | Prec. | IGIE | ixRT | IXUCA SDK | |---------------------|-------|----------------------------------------------------|------|-----------| | FastReID | FP16 | [✅](models/cv/multi_object_tracking/fastreid/igie) | | 4.3.0 | | DeepSort | FP16 | [✅](models/cv/multi_object_tracking/deepsort/igie) | | 4.3.0 | | | INT8 | [✅](models/cv/multi_object_tracking/deepsort/igie) | | 4.3.0 | | RepNet-Vehicle-ReID | FP16 | [✅](models/cv/multi_object_tracking/repnet/igie) | | 4.3.0 | ### Multimodal | Model | vLLM | IxFormer | IXUCA SDK | |---------------------|-----------------------------------------------------------------------|------------------------------------------------------------|-----------| | Aria | [✅](models/multimodal/vision_language_model/aria/vllm) | | 4.3.0 | | Chameleon-7B | [✅](models/multimodal/vision_language_model/chameleon_7b/vllm) | | 4.3.0 | | CLIP | | [✅](models/multimodal/vision_language_model/clip/ixformer) | 4.3.0 | | Fuyu-8B | [✅](models/multimodal/vision_language_model/fuyu_8b/vllm) | | 4.3.0 | | H2OVL Mississippi | [✅](models/multimodal/vision_language_model/h2vol/vllm) | | 4.3.0 | | Idefics3 | [✅](models/multimodal/vision_language_model/idefics3/vllm) | | 4.3.0 | | InternVL2-4B | [✅](models/multimodal/vision_language_model/intern_vl/vllm) | | 4.3.0 | | LLaVA | [✅](models/multimodal/vision_language_model/llava/vllm) | | 4.3.0 | | LLaVA-Next-Video-7B | [✅](models/multimodal/vision_language_model/llava_next_video_7b/vllm) | | 4.3.0 | | Llama-3.2 | [✅](models/multimodal/vision_language_model/llama-3.2/vllm) | | 4.3.0 | | MiniCPM-V 2 | [✅](models/multimodal/vision_language_model/minicpm_v/vllm) | | 4.3.0 | | Pixtral | [✅](models/multimodal/vision_language_model/pixtral/vllm) | | 4.3.0 | ### NLP #### PLM (Pre-trained Language Model) | Model | Prec. | IGIE | ixRT | IXUCA SDK | |------------------|-------|-------------------------------------------|-------------------------------------------|-----------| | ALBERT | FP16 | | [✅](models/nlp/plm/albert/ixrt) | 4.3.0 | | BERT Base NER | INT8 | [✅](models/nlp/plm/bert_base_ner/igie) | | 4.3.0 | | BERT Base SQuAD | FP16 | [✅](models/nlp/plm/bert_base_squad/igie) | [✅](models/nlp/plm/bert_base_squad/ixrt) | 4.3.0 | | | INT8 | | [✅](models/nlp/plm/bert_base_squad/ixrt) | 4.3.0 | | BERT Large SQuAD | FP16 | [✅](models/nlp/plm/bert_large_squad/igie) | [✅](models/nlp/plm/bert_large_squad/ixrt) | 4.3.0 | | | INT8 | [✅](models/nlp/plm/bert_large_squad/igie) | [✅](models/nlp/plm/bert_large_squad/ixrt) | 4.3.0 | | DeBERTa | FP16 | | [✅](models/nlp/plm/deberta/ixrt) | 4.3.0 | | RoBERTa | FP16 | | [✅](models/nlp/plm/roberta/ixrt) | 4.3.0 | | RoFormer | FP16 | | [✅](models/nlp/plm/roformer/ixrt) | 4.3.0 | | VideoBERT | FP16 | | [✅](models/nlp/plm/videobert/ixrt) | 4.2.0 | ### Audio #### Speech Recognition | Model | Prec. | IGIE | ixRT | IXUCA SDK | |-----------------|-------|-----------------------------------------------------|-----------------------------------------------------------|-----------| | Conformer | FP16 | [✅](models/audio/speech_recognition/conformer/igie) | [✅](models/audio/speech_recognition/conformer/ixrt) | 4.3.0 | | Transformer ASR | FP16 | | [✅](models/audio/speech_recognition/transformer_asr/ixrt) | 4.2.0 | ### Others #### Recommendation Systems | Model | Prec. | IGIE | ixRT | IXUCA SDK | |-------------|-------|------|------------------------------------------------------|-----------| | Wide & Deep | FP16 | | [✅](models/others/recommendation/wide_and_deep/ixrt) | 4.3.0 | --- ## Docker | Docker Installer | IXUCA SDK | Introduction | |-----------|--------|--------------| | corex-docker-installer-4.3.0-*-py3.10-x86_64.run | 4.3.0 | for small model inference | | corex-docker-installer-4.3.0-*-llm-py3.10-x86_64.run | 4.3.0 | for large model inference | ## Community ### Code of Conduct Please refer to DeepSpark Code of Conduct on [Gitee](https://gitee.com/deep-spark/deepspark/blob/master/CODE_OF_CONDUCT.md) or on [GitHub](https://github.com/Deep-Spark/deepspark/blob/main/CODE_OF_CONDUCT.md). ### Contact Please contact . ### Contribution Please refer to the [DeepSparkInference Contributing Guidelines](CONTRIBUTING.md). ### Disclaimers DeepSparkInference only provides download and preprocessing scripts for public datasets. These datasets do not belong to DeepSparkInference, and DeepSparkInference is not responsible for their quality or maintenance. Please ensure that you have the necessary usage licenses for these datasets. Models trained based on these datasets can only be used for non-commercial research and education purposes. To dataset owners: If you do not want your dataset to be published on DeepSparkInference or wish to update the dataset that belongs to you on DeepSparkInference, please submit an issue on Gitee or Github. We will delete or update it according to your issue. We sincerely appreciate your support and contributions to our community. ## License This project is released under [Apache-2.0](LICENSE) License.