▶️Use Model

To help developers to easily find and run the right model, Nexa AI Hub provide comprehensive filter system and SDK.

Explore in Model Hub

The goal of Nexa Model Hub is to help developers find the most suitable models. To achieve the goal, we provide the following filter options:

Model Type

  • Computer Vision

    • Image-to-Text

    • Image-to-Image

  • Audio

    • Text-to-Speech

    • Automatic Speech Recognition

  • Multimodal

    • Image-Text-to-text

  • NLP

    • Text Generation

    • Chat Completion

    • Question Answering

File Format Tag

  • GGUF

GGUF is an optimized binary format designed for efficient model loading and saving, particularly suited for inference tasks. It is compatible with GGML and other executors. Developed by @ggerganov, the creator of llama.cpp (a widely-used C/C++ LLM inference framework), GGUF forms the foundation of the Nexa SDK's GGML component.

  • ONNX

ONNX is an open standard format for representing machine learning models. It establishes a common set of operators and a unified file format, enabling AI developers to utilize models across various frameworks, tools, runtimes, and compilers. ONNX shows unique performance advantages on devices with limited ram(mobile, IoT). The Nexa SDK's ONNX component is built upon the onnxruntime framework.

Parameters

The Nexa Model Hub specializes in on-device models with parameter less than 10 billion.

RAM

This metric indicates the minimum random access memory (RAM) necessary for local model execution.

File Size

Displays the total storage space required for the model.

Use Model

Download Nexa SDK

Follow the Installation to download the appropriate SDK for your operating system.

Run Model using SDK

Nexa SDK enables developers to use one line of code to run the model that fits your specific requirement locally. The one line of code follows this pattern:

nexa model_type MODEL_PATH

To see example and popular MODEL_PATH, see Supported Popular Models below:

ModelTypeFormatCommand

NLP

GGUF

nexa run octopus-v2

NLP

GGUF

nexa run octopus-v4

NLP

GGUF

nexa run tinyllama

NLP

GGUF/ONNX

nexa run llama2

NLP

GGUF/ONNX

nexa run llama3

NLP

GGUF/ONNX

nexa run llama3.1

NLP

GGUF/ONNX

nexa run gemma

NLP

GGUF

nexa run gemma2

NLP

GGUF

nexa run qwen1.5

NLP

GGUF/ONNX

nexa run qwen2

NLP

GGUF

nexa run qwen2.5

NLP

GGUF

nexa run mathqwen

NLP

GGUF/ONNX

nexa run mistral

NLP

GGUF

nexa run codegemma

NLP

GGUF

nexa run codellama

NLP

GGUF

nexa run codeqwen

NLP

GGUF

nexa run deepseek-coder

NLP

GGUF

nexa run dolphin-mistral

NLP

GGUF

nexa run phi2

NLP

GGUF/ONNX

nexa run phi3

NLP

GGUF

nexa run llama2-uncensored

NLP

GGUF

nexa run llama3-uncensored

NLP

GGUF

nexa run llama2-function-calling

Multimodal

GGUF

nexa run nanollava

Multimodal

GGUF

nexa run llava-phi3

Multimodal

GGUF

nexa run llava-llama3

Multimodal

GGUF

nexa run llava1.6-mistral

Multimodal

GGUF

nexa run llava1.6-vicuna

Computer Vision

GGUF

nexa run sd1-4

Computer Vision

GGUF/ONNX

nexa run sd1-5

Computer Vision

GGUF/ONNX

nexa run lcm-dreamshaper

Computer Vision

GGUF

nexa run hassaku-lcm

Computer Vision

GGUF

nexa run anything-lcm

Audio

BIN

nexa run faster-whisper-tiny

Audio

BIN

nexa run faster-whisper-small

Audio

BIN

nexa run faster-whisper-medium

Audio

BIN

nexa run faster-whisper-base

Audio

BIN

nexa run faster-whisper-large

Download more official models from Nexa model hub

To find the right code following "nexa", there are two ways:

  1. Find the model in the model hub using search and filter, and click on "run this model" to copy the code to run model locally.

  1. Follow the run a model section in CLI Reference

Last updated