.. _Introduction to LLiMa: Introduction to LLiMa ===================== Overview -------- The GenAI Model Compilation feature streamlines the process of compiling GenAI models based on HF safetensors or GGUF model format. For a wide set of different models like ``Llama``, ``Gemma``, ``Phi``, ``Qwen`` or ``Mistral`` from Hugging Face, the SDK automatically generates all required binary/elf files along with the Python orchestration script, enabling direct execution on the Sima.ai Modalix platform. To quickly get started, Sima has precompiled several popular LLM models and published them on `Hugging Face `_. You can download and run these models immediately using the following commands: .. code-block:: console modalix:~$ cd /media/nvme && mkdir llima && cd llima modalix:~$ sima-cli install -v 2.0.0 samples/llima -t select Wait until installation completes then run: .. code-block:: console modalix:~$ cd simaai-genai-demo && ./run.sh This command prompts you to select and download a specific precompiled model for evaluating the Sima.ai Modalix platform. More information can be found in the `LLiMa demo application <../overview/hello_sima/run_demos.html#llm-demo>`_. Supported Models ---------------- The following table shows the supported model architectures and their capabilities: .. list-table:: :widths: 30 15 55 :header-rows: 1 * - Model Architecture - Type - Supported Sizes * - `Llama 2 `_ - LLM - `7b `_ * - `Llama 3.1 `_ - LLM - `8b `_ * - `Llama 3.2 `_ - LLM - 1b, `3b `_ * - `Gemma 1 `_ - LLM - 2b, 7b * - `Gemma 2 `_ - LLM - 2b, 9b * - `Gemma 3 `_ - LLM - `1b `_, `4b `_ * - `Phi 3.5 mini `_ - LLM - `3.8b `_ * - `Qwen 2.5 `_ - LLM - `0.5b `_, `1.5b `_, 3b, `7b `_ * - `Qwen 3 `_ - LLM - `0.6b `_, `1.7b `_, `4b `_, `8b `_ * - `Mistral 1 `_ - LLM - `7b `_ * - `Llava 1.5 `_ - VLM - `7b `_ * - `PaliGemma `_ - VLM - `3b `_ * - `Gemma 3 `_ - VLM - `4b `_ Limitations ----------- .. list-table:: :widths: 30 70 :header-rows: 1 :class: wrapped-table * - Limitation Type - Description * - Model Architecture - Only models based on the architectures listed above are supported. * - Model Parameters - Only models with parameter count less than 10B are supported. * - HF Models - Models must be downloaded from Hugging Face and contain: ``config.json``, ``tokenizer.json``, ``tokenizer_config.json``, ``generation_config.json`` and weights in safetensors format * - GGUF Models - GGUF format is supported for LLMs only. VLMs must be compiled from the Hugging Face safetensors format. * - Gemma3 VLM - Supported with modfied SigLip 448 vision encoder * - LLAMA 3.2 Vision - Vision models are not supported