by Radosław Śmigielski
GGUF model file format is a binary file Full GGUF format documentation GGUF file format
Safetensors file format documentation Safetensors file format
ERROR 03-02 10:01:12 engine.py:400] ValueError: Bfloat16 is only supported on GPUs with compute capability
of at least 8.0. Your NVIDIA GeForce GT 1030 GPU has compute capability 6.1.
You can use float16 instead by explicitly setting the`dtype` flag in CLI,
for example: --dtype=half.
ValueError: Failed to determine backend: Cannot determine which backend to use:
The model file /home/ai/.cache/instructlab/models is not a GGUF format
nor a directory containing huggingface safetensors files.
Cannot determine which backend to use.
Please use a GGUF file for llama-cpp or a directory containing huggingface safetensors files for vllm.
Note that vLLM is only supported on Linux.