Skip to Content

Troubleshooting

Common issues and their solutions.

Installation Issues

Command Not Found

Problem: inferno: command not found after installation.

Solutions:

Permission Denied

Problem: Permission errors when running Inferno.

Solutions:

# Make binary executable (Linux/macOS)
chmod +x /usr/local/bin/inferno
 
# Run with sudo if needed for installation
sudo mv inferno /usr/local/bin/

Model Issues

Model Not Found

Problem: Error: Model 'xyz' not found

Solutions:

  1. List available models:

    inferno models list
  2. Download the model:

    inferno models download model-name
  3. Check model path in config:

    inferno config show

Model Download Fails

Problem: Model download interrupted or fails.

Solutions:

GPU Issues

GPU Not Detected

Problem: Inferno not using GPU acceleration.

Solutions:

  1. NVIDIA (CUDA):

    # Check CUDA installation
    nvidia-smi
     
    # Install CUDA toolkit if missing
    # https://developer.nvidia.com/cuda-downloads
  2. AMD (ROCm):

    # Check ROCm installation
    rocm-smi
     
    # Install ROCm if missing
    # https://rocm.docs.amd.com/
  3. Apple Silicon:

  4. Force CPU mode if needed:

    inferno run --device cpu --model llama2 --prompt "test"

Out of Memory (OOM)

Problem: Error: Out of GPU memory

Solutions:

API Server Issues

Port Already in Use

Problem: Error: Address already in use

Solutions:

  1. Use a different port:

    inferno serve --port 3000
  2. Find and kill process using the port:

    # Linux/macOS
    lsof -i :8080
    kill -9 <PID>
     
    # Windows
    netstat -ano | findstr :8080
    taskkill /PID <PID> /F

Connection Refused

Problem: Cannot connect to API server.

Solutions:

Performance Issues

Slow Inference

Problem: Inference taking too long.

Solutions:

  1. Enable GPU acceleration (see GPU Issues above)
  2. Use a smaller/faster model
  3. Reduce max_tokens
  4. Check system resources:
    # CPU and memory usage
    top
    htop
     
    # GPU usage (NVIDIA)
    nvidia-smi

High Memory Usage

Problem: Inferno using too much RAM.

Solutions:

Docker Issues

Container Won’t Start

Problem: Docker container exits immediately.

Solutions:

  1. Check logs:

    docker logs <container-id>
  2. Run interactively:

    docker run -it ghcr.io/ringo380/inferno:latest /bin/bash
  3. Verify volume mounts:

    docker run -v ~/models:/models ghcr.io/ringo380/inferno:latest

GPU Not Available in Docker

Problem: GPU not detected in Docker container.

Solutions:

Debugging

Enable Verbose Logging

Get more detailed output:

inferno --verbose run --model llama2 --prompt "test"
 
# Or set log level
export INFERNO_LOG_LEVEL=debug
inferno serve

Check Configuration

View current configuration:

inferno config show

Verify Installation

Run diagnostics:

inferno --version
inferno models list
inferno config show

Getting Help

If you’re still experiencing issues:

  1. Check GitHub Issues  for similar problems
  2. Create a new issue with:

Common Error Messages

ErrorCauseSolution
Model not foundModel not downloadedRun inferno models download MODEL
CUDA errorGPU driver issueUpdate NVIDIA drivers
Connection refusedServer not runningStart server with inferno serve
Permission deniedFile permissionsRun with sudo or fix permissions
Out of memoryModel too largeUse smaller model or add RAM