OpenAD AI Assistant


  • Supported LLMs are IBM BAM and Ollama
  • Ollama requires an 8GB GPU.
  • WatsonX support is coming soon.


IBM BAM Setup

To use IBM BAM if you have access to it, simply provide your API key when prompted.

set llm bam
tell me <enter prompt>


Ollama Setup

  1. Install Ollama onto your platform.

  2. Download the appropriate models.

    ollama pull llama3:latest
    ollama pull nomic-embed-text
    
  3. Start the server if not already started.

    ollama serve
    

That’s it for local usage. If you want to run Ollama remotely, continue below.

Ollama Remote Setup with SkyPilot

  1. Check out our configuration file to launch ollama on SkyPilot: ollama_setup.yaml

    sky serve up ollama_setup.yaml
    
  2. Set up local environment variables

    • For windows setx OLLAMA_HOST=<sky-server-ip>:11434
    • For Linux and macOS export OLLAMA_HOST=<sky-server-ip>:11434
    • To reset to local use OLLAMA_HOST=0.0.0.0:11434

Run Ollama

Note: If prompted for an API key and none was setup, just leave the input empty.

set llm ollama
tell me <enter prompt>