Skip to main content

AI and Machine Learning

Methodology

Bluefin was created by engineers, but was brought to life by Jacob Schnurr and Andy Frazer. The artwork is free for you to use and will always be made by humans. It is there to remind us that open source is an ecosystem that needs to be sustained. The software we make has an effect on the world. Bluefin's AI integration will always be user controlled, with a focus on open source models and tools.

AI is an extension of cloud native

Bluefin's focus in AI is providing a generic API endpoint to the operating system that is controlled by the user. Just as Bluefin's operating system is built with CNCF tech like bootc and podman, this experience is powered by Agentic AI Foundation tech like goose. With a strong dash of the open source components that power RHEL Lightspeed.

Bluespeed

"Bluespeed" is our collection of Bluefin's developer experience tools and support for AI development workflows. We do this via community managed set of tool recommendations and configuration. We believe that the operating system should have more API endpoints for AI.

  • "Bring your own LLM" aproach, it should be easy to switch between local models and hosted ones
    • Goose as the primary interface to hosted and local models
  • Accelerate open standards in AI by shipping tools from the Agentic AI Foundation, CNCF, and other foundations
  • Local LLM service management
    • Model management via ramalama and Docker Model, your choice
  • GPU Acceleration for both Nvidia and AMD are included out of the box and usually do not require any extra setup
  • Highlight great AI/ML applications on FlatHub in our curated section in the App Store
  • A great reason to sell more swag

We work closely with the RHEL Lightspeed team by shipping their code, giving feedback, and pushing the envelope where we can.

AI Lab with Podman Desktop

The AI Lab extension can be installed inside the included Podman Desktop to provide a graphical interface for managing local models:

image

AI Command Line Tools

The following AI-focused command-line tools are available via homebrew, install individually or use this command to install them all: ujust bbrew and choose the ai menu option:

NameDescription
aichatAll-in-one AI-Powered CLI Chat & Copilot
block-goose-cliBlock Protocol AI agent CLI
claude-codeClaude coding agent with desktop integration
codexCode editor for OpenAI's coding agent that runs in your terminal
copilot-cliGitHub Copilot CLI for terminal assistance
crushAI coding agent for the terminal, from charm.sh
gemini-cliCommand-line interface for Google's Gemini API
kimi-cliCLI for Moonshot AI's Kimi models
llmAccess large language models from the command line
lm-studioDesktop app for running local LLMs
mistral-vibeCLI for Mistral AI models
modsAI on the command-line, from charm.sh
opencodeAI coding agent for the terminal
qwen-codeCLI for Qwen3-Coder models
ramalamaManage and run AI models locally with containers
whisper-cppHigh-performance inference of OpenAI's Whisper model

Ramalama

Install Ramalama via brew install ramalama: manage local models and is the preferred default experience. It's for people who work with local models frequently and need advanced features. It offers the ability to pull models from huggingface, ollama, and any container registry. By default it pulls from ollama.com, check the Ramalama documentation for more information.

Ramalama's command line experience is similar to Podman. Bluefin sets rl as an alias for ramalama, for brevity. Examples include:

rl pull llama3.2:latest
rl run llama3.2
rl run deepseek-r1

You can also serve the models locally:

rl serve deepseek-r1

Then go to http://127.0.0.0:8080 in your browser.

Ramalama will automatically pull in anything your host needs to do the workload. The images are also stored in the same container storage as your other containers. This allows for centralized management of the models and other podman images:

❯ podman images
REPOSITORY TAG IMAGE ID CREATED SIZE
quay.io/ramalama/rocm latest 8875feffdb87 5 days ago 6.92 GB

Integrating with Existing Tools

ramalama serve will serve an OpenAI compatible endpoint at http://0.0.0.0:8080, you can use this to configure tools that do not support ramalama directly:

Newelle

Other Ramalama tips

  • Force Vulkan instead of ROCm: ramalama serve --image quay.io/ramalama/ramalama gpt-oss:latest
  • Strix Halo users: ramalama serve --image docker.io/kyuz0/amd-strix-halo-toolboxes:vulkan-radv gpt-oss:latest

Running AI Agents in VS Code

Here is an example of using devcontainers to run agents inside containers for isolation:

Docker Model Runner

[TBD]

Alpaca Graphical Client

For light chatbot usage we recommend that users install Alpaca to manage and chat with your LLM models from within a native desktop application. Alpaca supports Nvidia and AMD[^1] acceleration natively.

Only a keystroke away

Bluefin binds Ctrl-Alt-Backspace as a quicklaunch for Alpaca automatically after you install it!

Configuration

Alpaca

image

Automated Troubleshooting (WIP)

Bluefin ships with automated troubleshooting tools: