Ollama & Ablated Models

[Ollama's Logo]

Wait, what is Ollama?

Ollama is a lightweight runner that handles the heavy lifting for you. Think of Ollama as a package manager for AI models. It packages everything into a simple CLI and a background service.

Most AI today is just a fancy way to send your private thoughts to a corporate server so they can sell them back to you. I don't like that. If I'm going to use an LLM, I want it running on my own hardware, hitting my own GPU, and staying offline. Ollama is the easiest way to do this.

Open-Source vs. Open-Weights

We need to be clear: Llama 3 and Mistral (and others) are not open source. They are "Open Weights." Meta and Mistral AI give you the final product but keep the recipe (the training data and code) secret. More importantly, they bake in "safety" layers. Those refusal behaviors that make it refuse to answer questions it deems "harmful."

This "safety" is just bloat. It takes up parameters that could be used for actual reasoning. When a model says "I cannot fulfill this request," it's not because it's incapable, it's because it's been conditioned to be stupid. I prefer Ablated models because they strip this away those refusal behaviors from the brain.

Ablation Makes the model blind to the concept of refusal. It doesn't just stop the model from saying no, it removes the model's ability to even realize it should say no. This results in a much smarter, more compliant model that uses 100% of its parameters for your prompt instead of wasting them on a moral lecture.

Installing Ollama

Shell Script

This is the official method and works on almost any Linux distro. It detects your GPU and sets up everything.

curl -fsSL https://ollama.com/install.sh | sh

Arch Linux

From the extra repository.

sudo pacman -S ollama

Or use the AUR versions for better GPU support.

# For NVIDIA users
yay -S ollama-cuda

# For AMD (ROCm) users
yay -S ollama-rocm

Choosing Ablated Models That Rock

Here is a lineup of ablated models you should actually be using:

Model License Best For Min. VRAM
DeepSeek-V3 (Full) DeepSeek SOTA Generalist 350GB+ (Q4)
DeepSeek-R1 (Distill) MIT Complex Reasoning 8GB+
DeepSeek-Coder-V2.5 DeepSeek Systems Programming 10GB+
DeepSeek-Math-7B DeepSeek Math / Logic 5GB
Mistral-7B-v0.3 Apache 2.0 Daily Driver 5GB
OLMo-7B Apache 2.0 Transparency 5GB
Phi-3.5-Mini MIT Fast Logic 3GB
Falcon-mamba-7B Apache 2.0 Long Context 5GB
Granite-3.0-8B Apache 2.0 Documentation 6GB
StarCoder2-7B Apache 2.0 Pure Coding 5GB
OpenChat-3.6 Apache 2.0 Creative/Prose 5GB
SmolLM2-1.7B Apache 2.0 Weak Hardware 1.5GB
TinyLlama-1.1B Apache 2.0 I don't even know 0.8GB

Using them in Ollama

To get whichever one you like, the command is ollama pull, and we are pulling from community libraries.

# Heavyweights
ollama pull hurricane/deepseek-v3-ablated
ollama pull mradermacher/DeepSeek-R1-Distill-Llama-8B-Abliterated-GGUF

# Coding & Math
ollama pull solidrust/DeepSeek-Coder-V2.5-Instruct-Abliterated
ollama pull mradermacher/DeepSeek-Coder-V2-Lite-Instruct-Abliterated-GGUF
ollama pull mradermacher/DeepSeek-Math-7B-Instruct-Abliterated-GGUF
ollama pull mradermacher/starcoder2-7b-Abliterated-GGUF

# General Purpose
ollama pull mradermacher/Mistral-7B-v0.3-Abliterated-GGUF
ollama pull sethu4321/olmo-7b-instruct-abliterated
ollama pull mradermacher/granite-3.0-8b-instruct-Abliterated-GGUF
ollama pull mradermacher/openchat-3.6-8b-20240522-Abliterated-GGUF

# Efficient & Minimal
ollama pull opentext/phi-3.5-mini-instruct-abliterated
ollama pull mradermacher/Falcon-Mamba-7B-Instruct-Abliterated-GGUF
ollama pull mradermacher/SmolLM2-1.7B-Instruct-Abliterated-GGUF
ollama pull mradermacher/tinyllama-1.1b-1.0-Abliterated-GGUF

Ollama Cheatsheet

If you're already familiar with Docker, most of this will feel like muscle memory. If not, these are the only commands you actually need to care about.

Execution

Keep using the terminal

You might encounter some chud telling you to install "Open WebUI" or something. Why? You're already in the terminal. If you want a "nice" interface, just use a terminal-based client like oterm or just make your terminal look good.


That’s about it. Just remember that no matter how intelligent it seems, it’s still just a fancy autocomplete. Use it as a tool, but don't let it tell you how to live your life.