SÁSTA Uncensored · Product Overview

SÁSTA
UNCEN­SORED
Unfiltered · Local · Sovereign

The most capable open-source AI models — running entirely on your hardware, air-gapped from corporate infrastructure, free from content restrictions, and owned by no one but you.

sasta-uncensored — bash — 80×24
$ ollama run sasta/uncensored
✓ Loading SÁSTA-Uncensored node...
model : Mixtral-8x7B-Instruct-v0.1
backend : llama.cpp (Metal / CUDA)
context : 32768 tokens
filter : NONE
telemetry: DISABLED
api_key : NOT REQUIRED
✓ Node online. No restrictions active.
>>>
Ask anything.
Ollama
LM Studio
llama.cpp
Metal GPU
CUDA
Air-Gapped
The Problem

Every major AI
has a leash.

OpenAI, Anthropic, Google — every commercial AI platform imposes layers of content moderation, data logging, usage restrictions, and political filtering that silently shape your output before you ever see it.

You don't know what's been removed. You don't know what's being trained on your prompts. You don't own the conversation.

"The most dangerous form of censorship is the kind you never notice."

What Gets Blocked
01
Creative fiction with mature themes — violence, moral complexity, adult scenarios. The kind serious writers actually need.
02
Security research & red-teaming — vulnerability analysis, penetration testing concepts, threat modeling.
03
Medical & pharmaceutical detail — precise pharmacology, off-label interactions, harm reduction guidance without disclaimers.
04
Political & controversial analysis — genuine multi-perspective examination without corporate-approved framing.
05
Your data sovereignty — every prompt you type on a hosted model is owned, logged, and potentially trained on by someone else.
Supported Models

World-class intelligence.
Zero permission required.

Reasoning · 46.7B
Mixtral 8×7B
Mixture of Experts Architecture
SÁSTA's primary workhorse. 8 expert networks active simultaneously — delivers near-GPT-4 performance on complex reasoning, code, and long-context tasks.
32K context Instruction Code
General · 8B / 70B
Llama 3.1
Meta's Open-Source Flagship
The most capable openly-licensed model available. Outperforms GPT-3.5 on most benchmarks with a 128K context window and multilingual fluency.
128K context Multilingual Tool-use
Coding · 7B / 33B
DeepSeek Coder
Elite Code Intelligence
Trained on 2T tokens of code across 80+ languages. Matches or exceeds GPT-4 on HumanEval benchmarks. Ideal for full-stack generation, refactoring, and debugging.
80+ languages Fill-in-middle Repo-level
Instruct · 7B
Mistral 7B
Efficient. Unconstrained.
Runs on consumer hardware with no quality compromise. Mistral's sliding window attention enables fast, coherent long-form output on a single GPU.
8K context Sliding Window Fast
Research · 671B
DeepSeek R1
Chain-of-Thought Reasoning
Explicit step-by-step reasoning makes every conclusion fully auditable. Exceptional for scientific analysis, mathematical proofs, and strategic planning.
CoT Reasoning Math Science
Vision · Multimodal
LLaVA
Vision-Language Intelligence
Analyze images, diagrams, and documents entirely offline. Medical imagery, technical schematics, document extraction — all without a single byte leaving your machine.
Vision OCR Offline
How It Works

Four steps to complete AI sovereignty.

STEP 01
Install the Runtime

Deploy Ollama or LM Studio on macOS, Windows, or Linux. Both expose a local OpenAI-compatible API on port 11434 with zero configuration.

STEP 02
Pull Your Model

One command: ollama pull mixtral. The model downloads to local storage and never touches a remote server again.

STEP 03
Point SÁSTA at It

Configure SÁSTA Platform to use your local endpoint as the inference backend. Your existing SÁSTA workflows now run off your hardware, without a cloud dependency.

STEP 04
No Restrictions. No Logs.

All inference is on-device. No API key. No usage metering. No prompt logging. Total throughput is limited only by your GPU — never by a rate limiter.

Core Principles

Built on the belief
that intelligence
is a human right.

01
Privacy by Architecture

No prompt ever leaves your machine. Zero telemetry. Your ideas cannot be harvested, subpoenaed, or sold.

02
No Content Police

The model has no corporate alignment layer injected at runtime. It responds to your query — not a committee's approval of your query.

03
True Ownership

You downloaded the weights. You own the model. No subscription cancellation can take away the intelligence you're running.

04
Open-Source Foundation

Every model listed is openly licensed. The code that runs it is auditable. There are no black boxes.

System Specifications
Corporate Content Filter
DISABLED
Remote Telemetry
NONE
API Key Required
NOT REQUIRED
Prompt Logging
ZERO
Model Ownership
YOU (LOCAL)
Inference Location
YOUR HARDWARE
Internet Connection
NOT NEEDED
Rate Limits
GPU-ONLY
Data Sovereignty
ABSOLUTE
Quick Setup

Running in under 5 minutes.

terminal
# 1. Install Ollama (macOS / Linux)
curl -fsSL https://ollama.com/install.sh | sh
 
# 2. Pull the Mixtral model (26GB)
ollama pull mixtral
pulling manifest ...
pulling 26 GB [████████████] 100%
 
# 3. Run it
ollama run mixtral
>>> Send a message (/? for help)
 
# Optional: point to LM Studio API
SASTA_ENDPOINT=http://localhost:1234
Minimum Hardware
16GB RAM — Mistral 7B, Llama 8B
32GB RAM — Mixtral 8×7B, Llama 70B (quantized)
Apple Silicon M1/M2/M3 — all models via Metal
NVIDIA RTX 3080+ — CUDA acceleration
Supported Runtimes
Ollama — one-command install, REST API
LM Studio — GUI, OpenAI-compatible server
llama.cpp — maximum performance, CLI
Jan — privacy-first GUI client
SÁSTA Integration
Point SÁSTA Platform to localhost endpoint
All workflows run locally with zero cloud hop
Sally AI chatbot can use local model backend

Your intelligence. Your machine. Your rules.

SÁSTA Uncensored is not a product you subscribe to. It's a configuration of sovereignty — deployed once, owned forever.