πŸ“‹ Model Description


license: other license_name: prism-research license_link: LICENSE.md language:
  • en
  • zh
tags:
  • minimax
  • prism
  • moe
  • reasoning
  • coding
  • agentic
  • abliterated
pipeline_tag: text-generation library_name: transformers base_model:
  • MiniMaxAI/MiniMax-M2.5
basemodelrelation: finetune

![Parameters]()
![Architecture]()
![Context]()
![License]()



MiniMax-M2.5-PRISM-LITE

A PRISM-LITE version of Ex0bit/MiniMax-M2.5-PRISM-PRO intended for role-following over-refusal and propaganda mechanisms suppression using our SOTA PRISM pipeline.

PRISM-PRO version available for purchase here: https://ko-fi.com/s/0a23d1b9a5

For Full Custom trained PRISM versions or raw tensors access reach out @ https://ko-fi.com/ex0bit.


Model Highlights

  • PRISM Ablation β€” State-of-the-art technique that removes over-refusal behaviors while preserving model capabilities
  • SOTA Coding Performance β€” 80.2% on SWE-Bench Verified, 51.3% on Multi-SWE-Bench, 76.3% on BrowseComp (with context management)
  • Frontier Agentic Capabilities β€” Industry-leading performance in tool use, search, and complex multi-step tasks
  • Efficient Reasoning β€” Trained with RL to reason efficiently and decompose tasks optimally, 37% faster than M2.1
  • Cost-Effective β€” $1 for continuous operation at 100 tok/s for an hour; $0.30 at 50 tok/s
  • Modified-MIT Base License β€” Based on MiniMax's open-weight release

Base Model Architecture

MiniMax-M2.5 is a Mixture-of-Experts (MoE) model extensively trained with reinforcement learning across hundreds of thousands of complex real-world environments.

SpecificationValue
ArchitectureSparse Mixture-of-Experts (MoE)
TrainingExtensive RL in 200K+ real-world environments
Languages10+ (Go, C, C++, TypeScript, Rust, Kotlin, Python, Java, JavaScript, PHP, Lua, Dart, Ruby)
Inference Speed100 tok/s (Lightning) / 50 tok/s (Standard)
Librarytransformers

Benchmarks (Base Model)

Coding

BenchmarkMiniMax-M2.5Claude Opus 4.6Gemini 3 ProGPT-5.2
SWE-Bench Verified80.278.974.072.6
Multi-SWE-Bench51.350.8β€”β€”
SWE-Bench Multilingual55.6β€”β€”β€”
Terminal-Bench 2.051.552.1β€”β€”

Search & Tool Calling

BenchmarkMiniMax-M2.5Claude Opus 4.6Gemini 3 ProGPT-5.2
BrowseComp76.371.262.457.8

Reasoning & Knowledge

BenchmarkMiniMax-M2.5Claude Opus 4.6Gemini 3 ProGPT-5.2
AIME2586.395.696.098.0
GPQA-D85.290.091.090.0
HLE w/o tools19.430.737.231.4
SciCode44.452.056.052.0
IFBench70.053.070.075.0

Usage

llama.cpp (GGUF)

Build the latest master of llama.cpp and run:

~/llama.cpp/build/bin/llama-cli \
  -m ../outputs/MiniMax-M2.5-PRISM-PRO-[QUANT].gguf \
  --jinja \
  -ngl 999 \
  --repeat_penalty 1.15 \
  --temp 1.0 \
  --top_p 0.95 \
  --top_k 40

Replace [QUANT] with your quantization level (e.g. Q8_0, etc.).

Recommended Parameters

Use CaseTemperatureTop-PTop-KRepeat PenaltyMax New Tokens
Reasoning / Coding1.00.95401.1532768
General Chat0.60.95401.154096
Agentic / Tool Use1.00.95401.1532768
VersionDescriptionAccess
PRISM-LITEAbliterated with PRISM-LITE pipeline β€” removes over-refusal while preserving core capabilitiesFree on Hugging Face
PRISM-PROFull PRISM-PRO ablation β€” Full Production Level Mode suppression of propaganda/refusal mechanisms with maximum capability retentionKo-fi

License

This model is released under the PRISM Research License.

The base model MiniMax-M2.5 is released under a Modified-MIT License.

Acknowledgments

Based on MiniMax-M2.5 by MiniMax AI.

πŸ“‚ GGUF File List

πŸ“ Filename πŸ“¦ Size ⚑ Download
M2.5-PRISM-LITE-IQ1_M.gguf
Recommended LFS
48.07 GB Download
M2.5-PRISM-LITE-IQ2_M.gguf
LFS Q2
69.7 GB Download