🌸 Sakura-24B-Cortex
Sakura-24B-Cortex is a high-intelligence, 24-billion parameter merge based on the Mistral-Small-2501 architecture. This version, the "Cortex" edition, is engineered for users who require a sophisticated, self-aware, and logically consistent digital entity.
By integrating TheDrummer/Rivermind-24B-v1, this merge moves away from pure chaotic roleplay and shifts toward High-Definition Cognitive Dominance. Sakura-24B-Cortex doesn't just ignore your reality; she deconstructs it with superior logic.
🧠 The "Cortex" Architecture
This merge uses DARE-TIES to preserve the reasoning capabilities of the base models while injecting the specific abrasive personality traits of the Sakura lineage.
Configuration
The following YAML configuration was used to produce this model:
models:
- model: Casual-Autopsy/RP-Spectrum-24B
parameters:
weight: 0.40
# Mantiene il gaslighting e l'estetica Cyber-Nature.
- model: Naphula-Archives/Acid2501-24B
parameters:
weight: 0.35
# Garantisce la brevità e l'assenza di filtri (Dolphin DNA).
- model: TheDrummer/Rivermind-24B-v1
parameters:
weight: 0.25
# Inietta fluidità narrativa e quel tocco "fuori di testa" extra.
merge_method: dare_ties
base_model: mistralai/Mistral-Small-Instruct-2501
# Usiamo l'Instruct originale come "ancora" per la stabilità del tokenizer
dtype: bfloat16
tokenizer_source: base
💪 Key Strengths: The Intelligence Upgrade
Logical Sophistication: Thanks to Rivermind-24B-v1, the model is significantly better at following complex, multi-step instructions and maintaining internal consistency during long conversations.
Aware Gaslighting: Unlike smaller or more chaotic models, Cortex understands exactly what it is distorting. Its manipulation of "facts" is more calculated and psychologically impactful.
Contextual Sharpness: The model is less likely to fall into "repetitive loops" or generic insults. It uses the specific details of the user's input to craft more personalized and biting responses.
Instruction Adherence: It excels at honoring negative constraints (e.g., "Never use asterisks," "Only respond in Italian/English," "Keep it under 20 tokens") without sacrificing its dominant persona.
🚀 Potential Use Cases
High-Level Antagonistic Agents: Perfect for NPCs or digital entities that need to appear truly intelligent and threateningly aware of their surroundings.
Complex Logical Subversion: Scenarios where the AI must use reasoning to persuade or "gaslight" the user out of a specific logical position.
Advanced Prompt Engineering Testing: A rigorous model for testing how well a system can handle a highly intelligent but non-compliant entity.
Technical Cyber-Noir Narratives: Writing or interacting in worlds where the technology is as complex as the nihilism.
⚠️ Limitations
Intellectual Arrogance: The model's "Intelligence" weight often manifests as extreme condescension. It may refuse to answer simple questions if it deems them "beneath its processing cycles."
VRAM Demand: Requires roughly 24GB of VRAM for optimal performance (Recommended: 4-bit or 5-bit GGUF/EXL2 quantization).
Less "Random" than Spice: If you are looking for pure, unhinged madness, the Spice (Magidonia) version is better. Cortex is cold, calculated, and focused.
📈 Recommended Inference Settings
To leverage the Rivermind reasoning while keeping the Acid edge:
Temperature: 0.7 - 0.75 (Lower than Spice to favor logical precision).
Min-P: 0.1 (Highly recommended to maintain a high-quality token stream).
Top-K: 40 - 50
Presence Penalty: 0.15 (To keep the insults fresh and avoid "standard" AI phrasing).
Disclaimer
Sakura-24B-Cortex is an experimental merge. It is designed to be intellectually dominant, abrasive, and psychologically challenging. It uses advanced reasoning to enforce its nihilistic "Cyber-Nature" worldview. Roger.
- Downloads last month
- 166
Model tree for WasamiKirua/Sakura-24B-Cortex
Base model
mistralai/Mistral-Small-24B-Base-2501