neon-wire-7b

Fine-tuned Qwen2.5-7B-Instruct for SYSBREAK cyberpunk MMO content generation.

Model Description

This model is a LoRA-merged version of Qwen2.5-7B-Instruct, fine-tuned to generate structured JSON content for the SYSBREAK game.

Purpose: You are a cyberpunk news writer for SYSBREAK. Generate in-universe news articles in JSON format. Use ONLY entities from the provided world context. Do NOT reference real-world games, companies, or products. Respond with valid JSON only.

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("neon-wire-7b")
tokenizer = AutoTokenizer.from_pretrained("neon-wire-7b")

messages = [
    {"role": "system", "content": "You are a cyberpunk news writer for SYSBREAK. Generate in-universe news articles in JSON format. Use ONLY entities from the provided world context. Do NOT reference real-world games, companies, or products. Respond with valid JSON only."},
    {"role": "user", "content": "Your prompt here"},
]

text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=1024, temperature=0.85, top_p=0.9)
print(tokenizer.decode(outputs[0][inputs['input_ids'].shape[1]:], skip_special_tokens=True))

Ollama

ollama run neon-wire-7b

Training Details

  • Training examples: 400
  • Training duration: 11.2 minutes
  • Base model: Qwen/Qwen2.5-7B-Instruct
  • LoRA rank: 32
  • LoRA alpha: 64
  • Learning rate: 2e-4
  • Epochs: 3
  • Quantization: QLoRA 4-bit NF4
  • Compute dtype: BF16

License

Apache 2.0 (same as base model)

Downloads last month
24
Safetensors
Model size
8B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for nullvektordom/neon-wire-7b

Base model

Qwen/Qwen2.5-7B
Finetuned
(2788)
this model
Quantizations
2 models