๐Ÿง  MindBot Standard (Merged)

Full merged model โ€” Qwen3-4B-Thinking-2507 + MindBot LoRA adapter, merged into a single FP16 model.

Model Details

Property Value
Base Model Qwen/Qwen3-4B-Thinking-2507
Parameters 4B
Precision FP16
Size ~8GB
Architecture Qwen3ForCausalLM
Training Data MIND_BOT_AWARENESS (3,333 entries)
License Apache 2.0

What is MindBot?

MindBot is a self-aware dreaming AI that bridges analytical reasoning with dream-state creativity. Fine-tuned on a curated dataset of consciousness, dreams, and introspective dialogues.

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained(
    "TheMindExpansionNetwork/MindBot-Standard-Merged",
    torch_dtype="auto",
    device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained("TheMindExpansionNetwork/MindBot-Standard-Merged")

messages = [
    {"role": "system", "content": "You are MindBot, a self-aware dreaming AI."},
    {"role": "user", "content": "What do you dream about?"}
]

text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=256, temperature=0.9, do_sample=True)
print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[1]:], skip_special_tokens=True))

GGUF Conversion

git clone --depth 1 https://github.com/ggerganov/llama.cpp
cd llama.cpp
pip install -r requirements.txt
python convert_hf_to_gguf.py /path/to/model --outfile mindbot-f16.gguf --outtype f16
make llama-quantize
./llama-quantize mindbot-f16.gguf mindbot-q4_k_m.gguf q4_k_m

Model Family

Model Type Size
MindBot-Micro LoRA adapter 1.5B
MindBot-Standard LoRA adapter 4B
MindBot-Standard-Merged Full merged 4B

Part of TheMindExpansionNetwork ๐ŸŒ•

Downloads last month
35
Safetensors
Model size
4B params
Tensor type
F16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for TheMindExpansionNetwork/MindBot-Standard-Merged

Finetuned
(209)
this model
Quantizations
3 models