Instructions to use Vortex5/Prototype-X-12b with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use Vortex5/Prototype-X-12b with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="Vortex5/Prototype-X-12b") messages = [ {"role": "user", "content": "Who are you?"}, ] pipe(messages)# Load model directly from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("Vortex5/Prototype-X-12b") model = AutoModelForCausalLM.from_pretrained("Vortex5/Prototype-X-12b") messages = [ {"role": "user", "content": "Who are you?"}, ] inputs = tokenizer.apply_chat_template( messages, add_generation_prompt=True, tokenize=True, return_dict=True, return_tensors="pt", ).to(model.device) outputs = model.generate(**inputs, max_new_tokens=40) print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:])) - Notebooks
- Google Colab
- Kaggle
- Local Apps
- vLLM
How to use Vortex5/Prototype-X-12b with vLLM:
Install from pip and serve model
# Install vLLM from pip: pip install vllm # Start the vLLM server: vllm serve "Vortex5/Prototype-X-12b" # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:8000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "Vortex5/Prototype-X-12b", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }'Use Docker
docker model run hf.co/Vortex5/Prototype-X-12b
- SGLang
How to use Vortex5/Prototype-X-12b with SGLang:
Install from pip and serve model
# Install SGLang from pip: pip install sglang # Start the SGLang server: python3 -m sglang.launch_server \ --model-path "Vortex5/Prototype-X-12b" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "Vortex5/Prototype-X-12b", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }'Use Docker images
docker run --gpus all \ --shm-size 32g \ -p 30000:30000 \ -v ~/.cache/huggingface:/root/.cache/huggingface \ --env "HF_TOKEN=<secret>" \ --ipc=host \ lmsysorg/sglang:latest \ python3 -m sglang.launch_server \ --model-path "Vortex5/Prototype-X-12b" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "Vortex5/Prototype-X-12b", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }' - Docker Model Runner
How to use Vortex5/Prototype-X-12b with Docker Model Runner:
docker model run hf.co/Vortex5/Prototype-X-12b
âš¡ Prototype-X-12b âš¡
01 // Overview
Prototype-X-12b is a model merged using a custom flowforge method — it merges KansenSakura-Eclipse-RP-12B and KansenSakura-Radiance-RP-12B, with KansenSakura-Erosion-RP-12B as the base model.
02 // Custom Merge Method
It is a directional, coherence-aware merge algorithm that moves a base model along the weighted consensus direction defined by its donors rather than averaging them directly. Each donor’s influence is determined by its relative energy (the magnitude of its weight differences from the base), and the method normalizes and scales these offsets to preserve numerical stability. A small orthogonal adjustment prevents collapse when donors are highly similar, while the strength, trust, and top_k parameters control how far and how selectively the merge travels through parameter space. The result is a controlled shift in model behavior that reflects donor characteristics without discarding the base model’s underlying structure.
Show YAML
merge_method: flowforge models: - model: Retreatcost/KansenSakura-Eclipse-RP-12b - model: Retreatcost/KansenSakura-Radiance-RP-12b base_model: Retreatcost/KansenSakura-Erosion-RP-12b parameters: strength: 0.8 trust: 1.0 dtype: bfloat16 tokenizer: source: Retreatcost/KansenSakura-Erosion-RP-12b
03 // Acknowledgments
- Team Mradermacher — Static & imatrix quants
- DeathGodlike — EXL3 quants
- Original creators and model authors
- Downloads last month
- 44