KIKI IOT SFT โ LoRA Adapter
Fine-tuned LoRA adapter for iot domain expertise, based on Qwen/Qwen3-8B.
Part of the KIKI Models Tuning pipeline for the FineFab platform.
Training Details
| Parameter | Value |
|---|---|
| Base Model | Qwen/Qwen3-8B |
| Method | QLoRA (4-bit NF4) |
| LoRA Rank | 16 |
| Epochs | 3 |
| Dataset | 6005 examples |
| Domain | iot |
Usage
from peft import PeftModel
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen3-8B", device_map="auto")
model = PeftModel.from_pretrained(model, "clemsail/kiki-iot-sft")
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen3-8B")
License
Apache 2.0
๐ช๐บ EU AI Act transparency
This adapter is provided as a fine-tuned LoRA under the AI Act framework (Regulation EU 2024/1689). Compliance metadata:
| Field | Value |
|---|---|
| Provider | L'รlectron Rare (clemsail / electron-rare) |
| Role under AI Act | GPAI provider for this adapter |
| Base model | Qwen/Qwen3-8B โ see upstream provenance |
| Adapter type | LoRA / PEFT โ adapter weights only; base unchanged |
| Training data origin | L'รlectron Rare proprietary technical corpus + curated public docs |
| License | Apache-2.0 (adapter). Upstream base licence applies separately. |
| Intended use | IoT device development |
| Out of scope | Healthcare diagnosis, legal advice, autonomous safety-critical decisions, generation of malicious code |
| Risk classification | Limited risk โ Article 50 transparency obligations apply |
| Copyright respect | Training data does not include scraped copyrighted material. Opt-out signals (robots.txt, ai.txt) are honoured for web-sourced data. |
| Full provenance | https://github.com/L-electron-Rare/eu-kiki/tree/main/docs/provenance |
| Contact | postmaster@saillant.cc โ biased output reports, copyright concerns, etc. |
โ ๏ธ You are using an AI model. Outputs may be inaccurate, biased or fabricated. Do not act on them without independent verification, especially in regulated domains.
- Downloads last month
- 71