You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Acnoryx AI Release

Overview

  • Model family: Qwen/Qwen3.5-0.8B
  • Project name: acnoryx
  • Model size: 0.8b
  • GGUF quantizations: F16, Q8_0, Q5_K_M, Q4_K_M, Q4_0, IQ4_NL, IQ4_XS
  • Domain: acne, acne-prone skin, skincare, and dermatology guidance

App

Default behavior

  • The saved tokenizer/chat template injects a short default system prompt when no system message is provided.
  • That means the model can still understand its identity as Acnoryx AI in chat mode without a long system prompt.
  • For best results, write Vietnamese with full accents, or use natural English.

Prompt examples

  • Tiแบฟng Viแป‡t: Da em nhiแปu mแปฅn viรชm แปŸ mรก, routine hiแป‡n tแบกi chแป‰ cรณ sแปฏa rแปญa mแบทt vร  kem dฦฐแปกng. Em nรชn ฦฐu tiรชn gรฌ trฦฐแป›c?
  • Tiแบฟng Viแป‡t: Kแบฟt quแบฃ quรฉt cแปงa tรดi cรณ mแปฅn ฤ‘แบงu ฤ‘en 32%, mแปฅn mแปง 21%, thรขm mแปฅn 18%. Hรฃy tรณm tแบฏt ฤ‘รบng theo dแปฏ liแป‡u.
  • English: I have oily acne-prone skin with dark marks after breakouts. What should I prioritize first?

Included folders

  • gguf/: GGUF exports for llama.cpp runtimes
  • hf_transformers/: merged Hugging Face Transformers model

Training stack

  • Transformers + PEFT + TRL bf16 LoRA
  • Qwen3.5 hybrid architecture with fast linear path enabled when available

Prompting

  • See PROMPT_TEMPLATE.txt for usage guidance.

Evaluation Snapshot

Release GGUFs were retested on the curated release_eval_v1 set with 58 bilingual questions in both thinking and non-thinking modes.

Quant Think No-Think Avg Notes
Q8_0 86.2% 87.9% 87.0% Best overall score in the current release rerun
Q5_K_M 89.7% 82.8% 86.2% Strong think-mode quality
IQ4_NL 86.2% 86.2% 86.2% Best balanced sub-500 MB option
F16 87.9% 81.0% 84.4% Highest-fidelity source export
IQ4_XS 84.5% 81.0% 82.8% Smaller release option
Q4_K_M 82.8% 81.0% 81.9% Usable but clearly weaker than Q8_0 / Q5_K_M
Q4_0 77.6% 75.9% 76.8% Lowest-quality release quant

Deployment Guidance

  • Recommended default release quant: Q8_0
  • Best size/quality trade-off under 500 MB: IQ4_NL
  • Keep Q4_0 only for constrained experiments, not as a primary deployment target
  • Current release family remains below the older internal 96% gate, so these artifacts should be treated as interim bundles rather than a final quality-signoff build

Test Results

Latest automated GGUF test results are below:

  • acnoryx-0.8b-f16: Think mode 87.9%, No-Think mode 81.0%
  • acnoryx-0.8b-iq4_nl: Think mode 86.2%, No-Think mode 86.2%
  • acnoryx-0.8b-iq4_xs: Think mode 84.5%, No-Think mode 81.0%
  • acnoryx-0.8b-q4_0: Think mode 77.6%, No-Think mode 75.9%
  • acnoryx-0.8b-q4_k_m: Think mode 82.8%, No-Think mode 81.0%
  • acnoryx-0.8b-q5_k_m: Think mode 89.7%, No-Think mode 82.8%
  • acnoryx-0.8b-q8_0: Think mode 86.2%, No-Think mode 87.9%

Full detailed results in results/release_gguf_0.8b/TEST_RESULTS.json.

For cross-family comparison with research quants, see results/COMPARISON.md in the workspace.

Downloads last month
8
GGUF
Model size
0.8B params
Architecture
qwen35
Hardware compatibility
Log In to add your hardware

4-bit

5-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Acnoryx/Airy-Core-0.8B

Quantized
(106)
this model