Rock Paper Anything

A series of fine-tuned small models for an open ended version of the classic game. Premise suggested by an 8 year old.

The goal has been to have a model small enough to play the game offline or in a browser and on low end machines. Each version was fine-tuned using QLoRA via Unsloth, meaning only ~1% of the model's parameters were trained, with the rest frozen. The adapter was then merged back into the base weights and quantized to Q4_K_M GGUF format.

Usage

Ollama

ollama create rockpaperanything -f Modelfile
ollama run rockpaperanything '["caterpillar", "halitosis"]'
{"winner": "caterpillar", "loser": "halitosis", "reason": "The caterpillar's transformation from gnat food to butterfly beauty defies even the most persistent bad breath."}

Input / output

Input is best done as a JSON array of two items:

["arcade fire", "pie"]

Output is JSON:

{
    "winner": "arcade fire",
    "loser": "pie",
    "reason": "The Arcade Fire's infectious energy fills the entire venue, making even a pie feel like it needs to dance."
}

Models

  • v1: from Llama-3.2 3B โ€” 1,500 examples (2.02gb as gguf)
  • v2: from Llama-3.2 3B โ€” 2,200 examples (2.02gb as gguf)
  • v3: from Qwen3.5 2B โ€” 2,200 examples (1.27gb as gguf; 1.08gb as mlc)
  • v4: from Qwen3 1.7B โ€” 2,200 examples (1.11gb as gguf; 0.98gb as mlc))
Downloads last month
660
GGUF
Model size
3B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

We're not able to determine the quantization variants.

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for olaughter/rockpaperanything

Finetuned
Qwen/Qwen3-1.7B
Quantized
(256)
this model