whisper-large-v3-it

Fine-tuned openai/whisper-large-v3 (1.55B params) for Italian ASR.

Author: Ettore Di Giacinto

Brought to you by the LocalAI team. This model can be used directly with LocalAI.

Usage with LocalAI

This model is ready to use with LocalAI via the whisperx backend.

Save the following as whisperx-large-v3-it.yaml in your LocalAI models directory:

name: whisperx-large-v3-it
backend: whisperx
known_usecases:
  - transcript
parameters:
  model: LocalAI-io/whisper-large-v3-it-ct2-int8
  language: it

Then transcribe audio via the OpenAI-compatible endpoint:

curl http://localhost:8080/v1/audio/transcriptions \
  -H "Content-Type: multipart/form-data" \
  -F file="@audio.mp3" \
  -F model="whisperx-large-v3-it"

Training Details

  • Base model: openai/whisper-large-v3 (1.55B parameters)
  • Dataset: Common Voice 25.0 Italian (173k train, 15k dev, 15k test)
  • Steps: 10,000
  • Precision: bf16 on NVIDIA GB10
  • Note: trained with --no-eval; final WER to be measured separately

Usage

Transformers

from transformers import pipeline

pipe = pipeline("automatic-speech-recognition", model="LocalAI-io/whisper-large-v3-it")
result = pipe("audio.mp3", generate_kwargs={"language": "it", "task": "transcribe"})
print(result["text"])

CTranslate2 / faster-whisper

For optimized CPU inference: LocalAI-io/whisper-large-v3-it-ct2-int8

Links

Downloads last month
40
Safetensors
Model size
2B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for LocalAI-io/whisper-large-v3-it

Finetuned
(834)
this model