Skip to main content

Basic Audio Transcription

Transcribe audio files to text:
import createRvencClient from "./index";
import fs from "fs";

const client = await createRvencClient({
  apiKey: "your-api-key",
});

const audioFile = fs.createReadStream("./audio.mp3");

const transcription = await client.audio.transcriptions.create({
  file: audioFile,
  model: "openai/whisper-large-v3",
});

console.log(transcription.text);

Transcription with Language Specification

Improve accuracy by specifying the language:
const transcription = await client.audio.transcriptions.create({
  file: audioFile,
  model: "openai/whisper-large-v3",
  language: "en",
});

console.log(transcription.text);

Transcription with Timestamps

Get word-level or segment-level timestamps:
const transcription = await client.audio.transcriptions.create({
  file: audioFile,
  model: "openai/whisper-large-v3",
  language: "en",
  response_format: "verbose_json",
  timestamp_granularities: ["word", "segment"],
});

console.log(transcription);

Transcription with Custom Prompt

Provide context to improve accuracy:
const transcription = await client.audio.transcriptions.create({
  file: audioFile,
  model: "openai/whisper-large-v3",
  prompt: "This is a medical consultation discussing patient symptoms and treatment options.",
});

console.log(transcription.text);