LLM-Comparison

STT Comparison whisper-large-v3 | whisper-1

🇨🇳简体中文

In this paper, we briefly compare the two models in terms of average processing time, failure rate, and the difference between query length and processing time growth.

Parameter Settings

The three models compared are

models = ["whisper-large-v3", "whisper-1"]

The whisper-large-v3 uses the Groq API and the whisper-1 uses the OpenAI API.The API usage is Python Client Library as follows

groq_client = Groq(
    api_key=os.environ["GROQ_API_KEY"],
)

openai_client = OpenAI(api_key=os.environ.get("OPENAI_API_KEY"))

The test set will generate languages of different lengths and in both Chinese and English using the 20 TTS named.

Test Procedure

The test will traverse the test set one by one using a loop and record the query length, processing time and failure rate. Every 10 queries tested will be paused for 60 seconds to avoid issues such as exceeding the limited request rate. The test set does not include identical text, so theoretically there are no caching issues.

Results

Summary of data:

output4

output5

output6

Model Average time Fail rate Overall query
whisper-large-v3 1.998681701719761 0.0 136.296
whisper-1 0.8740840554237366 0.0 136.296

Summary of data analysis

  1. Average time comparison:
    • whisper-large-v3 average time is 1.9987 seconds.
    • The average time for whisper-1 is 0.8741 seconds.
    • The response time of whisper-1 is significantly lower than that of whisper-large-v3.
  2. Comparison of failure rates:
    • The failure rate of both models is 0%.
  3. Query length vs. time:
    • whisper-large-v3 Time hash is too large and unstable, and the variation relationship is not informative.
    • whisper-1 time does not vary much with query length, remains stable and is short.
  4. Total number of queries:
    • The total number of queries for both models is 136.296.

Summary


LLM-Comparison by Haozhe Li is licensed under CC BY-NC 4.0