AudioQnA Accuracy¶
AudioQnA is an example that demonstrates the integration of Generative AI (GenAI) models for performing question-answering (QnA) on audio scene, which contains Automatic Speech Recognition (ASR) and Text-to-Speech (TTS). The following is the piepline for evaluating the ASR accuracy.
Dataset¶
We evaluate the ASR accuracy on the test set of librispeech dataset, which contains 2620 records of audio and texts.
Metrics¶
We evaluate the WER (Word Error Rate) metric of the ASR microservice.
Evaluation¶
Launch ASR microservice¶
Launch the ASR microserice with the following commands. For more details please refer to doc.
git clone https://github.com/opea-project/GenAIComps
cd GenAIComps
docker build -t opea/whisper:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/asr/whisper/Dockerfile .
# change the name of model by editing model_name_or_path you want to evaluate
docker run -p 7066:7066 --ipc=host -e http_proxy=$http_proxy -e https_proxy=$https_proxy opea/whisper:latest --model_name_or_path "openai/whisper-tiny"
Evaluate¶
Install dependencies:
pip install -r requirements.txt
Evaluate the performance with the LLM:
# validate the offline model
# python offline_eval.py
# validate the online asr microservice accuracy
python online_eval.py
Performance Result¶
Here is the tested result for your reference
WER |
|
---|---|
whisper-large-v2 |
2.87 |
whisper-large |
2.7 |
whisper-medium |
3.45 |