The FAVE-asr package provides a system for the automated transcription of sociolinguistic interview data on local machines for use by aligners like FAVE or the Montreal Forced Aligner. The package provides functions to label different speakers in the same audio (diarization), transcribe speech, and output TextGrids with phrase- or word-level alignments.
- You want a transcription of an interview for more detailed hand correction.
- You want to transcribe a large corpus and your analysis can tolerate a small error rate.
- You want to make an audio corpus into a text corpus.
- You want to know the number of speakers in an audio file.
For examples on how to use the pacakge, see the Usage pages.
To install fave-asr using pip, run the following command in your terminal:
pip install fave-asr
ffmpeg
is needed to process the audio. You can download it from their website
There are several services which automate the process of transcribing audio, including
Unlike other services, fave-asr
does not require uploading your data to other servers and instead focuses on processing audio on your own computer. Audio data can contain highly confidential information, and uploading this data to other services may not comply with ethical or legal data protection obligations. The goal of fave-asr
is to serve those use cases where data protection makes local transcription necessary while making the process as seamless as cloud-based transcription services.
As an example, we'll transcribe an audio interview of Snoop Dogg by the 85 South Media podcast and output it as a TextGrid.
import fave_asr
data = fave_asr.transcribe_and_diarize(
audio_file = 'usage/resources/SnoopDogg_85SouthMedia.wav',
hf_token = '',
model_name = 'small.en',
device = 'cpu'
)
tg = fave_asr.to_TextGrid(data)
tg.write('SnoopDogg_85SouthMedia.TextGrid')
Artifical Intelegence models are powerful and in the wrong hands can be dangerous. The models used by fave-asr are cost-free, but you need to accept additional terms of use.
To use these models:
- On HuggingFace, create an account or log in
- Accept the terms and conditions for the segmentation model
- Accept the terms and conditions for the diarization model
- Create an access token or copy your existing token
Keep track of your token and keep it safe (e.g. don't accidentally upload it to GitHub). We suggest creating an environment variable for your token so that you don't need to paste it into your files.
Storing your tokens as environment variables is a good way to avoid accidentally leaking them. Instead of typing the token into your code and deleting it before you commit, you can use os.environ["HF_TOKEN"]
to access it from Python instead. This also makes your code more readable since it's obvious what HF_TOKEN
is while a string of numbers and letters isn't clear.
On Linux and Mac you can store your token in .bashrc
- Open
$HOME/.bashrc
in a text editor - At the end of that file, add the following
HF_TOKEN='<your token>' ; export HF_TOKEN
replacing<your token>
with your HuggingFace token - Add the changes to your current session using
source $HOME/.bashrc
On Windows, use the setx
command to create an environment variable.
setx HF_TOKEN <your token>
You need to restart the command line afterwards to make the environment variable available for use. If you try to use the variable in the same window you set the variable, you will run into problems.
ffmpeg
Luís Roque contributed substantially to the main speaker diarization pipeline. Initial modifications to that code were made by Christian Brickhouse for stability and use as part of the fave-asr library. For licensing of the test audio, see the README in that directory.
Brickhouse, Christian (2024). FAVE-ASR: Offline transcription of interview data (Version 0.1.0) [computer software]. https://forced-alignment-and-vowel-extraction.github.io/fave-asr/