Skip to content

Are different samples in a batch processed independently? #142

Answered by snakers4
RuABraun asked this question in Q&A
Discussion options

You must be logged in to vote

The V3 models, even for large, long files, are to be used chunk-wise:

silero-vad/utils_vad.py

Lines 198 to 211 in f6b1294

model.reset_states()
min_speech_samples = sampling_rate * min_speech_duration_ms / 1000
min_silence_samples = sampling_rate * min_silence_duration_ms / 1000
speech_pad_samples = sampling_rate * speech_pad_ms / 1000
audio_length_samples = len(audio)
speech_probs = []
for current_start_sample in range(0, audio_length_samples, window_size_samples):
chunk = audio[current_start_sample: current_start_sample + window_size_samples]
if len(chunk) < window_size_samples:
chunk = torch.nn.functional.pad(chunk, (0, int

Replies: 1 comment 6 replies

Comment options

You must be logged in to vote
6 replies
@snakers4
Comment options

@RuABraun
Comment options

@snakers4
Comment options

@RuABraun
Comment options

@snakers4
Comment options

Answer selected by snakers4
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants