You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Can't use tarred_audio_filepaths with use_lhoste: true
[NeMo I 2025-02-07 15:43:40 nemo_logging:393] Creating a Lhotse DynamicBucketingSampler (max_batch_duration=1100.0 max_batch_size=1)
Error executing job with overrides: []
Traceback (most recent call last):
File "train.py", line 31, in main
asr_model = setup_dataloaders(asr_model, cfg)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "train_utils.py", line 150, in setup_dataloaders
asr_model.setup_training_data(cfg.model.train_ds)
File "NeMo/nemo/collections/asr/models/ctc_models.py", line 419, in setup_training_data
* ceil((len(self._train_dl.dataset) / self.world_size) / train_data_config['batch_size'])
^^^^^^^^^^^^^^^^^^^^^^^^^^^
TypeError: object of type 'IterableDatasetWrapper' has no len()
Steps/Code to reproduce bug
I'm based on speech_to_text_finetune.py and I sue the following config:
init_from_nemo_model: ".cache/huggingface/hub/models--nvidia--parakeet-ctc-0.6b/snapshots/16ca39445465932bfbaeb5126933d5ce8bd43a77/parakeet-ctc-0.6b.nemo"# path to nemo modelmodel:
sample_rate: 16000compute_eval_loss: true # eval samples can be very long and exhaust memory. Disable computation of transducer loss during validation/testing with this flag.log_prediction: false # enables logging sample predictions in the output during trainingctc_reduction: 'mean_volume'skip_nan_grad: falseseed: 42train_ds:
sample_rate: ${model.sample_rate}batch_size: 1# must be 1 if using bucketing_batch_sizeshuffle: truepin_memory: truemax_duration: 30.1min_duration: 0.1shuffle_n: 2048num_workers: 8manifest_filepath:
- - bucket1/sharded_manifests/manifest__OP_0..23_CL_.json
- - bucket2/sharded_manifests/manifest__OP_0..23_CL_.jsontarred_audio_filepaths:
- - bucket1/audio__OP_0..23_CL_.tar
- - bucket2/audio__OP_0..23_CL_.taruse_lhotse: truebatch_duration: 1100quadratic_duration: 30num_buckets: 6num_cuts_for_bins_estimate: 10000bucket_buffer_size: 10000shuffle_buffer_size: 10000use_bucketing: true# tarred datasetsis_tarred: false
Environment overview (please complete the following information)
Installed with pip install nemo_toolkit['asr']: v2.2.0rc2
Environment details
PyTorch 2.5.1+cu121
Python 3.11
Additional context
GPU: A100
The text was updated successfully, but these errors were encountered:
Hi @pzelasko , Do you any clue? I followed your recommendations from #10084, but I can't make it work. Is there any example of finetuning using lhoste with tarred datasets?
Describe the bug
Can't use tarred_audio_filepaths with use_lhoste: true
Steps/Code to reproduce bug
I'm based on
speech_to_text_finetune.py
and I sue the following config:Environment overview (please complete the following information)
pip install nemo_toolkit['asr']
: v2.2.0rc2Environment details
Additional context
GPU: A100
The text was updated successfully, but these errors were encountered: