Spaces:
Running
on
Zero
Running
on
Zero
| # @package __global__ | |
| # WARNING: This is a base configuration file shared across ALL solvers in AudioCraft | |
| # Please don't update this file directly. Instead use distinct configuration files | |
| # to override the below configuration. | |
| solver: ??? | |
| fsdp: | |
| use: false # should we use FSDP. | |
| param_dtype: float16 # equivalent to autocast_dtype for FSDP. | |
| reduce_dtype: float32 # gradient averaging dtype, float32 will give max stability. | |
| buffer_dtype: float32 # dtype used for buffers, we don't have much buffers, so let's leave it. | |
| sharding_strategy: shard_grad_op # can be shard_grad_op or full_shard. | |
| # full_shard will use less memory but slower ?? | |
| per_block: true # If True, uses nested FSDP. | |
| profiler: | |
| enabled: false | |
| deadlock: | |
| use: false | |
| timeout: 600 | |
| dataset: | |
| batch_size: ??? | |
| num_workers: 10 | |
| segment_duration: null | |
| num_samples: null | |
| return_info: false | |
| shuffle: false | |
| sample_on_duration: true | |
| sample_on_weight: true | |
| min_segment_ratio: 0.5 | |
| train: | |
| num_samples: null | |
| shuffle: true | |
| shuffle_seed: 0 # if you want to sample the data differently. | |
| permutation_on_files: false | |
| valid: | |
| num_samples: null | |
| evaluate: | |
| num_samples: null | |
| generate: | |
| num_samples: null | |
| return_info: true | |
| checkpoint: | |
| save_last: true | |
| save_every: null | |
| keep_last: null | |
| keep_every_states: null | |
| generate: | |
| every: null | |
| path: 'samples' | |
| audio: | |
| format: 'mp3' | |
| strategy: 'clip' | |
| sample_rate: null | |
| lm: | |
| use_sampling: false | |
| temp: 1.0 | |
| top_k: 0 | |
| top_p: 0.0 | |
| evaluate: | |
| every: null | |
| num_workers: 5 | |
| truncate_audio: null | |
| fixed_generation_duration: null # in secs | |
| metrics: | |
| base: true # run default evaluation (e.g. like train/valid stage) | |
| optim: | |
| epochs: ??? | |
| updates_per_epoch: null | |
| lr: ??? | |
| optimizer: ??? | |
| adam: | |
| betas: [0.9, 0.999] | |
| weight_decay: 0. | |
| ema: | |
| use: false # whether to use EMA or not | |
| updates: ${optim.updates_per_epoch} # frequency of updates of the EMA | |
| device: cpu # device for EMA, can be put on GPU if more frequent updates | |
| decay: 0.99 # EMA decay value, if null, no EMA is used | |
| schedule: | |
| lr_scheduler: null | |
| step: | |
| step_size: null | |
| gamma: null | |
| exponential: | |
| lr_decay: null | |
| cosine: | |
| warmup: null | |
| lr_min_ratio: 0.0 | |
| cycle_length: 1.0 | |
| polynomial_decay: | |
| warmup: null | |
| zero_lr_warmup_steps: 0 | |
| end_lr: 0.0 | |
| power: 1 | |
| inverse_sqrt: | |
| warmup: null | |
| warmup_init_lr: 0.0 | |
| linear_warmup: | |
| warmup: null | |
| warmup_init_lr: 0.0 | |