# padertorch
[![Build Status](https://dev.azure.com/fgnt/fgnt/_apis/build/status/fgnt.padertorch?branchName=master)](https://dev.azure.com/fgnt/fgnt/_build/latest?definitionId=3&branchName=master)
[![Azure DevOps tests](https://img.shields.io/azure-devops/tests/fgnt/fgnt/3/master)](https://dev.azure.com/fgnt/fgnt/_build/latest?definitionId=3&branchName=master)
[![Azure DevOps coverage](https://img.shields.io/azure-devops/coverage/fgnt/fgnt/3/master)](https://dev.azure.com/fgnt/fgnt/_build/latest?definitionId=3&branchName=master)
[![License: MIT](https://img.shields.io/badge/License-MIT-blue.svg)](https://github.com/fgnt/lazy_dataset/blob/master/LICENSE)
Padertorch is designed to simplify the training of deep learning models written with [PyTorch](https://pytorch.org).
While focusing on speech and audio processing, it is not limited to these application areas.
[//]: <> (This repository is currently under construction.)
[//]: <> (The examples in contrib/examples are only working in the Paderborn NT environment)
![Schematic overview of padertorch](doc/padertorch.svg)
# Highlights
- **Easily extensible**: Write your own network modules and models based on `padertorch.Module` and `padertorch.Model`.
- **Seamless integration**: You provide your own data and model. We provide the trainer to wrap around your model and data. Calling `train` starts the training.
- **Fast prototyping**: The trainer and models are all compatible with [sacred](https://github.com/IDSIA/sacred) and easily allow hyperparameter changes over the command line. Check out the [configurable](padertorch/configurable.py) module and the [examples](padertorch/contrib/examples) for how it works.
- **Trainer**: Our `padertorch.Trainer` takes care of the repetitive training loop and allows you to focus on network tuning. Included are features such as:
- Periodically executed validation runs
- <details>
<summary>Automatic checkpointing</summary><p>
The parameters of the model and the state of the trainer are periodically saved. The checkpoint interval and number of total checkpoints are customizable: Keep one, some or all checkpoints. We also keep track of the best checkpoint on the validation data given some metric.
- Resume from the latest checkpoint if the training was interrupted.
</p></details>
- Learning rate scheduling
- Backoff: Restore the best checkpoint and change the learning rate if the loss is not decreasing.
- [Hooks](padertorch/train/hooks.py): Extend the basic features of the trainer with your own functionality.
- <details>
<summary><b>Test run</b></summary><p>
The trainer has a `test_run` function to train the model for few iterations and test if
- the model is executable (burn test),
- the validation is deterministic/reproducible, and
- the model changes the parameter during training.
<p></details>
- <details>
<summary><b>Virtual minibatch</b></summary><p>
- The `Trainer` usually does not know if the model is trained with a single example or multiple examples (minibatch), because the examples that are yielded from the dataset are directly forwarded to the model.
- When the `virtual_minibatch_size` option is larger than one, the trainer calls the forward and backward step `virtual_minibatch_size` times before applying the gradients. This increases the number of examples over which the gradient is calculated while the memory consumption stays similar, e.g., with `virtual_minibatch_size=2` the gradient is accumulated over two (batched) examples before calling `optimizer.step(); optimizer.zero_grad()`. See [here](doc/virtual_batch_size_multi_gpu.md) for a more thorough explanation.
</p></details>
- **Logging**: As logging backend, we use [tensorboardX](https://github.com/lanpa/tensorboardX) to generate a `tfevents` file that can be visualized from a [tensorboard](https://github.com/tensorflow/tensorboard). Custom values to be logged can be defined in subclasses of `padertorch.Model`.
- **Multi-GPU training**: Easily deploy your model onto multiple GPUs to increase the total batch size and speed up the training. See [here](doc/virtual_batch_size_multi_gpu.md#L68) for implementation details and the [example](padertorch/contrib/examples/multi_gpu) for how to enable it.
- **Support for lazy data preparation**: Ever tired of pre-computing features, taking up time and space on the hard disk? Padertorch works with lazy dataloaders (e.g., [lazy_dataset](https://github.com/fgnt/lazy_dataset)) which extract the features on the fly!
- **Hands-on examples**: We provide models and training scripts which help you getting started with padertorch.
# Installation
```bash
$ git clone https://github.com/fgnt/padertorch.git
$ cd padertorch && pip install -e .[all]
```
This will install all dependencies.
For a light installation, you can drop `[all]`.
**Requirements**
- Python 3
- matplotlib
- Installed by padertorch (core dependencies):
- torch
- tensorboardX
- einops
- tqdm
- natsort
- [lazy_dataset](https://github.com/fgnt/lazy_dataset)
- IPython
- [paderbox](https://github.com/fgnt/paderbox)
- To run the examples (included in `[all]`):
- [sacred](https://github.com/IDSIA/sacred)
- torchvision
- [pb_bss](http://github.com/fgnt/pb_bss)
# Getting Started
## A Short Explanation of `padertorch.Module` and `padertorch.Model`
You can build your models upon `padertorch.Module` and `padertorch.Model`.
Both expect a `forward` method which has the same functionality as the `forward` call of `torch.nn.Module`: It takes some data as input, applies some transformations, and returns the network output:
```python
class MyModel(pt.Module):
def forward(self, example):
x = example['x']
out = transform(x)
return out
```
Additionally, `padertorch.Model` expects a `review` method to be implemented which takes the input and output of the `forward` call as its inputs from which it computes the training loss and metrics for logging in tensorboard.
The following is an example for a classification problem using the cross-entropy loss:
```python
import torch
class MyModel(pt.Model):
def forward(self, example):
output = ... # e.g., output has shape (N, C), where C is the number of classes
return output
def review(self, example, output):
# loss computation, where example['label'] has shape (N,)
ce_loss = torch.nn.CrossEntropyLoss()(output, example['label'])
# compute additional metrics
with torch.no_grad():
prediction = torch.argmax(output, dim=1)
accuracy = (prediction == example['label']).float().mean()
return {
'loss': ce_loss,
'scalars': {'accuracy': accuracy}
}
```
See [padertorch.summary.tbx_utils.review_dict](padertorch/summary/tbx_utils.py#L213) for how the review dictionary should be constructed.
For each training step, the trainer calls `forward`, passes its output to `review` and performs a backpropagation step on the loss.
Typically, the input to the `forward` of a `Module` is a Tensor, while for a `Model`, it is a dictionary which contains additional entries, e.g., labels, which are needed in the `review`.
This is only a recommendation and there is no restriction for the input type.
While these two methods are mandatory, you are free to add any further methods to your models.
Since a `Module` does not need a `review` method, it can be used as a component of a `Model`.
## How to Integrate your Data and Model with the Trainer
The trainer works with any kind of iterable, e.g., `list`, `torch.utils.data.DataLoader` or `lazy_dataset.Dataset`.
The `train` method expects an iterable as input which yields training examples or minibatches of examples that are forwarded to the model without being interpreted by the trainer, i.e., the yielded entries can have any data type and only the model has to be designed to work with them.
In our [examples](padertorch/contrib/examples), the iterables always yield a `dict`.
The `Model` implements an `example_to_device` which is called by the trainer to move the data to a CPU or GPU.
Per default, `example_to_device` uses `padertorch.data.example_to_device` which recursively converts numpy arrays to Tensors and moves all Tensors to the available device.
The training device can be directly provided to the call of `Trainer.train`.
Otherwise, it is automatically set by the trainer according to `torch.cuda.is_available`.
Optionally, you can add an iterable with validation examples by using `Trainer.register_validation_hook`.
Some functionalities (e.g., keeping track of the best checkpoint) are then performed on the validation data.
A simple sketch for the trainer setup is given below:
```python
import torch
import padertorch as pt
train_dataset = ...
validation_dataset = ...
class MyModel(pt.Model):
def __init__(self):
super().__init__()
self.net = torch.nn.Sequential(...)
def forward(self, example):
output = self.net(example['observation'])
return output
def review(self, example, output):
loss = ... # calculate loss
with torch.no_grad():
... # calculate general metrics
if self.training:
... # calculate training specific metrics
else:
... # calculate validation specific metrics
return {
'loss': loss,
'scalars': {
'accuracy': ...,
...
},
} # Furthers keys: 'images', 'audios', 'histograms', 'texts', 'figures'
trainer = padertorch.Trainer(
model=MyModel(),
storage_dir=pt.io.get_new_storage_dir('my_experiment'), # checkpoints of the trained model are stored here
optimizer=pt.train.optimizer.Adam(),
loss_weights=None,
summary_trigger=(1, 'epoch'),
checkpoint_trigger=(1, 'epoch'),
stop_trigger=(1, 'epoch'),
virtual_minibatch_size=1,
)
trainer.test_run(train_dataset, validation_dataset)
trainer.register_validation_hook(validation_dataset)
trainer.train(train_dataset)
```
See the [trainer](padertorch/train/trainer.py#L40) for an explanation of its signature.
If you want to use `pt.io.get_new_storage_dir` to manage your experiments, you have to define an environment variable `STORAGE_ROOT` which points to the path where all your experiments will be stored, i.e., in the example above, a new directory under `$STORAGE_ROOT/my_experiment/1` will be created.
Otherwise, you can use `pt.io.get_new_subdir` where you can directly input the path to store your model without defining an environment variable.
# Features for Application in Deep Learning
Padertorch provides a selection of frequently used network architectures and functionalities such as activation and normalization, ready for you to integrate into your own models.
- [Multi-layer Feed-Forward](padertorch/modules/fully_connected.py): Multiple fully-connected layers with non-linearity and dropout.
- [CNN](padertorch/contrib/je/modules/conv.py) (currently subject to breaking changes and hardly any documentation): 1d- and 2d-CNNs with residual connections, dropout, gated activations, batch and sequence norm and correct handling of down- and upsampling.
- [Normalization](padertorch/modules/normalization.py): Perform normalization of arbitrary axes/dimensions of your features, keep track of running statistics and apply learnable affine transformation.
- [Collection of activation functions](padertorch/ops/mappings.py): Fast access of various activation functions with just a string.
- [Losses](padertorch/ops/losses): We provide an implementation of the [Kullback-Leibler divergence](padertorch/ops/losses/kl_divergence.py) and different [regression](padertorch/ops/losses/regression.py) objectives (SI-SNR, log-MSE and others).
## Support for Sequential and Speech Data
Padertorch especially offers support for training with [sequential data](padertorch/ops/sequence) such as:
- [Masking](padertorch/ops/sequence/mask.py): Calculate a mask which has non-zero entries for non-padded positions and zero entries for padded positions in the sequence.
- [Segmenting](padertorch/data/segment.py): Segment a sequence into smaller chunks of the same length.
- [Dual-Path RNN (DPRNN)](padertorch/modules/dual_path_rnn.py): See the [paper](https://arxiv.org/abs/1910.06379).
- [WaveNet](padertorch/modules/wavenet): Synthesize waveforms from spectrograms (supports fast inference with [nv_wavenet](https://github.com/NVIDIA/nv-wavenet)).
- [Visualization in tensorboard](padertorch/summary/tbx_utils.py): Prepare spectrograms and speech masks for visualization and audio for playback in tensorboard.
We also provide a [loss wrapper for permutation-invariant training (PIT)](padertorch/ops/losses/source_separation.py#L34) criteria which are, e.g., commonly used in (speech) source separation.
# Further Reading
Have a look at the following links to get the most out of your experience with padertorch:
- [Configurable](doc/configurable.md): Provides a thorough explanation of our [configurable](padertorch/configurable.py) module which allows to create model instances from a config dict.
- [Sacred](doc/sacred.md): Explains how to use sacred in combination with `pt.Configurable`.
- [contrib/](padertorch/contrib): Unordered collection of advanced and experimental features. Subject to breaking changes so be careful with relying on it too much. But it might provide ideas for your own implementations.
- [contrib/examples/](padertorch/contrib/examples): Shows advanced usage of padertorch with actual data and models. A good starting point to get ideas for writing own models and experiments. See [here](doc/examples.md) for a guide to navigate through the examples and recommendations for which examples to start with.
Raw data
{
"_id": null,
"home_page": "https://github.com/fgnt/padertorch/",
"name": "padertorch",
"maintainer": null,
"docs_url": null,
"requires_python": null,
"maintainer_email": null,
"keywords": "pytorch, audio, speech",
"author": "Department of Communications Engineering, Paderborn University",
"author_email": "sek@nt.upb.de",
"download_url": "https://files.pythonhosted.org/packages/3c/b4/974065542ad1b077e57fae3f8e3ad1dc3e185eedc5ecf0c340eb19edaabc/padertorch-0.0.1.tar.gz",
"platform": null,
"description": "# padertorch\n[![Build Status](https://dev.azure.com/fgnt/fgnt/_apis/build/status/fgnt.padertorch?branchName=master)](https://dev.azure.com/fgnt/fgnt/_build/latest?definitionId=3&branchName=master)\n[![Azure DevOps tests](https://img.shields.io/azure-devops/tests/fgnt/fgnt/3/master)](https://dev.azure.com/fgnt/fgnt/_build/latest?definitionId=3&branchName=master)\n[![Azure DevOps coverage](https://img.shields.io/azure-devops/coverage/fgnt/fgnt/3/master)](https://dev.azure.com/fgnt/fgnt/_build/latest?definitionId=3&branchName=master)\n[![License: MIT](https://img.shields.io/badge/License-MIT-blue.svg)](https://github.com/fgnt/lazy_dataset/blob/master/LICENSE)\n\nPadertorch is designed to simplify the training of deep learning models written with [PyTorch](https://pytorch.org).\nWhile focusing on speech and audio processing, it is not limited to these application areas.\n\n[//]: <> (This repository is currently under construction.)\n\n[//]: <> (The examples in contrib/examples are only working in the Paderborn NT environment)\n\n![Schematic overview of padertorch](doc/padertorch.svg)\n\n# Highlights\n\n- **Easily extensible**: Write your own network modules and models based on `padertorch.Module` and `padertorch.Model`.\n- **Seamless integration**: You provide your own data and model. We provide the trainer to wrap around your model and data. Calling `train` starts the training.\n- **Fast prototyping**: The trainer and models are all compatible with [sacred](https://github.com/IDSIA/sacred) and easily allow hyperparameter changes over the command line. Check out the [configurable](padertorch/configurable.py) module and the [examples](padertorch/contrib/examples) for how it works.\n- **Trainer**: Our `padertorch.Trainer` takes care of the repetitive training loop and allows you to focus on network tuning. Included are features such as:\n - Periodically executed validation runs\n - <details>\n <summary>Automatic checkpointing</summary><p>\n\n The parameters of the model and the state of the trainer are periodically saved. The checkpoint interval and number of total checkpoints are customizable: Keep one, some or all checkpoints. We also keep track of the best checkpoint on the validation data given some metric.\n - Resume from the latest checkpoint if the training was interrupted.\n\n </p></details>\n\n - Learning rate scheduling\n - Backoff: Restore the best checkpoint and change the learning rate if the loss is not decreasing.\n - [Hooks](padertorch/train/hooks.py): Extend the basic features of the trainer with your own functionality.\n - <details>\n <summary><b>Test run</b></summary><p>\n\n The trainer has a `test_run` function to train the model for few iterations and test if\n - the model is executable (burn test),\n - the validation is deterministic/reproducible, and\n - the model changes the parameter during training.\n\n <p></details>\n\n - <details>\n <summary><b>Virtual minibatch</b></summary><p>\n\n - The `Trainer` usually does not know if the model is trained with a single example or multiple examples (minibatch), because the examples that are yielded from the dataset are directly forwarded to the model.\n - When the `virtual_minibatch_size` option is larger than one, the trainer calls the forward and backward step `virtual_minibatch_size` times before applying the gradients. This increases the number of examples over which the gradient is calculated while the memory consumption stays similar, e.g., with `virtual_minibatch_size=2` the gradient is accumulated over two (batched) examples before calling `optimizer.step(); optimizer.zero_grad()`. See [here](doc/virtual_batch_size_multi_gpu.md) for a more thorough explanation.\n\n </p></details>\n\n- **Logging**: As logging backend, we use [tensorboardX](https://github.com/lanpa/tensorboardX) to generate a `tfevents` file that can be visualized from a [tensorboard](https://github.com/tensorflow/tensorboard). Custom values to be logged can be defined in subclasses of `padertorch.Model`.\n- **Multi-GPU training**: Easily deploy your model onto multiple GPUs to increase the total batch size and speed up the training. See [here](doc/virtual_batch_size_multi_gpu.md#L68) for implementation details and the [example](padertorch/contrib/examples/multi_gpu) for how to enable it.\n- **Support for lazy data preparation**: Ever tired of pre-computing features, taking up time and space on the hard disk? Padertorch works with lazy dataloaders (e.g., [lazy_dataset](https://github.com/fgnt/lazy_dataset)) which extract the features on the fly!\n- **Hands-on examples**: We provide models and training scripts which help you getting started with padertorch.\n\n# Installation\n\n```bash\n$ git clone https://github.com/fgnt/padertorch.git\n$ cd padertorch && pip install -e .[all]\n```\nThis will install all dependencies.\nFor a light installation, you can drop `[all]`.\n\n**Requirements**\n- Python 3\n- matplotlib\n- Installed by padertorch (core dependencies):\n - torch\n - tensorboardX\n - einops\n - tqdm\n - natsort\n - [lazy_dataset](https://github.com/fgnt/lazy_dataset)\n - IPython\n - [paderbox](https://github.com/fgnt/paderbox)\n- To run the examples (included in `[all]`):\n - [sacred](https://github.com/IDSIA/sacred)\n - torchvision\n - [pb_bss](http://github.com/fgnt/pb_bss)\n\n# Getting Started\n\n## A Short Explanation of `padertorch.Module` and `padertorch.Model`\n\nYou can build your models upon `padertorch.Module` and `padertorch.Model`.\nBoth expect a `forward` method which has the same functionality as the `forward` call of `torch.nn.Module`: It takes some data as input, applies some transformations, and returns the network output:\n\n```python\nclass MyModel(pt.Module):\n\n def forward(self, example):\n x = example['x']\n out = transform(x)\n return out\n```\n\nAdditionally, `padertorch.Model` expects a `review` method to be implemented which takes the input and output of the `forward` call as its inputs from which it computes the training loss and metrics for logging in tensorboard.\nThe following is an example for a classification problem using the cross-entropy loss:\n\n```python\nimport torch\n\nclass MyModel(pt.Model):\n\n def forward(self, example):\n output = ... # e.g., output has shape (N, C), where C is the number of classes\n return output\n\n def review(self, example, output):\n # loss computation, where example['label'] has shape (N,)\n ce_loss = torch.nn.CrossEntropyLoss()(output, example['label'])\n # compute additional metrics\n with torch.no_grad():\n prediction = torch.argmax(output, dim=1)\n accuracy = (prediction == example['label']).float().mean()\n return {\n 'loss': ce_loss,\n 'scalars': {'accuracy': accuracy}\n }\n```\n\nSee [padertorch.summary.tbx_utils.review_dict](padertorch/summary/tbx_utils.py#L213) for how the review dictionary should be constructed.\nFor each training step, the trainer calls `forward`, passes its output to `review` and performs a backpropagation step on the loss.\nTypically, the input to the `forward` of a `Module` is a Tensor, while for a `Model`, it is a dictionary which contains additional entries, e.g., labels, which are needed in the `review`.\nThis is only a recommendation and there is no restriction for the input type.\n\nWhile these two methods are mandatory, you are free to add any further methods to your models.\nSince a `Module` does not need a `review` method, it can be used as a component of a `Model`.\n\n## How to Integrate your Data and Model with the Trainer\n\nThe trainer works with any kind of iterable, e.g., `list`, `torch.utils.data.DataLoader` or `lazy_dataset.Dataset`.\nThe `train` method expects an iterable as input which yields training examples or minibatches of examples that are forwarded to the model without being interpreted by the trainer, i.e., the yielded entries can have any data type and only the model has to be designed to work with them.\nIn our [examples](padertorch/contrib/examples), the iterables always yield a `dict`.\n\nThe `Model` implements an `example_to_device` which is called by the trainer to move the data to a CPU or GPU.\nPer default, `example_to_device` uses `padertorch.data.example_to_device` which recursively converts numpy arrays to Tensors and moves all Tensors to the available device.\nThe training device can be directly provided to the call of `Trainer.train`.\nOtherwise, it is automatically set by the trainer according to `torch.cuda.is_available`.\n\nOptionally, you can add an iterable with validation examples by using `Trainer.register_validation_hook`.\nSome functionalities (e.g., keeping track of the best checkpoint) are then performed on the validation data.\n\nA simple sketch for the trainer setup is given below:\n\n```python\nimport torch\nimport padertorch as pt\n\ntrain_dataset = ...\nvalidation_dataset = ...\n\nclass MyModel(pt.Model):\n def __init__(self):\n super().__init__()\n self.net = torch.nn.Sequential(...)\n\n def forward(self, example):\n output = self.net(example['observation'])\n return output\n\n def review(self, example, output):\n loss = ... # calculate loss\n with torch.no_grad():\n ... # calculate general metrics\n if self.training:\n ... # calculate training specific metrics\n else:\n ... # calculate validation specific metrics\n return {\n 'loss': loss,\n 'scalars': {\n 'accuracy': ...,\n ...\n },\n } # Furthers keys: 'images', 'audios', 'histograms', 'texts', 'figures'\n\n\ntrainer = padertorch.Trainer(\n model=MyModel(),\n storage_dir=pt.io.get_new_storage_dir('my_experiment'), # checkpoints of the trained model are stored here\n optimizer=pt.train.optimizer.Adam(),\n loss_weights=None,\n summary_trigger=(1, 'epoch'),\n checkpoint_trigger=(1, 'epoch'),\n stop_trigger=(1, 'epoch'),\n virtual_minibatch_size=1,\n)\ntrainer.test_run(train_dataset, validation_dataset)\ntrainer.register_validation_hook(validation_dataset)\ntrainer.train(train_dataset)\n```\n\nSee the [trainer](padertorch/train/trainer.py#L40) for an explanation of its signature.\nIf you want to use `pt.io.get_new_storage_dir` to manage your experiments, you have to define an environment variable `STORAGE_ROOT` which points to the path where all your experiments will be stored, i.e., in the example above, a new directory under `$STORAGE_ROOT/my_experiment/1` will be created.\nOtherwise, you can use `pt.io.get_new_subdir` where you can directly input the path to store your model without defining an environment variable.\n\n# Features for Application in Deep Learning\n\nPadertorch provides a selection of frequently used network architectures and functionalities such as activation and normalization, ready for you to integrate into your own models.\n\n- [Multi-layer Feed-Forward](padertorch/modules/fully_connected.py): Multiple fully-connected layers with non-linearity and dropout.\n- [CNN](padertorch/contrib/je/modules/conv.py) (currently subject to breaking changes and hardly any documentation): 1d- and 2d-CNNs with residual connections, dropout, gated activations, batch and sequence norm and correct handling of down- and upsampling.\n- [Normalization](padertorch/modules/normalization.py): Perform normalization of arbitrary axes/dimensions of your features, keep track of running statistics and apply learnable affine transformation.\n- [Collection of activation functions](padertorch/ops/mappings.py): Fast access of various activation functions with just a string.\n- [Losses](padertorch/ops/losses): We provide an implementation of the [Kullback-Leibler divergence](padertorch/ops/losses/kl_divergence.py) and different [regression](padertorch/ops/losses/regression.py) objectives (SI-SNR, log-MSE and others).\n\n## Support for Sequential and Speech Data\n\nPadertorch especially offers support for training with [sequential data](padertorch/ops/sequence) such as:\n- [Masking](padertorch/ops/sequence/mask.py): Calculate a mask which has non-zero entries for non-padded positions and zero entries for padded positions in the sequence.\n- [Segmenting](padertorch/data/segment.py): Segment a sequence into smaller chunks of the same length.\n- [Dual-Path RNN (DPRNN)](padertorch/modules/dual_path_rnn.py): See the [paper](https://arxiv.org/abs/1910.06379).\n- [WaveNet](padertorch/modules/wavenet): Synthesize waveforms from spectrograms (supports fast inference with [nv_wavenet](https://github.com/NVIDIA/nv-wavenet)).\n- [Visualization in tensorboard](padertorch/summary/tbx_utils.py): Prepare spectrograms and speech masks for visualization and audio for playback in tensorboard.\n\nWe also provide a [loss wrapper for permutation-invariant training (PIT)](padertorch/ops/losses/source_separation.py#L34) criteria which are, e.g., commonly used in (speech) source separation.\n\n# Further Reading\n\nHave a look at the following links to get the most out of your experience with padertorch:\n\n- [Configurable](doc/configurable.md): Provides a thorough explanation of our [configurable](padertorch/configurable.py) module which allows to create model instances from a config dict.\n- [Sacred](doc/sacred.md): Explains how to use sacred in combination with `pt.Configurable`.\n- [contrib/](padertorch/contrib): Unordered collection of advanced and experimental features. Subject to breaking changes so be careful with relying on it too much. But it might provide ideas for your own implementations.\n- [contrib/examples/](padertorch/contrib/examples): Shows advanced usage of padertorch with actual data and models. A good starting point to get ideas for writing own models and experiments. See [here](doc/examples.md) for a guide to navigate through the examples and recommendations for which examples to start with.\n",
"bugtrack_url": null,
"license": "MIT",
"summary": "A collection of common functionality to simplify the design, training and evaluation of machine learning models based on pytorch with an emphasis on speech processing.",
"version": "0.0.1",
"project_urls": {
"Homepage": "https://github.com/fgnt/padertorch/"
},
"split_keywords": [
"pytorch",
" audio",
" speech"
],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "3cb4974065542ad1b077e57fae3f8e3ad1dc3e185eedc5ecf0c340eb19edaabc",
"md5": "198bcc515d2617f620cd7b9417894e6a",
"sha256": "d8101d6539bf1dc98f544f58e8a4023ca01ae1858d09db1c466867926c7cec8d"
},
"downloads": -1,
"filename": "padertorch-0.0.1.tar.gz",
"has_sig": false,
"md5_digest": "198bcc515d2617f620cd7b9417894e6a",
"packagetype": "sdist",
"python_version": "source",
"requires_python": null,
"size": 250907,
"upload_time": "2024-06-18T13:53:43",
"upload_time_iso_8601": "2024-06-18T13:53:43.586681Z",
"url": "https://files.pythonhosted.org/packages/3c/b4/974065542ad1b077e57fae3f8e3ad1dc3e185eedc5ecf0c340eb19edaabc/padertorch-0.0.1.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2024-06-18 13:53:43",
"github": true,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"github_user": "fgnt",
"github_project": "padertorch",
"travis_ci": false,
"coveralls": true,
"github_actions": true,
"lcname": "padertorch"
}