LTNtorch


NameLTNtorch JSON
Version 1.0.1 PyPI version JSON
download
home_pagehttps://github.com/bmxitalia/LTNtorch
SummaryLTNtorch: PyTorch implementation of Logic Tensor Networks
upload_time2023-03-28 12:38:35
maintainer
docs_urlNone
authorTommaso Carraro
requires_python>=3.7
licenseMIT
keywords pytorch machine-learning framework neural-symbolic-computing fuzzy-logic
VCS
bugtrack_url
requirements No requirements were recorded.
Travis-CI No Travis.
coveralls test coverage No coveralls.
            ![build status](https://github.com/tommasocarraro/LTNtorch/actions/workflows/build.yml/badge.svg)
[![coverage status](https://coveralls.io/repos/github/tommasocarraro/LTNtorch/badge.svg?branch=main)](https://coveralls.io/github/tommasocarraro/LTNtorch?branch=main)
[![PyPi](https://img.shields.io/pypi/v/LTNtorch.svg)](https://pypi.python.org/pypi/LTNtorch)
[![docs link](https://img.shields.io/badge/docs-github.io-blue)](https://tommasocarraro.github.io/LTNtorch/)
[![MIT license](https://img.shields.io/github/license/Naereen/StrapDown.js.svg)](https://github.com/Naereen/StrapDown.js/blob/master/LICENSE)
![python version](https://img.shields.io/badge/python-3.7|3.8|3.9-blue)
[![DOI BADGE](https://zenodo.org/badge/DOI/10.5281/zenodo.6394282.svg)](https://doi.org/10.5281/zenodo.6394282)

# LTNtorch: PyTorch implementation of Logic Tensor Networks

Welcome to the PyTorch's implementation of [Logic Tensor Networks](https://arxiv.org/abs/2012.13635)!

### Basic idea of the framework

Logic Tensor Network (LTN) is a Neural-Symbolic (NeSy) framework which supports learning of neural networks using the
satisfaction of a first-order logic knowledge base as an objective. In other words, LTN uses logical reasoning on the
knowledge base to guide the learning of a potentially deep neural network. 

The idea of the framework is simple: 
- we have a first-order logic knowledge base containing a set of axioms;
- we have some predicates, functions, or logical constants appearing in these axioms that we want to learn;
- we have some data available that we can use to learn the parameters of those symbols.

The idea is to use the logical axioms as a loss function for our Logic Tensor Network. The objective is to find solutions
in the hypothesis space that maximally satisfy all the axioms contained in our knowledge base.

### Learning in LTN

In LTN, the learnable parameters are contained in the predicates, functions, and possibly learnable logical constants
that appear in the logical axioms of the knowledge base. 

During the forward step of the back-propagation algorithm, LTN
computes the truth values of the logical formulas contained in the knowledge base, using the available data to ground
(or instantiate) the logical formulas. As we have already said, these formulas 
will contain some predicates and functions which are represented as learnable models. 

At the end of the forward phase,
the truth values computed for the formulas are aggregated and used in the loss function. Our objective is to maximize
the aggregated truth value, namely maximally satisfy all the axioms.

During the backward step, the
learnable parameters of predicates, functions, and possibly learnable logical constants are changed in such a way to move
towards a solution in the hypothesis space which better satisfies all the axioms in the knowledge base.

At the end of the training, the parameters of predicates, functions, and constants will have been updated in such a way the 
logical formulas in the knowledge base are maximally satisfied. In particular, the parameters will have been learned by using both
data (to ground the formulas) and logical reasoning (at the loss function).

After learning, it is possible to query predicates and functions on new data which was not available during training. Also,
it is possible to query the truth values of new formulas which were not included in the knowledge base during training. In addition,
if some logical constants have been learned, their parameters can be interpreted as embeddings.

### Real Logic logical language

To make this learning possible, LTN uses a differentiable first-order logic language, called Real Logic, which enables 
the incorporation of data and logic.

Real Logic defines the concept of `grounding` (different from the grounding of logic), which is a mapping from the logical domain (i.e., constants, variables, and logical symbols)
to tensors in the Real field or operations based on tensors. These operations could be, for instance, mathematical functions or learnable neural networks. In other words,
a `grounding`, denoted as 𝒒, is a function which maps a logical symbol into a real tensor or an operation on tensors.

In particular, the grounding is defined as follows. Let us assume that *c* is a constant, *x* is a logical 
variable, *P* is a predicate, and *f* is a logical function:
![Grounding_in LTN](https://github.com/tommasocarraro/LTNtorch/blob/main/images/grounding.png?raw=true)

The `grounding` defines also how the logical connectives (∧, ∨, Β¬, β‡’, ↔) and quantifiers
(βˆ€, βˆƒ) are mapped in the Real domain. In particular, logical connectives are grounded using fuzzy logic semantics, while
quantifiers are grounded using fuzzy aggregators. Please, **carefully** read the paper if you have some doubts on these notions.

Examples of possible groundings are showed in the following figure. In the figure, `friend(Mary, John)` is an
atomic formula (predicate), while `βˆ€x(friend(John, x) β‡’ friend(Mary, x))` is a closed formula (all the variables are
quantified). The letter 𝒒, again, is the grounding, the function which maps the logical domain into the Real domain.

![Grounding_illustration](https://github.com/tommasocarraro/LTNtorch/blob/main/images/framework_grounding.png?raw=true)

### LTN as PyTorch computational graphs

In practice, LTN converts Real Logic formulas (e.g., `βˆ€xβˆƒy(friend(x,y) ∧ italian(y))`, which states that everybody has a friend that is Italian) into [PyTorch](https://www.pytorch.org/) 
computational graphs. Such formulas can express complex queries about the data, prior knowledge to satisfy during 
learning, statements to prove, etc. The following figure shows an example of how LTN converts such formulas into PyTorch 
computational graphs.

![Computational_graph_illustration](https://github.com/tommasocarraro/LTNtorch/blob/main/images/framework_computational_graph.png?raw=true)

Let us assume we have 6 people which are denoted using 4 real-valued features.
The previous figure illustrates the following:
![Computational_graph_explanation](https://github.com/tommasocarraro/LTNtorch/blob/main/images/computational_graph_explanation.png?raw=true)

### Conclusion

Using LTN, one can represent and effectively compute some of the most important tasks of deep learning. Examples of such 
tasks are classification, regression, clustering, and so on.

The [Getting Started](#getting-started) section of the README links to tutorials and examples to learn how to code Logic
Tensor Networks in PyTorch.

However, we suggest to carefully read the [paper](https://arxiv.org/pdf/2012.13635.pdf) before going through the tutorials and examples.


# Installation

It is possible to install LTNtorch using `pip`.

`pip install LTNtorch`

Alternatively, it is possible to install LTNtorch by cloning this repository. In this case, make sure to install all the 
requirements.

`pip3 install -r requirements.txt`

# Structure of repository

- `ltn/core.py`: this module contains the implementation of the LTN framework. In particular, it contains the definition 
of constants, variables, predicates, functions, connectives and quantifiers;
- `ltn/fuzzy_ops.py`: this module contains the implementation of common fuzzy logic semantics using PyTorch primitives;
- `tutorials/`: this folder contains some important tutorials to getting started with coding in LTN;
- `examples/`: this folder contains various problems approached using LTN and presented in the "Reach of Logic Tensor 
Networks" section of the paper;
- `tests/`: this folder contains unit tests that have been used to test the `core` and `fuzzy_ops` modules. The coverage 
is 100%.

# Tests

The `core` and `fuzzy_ops` modules of this repository have been entirely tested using `pytest`, with a coverage of 100%.
The examples included in the documentation have also been tested, using `doctest`.

# Documentation

The [documentation](https://tommasocarraro.github.io/LTNtorch/) has been created with 
[Sphinx](https://www.sphinx-doc.org/en/master/index.html), using the 
[Read the Docs Sphinx Theme](https://sphinx-rtd-theme.readthedocs.io/en/stable/).

# Getting Started

## Tutorials

`tutorials/` contains some important tutorials to getting started with coding in LTN. We suggest completing the tutorials in order.
The tutorials cover the following topics:
1. [Grounding in LTN (part 1)](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/tutorials/1-grounding_non_logical_symbols.ipynb): Real Logic, constants, predicates, functions, variables;
2. [Grounding in LTN (part 2)](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/tutorials/2-grounding_connectives.ipynb): connectives and quantifiers (+ [complement](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/tutorials/2b-operators-and-gradients.ipynb): choosing appropriate operators for learning);
3. [Learning in LTN](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/tutorials/3-knowledgebase-and-learning.ipynb): using satisfiability of LTN formulas as a training objective.

The tutorials are implemented using jupyter notebooks.

## Examples

`examples/` contains the series of examples included in the "Reach of Logic Tensor Networks" section of the paper. Their objective 
is to show how the language of Real Logic can be used to specify a number of tasks that involve learning from data and 
reasoning about logical knowledge. 

The examples covered are the following:
1. [Binary classification](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/examples/1-binary_classification.ipynb): illustrates, in the simplest setting, how to ground a binary classifier as a predicate in LTN;
2. [Multi-class single-label classification](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/examples/2-multi_class_single_label_classification.ipynb): illustrate how to ground predicates that can classify samples in several mutually-exclusive classes;
3. [Multi-class multi-label classification](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/examples/3-multi_class_multi_label_classification.ipynb): illustrate how to ground predicates that can classify samples in several classes which are not mutually-exclusive;
4. [Semi-supervised pattern recognition](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/examples/4-semi-supervised_pattern_recognition.ipynb): showcases the power of LTN in dealing with semi-supervised learning tasks;
5. [Regression](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/examples/5-regression.ipynb): illustrates how to ground a regressor as a function symbol in LTN;
6. [Clustering](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/examples/6-clustering.ipynb): illustrates how LTN can solve a unsupervised tasks using first-order logical constraints;
7. [Learning embeddings with LTN](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/examples/7-learning_embeddings_with_LTN.ipynb): illustrates how LTN can learn embeddings using learnable logical constants.

The examples are presented using jupyter notebooks.

# License

This project is licensed under the MIT License - see the [LICENSE](https://nbviewer.jupyter.org/github/bmxitalia/LTNtorch/blob/main/LICENSE) file for details.

# Acknowledgements

LTN has been developed thanks to active contributions and discussions with the following people (in alphabetical order):
- Alessandro Daniele (FBK)
- Artur d’Avila Garcez (City)
- Benedikt Wagner (City)
- Emile van Krieken (VU Amsterdam)
- Francesco Giannini (UniSiena)
- Giuseppe Marra (UniSiena)
- Ivan Donadello (FBK)
- Lucas Bechberger (UniOsnabruck)
- Luciano Serafini (FBK)
- Marco Gori (UniSiena)
- Michael Spranger (Sony AI)
- Michelangelo Diligenti (UniSiena)
- Samy Badreddine (Sony AI)
- Tommaso Carraro (FBK)

# Citing this repo

If you are using **LTNtorch** in your work, please consider citing this repository.

```
@software{LTNtorch,
  author = {Tommaso Carraro},
  title = {{LTNtorch: PyTorch implementation of Logic Tensor Networks}},
  month = {mar},
  year = {2022},
  publisher = {Zenodo},
  version = {1.0.0},
  doi = {10.5281/zenodo.6394282},
  url = {https://doi.org/10.5281/zenodo.6394282}
}
```



            

Raw data

            {
    "_id": null,
    "home_page": "https://github.com/bmxitalia/LTNtorch",
    "name": "LTNtorch",
    "maintainer": "",
    "docs_url": null,
    "requires_python": ">=3.7",
    "maintainer_email": "",
    "keywords": "pytorch,machine-learning,framework,neural-symbolic-computing,fuzzy-logic",
    "author": "Tommaso Carraro",
    "author_email": "tommaso.carraro@studenti.unipd.it",
    "download_url": "https://files.pythonhosted.org/packages/ea/22/1100e3c81b27316a71ae79865d907c355684bfe45b5590c3848a1612114f/LTNtorch-1.0.1.tar.gz",
    "platform": null,
    "description": "![build status](https://github.com/tommasocarraro/LTNtorch/actions/workflows/build.yml/badge.svg)\n[![coverage status](https://coveralls.io/repos/github/tommasocarraro/LTNtorch/badge.svg?branch=main)](https://coveralls.io/github/tommasocarraro/LTNtorch?branch=main)\n[![PyPi](https://img.shields.io/pypi/v/LTNtorch.svg)](https://pypi.python.org/pypi/LTNtorch)\n[![docs link](https://img.shields.io/badge/docs-github.io-blue)](https://tommasocarraro.github.io/LTNtorch/)\n[![MIT license](https://img.shields.io/github/license/Naereen/StrapDown.js.svg)](https://github.com/Naereen/StrapDown.js/blob/master/LICENSE)\n![python version](https://img.shields.io/badge/python-3.7|3.8|3.9-blue)\n[![DOI BADGE](https://zenodo.org/badge/DOI/10.5281/zenodo.6394282.svg)](https://doi.org/10.5281/zenodo.6394282)\n\n# LTNtorch: PyTorch implementation of Logic Tensor Networks\n\nWelcome to the PyTorch's implementation of [Logic Tensor Networks](https://arxiv.org/abs/2012.13635)!\n\n### Basic idea of the framework\n\nLogic Tensor Network (LTN) is a Neural-Symbolic (NeSy) framework which supports learning of neural networks using the\nsatisfaction of a first-order logic knowledge base as an objective. In other words, LTN uses logical reasoning on the\nknowledge base to guide the learning of a potentially deep neural network. \n\nThe idea of the framework is simple: \n- we have a first-order logic knowledge base containing a set of axioms;\n- we have some predicates, functions, or logical constants appearing in these axioms that we want to learn;\n- we have some data available that we can use to learn the parameters of those symbols.\n\nThe idea is to use the logical axioms as a loss function for our Logic Tensor Network. The objective is to find solutions\nin the hypothesis space that maximally satisfy all the axioms contained in our knowledge base.\n\n### Learning in LTN\n\nIn LTN, the learnable parameters are contained in the predicates, functions, and possibly learnable logical constants\nthat appear in the logical axioms of the knowledge base. \n\nDuring the forward step of the back-propagation algorithm, LTN\ncomputes the truth values of the logical formulas contained in the knowledge base, using the available data to ground\n(or instantiate) the logical formulas. As we have already said, these formulas \nwill contain some predicates and functions which are represented as learnable models. \n\nAt the end of the forward phase,\nthe truth values computed for the formulas are aggregated and used in the loss function. Our objective is to maximize\nthe aggregated truth value, namely maximally satisfy all the axioms.\n\nDuring the backward step, the\nlearnable parameters of predicates, functions, and possibly learnable logical constants are changed in such a way to move\ntowards a solution in the hypothesis space which better satisfies all the axioms in the knowledge base.\n\nAt the end of the training, the parameters of predicates, functions, and constants will have been updated in such a way the \nlogical formulas in the knowledge base are maximally satisfied. In particular, the parameters will have been learned by using both\ndata (to ground the formulas) and logical reasoning (at the loss function).\n\nAfter learning, it is possible to query predicates and functions on new data which was not available during training. Also,\nit is possible to query the truth values of new formulas which were not included in the knowledge base during training. In addition,\nif some logical constants have been learned, their parameters can be interpreted as embeddings.\n\n### Real Logic logical language\n\nTo make this learning possible, LTN uses a differentiable first-order logic language, called Real Logic, which enables \nthe incorporation of data and logic.\n\nReal Logic defines the concept of `grounding` (different from the grounding of logic), which is a mapping from the logical domain (i.e., constants, variables, and logical symbols)\nto tensors in the Real field or operations based on tensors. These operations could be, for instance, mathematical functions or learnable neural networks. In other words,\na `grounding`, denoted as \ud835\udca2, is a function which maps a logical symbol into a real tensor or an operation on tensors.\n\nIn particular, the grounding is defined as follows. Let us assume that *c* is a constant, *x* is a logical \nvariable, *P* is a predicate, and *f* is a logical function:\n![Grounding_in LTN](https://github.com/tommasocarraro/LTNtorch/blob/main/images/grounding.png?raw=true)\n\nThe `grounding` defines also how the logical connectives (\u2227, \u2228, \u00ac, \u21d2, \u2194) and quantifiers\n(\u2200, \u2203) are mapped in the Real domain. In particular, logical connectives are grounded using fuzzy logic semantics, while\nquantifiers are grounded using fuzzy aggregators. Please, **carefully** read the paper if you have some doubts on these notions.\n\nExamples of possible groundings are showed in the following figure. In the figure, `friend(Mary, John)` is an\natomic formula (predicate), while `\u2200x(friend(John, x) \u21d2 friend(Mary, x))` is a closed formula (all the variables are\nquantified). The letter \ud835\udca2, again, is the grounding, the function which maps the logical domain into the Real domain.\n\n![Grounding_illustration](https://github.com/tommasocarraro/LTNtorch/blob/main/images/framework_grounding.png?raw=true)\n\n### LTN as PyTorch computational graphs\n\nIn practice, LTN converts Real Logic formulas (e.g., `\u2200x\u2203y(friend(x,y) \u2227 italian(y))`, which states that everybody has a friend that is Italian) into [PyTorch](https://www.pytorch.org/) \ncomputational graphs. Such formulas can express complex queries about the data, prior knowledge to satisfy during \nlearning, statements to prove, etc. The following figure shows an example of how LTN converts such formulas into PyTorch \ncomputational graphs.\n\n![Computational_graph_illustration](https://github.com/tommasocarraro/LTNtorch/blob/main/images/framework_computational_graph.png?raw=true)\n\nLet us assume we have 6 people which are denoted using 4 real-valued features.\nThe previous figure illustrates the following:\n![Computational_graph_explanation](https://github.com/tommasocarraro/LTNtorch/blob/main/images/computational_graph_explanation.png?raw=true)\n\n### Conclusion\n\nUsing LTN, one can represent and effectively compute some of the most important tasks of deep learning. Examples of such \ntasks are classification, regression, clustering, and so on.\n\nThe [Getting Started](#getting-started) section of the README links to tutorials and examples to learn how to code Logic\nTensor Networks in PyTorch.\n\nHowever, we suggest to carefully read the [paper](https://arxiv.org/pdf/2012.13635.pdf) before going through the tutorials and examples.\n\n\n# Installation\n\nIt is possible to install LTNtorch using `pip`.\n\n`pip install LTNtorch`\n\nAlternatively, it is possible to install LTNtorch by cloning this repository. In this case, make sure to install all the \nrequirements.\n\n`pip3 install -r requirements.txt`\n\n# Structure of repository\n\n- `ltn/core.py`: this module contains the implementation of the LTN framework. In particular, it contains the definition \nof constants, variables, predicates, functions, connectives and quantifiers;\n- `ltn/fuzzy_ops.py`: this module contains the implementation of common fuzzy logic semantics using PyTorch primitives;\n- `tutorials/`: this folder contains some important tutorials to getting started with coding in LTN;\n- `examples/`: this folder contains various problems approached using LTN and presented in the \"Reach of Logic Tensor \nNetworks\" section of the paper;\n- `tests/`: this folder contains unit tests that have been used to test the `core` and `fuzzy_ops` modules. The coverage \nis 100%.\n\n# Tests\n\nThe `core` and `fuzzy_ops` modules of this repository have been entirely tested using `pytest`, with a coverage of 100%.\nThe examples included in the documentation have also been tested, using `doctest`.\n\n# Documentation\n\nThe [documentation](https://tommasocarraro.github.io/LTNtorch/) has been created with \n[Sphinx](https://www.sphinx-doc.org/en/master/index.html), using the \n[Read the Docs Sphinx Theme](https://sphinx-rtd-theme.readthedocs.io/en/stable/).\n\n# Getting Started\n\n## Tutorials\n\n`tutorials/` contains some important tutorials to getting started with coding in LTN. We suggest completing the tutorials in order.\nThe tutorials cover the following topics:\n1. [Grounding in LTN (part 1)](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/tutorials/1-grounding_non_logical_symbols.ipynb): Real Logic, constants, predicates, functions, variables;\n2. [Grounding in LTN (part 2)](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/tutorials/2-grounding_connectives.ipynb): connectives and quantifiers (+ [complement](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/tutorials/2b-operators-and-gradients.ipynb): choosing appropriate operators for learning);\n3. [Learning in LTN](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/tutorials/3-knowledgebase-and-learning.ipynb): using satisfiability of LTN formulas as a training objective.\n\nThe tutorials are implemented using jupyter notebooks.\n\n## Examples\n\n`examples/` contains the series of examples included in the \"Reach of Logic Tensor Networks\" section of the paper. Their objective \nis to show how the language of Real Logic can be used to specify a number of tasks that involve learning from data and \nreasoning about logical knowledge. \n\nThe examples covered are the following:\n1. [Binary classification](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/examples/1-binary_classification.ipynb): illustrates, in the simplest setting, how to ground a binary classifier as a predicate in LTN;\n2. [Multi-class single-label classification](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/examples/2-multi_class_single_label_classification.ipynb): illustrate how to ground predicates that can classify samples in several mutually-exclusive classes;\n3. [Multi-class multi-label classification](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/examples/3-multi_class_multi_label_classification.ipynb): illustrate how to ground predicates that can classify samples in several classes which are not mutually-exclusive;\n4. [Semi-supervised pattern recognition](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/examples/4-semi-supervised_pattern_recognition.ipynb): showcases the power of LTN in dealing with semi-supervised learning tasks;\n5. [Regression](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/examples/5-regression.ipynb): illustrates how to ground a regressor as a function symbol in LTN;\n6. [Clustering](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/examples/6-clustering.ipynb): illustrates how LTN can solve a unsupervised tasks using first-order logical constraints;\n7. [Learning embeddings with LTN](https://nbviewer.jupyter.org/github/tommasocarraro/LTNtorch/blob/main/examples/7-learning_embeddings_with_LTN.ipynb): illustrates how LTN can learn embeddings using learnable logical constants.\n\nThe examples are presented using jupyter notebooks.\n\n# License\n\nThis project is licensed under the MIT License - see the [LICENSE](https://nbviewer.jupyter.org/github/bmxitalia/LTNtorch/blob/main/LICENSE) file for details.\n\n# Acknowledgements\n\nLTN has been developed thanks to active contributions and discussions with the following people (in alphabetical order):\n- Alessandro Daniele (FBK)\n- Artur d\u2019Avila Garcez (City)\n- Benedikt Wagner (City)\n- Emile van Krieken (VU Amsterdam)\n- Francesco Giannini (UniSiena)\n- Giuseppe Marra (UniSiena)\n- Ivan Donadello (FBK)\n- Lucas Bechberger (UniOsnabruck)\n- Luciano Serafini (FBK)\n- Marco Gori (UniSiena)\n- Michael Spranger (Sony AI)\n- Michelangelo Diligenti (UniSiena)\n- Samy Badreddine (Sony AI)\n- Tommaso Carraro (FBK)\n\n# Citing this repo\n\nIf you are using **LTNtorch** in your work, please consider citing this repository.\n\n```\n@software{LTNtorch,\n  author = {Tommaso Carraro},\n  title = {{LTNtorch: PyTorch implementation of Logic Tensor Networks}},\n  month = {mar},\n  year = {2022},\n  publisher = {Zenodo},\n  version = {1.0.0},\n  doi = {10.5281/zenodo.6394282},\n  url = {https://doi.org/10.5281/zenodo.6394282}\n}\n```\n\n\n",
    "bugtrack_url": null,
    "license": "MIT",
    "summary": "LTNtorch: PyTorch implementation of Logic Tensor Networks",
    "version": "1.0.1",
    "split_keywords": [
        "pytorch",
        "machine-learning",
        "framework",
        "neural-symbolic-computing",
        "fuzzy-logic"
    ],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "73111cc312c068f86c5408d9b71842634d8a89415c4419e599086a1d2e13503a",
                "md5": "36e165c610dc8bd5e4644845c08c12e1",
                "sha256": "028e3fd652ae7996735d0b35691924b11abdf22eed00ab730d53adf863f65284"
            },
            "downloads": -1,
            "filename": "LTNtorch-1.0.1-py3-none-any.whl",
            "has_sig": false,
            "md5_digest": "36e165c610dc8bd5e4644845c08c12e1",
            "packagetype": "bdist_wheel",
            "python_version": "py3",
            "requires_python": ">=3.7",
            "size": 29256,
            "upload_time": "2023-03-28T12:38:33",
            "upload_time_iso_8601": "2023-03-28T12:38:33.488992Z",
            "url": "https://files.pythonhosted.org/packages/73/11/1cc312c068f86c5408d9b71842634d8a89415c4419e599086a1d2e13503a/LTNtorch-1.0.1-py3-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        },
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "ea221100e3c81b27316a71ae79865d907c355684bfe45b5590c3848a1612114f",
                "md5": "605ebfca5ac16bc4438bfa792da73b67",
                "sha256": "b485dc518665898a2483c074111b725185274738b70fbc79ee829c59b5ad55f3"
            },
            "downloads": -1,
            "filename": "LTNtorch-1.0.1.tar.gz",
            "has_sig": false,
            "md5_digest": "605ebfca5ac16bc4438bfa792da73b67",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": ">=3.7",
            "size": 28633,
            "upload_time": "2023-03-28T12:38:35",
            "upload_time_iso_8601": "2023-03-28T12:38:35.165020Z",
            "url": "https://files.pythonhosted.org/packages/ea/22/1100e3c81b27316a71ae79865d907c355684bfe45b5590c3848a1612114f/LTNtorch-1.0.1.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2023-03-28 12:38:35",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "github_user": "bmxitalia",
    "github_project": "LTNtorch",
    "travis_ci": false,
    "coveralls": false,
    "github_actions": true,
    "requirements": [],
    "lcname": "ltntorch"
}
        
Elapsed time: 0.05023s