Metadata-Version: 2.1
Name: trphysx
Version: 0.0.2
Summary: Transformers for modeling physical systems
Home-page: https://github.com/zabaras/transformer-physx
Author: Nicholas Geneva
Author-email: ngeneva@nd.edu
License: UNKNOWN
Project-URL: Bug Reports, https://github.com/zabaras/transformer-physx/issues
Project-URL: Source, https://github.com/zabaras/transformer-physx
Description: # Transformer PhysX
         [![PyPI version](https://badge.fury.io/py/trphysx.svg)](https://pypi.org/project/trphysx/) [![CircleCI](https://circleci.com/gh/zabaras/transformer-physx.svg?style=shield&circle-token=1d8e1117961b1d63d754f90e6b526649607aad34&branch=main)](https://circleci.com/gh/zabaras/transformer-physx) [![Documentation Status](https://readthedocs.org/projects/transformer-physx/badge/?version=latest)](https://transformer-physx.readthedocs.io/en/latest/?badge=latest) [![Website](https://img.shields.io/website?url=https%3A%2F%2Fzabaras.github.io%2Ftransformer-physx%2F)](https://zabaras.github.io/transformer-physx/) ![liscense](https://img.shields.io/github/license/zabaras/transformer-physx)
        
        
        Transformer PhysX is a Python packaged modeled after the [Hugging Face repository](https://github.com/huggingface/transformers) designed for the use of transformers for modeling physical systems. Transformers have seen recent success in both natural language processing and vision fields but have yet to fully permute other machine learning areas. Originally proposed in [Transformers for Modeling Physical Systems](https://arxiv.org/abs/2010.03957), this projects goal is to make these deep learning advances including self-attention and Koopman embeddings more accessible for the scientific machine learning community.
        
        [Website](https://zabaras.github.io/transformer-physx/) | [Documentation](https://transformer-physx.readthedocs.io) |[Getting Started](https://transformer-physx.readthedocs.io/en/latest/install.html) (Coming Soon) | [Data]() (Coming Soon)
        
        ### Associated Papers
        
        Transformers for Modeling Physical Systems [ [ArXiV](https://arxiv.org/abs/2010.03957) ]
        
        
        ### Colab Quick Start
        
        | | Embedding Model | Transformer |
        |---|:-------:|:-------:|
        | Lorenz | [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/zabaras/transformer-physx/blob/main/examples/lorenz/train_lorenz_enn.ipynb) | [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/zabaras/transformer-physx/blob/main/examples/lorenz/train_lorenz_transformer.ipynb) |
        | Cylinder Flow | - | - |
        | Gray-Scott | - | - |
        | Rossler | Coming Soon | Coming Soon |
        
        ### Road Map
        
        This is an on going project, hence many parts are not fully developed and will be added in the near future. If you have any particular questions or features you are interested in, please make a issue request so it can be prioritized!
        Thanks for understanding.
        
        - Add Rossler number example and collab notebook
        - Set up data repository
        - Github tasks for building wheels and uploading?
        - Additional Unit Testing for Better Code Coverage
        - Parallel Data training for transformer
        - Unsupervised pretraining physics
        
        ### Additional Resources
        
        - [Huggingface Repository](https://github.com/huggingface/transformers)
        - [Transformer illustrated blog post](https://jalammar.github.io/illustrated-transformer/)
        - [Deep learning Koopman dynamics blog post](https://nicholasgeneva.com/deep-learning/koopman/dynamics/2020/05/30/intro-to-koopman.html)
        
        
        ### Contact
        Open an issue on the Github repository if you have any questions/concerns.
        
        ### Citation
        Find this useful or like this work? Cite us with:
        
        ```latex
        @article{geneva2020transformers,
            title={Transformers for Modeling Physical Systems},
            author={Geneva, Nicholas and Zabaras, Nicholas},
            journal={arXiv preprint arXiv:2010.03957},
            year={2020}
        }
        ```
Keywords: transformers,physics,surrogate,machine learning,deep learning
Platform: UNKNOWN
Classifier: Development Status :: 2 - Pre-Alpha
Classifier: Intended Audience :: Science/Research
Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
Classifier: License :: OSI Approved :: MIT License
Classifier: Programming Language :: Python :: 3
Classifier: Programming Language :: Python :: 3.6
Classifier: Programming Language :: Python :: 3.7
Classifier: Programming Language :: Python :: 3.8
Requires-Python: >=3.6, <4
Description-Content-Type: text/markdown
