Metadata-Version: 2.1
Name: onnxt5
Version: 0.1.4
Summary: Blazing fast summarization, translation, text-generation, Q&A and more using T5 in ONNX
Home-page: https://github.com/abelriboulot/onnxt5
Author: Abel Riboulot
Author-email: abel@kta.io
License: apache-2.0
Download-URL: https://github.com/abelriboulot/onnxt5/archive/0.1.4.tar.gz
Project-URL: Repo, https://github.com/abelriboulot/onnxt5
Description: ![ONNX T5](https://github.com/abelriboulot/onnxt5/blob/master/data/social_preview.png?raw=true)
        [![Actions Status](https://github.com/abelriboulot/onnxt5/workflows/Tests/badge.svg)](https://github.com/abelriboulot/onnxt5/actions)
        ![Actions Status](https://img.shields.io/github/license/abelriboulot/onnxt5)
        ![Version](https://img.shields.io/github/v/release/abelriboulot/onnxt5.svg)
        [![Downloads](https://pepy.tech/badge/onnxt5/week)](https://pepy.tech/project/onnxt5/week)
        [![Slack](https://img.shields.io/badge/slack-@onnxt5-yellow.svg?logo=slack)](https://join.slack.com/t/onnxt5/shared_invite/zt-gdjudd03-xutGvyQuYLMjBGnKH8fbLw)
        
        Summarization, translation, Q&A, text generation and more at blazing speed using a T5 version implemented in ONNX. 
        
        This package is still in alpha stage, therefore some functionalities such as beam searches are still in development.
        
        ## Installation
        ONNX-T5 is available on PyPi.
        ```bash
        pip install onnxt5
        ```
        
        For the dev version you can run the following.
        ```bash
        git clone https://github.com/abelriboulot/onnxt5
        cd onnxt5
        pip install -e .
        ```
        
        ## Usage
        
        The simplest way to get started for generation is to use the default pre-trained
        version of T5 on ONNX included in the package.
        
        > **_NOTE:_**  Please note that the first time you call get_encoder_decoder_tokenizer, the models are
        being downloaded which might take a minute or two.
        
        ```python
        from onnxt5 import GenerativeT5
        from onnxt5.api import get_encoder_decoder_tokenizer
        decoder_sess, encoder_sess, tokenizer = get_encoder_decoder_tokenizer()
        generative_t5 = GenerativeT5(encoder_sess, decoder_sess, tokenizer, onnx=True)
        prompt = 'translate English to French: I was a victim of a series of accidents.'
        
        output_text, output_logits = generative_t5(prompt, max_length=100, temperature=0.)
        # output_text: "J'ai été victime d'une série d'accidents."
        ```
        
        Other tasks just require to change the prefix in your prompt, for instance for summarization:
        ```python
        prompt = 'summarize: <PARAGRAPH>'
        output_text, output_logits = generative_t5(prompt, max_length=100, temperature=0.)
        ```
        
        If you want to get the embeddings of text, you can run the following
        ```python
        from onnxt5.api import get_encoder_decoder_tokenizer, run_embeddings_text
        
        decoder_sess, encoder_sess, tokenizer = get_encoder_decoder_tokenizer()
        prompt = 'Listen, Billy Pilgrim has come unstuck in time.'
        encoder_embeddings, decoder_embeddings = run_embeddings_text(encoder_sess, decoder_sess, tokenizer, prompt)
        ```
        
        ONNXT5 also lets you export and use your own models. See the `examples\` folder for more detailed examples.
        
        T5 works with tokens such as `summarize:`, `translate English to German:`, or `question: ... context:`. You can see a 
        list of the pretrained tasks and token in the appendix D of the [original paper](https://arxiv.org/pdf/1910.10683.pdf).
        
        ## Functionalities
        * Run any of the T5 trained tasks in a line (translation, summarization, sentiment analysis, completion, generation)
        * Export your own T5 models to ONNX easily
        * Utility functions to generate what you need quickly
        * Up to 4X speedup compared to PyTorch execution for smaller contexts
        
        ## Benchmarks
        The outperformance varies heavily based on the length of the context. For contexts less than ~500 words,
        ONNX outperforms greatly, going up to a 4X speedup compared to PyTorch. However, the longer the context, the smaller the 
        speedup of ONNX, with Pytorch being faster above 500 words.
        
        #### GPU Benchmark, Embedding Task
        
        ![Benchmark Embedding](https://github.com/abelriboulot/onnxt5/blob/master/data/Embedding_benchmark.png?raw=true)
        #### GPU Benchmark, Generation Task
        
        ![Benchmark Generation](https://github.com/abelriboulot/onnxt5/blob/master/data/Generation_benchmark.png?raw=true)
        
        ## Contributing
        The project is still in its infancy, so I would love your feedback, to know what problems you are trying to solve, hear issues you're encountering, 
        and discuss features that would help you. Therefore feel free to shoot me an e-mail (see [my profile](https://github.com/abelriboulot) for the address!) or 
        join our [slack community](https://join.slack.com/t/onnxt5/shared_invite/zt-gdjudd03-xutGvyQuYLMjBGnKH8fbLw).
        
        ## Acknowledgements
        This repo is based on the work of Colin Raffel and Noam Shazeer and Adam Roberts and Katherine Lee and Sharan Narang and 
        Michael Matena and Yanqi Zhou and Wei Li and Peter J. Liu from Google, as well as the implementation of T5 from the 
        huggingface team, the work of the Microsoft ONNX and onnxruntime teams, in particular Tianlei Wu, and the work of Thomas Wolf on generation of text.
        
        [Original T5 Paper](https://arxiv.org/pdf/1910.10683.pdf)
        ```
        @article{2019t5,
          author = {Colin Raffel and Noam Shazeer and Adam Roberts and Katherine Lee and Sharan Narang and Michael Matena and Yanqi Zhou and Wei Li and Peter J. Liu},
          title = {Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer},
          journal = {arXiv e-prints},
          year = {2019},
          archivePrefix = {arXiv},
          eprint = {1910.10683},
        }
        ```
        [Microsoft onnxruntime repo](https://github.com/microsoft/onnxruntime)
        
        [HuggingFace implementation of T5](https://huggingface.co/transformers/model_doc/t5.html)
Keywords: T5,ONNX,onnxruntime,NLP,transformer,generate text,summarization,translation,q&a,machine learning,inference,fast inference
Platform: UNKNOWN
Classifier: Development Status :: 3 - Alpha
Classifier: Intended Audience :: Developers
Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
Classifier: License :: OSI Approved :: Apache Software License
Classifier: Programming Language :: Python :: 3
Classifier: Programming Language :: Python :: 3.5
Classifier: Programming Language :: Python :: 3.6
Classifier: Programming Language :: Python :: 3.7
Classifier: Programming Language :: Python :: 3.8
Requires-Python: >=3.5, <4
Description-Content-Type: text/markdown
Provides-Extra: dev
Provides-Extra: test
