Metadata-Version: 2.1
Name: fitbert
Version: 0.4.0
Summary: Use BERT to Fill in the Blanks
Home-page: https://github.com/Qordobacode/fitbert
Author: Qordoba
Author-email: sam.havens@qordoba.com
License: unlicensed
Description: # FitBERT
        
        ![buff bert](img/fitbert.png)
        
        FitBert ((F)ill (i)n (t)he blanks, (BERT)) is a library for using [BERT](https://arxiv.org/abs/1810.04805) to fill in the blank(s) in a section of text from a list of options. Here is the envisioned usecase for FitBert:
        
        1. A service (statistical model or something simpler) suggests replacements/corrections for a segment of text
        2. That service is specialized to a domain, and isn't good at the big picture, e.g. grammar
        3. That service passes the segment of text, with the words to be replaced identified, and the list of suggestions
        4. FitBert _crushes_ all but the best suggestion :muscle:
        
        ## Installation
        
        **This software is currently unlicensed. If you do not work at Qordoba, you are not able to legally use it.** However, we are working to make it available, probably under Apache 2.0. It will be available as a PyPi package.
        
        ## Usage
        
        [A Jupyter notebook with a short introduction is available here.](https://colab.research.google.com/drive/1NHYRMR7pKt9uY_KfFhIKXwPwvuX-VQLW)
        
        FitBert will automatically use GPU if `torch.cuda.is_available()`. Which is a good thing, because CPU inference time is really bad:
        
        ![cpu inference time chart showing roughly 1 prediction per second](https://imgur.com/3u1U9P8.png)
        
        Here is what GPU inference times are looking like
        
        ![gpu inference time chart showing roughly 100x speedup over CPU to approximately 100 per second](https://imgur.com/aEUmJXn.png)
        
        ### Usage as a library / in a server
        
        ```python
        from fitbert import FitBert
        
        
        # in theory you can pass a model_name and tokenizer, but currently only
        # bert-large-uncased and BertTokenizer are available
        # this takes a while and loads a whole big BERT into memory
        fb = FitBert()
        
        masked_string = "Why Bert, you're looking ***mask*** today!"
        options = ['buff', 'handsome', 'strong']
        
        ranked_options = fb.rank(masked_string, options=options)
        # >>> ['handsome', 'strong', 'buff']
        # or
        filled_in = fb.fitb(masked_string, options=options)
        # >>> "Why Bert, you're looking handsome today!"
        ```
        
        We commonly find ourselves knowing what verb to suggest, but not what conjugation:
        
        ```python
        from fitbert import FitBert
        
        
        fb = FitBert()
        
        masked_string = "Why Bert, you're ***mask*** handsome today!"
        options = ['looks']
        
        filled_in = fb.fitb(masked_string, options=options)
        # >>> "Why Bert, you're looking handsome today!"
        
        # under the hood, we notice there is only one suggestion and act as if
        # fitb was called with delemmatize=True:
        filled_in = fb.fitb(masked_string, options=options, delemmatize=True)
        ```
        
        If you are already using `pytorch_pretrained_bert.BertForMaskedLM`, and have an instance of BertForMaskedLM already instantiated, you can pass pass it in to reuse it:
        
        ```python
        BLM = pytorch_pretrained_bert.BertForMaskedLM.from_pretrained(model_name)
        
        fb = FitBert(model=BLM)
        ```
        
        You can also have FitBert mask the string for you
        
        ```python
        from fitbert import FitBert
        
        
        fb = FitBert()
        
        unmasked_string = "Why Bert, you're looks handsome today!"
        span_to_mask = (17, 22)
        masked_string, masked = fb.mask(unmasked_string, span_to_mask)
        # >>> "Why Bert, you're ***mask*** handsome today!", 'looks'
        
        # you can set options = [masked] or use any List[str]
        options = [masked]
        
        filled_in = fb.fitb(masked_string, options=options)
        # >>> "Why Bert, you're looking handsome today!"
        ```
        
        and there is a convenience method for doing this:
        
        ```python
        unmasked_string = "Why Bert, you're looks handsome today!"
        span_to_mask = (17, 22)
        
        filled_in = fb.mask_fitb(unmasked_string, span_to_mask)
        # >>> "Why Bert, you're looking handsome today!"
        ```
        
        ### Client
        
        If you are sending strings to a FitBert server, you need to either mask the string yourself, or identify the span you want masked:
        
        ```python
        from fitbert import FitBert
        
        s = "This might be justified as a means of signalling the connection between drunken driving and fatal accidents."
        
        better_string, span_to_change = MyRuleBasedNLPModel.remove_overly_fancy_language(s)
        
        assert better_string == "This might be justified to signalling the connection between drunken driving and fatal accidents.", "Notice 'as a means of' became 'to', but we didn't re-conjuagte signalling, or fix the spelling mistake"
        
        assert span_to_change == (27, 37), "This span is the start and stop of the characters for the substring 'signalling'."
        
        masked_string, replaced_substring = FitBert.mask(better_string, span_to_change)
        
        assert masked_string == "This might be justified to ***mask*** the connection between drunken driving and fatal accidents."
        
        assert replaced_substring == "signalling"
        
        FitBertServer.fitb(masked_string, options=[replaced_substring])
        ```
        
        The benefit to doing this over masking yourself is that if the internally used masking token changes, you don't have to know about that. Also, you don't need to make an instance of FitBert, so you don't have to incur the cost of downloading a pretrained Bert model.
        
        However, you could also write your `CallFitBertServer` function to take an unmasked string and a span, something like:
        
        ```python
        FitBertServer.mask_fitb(better_string, span_to_change)
        ```
        
        And then not need to install `FitBert` in your client at all.
        
        ## Development
        
        Run tests with `python -m pytest` or `python -m pytest -m "not slow"` to skip the 20 seconds of loading pretrained bert.
        
        ### Acknowledgement
        
        I am trying to get in touch with [NodoBird](https://drawception.com/player/546330/nodo-bird/), the artist of the fantastic portrait of Bert depicted above.
        
Platform: UNKNOWN
Classifier: Development Status :: 3 - Alpha
Classifier: Intended Audience :: Developers
Classifier: Intended Audience :: Science/Research
Classifier: License :: Other/Proprietary License
Classifier: Natural Language :: English
Classifier: Programming Language :: Python :: 3.5
Classifier: Programming Language :: Python :: 3.6
Classifier: Programming Language :: Python :: 3.7
Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
Classifier: Topic :: Text Processing :: Linguistic
Classifier: Typing :: Typed
Requires-Python: >=3.5
Description-Content-Type: text/markdown
