Metadata-Version: 2.1
Name: astred
Version: 0.9.6
Summary: A collection of syntactic metrics to calculate (dis)similarities between source and target sentences.
Home-page: https://github.com/BramVanroy/astred
Author: Bram Vanroy
Author-email: bramvanroy@hotmail.com
License: Apache 2.0
Project-URL: Issue tracker, https://github.com/BramVanroy/astred/issues
Project-URL: Source, https://github.com/BramVanroy/astred
Description: Easily compare two word-aligned sentences with ASTrED
        =====================================================
        
        Example notebooks
        -----------------
        
        A couple example notebooks exist, each with a different grade of automation for the initialisation of the aligned object. 
        Once an aligned object has been created, the functionality is identical.
        
        - `High automation`_: *automate all the things*. Tokenisation, parsing, and word alignment is done automatically
          [`Try on Colab <https://colab.research.google.com/github/BramVanroy/astred/blob/master/examples/full-auto.ipynb>`__]
        - `Normal automation`_: the typical scenario where you have tokenised and aligned text that is not parsed yet
          [`Try on Colab <https://colab.research.google.com/github/BramVanroy/astred/blob/master/examples/automatic-parsing.ipynb>`__]
        - `No automation`_: full-manual mode, where you provide all the required information, including dependency labels
          and heads [`Try on Colab <https://colab.research.google.com/github/BramVanroy/astred/blob/master/examples/full-manual.ipynb>`__]
        - `Monolingual`_: in this example we rely on spaCy to compare two English sentences and calculate semantic similarity
          between aligned words [`Try on Colab <https://colab.research.google.com/github/BramVanroy/astred/blob/master/examples/monolingual.ipynb>`__]
        
        .. _High automation: examples/full-auto.ipynb
        .. _Normal automation: examples/automatic-parsing.ipynb
        .. _No automation: examples/full-manual.ipynb
        .. _Monolingual: examples/monolingual.ipynb
        
        Installation
        ------------
        
        Requires Python 3.7 or higher. To keep the overhead low, a default parser is NOT installed. Currently both `spaCy`_ and
        `stanza`_ are supported and you can choose which one to use. Stanza is recommended for bilingual research (because it
        is ensured that all of its models use Universal Dependencies), but spaCy can be used as well. The latter is especially
        used for monolingual comparisons, or if you are not interested in the linguistic comparisons and only require word
        reordering metrics.
        
        A pre-release is available on PyPi. You can install it with pip as follows.
        
        .. code-block:: bash
        
            # Install with stanza (recommended)
            pip install astred[stanza]
            # ... or install with spacy
            pip install astred[spacy]
            # ... or install with both and decide later
            pip install astred[parsers]
        
        If you want to use spaCy, you have to make sure that you `install`_ the required models manually, which cannot be
        automated.
        
        .. _spaCy: https://spacy.io/
        .. _stanza: https://github.com/stanfordnlp/stanza
        .. _install: https://spacy.io/usage/models
        
        Automatic Word Alignment
        ------------------------
        
        Automatic word alignment is supported by using a modified version of `Awesome Align`_ under the hood. This is a neural
        word aligner that uses transfer learning with multilingual models to do word alignment. It does require
        some manual installation work. Specifically, you need to install the :code:`astred_compat` branch from `this fork`_.
        If you are using pip, you can run the following command:
        
        .. code-block:: bash
        
            pip install git+https://github.com/BramVanroy/awesome-align.git@astred_compat
        
        Awesome Align requires PyTorch, like :code:`stanza` above.
        
        If it is installed, you can initialize :code:`AlignedSentences` without providing word alignments. Those will be added
        automatically behind the scenes. See `this example notebook`_ [`Try on Colab <https://colab.research.google.com/github/BramVanroy/astred/blob/master/examples/full-auto.ipynb>`__] for more.
        
        .. code-block:: bash
        
        	sent_en = Sentence.from_text("I like eating cookies", "en")
        	sent_nl = Sentence.from_text("Ik eet graag koekjes", "nl")
        
        	# Word alignments do not need to be added on init:
        	aligned = AlignedSentences(sent_en, sent_nl)
        
        Keep in mind however that automatic alignment will never have the same quality as manual alignments. Use with caution!
        I highly suggest reading `the paper`_ of Awesome Align to see whether it is a good pick for you.
        
        .. _Awesome Align: https://github.com/neulab/awesome-align
        .. _this fork: https://github.com/BramVanroy/awesome-align/tree/astred_compat
        .. _this example notebook: examples/full-auto.ipynb
        .. _the paper: https://arxiv.org/abs/2101.08231
        
        License
        -------
        Licensed under Apache License Version 2.0. See the LICENSE file attached to this repository.
        
        Citation
        --------
        Please cite our `papers`_ if you use this library.
        
        Vanroy, B., De Clercq, O., Tezcan, A., Daems, J., & Macken, L. (2021). Metrics ofsyntactic equivalence to assess 
        translation difficulty. In M. Carl (Ed.), *Explorations in empirical translation process research* (Vol. 3, pp. 259–294).
        Cham, Switzerland: Springer International Publishing. https://doi.org/10.1007/978-3-030-69777-8_10
        
        .. code-block::
        
        	@incollection{vanroy2021metrics,
        	    title = {Metrics of syntactic equivalence to assess translation difficulty},
        	    booktitle = {Explorations in empirical translation process research},
        	    author = {Vanroy, Bram and De Clercq, Orph{\'e}e and Tezcan, Arda and Daems, Joke and Macken, Lieve},
        	    editor = {Carl, Michael},
        	    year = {2021},
        	    series = {Machine {{Translation}}: {{Technologies}} and {{Applications}}},
        	    volume = {3},
        	    pages = {259--294},
        	    publisher = {{Springer International Publishing}},
        	    address = {{Cham, Switzerland}},
        	    isbn = {978-3-030-69776-1},
        	    url = {https://link.springer.com/chapter/10.1007/978-3-030-69777-8_10},
        	    doi = {10.1007/978-3-030-69777-8_10}
        	}
        
        Vanroy, B., Schaeffer, M., & Macken, L. (2021). Comparing the Effect of Product-Based Metrics on the Translation Process. *Frontiers in Psychology*, 12. https://doi.org/10.3389/fpsyg.2021.681945
        
        .. code-block::
        
        	@article{vanroy2021comparing,
        	    publisher = {Frontiers},
        	    author = {Vanroy, Bram and Schaeffer, Moritz and Macken, Lieve},
        	    title = {Comparing the effect of product-based metrics on the translation process},
        	    year = {2021},
        	    journal = {Frontiers in Psychology},
        	    volume = {12}, 
        	    issn = {1664-1078}, 
        	    url = {https://www.frontiersin.org/article/10.3389/fpsyg.2021.681945},
        	    doi = {10.3389/fpsyg.2021.681945}, 
        	}
        
        
        .. _papers: CITATION
        
Keywords: nlp tree-edit-distance ted syntax compling computational-linguistics syntactic-distance translation
Platform: UNKNOWN
Classifier: Development Status :: 4 - Beta
Classifier: Intended Audience :: Developers
Classifier: Intended Audience :: Science/Research
Classifier: Topic :: Scientific/Engineering
Classifier: Topic :: Text Processing
Classifier: Programming Language :: Python :: 3.7
Classifier: Programming Language :: Python :: 3.8
Classifier: Programming Language :: Python :: 3.9
Classifier: Operating System :: OS Independent
Requires-Python: >=3.7
Description-Content-Type: text/x-rst
Provides-Extra: stanza
Provides-Extra: spacy
Provides-Extra: parsers
Provides-Extra: all
Provides-Extra: dev
