Metadata-Version: 2.1
Name: pyhacrf-datamade
Version: 0.2.6
Summary: Hidden alignment conditional random field, a discriminative string edit distance
Home-page: https://github.com/datamade/pyhacrf
Author: Dirko Coetsee
Author-email: dpcoetsee@gmail.com
Maintainer: Forest Gregg
Maintainer-email: fgregg@gmail.com
License: UNKNOWN
Platform: UNKNOWN
Classifier: Intended Audience :: Science/Research
Classifier: License :: OSI Approved :: BSD License
Classifier: Operating System :: OS Independent
Classifier: Programming Language :: Python
Classifier: Topic :: Scientific/Engineering
License-File: LICENSE

pyhacrf
=======
.. image:: https://travis-ci.org/dedupeio/pyhacrf.svg?branch=master
    :target: https://travis-ci.org/dedupeio/pyhacrf
    
.. image:: https://ci.appveyor.com/api/projects/status/kibqrd7wnsk2ilpf/branch/master?svg=true
    :target: https://ci.appveyor.com/project/fgregg/pyhacrf

Hidden alignment conditional random field for classifying string pairs -
a learnable edit distance.

Part of the Dedupe.io cloud service and open source toolset for de-duplicating and finding fuzzy matches in your data: https://dedupe.io

This package aims to implement the HACRF machine learning model with a
``sklearn``-like interface. It includes ways to fit a model to training
examples and score new example.

The model takes string pairs as input and classify them into any number
of classes. In McCallum's original paper the model was applied to the
database deduplication problem. Each database entry was paired with
every other entry and the model then classified whether the pair was a
'match' or a 'mismatch' based on training examples of matches and
mismatches.

I also tried to use it as learnable string edit distance for normalizing
noisy text. See *A Conditional Random Field for Discriminatively-trained
Finite-state String Edit Distance* by McCallum, Bellare, and Pereira,
and the report *Conditional Random Fields for Noisy text normalisation*
by Dirko Coetsee.

Example
-------

.. code:: python

    from pyhacrf import StringPairFeatureExtractor, Hacrf

    training_X = [('helloooo', 'hello'), # Matching examples
                  ('h0me', 'home'),
                  ('krazii', 'crazy'),
                  ('non matching string example', 'no really'), # Non-matching examples
                  ('and another one', 'yep')]
    training_y = ['match',
                  'match',
                  'match',
                  'non-match',
                  'non-match']

    # Extract features
    feature_extractor = StringPairFeatureExtractor(match=True, numeric=True)
    training_X_extracted = feature_extractor.fit_transform(training_X)

    # Train model
    model = Hacrf(l2_regularization=1.0)
    model.fit(training_X_extracted, training_y)

    # Evaluate
    from sklearn.metrics import confusion_matrix
    predictions = model.predict(training_X_extracted)

    print(confusion_matrix(training_y, predictions))
    > [[0 3]
    >  [2 0]]

    print(model.predict_proba(training_X_extracted))
    > [[ 0.94914812  0.05085188]
    >  [ 0.92397711  0.07602289]
    >  [ 0.86756034  0.13243966]
    >  [ 0.05438812  0.94561188]
    >  [ 0.02641275  0.97358725]]

Dependencies
------------

This package depends on ``numpy``. The LBFGS optimizer in ``pylbfgs`` is
used, but alternative optimizers can be passed.

Install
-------

Install by running:

::

    python setup.py install

or from pypi:

::

    pip install pyhacrf

Developing
----------
Clone from repository, then

::

    pip install -r requirements.txt
    cython pyhacrf/*.pyx
    python setup.py install

To deploy to pypi, make sure you have compiled the \*.pyx files to \*.c


