Metadata-Version: 2.1
Name: tlidb
Version: 0.0.2
Summary: The Transfer Learning in Dialogue Baselines Toolkit
Home-page: https://github.com/alon-albalak/TLiDB
Author: Alon Albalak
Author-email: alon_albalak@ucsb.edu
License: MIT
Description: # The Transfer Learning in Dialogue Benchmarking Toolkit
        
        ## Overview
        ---
        TLiDB is a tool used to benchmark methods of transfer learning in conversational AI.
        TLiDB can easily handle domain adaptation, task transfer, multitasking, continual learning, and other transfer learning settings.
        
        The main features of TLiDB are:
        
        1. Dataset class to easily load a dataset for use across models
        2. Unified metrics to standardize evaluation across datasets
        3. Extensible Model and Algorithm classes to support fast prototyping
        
        ## Installation
        ---
        To use TLiDB, you can simply isntall via pip:
        ```bash
        pip install tlidb
        ```
        
        OR, if you would like to edit or contribute, you can clone the repository and install from source:
        ```bash
        git clone git@github.com:alon-albalak/TLiDB.git
        cd TLiDB
        pip install -e .
        ```
        
        `examples/` contains sample scripts for:
        
        1. Training/Evaluating models in transfer learning settings
        2. 3 example models: BERT, GPT-2, T5, and training algorithms for each
        
        ## How to use TLiDB
        
        TODO:
        - Add examples for data loading/training
        - Add examples for using examples/run_experiment.py
        
        ### Folder descriptions:
        ---
        - /TLiDB is the main folder holding the code for data
            - /TLiDB/data_loaders contains code for data_loaders
            - /TLiDB/data is the destination folder for downloaded datasets
            - /TLiDB/datasets contains code for datasets
            - /TLiDB/metrics contains code for loss and evaluation metrics
            - /TLiDB/utils contains utility files
        - /examples contains sample code for training models
            - /examples/algorithms contains code which trains and evaluates a model
            - /examples/models contains code to define a model
            - /examples/configs contains code for model configurations
            - /examples/logs_and_models is the destination folder for training logs and model checkpoints
        - /dataset_preprocessing is for reproducability purposes, not required for end users. It contains scripts used to preprocess the TLiDB datasets from their original form into the TLiDB form
Platform: UNKNOWN
Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
Classifier: Intended Audience :: Science/Research
Classifier: Programming Language :: Python :: 3
Classifier: License :: OSI Approved :: MIT License
Requires-Python: >=3.6
Description-Content-Type: text/markdown
