Metadata-Version: 1.2
Name: zero-downtime-migrations
Version: 0.11
Summary: django migrations without long locks
Home-page: https://github.com/yandex/zero-downtime-migrations
Author: Vladimir Koljasinskij
Author-email: smosker@yandex-team.ru
License: BSD-3-Clause
Description: .. image:: https://img.shields.io/pypi/v/zero-downtime-migrations.svg?style=flat
            :alt: PyPI Version
            :target: https://pypi.python.org/pypi/zero-downtime-migrations
        
        .. image:: https://img.shields.io/pypi/pyversions/zero-downtime-migrations.svg
            :alt: Supported Python versions
            :target: https://pypi.python.org/pypi/zero-downtime-migrations
        
        .. image:: https://travis-ci.org/yandex/zero-downtime-migrations.svg?branch=master
            :alt: Build Status
            :target: https://travis-ci.org/yandex/zero-downtime-migrations
        
        
        Zero-Downtime-Migrations
        ========================
        
        Description
        -----------
        *Zero-Downtime-Migrations (ZDM)* -- this is application which allow you to avoid long locks (and rewriting the whole table)
        while applying Django migrations using PostgreSql as database.
        
        Current possibilities
        ---------------------
        * add field with default value (nullable or not)
        * create index concurrently (and check index status after creation in case it was created with INVALID status)
        * add unique property to existing field through creating unique index concurrently and creating constraint using this index
        
        Why use it
        ----------
        We face such a problem - performing some django migrations (such as add column with default value) lock the table on
        read/write, so its impossible for our services to work properly during this periods. It can be acceptable on rather small
        tables (less than million rows), but even on them it can be painful if service is high loaded.
        But we have a lot of tables with more than 50 millions rows, and applying migrations on such a table lock it for
        more than an hour, which is totally unacceptable. Also, during this time consuming operations, migration rather often fail
        because of different errors (such as TimeoutError) and we have to start it from scratch or run sql manually thought
        psql and when fake migration.
        
        So in the end we have an idea of writing this package so it can prevent long locks on table and also
        provide more stable migration process which can be continued if operation fall for some reason.
        
        Installation
        ------------
        To install :code:`ZDM`, simply run:
        
        .. code:: bash
        
            pip install zero-downtime-migrations
        
        Usage
        -----
        If you are currently using default postresql backend change it to:
        
        .. code:: python
        
            DATABASES = {
                 'default': {
                     'ENGINE': 'zero_downtime_migrations.backend',
                     ...
                 }
                 ...
             }
        
        
        If you are using your own custom backend you can:
        
        * Set :code:`SchemaEditorClass` if you are currently using default one:
        
        .. code:: python
        
            from zero_downtime_migrations.backend.schema import DatabaseSchemaEditor
        
            class DatabaseWrapper(BaseWrapper):
                SchemaEditorClass = DatabaseSchemaEditor
        
        
        * Add :code:`ZeroDownTimeMixin` to base classes of your :code:`DatabaseSchemaEditor` if you are using custom one:
        
        .. code:: python
        
            from zero_downtime_migrations.backend.schema import ZeroDownTimeMixin
        
            class YourCustomSchemaEditor(ZeroDownTimeMixin, ...):
                ...
        
        Note about indexes
        ------------------
        Library will always force CONCURRENTLY index creation and after that check index status - if index was
        created with INVALID status it will be deleted and error will be raised.
        In this case you should fix problem if needed and restart migration.
        For example if creating unique index was failed you should make sure that there are only unique values
        in column on which index is creating.
        Usually index creating with invalid status due to deadlock so you need just restart migration.
        
        Example
        -------
        When adding not null column with default django will perform such sql query:
        
        .. code:: sql
        
            ALTER TABLE "test" ADD COLUMN "field" boolean DEFAULT True NOT NULL;
        
        Which cause postgres to rewrite the whole table and when swap it with existing one (`note from django documentation <https://docs.djangoproject.com/en/dev/topics/migrations/#postgresql>`_)
        and during this period it will hold exclusive lock on write/read on this table.
        
        This package will break sql above in separate commands not only to prevent the rewriting of whole
        table but also to add column with as small lock times as possible.
        
        First of all we will add nullable column without default and add default value to it in separate command in one transaction:
        
        .. code:: sql
        
            ALTER TABLE "test" ADD COLUMN "field" boolean NULL;
            ALTER TABLE "test" ALTER COLUMN "field" SET DEFAULT true;
        
        This will add default for all new rows in table but all existing ones will be with null value in this column for now,
        this operation will be quick because postgres doesn't have to fill all existing rows with default.
        
        Next we will count objects in table and if result if more than zero - calculate the
        size of batch in witch we will update existing rows. After that while where are still objects with null in this
        column - we will update them.
        
        While result of following statement is more than zero:
        
        .. code:: sql
        
            WITH cte AS (
            SELECT <table_pk_column> as pk
            FROM "test"
            WHERE  "field" is null
            LIMIT  <size_calculated_on_previous_step>
            )
            UPDATE "test" table_
            SET "field" = true
            FROM   cte
            WHERE  table_.<table_pk_column> = cte.pk
        
        When we have no more rows with null in this column we can set not null and drop default (which is django default
        behavior):
        
        .. code:: sql
        
            ALTER TABLE "test" ALTER COLUMN "field" SET NOT NULL;
            ALTER TABLE "test" ALTER COLUMN "field" DROP DEFAULT;
        
        So we finish add field process.
        It will be definitely more time consuming than basic variant with one sql statement, but in this approach
        there are no long locks on table so service can work normally during this migrations process.
        
        Run tests
        ---------
        
        .. code:: bash
        
            ./run_tests.sh
        
Keywords: django postgresql migrations
Platform: UNKNOWN
Classifier: Development Status :: 4 - Beta
Classifier: Framework :: Django
Classifier: Framework :: Django :: 1.8
Classifier: Framework :: Django :: 1.9
Classifier: Framework :: Django :: 1.10
Classifier: Framework :: Django :: 1.11
Classifier: Framework :: Django :: 2.0
Classifier: Framework :: Django :: 2.1
Classifier: Intended Audience :: Developers
Classifier: License :: OSI Approved :: BSD License
Classifier: Operating System :: OS Independent
Classifier: Topic :: Software Development :: Libraries :: Python Modules
Classifier: Programming Language :: Python
Classifier: Programming Language :: Python :: 2.7
Classifier: Programming Language :: Python :: 3.4
Classifier: Programming Language :: Python :: 3.5
Classifier: Programming Language :: Python :: 3.6
Requires-Python: >=2.7,!=3.1.*,!=3.0.*,!=3.2.*,!=3.3.*,<4.0
