summaryrefslogtreecommitdiff
path: root/oslo_db/sqlalchemy
diff options
context:
space:
mode:
authorMike Bayer <mike_mp@zzzcomputing.com>2022-08-12 18:47:57 -0400
committerMike Bayer <mike_mp@zzzcomputing.com>2023-03-21 13:59:26 -0400
commit94d6e24ca19b0116eed00d5ccdb8a538918c6dcf (patch)
tree7006674c4fcd747ae7c679ac4d2a79f77075866a /oslo_db/sqlalchemy
parent877bcfc6a6ed16ba6885f47824df6b1f5ac60b4e (diff)
downloadoslo-db-94d6e24ca19b0116eed00d5ccdb8a538918c6dcf.tar.gz
Remove sqlalchemy-migrate
sqlalchemy-migrate does not support SQLAlchemy 2.x and we're not going to invest the time in changing this. Remove integration of sqlalchemy-migrate in oslo.db, allowing us to support SQLAlchemy 2.x fully. Note that we do not remove the 'migration_cli' module entirely yet. While this is deprecated, it is possible to use this with alembic. New users shouldn't be switching to it, but any existing users can continue to use this module for some time yet. Change-Id: Ic3d6bd318038d723b0d50d39e45f8e26289e9a57 Sem-Ver: api-break
Diffstat (limited to 'oslo_db/sqlalchemy')
-rw-r--r--oslo_db/sqlalchemy/migration.py183
-rw-r--r--oslo_db/sqlalchemy/migration_cli/ext_migrate.py79
-rw-r--r--oslo_db/sqlalchemy/test_migrations.py217
-rw-r--r--oslo_db/sqlalchemy/utils.py237
4 files changed, 0 insertions, 716 deletions
diff --git a/oslo_db/sqlalchemy/migration.py b/oslo_db/sqlalchemy/migration.py
deleted file mode 100644
index b29b4da..0000000
--- a/oslo_db/sqlalchemy/migration.py
+++ /dev/null
@@ -1,183 +0,0 @@
-# coding=utf-8
-
-# Copyright (c) 2013 OpenStack Foundation
-# All Rights Reserved.
-#
-# Licensed under the Apache License, Version 2.0 (the "License"); you may
-# not use this file except in compliance with the License. You may obtain
-# a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
-# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
-# License for the specific language governing permissions and limitations
-# under the License.
-#
-# Base on code in migrate/changeset/databases/sqlite.py which is under
-# the following license:
-#
-# The MIT License
-#
-# Copyright (c) 2009 Evan Rosson, Jan Dittberner, Domen Kožar
-#
-# Permission is hereby granted, free of charge, to any person obtaining a copy
-# of this software and associated documentation files (the "Software"), to deal
-# in the Software without restriction, including without limitation the rights
-# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
-# copies of the Software, and to permit persons to whom the Software is
-# furnished to do so, subject to the following conditions:
-# The above copyright notice and this permission notice shall be included in
-# all copies or substantial portions of the Software.
-#
-# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
-# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
-# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
-# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
-# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
-# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
-# THE SOFTWARE.
-
-import os
-
-from debtcollector import removals
-from migrate import exceptions as versioning_exceptions
-from migrate.versioning import api as versioning_api
-from migrate.versioning.repository import Repository
-import sqlalchemy
-
-from oslo_db._i18n import _
-from oslo_db import exception
-
-
-_removed_msg = (
- 'sqlalchemy-migrate support in oslo_db is deprecated; consider '
- 'migrating to alembic'
-)
-
-
-@removals.remove(message=_removed_msg, version='8.3.0')
-def db_sync(engine, abs_path, version=None, init_version=0, sanity_check=True):
- """Upgrade or downgrade a database.
-
- Function runs the upgrade() or downgrade() functions in change scripts.
-
- :param engine: SQLAlchemy engine instance for a given database
- :param abs_path: Absolute path to migrate repository.
- :param version: Database will upgrade/downgrade until this version.
- If None - database will update to the latest
- available version.
- :param init_version: Initial database version
- :param sanity_check: Require schema sanity checking for all tables
- """
-
- if version is not None:
- try:
- version = int(version)
- except ValueError:
- raise exception.DBMigrationError(_("version should be an integer"))
-
- current_version = db_version(engine, abs_path, init_version)
- repository = _find_migrate_repo(abs_path)
- if sanity_check:
- _db_schema_sanity_check(engine)
- if version is None or version > current_version:
- try:
- migration = versioning_api.upgrade(engine, repository, version)
- except Exception as ex:
- raise exception.DBMigrationError(ex)
- else:
- migration = versioning_api.downgrade(engine, repository,
- version)
- if sanity_check:
- _db_schema_sanity_check(engine)
-
- return migration
-
-
-def _db_schema_sanity_check(engine):
- """Ensure all database tables were created with required parameters.
-
- :param engine: SQLAlchemy engine instance for a given database
-
- """
-
- if engine.name == 'mysql':
- onlyutf8_sql = ('SELECT TABLE_NAME,TABLE_COLLATION '
- 'from information_schema.TABLES '
- 'where TABLE_SCHEMA=%s and '
- 'TABLE_COLLATION NOT LIKE \'%%utf8%%\'')
-
- # NOTE(morganfainberg): exclude the sqlalchemy-migrate and alembic
- # versioning tables from the tables we need to verify utf8 status on.
- # Non-standard table names are not supported.
- EXCLUDED_TABLES = ['migrate_version', 'alembic_version']
-
- table_names = [res[0] for res in
- engine.execute(onlyutf8_sql, engine.url.database) if
- res[0].lower() not in EXCLUDED_TABLES]
-
- if len(table_names) > 0:
- raise ValueError(_('Tables "%s" have non utf8 collation, '
- 'please make sure all tables are CHARSET=utf8'
- ) % ','.join(table_names))
-
-
-@removals.remove(message=_removed_msg, version='8.3.0')
-def db_version(engine, abs_path, init_version):
- """Show the current version of the repository.
-
- :param engine: SQLAlchemy engine instance for a given database
- :param abs_path: Absolute path to migrate repository
- :param init_version: Initial database version
- """
- repository = _find_migrate_repo(abs_path)
- try:
- return versioning_api.db_version(engine, repository)
- except versioning_exceptions.DatabaseNotControlledError:
- meta = sqlalchemy.MetaData()
- meta.reflect(bind=engine)
- tables = meta.tables
- if (len(tables) == 0 or 'alembic_version' in tables or
- 'migrate_version' in tables):
- db_version_control(engine, abs_path, version=init_version)
- return versioning_api.db_version(engine, repository)
- else:
- raise exception.DBMigrationError(
- _("The database is not under version control, but has "
- "tables. Please stamp the current version of the schema "
- "manually."))
-
-
-@removals.remove(message=_removed_msg, version='8.3.0')
-def db_version_control(engine, abs_path, version=None):
- """Mark a database as under this repository's version control.
-
- Once a database is under version control, schema changes should
- only be done via change scripts in this repository.
-
- :param engine: SQLAlchemy engine instance for a given database
- :param abs_path: Absolute path to migrate repository
- :param version: Initial database version
- """
- repository = _find_migrate_repo(abs_path)
-
- try:
- versioning_api.version_control(engine, repository, version)
- except versioning_exceptions.InvalidVersionError as ex:
- raise exception.DBMigrationError("Invalid version : %s" % ex)
- except versioning_exceptions.DatabaseAlreadyControlledError:
- raise exception.DBMigrationError("Database is already controlled.")
-
- return version
-
-
-def _find_migrate_repo(abs_path):
- """Get the project's change script repository
-
- :param abs_path: Absolute path to migrate repository
- """
- if not os.path.exists(abs_path):
- raise exception.DBMigrationError("Path %s not found" % abs_path)
- return Repository(abs_path)
diff --git a/oslo_db/sqlalchemy/migration_cli/ext_migrate.py b/oslo_db/sqlalchemy/migration_cli/ext_migrate.py
deleted file mode 100644
index 09f63d3..0000000
--- a/oslo_db/sqlalchemy/migration_cli/ext_migrate.py
+++ /dev/null
@@ -1,79 +0,0 @@
-# Licensed under the Apache License, Version 2.0 (the "License"); you may
-# not use this file except in compliance with the License. You may obtain
-# a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
-# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
-# License for the specific language governing permissions and limitations
-# under the License.
-
-import logging
-import os
-
-from migrate.versioning import version as migrate_version
-
-from oslo_db.sqlalchemy import migration
-from oslo_db.sqlalchemy.migration_cli import ext_base
-
-
-LOG = logging.getLogger(__name__)
-
-
-class MigrateExtension(ext_base.MigrationExtensionBase):
- """Extension to provide sqlalchemy-migrate features.
-
- :param migration_config: Stores specific configuration for migrations
- :type migration_config: dict
- """
-
- order = 1
-
- def __init__(self, engine, migration_config):
- self.engine = engine
- self.repository = migration_config.get('migration_repo_path', '')
- self.init_version = migration_config.get('init_version', 0)
-
- @property
- def enabled(self):
- return os.path.exists(self.repository)
-
- def upgrade(self, version):
- version = None if version == 'head' else version
- return migration.db_sync(
- self.engine, self.repository, version,
- init_version=self.init_version)
-
- def downgrade(self, version):
- try:
- # version for migrate should be valid int - else skip
- if version in ('base', None):
- version = self.init_version
- version = int(version)
- return migration.db_sync(
- self.engine, self.repository, version,
- init_version=self.init_version)
- except ValueError:
- LOG.error(
- 'Migration number for migrate plugin must be valid '
- 'integer or empty, if you want to downgrade '
- 'to initial state'
- )
- raise
-
- def version(self):
- return migration.db_version(
- self.engine, self.repository, init_version=self.init_version)
-
- def has_revision(self, rev_id):
- collection = migrate_version.Collection(self.repository)
- try:
- collection.version(rev_id)
- return True
- except (KeyError, ValueError):
- # NOTE(breton): migrate raises KeyError if an int is passed but not
- # found in the list of revisions and ValueError if non-int is
- # passed. Both mean there is no requested revision.
- return False
diff --git a/oslo_db/sqlalchemy/test_migrations.py b/oslo_db/sqlalchemy/test_migrations.py
index a0b5591..22ace67 100644
--- a/oslo_db/sqlalchemy/test_migrations.py
+++ b/oslo_db/sqlalchemy/test_migrations.py
@@ -27,229 +27,12 @@ import sqlalchemy.exc
import sqlalchemy.sql.expression as expr
import sqlalchemy.types as types
-from oslo_db import exception as exc
from oslo_db.sqlalchemy import provision
from oslo_db.sqlalchemy import utils
LOG = logging.getLogger(__name__)
-class WalkVersionsMixin(object, metaclass=abc.ABCMeta):
- """Test mixin to check upgrade and downgrade ability of migration.
-
- This is only suitable for testing of migrate_ migration scripts. An
- abstract class mixin. `INIT_VERSION`, `REPOSITORY` and `migration_api`
- attributes must be implemented in subclasses.
-
- .. _auxiliary-dynamic-methods:
-
- Auxiliary Methods:
-
- `migrate_up` and `migrate_down` instance methods of the class can be
- used with auxiliary methods named `_pre_upgrade_<revision_id>`,
- `_check_<revision_id>`, `_post_downgrade_<revision_id>`. The methods
- intended to check applied changes for correctness of data operations.
- This methods should be implemented for every particular revision
- which you want to check with data. Implementation recommendations for
- `_pre_upgrade_<revision_id>`, `_check_<revision_id>`,
- `_post_downgrade_<revision_id>` implementation:
-
- * `_pre_upgrade_<revision_id>`: provide a data appropriate to
- a next revision. Should be used an id of revision which
- going to be applied.
-
- * `_check_<revision_id>`: Insert, select, delete operations
- with newly applied changes. The data provided by
- `_pre_upgrade_<revision_id>` will be used.
-
- * `_post_downgrade_<revision_id>`: check for absence
- (inability to use) changes provided by reverted revision.
-
- Execution order of auxiliary methods when revision is upgrading:
-
- `_pre_upgrade_###` => `upgrade` => `_check_###`
-
- Execution order of auxiliary methods when revision is downgrading:
-
- `downgrade` => `_post_downgrade_###`
-
- .. _migrate: https://sqlalchemy-migrate.readthedocs.org/en/latest/
-
- """
-
- @property
- @abc.abstractmethod
- def INIT_VERSION(self):
- """Initial version of a migration repository.
-
- Can be different from 0, if a migrations were squashed.
-
- :rtype: int
- """
- pass
-
- @property
- @abc.abstractmethod
- def REPOSITORY(self):
- """Allows basic manipulation with migration repository.
-
- :returns: `migrate.versioning.repository.Repository` subclass.
- """
- pass
-
- @property
- @abc.abstractmethod
- def migration_api(self):
- """Provides API for upgrading, downgrading and version manipulations.
-
- :returns: `migrate.api` or overloaded analog.
- """
- pass
-
- @property
- @abc.abstractmethod
- def migrate_engine(self):
- """Provides engine instance.
-
- Should be the same instance as used when migrations are applied. In
- most cases, the `engine` attribute provided by the test class in a
- `setUp` method will work.
-
- Example of implementation:
-
- def migrate_engine(self):
- return self.engine
-
- :returns: sqlalchemy engine instance
- """
- pass
-
- def walk_versions(self, snake_walk=False, downgrade=True):
- """Check if migration upgrades and downgrades successfully.
-
- Determine the latest version script from the repo, then
- upgrade from 1 through to the latest, with no data
- in the databases. This just checks that the schema itself
- upgrades successfully.
-
- `walk_versions` calls `migrate_up` and `migrate_down` with
- `with_data` argument to check changes with data, but these methods
- can be called without any extra check outside of `walk_versions`
- method.
-
- :param snake_walk: enables checking that each individual migration can
- be upgraded/downgraded by itself.
-
- If we have ordered migrations 123abc, 456def, 789ghi and we run
- upgrading with the `snake_walk` argument set to `True`, the
- migrations will be applied in the following order::
-
- `123abc => 456def => 123abc =>
- 456def => 789ghi => 456def => 789ghi`
-
- :type snake_walk: bool
- :param downgrade: Check downgrade behavior if True.
- :type downgrade: bool
- """
-
- # Place the database under version control
- self.migration_api.version_control(self.migrate_engine,
- self.REPOSITORY,
- self.INIT_VERSION)
- self.assertEqual(self.INIT_VERSION,
- self.migration_api.db_version(self.migrate_engine,
- self.REPOSITORY))
-
- LOG.debug('latest version is %s', self.REPOSITORY.latest)
- versions = range(int(self.INIT_VERSION) + 1,
- int(self.REPOSITORY.latest) + 1)
-
- for version in versions:
- # upgrade -> downgrade -> upgrade
- self.migrate_up(version, with_data=True)
- if snake_walk:
- downgraded = self.migrate_down(version - 1, with_data=True)
- if downgraded:
- self.migrate_up(version)
-
- if downgrade:
- # Now walk it back down to 0 from the latest, testing
- # the downgrade paths.
- for version in reversed(versions):
- # downgrade -> upgrade -> downgrade
- downgraded = self.migrate_down(version - 1)
-
- if snake_walk and downgraded:
- self.migrate_up(version)
- self.migrate_down(version - 1)
-
- def migrate_down(self, version, with_data=False):
- """Migrate down to a previous version of the db.
-
- :param version: id of revision to downgrade.
- :type version: str
- :keyword with_data: Whether to verify the absence of changes from
- migration(s) being downgraded, see
- :ref:`Auxiliary Methods <auxiliary-dynamic-methods>`.
- :type with_data: Bool
- """
-
- try:
- self.migration_api.downgrade(self.migrate_engine,
- self.REPOSITORY, version)
- except NotImplementedError:
- # NOTE(sirp): some migrations, namely release-level
- # migrations, don't support a downgrade.
- return False
-
- self.assertEqual(version, self.migration_api.db_version(
- self.migrate_engine, self.REPOSITORY))
-
- # NOTE(sirp): `version` is what we're downgrading to (i.e. the 'target'
- # version). So if we have any downgrade checks, they need to be run for
- # the previous (higher numbered) migration.
- if with_data:
- post_downgrade = getattr(
- self, "_post_downgrade_%03d" % (version + 1), None)
- if post_downgrade:
- post_downgrade(self.migrate_engine)
-
- return True
-
- def migrate_up(self, version, with_data=False):
- """Migrate up to a new version of the db.
-
- :param version: id of revision to upgrade.
- :type version: str
- :keyword with_data: Whether to verify the applied changes with data,
- see :ref:`Auxiliary Methods <auxiliary-dynamic-methods>`.
- :type with_data: Bool
- """
- # NOTE(sdague): try block is here because it's impossible to debug
- # where a failed data migration happens otherwise
- try:
- if with_data:
- data = None
- pre_upgrade = getattr(
- self, "_pre_upgrade_%03d" % version, None)
- if pre_upgrade:
- data = pre_upgrade(self.migrate_engine)
-
- self.migration_api.upgrade(self.migrate_engine,
- self.REPOSITORY, version)
- self.assertEqual(version,
- self.migration_api.db_version(self.migrate_engine,
- self.REPOSITORY))
- if with_data:
- check = getattr(self, "_check_%03d" % version, None)
- if check:
- check(self.migrate_engine, data)
- except exc.DBMigrationError:
- msg = "Failed to migrate to version %(ver)s on engine %(eng)s"
- LOG.error(msg, {"ver": version, "eng": self.migrate_engine})
- raise
-
-
class ModelsMigrationsSync(object, metaclass=abc.ABCMeta):
"""A helper class for comparison of DB migration scripts and models.
diff --git a/oslo_db/sqlalchemy/utils.py b/oslo_db/sqlalchemy/utils.py
index 83a2cbd..ba0a607 100644
--- a/oslo_db/sqlalchemy/utils.py
+++ b/oslo_db/sqlalchemy/utils.py
@@ -39,13 +39,11 @@ from sqlalchemy import Index
from sqlalchemy import inspect
from sqlalchemy import Integer
from sqlalchemy import MetaData
-from sqlalchemy import PrimaryKeyConstraint
from sqlalchemy.sql.expression import cast
from sqlalchemy.sql.expression import literal_column
from sqlalchemy.sql import text
from sqlalchemy import String
from sqlalchemy import Table
-from sqlalchemy.types import NullType
from oslo_db._i18n import _
from oslo_db import exception
@@ -551,155 +549,6 @@ def _restore_indexes_on_deleted_columns(engine, table_name, indexes):
new_index.create(engine)
-@debtcollector.removals.remove(
- message='This API is intended for use with sqlalchemy-migrate, support '
- 'for which is deprecated for removal; it will be removed in a future '
- 'release',
- version='10.1.0',
-)
-def change_deleted_column_type_to_boolean(engine, table_name,
- **col_name_col_instance):
- if engine.name == "sqlite":
- return _change_deleted_column_type_to_boolean_sqlite(
- engine, table_name, **col_name_col_instance)
- indexes = get_indexes(engine, table_name)
- table = get_table(engine, table_name)
-
- old_deleted = Column('old_deleted', Boolean, default=False)
- table.metadata.bind = engine
- try:
- old_deleted.create(table, populate_default=False)
- finally:
- table.metadata.bind = None
-
- with engine.connect() as conn, conn.begin():
- conn.execute(
- table.update().where(
- table.c.deleted == table.c.id
- ).values(old_deleted=True)
- )
-
- table.metadata.bind = engine
- try:
- table.c.deleted.drop()
- table.c.old_deleted.alter(name="deleted")
- finally:
- table.metadata.bind = None
-
- _restore_indexes_on_deleted_columns(engine, table_name, indexes)
-
-
-def _change_deleted_column_type_to_boolean_sqlite(engine, table_name,
- **col_name_col_instance):
- table = get_table(engine, table_name)
- columns = []
- for column in table.columns:
- column_copy = None
- if column.name != "deleted":
- if isinstance(column.type, NullType):
- column_copy = _get_not_supported_column(col_name_col_instance,
- column.name)
- else:
- # FIXME(stephenfin): We shouldn't be using this private API;
- # figure out how else to copy an arbitrary column schema
- column_copy = column._copy()
- else:
- column_copy = Column('deleted', Boolean, default=0)
- columns.append(column_copy)
-
- # FIXME(stephenfin): We shouldn't be using this private API;
- # figure out how else to copy an arbitrary column schema
- # NOTE(stephenfin): We drop PrimaryKeyConstraint-type constraints since
- # these duplicate the 'primary_key=True' attribute on the speicified
- # column(s). This technically breaks things when the primary key covers
- # multiple columns but that's okay: these are deprecated APIs
- constraints = [
- constraint._copy() for constraint in table.constraints
- if not isinstance(constraint, PrimaryKeyConstraint)
- ]
-
- with engine.connect() as conn:
- meta = table.metadata
- new_table = Table(
- table_name + "__tmp__", meta,
- *(columns + constraints))
-
- with conn.begin():
- new_table.create(conn)
-
- indexes = []
- for index in get_indexes(engine, table_name):
- column_names = [new_table.c[c] for c in index['column_names']]
- indexes.append(
- Index(index["name"], *column_names, unique=index["unique"])
- )
-
- c_select = []
- for c in table.c:
- if c.name != "deleted":
- c_select.append(c)
- else:
- c_select.append(table.c.deleted == table.c.id)
-
- with conn.begin():
- table.drop(conn)
- for index in indexes:
- index.create(conn)
-
- table.metadata.bind = engine
- try:
- new_table.rename(table_name)
- finally:
- table.metadata.bind = None
-
- with conn.begin():
- conn.execute(
- new_table.update().where(
- new_table.c.deleted == new_table.c.id
- ).values(deleted=True)
- )
-
-
-@debtcollector.removals.remove(
- message='This API is intended for use with sqlalchemy-migrate, support '
- 'for which is deprecated for removal; it will be removed in a future '
- 'release',
- version='10.1.0',
-)
-def change_deleted_column_type_to_id_type(engine, table_name,
- **col_name_col_instance):
- if engine.name == "sqlite":
- return _change_deleted_column_type_to_id_type_sqlite(
- engine, table_name, **col_name_col_instance)
- indexes = get_indexes(engine, table_name)
- table = get_table(engine, table_name)
-
- new_deleted = Column('new_deleted', table.c.id.type,
- default=_get_default_deleted_value(table))
- table.metadata.bind = engine
- try:
- new_deleted.create(table, populate_default=True)
- finally:
- table.metadata.bind = None
-
- table.metadata.bind = engine
- try:
- with engine.connect() as conn, conn.begin():
- deleted = True # workaround for pyflakes
- conn.execute(
- table.update().where(
- table.c.deleted == deleted
- ).values(new_deleted=table.c.id)
- )
-
- table.c.deleted.drop()
- table.c.new_deleted.alter(name="deleted")
-
- _restore_indexes_on_deleted_columns(engine, table_name, indexes)
- finally:
- table.metadata.bind = None
-
-
def _is_deleted_column_constraint(constraint):
# NOTE(boris-42): There is no other way to check is CheckConstraint
# associated with deleted column.
@@ -713,92 +562,6 @@ def _is_deleted_column_constraint(constraint):
return bool(re.match(r".*deleted in \(.*\)", sqltext, re.I))
-def _change_deleted_column_type_to_id_type_sqlite(engine, table_name,
- **col_name_col_instance):
- # NOTE(boris-42): sqlalchemy-migrate can't drop column with check
- # constraints in sqlite DB and our `deleted` column has two check
- # constraints. There is only one way to remove these constraints:
- #
- # 1) Create new table with the same columns, constraints and indexes.
- # (except deleted column).
- # 2) Copy all data from old to new table.
- # 3) Drop old table.
- # 4) Rename new table to old table name.
- meta = MetaData()
- table = Table(table_name, meta, autoload_with=engine)
- default_deleted_value = _get_default_deleted_value(table)
-
- columns = []
- for column in table.columns:
- column_copy = None
- if column.name != "deleted":
- if isinstance(column.type, NullType):
- column_copy = _get_not_supported_column(col_name_col_instance,
- column.name)
- else:
- # FIXME(stephenfin): We shouldn't be using this private API;
- # figure out how else to copy an arbitrary column schema
- column_copy = column._copy()
- else:
- column_copy = Column('deleted', table.c.id.type,
- default=default_deleted_value)
- columns.append(column_copy)
-
- constraints = []
- for constraint in table.constraints:
- if not (
- _is_deleted_column_constraint(constraint) or
- isinstance(constraint, PrimaryKeyConstraint)
- ):
- # FIXME(stephenfin): We shouldn't be using this private API;
- # figure out how else to copy an arbitrary constraint schema
- constraints.append(constraint._copy())
-
- with engine.connect() as conn:
- # we need separate transactions, since we must create the table before
- # we can copy entries into it (later)
- with conn.begin():
- new_table = Table(
- table_name + "__tmp__", meta,
- *(columns + constraints),
- )
- new_table.create(conn)
-
- indexes = []
- for index in get_indexes(engine, table_name):
- column_names = [new_table.c[c] for c in index['column_names']]
- indexes.append(
- Index(index["name"], *column_names, unique=index["unique"])
- )
-
- with conn.begin():
- table.drop(conn)
- for index in indexes:
- index.create(conn)
-
- with conn.begin():
- new_table.metadata.bind = engine
- try:
- new_table.rename(table_name)
- finally:
- new_table.metadata.bind = None
-
- deleted = True # workaround for pyflakes
- conn.execute(
- new_table.update().where(
- new_table.c.deleted == deleted
- ).values(deleted=new_table.c.id)
- )
-
- # NOTE(boris-42): Fix value of deleted column: False -> "" or 0.
- deleted = False # workaround for pyflakes
- conn.execute(
- new_table.update().where(
- new_table.c.deleted == deleted
- ).values(deleted=default_deleted_value)
- )
-
-
def get_db_connection_info(conn_pieces):
database = conn_pieces.path.strip('/')
loc_pieces = conn_pieces.netloc.split('@')