postgreSQL database version control tool
Project description
Database pgvctrl is a tool designed to help deploy changes to postgres databases. All changes are stored in versioned folders with the order of execution set by the programmer.
Prerequisites:
Getting started:
pip install
pip install pgvctrl
github
Download or clone the repository
Install
python setup.py install
Get help:
pgvctrl -h
Get version:
pgvctrl -version
Running the tests
In the test directory:
pytest
Getting Started
If you don’t already have a database, create one on your postgres server.
Create pgvctrl dbRepoConfig:
Make a directory where you want you database repositories to live.
pgvctrl -mkconf
This will create a dbRepoConfig.json file.
Create database repository:
In the same directory as the dbRepoConfig.json file, run:
pgvctrl -mkrepo [repository name]
e.g
pgvctrl -mkrepo mydb
Output:
Repository created: mydb
What just happened?
There will be a folder structure: [my dir]/databases/[repository name]/ created.
The dbRepoConfig.json file will be updated to reflect the new repository.
Initialize database repository:
In the same directory as the dbRepoConfig.json file, run:
pgvctrl -init [db connection information] -repo [repository name]
For production databases:
pgvctrl -init [db connection information] -repo [repository name] -production
NOTE: Database connection information should include at a minimum.
Standard Information
-d [database name on server]
e.g.
pgvctrl -init -d mylocaldb -repo mydb
Other information as needed:
-host [postgres server host] -p [port] -u [database username] Use the -on-server flag when the above arguments are optional
Or
Service Information.pg_service
-svc [pg service information]
e.g.
pgvctrl -svc mydatabase:test -repo mydb
What just happened? After initialization is complete:
There will be a new table in your database named repository_version. This is where pgvctrl stores your repository name, version number with a version hash for each sql update file, environment name, revision (number of times the same version has been applied with different sql hash) and production flag.
Make repository version for repository: -mkv: Make version number:
pgvctrl -mkv [x.x.x.version_name] -repo [repository name]
e.g.:
pgvctrl -mkv 1.0.0.my_new_version -repo mydb
Output:
Version mydb/1.0.0.my_new_version created.
Remove repository version of repository: -rmv: Remove version number and files
pgvctrl -rmv [x.x.x.version_name] -repo [repository name]
e.g.:
pgvctrl -rmv 1.0.0.my_new_version -repo mydb
Output:
Do you want to remove the repository version? [YES/NO] Version mydb/1.0.0.my_new_version removed.
Create sql change files in the versioned directory! These files will be used to update your database and should have the naming convention of: [order number].[change name].sql e.g.: 100.AddedUserTable.sql
Notes:
For best results with sql files, wrap all statements in a Transactions.
BEGIN TRANSACTION; [Your sql changes] COMMIT;
List repositories and changes:
pgvctrl -rl
Output:
mydb v 1.0.0.my_new_version
Verbose:
pgvctrl -rlv
Output:
mydb v 0.0.0.my_new_version 100 AddUsersTable
List repository Schema Snapshots:
pgvctrl -lss or pgvctrl -list-schema-snapshots
Output:
mydb 1.0.0.my_new_version 5.21 KB
List repository database dumps:
pgvctrl -ldd or pgvctrl -list-database-dumps
Output:
mydb mydb.test.20190101 132.22 MB
When you are ready to apply your changes to your database:
pgvctrl -apply -v [version number] -repo [repository name] [db connection information]
e.g.
pgvctrl -apply -v 0.0.0 -repo mydb -d mylocaldb
Output:
Running: 100.AddUsersTable ... Running: 500.AddStatesTable
Notes:
If you are applying changes to a production database, you must use the -production flag.
What just happened?
All of the sql files with [number].[change name].sql were ran against your database.
The repository_version table was update with the new version hash.
SQL Error handling
SQL Error handling on -apply In the event of an SQL error, pgvctrl will attempt to run the rollback version of your sql.
e.g
100.AddUsers.sql
100.AddUsers_rollback.sql - rollback file for 100.AddUsers.sql
If your rollback file does not exist or fails, the -apply command fails and no sql after the first failing sql file will be ran.
If the rollback file succeeds, all other sql files will be ran until all files have been applied if they can be.
Working with environments:
Setting up environment versions in repositories help ensure versions get deployed to the proper database.
Making and setting environments.
-mkenv: Make environment type:
pgvctrl -mkenv [env_name] -repo [repository name]
e.g.:
pgvctrl -mkenv test -repo mydb
Output:
Repository environment created: mydb test
-setenv: Set environment type to a version:
pgvctrl -setenv [env_name] -v [x.x] -repo [repository name]
e.g.:
pgvctrl -setenv test -v 1.0.0 -repo mydb
Output:
Repository environment set: mydb test 1.0.0
-init database with environment:
pgvctrl -init [db connection information] -repo [repository name] -setenv [env_name]
For production databases:
pgvctrl -init [db connection information] -repo [repository name] -setenv [env_name] -production
Output:
Database initialized environment [env_name]
-apply using -env:
pgvctrl -apply -env [env_name] -repo [repository name] [db connection information]
e.g.
pgvctrl -apply -env test -repo mydb -d mylocaldb
Output:
Running: 100.AddUsersTable
...
Running: 500.AddStatesTable
Applied: mydb v 1.1.0.MyVersion.0
What else can pgvctrl do?
-chkver: Check the version and repo on a database:
pgvctrl -chkver -repo [repository name] [db connection information]
e.g:
pgvctrl -chkver -repo mydb -d mylocaldb
Output:
mydb: 0.0.0.first.0
-status: Check database repository version status:
pgvctrl -status -repo [repository name] [db connection information]
e.g:
pgvctrl -status -repo mydb -d mylocaldb
Output:
mydb
v 0.0.0.first ['test']
Applied 100.some_sql
Not Applied 200.some_sql
Different 300.some_sql
Missing 400.some_sql
Applied (whitish) - The sql file has been applied to the database.
Not Applied (green)- The sql file has not yet been applied to the database.
Different (orange) - The sql file has been applied to the database, but the file has been altered/updated.
Missing (red) - The file had been applied to the database, but was removed from the version.
-timer-on/-timer-off: Turn executions timer on/off for -apply, -applyss, -pulldata, -pushdata, -dump and -restore:
pgvctrl -timer-on
Output:
Execution Timer ON
pgvctrl -timer-off
Output:
Execution Timer OFF
What happens?
The “timeExecutions” value in dbRepoConfig.json is toggled
-rmenv: Remove environment type:
pgvctrl -rmenv [env_name] -repo [repository name]
e.g.:
pgvctrl -rmenv test -repo mydb
Output:
Repository environment removed: mydb test
-rmrepo: Remove Repository
pgvctrl -rmrepo [repository name]
e.g.:
pgvctrl -rmrepo test
Output:
Repository removed: test
Notes:
If this command does not remove the folder from database, you must remove it and its contents yourself. This is a safety measure.
Any repository folders left behind will be displayed as UNREGISTERED when the -rl option is used.
Manage schemas and tables in Schema Snapshots
Manage schemas (–schema, –exclude-schema, –rm-schema, –rm-exclude-schema):
Allows the user to say what schemas structures to include/exclude when Schema Snapshots are created.
The ‘rm’ arguments allow the user to remove schemas from the included and excluded lists.
To include a schema:
pgvctrl -n membership -repo pgvctrl_test
or
pgvctrl -schema membership -repo pgvctrl_test
Output:
Repository added: pgvctrl_test
include-schemas ['membership']
What happens?
The dbRepoConfig.json file with have the membership schema added to the includeSchemas list property of the “pgvctrl_test” repository
Manage table (–table, –exclude-table, –rm-table, –rmexclude-table):
Allows the user to say what tables structures to include/exclude when Schema Snapshots are created.
The ‘rm’ arguments allow the user to remove tables from the included and excluded lists.
To include a table:
pgvctrl -t membership.user -repo pgvctrl_test
or
pgvctrl -table membership.user -repo pgvctrl_test
Output:
Repository added: pgvctrl_test
include-table ['membership.user']
Notes:
If a table/schema is included and then later excluded, the table/schema is moved from included to exclude and vice versa.
Include table/schema works the same as with pg_dump.
Schema Snapshot (-getss, -applyss)
What are Schema Snapshots? Schema Snapshots are snapshots of the database structure (tables, views, functions ect.) at the time the snapshot was taken.
Notes:
There can be only one per repository version!
The table holding the repository information (repository_version) will be saved as an insert in the Schema Snapshot.
Currently, only the schema is saved with Schema Snapshots.
If there were database schema changes outside of pgvctrl, it will be captured in the Schema Snapshot.
Schema Snapshots should only be applied to empty databases.
-getss: Set version Schema Snapshot
-getss -repo [repository name] [db connection information] [--name=]
-applyss or -apply-schema-snapshot: Apply version Schema Snapshot
-applyss [Schema Snapshot Name] -repo [repository name] [db connection information]
Manage data (-pulldata, -pushdata)
There could be many reason why one would want to manage data:
Lookup tables.
Testing data.
Just because your boss wants you too.
-pulldata: Pull data from repository by table
-pulldata [-dt [table name]] -repo [repository name] [db connection information]
e.g.
-pulldata -dt error_set -dt membership.user_state -repo mydb -d mylocaldb
Output:
Pulling: error_set
Pulling: membership.user_state
What happens?
The data folder for the repository is created.
One sql file per table is created with the table name was the file name.
A data.json file is created in data folder as well.
Notes: If you are just setting up data pulls for the first time, you can add one or more tables with the [-t [table name]] option.
-pushdata: Push data from repository to database
Once you have your data in your repository, pushing data is easy.
-pushdata -repo [repository name] [db connection information] [--force]
e.g. For pushing by table(s).
-pushdata -dt error_set -dt process_state -repo mydb -d mylocaldb
e.g. For pushing all tables.
-pushdata -repo mydb -d mylocaldb
Output:
Pushing Data
Running: error_set.sql
Notes: For interdependent data pushes, create _pre_push.sql and _post_push.sql files in the data folder to have pgvctrl execute before and after the data push.
e.g.
data/
_pre_push.sql
error_set.sql
membership.user_state.sql
_post_push.sql
-dump: Dump the repositories database
You can dump the database based on the repository backing it. This means includes/excludes for schemas and tables are honored during the database backup.
-dump -repo [repository name] [db connection information] [--name=]
e.g. For dumping the database.
-dump -repo mydb -d mylocaldb
-dump -repo mydb -d mylocaldb --name mybackup
Output:
Do you want to dump the database? [YES/NO]
:[Type YES]
Repository mydb database backed up
What happens?
The _databaseBackup/[repository name] folder is created if it doesn’t exist.
The backup [repository name][.environment].[string date] file is created.
-restore: Restore a repositories database from -dump
You can restore a repositories database based on a previous repository database dump.
-restore [repository name][.environment].[string date] -repo [repository name] [db connection information]
e.g. For dumping the database.
-restore mylocaldb.test.20190101 -repo mydb -d mylocaldb
Output:
Do you want to restore the database? [YES/NO]
:[Type YES]
Database mylocaldb.20190101 from repository mydb restored ['-d', 'mylocaldb'].
What happens?
The _databaseBackup/[repository name]/[dump file] file is used to fill the empty database at [db connection information].
Notes:
Database for restore should an empty databases.
dbRepoConfig.json
The dbRepoConfig.json files is the configuration file for your repositories. The defaultVersionStorage object is used to build the table that stores your repository information in the database on initialization. Each repository can be set up with different repository table structures as you see fit. The root setting tells pgvctrl where to look for the repositories.
{
"defaultVersionStorage": {
"env": "env",
"isProduction": "is_production",
"repository": "repository_name",
"revision": "revision",
"table": "repository_version",
"tableOwner": null,
"version": "version",
"versionHash": "version_hash"
},
"dumpDatabaseOptionsDefault": "-Fc -Z4",
"repositories": [
{
"dumpDatabaseOptions": "-Fc -Z4",
"envs": {
"your_test": "1.0.1",
"your_qa": "1.0.0",
"your_prod": "0.9.0"
},
"name": "YouRepoName",
"restoreDatabaseOptions": "-Fc -j 8",
"versionStorage": {
"env": "env",
"isProduction": "is_production",
"repository": "repository_name",
"revision": "revision",
"table": "repository_version",
"tableOwner": null,
"version": "version",
"versionHash": "version_hash"
}
}
],
"restoreDatabaseOptionsDefault": "-Fc -j 8",
"root": "databases",
"timeExecutions": false
}
data.json
The data.json file holds the list of tables for pushing and pulling data to and from your database. The column-inserts setting tells pgvctrl to create the table with insert statements if set to true. If false, the table is created with postgres copy. When data is pushed to the database, all relationships are dropped and recreated when the copy is complete.
Example data.json file:
[
{
"apply-order": 0,
"column-inserts": true,
"table": "error_set"
},
{
"apply-order": 0,
"column-inserts": true,
"table": "membership.user_state"
}
]
License
This project is licensed under the MIT License, provided in repository.
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
File details
Details for the file pgvctrl-1.2.0.tar.gz
.
File metadata
- Download URL: pgvctrl-1.2.0.tar.gz
- Upload date:
- Size: 49.9 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: Python-urllib/3.6
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 29bff492c5c9b1b1801e4e922026c8ad2b8b5c2548733c22adfd1ab5a0feae1b |
|
MD5 | 4f56145e52bb3665610de21823dd57f4 |
|
BLAKE2b-256 | 539c3c99976e4d5dddc70b01951cf336f200d78ec11242e7052285546de5d325 |