Skip to main content

High speed database with key / data in python.

Project description

pydbase (see blockchain at the end)

High speed database with key / data

The motivation was to create a no frills way of saving / retrieving data. It is fast, and the time test shows that this is an order of magnitude faster than most mainstream databases. This is due to the engine's simplicity. It avoids expensive computations in favor of quickly saving data.

Fast data save / retrieve

Mostly ready for production. All tests pass. Please use caution, as this is new. The command line tester can drive most aspects of this API; and it is somewhat complete. It is also good way to see the API / Module in action.

API

The module 'twincore' uses two data files and a lock file. The file names are generated from the base name of the data file; name.pydb for data; name.pidx for the index, name.lock for the lock file. In case of frozen process the lock file times out in xx seconds and breaks the lock. If the locking process (id in lockfile) does not exist, the lock breaks immediately.

Setting verbosity and debug level:

twincore.core_quiet   = quiet
twincore.core_verbose = verbose
twincore.core_pgdebug = pgdebug
twincore.core_showdel = sdelx

(Setting before data creation will display mesages from the construtor)

Example DB creation:

core = twincore.TwinCore(datafile_name)

Some basic ops:

dbsize = core.getdbsize()

core.save_data(keyx, datax)
rec_arr = core.retrieve(keyx, ncount)
print("rec_arr", rec_arr)

Example chain DB creation:

core = twinchain.TwinChain(datafile_name)
core.append(keyx, datax)
recnum = core.getdbsize()
rec = core.get_payload(recnum)
print(recnum, rec)

Structure of the data:

32 byte header, starting with FILESIG

4 bytes    4 bytes          4 bytes         Variable
------------------------------------------------------------
RECSIG     Hash_of_key      Len_of_key      DATA_for_key
RECSEP     Hash_of_payload  Len_of_payload  DATA_for_payload

    .
    .

RECSIG     Hash_of_key      Len_of_key      DATA_for_key
RECSEP     Hash_of_payload  Len_of_payload  DATA_for_payload

where:

RECSIG="RECB" (record begin here)
RECSEP="RECS" (record separated here)
RECDEL="RECX" (record deleted)

Deleted records are marked with the RECSIG mutated from RECB to RECX

  Vacuum will remove the deleted records; Make sure your database has no
pending ops; or non atomic opts when vacuuming;

    (like: find keys - delete keys in two ops)

  New data is appended to the end, no duplicate filtering is done.
Retrieval is searched from reverse, the latest record with this key
is retrieved first. Most of the times this behavior is what we
want; also the record history is kept this way, also a desirable
behavior.

Usage:

The DB exerciser

The file dbaseadm.py exercises most of the twincore functionality. It also provides examples of how to drive it.

The command line utility's help response:

Usage: dbaseadm.py [options] [arg_key arg_data]
 Options: -h         help (this screen)   -|-  -G  num  get record by number
          -V         print version        -|-  -q  quiet on, less printting
          -d         debug level (0-10)   -|-  -v  increment verbosity level
          -r         randomize data       -|-  -w  write fixed record(s)
          -z         dump backwards(s)    -|-  -i  show deleted record(s)
          -U         Vacuum DB            -|-  -R  reindex / recover DB
          -I         DB Integrity check   -|-  -c  set check integrity flag
          -s         Skip to count recs   -|-  -K  list keys only
          -y  key    find by key          -|-  -m  dump data to console
          -o  offs   get data from offset -|-  -e  offs   delete at offset
          -F  subkey find by sub str      -|-  -g  num    get number of recs.
          -k  key    key to save          -|-  -a  str    data to save
          -S         print num recs       -|-  -D  key    delete by key
          -n  num    number of records    -|-  -t  key    retrieve by key
          -p  num    skip number of recs  -|-  -u  rec    delete at recnum
          -l  lim    limit number of recs -|-
          -x  max    limit max number of records to get
          -f  file   input or output file (default: 'pydbase.pydb')
The verbosity level influences the amount of data presented.
Use quotes for multi word arguments.

The chain adm utility:

Usage: pychain.py [options]
   Options: -a  data   append data to the end of chain
            -h         help (this screen)
            -m         dump chain data
            -c         check data integrity
            -i         check link integrity
            -v         increase verbosity

Comparison to other databases:

This comparison is to show the time it takes to write 500 records. In the tests the record size is about the same (Hello, 1 /vs/ "Hello", 1) Please see the sqlite_test.sql for details of data output;

The test can be repeated with running the 'time.sh' script file. Please note the the time.sh clears all files test_data/* for a fair test.

    sqlite time test, writing 500 records ...
    real	0m1.730s
    user	0m0.110s
    sys	0m0.455s

    pydbase time test, writing 500 records ...
    real	0m0.120s
    user	0m0.075s
    sys	0m0.044s

    -rw-r--r-- 1 peterglen users  4032 Feb  7 15:35 pydb_test.pidx
    -rw-r--r-- 1 peterglen users 15032 Feb  7 15:35 pydb_test.pydb
    -rw-r--r-- 1 peterglen users 20480 Feb  7 15:35 sqlite_test.db

Please mind the fact that the sqlite engine has to do a lot of parsing which we skip doing; That is why pydbase is more than an order of magnitude faster ...

Saving more complex data

The database saves a key / value pair. However, the key can be mutated to contain meta data. (for example adding a string in front of it.) [like: CUST_ for customer data / details] Also the key can be made unique by adding a UUID to it.

The data can consist of any text / binary. The library pyvpacker.py can pack any data into a string; A copy of pyvpacker can be obtained from pip or github

pyvpacker.py

This module can pack arbitrary python data into a string; which can be used to store anything in the pydbase key / data sections.

Example from running testpacker.py:

    org: (1, 2, 'aa', ['bb', b'dd'])
    packed: pg s4 'iisa' i4 1 i4 2 s2 'aa' a29 'pg s2 'sb' s2 'bb' b4 'ZGQ=' '
    unpacked: [1, 2, 'aa', ['bb', b'dd']]
    rec_arr: pg s4 'iisa' i4 1 i4 2 s2 'aa' a29 'pg s2 'sb' s2 'bb' b4 'ZGQ=' '
    rec_arr_upacked: [1, 2, 'aa', ['bb', b'dd']]
    (Note: the decode returns an array of data; use data[0] to get the original)

There is also the option of using pyvpacker on the key itself. Because the key is identified by its hash, there is no speed penalty; Note that the hash is a 32 bit one; collisions are possible, however unlikely; To compensate, make sure you compare the key proper with the returned key.

Maintenance

The DB can rebuild its index and purge all deleted records. In the test utility the options are:

    ./dbaseadm.py -U     for vacuum (add -v for verbosity)

The database is re-built, the deleted entries are purged, the damaged data (if any) is saved into a separate file, created with the same base name as the data base, with the '.perr' extension.

    ./dbaseadm.py -R     for re-index

The index is recreated; as of the current file contents. This is useful if the index is lost (like copying the data only)

If there is a data file without the index, the re-indexing is called automatically. In case of deleted data file, pydbase will recognize the dangling index and nuke it by renaming it to orgfilename.pidx.dangle (Tue 07.Feb.2023 just deleted it);

The database grows with every record added to it. It does not check if the particular record already exists. It adds the new copy of the record to the end; Retrieving starts from the end, and the data retrieved (for this particular key) is the last record saved. All the other records of this key are also there in chronological (save) order. Miracle of record history archived by default.

To clean the old record history, one may delete all the records with this same key, except the last one.

Blockchain implementation

The database is extended with a blockhcain implementation. The new class is called twinchain; and it is a class derived from twincore.

To drive the blockchain, just use the append method. Th database will calculate all the hashes, integrate it into the existing chain with the new item getting a backlink field. This field is calulated based upon the previous record's hash and the previous record's frozen date. This assures that identical data will have a different hash, so data cannot be anticipated based upon its hash alone. The hash is done with 256 bits, and assumed to be very secure.

To drive it:

    core = twinchain.TwinChain()    # Takes an optional file name
    core.append("The payload")      # Arbitrary data

Block chain layer on top of twincore.

    prev     curr
        record
|   Time Now    |   Time  Now    |  Time Now     |
|   hash256   | |    hash256   | |   hash256   | |
|   Header    | |    Header    | |   Header    | |
|   Payload   | |    Payload   | |   Payload   | |
|   Backlink  | |    Backlink  | |   Backlink  | |
              |---->----|      |---->---|     |------ ...

The hashed sum of fields saved to the next backlink.

Integrity check

Two levels; Level one is checking if the record checksums are correct; Level two checks if the linkage is correct.

TODO

Speed this up by implementing this as a 'C' module

PyTest

The pytest passes with no errors; The following (and more) test are created / executed:

Test results:

============================= test session starts ==============================
platform linux -- Python 3.10.12, pytest-7.4.3, pluggy-1.0.0
rootdir: /home/xxxx/pgpygtk/pydbase
collected 33 items

test_bindata.py .                                                        [  3%]
test_chain.py .                                                          [  6%]
test_create.py .....                                                     [ 21%]
test_del.py .                                                            [ 24%]
test_dump.py .                                                           [ 27%]
test_find.py ..                                                          [ 33%]
test_findrec.py ..                                                       [ 39%]
test_getrec.py .                                                         [ 42%]
test_integrity.py .                                                      [ 45%]
test_list.py ..                                                          [ 51%]
test_lockrel.py ..                                                       [ 57%]
test_multi.py ..                                                         [ 63%]
test_packer.py ......                                                    [ 81%]
test_randdata.py .                                                       [ 84%]
test_reindex.py .                                                        [ 87%]
test_search.py ...                                                       [ 96%]
test_vacuum.py .                                                         [100%]

============================== 33 passed in 0.73s ==============================

History

1.1         Tue 20.Feb.2024     Initial release
1.2.0       Mon 26.Feb.2024     Moved pip home to pydbase/
1.4.0       Tue 27.Feb.2024     Addedd pgdebug
1.4.2       Wed 28.Feb.2024     Fixed multiple instances
1.4.3       Wed 28.Feb.2024     ChainAdm added

Errata

Chain is still in development, most of it functions well.
Not for production.

// EOF

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pydbase-1.4.3.tar.gz (29.6 kB view details)

Uploaded Source

Built Distribution

pydbase-1.4.3-py3-none-any.whl (24.6 kB view details)

Uploaded Python 3

File details

Details for the file pydbase-1.4.3.tar.gz.

File metadata

  • Download URL: pydbase-1.4.3.tar.gz
  • Upload date:
  • Size: 29.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.8.0 colorama/0.4.4 importlib-metadata/4.6.4 keyring/23.5.0 pkginfo/1.8.2 readme-renderer/34.0 requests-toolbelt/0.9.1 requests/2.28.2 rfc3986/1.5.0 tqdm/4.64.1 urllib3/1.26.5 CPython/3.10.12

File hashes

Hashes for pydbase-1.4.3.tar.gz
Algorithm Hash digest
SHA256 d503ac980c1b06d437bbdbaae22afc3f710fd54903799fc4fd684d942f485ec7
MD5 7f8d8d92f374fa3745aa3c4642b20806
BLAKE2b-256 a6fcdddd7dd4f36730a8cbb585ea1826f083c3d57bc4d49e68e431fc2ac1e3b2

See more details on using hashes here.

File details

Details for the file pydbase-1.4.3-py3-none-any.whl.

File metadata

  • Download URL: pydbase-1.4.3-py3-none-any.whl
  • Upload date:
  • Size: 24.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.8.0 colorama/0.4.4 importlib-metadata/4.6.4 keyring/23.5.0 pkginfo/1.8.2 readme-renderer/34.0 requests-toolbelt/0.9.1 requests/2.28.2 rfc3986/1.5.0 tqdm/4.64.1 urllib3/1.26.5 CPython/3.10.12

File hashes

Hashes for pydbase-1.4.3-py3-none-any.whl
Algorithm Hash digest
SHA256 81d068ade30e22eb490f12ea923e760b1c45532f18d16223ba84cf31b85c028d
MD5 1b4a51ee40e39d61d415fb372d9c8ac6
BLAKE2b-256 d96b67649e78311ac5877c5b74546983a3e7112c9515e5278faa6e19ae43dd56

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page