jobmanager/tests/test_persistentData.py

325 lines
9.9 KiB
Python
Raw Normal View History

#!/usr/bin/env python
# -*- coding: utf-8 -*-
from __future__ import division, print_function
import sys
import pickle
from os import rmdir, remove
from os.path import abspath, dirname, split, exists
# Add parent directory to beginning of path variable
sys.path = [split(dirname(abspath(__file__)))[0]] + sys.path
import jobmanager.persistentData as pd
from jobmanager.persistentData import PersistentDataStructure as PDS
VERBOSE = 1
if sys.version_info[0] == 2:
# fixes keyword problems with python 2.x
old_open = open
def new_open(file, mode):
old_open(name = file, mode = mode)
open = new_open
def test_pd():
with PDS(name='test_data', verbose=VERBOSE) as data:
key = 'a'
value = 1
data.setData(key=key, value=value)
assert data.getData(key) == value
key_sub = 'zz'
with data.getData(key_sub, create_sub_data=True) as sub_data:
sub_data.setData(key=key, value=3)
assert sub_data.getData(key) == 3
assert data.getData(key) == 1
with sub_data.getData(key_sub, create_sub_data=True) as sub_sub_data:
sub_sub_data.setData(key=key, value=4)
assert sub_sub_data.getData(key) == 4
assert sub_data.getData(key) == 3
assert data.getData(key) == 1
with sub_data.getData(key_sub, create_sub_data=True) as sub_sub_data:
assert sub_sub_data.getData(key) == 4
assert sub_data.getData(key) == 3
assert data.getData(key) == 1
data._consistency_check()
data.erase()
def test_reserved_key_catch():
with PDS(name='data', verbose=VERBOSE) as data:
# TRY TO READ RESERVED KEYS
try:
a = data[pd.KEY_COUNTER]
except RuntimeError:
pass
else:
assert False
try:
b = data[pd.KEY_SUB_DATA_KEYS]
except RuntimeError:
pass
else:
assert False
# TRY TO SET RESERVED KEYS
try:
data[pd.KEY_COUNTER] = 4
except RuntimeError:
pass
else:
assert False
try:
data[pd.KEY_SUB_DATA_KEYS] = 4
except RuntimeError:
pass
else:
assert False
# TRY TO REMOVE RESERVED KEYS
try:
del data[pd.KEY_COUNTER]
except RuntimeError:
pass
else:
assert False
try:
del data[pd.KEY_SUB_DATA_KEYS]
except RuntimeError:
pass
else:
assert False
data.erase()
def test_pd_bytes():
import pickle
t1 = (3.4, 4.5, 5.6, 6.7, 7.8, 8.9)
t2 = (3.4, 4.5, 5.6, 6.7, 7.8, 8.9, 9,1)
b1 = pickle.dumps(t1)
b2 = pickle.dumps(t2)
with PDS(name='base', verbose=VERBOSE) as base_data:
with base_data.getData(key=b1, create_sub_data=True) as sub_data:
for i in range(2, 10):
sub_data[i] = t2
base_data[b2] = t1
if VERBOSE > 1:
print("\nCHECK\n")
with PDS(name='base', verbose=VERBOSE) as base_data:
with base_data.getData(key=b1) as sub_data:
for i in range(2, 10):
assert sub_data[i] == t2
assert base_data[b2] == t1
base_data._consistency_check()
base_data.erase()
def test_directory_removal():
with PDS(name='data', verbose=VERBOSE) as data:
with data.newSubData('s1') as s1:
s1['bla'] = 9
2015-01-27 19:48:14 +01:00
f = open(file=data._dirname + '/other_file', mode='w')
f.close()
data.erase()
2015-01-27 19:48:14 +01:00
assert exists(data._dirname)
remove(data._dirname + '/other_file')
rmdir(data._dirname)
def test_mp_read_from_sqlite():
import sqlitedict as sqd
import multiprocessing as mp
import time
d = sqd.SqliteDict('test.db', autocommit = True)
d.terminate()
def write(arg):
with sqd.SqliteDict('test.db', autocommit = True) as d:
for i in range(100):
d[i] = (i, arg)
def read():
with sqd.SqliteDict('test.db', autocommit = True) as d:
for i in range(len(d)):
print(i, d[i])
p1 = mp.Process(target = write, args=('p1', ))
time.sleep(0.1)
p2 = mp.Process(target = read)
p1.start()
p2.start()
p1.join()
p2.join()
from collections import namedtuple
a_tup_type = namedtuple('a_tup_type', ['a', 'b'])
def test_dict_dump():
a = {'a': 1, 'b': 2}
print(pickle.dumps(a))
a_tup = a_tup_type(**a)
print(pickle.dumps(a_tup))
print(hash(a_tup))
def test_from_existing_sub_data():
t1 = (3.4, 4.5, 5.6, 6.7, 7.8, 8.9)
t2 = (3.4, 4.5, 5.6, 6.7, 7.8, 8.9, 9,1)
b1 = pickle.dumps(t1)
b2 = pickle.dumps(t2)
with PDS(name='base', verbose=VERBOSE) as base_data:
with base_data.getData(key='sub1', create_sub_data = True) as sub_data:
sub_data[100] = t1
sub_data[200] = t2
with sub_data.getData(key = 'subsub1', create_sub_data = True) as sub_sub_data:
sub_sub_data['t'] = 'hallo Welt'
base_data.setDataFromSubData(key='sub2', subData = sub_data)
with PDS(name='base', verbose=VERBOSE) as base_data:
with base_data.getData(key='sub2', create_sub_data = False) as sub_data:
assert sub_data[100] == t1
assert sub_data[200] == t2
with sub_data.getData(key = 'subsub1', create_sub_data = False) as sub_sub_data:
assert sub_sub_data['t'] == 'hallo Welt'
with PDS(name='base', verbose=VERBOSE) as base_data:
with base_data.getData(key='sub1', create_sub_data = True) as sub_data:
base_data['sub2'] = sub_data
with PDS(name='base', verbose=VERBOSE) as base_data:
with base_data.getData(key='sub2', create_sub_data = False) as sub_data:
assert sub_data[100] == t1
assert sub_data[200] == t2
with sub_data.getData(key = 'subsub1', create_sub_data = False) as sub_sub_data:
assert sub_sub_data['t'] == 'hallo Welt'
sub_sub_data['t'] = 'sub2:hallo Welt'
sub_data[100] = "sub2:t1"
sub_data[200] = "sub2:t2"
with PDS(name='base', verbose=VERBOSE) as base_data:
with base_data.getData(key='sub1', create_sub_data = True) as sub_data:
assert sub_data[100] == t1
assert sub_data[200] == t2
with sub_data.getData(key = 'subsub1', create_sub_data = True) as sub_sub_data:
assert sub_sub_data['t'] == 'hallo Welt'
with PDS(name='base', verbose=VERBOSE) as base_data:
with base_data.getData(key='sub2', create_sub_data = False) as sub_data:
with sub_data.getData(key = 'subsub1', create_sub_data = False) as sub_sub_data:
assert sub_sub_data['t'] == 'sub2:hallo Welt'
assert sub_data[100] == "sub2:t1"
assert sub_data[200] == "sub2:t2"
base_data._consistency_check()
base_data.erase()
def test_remove_sub_data_and_check_len():
with PDS(name='base', verbose=VERBOSE) as base_data:
with base_data.getData(key='sub1', create_sub_data = True) as sub_data:
sub_data[100] = 't1'
sub_data[200] = 't2'
with sub_data.getData(key = 'subsub1', create_sub_data = True) as sub_sub_data:
sub_sub_data['t'] = 'hallo Welt'
assert len(sub_data) == 3
assert len(base_data) == 1
base_data['copy_of_sub1'] = sub_data
assert len(base_data) == 2
del base_data['sub1']
assert len(base_data) == 1
with base_data.getData(key='copy_of_sub1', create_sub_data = True) as sub_data:
assert len(sub_data) == 3
assert sub_data[100] == 't1'
assert sub_data[200] == 't2'
with sub_data.getData(key = 'subsub1', create_sub_data = True) as sub_sub_data:
assert sub_sub_data['t'] == 'hallo Welt'
assert ('sub1' not in base_data)
base_data._consistency_check()
base_data.erase()
2015-01-27 19:48:14 +01:00
def test_show_stat():
with PDS(name='test_data', verbose=VERBOSE) as data:
key = 'a'
value = 1
data.setData(key=key, value=value)
assert data.getData(key) == value
key_sub = 'zz'
with data.getData(key_sub, create_sub_data=True) as sub_data:
sub_data.setData(key=key, value=3)
assert sub_data.getData(key) == 3
assert data.getData(key) == 1
2015-01-27 19:48:14 +01:00
key_sub_bin = pickle.dumps(key_sub, protocol=2)
with sub_data.getData(key_sub_bin, create_sub_data=True) as sub_sub_data:
sub_sub_data.setData(key=key, value=4)
assert sub_sub_data.getData(key) == 4
assert sub_data.getData(key) == 3
assert data.getData(key) == 1
with sub_data.getData(key_sub_bin, create_sub_data=True) as sub_sub_data:
assert sub_sub_data.getData(key) == 4
assert sub_data.getData(key) == 3
assert data.getData(key) == 1
data._consistency_check()
data.show_stat(recursive=True)
data.erase()
if __name__ == "__main__":
2015-01-27 19:48:14 +01:00
test_reserved_key_catch()
test_pd()
test_pd_bytes()
2015-01-27 19:48:14 +01:00
test_directory_removal()
test_mp_read_from_sqlite()
test_dict_dump()
test_from_existing_sub_data()
test_remove_sub_data_and_check_len()
2015-01-27 19:48:14 +01:00
test_show_stat()