[dataimport] backport massive store from dataio cube
Some tweaks have been made to work with 3.21 foreign key constraints.
Also test are not executed in our own pg cluster instead of polluting the system
one.
Closes #5414760
[tox]
env = py27
[testenv]
sitepackages = True
commands = pytest -t {envname}/test {posargs}
[testenv:cubicweb]
deps =
-r{toxinidir}/test/requirements.txt
commands = pytest -t test {posargs}
[testenv:dataimport]
[testenv:devtools]
deps =
-r{toxinidir}/devtools/test/requirements.txt
[testenv:entities]
deps =
-r{toxinidir}/entities/test/requirements.txt
[testenv:etwist]
deps =
-r{toxinidir}/etwist/test/requirements.txt
[testenv:ext]
deps =
-r{toxinidir}/ext/test/requirements.txt
[testenv:hooks]
deps =
-r{toxinidir}/hooks/test/requirements.txt
[testenv:server]
deps =
-r{toxinidir}/server/test/requirements.txt
[testenv:sobjects]
deps =
-r{toxinidir}/sobjects/test/requirements.txt
[testenv:web]
deps =
-r{toxinidir}/web/test/requirements.txt
[testenv:wsgi]
deps =
-r{toxinidir}/wsgi/test/requirements.txt
[testenv:doc]
changedir = doc
whitelist_externals =
sphinx-build
deps =
sphinx
commands = sphinx-build -b html -d {envtmpdir}/doctrees . {envtmpdir}/html