Cli_results
Simple lib to test results or script runs from command line.
Install
Install from pypi:
pip install cli_result
Or install from github repo:
pip install git+https://github.com/ayasyrev/cli_result.git
Usage.
Main purpose test results of examples run. We run all scripts in examples folder and compare results with expected results. Check it at different python versions.
So we can be sure that all scripts work and has similar behaviors in different python versions.
It's not suitable to run script that supposed to run for a long time or resources are limited.
But it's good for quick tests, to check configs and shorts demos (examples).
Put your script in examples folder and expected results in results folder.
Arguments for tests at file name same as script name + __args.txt.
from cli_result import check_examples, Cfg
errors = check_examples()
This run all scripts in examples folder with arguments from __args.txt
file and compare with results at results/
folder.
assert errors is None
Examples
We can change examples folder.
cfg = Cfg(examples_path="../examples/examples_extra/")
Check examples at folder:
from cli_result import get_examples
examples = get_examples(cfg=cfg)
We got list of examples as named tuple example_name, files
example = examples[0]
print(example.name)
print(example.files[0])
print(example[1][1])
output
example_extra_1
../examples/examples_extra/example_extra_1.py
../examples/examples_extra/example_extra_1__alter.py
Run script
We can run script and look at result.
from cli_result import run_script
result = run_script(
filename=example[1][0],
args="--help",
)
print(result.stdout)
output
usage: example_extra_1.py [-h] [--echo ECHO]
options:
-h, --help show this help message and exit
--echo ECHO
assert result.stderr == ""
Load expected result.
from cli_result import read_result, get_args
Load arguments for example.
get_args
returns list of Args
args = get_args(example.name, cfg)
print(args[0])
output
Args(name='help', list=['--help'])
expected = read_result(
name=example.name,
arg_name=args[0].name,
cfg=cfg,
)
Now we can compare results.
assert result == expected
Check one example.
We can check one example.
from cli_result import run_check_example
errors = run_check_example(
example_name=example.name,
file_list=example.files,
cfg=cfg,
)
assert errors is None
Alternatively we can check one as:
errors = check_examples(
names=example.name,
cfg=cfg,
)
assert errors is None
Check all examples.
Or check all examples.
errors = check_examples(cfg=cfg)
assert errors is None
Check errors
Lets look at example with error.
cfg = Cfg(examples_path="../tests/examples/examples_errors/")
errors = check_examples(cfg=cfg)
assert errors is not None
print(f"Examples with errors: {len(errors)}, {examples[0].name}: {len(errors[0].list)} errors")
output
Examples with errors: 1, example_extra_1: 10 errors
Let look at one of errors.
We got file name that has error, result of run and expected result. Now we can look for what is wrong.
example_error = errors[0]
print(example_error.name)
output
example_1
error = example_error.list[4]
print(error.argname)
print(error.filename)
output
empty_str
../tests/examples/examples_errors/example_1__err_1.py
We can compare result with expected result.
print(error.res)
output
usage: example_1__err_1.py [-h] [--e E]
example_1__err_1.py: error: unrecognized arguments: ""
And expected is:
print(error.exp)
output
usage: example_1.py [-h] [--echo ECHO]
example_1.py: error: unrecognized arguments: ""