Other tools#
Debugging#
enumerate_ort_run#
- onnx_extended.tools.ort_debug.enumerate_ort_run(onx: str | ModelProto, feeds: Dict[str, Any], verbose: int = 0) Iterator[Tuple[List[str], List[Any]]] [source]#
Yields all the intermediate results produced by onnxruntime.
- Parameters:
onx – model
feeds – input tensors
verbose – prints out a summary of the results
- Returns:
intermediate results and names
Profiling#
js_profile_to_dataframe#
- onnx_extended.tools.js_profile.js_profile_to_dataframe(filename: str, as_df: bool = True, first_it_out: bool = False, agg: bool = False, agg_op_name: bool = False, with_shape: bool = False) List | DataFrame [source]#
Profiles the execution of an onnx graph with onnxruntime.
- Parameters:
filename – filename holding the profiling stored in json format
as_df – returns the
first_it_out – if aggregated, leaves the first iteration out
agg – aggregate by event
agg_op_name – aggregate on operator name or operator index
with_shape – keep the shape before aggregating
- Returns:
DataFrame or dictionary
plot_ort_profile#
- onnx_extended.tools.js_profile.plot_ort_profile(df: DataFrame, ax0: matplotlib.axes.Axes | None = None, ax1: matplotlib.axes.Axes | None = None, title: str | None = None) matplotlib.axes.Axes [source]#
Plots time spend in computation based on dataframe produced by function
js_profile_to_dataframe()
.- Parameters:
df – dataframe
ax0 – first axis to draw time
ax1 – second axis to draw occurences
title – graph title
- Returns:
the graph
Testing#
save_for_benchmark_or_test#
- onnx_extended.tools.run_onnx.save_for_benchmark_or_test(folder: str, test_name: str, model: ModelProto, inputs: List[ndarray], outputs: List[ndarray] | None = None, data_set: int = 0) str [source]#
Saves input, outputs on disk to later uses it as a backend test or a benchmark.
- Parameters:
folder – folder to save
test_name – test name or subfolder
model – model to save
inputs – inputs of the node
outputs – outputs of the node, supposedly the expected outputs, if not speficied, they are computed with the reference evaluator
data_set – to have multiple tests with the same model
- Returns:
test folder
bench_virtual#
- onnx_extended.tools.run_onnx.bench_virtual(test_path: str, virtual_path: str, runtimes: List[str] | str = 'ReferenceEvaluator', index: int = 0, warmup: int = 5, repeat: int = 10, modules: List[Dict[str, str | None]] | None = None, verbose: int = 0, save_as_dataframe: str | None = None, filter_fct: Callable[[str, Dict[str, str | None]], bool] | None = None) List[Dict[str, float | Dict[str, Tuple[int, ...]]]] [source]#
Runs the same benchmark over different versions of the same packages in a virtual environment.
- Parameters:
test_path – test path
virtual_path – path to the virtual environment
runtimes – runtimes to measure (ReferenceEvaluation, CReferenceEvaluator, onnxruntime)
index – test index to measure
warmup – number of iterations to run before starting to measure the model
repeat – number of iterations to measure
modules – modules to install, example: modules=[{“onnxruntime”: “1.16.0”, “onnx”: “1.15.0”}]
filter_fct – to disable some of the configuration based on the runtime and the installed modules
verbose – verbosity
save_as_dataframe – saves as dataframe
- Returns:
list of statistics
TestRun#
- class onnx_extended.tools.run_onnx.TestRun(folder: str)[source]#
Loads a test saved by
save_for_benchmark_or_test()
.- Parameters:
folder – test folder
It has the following attributes:
folder: str
proto: ModelProto
datasets: Dict[int, List[Tuple[int, np.array]]]
- bench(f_build: Callable[[ModelProto], Any], f_run: Callable[[Any, Dict[str, array]], List[array]], index: int = 0, warmup: int = 5, repeat: int = 10) Dict[str, float | str | Any] [source]#
Runs the model on the given inputs.
- Parameters:
f_build – function to call to build the inference class
f_run – function to call to run the inference
index – test index to measure
warmup – number of iterations to run before starting to measure the model
repeat – number of iterations to measure
- Returns:
dictionary with many metrics, any metric endings with “_time” is a duration
- property input_names#
Returns the input names of the model.
- property output_names#
Returns the output names of the model.
- test(f_build: Callable[[ModelProto], Any], f_run: Callable[[Any, Dict[str, array]], List[array]], index: int = 0, exc: bool = True, atol: float = 1e-05, rtol: float = 1e-05) Dict[str, Tuple[float, float, str]] | None [source]#
Runs the tests.
- Parameters:
f_build – function to call to build the inference class
f_run – function to call to run the inference
index – test index
exc – raises an exception if the verification fails
atol – absolute tolerance
rtol – relative tolerance
- Returns:
list of results with discrepancies, the absolute error, the relative one and a reason for the failure