Code from https://github.com/francesco-vaselli/FlowSim/tree/benchmark
\ No newline at end of file
The present benchmark replicates the inference step of the CMS FlashSim simulation framework.
A ML model is deployed to simulate an arbitrary number of reconstructed particle jets starting from a collection of generator-level jets.
Models can also be deployed sequentially to replicate the simulation of different, correlated object collections in FlashSim.
Code from https://github.com/francesco-vaselli/FlowSim/tree/benchmark
The figure of merit being tracked is the throughput measured in object simulated per second. A series of other stats related to CPU/GPU usage are also being logged.
The script accepts the following arguments:
'--n-objects' the Number of objects to simulate
'--batch-size', the Batch size for inference
'--device', the Device to run on (cpu/cuda/vulkan)
'--n-model-instances' the Number of model instances to run
'--gpu-memory-limit' the GPU memory limit in GB
'--num-threads' the Number of PyTorch threads to use
'--monitor-interval' the Resource monitoring interval in seconds