You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi everyone,
I'm working on a cache memory simulator written in C++. As part of the logic, I call a Python function using PyBind11 to integrate a machine learning model built with PyTorch. However, I've noticed that running the simulator with this Python integration is about 4x slower compared to running it without PyBind.
I suspect that the slowdown is due to the communication overhead between C++ and Python. How can I accurately measure this overhead? Are there any tools or profiling techniques that could help diagnose the issue?
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
-
Hi everyone,
I'm working on a cache memory simulator written in C++. As part of the logic, I call a Python function using PyBind11 to integrate a machine learning model built with PyTorch. However, I've noticed that running the simulator with this Python integration is about 4x slower compared to running it without PyBind.
I suspect that the slowdown is due to the communication overhead between C++ and Python. How can I accurately measure this overhead? Are there any tools or profiling techniques that could help diagnose the issue?
Any suggestions would be greatly appreciated
Beta Was this translation helpful? Give feedback.
All reactions