Alex Seceleanu

Results 3 comments of Alex Seceleanu

I tried reducing the block size in the profiler context, but with no luck. I get 1.9G torch_trace.json.

Tested the same steps with the latest version of Chakra, installed from repository, 16 Oct, the behavior is the same.

While looking into this issue I observed that the nccl:broadcast operation is a CPU operation and therefore is does not pass this check from pytorch_converter.py: ``` def get_protobuf_node_type_from_json_node( self, json_node_map:...