Skip to content

[benchmark_inference] Investigate the graph-breaks and split reasons encountered when running benchmark_inference.py #2681

@kshitij12345

Description

@kshitij12345

We observe graph-breaks when running TORCH_LOGS="graph_breaks" python thunder/benchmarks/benchmark_inference.py --input-length 32 --output-length 32 --mode thunder --num-iterations 10. This can result in increased latency in the decode stage.
Thanks @wujingyue for highlighting this.

Request: Investigate and try to fix these graph-breaks. Also, investigate and fix the split reasons.

cc: @IvanYashchuk for assignment.

cc @crcrpar

Metadata

Metadata

Assignees

Labels

benchmarkingthunderfxfor things that could be applicable to the dynamo+thunder frontend

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions