Conversation
|
@claude review this PR |
|
Claude finished @hanaol's task in 1m 26s —— View job Code Review:
|
Fix the silent-discard bug when --out is omitted (docstring promises stdout as default) and move status/peak_gb/epoch_s to module level since they don't close over loop variables. Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
forklady42
left a comment
There was a problem hiding this comment.
Addressed the bug and one of the minor comments. Okay to ignore the other minor comment and the nit for this script.
Summary
This adds a GPU comparison benchmark script that runs the same per-sample training experiment on both an A100 and an H200, recording peak GPU memory, forward/backward times, and OOM status for 10 large-grid Materials Project samples under f32 and bf16-mixed precision.
The 10 task IDs are Materials Project entries with relatively large charge-density grids, spanning 3.4 M – 46.7 M voxels across a variety of shapes and aspect ratios.
Benchmark results
Model:
resunet.ResUNet3D,n_channels=32,n_residual_blocks=1,kernel_size=5,depth=2,batch_size=1, single GPU, 3 epochs per experiment.f32
bf16-mixed
Summary
Key findings
Files
scripts/benchmark_gpus.py-- Reads two JSON result files (one per GPU) produced by