[ENHANCEMENT] Dynamic binning + distance-aware label smoothing #251
Killer3048
started this conversation in
Ideas
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Dynamic binning + distance-aware label smoothing
Why this matters
[-15..+15]for quantization after mean scaling. This can cause severe overflow (if the real data extends beyond +15) or underflow (if data is confined way below ±15) in zero-shot scenarios.chronos.py, the classMeanScaleUniformBinsspecifically relies onlow_limitandhigh_limitto buildself.centersandself.boundaries. While that works fine for data roughly within[-15..+15], new domains can push well outside that range.torch.bucketizewithout distinguishing “close” vs. “far” bins.Proposed enhancement
Dynamic range determination
MeanScaleUniformBinsto compute (at inference or per training batch) a local min/max (or percentiles, like p10/p90) for the scaled data.self.centersandself.boundariesso that we cover[actual_min..actual_max], clamping if needed (e.g.,[-50..50]).Distance-aware label smoothing
bis the correct bin, allocate ~80% probability tob, ~10% tob-1, and ~10% tob+1(splitting among neighbors).ChronosModel.forward, if we have direct control of the loss.Estimated impact on accuracy & inference
Zero-shot improvements:
Few-shot improvements:
Inference overhead:
Why it improves zero-/few-shot
Beta Was this translation helpful? Give feedback.
All reactions