Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BUG] Final prediction ESR doesn't match what's computed during the trainer's validation step #489

Open
2-dor opened this issue Oct 10, 2024 · 1 comment
Labels
bug Something isn't working priority:low Low-priority issues

Comments

@2-dor
Copy link

2-dor commented Oct 10, 2024

Hi Steve,

I've been re-training some models with trainer version v0.10.0 and encountered something I know happened a while back.

In the "checkpoints" folder, my lowest ESR seems to have been 0.01003; two other checkpoints with ESR 0.01004.

However, when the trainer is stopped (I hit Ctrl + C in the CLI), it reports an ESR of 0.01006.

image

@2-dor 2-dor added bug Something isn't working priority:low Low-priority issues unread This issue is new and hasn't been seen by the maintainers yet labels Oct 10, 2024
@sdatkinson
Copy link
Owner

Yeah, I've seen this.

There's a little discrepancy between how the validation step is computed and how that final prediction is run. In validation, the input is processed as-is and the output is left-cropped to match the model's receptive field; in the final prediction, the input is pre-padded with zeroes so that it's always the exact same output that's being reported on no matter what model is used (and what its receptive field happens to be)

I'll leave this Issue open because I might be able to do a bit of refactoring elsewhere that makes this all agree (I'm thinking of refactoring some of the data processing code), but I'll want to hang on a bit and do that instead of jumping in on this--using the tools that are in the code right now, the result might come out kind of ugly.

To take a step back, this discrepancy won't cause any real problems--if the difference between two models is their ability to predict a bit of silence, then that's probably not telling you which model is actually better for real 😉.

@sdatkinson sdatkinson removed the unread This issue is new and hasn't been seen by the maintainers yet label Oct 11, 2024
@sdatkinson sdatkinson changed the title [BUG] Lowest checkpoint ESR is not exported by the trainer in v0.10.0 [BUG] Final prediction ESR doesn't match what's computed during the trainer's validation step Oct 11, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working priority:low Low-priority issues
Projects
None yet
Development

No branches or pull requests

2 participants