Add handling of data series and scaling data#8
Conversation
3fedbb8 to
2275caf
Compare
Codecov Report✅ All modified and coverable lines are covered by tests. Additional details and impacted files@@ Coverage Diff @@
## main #8 +/- ##
=========================================
Coverage 100.00% 100.00%
=========================================
Files 7 8 +1
Lines 188 211 +23
=========================================
+ Hits 188 211 +23 ☔ View full report in Codecov by Sentry. 🚀 New features to boost your workflow:
|
33d48ac to
4d4e6d8
Compare
4d4e6d8 to
740480f
Compare
|
@manodeep This one should now be ready for review. |
740480f to
a8d4da9
Compare
|
@manodeep I've tried to address all your comments. Please have a look and let me know if you're happy with the changes. |
tests/test_parser.py
Outdated
There was a problem hiding this comment.
@micaeljtoliveira tmax on this line is smaller than tmin - is that supposed to happen?
There was a problem hiding this comment.
Related - do we have any sanity checks on the values that tmin <= tavg <= tmax etc?
There was a problem hiding this comment.
These numbers were originally added by Edward in the other repository and I don't think they were meant to be consistent, as they are only used for testing the parser.
I'm not sure we want to sanity check the numbers coming out of the logs, as that would be a lot of code and a lot of work. For now, let's just assume they are correct.
There was a problem hiding this comment.
That's fine to assume that the numbers are correct coming from the log files, but can we then swap the numbers in tmin[0] and tmax[0] in this mock profiling data?
There was a problem hiding this comment.
Well, if it bothers you that much, yes, we can swap them. But note that in the context of these unit tests, the actual numbers are quite meaningless.
There was a problem hiding this comment.
But then most of the other numbers are also problematic: tavg numbers for the "Total runtime" are clearly wrong and tavg for the other region in the 2 CPU case is greater than tmin and tmax...
There was a problem hiding this comment.
It probably shouldn't but it does actually bother me 😅
How about a compromise - can we add this an issue/feature request/test to make sure that the semantics of the parsed data are validated?
There was a problem hiding this comment.
I've decided to change the data entirely. Looking at the code being tested, there was no point in having so many metrics. Instead, it seemed more important that the dimensions of the dataset (npucs and number of regions) were different. As said above, the actual numbers are not important, but I made sure tmax values were greater than tmin 😉
manodeep
left a comment
There was a problem hiding this comment.
The linked comment is empty but so I can get another review in. There are two comments that I have made inline as a "reply"
cf73cae to
21d6a2f
Compare
manodeep
left a comment
There was a problem hiding this comment.
Thanks for the updates - everything looks good. One minor comment - would it be possible to swap the tmin[0] and tmax[0] numbers in the profiling data? That way a future maintainer/reader does not get confused by the semantics of the fake data.
(I am being extra careful since this feature is likely to be re-used by other teams and potentially external folks.)
tests/test_parser.py
Outdated
There was a problem hiding this comment.
That's fine to assume that the numbers are correct coming from the log files, but can we then swap the numbers in tmin[0] and tmax[0] in this mock profiling data?
21d6a2f to
3046f0a
Compare
3046f0a to
460c24c
Compare
|
@manodeep Thanks for the review! Feel free to open an issue about the data validation for further discussions. |
No description provided.