Skip to content

Conversation

@ivarflakstad
Copy link
Member

What does this PR do?

Adds expected results to certain tests that had slightly different results on AMD devices.

Comment on lines 2700 to 2705
cuda = [" Folks, I spend a lot of time right over there, night after night after night, actually. Carefully selecting for you the day's noosiest, most aerodynamic headlines, stress testing, and those topical anti-lock breaks and power steering, painstakingly stitching, leather seating so soft, it would make JD power and her associates blush to create the luxury sedan that is my nightly monologue. But sometimes, you sometimes, folks. I lurched a consciousness in the back of an abandoned school bus and slap myself awake."]
cuda1 = [" Folks, I spend a lot of time right over there, night after night after night, actually. Carefully selecting for you the day's noosiest, most aerodynamic headlines, stress testing, and those topical anti-lock breaks and power steering, painstakingly stitching, leather seating so soft, it would make JD power and her associates blush to create the luxury sedan that is my nightly monologue. But sometimes, you sometimes, folks. I lurched a consciousness in the back of an abandoned school bus and slap myself a wig."]
rocm = [" Folks, I spend a lot of time right over there, night after night after night, actually. Carefully selecting for you the day's noosiest, most aerodynamic headlines, stress testing, and those topical anti-lock breaks and power steering, painstakingly stitching, leather seating, so soft, it would make JD power and her associates blush to create the luxury sedan that is my nightly monologue. But sometimes, you sometimes, folks, I lurched a consciousness in the back of an abandoned school bus and slap myself awake."]
# fmt: on
expected_output = Expectations({("cuda", None): cuda, ("rocm", None): rocm}).get_expectation()
expected_output1 = Expectations({("cuda", None): cuda1, ("rocm", None): rocm}).get_expectation()
Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This one is interesting. Notice how cuda and cuda1 end with different words.
This did not happen when I ran it on the MI325. If the second result is supposed to end with "wig" then we have to look into it.
@eustlb thoughts? ☺️

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It seems like cuda output uses this generation config while cuda1 uses a different generation config so they are expected to have different results.

So I think with rocm you should generate another output with the same config as cuda1, to have something like rocm1.

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think they could generate different results because the generation config are not the same, yet it's possible that the two results be the same. Even with cuda and cuda1 the two generations are very close, so it's not a stretch to think that on rocm the output tokens are the same for the two generations.

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Right, but the purpose of the test is to verify that the generation config has the expected effect.
So we either have a rocm bug or need to adjust the config.

@ivarflakstad
Copy link
Member Author

run-slow: whisper

@github-actions
Copy link
Contributor

This comment contains run-slow, running the specified jobs:

models: ['models/whisper']
quantizations: [] ...

@HuggingFaceDocBuilderDev

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.

Copy link
Collaborator

@remi-or remi-or left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM just add precise versions for ROCm please!

cuda1 = [" Folks, I spend a lot of time right over there, night after night after night, actually. Carefully selecting for you the day's noosiest, most aerodynamic headlines, stress testing, and those topical anti-lock breaks and power steering, painstakingly stitching, leather seating so soft, it would make JD power and her associates blush to create the luxury sedan that is my nightly monologue. But sometimes, you sometimes, folks. I lurched a consciousness in the back of an abandoned school bus and slap myself a wig."]
rocm = [" Folks, I spend a lot of time right over there, night after night after night, actually. Carefully selecting for you the day's noosiest, most aerodynamic headlines, stress testing, and those topical anti-lock breaks and power steering, painstakingly stitching, leather seating, so soft, it would make JD power and her associates blush to create the luxury sedan that is my nightly monologue. But sometimes, you sometimes, folks, I lurched a consciousness in the back of an abandoned school bus and slap myself awake."]
# fmt: on
expected_output = Expectations({("cuda", None): cuda, ("rocm", None): rocm}).get_expectation()
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Specify version for rocm

Copy link
Contributor

@ahadnagy ahadnagy left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks Ivar! i think it's fine not specifying the exact platform versions until we need to differentiate.

@ivarflakstad
Copy link
Member Author

run-slow: whisper

@ivarflakstad ivarflakstad requested a review from remi-or August 27, 2025 11:15
@github-actions
Copy link
Contributor

This comment contains run-slow, running the specified jobs:

models: ['models/whisper']
quantizations: [] ...

Copy link
Collaborator

@remi-or remi-or left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

approved w/ one very minor nit

@github-actions
Copy link
Contributor

[For maintainers] Suggested jobs to run (before merge)

run-slow: whisper

@ivarflakstad ivarflakstad enabled auto-merge (squash) August 27, 2025 16:11
@ivarflakstad ivarflakstad merged commit 3c343c6 into main Aug 27, 2025
19 checks passed
@ivarflakstad ivarflakstad deleted the amd-whisper-expectations branch August 27, 2025 16:19
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

6 participants