Skip to content

opencl: broadcast for soft_max #14510

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 2 commits into from
Jul 3, 2025
Merged

Conversation

lhez
Copy link
Contributor

@lhez lhez commented Jul 3, 2025

Following ec68e84, this PR adds broadcast for soft_max. This fixes the failures in test-backend-ops.

Make sure to read the contributing guidelines before submitting a PR

@github-actions github-actions bot added ggml changes relating to the ggml tensor library for machine learning OpenCL Issues specific to the OpenCL backend labels Jul 3, 2025
@lhez lhez marked this pull request as ready for review July 3, 2025 06:38
@CISC
Copy link
Collaborator

CISC commented Jul 3, 2025

BTW, I see none of the other backends bother with passing along ne11, probably doesn't make any difference if you do or don't?

@CISC CISC merged commit bee2842 into ggml-org:master Jul 3, 2025
48 checks passed
@lhez
Copy link
Contributor Author

lhez commented Jul 3, 2025

Yeah, ne11 was not used; also removed other unused kernel args.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ggml changes relating to the ggml tensor library for machine learning OpenCL Issues specific to the OpenCL backend
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants