Skip to content

Conversation

reeselevine
Copy link
Collaborator

In my previous PR (#16357), I missed cherry-picking a commit and ended up missing the code to actually enable soft_max (oops). This PR actually adds it to supports_op and encode_node, as well as fixes a potential bug in rms_norm where it was using the wrong offset into the tensor.

@github-actions github-actions bot added the ggml changes relating to the ggml tensor library for machine learning label Oct 3, 2025
@CISC
Copy link
Collaborator

CISC commented Oct 3, 2025

Mind you, there appears to be an issue with SOFT_MAX:
https://github.com/ggml-org/llama.cpp/actions/runs/18212653903/job/51856062582?pr=16400#step:7:31662

@reeselevine
Copy link
Collaborator Author

fixed the soft max shader and added a temporary fix that blocks on each WebGPU queue submission until I figure out a more efficient way to wait for futures to complete.

@reeselevine reeselevine merged commit 3526657 into ggml-org:master Oct 5, 2025
66 of 68 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ggml changes relating to the ggml tensor library for machine learning
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants