You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Pull Request resolved: #14732
Collecting fixes for various models/ops in this diff/PR.
They have all been squashed into this single change to make it easier to cherry pick.
# Fixes
## Wav2Letter
Type: Output correctness failure
This is caused by a bug in swiftshader, and not reproducible on any other platform. Specifically, the issue is in the softmax shader; the exact cause of the issue is unknown, but it is related to using shared memory within shaders. The workaround for this issue is to use separate shared memory arrays for the shared max and shared sum.
## ConvNeXT
Type: Exception during runtime
This is caused by an incompatible memory layout being used for mean2d. More technically, the packed dimension of the tensor cannot be one of the dims being reduced. The current operator registry system did not have a way to select valid tensor representations based on the actual arguments of an op.
To fix, we have to introduce a mechanism for ops to specify valid representations once a node's arguments are known. Once the model is exported with supported memory layout, the model test passes.
## Inception_V3/ViT
Type: Exception during runtime
The root cause of this was an interaction betwen the fuse batch norm pass and how `vulkan_preprocess.py` was applying passes. Essentially, the fuse batch norm pass creates a new param node for the fused weight, which uses the original name of the tensor which may contain capital letters. However after re-tracing the graph, the node's name was being lowercased. `vulkan_preprocess` was using _copy_module to update the exported program's graph module in place, which was not updating the ep's graph signature with the new lowercase name after retracing.
The solution was to migrate vulkan_preprocess.py to use the _transform() API instead of using _copy_module.
There was also a small bug in Pool.cpp where `bool` was used to pass a UBO field that is received as an `int`.
## DenseNet 161 (w/ dynamic shapes)
Type: Output Mismatch
Cause: the native_batch_norm op doesn't support dynamic shapes. However, the backend test runner doesn't set the correct compile option to filter ops without dynamic shape support.
Since batch norm is easy to implement, fix by implementing resize for batch norm.
ghstack-source-id: 313794474
Differential Revision: [D83703496](https://our.internmc.facebook.com/intern/diff/D83703496/)
0 commit comments