You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Enabling the use of extensions (such as cooperative matrix) and special codepaths for various runtime-detectable capabilities would be a good way of evaluating the flexibility of the compiler. Most of this work is predicated on support for such features in MLIR's SPIR-V dialect (if we want to go that route) and the design of the integration, however we could do some proof of concept work for benchmarking independently.
The text was updated successfully, but these errors were encountered:
Also fixes some other odds and ends:
* Switches to larger runners.
* Uses the compiler's native support for memory outputs (vs direct use
of memfd). This removes special casing that would be needed when
building for old glibc versions (the compiler already does the right
thing in these cases).
* Adds a :plugins target to build everything.
* Fixes the CUDA SDK env var to include the "_DIR" suffix.
* Installs the needed parts of the CUDA SDK
* Forks the build_linux_packages.sh that everyone uses and builds binary
plugins. Will extend this later to build Python wheels that
auto-configure.
Enabling the use of extensions (such as cooperative matrix) and special codepaths for various runtime-detectable capabilities would be a good way of evaluating the flexibility of the compiler. Most of this work is predicated on support for such features in MLIR's SPIR-V dialect (if we want to go that route) and the design of the integration, however we could do some proof of concept work for benchmarking independently.
The text was updated successfully, but these errors were encountered: