Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merge main -> google #5622

Merged
merged 13 commits into from
Apr 26, 2021
Merged

Conversation

ThomasRaoux
Copy link
Contributor

asaadaldien and others added 13 commits April 23, 2021 16:02
* Build simple_embedding_run in c API.

The simple_embedding_run target has a
Run function that executes
the bytecode of simle_embedding_test.mlir with c APIs.
The function supports all available HAL drivers, and is
selectable in the function argument, e.g.,

> simple_embedding_run dylib

The function also supports cross-compilation, if the cross-compiled
bytecode module is available (a RISC-V 64 vmfb example is in the
`data/` directory.

The function can be used to replace the core function of
simple_embeding_test in a following patch.
* eeef486 Fix include sort order
* 04b1fce Merge pull request iree-org#5595 from GMNGeoffrey:main-to-google
* a81fb9b Synchronize submodules with LLVM at llvm/llvm-project@91d3f73937b6
* 798f6ba Integrate LLVM at llvm/llvm-project@91d3f73937b6
* 91c5f7b Integrate LLVM at llvm/llvm-project@37e145812855
* 22c33a2 Merge pull request iree-org#5575 from GMNGeoffrey:main-to-google
* 49dc151 Move ModelBuilder out of IREE and into iree-llvm-sandbox
* Enable SourceMgrDiagnosticHandler in importers.

* This is a strict improvement over not having a diagnostic handler.
* Also introduces a stub pass for cleaning up TensorFlow locations but I have not yet found an algorithm I like so just leaving it as a stub for later.
* Progress on iree-org#5295
- Depthwise convolutions
- O(N^2) like matmul (or say matvec).
- RFFT
It was a mistake to migrate to benchmark executables. Eventually, we
would like to have both benchmarks. One is for end-to-end, and another
is for CodeGen kernels. The PR disables batch mode because:

- There is an issue in GPU benchmarks with batch mode. See iree-org#5248
- Make CPU and GPU benchmarks consistent. It is weird that one is in
  batch mode and another is not. Plus, this can make mako tags have same
  naming rule if we'd like to enable the executables benchmarks in the
  future. (Otherwise, "cpu" tag would be with batch mode, and "gpu"
  would not.)

note: this will introduce "regression" on Mako dashboard.
IREE recently introduced new driver `dylib-sync` which is used for
single threaded. After discussion, this should run with `taskset 80`,
and the multithreaded version should run with `taskset f0`. The PR
also makes taskset configurable.

This is a step towards iree-org#4752
The flag is off by default, so the runtime build will not have the
dependency of the compiler.

To use the flag, the users also need to specify the
IREE_HOST_BINARY_ROOT path or build the compiler.
…ee-org#5596)

- An alternative is to apply this after dispatching into workgroup tiles but its running throw multiple hops of issues related to
reshapes semantics subtensor inserts.
Split the TEST_P into two tests.
Supports both bazel test and ctest.
* On our GCC builds, __has_builtin appears to not be feature detecting this.
* Should fix iree-org#5617
@google-cla google-cla bot added the cla: yes label Apr 26, 2021
@copybara-service copybara-service bot merged commit a8c20a5 into iree-org:google Apr 26, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

6 participants