Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[pull] master from tensorflow:master #233

Merged
merged 10,000 commits into from
Apr 17, 2024
Merged

[pull] master from tensorflow:master #233

merged 10,000 commits into from
Apr 17, 2024

Conversation

pull[bot]
Copy link

@pull pull bot commented Sep 17, 2022

See Commits and Changes for more details.


Created by pull[bot]

Can you help keep this open source service alive? 💖 Please sponsor : )

@pull pull bot added the ⤵️ pull label Sep 17, 2022
Matthew Fahrbach and others added 29 commits April 11, 2024 14:39
PiperOrigin-RevId: 623939572
…ation layout shape.

PiperOrigin-RevId: 623947612
PiperOrigin-RevId: 623955329
The `GFile` documentation does not mention what exceptions it raises:
https://github.com/tensorflow/tensorflow/blob/38b17d708344a91234dad879794735e79f9af42a/tensorflow/python/platform/gfile.py#L37.

It would be safer to explicitly check for file existence and return the
appropriate result.

PiperOrigin-RevId: 623961044
Breaks jax tests

PiperOrigin-RevId: 623961307
Updates LLVM usage to match
[ffb5bea2be9f](llvm/llvm-project@ffb5bea2be9f)

PiperOrigin-RevId: 623964285
PiperOrigin-RevId: 623975807
This is required for `index_flat_map` where running the captured
function may return a `Status`.

PiperOrigin-RevId: 624026405
More changes will follow to demonstrate selective quantization.

PiperOrigin-RevId: 624026901
Bumps [idna](https://github.com/kjd/idna) from 3.6 to 3.7.
- [Release notes](https://github.com/kjd/idna/releases)
- [Changelog](https://github.com/kjd/idna/blob/master/HISTORY.rst)
- [Commits](kjd/idna@v3.6...v3.7)

---
updated-dependencies:
- dependency-name: idna
  dependency-type: direct:production
...

Signed-off-by: dependabot[bot] <[email protected]>
PiperOrigin-RevId: 624045436
Imported from GitHub PR openxla/xla#11444

1. cuDNN attention mask is not doing masking with -inf but multiply which is not correct. Hence disable patterns with mask.
2. Follow up PR to clean up the remaining mask related logic.
Copybara import of the project:

--
acf95b6cc7e1084026eaf87c0119ba3801ba8f8c by cjkkkk <[email protected]>:

disable mask

Merging this change closes #11444

PiperOrigin-RevId: 624068883
When looking for ptxas and other NVIDIA binary tools we go through a list of potential locations until we find a suitable version of ptxas.

Changing this list or its order has been deemed difficult in the past because there are many users out there that rely on the fact that this list happens to choose the right binary for them.

So this is making a careful change that is only affecting Google environments and makes no change to any OSS build of TF, JAX or XLA.

PiperOrigin-RevId: 624078399
PiperOrigin-RevId: 624088621
PiperOrigin-RevId: 624088926
PiperOrigin-RevId: 624091492
Imported from GitHub PR openxla/xla#11053

This PR supports more patterns from TF->XLA GELU Approximate. It was found that the variations were resulting from associativity of multiply instructions like in the expression a * b * c. An utility function is added in this PR that captures patterns from two consecutive multiplies.
Copybara import of the project:

--
74a287c7784138da3905d618d65a8301c821d7f6 by mdfaijul <[email protected]>:

Patterns from tf.nn.gelu(approximate=True)

--
51b4962709de6cee8ec92c9fc225b5fc3371c276 by mdfaijul <[email protected]>:

Fix typos

Merging this change closes #11053

PiperOrigin-RevId: 624093341
PiperOrigin-RevId: 624095085
PiperOrigin-RevId: 624095561
zoranjovanovic-ns and others added 29 commits April 16, 2024 10:07
…test

Imported from GitHub PR openxla/xla#11352

Modified the test case to pass on rocm and cuda.
There will be at least one more XLA related PR - to switch on Triton usage for ROCm, but only after PR for adding Triton build files for ROCm on openxla/triton.
Copybara import of the project:

--
dd80a6f44295642efd3ae6af6ffd3e2a3302d36e by Zoran Jovanovic <[email protected]>:

[ROCm] Triton in XLA for ROCm - changed ir_emitter_triton_test to pass on rocm and cuda

Merging this change closes #11352

PiperOrigin-RevId: 625361694
Proto messages for IFRT types may be used for various goals other than
transferring them over wire. This change reflects it on the proto comments.

Also set up a linter directive for `DType::Kind` and `DTypeProto::Kind` whose
mirroring is easy to miss.

PiperOrigin-RevId: 625365783
…utures to communicate events completion

PiperOrigin-RevId: 625375855
…e sharding on a tuple op, by using the same sharding for all get-tuple-elements.

PiperOrigin-RevId: 625384767
PiperOrigin-RevId: 625394661
PiperOrigin-RevId: 625409989
…ng two conditions are true.

1. converting `mhlo::ReturnOp` or `mlir::func::ReturnOp`
2. `num_return_values == 1`
3. `!return_tuple || !is_entry_function`

PiperOrigin-RevId: 625411731
PiperOrigin-RevId: 625419138
…>= 8.9.4

Imported from GitHub PR openxla/xla#11478

* Build XLA with cuDNN 8.6 will cause compilation error because flash attention graph will use `CudnnfMHAUid` which is only defined with cuDNN > 8.8.
* Add a guard for flash attention graph with at least cuDNN 8.9.4. So the current logic would be only compile FMHA with cuDNN > 8.8 and only compile flash attention with cuDNN > 8.9.4.
Copybara import of the project:

--
a1aa585f4e6ce42c7486336549447151cd5f7690 by cjkkkk <[email protected]>:

add guards for flash attention graph with at least 8.9.4

Merging this change closes #11478

PiperOrigin-RevId: 625425491
Imported from GitHub PR openxla/xla#11353

This implements async dynamic-slice and dynamic-update-slice for host memory offloading on GPU.

Since the emitter does not understand dynamic slicing instructions in async computation, we wrap them in a fusion node and mark them for execution on a different stream. This is all we need to execute the offloading of slices asynchronously.
Copybara import of the project:

--
905db6df8130300014728bf4eceb2ee2fc23d6e1 by Jaroslav Sevcik <[email protected]>:

Wrap async dynamic slicing into fusion

Merging this change closes #11353

PiperOrigin-RevId: 625428675
Otherwise, I'm getting this error:
```
Tensor("Const:0", shape=(4, 3), dtype=int64) must be from the same graph
 as Tensor("TensorSliceDataset:0", shape=(), dtype=variant)
```

PiperOrigin-RevId: 625429037
This workflow is no longer relevant.

PiperOrigin-RevId: 625430005
PiperOrigin-RevId: 625439915
…es are not divisible by the mesh dimensions.

Also get rid of a duplicate debug print statement

PiperOrigin-RevId: 625454917
Updates LLVM usage to match
[694c444b5bbb](llvm/llvm-project@694c444b5bbb)

PiperOrigin-RevId: 625464884
This will allow tf_runtime to depend on XLA in a followup change

PiperOrigin-RevId: 625474344
Without this check, it could crash if `index_map_fn` returns an
out-of-bound index due to:
https://github.com/tensorflow/tensorflow/blob/69a908420c5c5b90027f23905cd842c76ca3955c/tensorflow/core/framework/tensor.cc#L1104

PiperOrigin-RevId: 625476301
This commit adds folders for the ops `tfl.maximum` and `tfl.minimum` for the case where one of the arguments is `-FLT_MAX` and `FLT_MAX`, respectively.

PiperOrigin-RevId: 625494071
…k with async operations

PiperOrigin-RevId: 625518793
…tore the `IfrtRestoreTensorRegistry` in IfrtServingExecutable for looking up the dtype and shape.

PiperOrigin-RevId: 625527802
@pull pull bot merged commit 603690b into GesuBackups:master Apr 17, 2024
1 check passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.