Skip to content
Merged
Show file tree
Hide file tree
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Next Next commit
Upgrade Bazel version and its config
  • Loading branch information
karllessard committed Mar 8, 2022
commit 62b3af13ac58824f8363edf868823d75e7d3e302
3 changes: 2 additions & 1 deletion tensorflow-core/tensorflow-core-api/.bazelrc
Original file line number Diff line number Diff line change
@@ -1,2 +1,3 @@
build --remote_cache=https://storage.googleapis.com/tensorflow-sigs-jvm
build --remote_upload_local_results=false
build --remote_upload_local_results=false
build --incompatible_restrict_string_escapes=false
2 changes: 1 addition & 1 deletion tensorflow-core/tensorflow-core-api/.bazelversion
Original file line number Diff line number Diff line change
@@ -1 +1 @@
3.7.2
4.2.1
56 changes: 42 additions & 14 deletions tensorflow-core/tensorflow-core-api/tensorflow.bazelrc
Original file line number Diff line number Diff line change
Expand Up @@ -115,6 +115,7 @@ build -c opt
# Make Bazel print out all options from rc files.
build --announce_rc

# TODO(mihaimaruseac): Document this option or remove if no longer needed
build --define=grpc_no_ares=true

# See https://github.com/bazelbuild/bazel/issues/7362 for information on what
Expand All @@ -131,19 +132,26 @@ build --define=grpc_no_ares=true
# archives in -whole_archive -no_whole_archive.
build --noincompatible_remove_legacy_whole_archive

# TODO(mihaimaruseac): Document this option or remove if no longer needed
build --enable_platform_specific_config

# Enable XLA support by default.
build --define=with_xla_support=true

# TODO(mihaimaruseac): Document this option or remove if no longer needed
build --config=short_logs

# TODO(mihaimaruseac): Document this option or remove if no longer needed
build --config=v2

# Disable AWS/HDFS support by default
build --define=no_aws_support=true
build --define=no_hdfs_support=true

# TF now has `cc_shared_library` targets, so it needs the experimental flag
# TODO(rostam): Remove when `cc_shared_library` is enabled by default
build --experimental_cc_shared_library

# Default options should come above this line.

# Allow builds using libc++ as a linker library
Expand Down Expand Up @@ -219,9 +227,8 @@ build:mkl_threadpool -c opt

# Config setting to build oneDNN with Compute Library for the Arm Architecture (ACL).
# This build is for the inference regime only.
build:mkl_aarch64 --define=build_with_mkl_aarch64=true --define=enable_mkl=true
build:mkl_aarch64 --define=build_with_mkl_aarch64=true
build:mkl_aarch64 --define=tensorflow_mkldnn_contraction_kernel=0
build:mkl_aarch64 --define=build_with_mkl_opensource=true
build:mkl_aarch64 --define=build_with_openmp=true
build:mkl_aarch64 -c opt

Expand Down Expand Up @@ -473,7 +480,7 @@ build:rbe_linux_cuda_nvcc_py38 --config=rbe_linux_cuda11.2_nvcc_py3.8
build:rbe_linux_cuda_nvcc_py39 --config=rbe_linux_cuda11.2_nvcc_py3.9

# Deprecated configs that people might still use.
build:rbe_linux_cuda_nvcc --config=rbe_linux_cuda_nvcc_py36
build:rbe_linux_cuda_nvcc --config=rbe_linux_cuda_nvcc_py39
build:rbe_gpu_linux --config=rbe_linux_cuda_nvcc

build:rbe_linux_cuda_clang_base --config=rbe_linux_cuda_base
Expand Down Expand Up @@ -586,6 +593,12 @@ build:release_cpu_linux --config=avx_linux
build:release_cpu_linux --crosstool_top="@ubuntu18.04-gcc7_manylinux2010-cuda11.2-cudnn8.1-tensorrt7.2_config_cuda//crosstool:toolchain"
test:release_cpu_linux --test_env=LD_LIBRARY_PATH

# manylinux2014 config for cpu
build:release_cpu_linux_manylinux2014 --config=release_base
build:release_cpu_linux_manylinux2014 --config=avx_linux
build:release_cpu_linux_manylinux2014 --crosstool_top="@ubuntu18.04-gcc8_manylinux2014-cuda11.2-cudnn8.1-tensorrt7.2_config_cuda//crosstool:toolchain"
test:release_cpu_linux_manylinux2014 --test_env=LD_LIBRARY_PATH

build:release_cpu_macos --config=release_base
build:release_cpu_macos --config=avx_linux

Expand All @@ -608,6 +621,12 @@ build:release_gpu_linux_11_4 --action_env=TF_CUDA_VERSION="11.4"
build:release_gpu_linux_11_4 --action_env=TF_CUDNN_VERSION="8.2"
build:release_gpu_linux_11_4 --crosstool_top=@ubuntu18.04-gcc7_manylinux2010-cuda11.4-cudnn8.2-tensorrt7.2_config_cuda//crosstool:toolchain

# manylinux2014 config for gpu
build:release_gpu_linux_manylinux2014 --config=release_gpu_linux
build:release_gpu_linux_manylinux2014 --action_env=GCC_HOST_COMPILER_PATH="/dt8/usr/bin/gcc"
build:release_gpu_linux_manylinux2014 --crosstool_top=@ubuntu18.04-gcc8_manylinux2014-cuda11.2-cudnn8.1-tensorrt7.2_config_cuda//crosstool:toolchain


build:release_cpu_windows --config=release_base
build:release_cpu_windows --config=avx_win
build:release_cpu_windows --define=no_tensorflow_py_deps=true
Expand Down Expand Up @@ -650,17 +669,26 @@ build:ubsan --linkopt -fsanitize=undefined
build:ubsan --linkopt -lubsan

# Disable TFRT integration for now unless --config=tfrt is specified.
build --deleted_packages=tensorflow/compiler/mlir/tfrt,tensorflow/compiler/mlir/tfrt/benchmarks,tensorflow/compiler/mlir/tfrt/jit/python_binding,tensorflow/compiler/mlir/tfrt/jit/transforms,tensorflow/compiler/mlir/tfrt/python_tests,tensorflow/compiler/mlir/tfrt/tests,tensorflow/compiler/mlir/tfrt/tests/saved_model,tensorflow/compiler/mlir/tfrt/transforms/lhlo_gpu_to_tfrt_gpu,tensorflow/core/runtime_fallback,tensorflow/core/runtime_fallback/conversion,tensorflow/core/runtime_fallback/kernel,tensorflow/core/runtime_fallback/opdefs,tensorflow/core/runtime_fallback/runtime,tensorflow/core/runtime_fallback/util,tensorflow/core/tfrt/common,tensorflow/core/tfrt/eager,tensorflow/core/tfrt/eager/backends/cpu,tensorflow/core/tfrt/eager/backends/gpu,tensorflow/core/tfrt/eager/core_runtime,tensorflow/core/tfrt/eager/cpp_tests/core_runtime,tensorflow/core/tfrt/fallback,tensorflow/core/tfrt/gpu,tensorflow/core/tfrt/run_handler_thread_pool,tensorflow/core/tfrt/runtime,tensorflow/core/tfrt/saved_model,tensorflow/core/tfrt/saved_model/tests,tensorflow/core/tfrt/tpu,tensorflow/core/tfrt/utils
build --deleted_packages=tensorflow/compiler/mlir/tfrt,tensorflow/compiler/mlir/tfrt/benchmarks,tensorflow/compiler/mlir/tfrt/jit/python_binding,tensorflow/compiler/mlir/tfrt/jit/transforms,tensorflow/compiler/mlir/tfrt/python_tests,tensorflow/compiler/mlir/tfrt/tests,tensorflow/compiler/mlir/tfrt/tests/analysis,tensorflow/compiler/mlir/tfrt/tests/jit,tensorflow/compiler/mlir/tfrt/tests/lhlo_to_tfrt,tensorflow/compiler/mlir/tfrt/tests/tf_to_corert,tensorflow/compiler/mlir/tfrt/tests/tf_to_tfrt_data,tensorflow/compiler/mlir/tfrt/tests/saved_model,tensorflow/compiler/mlir/tfrt/transforms/lhlo_gpu_to_tfrt_gpu,tensorflow/core/runtime_fallback,tensorflow/core/runtime_fallback/conversion,tensorflow/core/runtime_fallback/kernel,tensorflow/core/runtime_fallback/opdefs,tensorflow/core/runtime_fallback/runtime,tensorflow/core/runtime_fallback/util,tensorflow/core/tfrt/common,tensorflow/core/tfrt/eager,tensorflow/core/tfrt/eager/backends/cpu,tensorflow/core/tfrt/eager/backends/gpu,tensorflow/core/tfrt/eager/core_runtime,tensorflow/core/tfrt/eager/cpp_tests/core_runtime,tensorflow/core/tfrt/fallback,tensorflow/core/tfrt/gpu,tensorflow/core/tfrt/run_handler_thread_pool,tensorflow/core/tfrt/runtime,tensorflow/core/tfrt/saved_model,tensorflow/core/tfrt/saved_model/tests,tensorflow/core/tfrt/tpu,tensorflow/core/tfrt/utils
build:tfrt --deleted_packages=

# Experimental configuration for testing XLA GPU lowering to TFRT BEF thunks.
# bazel test --config=experimental_enable_bef_thunk \
# Experimental configuration for testing XLA GPU lowering to TFRT.
#
# To run using BEF thunks, combine with
# --//tensorflow/compiler/xla/service/gpu:enable_bef_thunk. For example,
# bazel test --config=experimental_tfrt_gpu \
# --//tensorflow/compiler/xla/service/gpu:enable_bef_thunk \
# //tensorflow/compiler/xla/service/gpu/tests:mlir_gemm_test
build:experimental_enable_bef_thunk --config=tfrt
build:experimental_enable_bef_thunk --//tensorflow/compiler/xla/service/gpu:enable_bef_thunk
build:experimental_enable_bef_thunk --@tf_runtime//:enable_gpu
build:experimental_enable_bef_thunk --@rules_cuda//cuda:enable_cuda
build:experimental_enable_bef_thunk --@rules_cuda//cuda:cuda_runtime=//tensorflow/compiler/xla/service/gpu:cuda_runtime_for_xlir
build:experimental_enable_bef_thunk --nocheck_visibility
build:experimental_enable_bef_thunk --incompatible_strict_action_env
build:experimental_enable_bef_thunk --config=monolithic
#
# To run using BEF executable, combine with
# --//tensorflow/compiler/xla/service/gpu:enable_bef_executable. For example,
# bazel test --config=experimental_tfrt_gpu \
# --//tensorflow/compiler/xla/service/gpu:enable_bef_executable \
# //tensorflow/compiler/xla/service/gpu/tests:mnist
build:experimental_tfrt_gpu --config=tfrt
build:experimental_tfrt_gpu --@tf_runtime//:enable_gpu
build:experimental_tfrt_gpu --@rules_cuda//cuda:enable_cuda
build:experimental_tfrt_gpu --@rules_cuda//cuda:cuda_runtime=//tensorflow/compiler/xla/service/gpu:cuda_runtime_for_xlir
build:experimental_tfrt_gpu --nocheck_visibility
build:experimental_tfrt_gpu --incompatible_strict_action_env
build:experimental_tfrt_gpu --config=monolithic