Skip to content
  •  
  •  
  •  
Prev Previous commit
Next Next commit
Op classifications for TF 2.18.
  • Loading branch information
Craigacp committed May 16, 2025
commit b533b21c820fc0c0bfa9252c305460cec8fb2b31
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
op {
graph_op_name: "CheckPinned"
endpoint {
name: "CheckPinned"
}
}
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
op {
graph_op_name: "ComputeDedupDataSize"
visibility: VISIBLE
visibility: SKIP
endpoint {
name: "tpu.ComputeDedupDataSize"
}
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
op {
graph_op_name: "ComputeDedupDataSizeV2"
endpoint {
name: "tpu.ComputeDedupDataSize"
}
}
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
op {
visibility: VISIBLE
visibility: SKIP
graph_op_name: "ComputeDedupDataTupleMask"
endpoint {
name: "tpu.ComputeDedupDataTupleMask"
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
op {
graph_op_name: "ComputeDedupDataTupleMaskV2"
endpoint {
name: "tpu.ComputeDedupDataTupleMask"
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,7 @@
op {
graph_op_name: "ConvertToListOfSparseCoreCooTensors"
visibility: VISIBLE
endpoint {
name: "sparse.ConvertToListOfSparseCoreCooTensors"
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,7 @@
op {
graph_op_name: "ConvertToSparseCoreCsrWrappedCooTensor"
visibility: VISIBLE
endpoint {
name: "sparse.ConvertToSparseCoreCsrWrappedCooTensor"
}
}
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
op {
visibility: VISIBLE
visibility: SKIP
graph_op_name: "FinalizeTPUEmbedding"
endpoint {
name: "tpu.FinalizeTPUEmbedding"
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
op {
graph_op_name: "FinalizeTPUEmbeddingV2"
endpoint {
name: "tpu.FinalizeTPUEmbedding"
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,7 @@
op {
graph_op_name: "GetStatsFromListOfSparseCoreCooTensors"
visibility: VISIBLE
endpoint {
name: "sparse.GetStatsFromListOfSparseCoreCooTensors"
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,7 @@
op {
graph_op_name: "GetTpuTaskId"
visibility: VISIBLE
endpoint {
name: "tpu.GetTpuTaskId"
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
op {
graph_op_name: "GlobalShuffleDataset"
endpoint {
name: "data.GlobalShuffleDataset"
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,7 @@
op {
graph_op_name: "IndexFlatMapDataset"
visibility: VISIBLE
endpoint {
name: "data.IndexFlatMapDataset"
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
op {
graph_op_name: "IteratorGetModelProto"
endpoint {
name: "data.IteratorGetModelProto"
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
op {
graph_op_name: "SortListOfSparseCoreCooTensors"
endpoint {
name: "sparse.SortListOfSparseCoreCooTensors"
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
op {
graph_op_name: "UpdateTaskIdAndGlobalCoreArray"
endpoint {
name: "tpu.UpdateTaskIdAndGlobalCoreArray"
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
op {
graph_op_name: "WeightedFlatMapDataset"
endpoint {
name: "data.WeightedFlatMapDataset"
}
}
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
op {
visibility: VISIBLE
visibility: SKIP
graph_op_name: "XlaRecvTPUEmbeddingActivations"
endpoint {
name: "xla.XlaRecvTPUEmbeddingActivations"
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
op {
graph_op_name: "XlaRecvTPUEmbeddingActivationsV2"
endpoint {
name: "xla.XlaRecvTPUEmbeddingActivations"
}
}
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
op {
visibility: VISIBLE
visibility: SKIP
graph_op_name: "XlaRecvTPUEmbeddingDeduplicationData"
endpoint {
name: "xla.XlaRecvTPUEmbeddingDeduplicationData"
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
op {
graph_op_name: "XlaRecvTPUEmbeddingDeduplicationDataV2"
endpoint {
name: "xla.XlaRecvTPUEmbeddingDeduplicationData"
}
}
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
op {
visibility: VISIBLE
visibility: SKIP
graph_op_name: "XlaSendTPUEmbeddingGradients"
endpoint {
name: "xla.XlaSendTPUEmbeddingGradients"
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
op {
graph_op_name: "XlaSendTPUEmbeddingGradientsV2"
endpoint {
name: "xla.XlaSendTPUEmbeddingGradients"
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
op {
graph_op_name: "XlaSparseDenseMatmulGradWithAdagradAndStaticBufferSize"
endpoint {
name: "xla.XlaSparseDenseMatmulGradWithAdagradAndStaticBufferSize"
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
op {
graph_op_name: "XlaSparseDenseMatmulGradWithAdagradMomentumAndStaticBufferSize"
endpoint {
name: "xla.XlaSparseDenseMatmulGradWithAdagradMomentumAndStaticBufferSize"
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
op {
graph_op_name: "XlaSparseDenseMatmulGradWithAdamAndStaticBufferSize"
endpoint {
name: "xla.XlaSparseDenseMatmulGradWithAdamAndStaticBufferSize"
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
op {
graph_op_name: "XlaSparseDenseMatmulGradWithCsrInput"
endpoint {
name: "xla.XlaSparseDenseMatmulGradWithCsrInput"
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
op {
graph_op_name: "XlaSparseDenseMatmulGradWithFtrlAndStaticBufferSize"
endpoint {
name: "xla.XlaSparseDenseMatmulGradWithFtrlAndStaticBufferSize"
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
op {
graph_op_name: "XlaSparseDenseMatmulGradWithSgdAndStaticBufferSize"
endpoint {
name: "xla.XlaSparseDenseMatmulGradWithSgdAndStaticBufferSize"
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
op {
graph_op_name: "XlaSparseDenseMatmulWithStaticBufferSize"
endpoint {
name: "xla.XlaSparseDenseMatmulWithStaticBufferSize"
}
}
Original file line number Diff line number Diff line change
Expand Up @@ -61,7 +61,6 @@ public final class BitwiseOps {
* tf.assert_equal(tf.cast(res, tf.float32), exp) # TRUE
* </pre>
*
* @param <T> data type for {@code z} output
* @param x The x value
* @param y The y value
* @param <T> data type for {@code BitwiseAnd} output and operands
Expand Down Expand Up @@ -91,7 +90,6 @@ public <T extends TNumber> BitwiseAnd<T> bitwiseAnd(Operand<T> x, Operand<T> y)
* tf.assert_equal(tf.cast(res, tf.float32), exp) # TRUE
* </pre>
*
* @param <T> data type for {@code z} output
* @param x The x value
* @param y The y value
* @param <T> data type for {@code BitwiseOr} output and operands
Expand Down Expand Up @@ -121,7 +119,6 @@ public <T extends TNumber> BitwiseOr<T> bitwiseOr(Operand<T> x, Operand<T> y) {
* tf.assert_equal(tf.cast(res, tf.float32), exp) # TRUE
* </pre>
*
* @param <T> data type for {@code z} output
* @param x The x value
* @param y The y value
* @param <T> data type for {@code BitwiseXor} output and operands
Expand Down Expand Up @@ -172,7 +169,6 @@ public <T extends TNumber> BitwiseXor<T> bitwiseXor(Operand<T> x, Operand<T> y)
* tf.assert_equal(tf.cast(inverted, tf.float32), tf.cast(expected, tf.float32))
* </pre>
*
* @param <T> data type for {@code y} output
* @param x The x value
* @param <T> data type for {@code Invert} output and operands
* @return a new instance of Invert
Expand Down Expand Up @@ -212,7 +208,6 @@ public <T extends TNumber> Invert<T> invert(Operand<T> x) {
* # &lt;tf.Tensor: shape=(4,), dtype=int8, numpy=array([ -2, 64, 101, 32], dtype=int8)&gt;
* </pre>
*
* @param <T> data type for {@code z} output
* @param x The x value
* @param y The y value
* @param <T> data type for {@code LeftShift} output and operands
Expand Down Expand Up @@ -255,7 +250,6 @@ public <T extends TNumber> LeftShift<T> leftShift(Operand<T> x, Operand<T> y) {
* # &lt;tf.Tensor: shape=(4,), dtype=int8, numpy=array([ -2, 64, 101, 32], dtype=int8)&gt;
* </pre>
*
* @param <T> data type for {@code z} output
* @param x The x value
* @param y The y value
* @param <T> data type for {@code RightShift} output and operands
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -49,7 +49,6 @@ public final class CollectiveOps {
/**
* Mutually exchanges multiple tensors of identical type and shape.
*
* @param <T> data type for {@code data} output
* @param input The input value
* @param communicator The communicator value
* @param groupAssignment The groupAssignment value
Expand Down Expand Up @@ -79,7 +78,6 @@ public CollectiveAssignGroup collectiveAssignGroup(Operand<TInt32> groupAssignme
/**
* Receives a tensor value broadcast from another device.
*
* @param <U> data type for {@code data} output
* @param groupSize The groupSize value
* @param groupKey The groupKey value
* @param instanceKey The instanceKey value
Expand All @@ -98,7 +96,6 @@ public <U extends TType> CollectiveBcastRecv<U> collectiveBcastRecv(Operand<TInt
/**
* Broadcasts a tensor value to one or more other devices.
*
* @param <T> data type for {@code data} output
* @param input The input value
* @param groupSize The groupSize value
* @param groupKey The groupKey value
Expand All @@ -119,7 +116,6 @@ public <T extends TType> CollectiveBcastSend<T> collectiveBcastSend(Operand<T> i
* collective ops. In this case, keys that are unique at runtime
* (e.g. {@code instance_key}) should be used to distinguish collective groups.
*
* @param <T> data type for {@code data} output
* @param input The input value
* @param groupSize The groupSize value
* @param groupKey The groupKey value
Expand Down Expand Up @@ -157,7 +153,6 @@ public CollectiveInitializeCommunicator collectiveInitializeCommunicator(Operand
* source_target_pairs={@code [[0,1],[1,2],[2,3],[3,0]]} gets the outputs:
* {@code [D, A, B, C]}.
*
* @param <T> data type for {@code output} output
* @param input The local input to be permuted. Currently only supports float and
* bfloat16.
* @param sourceTargetPairs A tensor with shape [num_pairs, 2].
Expand All @@ -172,7 +167,6 @@ public <T extends TType> CollectivePermute<T> collectivePermute(Operand<T> input
/**
* Mutually reduces multiple tensors of identical type and shape.
*
* @param <T> data type for {@code data} output
* @param input The input value
* @param communicator The communicator value
* @param groupAssignment The groupAssignment value
Expand All @@ -193,7 +187,6 @@ public <T extends TNumber> CollectiveReduce<T> collectiveReduce(Operand<T> input
* collective ops. In this case, keys that are unique at runtime
* (e.g. {@code instance_key}) should be used to distinguish collective groups.
*
* @param <T> data type for {@code data} output
* @param input The input value
* @param groupSize The groupSize value
* @param groupKey The groupKey value
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -59,6 +59,7 @@
import org.tensorflow.op.data.GroupByReducerDataset;
import org.tensorflow.op.data.GroupByWindowDataset;
import org.tensorflow.op.data.IgnoreErrorsDataset;
import org.tensorflow.op.data.IndexFlatMapDataset;
import org.tensorflow.op.data.InitializeTableFromDataset;
import org.tensorflow.op.data.InterleaveDataset;
import org.tensorflow.op.data.Iterator;
Expand Down Expand Up @@ -819,6 +820,28 @@ public IgnoreErrorsDataset ignoreErrorsDataset(Operand<? extends TType> inputDat
return IgnoreErrorsDataset.create(scope, inputDataset, outputTypes, outputShapes, options);
}

/**
* The IndexFlatMapDataset operation
*
* @param inputDataset The inputDataset value
* @param mapFuncOtherArgs The mapFuncOtherArgs value
* @param indexMapFuncOtherArgs The indexMapFuncOtherArgs value
* @param outputCardinality The outputCardinality value
* @param mapFunc The value of the mapFunc attribute
* @param indexMapFunc The value of the indexMapFunc attribute
* @param outputTypes The value of the outputTypes attribute
* @param outputShapes The value of the outputShapes attribute
* @param options carries optional attribute values
* @return a new instance of IndexFlatMapDataset
*/
public IndexFlatMapDataset indexFlatMapDataset(Operand<? extends TType> inputDataset,
Iterable<Operand<?>> mapFuncOtherArgs, Iterable<Operand<?>> indexMapFuncOtherArgs,
Operand<TInt64> outputCardinality, ConcreteFunction mapFunc, ConcreteFunction indexMapFunc,
List<Class<? extends TType>> outputTypes, List<Shape> outputShapes,
IndexFlatMapDataset.Options... options) {
return IndexFlatMapDataset.create(scope, inputDataset, mapFuncOtherArgs, indexMapFuncOtherArgs, outputCardinality, mapFunc, indexMapFunc, outputTypes, outputShapes, options);
}

/**
* The InitializeTableFromDataset operation
*
Expand Down Expand Up @@ -987,7 +1010,6 @@ public LatencyStatsDataset latencyStatsDataset(Operand<? extends TType> inputDat
/**
* Computes rectified linear gradients for a LeakyRelu operation.
*
* @param <T> data type for {@code backprops} output
* @param gradients The backpropagated gradients to the corresponding LeakyRelu operation.
* @param features The features passed as input to the corresponding LeakyRelu operation,
* OR the outputs of that operation (both work equivalently).
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -43,7 +43,6 @@ public final class DebuggingOps {
* tensor. Unlike CheckNumerics (V1), CheckNumericsV2 distinguishes -Inf and +Inf
* in the errors it throws.
*
* @param <T> data type for {@code output} output
* @param tensor The tensor value
* @param message Prefix of the error message.
* @param <T> data type for {@code CheckNumericsV2} output and operands
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,6 @@ public final class DistributeOps {
* num_devices: The number of devices participating in this reduction.
* shared_name: Identifier that shared between ops of the same reduction.
*
* @param <T> data type for {@code data} output
* @param input The input value
* @param reduction The value of the reduction attribute
* @param numDevices The value of the numDevices attribute
Expand All @@ -74,7 +73,6 @@ public <T extends TNumber> NcclAllReduce<T> ncclAllReduce(Operand<T> input, Stri
* output: The same as input.
* shape: The shape of the input tensor.
*
* @param <T> data type for {@code output} output
* @param input The input value
* @param shape The value of the shape attribute
* @param <T> data type for {@code NcclBroadcast} output and operands
Expand All @@ -93,7 +91,6 @@ public <T extends TNumber> NcclBroadcast<T> ncclBroadcast(Operand<T> input, Shap
* data: the value of the reduction across all {@code num_devices} devices.
* reduction: the reduction operation to perform.
*
* @param <T> data type for {@code data} output
* @param input The input value
* @param reduction The value of the reduction attribute
* @param <T> data type for {@code NcclReduce} output and operands
Expand Down
Loading