diff --git a/ttnn/cpp/ttnn/operations/eltwise/binary_backward/binary_backward_pybind.hpp b/ttnn/cpp/ttnn/operations/eltwise/binary_backward/binary_backward_pybind.hpp index 22e09bdaac3..b99897762e7 100644 --- a/ttnn/cpp/ttnn/operations/eltwise/binary_backward/binary_backward_pybind.hpp +++ b/ttnn/cpp/ttnn/operations/eltwise/binary_backward/binary_backward_pybind.hpp @@ -22,7 +22,7 @@ namespace binary_backward { namespace detail { template -void bind_binary_backward_ops(py::module& module, const binary_backward_operation_t& operation, const std::string_view description, const std::string_view supported_dtype = "") { +void bind_binary_backward_ops(py::module& module, const binary_backward_operation_t& operation, const std::string_view description, const std::string_view supported_dtype = "BFLOAT16") { auto doc = fmt::format( R"doc( {2} @@ -39,10 +39,19 @@ void bind_binary_backward_ops(py::module& module, const binary_backward_operatio List of ttnn.Tensor: the output tensor. Note: - {3} + Supported dtypes, layouts, and ranks: - Note : bfloat8_b/bfloat4_b is only supported on TILE_LAYOUT + .. list-table:: + :header-rows: 1 + * - Dtypes + - Layouts + - Ranks + * - {3} + - TILE + - 2, 3, 4 + + bfloat8_b/bfloat4_b is only supported on TILE_LAYOUT Example: >>> grad_tensor = ttnn.to_device(ttnn.from_torch(torch.tensor((1, 2), dtype=torch.bfloat16)), device=device) @@ -79,7 +88,7 @@ void bind_binary_backward_ops(py::module& module, const binary_backward_operatio } template -void bind_binary_backward_int_default(py::module& module, const binary_backward_operation_t& operation, const std::string& parameter_name, const std::string& parameter_doc, int parameter_value, const std::string_view description, const std::string_view supported_dtype = "") { +void bind_binary_backward_int_default(py::module& module, const binary_backward_operation_t& operation, const std::string& parameter_name, const std::string& parameter_doc, int parameter_value, const std::string_view description, const std::string_view supported_dtype = "BFLOAT16") { auto doc = fmt::format( R"doc( {5} @@ -104,9 +113,19 @@ void bind_binary_backward_int_default(py::module& module, const binary_backward_ Note: - {6} + Supported dtypes, layouts, and ranks: + + .. list-table:: + :header-rows: 1 - Note : bfloat8_b/bfloat4_b is only supported on TILE_LAYOUT + * - Dtypes + - Layouts + - Ranks + * - {6} + - TILE + - 2, 3, 4 + + bfloat8_b/bfloat4_b is only supported on TILE_LAYOUT Example: @@ -157,7 +176,7 @@ void bind_binary_backward_int_default(py::module& module, const binary_backward_ } template -void bind_binary_backward_opt_float_default(py::module& module, const binary_backward_operation_t& operation, const std::string& parameter_name, const std::string& parameter_doc, float parameter_value, const std::string_view description, const std::string_view supported_dtype = "") { +void bind_binary_backward_opt_float_default(py::module& module, const binary_backward_operation_t& operation, const std::string& parameter_name, const std::string& parameter_doc, float parameter_value, const std::string_view description, const std::string_view supported_dtype = "BFLOAT16") { auto doc = fmt::format( R"doc( {5} @@ -183,9 +202,19 @@ void bind_binary_backward_opt_float_default(py::module& module, const binary_bac Note: - {6} + Supported dtypes, layouts, and ranks: + + .. list-table:: + :header-rows: 1 - Note : bfloat8_b/bfloat4_b is only supported on TILE_LAYOUT + * - Dtypes + - Layouts + - Ranks + * - {6} + - TILE + - 2, 3, 4 + + bfloat8_b/bfloat4_b is only supported on TILE_LAYOUT Example: @@ -246,7 +275,7 @@ void bind_binary_backward_float_string_default( const std::string& parameter_b_doc, string parameter_b_value, const std::string& description, - const std::string& supported_dtype) { + const std::string& supported_dtype = "BFLOAT16") { auto doc = fmt::format( R"doc( @@ -265,9 +294,19 @@ void bind_binary_backward_float_string_default( List of ttnn.Tensor: the output tensor. Note: - {8} + Supported dtypes, layouts, and ranks: + + .. list-table:: + :header-rows: 1 - Note : bfloat8_b/bfloat4_b is only supported on TILE_LAYOUT + * - Dtypes + - Layouts + - Ranks + * - {8} + - TILE + - 2, 3, 4 + + bfloat8_b/bfloat4_b is only supported on TILE_LAYOUT Example: @@ -323,7 +362,7 @@ void bind_binary_backward_float_string_default( } template -void bind_binary_backward_sub_alpha(py::module& module, const binary_backward_operation_t& operation, const std::string& parameter_name, const std::string& parameter_doc, float parameter_value, const std::string_view description, const std::string_view supported_dtype = "") { +void bind_binary_backward_sub_alpha(py::module& module, const binary_backward_operation_t& operation, const std::string& parameter_name, const std::string& parameter_doc, float parameter_value, const std::string_view description, const std::string_view supported_dtype = "BFLOAT16") { auto doc = fmt::format( R"doc( @@ -343,9 +382,19 @@ void bind_binary_backward_sub_alpha(py::module& module, const binary_backward_op queue_id (int, optional): command queue id. Defaults to `0`. Note: - {6} + Supported dtypes, layouts, and ranks: + + .. list-table:: + :header-rows: 1 - Note : bfloat8_b/bfloat4_b is only supported on TILE_LAYOUT + * - Dtypes + - Layouts + - Ranks + * - {6} + - TILE + - 2, 3, 4 + + bfloat8_b/bfloat4_b is only supported on TILE_LAYOUT Example: @@ -394,7 +443,7 @@ void bind_binary_backward_sub_alpha(py::module& module, const binary_backward_op } template -void bind_binary_backward_rsub(py::module& module, const binary_backward_operation_t& operation, const std::string_view description, const std::string_view supported_dtype = "") { +void bind_binary_backward_rsub(py::module& module, const binary_backward_operation_t& operation, const std::string_view description, const std::string_view supported_dtype = "BFLOAT16") { auto doc = fmt::format( R"doc( @@ -414,9 +463,19 @@ void bind_binary_backward_rsub(py::module& module, const binary_backward_operati queue_id (int, optional): command queue id. Defaults to `0`. Note: - {3} + Supported dtypes, layouts, and ranks: + + .. list-table:: + :header-rows: 1 - Note : bfloat8_b/bfloat4_b is only supported on TILE_LAYOUT + * - Dtypes + - Layouts + - Ranks + * - {3} + - TILE + - 2, 3, 4 + + bfloat8_b/bfloat4_b is only supported on TILE_LAYOUT Example: @@ -460,7 +519,7 @@ void bind_binary_backward_rsub(py::module& module, const binary_backward_operati } template -void bind_binary_bw_mul(py::module& module, const binary_backward_operation_t& operation, const std::string_view description, const std::string_view supported_dtype = "") { +void bind_binary_bw_mul(py::module& module, const binary_backward_operation_t& operation, const std::string_view description, const std::string_view supported_dtype = "BFLOAT16") { auto doc = fmt::format( R"doc( {2} @@ -481,9 +540,19 @@ void bind_binary_bw_mul(py::module& module, const binary_backward_operation_t& o List of ttnn.Tensor: the output tensor. Note: - {3} + Supported dtypes, layouts, and ranks: + + .. list-table:: + :header-rows: 1 - Note : bfloat8_b/bfloat4_b is only supported on TILE_LAYOUT + * - Dtypes + - Layouts + - Ranks + * - {3} + - TILE + - 2, 3, 4 + + bfloat8_b/bfloat4_b is only supported on TILE_LAYOUT Example: >>> grad_tensor = ttnn.to_device(ttnn.from_torch(torch.tensor((1, 2), dtype=torch.bfloat16)), device) @@ -561,7 +630,7 @@ void bind_binary_bw_mul(py::module& module, const binary_backward_operation_t& o template -void bind_binary_bw(py::module& module, const binary_backward_operation_t& operation, const std::string_view description, const std::string_view supported_dtype = "") { +void bind_binary_bw(py::module& module, const binary_backward_operation_t& operation, const std::string_view description, const std::string_view supported_dtype = "BFLOAT16") { auto doc = fmt::format( R"doc( @@ -581,9 +650,19 @@ void bind_binary_bw(py::module& module, const binary_backward_operation_t& opera queue_id (int, optional): command queue id. Defaults to `0`. Note: - {3} + Supported dtypes, layouts, and ranks: + + .. list-table:: + :header-rows: 1 - Note : bfloat8_b/bfloat4_b is only supported on TILE_LAYOUT + * - Dtypes + - Layouts + - Ranks + * - {3} + - TILE + - 2, 3, 4 + + bfloat8_b/bfloat4_b is only supported on TILE_LAYOUT Example: >>> grad_tensor = ttnn.to_device(ttnn.from_torch(torch.tensor((1, 2), dtype=torch.bfloat16)), device) @@ -662,7 +741,7 @@ void bind_binary_bw(py::module& module, const binary_backward_operation_t& opera } template -void bind_binary_bw_div(py::module& module, const binary_backward_operation_t& operation, const std::string_view description, const std::string_view supported_dtype = "") { +void bind_binary_bw_div(py::module& module, const binary_backward_operation_t& operation, const std::string_view description, const std::string_view supported_dtype = "BFLOAT16") { auto doc = fmt::format( R"doc( @@ -687,9 +766,19 @@ void bind_binary_bw_div(py::module& module, const binary_backward_operation_t& o Supports broadcasting. Note: - {3} + Supported dtypes, layouts, and ranks: + + .. list-table:: + :header-rows: 1 - Note : bfloat8_b/bfloat4_b is only supported on TILE_LAYOUT + * - Dtypes + - Layouts + - Ranks + * - {3} + - TILE + - 2, 3, 4 + + bfloat8_b/bfloat4_b is only supported on TILE_LAYOUT Example: >>> grad_tensor = ttnn.to_device(ttnn.from_torch(torch.tensor((1, 2), dtype=torch.bfloat16)), device) @@ -770,7 +859,7 @@ void bind_binary_bw_div(py::module& module, const binary_backward_operation_t& o } template -void bind_binary_backward_overload(py::module& module, const binary_backward_operation_t& operation, const std::string& description, const std::string& supported_dtype) { +void bind_binary_backward_overload(py::module& module, const binary_backward_operation_t& operation, const std::string& description, const std::string& supported_dtype = "BFLOAT16", const std::string& note = "") { auto doc = fmt::format( R"doc( @@ -788,7 +877,19 @@ void bind_binary_backward_overload(py::module& module, const binary_backward_ope List of ttnn.Tensor: the output tensor. Note: - {3} + Supported dtypes, layouts, and ranks: + + .. list-table:: + :header-rows: 1 + + * - Dtypes + - Layouts + - Ranks + * - {3} + - TILE + - 2, 3, 4 + + {4} Example: @@ -800,7 +901,7 @@ void bind_binary_backward_overload(py::module& module, const binary_backward_ope operation.base_name(), operation.python_fully_qualified_name(), description, - supported_dtype); + supported_dtype, note); bind_registered_operation( module, @@ -838,7 +939,7 @@ void bind_binary_backward_overload(py::module& module, const binary_backward_ope } template -void bind_binary_backward_assign(py::module& module, const binary_backward_operation_t& operation, const std::string_view description, const std::string_view supported_dtype = "") { +void bind_binary_backward_assign(py::module& module, const binary_backward_operation_t& operation, const std::string_view description, const std::string_view supported_dtype = "BFLOAT16") { auto doc = fmt::format( R"doc( @@ -858,7 +959,17 @@ void bind_binary_backward_assign(py::module& module, const binary_backward_opera round_mode (str, optional): Round mode for the operation. Defaults to `None`. Note: - {3} + Supported dtypes, layouts, and ranks: + + .. list-table:: + :header-rows: 1 + + * - Dtypes + - Layouts + - Ranks + * - {3} + - TILE + - 2, 3, 4 Example: >>> grad_tensor = ttnn.to_device(ttnn.from_torch(torch.tensor((1, 2), dtype=torch.bfloat16)), device) @@ -924,288 +1035,126 @@ void py_module(py::module& module) { module, ttnn::mul_bw, R"doc(Performs backward operations for multiply on :attr:`input_tensor_a`, :attr:`input_tensor_b`, with given :attr:`grad_tensor`.)doc", - R"doc(Supported dtypes, layouts, and ranks: - - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16, BFLOAT8_B | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ - - )doc"); + R"doc(BFLOAT16, BFLOAT8_B)doc"); detail::bind_binary_bw( module, ttnn::add_bw, R"doc(Performs backward operations for add of :attr:`input_tensor_a` and :attr:`input_tensor_b` or :attr:`scalar` with given :attr:`grad_tensor`.)doc", - R"doc(Supported dtypes, layouts, and ranks: - - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16, BFLOAT8_B | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ - - )doc"); + R"doc(BFLOAT16, BFLOAT8_B)doc"); detail::bind_binary_bw( module, ttnn::sub_bw, R"doc(Performs backward operations for subtract of :attr:`input_tensor_a` and :attr:`input_tensor_b` or :attr:`scalar` with given :attr:`grad_tensor`.)doc", - R"doc(Supported dtypes, layouts, and ranks: - - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16, BFLOAT8_B | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ - - )doc"); + R"doc(BFLOAT16, BFLOAT8_B)doc"); detail::bind_binary_bw_div( module, ttnn::div_bw, R"doc(Performs backward operations for divide on :attr:`input_tensor`, :attr:`alpha` or :attr:`input_tensor_a`, :attr:`input_tensor_b`, :attr:`round_mode`, with given :attr:`grad_tensor`.)doc", - R"doc(Supported dtypes, layouts, and ranks: - - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16, BFLOAT8_B | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ - - )doc"); + R"doc(BFLOAT16, BFLOAT8_B)doc"); detail::bind_binary_backward_overload( module, ttnn::remainder_bw, R"doc(Performs backward operations for remainder of :attr:`input_tensor_a`, :attr:`scalar` or :attr:`input_tensor_b` with given :attr:`grad_tensor`.)doc", - R"doc(Supported dtypes, layouts, and ranks: - - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16 | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ - - supported only in WHB0. - - )doc"); + R"doc(BFLOAT16)doc", R"doc(Supported only in WHB0.)doc"); detail::bind_binary_backward_overload( module, ttnn::fmod_bw, R"doc(Performs backward operations for fmod of :attr:`input_tensor_a`, :attr:`scalar` or :attr:`input_tensor_b` with given :attr:`grad_tensor`.)doc", - - R"doc(Supported dtypes, layouts, and ranks: - - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16 | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ - - )doc"); + R"doc(BFLOAT16)doc"); detail::bind_binary_backward_assign( module, ttnn::assign_bw, - R"doc(Performs backward operations for assign of :attr:`input_tensor_a`, :attr:`input_tensor_b` with given :attr:`grad_tensor`.)doc", - - R"doc(Supported dtypes, layouts, and ranks: - - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16, BFLOAT8_B | ROW_MAJOR, TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ - - )doc"); + R"doc(Performs backward operations for assign of :attr:`input_tensor_a`, :attr:`input_tensor_b` with given :attr:`grad_tensor`.)doc", R"doc(BFLOAT16, BFLOAT8_B)doc"); detail::bind_binary_backward_ops( module, ttnn::atan2_bw, R"doc(Performs backward operations for atan2 of :attr:`input_tensor_a` and :attr:`input_tensor_b` with given :attr:`grad_tensor`.)doc", - R"doc(Supported dtypes, layouts, and ranks: - - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16, BFLOAT8_B | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ - - )doc"); + R"doc(BFLOAT16, BFLOAT8_B)doc"); detail::bind_binary_backward_sub_alpha( module, ttnn::subalpha_bw, "alpha", "Alpha value", 1.0f, R"doc(Performs backward operations for subalpha of :attr:`input_tensor_a` and :attr:`input_tensor_b` with given :attr:`grad_tensor`.)doc", - R"doc(Supported dtypes, layouts, and ranks: - - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16, BFLOAT8_B | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ - - )doc"); + R"doc(BFLOAT16, BFLOAT8_B)doc"); detail::bind_binary_backward_opt_float_default( module, ttnn::addalpha_bw, "alpha", "Alpha value", 1.0f, R"doc(Performs backward operations for addalpha on :attr:`input_tensor_b` , :attr:`input_tensor_a` and :attr:`alpha` with given :attr:`grad_tensor`.)doc", - R"doc(Supported dtypes, layouts, and ranks: - - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16, BFLOAT8_B | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ - - )doc"); + R"doc(BFLOAT16, BFLOAT8_B)doc"); detail::bind_binary_backward_ops( module, ttnn::xlogy_bw, R"doc(Performs backward operations for xlogy of :attr:`input_tensor_a` and :attr:`input_tensor_b` with given :attr:`grad_tensor`.)doc", - R"doc(Supported dtypes, layouts, and ranks: + R"doc(BFLOAT16, BFLOAT8_B)doc"); - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16 | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ - - )doc"); detail::bind_binary_backward_ops( module, ttnn::hypot_bw, R"doc(Performs backward operations for hypot of :attr:`input_tensor_a` and :attr:`input_tensor_b` with given :attr:`grad_tensor`.)doc", - R"doc(Supported dtypes, layouts, and ranks: - - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16, BFLOAT8_B | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ + R"doc(BFLOAT16, BFLOAT8_B)doc"); - )doc"); detail::bind_binary_backward_ops( module, ttnn::ldexp_bw, R"doc(Performs backward operations for ldexp of :attr:`input_tensor_a` and :attr:`input_tensor_b` with given :attr:`grad_tensor`.)doc", - R"doc(Supported dtypes, layouts, and ranks: + R"doc(BFLOAT16)doc"); - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16, BFLOAT8_B | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ - - )doc"); detail::bind_binary_backward_ops( module, ttnn::logaddexp_bw, R"doc(Performs backward operations for logaddexp of :attr:`input_tensor_a` and :attr:`input_tensor_b` with given :attr:`grad_tensor`.)doc", - R"doc(Supported dtypes, layouts, and ranks: - - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16, BFLOAT8_B | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ + R"doc(BFLOAT16)doc"); - )doc"); detail::bind_binary_backward_ops( module, ttnn::logaddexp2_bw, R"doc(Performs backward operations for logaddexp2 of :attr:`input_tensor_a` and :attr:`input_tensor_b` with given :attr:`grad_tensor`.)doc", - R"doc(Supported dtypes, layouts, and ranks: - - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16, BFLOAT8_B | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ - - )doc"); + R"doc(BFLOAT16)doc"); detail::bind_binary_backward_ops( module, ttnn::squared_difference_bw, R"doc(Performs backward operations for squared_difference of :attr:`input_tensor_a` and :attr:`input_tensor_b` with given :attr:`grad_tensor`.)doc", - R"doc(Supported dtypes, layouts, and ranks: + R"doc(BFLOAT16, BFLOAT8_B)doc"); - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16, BFLOAT8_B | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ - - )doc"); detail::bind_binary_backward_int_default( module, ttnn::concat_bw, "dim", "Dimension to concatenate", 0, R"doc(Performs backward operations for concat on :attr:`input_tensor_a` and :attr:`input_tensor_b` with given :attr:`grad_tensor`.)doc", - R"doc(Supported dtypes, layouts, and ranks: - - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16 | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ - - )doc"); + R"doc(BFLOAT16)doc"); detail::bind_binary_backward_rsub( module, ttnn::rsub_bw, R"doc(Performs backward operations for subraction of :attr:`input_tensor_a` from :attr:`input_tensor_b` with given :attr:`grad_tensor` (reversed order of subtraction operator).)doc", - R"doc(Supported dtypes, layouts, and ranks: - - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16, BFLOAT8_B | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ - - )doc"); + R"doc(BFLOAT16, BFLOAT8_B)doc"); detail::bind_binary_backward_ops( module, ttnn::min_bw, - R"doc(Performs backward operations for minimum of :attr:`input_tensor_a` and :attr:`input_tensor_b` with given :attr:`grad_tensor`.)doc", - R"doc(Supported dtypes, layouts, and ranks: - - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16 | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ - - )doc"); + R"doc(Performs backward operations for minimum of :attr:`input_tensor_a` and :attr:`input_tensor_b` with given :attr:`grad_tensor`.)doc"); detail::bind_binary_backward_ops( module, ttnn::max_bw, - R"doc(Performs backward operations for maximum of :attr:`input_tensor_a` and :attr:`input_tensor_b` with given :attr:`grad_tensor`.)doc", - R"doc(Supported dtypes, layouts, and ranks: - - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16 | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ - - )doc"); + R"doc(Performs backward operations for maximum of :attr:`input_tensor_a` and :attr:`input_tensor_b` with given :attr:`grad_tensor`.)doc"); detail::bind_binary_backward_float_string_default( module, @@ -1217,15 +1166,7 @@ void py_module(py::module& module) { "none", R"doc(Performs backward operations for bias_gelu on :attr:`input_tensor_a` and :attr:`input_tensor_b` or :attr:`input_tensor` and :attr:`bias`, with given :attr:`grad_tensor` using given :attr:`approximate` mode. :attr:`approximate` mode can be 'none', 'tanh'.)doc", - R"doc(Supported dtypes, layouts, and ranks: - - +----------------------------+---------------------------------+-------------------+ - | Dtypes | Layouts | Ranks | - +----------------------------+---------------------------------+-------------------+ - | BFLOAT16, BFLOAT8_B | TILE | 2, 3, 4 | - +----------------------------+---------------------------------+-------------------+ - - )doc"); + R"doc(BFLOAT16, BFLOAT8_B)doc"); } } // namespace binary_backward