Mojo function
batched_matmul
batched_matmul[*, transpose_a: Bool = False, transpose_b: Bool = False, elementwise_epilogue_fn: Optional[elementwise_epilogue_type] = None, saturated_vnni: Bool = False, target: StringSlice[StaticConstantOrigin] = StringSlice("cpu")](c_buf: TileTensor[c_buf.dtype, c_buf.LayoutType, c_buf.origin, linear_idx_type=c_buf.linear_idx_type, element_size=c_buf.element_size], a_buf: TileTensor[a_buf.dtype, a_buf.LayoutType, a_buf.origin, linear_idx_type=a_buf.linear_idx_type, element_size=a_buf.element_size], b_buf: TileTensor[b_buf.dtype, b_buf.LayoutType, b_buf.origin, linear_idx_type=b_buf.linear_idx_type, element_size=b_buf.element_size], *, context: DeviceContextPtr = DeviceContextPtr())
TileTensor primary implementation of batched_matmul.
Was this page helpful?
Thank you! We'll create more content like this.
Thank you for helping us improve!