View source on GitHub |
Solves tridiagonal systems of equations.
tf.linalg.tridiagonal_solve( diagonals, rhs, diagonals_format='compact', transpose_rhs=False, conjugate_rhs=False, name=None, partial_pivoting=True )
The input can be supplied in various formats: matrix
, sequence
and compact
, specified by the diagonals_format
arg.
In matrix
format, diagonals
must be a tensor of shape [..., M, M]
, with two inner-most dimensions representing the square tridiagonal matrices. Elements outside of the three diagonals will be ignored.
In sequence
format, diagonals
are supplied as a tuple or list of three tensors of shapes [..., N]
, [..., M]
, [..., N]
representing superdiagonals, diagonals, and subdiagonals, respectively. N
can be either M-1
or M
; in the latter case, the last element of superdiagonal and the first element of subdiagonal will be ignored.
In compact
format the three diagonals are brought together into one tensor of shape [..., 3, M]
, with last two dimensions containing superdiagonals, diagonals, and subdiagonals, in order. Similarly to sequence
format, elements diagonals[..., 0, M-1]
and diagonals[..., 2, 0]
are ignored.
The compact
format is recommended as the one with best performance. In case you need to cast a tensor into a compact format manually, use tf.gather_nd
. An example for a tensor of shape [m, m]:
rhs = tf.constant([...]) matrix = tf.constant([[...]]) m = matrix.shape[0] dummy_idx = [0, 0] # An arbitrary element to use as a dummy indices = [[[i, i + 1] for i in range(m - 1)] + [dummy_idx], # Superdiagonal [[i, i] for i in range(m)], # Diagonal [dummy_idx] + [[i + 1, i] for i in range(m - 1)]] # Subdiagonal diagonals=tf.gather_nd(matrix, indices) x = tf.linalg.tridiagonal_solve(diagonals, rhs)
Regardless of the diagonals_format
, rhs
is a tensor of shape [..., M]
or [..., M, K]
. The latter allows to simultaneously solve K systems with the same left-hand sides and K different right-hand sides. If transpose_rhs
is set to True
the expected shape is [..., M]
or [..., K, M]
.
The batch dimensions, denoted as ...
, must be the same in diagonals
and rhs
.
The output is a tensor of the same shape as rhs
: either [..., M]
or [..., M, K]
.
The op isn't guaranteed to raise an error if the input matrix is not invertible. tf.debugging.check_numerics
can be applied to the output to detect invertibility problems.
Note: with large batch sizes, the computation on the GPU may be slow, if eitherpartial_pivoting=True
or there are multiple right-hand sides (K > 1
). If this issue arises, consider if it's possible to disable pivoting and haveK = 1
, or, alternatively, consider using CPU.
On CPU, solution is computed via Gaussian elimination with or without partial pivoting, depending on partial_pivoting
parameter. On GPU, Nvidia's cuSPARSE library is used: https://docs.nvidia.com/cuda/cusparse/index.html#gtsv
Args | |
---|---|
diagonals | A Tensor or tuple of Tensor s describing left-hand sides. The shape depends of diagonals_format , see description above. Must be float32 , float64 , complex64 , or complex128 . |
rhs | A Tensor of shape [..., M] or [..., M, K] and with the same dtype as diagonals . Note that if the shape of rhs and/or diags isn't known statically, rhs will be treated as a matrix rather than a vector. |
diagonals_format | one of matrix , sequence , or compact . Default is compact . |
transpose_rhs | If True , rhs is transposed before solving (has no effect if the shape of rhs is [..., M]). |
conjugate_rhs | If True , rhs is conjugated before solving. |
name | A name to give this Op (optional). |
partial_pivoting | whether to perform partial pivoting. True by default. Partial pivoting makes the procedure more stable, but slower. Partial pivoting is unnecessary in some cases, including diagonally dominant and symmetric positive definite matrices (see e.g. theorem 9.12 in [1]). |
Returns | |
---|---|
A Tensor of shape [..., M] or [..., M, K] containing the solutions. |
Raises | |
---|---|
ValueError | An unsupported type is provided as input, or when the input tensors have incorrect shapes. |
UnimplementedError | Whenever partial_pivoting is true and the backend is XLA. |
[1] Nicholas J. Higham (2002). Accuracy and Stability of Numerical Algorithms: Second Edition. SIAM. p. 175. ISBN 978-0-89871-802-7.
© 2020 The TensorFlow Authors. All rights reserved.
Licensed under the Creative Commons Attribution License 3.0.
Code samples licensed under the Apache 2.0 License.
https://www.tensorflow.org/versions/r2.4/api_docs/python/tf/linalg/tridiagonal_solve