Initializes a distributed TPU system for use with TensorFlow.
tf.compat.v1.tpu.initialize_system( embedding_config=None, job=None, compilation_failure_closes_chips=True )
Args | |
---|---|
embedding_config | If not None, a TPUEmbeddingConfiguration proto describing the desired configuration of the hardware embedding lookup tables. If embedding_config is None, no hardware embeddings can be used. |
job | The job (the XXX in TensorFlow device specification /job:XXX) that contains the TPU devices that will be initialized. If job=None it is assumed there is only one job in the TensorFlow flock, and an error will be returned if this assumption does not hold. |
compilation_failure_closes_chips | Set the configuration whether we want to close TPU chips when there is a compilation failure. |
Returns | |
---|---|
A serialized TopologyProto that describes the TPU system. Note: the topology must be evaluated using Session.run before it can be used. |
© 2020 The TensorFlow Authors. All rights reserved.
Licensed under the Creative Commons Attribution License 3.0.
Code samples licensed under the Apache 2.0 License.
https://www.tensorflow.org/versions/r2.4/api_docs/python/tf/compat/v1/tpu/initialize_system