| View source on GitHub | 
This class specifies the configurations for an Estimator run.
tf.estimator.RunConfig(
    model_dir=None,
    tf_random_seed=None,
    save_summary_steps=100,
    save_checkpoints_steps=_USE_DEFAULT,
    save_checkpoints_secs=_USE_DEFAULT,
    session_config=None,
    keep_checkpoint_max=5,
    keep_checkpoint_every_n_hours=10000,
    log_step_count_steps=100,
    train_distribute=None,
    device_fn=None,
    protocol=None,
    eval_distribute=None,
    experimental_distribute=None,
    experimental_max_worker_delay_secs=None,
    session_creation_timeout_secs=7200,
    checkpoint_save_graph_def=True
)
   
| Args | |
|---|---|
| model_dir | directory where model parameters, graph, etc are saved. If PathLikeobject, the path will be resolved. IfNone, will use a default value set by the Estimator. | 
| tf_random_seed | Random seed for TensorFlow initializers. Setting this value allows consistency between reruns. | 
| save_summary_steps | Save summaries every this many steps. | 
| save_checkpoints_steps | Save checkpoints every this many steps. Can not be specified with save_checkpoints_secs. | 
| save_checkpoints_secs | Save checkpoints every this many seconds. Can not be specified with save_checkpoints_steps. Defaults to 600 seconds if bothsave_checkpoints_stepsandsave_checkpoints_secsare not set in constructor. If bothsave_checkpoints_stepsandsave_checkpoints_secsareNone, then checkpoints are disabled. | 
| session_config | a ConfigProto used to set session parameters, or None. | 
| keep_checkpoint_max | The maximum number of recent checkpoint files to keep. As new files are created, older files are deleted. If Noneor 0, all checkpoint files are kept. Defaults to 5 (that is, the 5 most recent checkpoint files are kept). If a saver is passed to the estimator, this argument will be ignored. | 
| keep_checkpoint_every_n_hours | Number of hours between each checkpoint to be saved. The default value of 10,000 hours effectively disables the feature. | 
| log_step_count_steps | The frequency, in number of global steps, that the global step and the loss will be logged during training. Also controls the frequency that the global steps / s will be logged (and written to summary) during training. | 
| train_distribute | An optional instance of tf.distribute.Strategy. If specified, then Estimator will distribute the user's model during training, according to the policy specified by that strategy. Settingexperimental_distribute.train_distributeis preferred. | 
| device_fn | A callable invoked for every Operationthat takes theOperationand returns the device string. IfNone, defaults to the device function returned bytf.train.replica_device_setterwith round-robin strategy. | 
| protocol | An optional argument which specifies the protocol used when starting server. Nonemeans default to grpc. | 
| eval_distribute | An optional instance of tf.distribute.Strategy. If specified, then Estimator will distribute the user's model during evaluation, according to the policy specified by that strategy. Settingexperimental_distribute.eval_distributeis preferred. | 
| experimental_distribute | An optional tf.contrib.distribute.DistributeConfigobject specifying DistributionStrategy-related configuration. Thetrain_distributeandeval_distributecan be passed as parameters toRunConfigor set inexperimental_distributebut not both. | 
| experimental_max_worker_delay_secs | An optional integer specifying the maximum time a worker should wait before starting. By default, workers are started at staggered times, with each worker being delayed by up to 60 seconds. This is intended to reduce the risk of divergence, which can occur when many workers simultaneously update the weights of a randomly initialized model. Users who warm-start their models and train them for short durations (a few minutes or less) should consider reducing this default to improve training times. | 
| session_creation_timeout_secs | Max time workers should wait for a session to become available (on initialization or when recovering a session) with MonitoredTrainingSession. Defaults to 7200 seconds, but users may want to set a lower value to detect problems with variable / session (re)-initialization more quickly. | 
| checkpoint_save_graph_def | Whether to save the GraphDef and MetaGraphDef to checkpoint_dir. The GraphDef is saved after the session is created asgraph.pbtxt. MetaGraphDefs are saved out for every checkpoint asmodel.ckpt-*.meta. | 
| Raises | |
|---|---|
| ValueError | If both save_checkpoints_stepsandsave_checkpoints_secsare set. | 
| Attributes | |
|---|---|
| checkpoint_save_graph_def | |
| cluster_spec | |
| device_fn | Returns the device_fn. If device_fn is not  | 
| eval_distribute | Optional tf.distribute.Strategyfor evaluation. | 
| evaluation_master | |
| experimental_max_worker_delay_secs | |
| global_id_in_cluster | The global id in the training cluster. All global ids in the training cluster are assigned from an increasing sequence of consecutive integers. The first id is 0. 
Note: Task id (the property field cluster = {'chief': ['host0:2222'],
           'ps': ['host1:2222', 'host2:2222'],
           'worker': ['host3:2222', 'host4:2222', 'host5:2222']}
Nodes with task type  Global id, i.e., this field, is tracking the index of the node among ALL nodes in the cluster. It is uniquely assigned. For example, for the cluster spec given above, the global ids are assigned as: task_type | task_id | global_id -------------------------------- chief | 0 | 0 worker | 0 | 1 worker | 1 | 2 worker | 2 | 3 ps | 0 | 4 ps | 1 | 5 | 
| is_chief | |
| keep_checkpoint_every_n_hours | |
| keep_checkpoint_max | |
| log_step_count_steps | |
| master | |
| model_dir | |
| num_ps_replicas | |
| num_worker_replicas | |
| protocol | Returns the optional protocol value. | 
| save_checkpoints_secs | |
| save_checkpoints_steps | |
| save_summary_steps | |
| service | Returns the platform defined (in TF_CONFIG) service dict. | 
| session_config | |
| session_creation_timeout_secs | |
| task_id | |
| task_type | |
| tf_random_seed | |
| train_distribute | Optional tf.distribute.Strategyfor training. | 
replace
replace(
    **kwargs
)
 Returns a new instance of RunConfig replacing specified properties.
Only the properties in the following list are allowed to be replaced:
model_dir,tf_random_seed,save_summary_steps,save_checkpoints_steps,save_checkpoints_secs,session_config,keep_checkpoint_max,keep_checkpoint_every_n_hours,log_step_count_steps,train_distribute,device_fn,protocol.eval_distribute,experimental_distribute,experimental_max_worker_delay_secs,In addition, either save_checkpoints_steps or save_checkpoints_secs can be set (should not be both).
| Args | |
|---|---|
| **kwargs | keyword named properties with new values. | 
| Raises | |
|---|---|
| ValueError | If any property name in kwargsdoes not exist or is not allowed to be replaced, or bothsave_checkpoints_stepsandsave_checkpoints_secsare set. | 
| Returns | |
|---|---|
| a new instance of RunConfig. | 
    © 2022 The TensorFlow Authors. All rights reserved.
Licensed under the Creative Commons Attribution License 4.0.
Code samples licensed under the Apache 2.0 License.
    https://www.tensorflow.org/versions/r2.9/api_docs/python/tf/estimator/RunConfig