tf.test.Benchmark
Abstract class that provides helpers for TensorFlow benchmarks.
tf.test.Benchmark()
Methods
evaluate
View source
evaluate(
tensors
)
Evaluates tensors and returns numpy values.
Args |
tensors | A Tensor or a nested list/tuple of Tensors. |
Returns |
tensors numpy values. |
is_abstract
View source
@classmethod
is_abstract()
report_benchmark
View source
report_benchmark(
iters=None, cpu_time=None, wall_time=None, throughput=None, extras=None,
name=None, metrics=None
)
Report a benchmark.
Args |
iters | (optional) How many iterations were run |
cpu_time | (optional) Median or mean cpu time in seconds. |
wall_time | (optional) Median or mean wall time in seconds. |
throughput | (optional) Throughput (in MB/s) |
extras | (optional) Dict mapping string keys to additional benchmark info. Values may be either floats or values that are convertible to strings. |
name | (optional) Override the BenchmarkEntry name with name . Otherwise it is inferred from the top-level method name. |
metrics | (optional) A list of dict, where each dict has the keys below name (required), string, metric name value (required), double, metric value min_value (optional), double, minimum acceptable metric value max_value (optional), double, maximum acceptable metric value |
run_op_benchmark
View source
run_op_benchmark(
sess, op_or_tensor, feed_dict=None, burn_iters=2, min_iters=10,
store_trace=False, store_memory_usage=True, name=None, extras=None, mbs=0
)
Run an op or tensor in the given session. Report the results.
Args |
sess | Session object to use for timing. |
op_or_tensor | Operation or Tensor to benchmark. |
feed_dict | A dict of values to feed for each op iteration (see the feed_dict parameter of Session.run ). |
burn_iters | Number of burn-in iterations to run. |
min_iters | Minimum number of iterations to use for timing. |
store_trace | Boolean, whether to run an extra untimed iteration and store the trace of iteration in returned extras. The trace will be stored as a string in Google Chrome trace format in the extras field "full_trace_chrome_format". Note that trace will not be stored in test_log_pb2.TestResults proto. |
store_memory_usage | Boolean, whether to run an extra untimed iteration, calculate memory usage, and store that in extras fields. |
name | (optional) Override the BenchmarkEntry name with name . Otherwise it is inferred from the top-level method name. |
extras | (optional) Dict mapping string keys to additional benchmark info. Values may be either floats or values that are convertible to strings. |
mbs | (optional) The number of megabytes moved by this op, used to calculate the ops throughput. |
Returns |
A dict containing the key-value pairs that were passed to report_benchmark . If store_trace option is used, then full_chrome_trace_format will be included in return dictionary even though it is not passed to report_benchmark with extras . |