tf.contrib.learn.monitors.BaseMonitor.step_begin()

tf.contrib.learn.monitors.BaseMonitor.step_begin(step) Callback before training step begins. You may use this callback to request evaluation of additional tensors in the graph. Args: step: int, the current value of the global step. Returns: List of Tensor objects or string tensor names to be run. Raises: ValueError: if we've already begun a step, or step < 0, or step > max_steps.

tf.contrib.bayesflow.stochastic_tensor.LaplaceWithSoftplusScaleTensor

class tf.contrib.bayesflow.stochastic_tensor.LaplaceWithSoftplusScaleTensor LaplaceWithSoftplusScaleTensor is a StochasticTensor backed by the distribution LaplaceWithSoftplusScale.

tf.contrib.distributions.Binomial.sample()

tf.contrib.distributions.Binomial.sample(sample_shape=(), seed=None, name='sample') Generate samples of the specified shape. Note that a call to sample() without arguments will generate a single sample. Args: sample_shape: 0D or 1D int32 Tensor. Shape of the generated samples. seed: Python integer seed for RNG name: name to give to the op. Returns: samples: a Tensor with prepended dimensions sample_shape.

tf.contrib.learn.monitors.StepCounter.epoch_end()

tf.contrib.learn.monitors.StepCounter.epoch_end(epoch) End epoch. Args: epoch: int, the epoch number. Raises: ValueError: if we've not begun an epoch, or epoch number does not match.

tf.contrib.learn.train()

tf.contrib.learn.train(graph, output_dir, train_op, loss_op, global_step_tensor=None, init_op=None, init_feed_dict=None, init_fn=None, log_every_steps=10, supervisor_is_chief=True, supervisor_master='', supervisor_save_model_secs=600, keep_checkpoint_max=5, supervisor_save_summaries_steps=100, feed_fn=None, steps=None, fail_on_nan_loss=True, monitors=None, max_steps=None) Train a model. Given graph, a directory to write outputs to (output_dir), and some ops, run a training loop. The given trai

tf.contrib.bayesflow.stochastic_tensor.BinomialTensor.input_dict

tf.contrib.bayesflow.stochastic_tensor.BinomialTensor.input_dict

tf.contrib.graph_editor.matcher.input_ops()

tf.contrib.graph_editor.matcher.input_ops(*args) Add input matches.

tf.sub()

tf.sub(x, y, name=None) Returns x - y element-wise. NOTE: Sub supports broadcasting. More about broadcasting here Args: x: A Tensor. Must be one of the following types: half, float32, float64, int32, int64, complex64, complex128. y: A Tensor. Must have the same type as x. name: A name for the operation (optional). Returns: A Tensor. Has the same type as x.

tf.contrib.graph_editor.SubGraphView.__enter__()

tf.contrib.graph_editor.SubGraphView.__enter__() Allow Python context to minize the life time of a subgraph view. A subgraph view is meant to be a lightweight and transient object. A short lifetime will alleviate the "out-of-sync" issue mentioned earlier. For that reason, a SubGraphView instance can be used within a Python context. For example: from tensorflow.contrib import graph_editor as ge with ge.make_sgv(...) as sgv: print(sgv) Returns: Itself.

tf.contrib.distributions.LaplaceWithSoftplusScale.sample()

tf.contrib.distributions.LaplaceWithSoftplusScale.sample(sample_shape=(), seed=None, name='sample') Generate samples of the specified shape. Note that a call to sample() without arguments will generate a single sample. Args: sample_shape: 0D or 1D int32 Tensor. Shape of the generated samples. seed: Python integer seed for RNG name: name to give to the op. Returns: samples: a Tensor with prepended dimensions sample_shape.