tf.contrib.learn.monitors.CheckpointSaver.step_end()

tf.contrib.learn.monitors.CheckpointSaver.step_end(step, output) Callback after training step finished. This callback provides access to the tensors/ops evaluated at this step, including the additional tensors for which evaluation was requested in step_begin. In addition, the callback has the opportunity to stop training by returning True. This is useful for early stopping, for example. Note that this method is not called if the call to Session.run() that followed the last call to step_begin()

tf.contrib.learn.monitors.ExportMonitor.export_dir

tf.contrib.learn.monitors.ExportMonitor.export_dir

tf.contrib.graph_editor.reroute_b2a_inputs()

tf.contrib.graph_editor.reroute_b2a_inputs(sgv0, sgv1) Re-route all the inputs of sgv1 to sgv0 (see reroute_inputs).

tf.nn.rnn_cell.InputProjectionWrapper

class tf.nn.rnn_cell.InputProjectionWrapper Operator adding an input projection to the given cell. Note: in many cases it may be more efficient to not use this wrapper, but instead concatenate the whole sequence of your inputs in time, do the projection on this batch-concatenated sequence, then split it.

tf.contrib.distributions.StudentTWithAbsDfSoftplusSigma.log_prob()

tf.contrib.distributions.StudentTWithAbsDfSoftplusSigma.log_prob(value, name='log_prob') Log probability density/mass function (depending on is_continuous). Args: value: float or double Tensor. name: The name to give this op. Returns: log_prob: a Tensor of shape sample_shape(x) + self.batch_shape with values of type self.dtype.

tf.nn.rnn_cell.InputProjectionWrapper.state_size

tf.nn.rnn_cell.InputProjectionWrapper.state_size

tensorflow::Tensor::operator=()

Tensor& tensorflow::Tensor::operator=(const Tensor &other) Assign operator. This tensor shares other's underlying storage.

tf.contrib.rnn.GRUBlockCell.output_size

tf.contrib.rnn.GRUBlockCell.output_size

tf.FixedLenSequenceFeature.__getnewargs__()

tf.FixedLenSequenceFeature.__getnewargs__() Return self as a plain tuple. Used by copy and pickle.

tf.contrib.distributions.WishartCholesky

class tf.contrib.distributions.WishartCholesky The matrix Wishart distribution on positive definite matrices. This distribution is defined by a scalar degrees of freedom df and a lower, triangular Cholesky factor which characterizes the scale matrix. Using WishartCholesky is a constant-time improvement over WishartFull. It saves an O(nbk^3) operation, i.e., a matrix-product operation for sampling and a Cholesky factorization in log_prob. For most use-cases it often saves another O(nbk^3) opera