Git Product home page Git Product logo

bruno's People

Contributors

christabella avatar irakorshunova avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar

bruno's Issues

Purpose of inverse softplus and square root of GP/TP variance?

Why is the trainable variable of the variance of the GP/TP kernel (v in Eq. 9 of the BRUNO paper, diagonal values on the kernel matrix) expressed as the inverse softplus of the square root of the actual variance?

tf.constant_initializer(inv_softplus(np.sqrt(var_init)))

And then actual variance is recovered like this:

self.var = tf.square(tf.nn.softplus(self.var_vbl))

Is it simply to ensure v is non-negative? But then taking the square of self.var_vbl would suffice for that. Is the softplus some kind of trick for stability or convergence (and if so is it documented anywhere)? Thank you!

P.S. What does the m1 in m1_shapenet stand for? ๐Ÿค”

Drawbacks of conditional BRUNO compared to RNN BRUNO (not a bug)

Hello, thank you for open-sourcing the code! I have a few high-level questions about the models:

1. Why is validation only done for RNN and not conditional?

In the original RNN version, there is validation done during training:

bruno/config_rnn/train.py

Lines 201 to 208 in c631d3d

if hasattr(config, 'validate_every') and (iteration + 1) % config.validate_every == 0:
print('\n Validating ...')
losses = []
rng = np.random.RandomState(42)
for _, x_valid_batch in zip(range(0, config.n_valid_batches),
config.valid_data_iter.generate(rng)):
feed_dict = {x_in_eval: x_valid_batch}
l = sess.run([eval_loss], feed_dict)

Whereas in the conditional version, eval_loss is never used:

# evaluation in case we want to validate
x_in_eval = tf.placeholder(tf.float32, shape=(config.batch_size,) + config.obs_shape)
y_in_eval = tf.placeholder(tf.float32, shape=(config.batch_size,) + config.label_shape)
log_probs = model(x_in_eval, y_in_eval)[0]
eval_loss = config.eval_loss(log_probs) if hasattr(config, 'eval_loss') else config.loss(log_probs)

2. Is conditional BRUNO not maximizing joint (conditional) log likelihood?

BRUNO is clearly maximizing the joint log likelihood:
image

However, conditional BRUNO does not seem to be maximizing the joint conditional log likelihood... or is it?
image

3. "Conditional de Finetti" is not guaranteed

Do you think this is this a problem, or not really since in practice it works nonetheless?

Thank you very much!

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.