trentbrick / pytorchdiscreteflows Goto Github PK
View Code? Open in Web Editor NEWDiscrete Normalizing Flows implemented in PyTorch
Discrete Normalizing Flows implemented in PyTorch
Hi,
I was able to more or less get the reproduction of the second figure. It looks something like this:
I guess minor differences are due to different hyperparameters.
Thanks for your code. I guess the main issue was that one_hot_add implementation was missing (as one_hot_add and one_hot_minus are exactly the same in your current version). I have corrected it in my fork, and have added some tests to cross-check with tf implementation. I have not tested with MADE network or bipartite case, but hopefully should work.
Hello Trenton,
Thank you for the awesome work and effort here. Have you tried to turn the scale update enabled(having
When I try to update the scale, I have this error reported: ValueError: Inverse for 0 modulo 37 does not exist.
In the paper, the author mentioned that "For the scale, we add a mask whose elements are negative infinity on non-invertible values such as 0." (right above equation (6)), but I cannot find the related implementation in either of your or the TensorFlow implementations.
I have two additional observations:
Thanks,
Zhaoxuan
Hi,
I'm trying to run the example 'Fig2Replication', but I get some errors in the part 'Fit model'.
The first is TypeError: 'module' object is not callable
that comes from line 119 of disc_utils.py
.
This can be fixed (I suppose) by replacing torch.fft
with torch.fft.fft
(also at line 120).
After that, however, there is an error at line 121: IndexError: index 1 is out of bounds for dimension 1 with size 1
.
Any idea how to fix it?
Thanks
Curious if you were able to get the lm examples from the paper working.
Hi, Trent,
Thanks for your implementation! I am trying to go through the code, but did not find 'DiscreteFlowDemo.ipynb' which is stated in the Readme.md. Also 'Figure2Replication.ipynb' causes dead kernel.
Could you help to fix this? Thanks!
Best,
Yali
Hey!
I was looking into the replication of some of the other experiments in the paper and found that they have always used autoregressive base with autoregressive flow. Do you have any idea how they are exactly implemented? (The only thing the paper says is that its a autoregressive categorical distribution and for experiments with full-rank discrete distribution they have said they use transformers).
A declarative, efficient, and flexible JavaScript library for building user interfaces.
๐ Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. ๐๐๐
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google โค๏ธ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.