-
Notifications
You must be signed in to change notification settings - Fork 759
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
support for pymc3 as a modeling language? #20
Comments
Pymc3 has the most appealing DSL in my opinion. |
My biased opinion is that this is a great idea :). CC @jsalvatier @fonnesbeck. @dustinvtran Are you in touch with Kevin Murphy by any chance? We have discussed either allowing for an additional backend like TensorFlow (probably results in a For the first option, it shouldn't be too difficult to change out the backend. Theano is used to compute the model logp and its gradient. The step methods and variational inference are pure python (only some parts we sped up by Or were you imagining a different way of combining them? |
Keep in mind that PyMC isn't really another language -- one is essentially coding models in pure Python, by instantiating high-level PyMC objects. We have studiously avoided exposing most of Theano to users. The DSL (if you want to call it that) just uses standard Python idioms, for the most part (e.g. context managers for specifying models). |
Hi @twiecki @fonnesbeck. Thanks for the description! Yes, it would be great to have PyMC as a modeling language if it's not much effort to get it in. Generally, the library just needs a way to calculate the model's log joint density and its gradient, The easiest way to get PyMC into the library is just to treat the model's log density function as one node in TensorFlow's graph. There's some black box in how it's computed, and it need only take the right inputs and spit out the right output, as well as a way to differentiate this density. If there's a way to get log density evaluations and their gradient from PyMC programs, then this can be easily achieved. Further integration would be as you said, where we treat the PyMC program not as one node but its own corresponding graph from Theano ported over to TensorFlow. |
OK, that should be pretty simple actually, as we expose |
Perfect. Then the first step would be to write a wrapper in the same way one for Stan is written. It will take a vector of parameters from TensorFlow and converts it to the right dictionary format to plug into |
nice! |
This seems to work: import edward as ed
import numpy as np
import tensorflow as tf
from edward import PythonModel, Variational, Beta
from scipy.stats import beta, bernoulli
import pymc3 as pm
import theano
import numpy as np
data_shared = theano.shared(np.zeros(1))
with pm.Model() as model:
beta = pm.Beta('beta', 1, 1, transform=None)
out = pm.Bernoulli('data',
beta,
observed=data_shared)
class PyMC3Model:
"""
Model wrapper for models written in PyMC3.
Arguments
----------
model : pymc3.Model object
observed : The shared theano tensor passed to the model likelihood
"""
def __init__(self, model, observed):
self.model = model
self.observed = observed
vars = pm.inputvars(model.cont_vars)
bij = pm.DictToArrayBijection(pm.ArrayOrdering(vars), model.test_point)
self.logp = bij.mapf(model.fastlogp)
self.dlogp = bij.mapf(model.fastdlogp(vars))
self.num_vars = len(vars)
def log_prob(self, xs, zs):
return tf.py_func(self._py_log_prob, [xs, zs], [tf.float32])[0]
def _py_log_prob(self, xs, zs):
n_minibatch = zs.shape[0]
lp = np.zeros(n_minibatch, dtype=np.float32)
self.observed.set_value(xs)
for s in range(n_minibatch):
lp[s] = self.logp(zs[s, :])
return lp
data = ed.Data(np.array([0, 1, 0, 0, 0, 0, 0, 0, 0, 1]))
m = PyMC3Model(model, data_shared)
variational = Variational()
variational.add(Beta(m.num_vars))
inference = ed.MFVI(m, variational, data)
inference.run(n_iter=10000) |
Does it not require the gradient somewhere? |
this is cool. i'll try to play around with this sometime soon. @twiecki we implicitly end up calling tensorflow's autodiff in |
But how can it compute the gradient over the pymc3 blackbox node?
|
in this particular model, we use the score function gradient estimator. thus, we only need to evaluate |
Awesome! This is excellent. Yes, for this Beta-Bernoulli model we just use the score function gradient which does not require model gradients. When I run the code, it seems to converge to roughly -9.6. The model codes converge to around -6.2. Is this a bug or something different in how this model is written compared to the other Beta-Bernoulli examples? |
Ah, this is due to the autotransformation we apply to beta. Setting it to Regarding the autodiff -- is there an API for the model gradient? Also, is mini-batching supported right now? Seems like |
That's really cool. So I'm not too familiar with PyMC3 syntax. Is there a reason that data has to appear both when defining the model and when calculating the joint log density? Also, do you mind submitting a pull request so we can officially merge this in? :)
It's mentioned here in TensorFlow's API but I haven't spent much time figuring out how to register gradients. It would be excellent to do this though.
Yes, the default is that we subsample from the first slice of the data structure, so random elements from a vector in this case, or more generally rows in a matrix or matrices from a tensor. You can see this in |
It's not required, you only have to set something for the shape. I changed the above example to do so.
Yeah, I'll do that. |
Here we go: #75 |
We want to support modeling languages which are either popular or are useful for certain tasks over the alternatives we support. With that in mind, pymc3 seems appealing for specifying large discrete latent variable models. You can't write them as a Stan program, and it could be rather annoying to code them up in raw TensorFlow or NumPy/SciPy.
On the other hand, it's one more modeling language to maintain; pymc3 actually uses Theano as a backend, which may lead to some bugs(?); and I don't know how popular pymc3 would be as a use case over the other supported languages.
The text was updated successfully, but these errors were encountered: