Skip to content
Snippets Groups Projects
Commit 1d0a42ca authored by Matthew Johnson's avatar Matthew Johnson
Browse files

update readme image links

parent cdb3a911
No related branches found
No related tags found
No related merge requests found
......@@ -44,7 +44,7 @@ plt.plot(data[:,0],data[:,1],'kx')
plt.title('data')
```
![randomly generated mixture model data](http://www.mit.edu/~mattjj/github/pybasicbayes/datapoints.png)
![randomly generated mixture model data](https://raw.githubusercontent.com/mattjj/pybasicbayes/master/images/data.png)
Imagine we loaded these data from some measurements file and we wanted to fit a
mixture model to it. We can create a new `Mixture` and run inference to get a
......@@ -95,7 +95,7 @@ for scores in allscores:
plt.title('model vlb scores vs iteration')
```
![model vlb scores vs iteration](http://www.mit.edu/~mattjj/github/pybasicbayes/scores.png)
![model vlb scores vs iteration](https://raw.githubusercontent.com/mattjj/pybasicbayes/master/images/model-vlb-vs-iteration.png)
And show the point estimate of the best model by calling the convenient `Mixture.plot()`:
......@@ -104,7 +104,7 @@ models_and_scores[0][0].plot()
plt.title('best model')
```
![best fit model and data](http://www.mit.edu/~mattjj/github/pybasicbayes/fit.png)
![best fit model and data](https://raw.githubusercontent.com/mattjj/pybasicbayes/master/images/best-model.png)
Since these are Bayesian methods, we have much more than just a point estimate
for plotting: we have fit entire distributions, so we can query any confidence
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment