Krama wrote:It would probably be too silly of me to expect deepmind to release the weights of their NN.
What would they lose by doing that?
If nothing then why won't they do it?
I think this is not so simple. It's not just the weights that are in question, but several implementation details. DM probably experimented with quite a few NN structures / parameter tweaks before their network reached the levels published. It seems perfectly possible that even with decent expertise the first few implementations will be inferior to theirs. And while they may not lose much by publishing the weights, the implementation details are a different matter. That knowledge is basically the reason Google invested into RL the first place...
(BTW, for this reason I think the public/distributed "Zero" project will also need some kind of experimental/development channel, constantly trying new and new implementations, not just training a single net.)