Dataflow matrix machines as generalized recurrent neural networks

Dec 27, 2016 01:29

A year ago I posted about dataflow programming and linear models of computation:

http://anhinga-anhinga.livejournal.com/82757.html

It turns out that those dataflow matrix machines are a fairly powerful generalization of recurrent neural networks.( Read more... )

advanced software, artificial intelligence, machine learning, strange technology, computer science, dataflow matrix machines, neural nets, software continualization

Leave a comment

datjko December 29 2016, 10:40:10 UTC
Misha, thank you, looks very interesting but I'm rather a newbie here. Could you answer a first bunch of my silly questions here or give me a link to some introduction? I started learning ml more or less seriously only recently and some looking into the papers you gave makes it clear for me: it's yet another interesting story I started to listen from the middle of the book ( ... )

Reply

anhinga_anhinga January 3 2017, 16:17:07 UTC
> Do you have a simple hint of what is that matrix transformation which would apply the neuron functions to appropriate data and produce a new computational matrix. Or, may be an example?Currently, we use a very simple form of Self. It has two arguments, and on the up movement it simply ( ... )

Reply

anhinga_anhinga January 3 2017, 17:08:06 UTC
4 in an interesting question ( ... )

Reply

datjko January 3 2017, 18:23:32 UTC
Misha, thank you a lot for so detailed answers! I'll keep making small steps and asking new questions.

Reply

anhinga_drafts February 4 2017, 08:35:13 UTC
Privet!

This is one of the most interesting talks I've ever heard in my life:

http://anhinga-drafts.livejournal.com/29954.html

Reply

datjko January 4 2017, 11:11:06 UTC
Misha, they say that "There's no sparse variable in TensorFlow ( ... )

Reply

anhinga_anhinga January 4 2017, 20:28:40 UTC
I'll also try to think about this...

Reply

anhinga_anhinga December 30 2016, 05:57:49 UTC
(more on 1. Of course, if one has unary neurons with "log" and "exp" activation functions, then multiplication is doable at a relatively moderate cost. Multiplication only becomes difficult, if one has the usual (rather impoverished) set of activation functions.

But even with "log" and "exp", there is some difficulty, because the most important case of multiplication is multiplication by zero. So, a bit more thought would be needed on handling this near zero. And, of course, sign reversal would be a separate unary operation. Still the idea of presenting multiplication in this fashion is neat, might be useful for some things, so let it be recorded here.)

Reply


Leave a comment

Up