I'm playing with Keras a little bit and I'm thinking about what is the difference between linear activation layer and no activation layer at all? Doesn't it have the same behavior? If so, what's the point of linear activation then?
I mean the difference between these two code snippets:
model.add(Dense(1500))
model.add(Activation('linear'))
model.add(Dense(1500))
and
model.add(Dense(1500))
model.add(Dense(1500))