Hierarchical RNNs, training bottlenecks and the future.
Hierarchy is inevitable. [Source: Google Images] As we know, the standard backpropagation algorithm is the most efficient procedure to compute the exact gradients of a loss function in a neural network … Read more