In Keras, what exactly am I configuring when I create a stateful `LSTM` layer with N `units`?
The first arguments in a normal Dense layer is also units , and is the number of neurons/nodes in that layer. A standard LSTM unit however looks like the following: (This is a reworked version of " Understanding LSTM Networks ") In Keras, when I create an LSTM object like this LSTM(units=N, ...) , am I actually creating N of these LSTM units? Or is it the size of the "Neural Network" layers inside the LSTM unit, i.e., the W 's in the formulas? Or is it something else? For context, I'm working based on this example code . The following is the documentation: https://keras.io/layers/recurrent/ It