LSTM-Attention Layer Network dimensions for classification task
问题 I figured I build an attention model, but got confused (again) regarding each layers dimension. So say, I have 90 documents, each being composed by 200 sentence-vectors. The sentence-vectors are of size 500 (each sentence embedded as 1x500). The task is a classification of each document and the sentence-vectors are already embedded ! #Creating randm features xx = np.random.randint(100, size=(90,200,500)) y = np.random.randint(2, size=(90,1)) In the end, the attention-layer should return the