I have a question about torch.stack
I have 2 tensors, a.shape=(2, 3, 4) and b.shape=(2, 3). How to stack them without in-place operation?
suppose you have two tensors a, b which are equal in dimensions i.e a ( A, B, C) so b (A, B , C) an example
a=torch.randn(2,3,4)
b=torch.randn(2,3,4)
print(a.size()) # 2, 3, 4
print(b.size()) # 2, 3, 4
f=torch.stack([a, b], dim=2) # 2, 3, 2, 4
f
it wont act if they wouldn't be the same dim. Be careful!!