How to use CUDA pinned “zero-copy” memory for a memory mapped file?
Objective/Problem In Python, I am looking for a fast way to read/write data from a memory mapped file to a GPU. In a previous SO overflow post [ Cupy OutOfMemoryError when trying to cupy.load larger dimension .npy files in memory map mode, but np.load works fine ] Where it is mentioned this is possible using CUDA pinned "zero-copy" memory. Furthermore, it seems that this method was developed by this person [ cuda - Zero-copy memory, memory-mapped file ] though that person was working in C++. My previous attempts have been with Cupy, but I am open to any cuda methods. What I have tried so far I