问题
I have data analysis module that contains functions which call on Matplotlib pyplot API multiple times to generate up to 30 figures in each run. These figures get immediately written to disk after they are generated, and so I need to clear them from memory. Currently, at the end of each of my function, I do
import matplotlib.pyplot as plt
plt.clf()
however I'm not too sure if this statement can actually clear the memory. I'm especially concerned since I see that each time I run my module for debugging, my free memory space keeps decreasing. Can anyone please enlighten me on what I need to do to really clear my memory each time after I have written the plots to disk?
Thank you.
回答1:
Especially when you are running multiple processes or threads, it is much better to define your figure variable and work with it directly:
from matplotlib import pyplot as plt
f = plt.figure()
f.clear()
plt.close(f)
In any case, you must combine the use of plt.clear() and plt.close()
UPDATE (2021/01/21)
If you are using a MacOS system along with its default backend (referred as 'MacOSX'), this does NOT work (at least in Big Sur). The only solution I have found is to switch to other of the well-known backends, such as TkAgg, Cairo, etc. To do it, just type:
import matplotlib
matplotlib.use('TkAgg') # Your favorite interactive or non-interactive backend
回答2:
I have data analysis module that contains functions which call on Matplotlib pyplot API multiple
Can you edit your functions which is calling matplotlib? I was facing the same issue, I tried following command but none of it worked.
plt.close(fig)
fig.clf()
gc.collect()
%reset_selective -f fig
Then one trick worked for me, instead of creating a new figure every time, I pass the same fig object to the function and this solved my issue.
for example use,
fig = plt.figure()
for i in range(100):
plt.plot(x,y)
instead of,
for i in range(100):
fig = plt.figure()
plt.plot(x,y)
回答3:
Late answer, but this worked for me. I had a long sequential code generating many plots, and it would always end up eating all the RAM by the end of the process. Rather than calling fig.close() after each figure is complete, I have simply redefined the plt.figure function as follows, so that it is done automatically:
import matplotlib.pyplot as plt
import copy
try:
# if script it run multiple times, only redefine once
plt.old_figure
except:
# matplotlib is imported for the first time --> redefine
plt.old_figure = copy.deepcopy(plt.figure)
def newfig(*args):
plt.show()
plt.close("all")
return plt.old_figure(*args)
plt.figure = newfig
I am well aware that this is not a nice solution, however it easy, quick, and did the trick for me ! Maybe there's a way to decorate plt.figure instead of redefining it.
回答4:
After one week trials, I got my solution! Hope it can help you. My demo is attached.
import matplotlib.pyplot as plt
import numpy as np
A = np.arange(1,5)
B = A**2
cnt=0
while(1):
cnt = cnt+1
print("########### test %d ###########" % cnt)
# here is the trick:
# set the figure a 'num' to prevent from re-malloc of a figure in the next loop
# and set "clear=True" to make the figure clear
# I never use plt.close() to kill the figure, because I found it doesn't work.
# Only one figure is allocated, which can be self-released when the program quits.
# Before: 6000 times calling of plt.figure() ~ about 1.6GB of memory leak
# Now: the memory keeps in a stable level
fig = plt.figure(num=1, clear=True)
ax = fig.add_subplot()
# alternatively use an other function in one line
# fig, ax = plt.subplots(num=1,clear=True)
ax.plot(A,B)
ax.plot(B,A)
# Here add the functions you need
# plt.show()
fig.savefig('%d.png' % cnt)
来源:https://stackoverflow.com/questions/57000313/why-simple-python-program-slows-down-as-time-progresses