Reinforcement Learning Memory Error

4 Ansichten (letzte 30 Tage)
Tech Logg Ding
Tech Logg Ding am 6 Dez. 2020
Kommentiert: Hasan Khanzada am 13 Apr. 2022
When I turn on the SaveExperienceBufferWithAgent, I get the following error:
Warning: Unable to save the agent to the directory "savedAgents". Increase the
disk space or check SaveAgentCriteriaValue in training options.
> In rl.train/TrainingManager/saveAgentToDisk (line 653)
In rl.train/TrainingManager/updateDisplaysFromTrainingInfo (line 717)
In rl.train/TrainingManager/update (line 147)
In rl.train.TrainingManager>@(info)update(this,info) (line 437)
In rl.train/Trainer/notifyEpisodeFinishedAndCheckStopTrain (line 56)
In rl.train.SeriesTrainer>iUpdateEpisodeFinished (line 31)
In rl.train.SeriesTrainer>@(src,ed)iUpdateEpisodeFinished(this,ed) (line 17)
In rl.env/AbstractEnv/notifyEpisodeFinished (line 324)
In rl.env.SimulinkEnvWithAgent.executeSimsWrapper/nestedSimFinishedBC (line 222)
In rl.env.SimulinkEnvWithAgent>@(src,ed)nestedSimFinishedBC(ed) (line 232)
In Simulink/SimulationManager/handleSimulationOutputAvailable
In Simulink.SimulationManager>@(varargin)obj.handleSimulationOutputAvailable(varargin{:})
In MultiSim.internal/SimulationRunnerSerial/executeImplSingle
In MultiSim.internal/SimulationRunnerSerial/executeImpl
In Simulink/SimulationManager/executeSims
In Simulink/SimulationManagerEngine/executeSims
In rl.env/SimulinkEnvWithAgent/executeSimsWrapper (line 244)
In rl.env/SimulinkEnvWithAgent/simWrapper (line 267)
In rl.env/SimulinkEnvWithAgent/simWithPolicyImpl (line 424)
In rl.env/AbstractEnv/simWithPolicy (line 82)
In rl.task/SeriesTrainTask/runImpl (line 33)
In rl.task/Task/run (line 21)
In rl.task/TaskSpec/internal_run (line 166)
In rl.task/TaskSpec/runDirect (line 170)
In rl.task/TaskSpec/runScalarTask (line 194)
In rl.task/TaskSpec/run (line 69)
In rl.train/SeriesTrainer/run (line 24)
In rl.train/TrainingManager/train (line 421)
In rl.train/TrainingManager/run (line 211)
In rl.agent.AbstractAgent/train (line 78)
The disk that I am running it on still has 100 GB of space. What is causing this issue?
  5 Kommentare
Tech Logg Ding
Tech Logg Ding am 12 Dez. 2020
Okay, I'll create a technical support case.
I created the experience buffer initially to have a 'save point' if the training crashes. However, i have turned it off and I still have the same issue.
Hasan Khanzada
Hasan Khanzada am 13 Apr. 2022
Your problem is solved ? because i am having the same issue as you were.

Melden Sie sich an, um zu kommentieren.

Antworten (0)

Produkte


Version

R2020b

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by