How to run the simulink model when implementing custom RL training?
1 Ansicht (letzte 30 Tage)
Ältere Kommentare anzeigen
Yihao Wan
am 25 Mai 2023
Kommentiert: Emmanouil Tzorakoleftherakis
am 25 Mai 2023
Hello, I am developing a custom training of RL DQN agent based on the link, however, how should I adapt it to the simulink environment?
Especially for the code below, when applying an action to the environment, the step is not applicable for a simulink model. How should I solve this issue? Thanks in advance.
% Apply the action to the environment
% and obtain the resulting observation and reward.
[nextObs,reward,isdone] = step(env,action{1});
0 Kommentare
Akzeptierte Antwort
Emmanouil Tzorakoleftherakis
am 25 Mai 2023
The way to do it would be to use runEpisode
2 Kommentare
Emmanouil Tzorakoleftherakis
am 25 Mai 2023
The example you are showing is model-based RL, it's different from what you mentioned at the beginning.
Weitere Antworten (0)
Siehe auch
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!