Answered
How to create an custom Reinforcement Learning Environment + DDPG agent
Hi Kai, What the very first error is telling you is that there is an issue with the dimensions of either your observation, rewa...

3 months ago | 1

| accepted

Answered
Reinforcement Learning Sample Time
Hi Braydon, The agent sample time effectively determines how often the agent will output a decision/action. Think of it as the...

3 months ago | 0

Answered
2 out of 7 Observations Defined in MATLAB DDPG Reinforcement Learning Environment. Are the rest given random values?
Hello, I am assuming you are referring to the initialization of x and y inside the "flyingRobotResetFcn" function. Basically, i...

3 months ago | 0

| accepted

Answered
RL Agent does not learn
Hello, It is really hard to say just by looking at the training plot. The first thing I would try is 1) a different agent (mayb...

3 months ago | 0

| accepted

Answered
Regarding Reinforcement Learning with GridWorld
Hello, You would need to create a custom Grid world for this. Have a look at this link, and particularly the 'Actions' property...

3 months ago | 0

| accepted

Answered
Why is the DDPG episode rewards never change during the whole training process?
Looks like the scale between Q0 and episode reward is very different. Try unchecking "Show Episode Q0" to see of the episode rew...

4 months ago | 0

| accepted

Answered
Reinforcement Learning -- Rocket Lander
Hi Averill, Here is a version that converges in ~18-20k episodes - thank you for pointing out that this example was not converg...

4 months ago | 0

| accepted

Answered
QTable reset when using train
If you stop training, you should be able to continue from where you left off. I called 'train' on the basic grid world example a...

4 months ago | 0

| accepted

Answered
How to test critic network in DDPG agent?
You can use getValue for inference on critic networks.

4 months ago | 1

| accepted

Answered
Questions about Reinforcement Learning
Hi Averill, Can you please let me know which examples do not converge? I will share this information with the development team....

4 months ago | 0

Answered
Error while designing MPC controller where output of vehicle body block is using as measured output in MPC
Seems the error is self-explanatory - have you checked the dimensions of the signal you are connecting to the 'mo' port of the M...

5 months ago | 0

| accepted

Answered
Simulink MPC Controller Constraints
Hi James, By "controlled variables" I am assuming you are referring to measured outputs. I believe that with linear MPC you can...

5 months ago | 0

| accepted

Answered
Can I Disable the Kalman Filter for MPC?
The following example shows how to disable the built-in estimator: https://www.mathworks.com/help/mpc/ug/custom-state-estimatio...

5 months ago | 0

Answered
How to apply the range control with Model Predictive Control in Simulink?
It seems to me that the following two examples correspond to the two scenarios you are describing: https://www.mathworks.com/he...

5 months ago | 0

Answered
How to make MPC controller act before disturbance hits?
Hi Nathan, I am assuming you have seen this example here? If the controller acted earlier, would it not violate the rotation pe...

5 months ago | 0

Answered
Help me understand the Architecture of DQN cor Cartpole problem in RL
Hi Michael, There are various architectures you can use when setting up the Q-network. In the example you mentioned and most ex...

5 months ago | 0

Answered
Episode simulation doesn't run while training DDPG
Hi Alice, This example has not been set up to update the visualization during training. If you follow add a MATLAB Function blo...

5 months ago | 0

| accepted

Answered
Unable to run 'rlwatertank' example in R2020a
Hi Nima, This is the plot I got when running the watertank example in 20a: While this is not exactly the same as the one show...

5 months ago | 0

| accepted

Answered
In the process of reinforcement learning training, the training was terminated erroneously because the data dimensions did not match.
Are you using GPU to train? If yes, there is an indentified bug in R2020a which will be fixed when the first update for this rel...

5 months ago | 0

| accepted

Answered
Deep Deterministic Policy Gradient Agents (DDPG at Reinforcement Learning), actor output is oscilating a few times then got stuck on the minimum.
Hi Samir, After reviewing your model, if you check the actions the agent outputs, they blow up to infinity. That should not b...

6 months ago | 0

Answered
Reinforcement learning for adaptive cruise control
Hello, Try the following load('SimulinkACCDDPG.mat','agent') actor = getActor(agent); weights = getLearnableParameterValue...

6 months ago | 0

| accepted

Answered
Policy Gradient with Baseline Reward Oscillation (MATLAB Reinforcement Learning Toolbox)
Hello, Some suggestions: 1) For a 13 to 2 mapping, maybe you need another set of FCl+Relu layers in your actor 2) Since you h...

6 months ago | 0

Answered
Do we need a delay block in reinforcement learning simulink?
A delay block may be necessary to break algebraic loops or to rely on observations from previous time steps (see e.g. this examp...

6 months ago | 1

Answered
Reinforcement Learning Toolbox: defining Size of layers and other parameters.
Hello, I would recomment looking at a similar example in thw documentation (e.g. this one). To your questions: Yes. Number of ...

6 months ago | 0

Answered
training an inverted pendel with DDPG agent
Can you try saving to a different folder? It looks like a permissions issue.

6 months ago | 0

| accepted

Answered
Custom Action Space DDPG Reinforcement Learning Agent
To my knowledge, you cannot implement a custom action space with rlNumericSpec, but what you could possibly do (since adding pen...

7 months ago | 0

| accepted

Answered
Generate Cuda code from a pretrained rlDDPGAgent object for NVIDIA board
If you see here, tha tanhLayer support code generation with GPU Coder starting in R2019b.

7 months ago | 0

| accepted

Answered
Binary Decision Variable in MPC
This should be doable with custom constraints in nonlinear MPC. You can create your own function that decides how the constraint...

7 months ago | 0

Answered
Export the reinforcement learning result - to see the weights of critic network and actor network
Hello, You can see the values of the neural network weights using this function. Yes, you can apply DDPG and RL in general to ...

7 months ago | 0

Answered
Deploy trained policy to simulink model
Hello, Looks like the dimensions cannot be determined automatically. If you double click the MATLAB Fcn block and then click "E...

7 months ago | 0

| accepted

Load more