https://github.com/Klimmasch/AEC
Tip revision: 96e9ae2336937469a8f1602c178ea5e0cb8564b6 authored by Lukas Klimmasch on 13 August 2021, 14:16:04 UTC
Merge branch 'alternateRearing' of https://github.com/Klimmasch/AEC into alternateRearing
Merge branch 'alternateRearing' of https://github.com/Klimmasch/AEC into alternateRearing
Tip revision: 96e9ae2
CACLAActor.m
%%%
% Continuous Actor Critic Learning Automaton Actor
% DEPRICATED
%%%
classdef CACLAActor < handle
properties
% Network parameters
input_dim;
hidden_dim;
output_dim;
wp_ji; % input -> hidden weights
wp_kj; % hidden weights -> output
w_init_range; % weight initialization
% Reinforcement Learning parameters
beta_p; % step-size schedule of weight update "learning rate"
variance; % variance of perturbation distribution
varianceRange;
varDec;
% Model state tracking of previous time step
z_i_prev; % input layer activation
z_j_prev; % hidden layer activation
z_k_prev; % output layer activation
command_prev; % resulted action
% model history tracking
params;
end
methods
function obj = CACLAActor(PARAM)
obj.input_dim = PARAM{1}(1);
obj.hidden_dim = PARAM{1}(2);
obj.output_dim = PARAM{1}(3);
obj.w_init_range = PARAM{2};
obj.wp_ji = (2 * rand(obj.hidden_dim, obj.input_dim) - 1) * obj.w_init_range(1); % [-1, 1] * w_init_range
obj.wp_kj = (2 * rand(obj.output_dim, obj.hidden_dim) - 1) * obj.w_init_range(2); % [-1, 1] * w_init_range
obj.beta_p = PARAM{3};
obj.varianceRange = PARAM{4};
obj.variance = obj.varianceRange(1);
obj.varDec = PARAM{5};
obj.params = zeros(1, 2);
obj.z_i_prev = zeros(obj.input_dim, 1);
obj.z_j_prev = zeros(obj.hidden_dim, 1);
obj.z_k_prev = zeros(obj.output_dim, 1);
obj.command_prev = 0;
end
function update(this)
% delta_weights(hidden -> output)
dwp_kj = (this.command_prev - this.z_k_prev) * this.z_j_prev'; %1 50
% delta_weights(input -> hidden) [standard backprop]
dwp_ji = ((1 - this.z_j_prev .^ 2) * this.z_i_prev') * (this.wp_kj * dwp_kj') * this.z_i_prev;
this.wp_kj = this.wp_kj + this.beta_p * dwp_kj;
this.wp_ji = this.wp_ji + this.beta_p * dwp_ji * this.z_i_prev'; %50 578
% 50 1 1 578
end
function command = act(this, z_i)
z_j = tanh(this.wp_ji * z_i); % activity of hidden layer
z_k = this.wp_kj * z_j; % activity of output layer
command = mvnrnd(z_k, this.variance); % perturbation of actor's output multivariate version
% model state tracking
this.z_i_prev = z_i;
this.z_j_prev = z_j;
this.z_k_prev = z_k;
this.command_prev = command;
end
function command = actHard(this, z_i)
z_j = tanh(this.wp_ji * z_i); % activity of hidden layer
command = this.wp_kj * z_j; % activity of output layer
end
function command = train(this, feature, delta, flag_update)
if (flag_update && delta > 0)
this.update();
end
command = this.act(feature);
% model state change tracking
this.params(1) = sum(sum(abs(this.wp_ji)));
this.params(2) = sum(sum(abs(this.wp_kj)));
end
end
end