Limitations
Future Work
A Comparison of the two approaches demonstrates that the more complex the
problem formulation is, the better the nature of solution obtained but
the more difficult it is achieve learning.
The first approach which involved a state space decomposition was a simpler
strategy which showed that the Q-Learning Network Algorithm works
well for a problem of this stature. However the second approach which provides
a more realistic solution is more difficult from the Q-learning point of
view . This is because the reward function , state space description etc
are quite complex.
- state space description ,
- action set
- reward functions
- probabilistic model ( exploration vs exploitation)