A simple library for experimenting with Markov Decision Process (MDP). Designed specifically for studying Navigation problems.
MDP is defined by states S, dynamics T, actions A, and rewards R. The design philosophy of this library is that each entity in the MDP is a separate object. State is the central entity. Everything else is optional, and can be defined and attached to the states as per the need.
To install current release with pip:
pip install navigation-mdp
To install from source:
python setup.py install
If you're curious what it can do:
Create a 3 x 3 state space:
S = DiscreteStateSpace(3,3)
Attach indicator features:
S.attach_feature_spec(FeatureStateIndicatorOneHot("ind"))
Visualize the world:
p = NavGridView(S.features(key="ind", gridded=True)[..., np.newaxis, np.newaxis]).render().ticks().grid()
plt.colorbar(p.im)
For visualizations: https://github.com/yrevar/navigation_vis
- Thanks to Prof. Michael L. Littman, Dr. Lucas Lehnert, and Dr. David Abel for all the discussions which were very helpful in developing concepts.
- State class is inspired from Dr. Abel's Simple RL framework: https://github.com/david-abel/simple_rl