Markov Decision Process
Project description
Markov Decision Process
Markov Decision Process
- Markov Decision Process
Installation
pip install md-pro
Usage
##################
### Parameters ###
##################
parser = argparse.ArgumentParser()
parser.add_argument('--sample_time', '-Ts', type=float, help='Ts=0.1',
default='0.1', required=False)
parser.add_argument('--gamma', '-gam', type=float, help='gamma=0.9',
default='0.9', required=False)
parser.add_argument('--x_grid', '-xgr', type=int, help='x_grid=5',
default='8', required=False)
parser.add_argument('--y_grid', '-ygr', type=int, help='y_grid=5',
default='5', required=False)
args = parser.parse_args()
params = vars(args)
####################################################
### Challenge with Markov Decision Process (MDP) ###
####################################################
# points
P=get_meshgrid_points(params)
# Topology
T, S = get_simple_topology_for_regular_grid(params, P)
# rewards
R = {'38': 100}
mdp_challenge = {'S': S, 'R': R, 'T': T, 'P': P}
dict_mdp=start_mdp(params, mdp_challenge)
plot_the_result(dict_mdp, mdp_challenge)
... should produce:
Citation
Please cite following document if you use this python package:
TODO
Image source: https://www.pexels.com/photo/photo-of-black-and-beige-wooden-chess-pieces-with-white-background-1083355/
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
md_pro-0.0.4.tar.gz
(9.8 kB
view hashes)
Built Distribution
md_pro-0.0.4-py3-none-any.whl
(29.1 kB
view hashes)