Skip to content

galrapo/markov-chain

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Markov Chain Transition Diagrams in Python

Simple Markov Chain visualization module in Python. Only requires matplotlib and numpy to work.

Description

The current version works with 2 to 4 states.

Getting Started

Dependencies

  • matplotlib
  • numpy

Installation

Copy the files src/node.py and src/markovchain.py in your script directory. Then

from markovchain import MarkovChain

2-state Markov chain demo

P = np.array([[0.8, 0.2], [0.1, 0.9]]) # Transition matrix
mc = MarkovChain(P, ['1', '2'])
mc.draw("../img/markov-chain-two-states.png")

two state markov chain transition diagram python

3-state Markov chain demo

P = np.array([
    [0.8, 0.1, 0.1],
    [0.1, 0.7, 0.2],
    [0.1, 0.7, 0.2],
])
mc = MarkovChain(P, ['A', 'B', 'C'])
mc.draw("../img/markov-chain-three-states.png")

three state markov chain transition diagram python

4-state Markov chain demo

P = np.array([
    [0.8, 0.1, 0.1, 0.0],
    [0.1, 0.7, 0.0, 0.2],
    [0.1, 0.0, 0.7, 0.2],
    [0.1, 0.0, 0.7, 0.2]
])
mc = MarkovChain(P, ['1', '2', '3', '4'])
mc.draw("../img/markov-chain-four-states.png")

four state markov chain transition diagram python

Author

Naysan Saran

Link to my blog.

License

This project is licensed under the GPL V3 licence.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 94.1%
  • Python 5.9%