Bellman Value Iteration (VI) (Optimal Policies for MDPs Optimal Policies for MDPs): Revision history

Jump to navigation Jump to search

Diff selection: Mark the radio buttons of the revisions to compare and hit enter or the button at the bottom.
Legend: (cur) = difference with latest revision, (prev) = difference with preceding revision, m = minor edit.

    15 February 2023

    • curprev 10:5310:53, 15 February 2023Admin talk contribs 348 bytes +348 Created page with "== Time Complexity == $O({2}^n)$ == Space Complexity == $O(n)$ words (Only needs to store values (V) and policy (pi), both size O(n)) == Description == == Approximate? == Exact == Randomized? == No, deterministic == Model of Computation == Word/Real RAM == Year == 1957 == Reference == https://www.jstor.org/stable/24900506"