/flipv2/20121113-091142-2.5K-ReLST-asneeded_epochs_50_5runs_noalphadecay/stdout-flip-2.5K_4.txt
https://bitbucket.org/evan13579b/soar-ziggurat · Plain Text · 34662 lines · 32610 code · 2052 blank · 0 comment · 0 complexity · d3a0179b931f9256445520fb852fb030 MD5 · raw file
- Seeding... 4
- dir: dir isL
- Python-Soar Flip environment.
- To accept commands from an external sml process, you'll need to
- type 'slave <log file> <n decisons>' at the prompt...
- sourcing 'flip_predict.soar'
- ***********
- Total: 11 productions sourced.
- seeding Soar with 4 ...
- soar> Entering slave mode:
- - log file 'rl-slave-2.5K_4.log'....
- - will exit slave mode after 2500 decisions
- waiting for commands from an externally connected sml process...
- -/|sleeping...
- \sleeping...
- -sleeping...
- /sleeping...
- |sleeping...
- \-/|\-/|\-/|sleeping...
- \-/|\-/sleeping...
- |1: O: O2 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- rule alias: '*'
- rule alias: '*'
- \-/|\-/2: O: O3 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isR
- |\-3: O: O6 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, False)
- predict error 1
- dir: dir isL
- /|\4: O: O8 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, False)
- predict error 1
- dir: dir isL
- -/5: O: O10 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- |\-6: O: O11 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- /|7: O: O14 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- \-/8: O: O15 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isU
- |\-9: O: O17 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isL
- /|\-10: O: O19 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- /|\11: O: O21 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- -12: O: O24 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, False)
- predict error 1
- dir: dir isL
- /|\13: O: O26 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/|14: O: O28 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-/15: O: O30 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\-16: O: O32 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|17: O: O34 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-18: O: O36 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|\19: O: O38 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/|20: O: O39 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isL
- \-/|sleeping...
- \21: O: O42 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -22: O: O43 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- /|23: O: O45 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isU
- \-/24: O: O48 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\-25: O: O50 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- /|\-26: O: O52 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, False)
- predict error 1
- dir: dir isR
- /|27: O: O53 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- \-/28: O: O56 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\-29: O: O57 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isL
- /|\30: O: O60 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, False)
- predict error 1
- dir: dir isR
- -31: O: O61 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- /32: O: O64 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, False)
- predict error 1
- dir: dir isU
- |\-33: O: O66 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|\34: O: O68 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/|35: O: O69 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- \-36: O: O72 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, False)
- predict error 1
- dir: dir isU
- /|\37: O: O74 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/|38: O: O76 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, False)
- predict error 1
- dir: dir isU
- \-/39: O: O78 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\-40: O: O80 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- /|\-41: O: O81 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isR
- /42: O: O83 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isR
- |\-/43: O: O85 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isR
- |\-44: O: O87 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isL
- /|\45: O: O89 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- -/|46: O: O91 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isU
- \-/47: O: O94 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\48: O: O95 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isL
- -/49: O: O97 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isR
- |\-/50: O: O99 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- |\-/|\-sleeping...
- /sleeping...
- |51: O: O102 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, False)
- predict error 1
- dir: dir isR
- rule alias: '*'
- rule alias: '*'
- \52: O: O103 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- -/53: O: O106 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\-54: O: O107 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isU
- /|\55: O: O110 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -/|\sleeping...
- -56: O: O112 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, False)
- predict error 1
- dir: dir isR
- /|57: O: O113 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- \-58: O: O115 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- /|59: O: O117 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- \-/60: O: O119 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- |\-/61: O: O121 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- |62: O: O124 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-/63: O: O125 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- |\-64: O: O127 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- /|65: O: O130 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-/66: O: O131 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- |\-67: O: O133 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isL
- /|\68: O: O135 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isU
- -/|69: O: O138 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-70: O: O140 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, False)
- predict error 1
- dir: dir isL
- /|\-71: O: O141 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- /72: O: O143 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isL
- |\-73: O: O145 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isR
- /|\74: O: O147 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- -/75: O: O149 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- |\-76: O: O151 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isU
- /|\77: O: O154 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|\78: O: O156 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -79: O: O158 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|\80: O: O159 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isL
- -/|\81: O: O161 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isL
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- -82: O: O163 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isU
- /|\83: O: O166 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|\84: O: O168 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/85: O: O169 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- |\-/86: O: O172 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\87: O: O173 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isU
- -/88: O: O176 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |\-89: O: O177 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- /|90: O: O180 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-/91: O: O181 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- |92: O: O183 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- \-/93: O: O185 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- |\-94: O: O188 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- /|\95: O: O189 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- -/|\96: O: O192 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/|97: O: O194 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-/98: O: O196 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\99: O: O198 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -100: O: O199 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- /|\101: O: O201 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- -/|\-/|\-/|\-/|\-/|\-/|\-/|\-/|\-/|sleeping...
- \102: O: O204 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|\103: O: O206 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/|104: O: O208 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, False)
- predict error 1
- dir: dir isL
- \-105: O: O209 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- /106: O: O211 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isL
- |\107: O: O214 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/108: O: O216 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\-109: O: O218 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|110: O: O220 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-111: O: O221 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isR
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- /112: O: O223 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- |\-113: O: O226 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, False)
- predict error 1
- dir: dir isU
- /|\114: O: O228 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/115: O: O229 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isR
- |\-116: O: O231 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- /|\117: O: O234 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/|118: O: O235 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isU
- \-/119: O: O238 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\-120: O: O239 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isL
- /|\121: O: O242 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, False)
- predict error 1
- dir: dir isR
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- rule alias: '*'
- -122: O: O243 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- /|\-123: O: O245 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- /|\124: O: O248 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, False)
- predict error 1
- dir: dir isL
- -125: O: O249 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- /|\126: O: O252 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|\127: O: O254 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/|\128: O: O255 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- -/129: O: O257 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- |\-130: O: O259 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isL
- /|131: O: O262 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- rule alias: '*'
- rule alias: '*'
- \132: O: O264 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/133: O: O266 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\-134: O: O268 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|135: O: O270 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-/|136: O: O272 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-137: O: O274 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|\138: O: O275 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- -/|\139: O: O278 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/140: O: O279 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isU
- |\141: O: O282 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -142: O: O283 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isU
- /|\143: O: O286 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/144: O: O288 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\-145: O: O289 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isL
- /|146: O: O291 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- \-/|147: O: O293 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isU
- \-/|148: O: O296 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-149: O: O298 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|150: O: O300 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-151: O: O301 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- /152: O: O303 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isU
- |\-/153: O: O306 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\-/154: O: O307 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isU
- |\-/sleeping...
- |155: O: O310 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- \-/156: O: O312 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\157: O: O314 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/|158: O: O315 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isL
- \-/159: O: O318 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, False)
- predict error 1
- dir: dir isU
- |\-/160: O: O319 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isU
- |\-161: O: O321 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isL
- /162: O: O323 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isU
- |\-163: O: O326 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|\-164: O: O327 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isU
- /|\165: O: O330 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/166: O: O332 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\167: O: O334 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/|168: O: O336 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-/169: O: O338 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\-170: O: O340 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|\-171: O: O342 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /172: O: O344 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- |\173: O: O345 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- -/174: O: O348 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\-175: O: O350 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|176: O: O352 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-/177: O: O353 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- |\178: O: O356 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|179: O: O358 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-/180: O: O360 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\-181: O: O362 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /182: O: O364 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\-183: O: O366 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|184: O: O368 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-/185: O: O370 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- |\186: O: O371 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- -/|187: O: O374 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-188: O: O375 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- /|\189: O: O378 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/|190: O: O380 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-191: O: O381 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isU
- /192: O: O384 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- |\193: O: O385 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- -/|194: O: O387 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isL
- \-/|sleeping...
- \195: O: O390 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, False)
- predict error 1
- dir: dir isL
- -/|\196: O: O392 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/|197: O: O394 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, False)
- predict error 1
- dir: dir isR
- \-/198: O: O396 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\-199: O: O398 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- /|\200: O: O400 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -/|201: O: O401 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- \-202: O: O403 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- /|\203: O: O406 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/|204: O: O408 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-205: O: O409 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- /|\-206: O: O412 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|\207: O: O414 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -208: O: O416 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|\209: O: O417 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- -/|210: O: O420 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- \-/211: O: O422 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |212: O: O423 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- \-/213: O: O426 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- |\-214: O: O428 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, False)
- predict error 1
- dir: dir isU
- /|\215: O: O430 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -/216: O: O431 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- |\-217: O: O434 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|\218: O: O435 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- -/|219: O: O438 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- \-/220: O: O440 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\221: O: O442 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -222: O: O444 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|\223: O: O446 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -/224: O: O447 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- |225: O: O450 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-226: O: O452 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|\227: O: O454 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|228: O: O456 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-229: O: O457 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- /|\230: O: O460 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -/|\231: O: O461 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- -232: O: O463 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isL
- /|\233: O: O466 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/|234: O: O468 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-/|235: O: O469 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- \-/236: O: O472 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\-/237: O: O474 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\-238: O: O475 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isU
- /|\-239: O: O477 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isL
- /|\240: O: O479 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- -/241: O: O482 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- |242: O: O483 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- \-/243: O: O485 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isU
- |\244: O: O488 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/|\245: O: O490 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -/|246: O: O491 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- \-/|247: O: O494 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-248: O: O496 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|249: O: O497 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- \-/250: O: O500 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\-/251: O: O502 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |252: O: O504 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- \-/253: O: O506 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\-254: O: O508 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- /|\-255: O: O509 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- /|\256: O: O512 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/|257: O: O513 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- \-/|258: O: O516 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-/259: O: O517 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- |\-260: O: O520 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|\261: O: O522 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -262: O: O523 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- /|263: O: O526 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-/264: O: O527 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- |\-/265: O: O530 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- |\-266: O: O531 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- /267: O: O534 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |268: O: O535 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- \-/|269: O: O538 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-/270: O: O540 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\271: O: O542 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -272: O: O543 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isU
- /|\273: O: O546 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/|274: O: O547 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- \-/275: O: O550 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\276: O: O552 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/|277: O: O554 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-/278: O: O555 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- |\-279: O: O557 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- /|\280: O: O560 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -/|281: O: O561 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- \282: O: O564 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, False)
- predict error 1
- dir: dir isL
- -/|283: O: O565 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- \-/284: O: O568 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\285: O: O570 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/|286: O: O572 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-/287: O: O574 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\288: O: O576 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|289: O: O577 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isU
- \290: O: O579 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isU
- -/291: O: O582 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |292: O: O584 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-/293: O: O585 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- |\-294: O: O588 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- /|295: O: O590 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- \-296: O: O592 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|297: O: O594 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- \-/298: O: O596 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |\-/299: O: O597 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- |\-300: O: O599 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- /|\-/|301: O: O602 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- \302: O: O604 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -303: O: O606 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|\304: O: O608 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -/|305: O: O609 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- \-/|306: O: O612 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-/307: O: O613 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- |\-308: O: O616 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|\309: O: O618 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/|310: O: O620 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-311: O: O621 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- /312: O: O624 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\313: O: O626 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/|314: O: O627 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- \-315: O: O630 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|\316: O: O632 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/|317: O: O634 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- \-/318: O: O636 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\-319: O: O638 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- /|320: O: O639 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- \-/|321: O: O641 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- \322: O: O643 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- -323: O: O646 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|\324: O: O647 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- -/|325: O: O650 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- \-/326: O: O652 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\-/327: O: O653 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isU
- |\-328: O: O656 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- /|\329: O: O658 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -/330: O: O659 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- |\331: O: O662 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -332: O: O664 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|\-333: O: O666 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|\334: O: O668 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/|335: O: O669 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- \-/336: O: O671 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- |\-/337: O: O674 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\-338: O: O676 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|\339: O: O678 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/340: O: O680 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\-/341: O: O682 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |342: O: O684 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-/|343: O: O686 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-/344: O: O688 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\-/345: O: O690 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\-346: O: O691 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isL
- /|347: O: O694 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-348: O: O696 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|\-sleeping...
- /349: O: O698 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- |\-/350: O: O699 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- |\351: O: O702 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -352: O: O703 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- /|353: O: O705 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- \-354: O: O707 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- /|\355: O: O710 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/|\356: O: O712 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/|\sleeping...
- -357: O: O714 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|\-358: O: O716 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|\359: O: O718 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|\360: O: O720 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/|361: O: O722 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \362: O: O724 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|363: O: O726 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-364: O: O728 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|\365: O: O729 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- -/|366: O: O732 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- \-/367: O: O734 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\-/368: O: O736 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\-369: O: O738 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|\370: O: O740 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/|371: O: O742 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- \372: O: O744 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/373: O: O746 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |\-374: O: O747 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- /|\375: O: O750 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/|376: O: O751 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- \-377: O: O753 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- /|\378: O: O756 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|\379: O: O758 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/|380: O: O760 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-381: O: O761 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- /382: O: O764 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\-383: O: O766 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- /|\-384: O: O767 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- /|\385: O: O769 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- -/|\386: O: O772 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -/387: O: O773 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- |\-388: O: O776 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|\389: O: O778 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|390: O: O780 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-/|391: O: O782 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \392: O: O784 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|393: O: O786 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-/|394: O: O787 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, False)
- predict error 1
- dir: dir isL
- \-/395: O: O790 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- |\-396: O: O791 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- /|\397: O: O794 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/|398: O: O795 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isL
- \-/399: O: O797 (predict-yes)
- I see 0 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- |\400: O: O800 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|\401: O: O802 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -402: O: O804 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|\403: O: O806 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|404: O: O808 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-/405: O: O809 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- |\-406: O: O812 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /407: O: O814 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\-408: O: O816 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- /|\409: O: O817 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- -/410: O: O819 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- |\-411: O: O822 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- /412: O: O823 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- |\-413: O: O825 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- /|\414: O: O828 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/|415: O: O830 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-/416: O: O831 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- |\-417: O: O833 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- /|\418: O: O835 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- -/|419: O: O838 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-420: O: O839 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- /|\421: O: O841 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- -422: O: O844 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|\423: O: O846 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, False)
- predict error 1
- dir: dir isR
- -/|\424: O: O848 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/|\425: O: O850 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/426: O: O852 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\-427: O: O854 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|\428: O: O856 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/|\429: O: O858 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/430: O: O860 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\431: O: O862 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -432: O: O864 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|\-433: O: O866 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- /|\434: O: O868 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -/|\435: O: O869 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- -/|436: O: O872 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-437: O: O873 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- /|\438: O: O876 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/|\439: O: O878 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -/|\440: O: O879 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- -/|441: O: O882 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \442: O: O884 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/443: O: O886 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\444: O: O888 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -445: O: O890 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|\446: O: O892 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/447: O: O894 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\-448: O: O896 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|\449: O: O897 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- -/|\450: O: O899 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- -/|451: O: O902 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \452: O: O903 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- -/453: O: O906 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |\-454: O: O907 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- /|\-455: O: O910 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|456: O: O912 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-457: O: O913 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- /|458: O: O915 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- \-/459: O: O917 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- |\460: O: O920 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/461: O: O921 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, False)
- predict error 1
- dir: dir isU
- |462: O: O924 (predict-no)
- I see 0 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-/463: O: O925 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- |\-464: O: O927 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- /|465: O: O930 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-466: O: O931 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- /|\467: O: O933 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- -/|468: O: O936 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- \-469: O: O938 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|470: O: O940 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-471: O: O941 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- /472: O: O944 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\473: O: O946 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|474: O: O948 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-/|475: O: O949 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- \-/476: O: O951 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- |\477: O: O954 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|478: O: O956 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-/479: O: O958 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\-480: O: O960 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|\481: O: O962 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -482: O: O963 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- /|\-sleeping...
- /483: O: O966 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\-484: O: O968 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- /|485: O: O969 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- \-/486: O: O971 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- |\-487: O: O973 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- /|\488: O: O975 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- -/|489: O: O977 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- \-490: O: O980 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|491: O: O982 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \492: O: O984 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/493: O: O986 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\-494: O: O988 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|\495: O: O990 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|496: O: O992 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-/|497: O: O994 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-498: O: O996 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|\499: O: O998 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|500: O: O1000 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-/|\-501: O: O1002 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /502: O: O1004 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- |\-503: O: O1005 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- /|\504: O: O1007 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- -505: O: O1009 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- /|\506: O: O1012 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/|507: O: O1014 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-508: O: O1015 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- /|509: O: O1018 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-/510: O: O1020 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\-511: O: O1022 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /512: O: O1024 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\-513: O: O1026 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|\514: O: O1027 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- -/515: O: O1030 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\-/516: O: O1032 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |\517: O: O1033 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- -/|518: O: O1036 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-519: O: O1038 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|\520: O: O1040 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/521: O: O1042 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |522: O: O1044 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-/523: O: O1046 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\-/524: O: O1048 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\-525: O: O1050 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|\-526: O: O1052 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|\527: O: O1054 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/528: O: O1056 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- |\-/529: O: O1057 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- |\-530: O: O1060 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|\-531: O: O1062 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- /532: O: O1063 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- |533: O: O1066 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-/534: O: O1068 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\535: O: O1070 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|536: O: O1072 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-537: O: O1074 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|538: O: O1076 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-/539: O: O1078 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\-540: O: O1080 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|\541: O: O1082 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -542: O: O1084 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|\-543: O: O1086 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|\-544: O: O1088 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|545: O: O1090 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-/546: O: O1091 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- |\-547: O: O1094 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- /|\548: O: O1096 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/|549: O: O1098 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- \-/550: O: O1100 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\-551: O: O1102 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /552: O: O1104 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |\553: O: O1105 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- -/|554: O: O1108 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-/|555: O: O1110 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-/556: O: O1112 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- |\-557: O: O1113 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- /|\558: O: O1115 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- -/559: O: O1117 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- |\-/560: O: O1120 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\-561: O: O1122 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /562: O: O1124 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\-563: O: O1126 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- /|\564: O: O1127 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- -/|565: O: O1129 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- \-566: O: O1132 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- /|\567: O: O1133 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- -/568: O: O1136 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\569: O: O1138 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|570: O: O1140 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-/571: O: O1142 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |572: O: O1144 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-/573: O: O1145 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- |\-574: O: O1147 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- /|\575: O: O1150 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|576: O: O1152 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-/|577: O: O1153 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- \-/|578: O: O1155 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- \-/579: O: O1157 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- |\580: O: O1160 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/581: O: O1162 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |582: O: O1164 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-583: O: O1165 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- /|\584: O: O1168 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/585: O: O1170 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\-586: O: O1172 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|587: O: O1173 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- \-/588: O: O1175 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- |\-589: O: O1178 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|\590: O: O1179 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- -/|591: O: O1182 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- \592: O: O1184 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/593: O: O1186 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\-594: O: O1188 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- /|\-595: O: O1189 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- /|\596: O: O1191 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- -/|597: O: O1194 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- \-598: O: O1196 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|\599: O: O1198 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -/|600: O: O1199 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- \-/601: O: O1201 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- |602: O: O1204 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-/603: O: O1205 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- |\604: O: O1208 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/|605: O: O1210 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-606: O: O1212 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|\-607: O: O1213 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- /|\608: O: O1215 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- -/609: O: O1217 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- |\610: O: O1219 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- -/|\611: O: O1221 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- -612: O: O1224 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- /|\-613: O: O1226 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|\614: O: O1228 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/|615: O: O1230 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-616: O: O1231 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- /|\-617: O: O1233 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- /|\618: O: O1236 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/619: O: O1238 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\-620: O: O1240 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|\621: O: O1242 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -622: O: O1243 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- /|623: O: O1245 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- \-/624: O: O1247 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- |\-/625: O: O1250 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- |\626: O: O1251 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- -/|627: O: O1254 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-/|628: O: O1255 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- \-/629: O: O1258 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- |\-630: O: O1259 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- /|\631: O: O1261 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- -632: O: O1264 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|\633: O: O1266 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/|634: O: O1268 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-/635: O: O1270 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\-636: O: O1272 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|637: O: O1273 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- \-/638: O: O1276 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |\-/639: O: O1277 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- |\-640: O: O1280 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|641: O: O1281 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- \642: O: O1284 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -/|643: O: O1285 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- \-644: O: O1287 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- /|\645: O: O1289 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- -/|\646: O: O1292 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/647: O: O1294 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\-/648: O: O1296 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\649: O: O1298 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/|\650: O: O1299 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- -/651: O: O1301 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- |652: O: O1303 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- \-653: O: O1306 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|654: O: O1308 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-655: O: O1309 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- /|\656: O: O1311 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- -/657: O: O1313 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- |\-658: O: O1315 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- /|\659: O: O1318 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -/|660: O: O1319 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- \-/661: O: O1322 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |662: O: O1324 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-/663: O: O1326 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\664: O: O1328 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -665: O: O1330 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|666: O: O1332 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-/667: O: O1334 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\-668: O: O1336 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|669: O: O1338 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-/|670: O: O1340 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-/671: O: O1341 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- |672: O: O1344 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-673: O: O1345 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- /|\674: O: O1347 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- -/|675: O: O1350 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- \-/676: O: O1352 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\-677: O: O1354 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- /|\678: O: O1356 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/679: O: O1358 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |\-/680: O: O1359 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- |\681: O: O1361 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- -682: O: O1363 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- /|\-683: O: O1366 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|\684: O: O1368 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/|685: O: O1370 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-/686: O: O1371 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- |\-687: O: O1373 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- /|688: O: O1375 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- \-689: O: O1377 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- /|\690: O: O1379 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- -/|691: O: O1381 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- \692: O: O1384 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/693: O: O1386 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\-694: O: O1388 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|\-695: O: O1389 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- /|\696: O: O1392 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -/|697: O: O1393 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- \-698: O: O1395 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- /|\699: O: O1398 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/|\700: O: O1400 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/|701: O: O1402 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \702: O: O1403 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- -/703: O: O1405 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- |\704: O: O1407 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- -/|705: O: O1409 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- \-/706: O: O1412 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\707: O: O1414 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/|708: O: O1416 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- \-/709: O: O1418 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |\-710: O: O1419 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- /|\-711: O: O1422 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /712: O: O1424 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\-/713: O: O1426 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |714: O: O1428 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-/|715: O: O1429 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- \-/716: O: O1431 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- |\-717: O: O1434 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|718: O: O1436 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-/719: O: O1438 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |720: O: O1440 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-/721: O: O1442 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- |722: O: O1443 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- \-723: O: O1445 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- /|724: O: O1447 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- \-/725: O: O1450 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |\726: O: O1451 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- -/|\727: O: O1454 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|728: O: O1456 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-729: O: O1458 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|730: O: O1460 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \731: O: O1462 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -732: O: O1464 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|733: O: O1466 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-/734: O: O1468 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\735: O: O1470 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|736: O: O1472 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-/737: O: O1474 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- |\738: O: O1475 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- -/739: O: O1478 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |\-/sleeping...
- |740: O: O1479 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- \-741: O: O1482 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /742: O: O1484 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- |\743: O: O1485 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- -/|744: O: O1487 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- \-/|745: O: O1489 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- \-/746: O: O1492 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |\747: O: O1493 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- -/|\748: O: O1496 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/749: O: O1498 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\750: O: O1500 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/|\751: O: O1502 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -752: O: O1504 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|\753: O: O1506 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/|754: O: O1507 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- \-/755: O: O1510 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\-/756: O: O1512 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\757: O: O1514 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/|758: O: O1516 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- \-/759: O: O1518 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\760: O: O1520 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -761: O: O1521 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- /762: O: O1523 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- |\-763: O: O1526 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- /|\-764: O: O1528 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- /|\765: O: O1530 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/|766: O: O1532 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- \-767: O: O1534 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|\768: O: O1536 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/769: O: O1538 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |\770: O: O1539 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- -/|771: O: O1541 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- \772: O: O1543 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- -/773: O: O1546 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\-774: O: O1548 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|775: O: O1550 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-/776: O: O1551 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- |\777: O: O1554 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/|\778: O: O1556 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/|779: O: O1558 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-/780: O: O1559 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- |\781: O: O1562 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -782: O: O1564 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|\783: O: O1565 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- -/|784: O: O1568 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-/785: O: O1569 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- |\-786: O: O1572 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|787: O: O1574 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-788: O: O1576 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|789: O: O1578 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-790: O: O1580 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|\-sleeping...
- /791: O: O1581 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- |792: O: O1584 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- \-/793: O: O1586 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\-794: O: O1588 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- /|\-795: O: O1590 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|\796: O: O1592 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -/|797: O: O1593 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- \-/798: O: O1595 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- |\-799: O: O1597 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- /|\800: O: O1600 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/|801: O: O1601 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- \802: O: O1604 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -/|803: O: O1605 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- \-/804: O: O1608 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- |\805: O: O1609 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- -/|806: O: O1612 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- \-/807: O: O1614 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\808: O: O1616 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/|809: O: O1618 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- \-/810: O: O1620 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |\-811: O: O1621 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- /812: O: O1624 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\-/813: O: O1626 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\-814: O: O1628 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|\815: O: O1630 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/|816: O: O1631 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- \-/|817: O: O1634 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-/818: O: O1635 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- |\-819: O: O1638 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|\-820: O: O1639 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- /|\821: O: O1641 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- -822: O: O1644 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|823: O: O1645 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- \-/|824: O: O1648 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- \-/825: O: O1650 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\826: O: O1652 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/|\827: O: O1654 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -828: O: O1656 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|829: O: O1658 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- \-/|830: O: O1660 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- \-/831: O: O1662 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |832: O: O1664 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- \-/|833: O: O1666 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- \-/834: O: O1668 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\-835: O: O1670 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|836: O: O1672 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- \-/837: O: O1674 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |\-838: O: O1675 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- /|839: O: O1677 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- \-840: O: O1679 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- /|\841: O: O1681 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- -842: O: O1683 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- /|\-843: O: O1686 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|844: O: O1688 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-/845: O: O1690 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\846: O: O1692 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/847: O: O1694 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\848: O: O1696 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/|849: O: O1698 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-/850: O: O1700 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |851: O: O1702 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \852: O: O1704 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/|\853: O: O1705 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- -/|854: O: O1707 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- \-/855: O: O1710 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\856: O: O1712 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/857: O: O1713 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- |\-858: O: O1716 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- /|859: O: O1718 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-860: O: O1719 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- /|\861: O: O1722 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -862: O: O1724 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|863: O: O1725 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- \-/|864: O: O1727 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- \-/865: O: O1730 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\-/866: O: O1732 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\867: O: O1734 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/|868: O: O1736 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- \-869: O: O1738 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|870: O: O1739 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- \-/871: O: O1742 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |872: O: O1744 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- \-873: O: O1746 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|874: O: O1748 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- \-/875: O: O1750 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- |\-876: O: O1752 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- /|877: O: O1754 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-878: O: O1755 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- /|\879: O: O1758 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- -/|880: O: O1760 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- \-881: O: O1762 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /882: O: O1763 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- |\-883: O: O1766 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- /|884: O: O1768 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- \-/885: O: O1770 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |\-/886: O: O1771 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- |\-887: O: O1774 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|\-888: O: O1775 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- /|\889: O: O1778 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- -/|890: O: O1779 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- \-891: O: O1782 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /892: O: O1783 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- |\893: O: O1786 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/894: O: O1788 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\895: O: O1790 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/896: O: O1792 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\897: O: O1794 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/898: O: O1796 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- |\899: O: O1798 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- -/|900: O: O1800 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-901: O: O1801 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- /902: O: O1804 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\-903: O: O1806 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- /|\904: O: O1807 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- -/|905: O: O1810 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- \-/|906: O: O1812 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- \-907: O: O1814 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- /|\908: O: O1816 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/909: O: O1818 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |\910: O: O1819 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- -/|911: O: O1821 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- \912: O: O1823 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- -/|\913: O: O1826 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -914: O: O1827 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- /|915: O: O1830 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- \-/916: O: O1832 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- |\917: O: O1833 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- -/918: O: O1836 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\-/919: O: O1838 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- |\-920: O: O1840 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|\921: O: O1842 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -922: O: O1844 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- /|\923: O: O1846 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- -/924: O: O1848 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- |\925: O: O1849 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- -926: O: O1852 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- /|\927: O: O1854 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- -/|928: O: O1856 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- \-/929: O: O1857 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- |\930: O: O1860 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/931: O: O1861 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- |932: O: O1863 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- \-933: O: O1866 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|\934: O: O1868 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/|935: O: O1869 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- \-/936: O: O1871 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- |\937: O: O1873 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- -/|938: O: O1875 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- \-/939: O: O1878 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |\-940: O: O1880 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|\-sleeping...
- /941: O: O1882 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- |942: O: O1884 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- \-/943: O: O1885 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- |\-944: O: O1887 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- /|\-945: O: O1890 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- /|\946: O: O1892 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- -/|947: O: O1893 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- \-948: O: O1896 (predict-no)
- I see 1 and I'm going to do: predict-no
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- /|949: O: O1897 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- \-/950: O: O1899 (predict-yes)
- I see 1 and I'm going to do: predict-yes
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- |\-/|\-/|--- Input Phase ---
- =>WM: (13313: I2 ^dir U)
- =>WM: (13312: I2 ^reward 1)
- =>WM: (13311: I2 ^see 1)
- =>WM: (13310: N950 ^status complete)
- <=WM: (13298: I2 ^dir R)
- <=WM: (13297: I2 ^reward 1)
- <=WM: (13296: I2 ^see 1)
- =>WM: (13314: I2 ^level-1 R1-root)
- <=WM: (13299: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R954 ^value 1 +)
- (R1 ^reward R954 +)
- Firing propose*predict-yes
- -->
- (O1901 ^name predict-yes +)
- (S1 ^operator O1901 +)
- Firing propose*predict-no
- -->
- (O1902 ^name predict-no +)
- (S1 ^operator O1902 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1900 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1899 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1900 ^name predict-no +)
- (S1 ^operator O1900 +)
- Retracting propose*predict-yes
- -->
- (O1899 ^name predict-yes +)
- (S1 ^operator O1899 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R953 ^value 1 +)
- (R1 ^reward R953 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O1900 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1900 = 0.4476192676183378)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1899 = 0.8155729125006117)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1899 = 0.1844075378173239)
- =>WM: (13321: S1 ^operator O1902 +)
- =>WM: (13320: S1 ^operator O1901 +)
- =>WM: (13319: I3 ^dir U)
- =>WM: (13318: O1902 ^name predict-no)
- =>WM: (13317: O1901 ^name predict-yes)
- =>WM: (13316: R954 ^value 1)
- =>WM: (13315: R1 ^reward R954)
- <=WM: (13306: S1 ^operator O1899 +)
- <=WM: (13308: S1 ^operator O1899)
- <=WM: (13307: S1 ^operator O1900 +)
- <=WM: (13305: I3 ^dir R)
- <=WM: (13301: R1 ^reward R953)
- <=WM: (13304: O1900 ^name predict-no)
- <=WM: (13303: O1899 ^name predict-yes)
- <=WM: (13302: R953 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1901 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1902 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1900 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1899 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675409 -0.491002 0.184408 -> 0.675413 -0.491002 0.18441(R,m,v=1,0.89441,0.0950311)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*46 0.324566 0.491007 0.815573 -> 0.324569 0.491006 0.815576(R,m,v=1,1,0)
- =>WM: (13322: S1 ^operator O1902)
- 951: O: O1902 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N951 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N950 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13323: I3 ^predict-no N951)
- <=WM: (13310: N950 ^status complete)
- <=WM: (13309: I3 ^predict-yes N950)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- \--- Input Phase ---
- =>WM: (13327: I2 ^dir R)
- =>WM: (13326: I2 ^reward 1)
- =>WM: (13325: I2 ^see 0)
- =>WM: (13324: N951 ^status complete)
- <=WM: (13313: I2 ^dir U)
- <=WM: (13312: I2 ^reward 1)
- <=WM: (13311: I2 ^see 1)
- =>WM: (13328: I2 ^level-1 R1-root)
- <=WM: (13314: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1901 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1902 = 0.5523833737960075)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R955 ^value 1 +)
- (R1 ^reward R955 +)
- Firing propose*predict-yes
- -->
- (O1903 ^name predict-yes +)
- (S1 ^operator O1903 +)
- Firing propose*predict-no
- -->
- (O1904 ^name predict-no +)
- (S1 ^operator O1904 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1902 = 0.4476192676183378)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1901 = 0.1844104702696336)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1902 ^name predict-no +)
- (S1 ^operator O1902 +)
- Retracting propose*predict-yes
- -->
- (O1901 ^name predict-yes +)
- (S1 ^operator O1901 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R954 ^value 1 +)
- (R1 ^reward R954 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1902 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1901 = 0.)
- =>WM: (13336: S1 ^operator O1904 +)
- =>WM: (13335: S1 ^operator O1903 +)
- =>WM: (13334: I3 ^dir R)
- =>WM: (13333: O1904 ^name predict-no)
- =>WM: (13332: O1903 ^name predict-yes)
- =>WM: (13331: R955 ^value 1)
- =>WM: (13330: R1 ^reward R955)
- =>WM: (13329: I3 ^see 0)
- <=WM: (13320: S1 ^operator O1901 +)
- <=WM: (13321: S1 ^operator O1902 +)
- <=WM: (13322: S1 ^operator O1902)
- <=WM: (13319: I3 ^dir U)
- <=WM: (13315: R1 ^reward R954)
- <=WM: (13300: I3 ^see 1)
- <=WM: (13318: O1902 ^name predict-no)
- <=WM: (13317: O1901 ^name predict-yes)
- <=WM: (13316: R954 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1903 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1903 = 0.1844104702696336)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1904 = 0.5523833737960075)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1904 = 0.4476192676183378)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1902 = 0.4476192676183378)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1902 = 0.5523833737960075)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1901 = 0.1844104702696336)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1901 = 0.1398795999120246)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (13337: S1 ^operator O1904)
- 952: O: O1904 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N952 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N951 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13338: I3 ^predict-no N952)
- <=WM: (13324: N951 ^status complete)
- <=WM: (13323: I3 ^predict-no N951)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (13342: I2 ^dir U)
- =>WM: (13341: I2 ^reward 1)
- =>WM: (13340: I2 ^see 0)
- =>WM: (13339: N952 ^status complete)
- <=WM: (13327: I2 ^dir R)
- <=WM: (13326: I2 ^reward 1)
- <=WM: (13325: I2 ^see 0)
- =>WM: (13343: I2 ^level-1 R0-root)
- <=WM: (13328: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R956 ^value 1 +)
- (R1 ^reward R956 +)
- Firing propose*predict-yes
- -->
- (O1905 ^name predict-yes +)
- (S1 ^operator O1905 +)
- Firing propose*predict-no
- -->
- (O1906 ^name predict-no +)
- (S1 ^operator O1906 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1904 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1903 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1904 ^name predict-no +)
- (S1 ^operator O1904 +)
- Retracting propose*predict-yes
- -->
- (O1903 ^name predict-yes +)
- (S1 ^operator O1903 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R955 ^value 1 +)
- (R1 ^reward R955 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1904 = 0.4476192676183378)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1904 = 0.5523833737960075)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1903 = 0.1844104702696336)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1903 = 0.1398795999120246)
- =>WM: (13350: S1 ^operator O1906 +)
- =>WM: (13349: S1 ^operator O1905 +)
- =>WM: (13348: I3 ^dir U)
- =>WM: (13347: O1906 ^name predict-no)
- =>WM: (13346: O1905 ^name predict-yes)
- =>WM: (13345: R956 ^value 1)
- =>WM: (13344: R1 ^reward R956)
- <=WM: (13335: S1 ^operator O1903 +)
- <=WM: (13336: S1 ^operator O1904 +)
- <=WM: (13337: S1 ^operator O1904)
- <=WM: (13334: I3 ^dir R)
- <=WM: (13330: R1 ^reward R955)
- <=WM: (13333: O1904 ^name predict-no)
- <=WM: (13332: O1903 ^name predict-yes)
- <=WM: (13331: R955 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1905 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1906 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1904 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1903 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622532 -0.174914 0.447619(R,m,v=1,0.925,0.069958)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*39 0.377469 0.174914 0.552383 -> 0.377469 0.174914 0.552383(R,m,v=1,1,0)
- =>WM: (13351: S1 ^operator O1906)
- 953: O: O1906 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N953 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N952 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13352: I3 ^predict-no N953)
- <=WM: (13339: N952 ^status complete)
- <=WM: (13338: I3 ^predict-no N952)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (13356: I2 ^dir L)
- =>WM: (13355: I2 ^reward 1)
- =>WM: (13354: I2 ^see 0)
- =>WM: (13353: N953 ^status complete)
- <=WM: (13342: I2 ^dir U)
- <=WM: (13341: I2 ^reward 1)
- <=WM: (13340: I2 ^see 0)
- =>WM: (13357: I2 ^level-1 R0-root)
- <=WM: (13343: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1905 = 0.6104621686166466)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1906 = 0.1063475139796038)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R957 ^value 1 +)
- (R1 ^reward R957 +)
- Firing propose*predict-yes
- -->
- (O1907 ^name predict-yes +)
- (S1 ^operator O1907 +)
- Firing propose*predict-no
- -->
- (O1908 ^name predict-no +)
- (S1 ^operator O1908 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1906 = 0.3873365065796835)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1905 = 0.3895397770301633)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1906 ^name predict-no +)
- (S1 ^operator O1906 +)
- Retracting propose*predict-yes
- -->
- (O1905 ^name predict-yes +)
- (S1 ^operator O1905 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R956 ^value 1 +)
- (R1 ^reward R956 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1906 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1905 = 0.)
- =>WM: (13364: S1 ^operator O1908 +)
- =>WM: (13363: S1 ^operator O1907 +)
- =>WM: (13362: I3 ^dir L)
- =>WM: (13361: O1908 ^name predict-no)
- =>WM: (13360: O1907 ^name predict-yes)
- =>WM: (13359: R957 ^value 1)
- =>WM: (13358: R1 ^reward R957)
- <=WM: (13349: S1 ^operator O1905 +)
- <=WM: (13350: S1 ^operator O1906 +)
- <=WM: (13351: S1 ^operator O1906)
- <=WM: (13348: I3 ^dir U)
- <=WM: (13344: R1 ^reward R956)
- <=WM: (13347: O1906 ^name predict-no)
- <=WM: (13346: O1905 ^name predict-yes)
- <=WM: (13345: R956 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1907 = 0.6104621686166466)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1907 = 0.3895397770301633)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1908 = 0.1063475139796038)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1908 = 0.3873365065796835)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1906 = 0.3873365065796835)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1906 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1905 = 0.3895397770301633)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1905 = 0.6104621686166466)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (13365: S1 ^operator O1907)
- 954: O: O1907 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N954 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N953 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13366: I3 ^predict-yes N954)
- <=WM: (13353: N953 ^status complete)
- <=WM: (13352: I3 ^predict-no N953)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- |\--- Input Phase ---
- =>WM: (13370: I2 ^dir U)
- =>WM: (13369: I2 ^reward 1)
- =>WM: (13368: I2 ^see 1)
- =>WM: (13367: N954 ^status complete)
- <=WM: (13356: I2 ^dir L)
- <=WM: (13355: I2 ^reward 1)
- <=WM: (13354: I2 ^see 0)
- =>WM: (13371: I2 ^level-1 L1-root)
- <=WM: (13357: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R958 ^value 1 +)
- (R1 ^reward R958 +)
- Firing propose*predict-yes
- -->
- (O1909 ^name predict-yes +)
- (S1 ^operator O1909 +)
- Firing propose*predict-no
- -->
- (O1910 ^name predict-no +)
- (S1 ^operator O1910 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1908 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1907 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1908 ^name predict-no +)
- (S1 ^operator O1908 +)
- Retracting propose*predict-yes
- -->
- (O1907 ^name predict-yes +)
- (S1 ^operator O1907 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R957 ^value 1 +)
- (R1 ^reward R957 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1908 = 0.3873365065796835)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1908 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1907 = 0.3895397770301633)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1907 = 0.6104621686166466)
- =>WM: (13379: S1 ^operator O1910 +)
- =>WM: (13378: S1 ^operator O1909 +)
- =>WM: (13377: I3 ^dir U)
- =>WM: (13376: O1910 ^name predict-no)
- =>WM: (13375: O1909 ^name predict-yes)
- =>WM: (13374: R958 ^value 1)
- =>WM: (13373: R1 ^reward R958)
- =>WM: (13372: I3 ^see 1)
- <=WM: (13363: S1 ^operator O1907 +)
- <=WM: (13365: S1 ^operator O1907)
- <=WM: (13364: S1 ^operator O1908 +)
- <=WM: (13362: I3 ^dir L)
- <=WM: (13358: R1 ^reward R957)
- <=WM: (13329: I3 ^see 0)
- <=WM: (13361: O1908 ^name predict-no)
- <=WM: (13360: O1907 ^name predict-yes)
- <=WM: (13359: R957 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1909 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1910 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1908 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1907 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.389539(R,m,v=1,0.886792,0.101027)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*44 0.288049 0.322413 0.610462 -> 0.288049 0.322413 0.610462(R,m,v=1,1,0)
- =>WM: (13380: S1 ^operator O1910)
- 955: O: O1910 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N955 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N954 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13381: I3 ^predict-no N955)
- <=WM: (13367: N954 ^status complete)
- <=WM: (13366: I3 ^predict-yes N954)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- -/--- Input Phase ---
- =>WM: (13385: I2 ^dir U)
- =>WM: (13384: I2 ^reward 1)
- =>WM: (13383: I2 ^see 0)
- =>WM: (13382: N955 ^status complete)
- <=WM: (13370: I2 ^dir U)
- <=WM: (13369: I2 ^reward 1)
- <=WM: (13368: I2 ^see 1)
- =>WM: (13386: I2 ^level-1 L1-root)
- <=WM: (13371: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R959 ^value 1 +)
- (R1 ^reward R959 +)
- Firing propose*predict-yes
- -->
- (O1911 ^name predict-yes +)
- (S1 ^operator O1911 +)
- Firing propose*predict-no
- -->
- (O1912 ^name predict-no +)
- (S1 ^operator O1912 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1910 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1909 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1910 ^name predict-no +)
- (S1 ^operator O1910 +)
- Retracting propose*predict-yes
- -->
- (O1909 ^name predict-yes +)
- (S1 ^operator O1909 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R958 ^value 1 +)
- (R1 ^reward R958 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1910 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1909 = 0.)
- =>WM: (13393: S1 ^operator O1912 +)
- =>WM: (13392: S1 ^operator O1911 +)
- =>WM: (13391: O1912 ^name predict-no)
- =>WM: (13390: O1911 ^name predict-yes)
- =>WM: (13389: R959 ^value 1)
- =>WM: (13388: R1 ^reward R959)
- =>WM: (13387: I3 ^see 0)
- <=WM: (13378: S1 ^operator O1909 +)
- <=WM: (13379: S1 ^operator O1910 +)
- <=WM: (13380: S1 ^operator O1910)
- <=WM: (13373: R1 ^reward R958)
- <=WM: (13372: I3 ^see 1)
- <=WM: (13376: O1910 ^name predict-no)
- <=WM: (13375: O1909 ^name predict-yes)
- <=WM: (13374: R958 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1911 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1912 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1910 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1909 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (13394: S1 ^operator O1912)
- 956: O: O1912 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N956 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N955 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13395: I3 ^predict-no N956)
- <=WM: (13382: N955 ^status complete)
- <=WM: (13381: I3 ^predict-no N955)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (13399: I2 ^dir L)
- =>WM: (13398: I2 ^reward 1)
- =>WM: (13397: I2 ^see 0)
- =>WM: (13396: N956 ^status complete)
- <=WM: (13385: I2 ^dir U)
- <=WM: (13384: I2 ^reward 1)
- <=WM: (13383: I2 ^see 0)
- =>WM: (13400: I2 ^level-1 L1-root)
- <=WM: (13386: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1912 = 0.6126622914849755)
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1911 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R960 ^value 1 +)
- (R1 ^reward R960 +)
- Firing propose*predict-yes
- -->
- (O1913 ^name predict-yes +)
- (S1 ^operator O1913 +)
- Firing propose*predict-no
- -->
- (O1914 ^name predict-no +)
- (S1 ^operator O1914 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1912 = 0.3873365065796835)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1911 = 0.3895394851831418)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1912 ^name predict-no +)
- (S1 ^operator O1912 +)
- Retracting propose*predict-yes
- -->
- (O1911 ^name predict-yes +)
- (S1 ^operator O1911 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R959 ^value 1 +)
- (R1 ^reward R959 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1912 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1911 = 0.)
- =>WM: (13407: S1 ^operator O1914 +)
- =>WM: (13406: S1 ^operator O1913 +)
- =>WM: (13405: I3 ^dir L)
- =>WM: (13404: O1914 ^name predict-no)
- =>WM: (13403: O1913 ^name predict-yes)
- =>WM: (13402: R960 ^value 1)
- =>WM: (13401: R1 ^reward R960)
- <=WM: (13392: S1 ^operator O1911 +)
- <=WM: (13393: S1 ^operator O1912 +)
- <=WM: (13394: S1 ^operator O1912)
- <=WM: (13377: I3 ^dir U)
- <=WM: (13388: R1 ^reward R959)
- <=WM: (13391: O1912 ^name predict-no)
- <=WM: (13390: O1911 ^name predict-yes)
- <=WM: (13389: R959 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1913 = -0.02274740735326741)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1913 = 0.3895394851831418)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1914 = 0.6126622914849755)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1914 = 0.3873365065796835)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1912 = 0.3873365065796835)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1912 = 0.6126622914849755)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1911 = 0.3895394851831418)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1911 = -0.02274740735326741)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (13408: S1 ^operator O1914)
- 957: O: O1914 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N957 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N956 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13409: I3 ^predict-no N957)
- <=WM: (13396: N956 ^status complete)
- <=WM: (13395: I3 ^predict-no N956)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (13413: I2 ^dir U)
- =>WM: (13412: I2 ^reward 1)
- =>WM: (13411: I2 ^see 0)
- =>WM: (13410: N957 ^status complete)
- <=WM: (13399: I2 ^dir L)
- <=WM: (13398: I2 ^reward 1)
- <=WM: (13397: I2 ^see 0)
- =>WM: (13414: I2 ^level-1 L0-root)
- <=WM: (13400: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R961 ^value 1 +)
- (R1 ^reward R961 +)
- Firing propose*predict-yes
- -->
- (O1915 ^name predict-yes +)
- (S1 ^operator O1915 +)
- Firing propose*predict-no
- -->
- (O1916 ^name predict-no +)
- (S1 ^operator O1916 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1914 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1913 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1914 ^name predict-no +)
- (S1 ^operator O1914 +)
- Retracting propose*predict-yes
- -->
- (O1913 ^name predict-yes +)
- (S1 ^operator O1913 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R960 ^value 1 +)
- (R1 ^reward R960 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1914 = 0.3873365065796835)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1914 = 0.6126622914849755)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1913 = 0.3895394851831418)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1913 = -0.02274740735326741)
- =>WM: (13421: S1 ^operator O1916 +)
- =>WM: (13420: S1 ^operator O1915 +)
- =>WM: (13419: I3 ^dir U)
- =>WM: (13418: O1916 ^name predict-no)
- =>WM: (13417: O1915 ^name predict-yes)
- =>WM: (13416: R961 ^value 1)
- =>WM: (13415: R1 ^reward R961)
- <=WM: (13406: S1 ^operator O1913 +)
- <=WM: (13407: S1 ^operator O1914 +)
- <=WM: (13408: S1 ^operator O1914)
- <=WM: (13405: I3 ^dir L)
- <=WM: (13401: R1 ^reward R960)
- <=WM: (13404: O1914 ^name predict-no)
- <=WM: (13403: O1913 ^name predict-yes)
- <=WM: (13402: R960 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1915 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1916 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1914 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1913 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.71908 -0.331744 0.387337 -> 0.719081 -0.331744 0.387337(R,m,v=1,0.930233,0.0652795)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*32 0.280918 0.331744 0.612662 -> 0.280918 0.331744 0.612662(R,m,v=1,1,0)
- =>WM: (13422: S1 ^operator O1916)
- 958: O: O1916 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N958 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N957 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13423: I3 ^predict-no N958)
- <=WM: (13410: N957 ^status complete)
- <=WM: (13409: I3 ^predict-no N957)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- -/--- Input Phase ---
- =>WM: (13427: I2 ^dir R)
- =>WM: (13426: I2 ^reward 1)
- =>WM: (13425: I2 ^see 0)
- =>WM: (13424: N958 ^status complete)
- <=WM: (13413: I2 ^dir U)
- <=WM: (13412: I2 ^reward 1)
- <=WM: (13411: I2 ^see 0)
- =>WM: (13428: I2 ^level-1 L0-root)
- <=WM: (13414: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1915 = 0.8155985324859676)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1916 = -0.00558448899823713)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R962 ^value 1 +)
- (R1 ^reward R962 +)
- Firing propose*predict-yes
- -->
- (O1917 ^name predict-yes +)
- (S1 ^operator O1917 +)
- Firing propose*predict-no
- -->
- (O1918 ^name predict-no +)
- (S1 ^operator O1918 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1916 = 0.4476188714061859)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1915 = 0.1844104702696336)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1916 ^name predict-no +)
- (S1 ^operator O1916 +)
- Retracting propose*predict-yes
- -->
- (O1915 ^name predict-yes +)
- (S1 ^operator O1915 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R961 ^value 1 +)
- (R1 ^reward R961 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1916 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1915 = 0.)
- =>WM: (13435: S1 ^operator O1918 +)
- =>WM: (13434: S1 ^operator O1917 +)
- =>WM: (13433: I3 ^dir R)
- =>WM: (13432: O1918 ^name predict-no)
- =>WM: (13431: O1917 ^name predict-yes)
- =>WM: (13430: R962 ^value 1)
- =>WM: (13429: R1 ^reward R962)
- <=WM: (13420: S1 ^operator O1915 +)
- <=WM: (13421: S1 ^operator O1916 +)
- <=WM: (13422: S1 ^operator O1916)
- <=WM: (13419: I3 ^dir U)
- <=WM: (13415: R1 ^reward R961)
- <=WM: (13418: O1916 ^name predict-no)
- <=WM: (13417: O1915 ^name predict-yes)
- <=WM: (13416: R961 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1917 = 0.8155985324859676)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1917 = 0.1844104702696336)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1918 = -0.00558448899823713)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1918 = 0.4476188714061859)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1916 = 0.4476188714061859)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1916 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1915 = 0.1844104702696336)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1915 = 0.8155985324859676)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (13436: S1 ^operator O1917)
- 959: O: O1917 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N959 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N958 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13437: I3 ^predict-yes N959)
- <=WM: (13424: N958 ^status complete)
- <=WM: (13423: I3 ^predict-no N958)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (13441: I2 ^dir L)
- =>WM: (13440: I2 ^reward 1)
- =>WM: (13439: I2 ^see 1)
- =>WM: (13438: N959 ^status complete)
- <=WM: (13427: I2 ^dir R)
- <=WM: (13426: I2 ^reward 1)
- <=WM: (13425: I2 ^see 0)
- =>WM: (13442: I2 ^level-1 R1-root)
- <=WM: (13428: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O1917 = 0.6104587229728515)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O1918 = 0.2714993082286609)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R963 ^value 1 +)
- (R1 ^reward R963 +)
- Firing propose*predict-yes
- -->
- (O1919 ^name predict-yes +)
- (S1 ^operator O1919 +)
- Firing propose*predict-no
- -->
- (O1920 ^name predict-no +)
- (S1 ^operator O1920 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1918 = 0.3873366868699847)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1917 = 0.3895394851831418)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1918 ^name predict-no +)
- (S1 ^operator O1918 +)
- Retracting propose*predict-yes
- -->
- (O1917 ^name predict-yes +)
- (S1 ^operator O1917 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R962 ^value 1 +)
- (R1 ^reward R962 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1918 = 0.4476188714061859)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1918 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1917 = 0.1844104702696336)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1917 = 0.8155985324859676)
- =>WM: (13450: S1 ^operator O1920 +)
- =>WM: (13449: S1 ^operator O1919 +)
- =>WM: (13448: I3 ^dir L)
- =>WM: (13447: O1920 ^name predict-no)
- =>WM: (13446: O1919 ^name predict-yes)
- =>WM: (13445: R963 ^value 1)
- =>WM: (13444: R1 ^reward R963)
- =>WM: (13443: I3 ^see 1)
- <=WM: (13434: S1 ^operator O1917 +)
- <=WM: (13436: S1 ^operator O1917)
- <=WM: (13435: S1 ^operator O1918 +)
- <=WM: (13433: I3 ^dir R)
- <=WM: (13429: R1 ^reward R962)
- <=WM: (13387: I3 ^see 0)
- <=WM: (13432: O1918 ^name predict-no)
- <=WM: (13431: O1917 ^name predict-yes)
- <=WM: (13430: R962 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1919 = 0.3895394851831418)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O1919 = 0.6104587229728515)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1920 = 0.3873366868699847)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O1920 = 0.2714993082286609)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1918 = 0.3873366868699847)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O1918 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1917 = 0.3895394851831418)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O1917 = 0.6104587229728515)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675413 -0.491002 0.18441 -> 0.675411 -0.491002 0.184409(R,m,v=1,0.895062,0.0945096)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*34 0.324599 0.491 0.815599 -> 0.324597 0.491 0.815597(R,m,v=1,1,0)
- =>WM: (13451: S1 ^operator O1919)
- 960: O: O1919 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N960 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N959 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13452: I3 ^predict-yes N960)
- <=WM: (13438: N959 ^status complete)
- <=WM: (13437: I3 ^predict-yes N959)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (13456: I2 ^dir L)
- =>WM: (13455: I2 ^reward 1)
- =>WM: (13454: I2 ^see 1)
- =>WM: (13453: N960 ^status complete)
- <=WM: (13441: I2 ^dir L)
- <=WM: (13440: I2 ^reward 1)
- <=WM: (13439: I2 ^see 1)
- =>WM: (13457: I2 ^level-1 L1-root)
- <=WM: (13442: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1920 = 0.6126624717752767)
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1919 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R964 ^value 1 +)
- (R1 ^reward R964 +)
- Firing propose*predict-yes
- -->
- (O1921 ^name predict-yes +)
- (S1 ^operator O1921 +)
- Firing propose*predict-no
- -->
- (O1922 ^name predict-no +)
- (S1 ^operator O1922 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1920 = 0.3873366868699847)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1919 = 0.3895394851831418)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1920 ^name predict-no +)
- (S1 ^operator O1920 +)
- Retracting propose*predict-yes
- -->
- (O1919 ^name predict-yes +)
- (S1 ^operator O1919 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R963 ^value 1 +)
- (R1 ^reward R963 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O1920 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1920 = 0.3873366868699847)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O1919 = 0.6104587229728515)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1919 = 0.3895394851831418)
- =>WM: (13463: S1 ^operator O1922 +)
- =>WM: (13462: S1 ^operator O1921 +)
- =>WM: (13461: O1922 ^name predict-no)
- =>WM: (13460: O1921 ^name predict-yes)
- =>WM: (13459: R964 ^value 1)
- =>WM: (13458: R1 ^reward R964)
- <=WM: (13449: S1 ^operator O1919 +)
- <=WM: (13451: S1 ^operator O1919)
- <=WM: (13450: S1 ^operator O1920 +)
- <=WM: (13444: R1 ^reward R963)
- <=WM: (13447: O1920 ^name predict-no)
- <=WM: (13446: O1919 ^name predict-yes)
- <=WM: (13445: R963 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1921 = 0.3895394851831418)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1921 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1922 = 0.3873366868699847)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1922 = 0.6126624717752767)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1920 = 0.3873366868699847)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1920 = 0.6126624717752767)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1919 = 0.3895394851831418)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1919 = -0.02274740735326741)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.389539 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.8875,0.100472)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*36 0.288049 0.32241 0.610459 -> 0.288049 0.32241 0.610459(R,m,v=1,1,0)
- =>WM: (13464: S1 ^operator O1922)
- 961: O: O1922 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N961 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N960 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13465: I3 ^predict-no N961)
- <=WM: (13453: N960 ^status complete)
- <=WM: (13452: I3 ^predict-yes N960)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- ---- Input Phase ---
- =>WM: (13469: I2 ^dir U)
- =>WM: (13468: I2 ^reward 1)
- =>WM: (13467: I2 ^see 0)
- =>WM: (13466: N961 ^status complete)
- <=WM: (13456: I2 ^dir L)
- <=WM: (13455: I2 ^reward 1)
- <=WM: (13454: I2 ^see 1)
- =>WM: (13470: I2 ^level-1 L0-root)
- <=WM: (13457: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R965 ^value 1 +)
- (R1 ^reward R965 +)
- Firing propose*predict-yes
- -->
- (O1923 ^name predict-yes +)
- (S1 ^operator O1923 +)
- Firing propose*predict-no
- -->
- (O1924 ^name predict-no +)
- (S1 ^operator O1924 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1922 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1921 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1922 ^name predict-no +)
- (S1 ^operator O1922 +)
- Retracting propose*predict-yes
- -->
- (O1921 ^name predict-yes +)
- (S1 ^operator O1921 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R964 ^value 1 +)
- (R1 ^reward R964 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1922 = 0.6126624717752767)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1922 = 0.3873366868699847)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1921 = -0.02274740735326741)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1921 = 0.3895397539597428)
- =>WM: (13478: S1 ^operator O1924 +)
- =>WM: (13477: S1 ^operator O1923 +)
- =>WM: (13476: I3 ^dir U)
- =>WM: (13475: O1924 ^name predict-no)
- =>WM: (13474: O1923 ^name predict-yes)
- =>WM: (13473: R965 ^value 1)
- =>WM: (13472: R1 ^reward R965)
- =>WM: (13471: I3 ^see 0)
- <=WM: (13462: S1 ^operator O1921 +)
- <=WM: (13463: S1 ^operator O1922 +)
- <=WM: (13464: S1 ^operator O1922)
- <=WM: (13448: I3 ^dir L)
- <=WM: (13458: R1 ^reward R964)
- <=WM: (13443: I3 ^see 1)
- <=WM: (13461: O1922 ^name predict-no)
- <=WM: (13460: O1921 ^name predict-yes)
- <=WM: (13459: R964 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1923 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1924 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1922 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1921 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.719081 -0.331744 0.387337 -> 0.719081 -0.331744 0.387337(R,m,v=1,0.930636,0.0649281)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*32 0.280918 0.331744 0.612662 -> 0.280918 0.331744 0.612663(R,m,v=1,1,0)
- =>WM: (13479: S1 ^operator O1924)
- 962: O: O1924 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N962 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N961 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13480: I3 ^predict-no N962)
- <=WM: (13466: N961 ^status complete)
- <=WM: (13465: I3 ^predict-no N961)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- /|\---- Input Phase ---
- =>WM: (13484: I2 ^dir R)
- =>WM: (13483: I2 ^reward 1)
- =>WM: (13482: I2 ^see 0)
- =>WM: (13481: N962 ^status complete)
- <=WM: (13469: I2 ^dir U)
- <=WM: (13468: I2 ^reward 1)
- <=WM: (13467: I2 ^see 0)
- =>WM: (13485: I2 ^level-1 L0-root)
- <=WM: (13470: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1923 = 0.8155971820726273)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1924 = -0.00558448899823713)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R966 ^value 1 +)
- (R1 ^reward R966 +)
- Firing propose*predict-yes
- -->
- (O1925 ^name predict-yes +)
- (S1 ^operator O1925 +)
- Firing propose*predict-no
- -->
- (O1926 ^name predict-no +)
- (S1 ^operator O1926 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1924 = 0.4476188714061859)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1923 = 0.1844091198562935)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1924 ^name predict-no +)
- (S1 ^operator O1924 +)
- Retracting propose*predict-yes
- -->
- (O1923 ^name predict-yes +)
- (S1 ^operator O1923 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R965 ^value 1 +)
- (R1 ^reward R965 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1924 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1923 = 0.)
- =>WM: (13492: S1 ^operator O1926 +)
- =>WM: (13491: S1 ^operator O1925 +)
- =>WM: (13490: I3 ^dir R)
- =>WM: (13489: O1926 ^name predict-no)
- =>WM: (13488: O1925 ^name predict-yes)
- =>WM: (13487: R966 ^value 1)
- =>WM: (13486: R1 ^reward R966)
- <=WM: (13477: S1 ^operator O1923 +)
- <=WM: (13478: S1 ^operator O1924 +)
- <=WM: (13479: S1 ^operator O1924)
- <=WM: (13476: I3 ^dir U)
- <=WM: (13472: R1 ^reward R965)
- <=WM: (13475: O1924 ^name predict-no)
- <=WM: (13474: O1923 ^name predict-yes)
- <=WM: (13473: R965 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1925 = 0.8155971820726273)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1925 = 0.1844091198562935)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1926 = -0.00558448899823713)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1926 = 0.4476188714061859)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1924 = 0.4476188714061859)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1924 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1923 = 0.1844091198562935)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1923 = 0.8155971820726273)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (13493: S1 ^operator O1925)
- 963: O: O1925 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N963 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N962 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13494: I3 ^predict-yes N963)
- <=WM: (13481: N962 ^status complete)
- <=WM: (13480: I3 ^predict-no N962)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (13498: I2 ^dir R)
- =>WM: (13497: I2 ^reward 1)
- =>WM: (13496: I2 ^see 1)
- =>WM: (13495: N963 ^status complete)
- <=WM: (13484: I2 ^dir R)
- <=WM: (13483: I2 ^reward 1)
- <=WM: (13482: I2 ^see 0)
- =>WM: (13499: I2 ^level-1 R1-root)
- <=WM: (13485: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1925 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1926 = 0.5523829775838558)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R967 ^value 1 +)
- (R1 ^reward R967 +)
- Firing propose*predict-yes
- -->
- (O1927 ^name predict-yes +)
- (S1 ^operator O1927 +)
- Firing propose*predict-no
- -->
- (O1928 ^name predict-no +)
- (S1 ^operator O1928 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1926 = 0.4476188714061859)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1925 = 0.1844091198562935)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1926 ^name predict-no +)
- (S1 ^operator O1926 +)
- Retracting propose*predict-yes
- -->
- (O1925 ^name predict-yes +)
- (S1 ^operator O1925 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R966 ^value 1 +)
- (R1 ^reward R966 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1926 = 0.4476188714061859)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1926 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1925 = 0.1844091198562935)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1925 = 0.8155971820726273)
- =>WM: (13506: S1 ^operator O1928 +)
- =>WM: (13505: S1 ^operator O1927 +)
- =>WM: (13504: O1928 ^name predict-no)
- =>WM: (13503: O1927 ^name predict-yes)
- =>WM: (13502: R967 ^value 1)
- =>WM: (13501: R1 ^reward R967)
- =>WM: (13500: I3 ^see 1)
- <=WM: (13491: S1 ^operator O1925 +)
- <=WM: (13493: S1 ^operator O1925)
- <=WM: (13492: S1 ^operator O1926 +)
- <=WM: (13486: R1 ^reward R966)
- <=WM: (13471: I3 ^see 0)
- <=WM: (13489: O1926 ^name predict-no)
- <=WM: (13488: O1925 ^name predict-yes)
- <=WM: (13487: R966 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1927 = 0.1844091198562935)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1927 = 0.1398795999120246)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1928 = 0.4476188714061859)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1928 = 0.5523829775838558)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1926 = 0.4476188714061859)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1926 = 0.5523829775838558)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1925 = 0.1844091198562935)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1925 = 0.1398795999120246)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675411 -0.491002 0.184409 -> 0.67541 -0.491002 0.184408(R,m,v=1,0.895706,0.0939938)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*34 0.324597 0.491 0.815597 -> 0.324596 0.491 0.815596(R,m,v=1,1,0)
- =>WM: (13507: S1 ^operator O1928)
- 964: O: O1928 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N964 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N963 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13508: I3 ^predict-no N964)
- <=WM: (13495: N963 ^status complete)
- <=WM: (13494: I3 ^predict-yes N963)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (13512: I2 ^dir L)
- =>WM: (13511: I2 ^reward 1)
- =>WM: (13510: I2 ^see 0)
- =>WM: (13509: N964 ^status complete)
- <=WM: (13498: I2 ^dir R)
- <=WM: (13497: I2 ^reward 1)
- <=WM: (13496: I2 ^see 1)
- =>WM: (13513: I2 ^level-1 R0-root)
- <=WM: (13499: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1927 = 0.6104618767696252)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1928 = 0.1063475139796038)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R968 ^value 1 +)
- (R1 ^reward R968 +)
- Firing propose*predict-yes
- -->
- (O1929 ^name predict-yes +)
- (S1 ^operator O1929 +)
- Firing propose*predict-no
- -->
- (O1930 ^name predict-no +)
- (S1 ^operator O1930 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1928 = 0.3873368130731955)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1927 = 0.3895397539597428)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1928 ^name predict-no +)
- (S1 ^operator O1928 +)
- Retracting propose*predict-yes
- -->
- (O1927 ^name predict-yes +)
- (S1 ^operator O1927 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R967 ^value 1 +)
- (R1 ^reward R967 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1928 = 0.5523829775838558)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1928 = 0.4476188714061859)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1927 = 0.1398795999120246)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1927 = 0.1844081745669553)
- =>WM: (13521: S1 ^operator O1930 +)
- =>WM: (13520: S1 ^operator O1929 +)
- =>WM: (13519: I3 ^dir L)
- =>WM: (13518: O1930 ^name predict-no)
- =>WM: (13517: O1929 ^name predict-yes)
- =>WM: (13516: R968 ^value 1)
- =>WM: (13515: R1 ^reward R968)
- =>WM: (13514: I3 ^see 0)
- <=WM: (13505: S1 ^operator O1927 +)
- <=WM: (13506: S1 ^operator O1928 +)
- <=WM: (13507: S1 ^operator O1928)
- <=WM: (13490: I3 ^dir R)
- <=WM: (13501: R1 ^reward R967)
- <=WM: (13500: I3 ^see 1)
- <=WM: (13504: O1928 ^name predict-no)
- <=WM: (13503: O1927 ^name predict-yes)
- <=WM: (13502: R967 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1929 = 0.3895397539597428)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1929 = 0.6104618767696252)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1930 = 0.3873368130731955)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1930 = 0.1063475139796038)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1928 = 0.3873368130731955)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1928 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1927 = 0.3895397539597428)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1927 = 0.6104618767696252)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622532 -0.174914 0.447619 -> 0.622532 -0.174914 0.447619(R,m,v=1,0.92562,0.0694215)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*39 0.377469 0.174914 0.552383 -> 0.377469 0.174914 0.552383(R,m,v=1,1,0)
- =>WM: (13522: S1 ^operator O1929)
- 965: O: O1929 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N965 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N964 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13523: I3 ^predict-yes N965)
- <=WM: (13509: N964 ^status complete)
- <=WM: (13508: I3 ^predict-no N964)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- \---- Input Phase ---
- =>WM: (13527: I2 ^dir L)
- =>WM: (13526: I2 ^reward 1)
- =>WM: (13525: I2 ^see 1)
- =>WM: (13524: N965 ^status complete)
- <=WM: (13512: I2 ^dir L)
- <=WM: (13511: I2 ^reward 1)
- <=WM: (13510: I2 ^see 0)
- =>WM: (13528: I2 ^level-1 L1-root)
- <=WM: (13513: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1930 = 0.6126625979784875)
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1929 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R969 ^value 1 +)
- (R1 ^reward R969 +)
- Firing propose*predict-yes
- -->
- (O1931 ^name predict-yes +)
- (S1 ^operator O1931 +)
- Firing propose*predict-no
- -->
- (O1932 ^name predict-no +)
- (S1 ^operator O1932 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1930 = 0.3873368130731955)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1929 = 0.3895397539597428)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1930 ^name predict-no +)
- (S1 ^operator O1930 +)
- Retracting propose*predict-yes
- -->
- (O1929 ^name predict-yes +)
- (S1 ^operator O1929 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R968 ^value 1 +)
- (R1 ^reward R968 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1930 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1930 = 0.3873368130731955)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1929 = 0.6104618767696252)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1929 = 0.3895397539597428)
- =>WM: (13535: S1 ^operator O1932 +)
- =>WM: (13534: S1 ^operator O1931 +)
- =>WM: (13533: O1932 ^name predict-no)
- =>WM: (13532: O1931 ^name predict-yes)
- =>WM: (13531: R969 ^value 1)
- =>WM: (13530: R1 ^reward R969)
- =>WM: (13529: I3 ^see 1)
- <=WM: (13520: S1 ^operator O1929 +)
- <=WM: (13522: S1 ^operator O1929)
- <=WM: (13521: S1 ^operator O1930 +)
- <=WM: (13515: R1 ^reward R968)
- <=WM: (13514: I3 ^see 0)
- <=WM: (13518: O1930 ^name predict-no)
- <=WM: (13517: O1929 ^name predict-yes)
- <=WM: (13516: R968 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1931 = 0.3895397539597428)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1931 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1932 = 0.3873368130731955)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1932 = 0.6126625979784875)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1930 = 0.3873368130731955)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1930 = 0.6126625979784875)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1929 = 0.3895397539597428)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1929 = -0.02274740735326741)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.888199,0.0999224)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*44 0.288049 0.322413 0.610462 -> 0.288049 0.322413 0.610462(R,m,v=1,1,0)
- =>WM: (13536: S1 ^operator O1932)
- 966: O: O1932 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N966 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N965 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13537: I3 ^predict-no N966)
- <=WM: (13524: N965 ^status complete)
- <=WM: (13523: I3 ^predict-yes N965)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (13541: I2 ^dir R)
- =>WM: (13540: I2 ^reward 1)
- =>WM: (13539: I2 ^see 0)
- =>WM: (13538: N966 ^status complete)
- <=WM: (13527: I2 ^dir L)
- <=WM: (13526: I2 ^reward 1)
- <=WM: (13525: I2 ^see 1)
- =>WM: (13542: I2 ^level-1 L0-root)
- <=WM: (13528: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1931 = 0.8155962367832892)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1932 = -0.00558448899823713)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R970 ^value 1 +)
- (R1 ^reward R970 +)
- Firing propose*predict-yes
- -->
- (O1933 ^name predict-yes +)
- (S1 ^operator O1933 +)
- Firing propose*predict-no
- -->
- (O1934 ^name predict-no +)
- (S1 ^operator O1934 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1932 = 0.4476185940576797)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1931 = 0.1844081745669553)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1932 ^name predict-no +)
- (S1 ^operator O1932 +)
- Retracting propose*predict-yes
- -->
- (O1931 ^name predict-yes +)
- (S1 ^operator O1931 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R969 ^value 1 +)
- (R1 ^reward R969 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1932 = 0.6126625979784875)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1932 = 0.3873368130731955)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1931 = -0.02274740735326741)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1931 = 0.3895395093503376)
- =>WM: (13550: S1 ^operator O1934 +)
- =>WM: (13549: S1 ^operator O1933 +)
- =>WM: (13548: I3 ^dir R)
- =>WM: (13547: O1934 ^name predict-no)
- =>WM: (13546: O1933 ^name predict-yes)
- =>WM: (13545: R970 ^value 1)
- =>WM: (13544: R1 ^reward R970)
- =>WM: (13543: I3 ^see 0)
- <=WM: (13534: S1 ^operator O1931 +)
- <=WM: (13535: S1 ^operator O1932 +)
- <=WM: (13536: S1 ^operator O1932)
- <=WM: (13519: I3 ^dir L)
- <=WM: (13530: R1 ^reward R969)
- <=WM: (13529: I3 ^see 1)
- <=WM: (13533: O1932 ^name predict-no)
- <=WM: (13532: O1931 ^name predict-yes)
- <=WM: (13531: R969 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1933 = 0.1844081745669553)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1933 = 0.8155962367832892)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1934 = 0.4476185940576797)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1934 = -0.00558448899823713)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1932 = 0.4476185940576797)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1932 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1931 = 0.1844081745669553)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1931 = 0.8155962367832892)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.719081 -0.331744 0.387337 -> 0.719081 -0.331744 0.387337(R,m,v=1,0.931034,0.0645804)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*32 0.280918 0.331744 0.612663 -> 0.280919 0.331744 0.612663(R,m,v=1,1,0)
- =>WM: (13551: S1 ^operator O1933)
- 967: O: O1933 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N967 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N966 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13552: I3 ^predict-yes N967)
- <=WM: (13538: N966 ^status complete)
- <=WM: (13537: I3 ^predict-no N966)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (13556: I2 ^dir R)
- =>WM: (13555: I2 ^reward 1)
- =>WM: (13554: I2 ^see 1)
- =>WM: (13553: N967 ^status complete)
- <=WM: (13541: I2 ^dir R)
- <=WM: (13540: I2 ^reward 1)
- <=WM: (13539: I2 ^see 0)
- =>WM: (13557: I2 ^level-1 R1-root)
- <=WM: (13542: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1933 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1934 = 0.5523827002353495)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R971 ^value 1 +)
- (R1 ^reward R971 +)
- Firing propose*predict-yes
- -->
- (O1935 ^name predict-yes +)
- (S1 ^operator O1935 +)
- Firing propose*predict-no
- -->
- (O1936 ^name predict-no +)
- (S1 ^operator O1936 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1934 = 0.4476185940576797)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1933 = 0.1844081745669553)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1934 ^name predict-no +)
- (S1 ^operator O1934 +)
- Retracting propose*predict-yes
- -->
- (O1933 ^name predict-yes +)
- (S1 ^operator O1933 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R970 ^value 1 +)
- (R1 ^reward R970 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1934 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1934 = 0.4476185940576797)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1933 = 0.8155962367832892)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1933 = 0.1844081745669553)
- =>WM: (13564: S1 ^operator O1936 +)
- =>WM: (13563: S1 ^operator O1935 +)
- =>WM: (13562: O1936 ^name predict-no)
- =>WM: (13561: O1935 ^name predict-yes)
- =>WM: (13560: R971 ^value 1)
- =>WM: (13559: R1 ^reward R971)
- =>WM: (13558: I3 ^see 1)
- <=WM: (13549: S1 ^operator O1933 +)
- <=WM: (13551: S1 ^operator O1933)
- <=WM: (13550: S1 ^operator O1934 +)
- <=WM: (13544: R1 ^reward R970)
- <=WM: (13543: I3 ^see 0)
- <=WM: (13547: O1934 ^name predict-no)
- <=WM: (13546: O1933 ^name predict-yes)
- <=WM: (13545: R970 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1935 = 0.1844081745669553)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1935 = 0.1398795999120246)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1936 = 0.4476185940576797)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1936 = 0.5523827002353495)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1934 = 0.4476185940576797)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1934 = 0.5523827002353495)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1933 = 0.1844081745669553)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1933 = 0.1398795999120246)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.67541 -0.491002 0.184408 -> 0.675409 -0.491002 0.184408(R,m,v=1,0.896341,0.0934835)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*34 0.324596 0.491 0.815596 -> 0.324595 0.491001 0.815596(R,m,v=1,1,0)
- =>WM: (13565: S1 ^operator O1936)
- 968: O: O1936 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N968 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N967 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13566: I3 ^predict-no N968)
- <=WM: (13553: N967 ^status complete)
- <=WM: (13552: I3 ^predict-yes N967)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (13570: I2 ^dir U)
- =>WM: (13569: I2 ^reward 1)
- =>WM: (13568: I2 ^see 0)
- =>WM: (13567: N968 ^status complete)
- <=WM: (13556: I2 ^dir R)
- <=WM: (13555: I2 ^reward 1)
- <=WM: (13554: I2 ^see 1)
- =>WM: (13571: I2 ^level-1 R0-root)
- <=WM: (13557: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R972 ^value 1 +)
- (R1 ^reward R972 +)
- Firing propose*predict-yes
- -->
- (O1937 ^name predict-yes +)
- (S1 ^operator O1937 +)
- Firing propose*predict-no
- -->
- (O1938 ^name predict-no +)
- (S1 ^operator O1938 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1936 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1935 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1936 ^name predict-no +)
- (S1 ^operator O1936 +)
- Retracting propose*predict-yes
- -->
- (O1935 ^name predict-yes +)
- (S1 ^operator O1935 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R971 ^value 1 +)
- (R1 ^reward R971 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1936 = 0.5523827002353495)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1936 = 0.4476185940576797)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1935 = 0.1398795999120246)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1935 = 0.1844075128644186)
- =>WM: (13579: S1 ^operator O1938 +)
- =>WM: (13578: S1 ^operator O1937 +)
- =>WM: (13577: I3 ^dir U)
- =>WM: (13576: O1938 ^name predict-no)
- =>WM: (13575: O1937 ^name predict-yes)
- =>WM: (13574: R972 ^value 1)
- =>WM: (13573: R1 ^reward R972)
- =>WM: (13572: I3 ^see 0)
- <=WM: (13563: S1 ^operator O1935 +)
- <=WM: (13564: S1 ^operator O1936 +)
- <=WM: (13565: S1 ^operator O1936)
- <=WM: (13548: I3 ^dir R)
- <=WM: (13559: R1 ^reward R971)
- <=WM: (13558: I3 ^see 1)
- <=WM: (13562: O1936 ^name predict-no)
- <=WM: (13561: O1935 ^name predict-yes)
- <=WM: (13560: R971 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1937 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1938 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1936 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1935 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622532 -0.174914 0.447619 -> 0.622532 -0.174914 0.447618(R,m,v=1,0.92623,0.0688931)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*39 0.377469 0.174914 0.552383 -> 0.377469 0.174914 0.552383(R,m,v=1,1,0)
- =>WM: (13580: S1 ^operator O1938)
- 969: O: O1938 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N969 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N968 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13581: I3 ^predict-no N969)
- <=WM: (13567: N968 ^status complete)
- <=WM: (13566: I3 ^predict-no N968)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (13585: I2 ^dir R)
- =>WM: (13584: I2 ^reward 1)
- =>WM: (13583: I2 ^see 0)
- =>WM: (13582: N969 ^status complete)
- <=WM: (13570: I2 ^dir U)
- <=WM: (13569: I2 ^reward 1)
- <=WM: (13568: I2 ^see 0)
- =>WM: (13586: I2 ^level-1 R0-root)
- <=WM: (13571: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O1937 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O1938 = 0.5523777234651187)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R973 ^value 1 +)
- (R1 ^reward R973 +)
- Firing propose*predict-yes
- -->
- (O1939 ^name predict-yes +)
- (S1 ^operator O1939 +)
- Firing propose*predict-no
- -->
- (O1940 ^name predict-no +)
- (S1 ^operator O1940 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1938 = 0.4476183999137253)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1937 = 0.1844075128644186)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1938 ^name predict-no +)
- (S1 ^operator O1938 +)
- Retracting propose*predict-yes
- -->
- (O1937 ^name predict-yes +)
- (S1 ^operator O1937 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R972 ^value 1 +)
- (R1 ^reward R972 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1938 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1937 = 0.)
- =>WM: (13593: S1 ^operator O1940 +)
- =>WM: (13592: S1 ^operator O1939 +)
- =>WM: (13591: I3 ^dir R)
- =>WM: (13590: O1940 ^name predict-no)
- =>WM: (13589: O1939 ^name predict-yes)
- =>WM: (13588: R973 ^value 1)
- =>WM: (13587: R1 ^reward R973)
- <=WM: (13578: S1 ^operator O1937 +)
- <=WM: (13579: S1 ^operator O1938 +)
- <=WM: (13580: S1 ^operator O1938)
- <=WM: (13577: I3 ^dir U)
- <=WM: (13573: R1 ^reward R972)
- <=WM: (13576: O1938 ^name predict-no)
- <=WM: (13575: O1937 ^name predict-yes)
- <=WM: (13574: R972 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O1939 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1939 = 0.1844075128644186)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O1940 = 0.5523777234651187)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1940 = 0.4476183999137253)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1938 = 0.4476183999137253)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O1938 = 0.5523777234651187)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1937 = 0.1844075128644186)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O1937 = 0.1664311307472832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (13594: S1 ^operator O1940)
- 970: O: O1940 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N970 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N969 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13595: I3 ^predict-no N970)
- <=WM: (13582: N969 ^status complete)
- <=WM: (13581: I3 ^predict-no N969)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (13599: I2 ^dir L)
- =>WM: (13598: I2 ^reward 1)
- =>WM: (13597: I2 ^see 0)
- =>WM: (13596: N970 ^status complete)
- <=WM: (13585: I2 ^dir R)
- <=WM: (13584: I2 ^reward 1)
- <=WM: (13583: I2 ^see 0)
- =>WM: (13600: I2 ^level-1 R0-root)
- <=WM: (13586: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1939 = 0.61046163216022)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1940 = 0.1063475139796038)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R974 ^value 1 +)
- (R1 ^reward R974 +)
- Firing propose*predict-yes
- -->
- (O1941 ^name predict-yes +)
- (S1 ^operator O1941 +)
- Firing propose*predict-no
- -->
- (O1942 ^name predict-no +)
- (S1 ^operator O1942 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1940 = 0.387336901415443)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1939 = 0.3895395093503376)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1940 ^name predict-no +)
- (S1 ^operator O1940 +)
- Retracting propose*predict-yes
- -->
- (O1939 ^name predict-yes +)
- (S1 ^operator O1939 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R973 ^value 1 +)
- (R1 ^reward R973 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1940 = 0.4476183999137253)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O1940 = 0.5523777234651187)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1939 = 0.1844075128644186)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O1939 = 0.1664311307472832)
- =>WM: (13607: S1 ^operator O1942 +)
- =>WM: (13606: S1 ^operator O1941 +)
- =>WM: (13605: I3 ^dir L)
- =>WM: (13604: O1942 ^name predict-no)
- =>WM: (13603: O1941 ^name predict-yes)
- =>WM: (13602: R974 ^value 1)
- =>WM: (13601: R1 ^reward R974)
- <=WM: (13592: S1 ^operator O1939 +)
- <=WM: (13593: S1 ^operator O1940 +)
- <=WM: (13594: S1 ^operator O1940)
- <=WM: (13591: I3 ^dir R)
- <=WM: (13587: R1 ^reward R973)
- <=WM: (13590: O1940 ^name predict-no)
- <=WM: (13589: O1939 ^name predict-yes)
- <=WM: (13588: R973 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1941 = 0.61046163216022)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1941 = 0.3895395093503376)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1942 = 0.1063475139796038)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1942 = 0.387336901415443)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1940 = 0.387336901415443)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1940 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1939 = 0.3895395093503376)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1939 = 0.61046163216022)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622532 -0.174914 0.447618 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.926829,0.0683727)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*41 0.377465 0.174913 0.552378 -> 0.377465 0.174913 0.552378(R,m,v=1,1,0)
- =>WM: (13608: S1 ^operator O1941)
- 971: O: O1941 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N971 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N970 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13609: I3 ^predict-yes N971)
- <=WM: (13596: N970 ^status complete)
- <=WM: (13595: I3 ^predict-no N970)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- ---- Input Phase ---
- =>WM: (13613: I2 ^dir U)
- =>WM: (13612: I2 ^reward 1)
- =>WM: (13611: I2 ^see 1)
- =>WM: (13610: N971 ^status complete)
- <=WM: (13599: I2 ^dir L)
- <=WM: (13598: I2 ^reward 1)
- <=WM: (13597: I2 ^see 0)
- =>WM: (13614: I2 ^level-1 L1-root)
- <=WM: (13600: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R975 ^value 1 +)
- (R1 ^reward R975 +)
- Firing propose*predict-yes
- -->
- (O1943 ^name predict-yes +)
- (S1 ^operator O1943 +)
- Firing propose*predict-no
- -->
- (O1944 ^name predict-no +)
- (S1 ^operator O1944 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1942 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1941 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1942 ^name predict-no +)
- (S1 ^operator O1942 +)
- Retracting propose*predict-yes
- -->
- (O1941 ^name predict-yes +)
- (S1 ^operator O1941 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R974 ^value 1 +)
- (R1 ^reward R974 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1942 = 0.387336901415443)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1942 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1941 = 0.3895395093503376)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1941 = 0.61046163216022)
- =>WM: (13622: S1 ^operator O1944 +)
- =>WM: (13621: S1 ^operator O1943 +)
- =>WM: (13620: I3 ^dir U)
- =>WM: (13619: O1944 ^name predict-no)
- =>WM: (13618: O1943 ^name predict-yes)
- =>WM: (13617: R975 ^value 1)
- =>WM: (13616: R1 ^reward R975)
- =>WM: (13615: I3 ^see 1)
- <=WM: (13606: S1 ^operator O1941 +)
- <=WM: (13608: S1 ^operator O1941)
- <=WM: (13607: S1 ^operator O1942 +)
- <=WM: (13605: I3 ^dir L)
- <=WM: (13601: R1 ^reward R974)
- <=WM: (13572: I3 ^see 0)
- <=WM: (13604: O1942 ^name predict-no)
- <=WM: (13603: O1941 ^name predict-yes)
- <=WM: (13602: R974 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1943 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1944 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1942 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1941 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.389539(R,m,v=1,0.888889,0.0993789)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*44 0.288049 0.322413 0.610462 -> 0.288049 0.322413 0.610461(R,m,v=1,1,0)
- =>WM: (13623: S1 ^operator O1944)
- 972: O: O1944 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N972 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N971 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13624: I3 ^predict-no N972)
- <=WM: (13610: N971 ^status complete)
- <=WM: (13609: I3 ^predict-yes N971)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (13628: I2 ^dir U)
- =>WM: (13627: I2 ^reward 1)
- =>WM: (13626: I2 ^see 0)
- =>WM: (13625: N972 ^status complete)
- <=WM: (13613: I2 ^dir U)
- <=WM: (13612: I2 ^reward 1)
- <=WM: (13611: I2 ^see 1)
- =>WM: (13629: I2 ^level-1 L1-root)
- <=WM: (13614: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R976 ^value 1 +)
- (R1 ^reward R976 +)
- Firing propose*predict-yes
- -->
- (O1945 ^name predict-yes +)
- (S1 ^operator O1945 +)
- Firing propose*predict-no
- -->
- (O1946 ^name predict-no +)
- (S1 ^operator O1946 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1944 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1943 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1944 ^name predict-no +)
- (S1 ^operator O1944 +)
- Retracting propose*predict-yes
- -->
- (O1943 ^name predict-yes +)
- (S1 ^operator O1943 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R975 ^value 1 +)
- (R1 ^reward R975 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1944 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1943 = 0.)
- =>WM: (13636: S1 ^operator O1946 +)
- =>WM: (13635: S1 ^operator O1945 +)
- =>WM: (13634: O1946 ^name predict-no)
- =>WM: (13633: O1945 ^name predict-yes)
- =>WM: (13632: R976 ^value 1)
- =>WM: (13631: R1 ^reward R976)
- =>WM: (13630: I3 ^see 0)
- <=WM: (13621: S1 ^operator O1943 +)
- <=WM: (13622: S1 ^operator O1944 +)
- <=WM: (13623: S1 ^operator O1944)
- <=WM: (13616: R1 ^reward R975)
- <=WM: (13615: I3 ^see 1)
- <=WM: (13619: O1944 ^name predict-no)
- <=WM: (13618: O1943 ^name predict-yes)
- <=WM: (13617: R975 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1945 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1946 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1944 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1943 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (13637: S1 ^operator O1946)
- 973: O: O1946 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N973 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N972 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13638: I3 ^predict-no N973)
- <=WM: (13625: N972 ^status complete)
- <=WM: (13624: I3 ^predict-no N972)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- -/--- Input Phase ---
- =>WM: (13642: I2 ^dir U)
- =>WM: (13641: I2 ^reward 1)
- =>WM: (13640: I2 ^see 0)
- =>WM: (13639: N973 ^status complete)
- <=WM: (13628: I2 ^dir U)
- <=WM: (13627: I2 ^reward 1)
- <=WM: (13626: I2 ^see 0)
- =>WM: (13643: I2 ^level-1 L1-root)
- <=WM: (13629: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R977 ^value 1 +)
- (R1 ^reward R977 +)
- Firing propose*predict-yes
- -->
- (O1947 ^name predict-yes +)
- (S1 ^operator O1947 +)
- Firing propose*predict-no
- -->
- (O1948 ^name predict-no +)
- (S1 ^operator O1948 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1946 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1945 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1946 ^name predict-no +)
- (S1 ^operator O1946 +)
- Retracting propose*predict-yes
- -->
- (O1945 ^name predict-yes +)
- (S1 ^operator O1945 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R976 ^value 1 +)
- (R1 ^reward R976 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1946 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1945 = 0.)
- =>WM: (13649: S1 ^operator O1948 +)
- =>WM: (13648: S1 ^operator O1947 +)
- =>WM: (13647: O1948 ^name predict-no)
- =>WM: (13646: O1947 ^name predict-yes)
- =>WM: (13645: R977 ^value 1)
- =>WM: (13644: R1 ^reward R977)
- <=WM: (13635: S1 ^operator O1945 +)
- <=WM: (13636: S1 ^operator O1946 +)
- <=WM: (13637: S1 ^operator O1946)
- <=WM: (13631: R1 ^reward R976)
- <=WM: (13634: O1946 ^name predict-no)
- <=WM: (13633: O1945 ^name predict-yes)
- <=WM: (13632: R976 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1947 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1948 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1946 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1945 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (13650: S1 ^operator O1948)
- 974: O: O1948 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N974 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N973 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13651: I3 ^predict-no N974)
- <=WM: (13639: N973 ^status complete)
- <=WM: (13638: I3 ^predict-no N973)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (13655: I2 ^dir U)
- =>WM: (13654: I2 ^reward 1)
- =>WM: (13653: I2 ^see 0)
- =>WM: (13652: N974 ^status complete)
- <=WM: (13642: I2 ^dir U)
- <=WM: (13641: I2 ^reward 1)
- <=WM: (13640: I2 ^see 0)
- =>WM: (13656: I2 ^level-1 L1-root)
- <=WM: (13643: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R978 ^value 1 +)
- (R1 ^reward R978 +)
- Firing propose*predict-yes
- -->
- (O1949 ^name predict-yes +)
- (S1 ^operator O1949 +)
- Firing propose*predict-no
- -->
- (O1950 ^name predict-no +)
- (S1 ^operator O1950 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1948 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1947 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1948 ^name predict-no +)
- (S1 ^operator O1948 +)
- Retracting propose*predict-yes
- -->
- (O1947 ^name predict-yes +)
- (S1 ^operator O1947 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R977 ^value 1 +)
- (R1 ^reward R977 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1948 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1947 = 0.)
- =>WM: (13662: S1 ^operator O1950 +)
- =>WM: (13661: S1 ^operator O1949 +)
- =>WM: (13660: O1950 ^name predict-no)
- =>WM: (13659: O1949 ^name predict-yes)
- =>WM: (13658: R978 ^value 1)
- =>WM: (13657: R1 ^reward R978)
- <=WM: (13648: S1 ^operator O1947 +)
- <=WM: (13649: S1 ^operator O1948 +)
- <=WM: (13650: S1 ^operator O1948)
- <=WM: (13644: R1 ^reward R977)
- <=WM: (13647: O1948 ^name predict-no)
- <=WM: (13646: O1947 ^name predict-yes)
- <=WM: (13645: R977 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1949 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1950 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1948 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1947 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (13663: S1 ^operator O1950)
- 975: O: O1950 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N975 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N974 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13664: I3 ^predict-no N975)
- <=WM: (13652: N974 ^status complete)
- <=WM: (13651: I3 ^predict-no N974)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (13668: I2 ^dir R)
- =>WM: (13667: I2 ^reward 1)
- =>WM: (13666: I2 ^see 0)
- =>WM: (13665: N975 ^status complete)
- <=WM: (13655: I2 ^dir U)
- <=WM: (13654: I2 ^reward 1)
- <=WM: (13653: I2 ^see 0)
- =>WM: (13669: I2 ^level-1 L1-root)
- <=WM: (13656: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O1950 = -0.02155734064455064)
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1949 = 0.8155758449529213)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R979 ^value 1 +)
- (R1 ^reward R979 +)
- Firing propose*predict-yes
- -->
- (O1951 ^name predict-yes +)
- (S1 ^operator O1951 +)
- Firing propose*predict-no
- -->
- (O1952 ^name predict-no +)
- (S1 ^operator O1952 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1950 = 0.4476189814068987)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1949 = 0.1844075128644186)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1950 ^name predict-no +)
- (S1 ^operator O1950 +)
- Retracting propose*predict-yes
- -->
- (O1949 ^name predict-yes +)
- (S1 ^operator O1949 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R978 ^value 1 +)
- (R1 ^reward R978 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1950 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1949 = 0.)
- =>WM: (13676: S1 ^operator O1952 +)
- =>WM: (13675: S1 ^operator O1951 +)
- =>WM: (13674: I3 ^dir R)
- =>WM: (13673: O1952 ^name predict-no)
- =>WM: (13672: O1951 ^name predict-yes)
- =>WM: (13671: R979 ^value 1)
- =>WM: (13670: R1 ^reward R979)
- <=WM: (13661: S1 ^operator O1949 +)
- <=WM: (13662: S1 ^operator O1950 +)
- <=WM: (13663: S1 ^operator O1950)
- <=WM: (13620: I3 ^dir U)
- <=WM: (13657: R1 ^reward R978)
- <=WM: (13660: O1950 ^name predict-no)
- <=WM: (13659: O1949 ^name predict-yes)
- <=WM: (13658: R978 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1951 = 0.8155758449529213)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1951 = 0.1844075128644186)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O1952 = -0.02155734064455064)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1952 = 0.4476189814068987)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1950 = 0.4476189814068987)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O1950 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1949 = 0.1844075128644186)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1949 = 0.8155758449529213)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (13677: S1 ^operator O1951)
- 976: O: O1951 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N976 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N975 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13678: I3 ^predict-yes N976)
- <=WM: (13665: N975 ^status complete)
- <=WM: (13664: I3 ^predict-no N975)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (13682: I2 ^dir L)
- =>WM: (13681: I2 ^reward 1)
- =>WM: (13680: I2 ^see 1)
- =>WM: (13679: N976 ^status complete)
- <=WM: (13668: I2 ^dir R)
- <=WM: (13667: I2 ^reward 1)
- <=WM: (13666: I2 ^see 0)
- =>WM: (13683: I2 ^level-1 R1-root)
- <=WM: (13669: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O1951 = 0.6104589917494525)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O1952 = 0.2714993082286609)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R980 ^value 1 +)
- (R1 ^reward R980 +)
- Firing propose*predict-yes
- -->
- (O1953 ^name predict-yes +)
- (S1 ^operator O1953 +)
- Firing propose*predict-no
- -->
- (O1954 ^name predict-no +)
- (S1 ^operator O1954 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1952 = 0.387336901415443)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1951 = 0.389539338123754)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1952 ^name predict-no +)
- (S1 ^operator O1952 +)
- Retracting propose*predict-yes
- -->
- (O1951 ^name predict-yes +)
- (S1 ^operator O1951 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R979 ^value 1 +)
- (R1 ^reward R979 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1952 = 0.4476189814068987)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O1952 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1951 = 0.1844075128644186)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1951 = 0.8155758449529213)
- =>WM: (13691: S1 ^operator O1954 +)
- =>WM: (13690: S1 ^operator O1953 +)
- =>WM: (13689: I3 ^dir L)
- =>WM: (13688: O1954 ^name predict-no)
- =>WM: (13687: O1953 ^name predict-yes)
- =>WM: (13686: R980 ^value 1)
- =>WM: (13685: R1 ^reward R980)
- =>WM: (13684: I3 ^see 1)
- <=WM: (13675: S1 ^operator O1951 +)
- <=WM: (13677: S1 ^operator O1951)
- <=WM: (13676: S1 ^operator O1952 +)
- <=WM: (13674: I3 ^dir R)
- <=WM: (13670: R1 ^reward R979)
- <=WM: (13630: I3 ^see 0)
- <=WM: (13673: O1952 ^name predict-no)
- <=WM: (13672: O1951 ^name predict-yes)
- <=WM: (13671: R979 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1953 = 0.389539338123754)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O1953 = 0.6104589917494525)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1954 = 0.387336901415443)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O1954 = 0.2714993082286609)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1952 = 0.387336901415443)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O1952 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1951 = 0.389539338123754)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O1951 = 0.6104589917494525)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675409 -0.491002 0.184408 -> 0.675412 -0.491002 0.18441(R,m,v=1,0.89697,0.0929786)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*46 0.324569 0.491006 0.815576 -> 0.324573 0.491006 0.815578(R,m,v=1,1,0)
- =>WM: (13692: S1 ^operator O1953)
- 977: O: O1953 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N977 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N976 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13693: I3 ^predict-yes N977)
- <=WM: (13679: N976 ^status complete)
- <=WM: (13678: I3 ^predict-yes N976)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (13697: I2 ^dir L)
- =>WM: (13696: I2 ^reward 1)
- =>WM: (13695: I2 ^see 1)
- =>WM: (13694: N977 ^status complete)
- <=WM: (13682: I2 ^dir L)
- <=WM: (13681: I2 ^reward 1)
- <=WM: (13680: I2 ^see 1)
- =>WM: (13698: I2 ^level-1 L1-root)
- <=WM: (13683: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1954 = 0.6126626863207351)
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1953 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R981 ^value 1 +)
- (R1 ^reward R981 +)
- Firing propose*predict-yes
- -->
- (O1955 ^name predict-yes +)
- (S1 ^operator O1955 +)
- Firing propose*predict-no
- -->
- (O1956 ^name predict-no +)
- (S1 ^operator O1956 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1954 = 0.387336901415443)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1953 = 0.389539338123754)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1954 ^name predict-no +)
- (S1 ^operator O1954 +)
- Retracting propose*predict-yes
- -->
- (O1953 ^name predict-yes +)
- (S1 ^operator O1953 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R980 ^value 1 +)
- (R1 ^reward R980 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O1954 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1954 = 0.387336901415443)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O1953 = 0.6104589917494525)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1953 = 0.389539338123754)
- =>WM: (13704: S1 ^operator O1956 +)
- =>WM: (13703: S1 ^operator O1955 +)
- =>WM: (13702: O1956 ^name predict-no)
- =>WM: (13701: O1955 ^name predict-yes)
- =>WM: (13700: R981 ^value 1)
- =>WM: (13699: R1 ^reward R981)
- <=WM: (13690: S1 ^operator O1953 +)
- <=WM: (13692: S1 ^operator O1953)
- <=WM: (13691: S1 ^operator O1954 +)
- <=WM: (13685: R1 ^reward R980)
- <=WM: (13688: O1954 ^name predict-no)
- <=WM: (13687: O1953 ^name predict-yes)
- <=WM: (13686: R980 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1955 = 0.389539338123754)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1955 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1956 = 0.387336901415443)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1956 = 0.6126626863207351)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1954 = 0.387336901415443)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1954 = 0.6126626863207351)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1953 = 0.389539338123754)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1953 = -0.02274740735326741)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.389539 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.889571,0.0988412)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*36 0.288049 0.32241 0.610459 -> 0.288049 0.322411 0.610459(R,m,v=1,1,0)
- =>WM: (13705: S1 ^operator O1956)
- 978: O: O1956 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N978 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N977 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13706: I3 ^predict-no N978)
- <=WM: (13694: N977 ^status complete)
- <=WM: (13693: I3 ^predict-yes N977)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (13710: I2 ^dir R)
- =>WM: (13709: I2 ^reward 1)
- =>WM: (13708: I2 ^see 0)
- =>WM: (13707: N978 ^status complete)
- <=WM: (13697: I2 ^dir L)
- <=WM: (13696: I2 ^reward 1)
- <=WM: (13695: I2 ^see 1)
- =>WM: (13711: I2 ^level-1 L0-root)
- <=WM: (13698: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1955 = 0.8155955750807526)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1956 = -0.00558448899823713)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R982 ^value 1 +)
- (R1 ^reward R982 +)
- Firing propose*predict-yes
- -->
- (O1957 ^name predict-yes +)
- (S1 ^operator O1957 +)
- Firing propose*predict-no
- -->
- (O1958 ^name predict-no +)
- (S1 ^operator O1958 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1956 = 0.4476189814068987)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1955 = 0.1844100091918176)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1956 ^name predict-no +)
- (S1 ^operator O1956 +)
- Retracting propose*predict-yes
- -->
- (O1955 ^name predict-yes +)
- (S1 ^operator O1955 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R981 ^value 1 +)
- (R1 ^reward R981 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1956 = 0.6126626863207351)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1956 = 0.387336901415443)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1955 = -0.02274740735326741)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1955 = 0.389539588642773)
- =>WM: (13719: S1 ^operator O1958 +)
- =>WM: (13718: S1 ^operator O1957 +)
- =>WM: (13717: I3 ^dir R)
- =>WM: (13716: O1958 ^name predict-no)
- =>WM: (13715: O1957 ^name predict-yes)
- =>WM: (13714: R982 ^value 1)
- =>WM: (13713: R1 ^reward R982)
- =>WM: (13712: I3 ^see 0)
- <=WM: (13703: S1 ^operator O1955 +)
- <=WM: (13704: S1 ^operator O1956 +)
- <=WM: (13705: S1 ^operator O1956)
- <=WM: (13689: I3 ^dir L)
- <=WM: (13699: R1 ^reward R981)
- <=WM: (13684: I3 ^see 1)
- <=WM: (13702: O1956 ^name predict-no)
- <=WM: (13701: O1955 ^name predict-yes)
- <=WM: (13700: R981 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1957 = 0.1844100091918176)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1957 = 0.8155955750807526)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1958 = 0.4476189814068987)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1958 = -0.00558448899823713)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1956 = 0.4476189814068987)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1956 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1955 = 0.1844100091918176)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1955 = 0.8155955750807526)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.719081 -0.331744 0.387337 -> 0.719081 -0.331744 0.387337(R,m,v=1,0.931429,0.0642365)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*32 0.280919 0.331744 0.612663 -> 0.280919 0.331744 0.612663(R,m,v=1,1,0)
- =>WM: (13720: S1 ^operator O1957)
- 979: O: O1957 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N979 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N978 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13721: I3 ^predict-yes N979)
- <=WM: (13707: N978 ^status complete)
- <=WM: (13706: I3 ^predict-no N978)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (13725: I2 ^dir U)
- =>WM: (13724: I2 ^reward 1)
- =>WM: (13723: I2 ^see 1)
- =>WM: (13722: N979 ^status complete)
- <=WM: (13710: I2 ^dir R)
- <=WM: (13709: I2 ^reward 1)
- <=WM: (13708: I2 ^see 0)
- =>WM: (13726: I2 ^level-1 R1-root)
- <=WM: (13711: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R983 ^value 1 +)
- (R1 ^reward R983 +)
- Firing propose*predict-yes
- -->
- (O1959 ^name predict-yes +)
- (S1 ^operator O1959 +)
- Firing propose*predict-no
- -->
- (O1960 ^name predict-no +)
- (S1 ^operator O1960 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1958 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1957 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1958 ^name predict-no +)
- (S1 ^operator O1958 +)
- Retracting propose*predict-yes
- -->
- (O1957 ^name predict-yes +)
- (S1 ^operator O1957 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R982 ^value 1 +)
- (R1 ^reward R982 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1958 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1958 = 0.4476189814068987)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1957 = 0.8155955750807526)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1957 = 0.1844100091918176)
- =>WM: (13734: S1 ^operator O1960 +)
- =>WM: (13733: S1 ^operator O1959 +)
- =>WM: (13732: I3 ^dir U)
- =>WM: (13731: O1960 ^name predict-no)
- =>WM: (13730: O1959 ^name predict-yes)
- =>WM: (13729: R983 ^value 1)
- =>WM: (13728: R1 ^reward R983)
- =>WM: (13727: I3 ^see 1)
- <=WM: (13718: S1 ^operator O1957 +)
- <=WM: (13720: S1 ^operator O1957)
- <=WM: (13719: S1 ^operator O1958 +)
- <=WM: (13717: I3 ^dir R)
- <=WM: (13713: R1 ^reward R982)
- <=WM: (13712: I3 ^see 0)
- <=WM: (13716: O1958 ^name predict-no)
- <=WM: (13715: O1957 ^name predict-yes)
- <=WM: (13714: R982 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1959 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1960 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1958 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1957 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675412 -0.491002 0.18441 -> 0.675411 -0.491002 0.184409(R,m,v=1,0.89759,0.092479)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*34 0.324595 0.491001 0.815596 -> 0.324594 0.491001 0.815595(R,m,v=1,1,0)
- =>WM: (13735: S1 ^operator O1960)
- 980: O: O1960 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N980 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N979 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13736: I3 ^predict-no N980)
- <=WM: (13722: N979 ^status complete)
- <=WM: (13721: I3 ^predict-yes N979)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (13740: I2 ^dir U)
- =>WM: (13739: I2 ^reward 1)
- =>WM: (13738: I2 ^see 0)
- =>WM: (13737: N980 ^status complete)
- <=WM: (13725: I2 ^dir U)
- <=WM: (13724: I2 ^reward 1)
- <=WM: (13723: I2 ^see 1)
- =>WM: (13741: I2 ^level-1 R1-root)
- <=WM: (13726: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R984 ^value 1 +)
- (R1 ^reward R984 +)
- Firing propose*predict-yes
- -->
- (O1961 ^name predict-yes +)
- (S1 ^operator O1961 +)
- Firing propose*predict-no
- -->
- (O1962 ^name predict-no +)
- (S1 ^operator O1962 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1960 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1959 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1960 ^name predict-no +)
- (S1 ^operator O1960 +)
- Retracting propose*predict-yes
- -->
- (O1959 ^name predict-yes +)
- (S1 ^operator O1959 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R983 ^value 1 +)
- (R1 ^reward R983 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1960 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1959 = 0.)
- =>WM: (13748: S1 ^operator O1962 +)
- =>WM: (13747: S1 ^operator O1961 +)
- =>WM: (13746: O1962 ^name predict-no)
- =>WM: (13745: O1961 ^name predict-yes)
- =>WM: (13744: R984 ^value 1)
- =>WM: (13743: R1 ^reward R984)
- =>WM: (13742: I3 ^see 0)
- <=WM: (13733: S1 ^operator O1959 +)
- <=WM: (13734: S1 ^operator O1960 +)
- <=WM: (13735: S1 ^operator O1960)
- <=WM: (13728: R1 ^reward R983)
- <=WM: (13727: I3 ^see 1)
- <=WM: (13731: O1960 ^name predict-no)
- <=WM: (13730: O1959 ^name predict-yes)
- <=WM: (13729: R983 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1961 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1962 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1960 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1959 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (13749: S1 ^operator O1962)
- 981: O: O1962 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N981 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N980 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13750: I3 ^predict-no N981)
- <=WM: (13737: N980 ^status complete)
- <=WM: (13736: I3 ^predict-no N980)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- \--- Input Phase ---
- =>WM: (13754: I2 ^dir U)
- =>WM: (13753: I2 ^reward 1)
- =>WM: (13752: I2 ^see 0)
- =>WM: (13751: N981 ^status complete)
- <=WM: (13740: I2 ^dir U)
- <=WM: (13739: I2 ^reward 1)
- <=WM: (13738: I2 ^see 0)
- =>WM: (13755: I2 ^level-1 R1-root)
- <=WM: (13741: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R985 ^value 1 +)
- (R1 ^reward R985 +)
- Firing propose*predict-yes
- -->
- (O1963 ^name predict-yes +)
- (S1 ^operator O1963 +)
- Firing propose*predict-no
- -->
- (O1964 ^name predict-no +)
- (S1 ^operator O1964 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1962 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1961 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1962 ^name predict-no +)
- (S1 ^operator O1962 +)
- Retracting propose*predict-yes
- -->
- (O1961 ^name predict-yes +)
- (S1 ^operator O1961 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R984 ^value 1 +)
- (R1 ^reward R984 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1962 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1961 = 0.)
- =>WM: (13761: S1 ^operator O1964 +)
- =>WM: (13760: S1 ^operator O1963 +)
- =>WM: (13759: O1964 ^name predict-no)
- =>WM: (13758: O1963 ^name predict-yes)
- =>WM: (13757: R985 ^value 1)
- =>WM: (13756: R1 ^reward R985)
- <=WM: (13747: S1 ^operator O1961 +)
- <=WM: (13748: S1 ^operator O1962 +)
- <=WM: (13749: S1 ^operator O1962)
- <=WM: (13743: R1 ^reward R984)
- <=WM: (13746: O1962 ^name predict-no)
- <=WM: (13745: O1961 ^name predict-yes)
- <=WM: (13744: R984 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1963 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1964 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1962 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1961 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (13762: S1 ^operator O1964)
- 982: O: O1964 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N982 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N981 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13763: I3 ^predict-no N982)
- <=WM: (13751: N981 ^status complete)
- <=WM: (13750: I3 ^predict-no N981)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (13767: I2 ^dir R)
- =>WM: (13766: I2 ^reward 1)
- =>WM: (13765: I2 ^see 0)
- =>WM: (13764: N982 ^status complete)
- <=WM: (13754: I2 ^dir U)
- <=WM: (13753: I2 ^reward 1)
- <=WM: (13752: I2 ^see 0)
- =>WM: (13768: I2 ^level-1 R1-root)
- <=WM: (13755: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1963 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1964 = 0.5523825060913952)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R986 ^value 1 +)
- (R1 ^reward R986 +)
- Firing propose*predict-yes
- -->
- (O1965 ^name predict-yes +)
- (S1 ^operator O1965 +)
- Firing propose*predict-no
- -->
- (O1966 ^name predict-no +)
- (S1 ^operator O1966 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1964 = 0.4476189814068987)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1963 = 0.1844091715509321)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1964 ^name predict-no +)
- (S1 ^operator O1964 +)
- Retracting propose*predict-yes
- -->
- (O1963 ^name predict-yes +)
- (S1 ^operator O1963 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R985 ^value 1 +)
- (R1 ^reward R985 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1964 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1963 = 0.)
- =>WM: (13775: S1 ^operator O1966 +)
- =>WM: (13774: S1 ^operator O1965 +)
- =>WM: (13773: I3 ^dir R)
- =>WM: (13772: O1966 ^name predict-no)
- =>WM: (13771: O1965 ^name predict-yes)
- =>WM: (13770: R986 ^value 1)
- =>WM: (13769: R1 ^reward R986)
- <=WM: (13760: S1 ^operator O1963 +)
- <=WM: (13761: S1 ^operator O1964 +)
- <=WM: (13762: S1 ^operator O1964)
- <=WM: (13732: I3 ^dir U)
- <=WM: (13756: R1 ^reward R985)
- <=WM: (13759: O1964 ^name predict-no)
- <=WM: (13758: O1963 ^name predict-yes)
- <=WM: (13757: R985 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1965 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1965 = 0.1844091715509321)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1966 = 0.5523825060913952)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1966 = 0.4476189814068987)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1964 = 0.4476189814068987)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1964 = 0.5523825060913952)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1963 = 0.1844091715509321)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1963 = 0.1398795999120246)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (13776: S1 ^operator O1966)
- 983: O: O1966 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N983 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N982 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13777: I3 ^predict-no N983)
- <=WM: (13764: N982 ^status complete)
- <=WM: (13763: I3 ^predict-no N982)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- \-/|--- Input Phase ---
- =>WM: (13781: I2 ^dir R)
- =>WM: (13780: I2 ^reward 1)
- =>WM: (13779: I2 ^see 0)
- =>WM: (13778: N983 ^status complete)
- <=WM: (13767: I2 ^dir R)
- <=WM: (13766: I2 ^reward 1)
- <=WM: (13765: I2 ^see 0)
- =>WM: (13782: I2 ^level-1 R0-root)
- <=WM: (13768: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O1965 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O1966 = 0.5523783049582921)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R987 ^value 1 +)
- (R1 ^reward R987 +)
- Firing propose*predict-yes
- -->
- (O1967 ^name predict-yes +)
- (S1 ^operator O1967 +)
- Firing propose*predict-no
- -->
- (O1968 ^name predict-no +)
- (S1 ^operator O1968 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1966 = 0.4476189814068987)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1965 = 0.1844091715509321)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1966 ^name predict-no +)
- (S1 ^operator O1966 +)
- Retracting propose*predict-yes
- -->
- (O1965 ^name predict-yes +)
- (S1 ^operator O1965 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R986 ^value 1 +)
- (R1 ^reward R986 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1966 = 0.4476189814068987)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1966 = 0.5523825060913952)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1965 = 0.1844091715509321)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1965 = 0.1398795999120246)
- =>WM: (13788: S1 ^operator O1968 +)
- =>WM: (13787: S1 ^operator O1967 +)
- =>WM: (13786: O1968 ^name predict-no)
- =>WM: (13785: O1967 ^name predict-yes)
- =>WM: (13784: R987 ^value 1)
- =>WM: (13783: R1 ^reward R987)
- <=WM: (13774: S1 ^operator O1965 +)
- <=WM: (13775: S1 ^operator O1966 +)
- <=WM: (13776: S1 ^operator O1966)
- <=WM: (13769: R1 ^reward R986)
- <=WM: (13772: O1966 ^name predict-no)
- <=WM: (13771: O1965 ^name predict-yes)
- <=WM: (13770: R986 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1967 = 0.1844091715509321)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O1967 = 0.1664311307472832)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1968 = 0.4476189814068987)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O1968 = 0.5523783049582921)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1966 = 0.4476189814068987)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O1966 = 0.5523783049582921)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1965 = 0.1844091715509321)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O1965 = 0.1664311307472832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622532 -0.174914 0.447619(R,m,v=1,0.927419,0.06786)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*39 0.377469 0.174914 0.552383 -> 0.377468 0.174914 0.552382(R,m,v=1,1,0)
- =>WM: (13789: S1 ^operator O1968)
- 984: O: O1968 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N984 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N983 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13790: I3 ^predict-no N984)
- <=WM: (13778: N983 ^status complete)
- <=WM: (13777: I3 ^predict-no N983)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (13794: I2 ^dir U)
- =>WM: (13793: I2 ^reward 1)
- =>WM: (13792: I2 ^see 0)
- =>WM: (13791: N984 ^status complete)
- <=WM: (13781: I2 ^dir R)
- <=WM: (13780: I2 ^reward 1)
- <=WM: (13779: I2 ^see 0)
- =>WM: (13795: I2 ^level-1 R0-root)
- <=WM: (13782: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R988 ^value 1 +)
- (R1 ^reward R988 +)
- Firing propose*predict-yes
- -->
- (O1969 ^name predict-yes +)
- (S1 ^operator O1969 +)
- Firing propose*predict-no
- -->
- (O1970 ^name predict-no +)
- (S1 ^operator O1970 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1968 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1967 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1968 ^name predict-no +)
- (S1 ^operator O1968 +)
- Retracting propose*predict-yes
- -->
- (O1967 ^name predict-yes +)
- (S1 ^operator O1967 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R987 ^value 1 +)
- (R1 ^reward R987 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O1968 = 0.5523783049582921)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1968 = 0.4476187582821546)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O1967 = 0.1664311307472832)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1967 = 0.1844091715509321)
- =>WM: (13802: S1 ^operator O1970 +)
- =>WM: (13801: S1 ^operator O1969 +)
- =>WM: (13800: I3 ^dir U)
- =>WM: (13799: O1970 ^name predict-no)
- =>WM: (13798: O1969 ^name predict-yes)
- =>WM: (13797: R988 ^value 1)
- =>WM: (13796: R1 ^reward R988)
- <=WM: (13787: S1 ^operator O1967 +)
- <=WM: (13788: S1 ^operator O1968 +)
- <=WM: (13789: S1 ^operator O1968)
- <=WM: (13773: I3 ^dir R)
- <=WM: (13783: R1 ^reward R987)
- <=WM: (13786: O1968 ^name predict-no)
- <=WM: (13785: O1967 ^name predict-yes)
- <=WM: (13784: R987 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1969 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1970 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1968 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1967 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622532 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.928,0.0673548)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*41 0.377465 0.174913 0.552378 -> 0.377466 0.174913 0.552379(R,m,v=1,1,0)
- =>WM: (13803: S1 ^operator O1970)
- 985: O: O1970 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N985 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N984 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13804: I3 ^predict-no N985)
- <=WM: (13791: N984 ^status complete)
- <=WM: (13790: I3 ^predict-no N984)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- |\-/--- Input Phase ---
- =>WM: (13808: I2 ^dir L)
- =>WM: (13807: I2 ^reward 1)
- =>WM: (13806: I2 ^see 0)
- =>WM: (13805: N985 ^status complete)
- <=WM: (13794: I2 ^dir U)
- <=WM: (13793: I2 ^reward 1)
- <=WM: (13792: I2 ^see 0)
- =>WM: (13809: I2 ^level-1 R0-root)
- <=WM: (13795: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1969 = 0.6104614609336363)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1970 = 0.1063475139796038)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R989 ^value 1 +)
- (R1 ^reward R989 +)
- Firing propose*predict-yes
- -->
- (O1971 ^name predict-yes +)
- (S1 ^operator O1971 +)
- Firing propose*predict-no
- -->
- (O1972 ^name predict-no +)
- (S1 ^operator O1972 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1970 = 0.3873369632550164)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1969 = 0.389539588642773)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1970 ^name predict-no +)
- (S1 ^operator O1970 +)
- Retracting propose*predict-yes
- -->
- (O1969 ^name predict-yes +)
- (S1 ^operator O1969 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R988 ^value 1 +)
- (R1 ^reward R988 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O1970 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1969 = 0.)
- =>WM: (13816: S1 ^operator O1972 +)
- =>WM: (13815: S1 ^operator O1971 +)
- =>WM: (13814: I3 ^dir L)
- =>WM: (13813: O1972 ^name predict-no)
- =>WM: (13812: O1971 ^name predict-yes)
- =>WM: (13811: R989 ^value 1)
- =>WM: (13810: R1 ^reward R989)
- <=WM: (13801: S1 ^operator O1969 +)
- <=WM: (13802: S1 ^operator O1970 +)
- <=WM: (13803: S1 ^operator O1970)
- <=WM: (13800: I3 ^dir U)
- <=WM: (13796: R1 ^reward R988)
- <=WM: (13799: O1970 ^name predict-no)
- <=WM: (13798: O1969 ^name predict-yes)
- <=WM: (13797: R988 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1971 = 0.6104614609336363)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1971 = 0.389539588642773)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1972 = 0.1063475139796038)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1972 = 0.3873369632550164)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1970 = 0.3873369632550164)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1970 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1969 = 0.389539588642773)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1969 = 0.6104614609336363)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (13817: S1 ^operator O1971)
- 986: O: O1971 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N986 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N985 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13818: I3 ^predict-yes N986)
- <=WM: (13805: N985 ^status complete)
- <=WM: (13804: I3 ^predict-no N985)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (13822: I2 ^dir R)
- =>WM: (13821: I2 ^reward 1)
- =>WM: (13820: I2 ^see 1)
- =>WM: (13819: N986 ^status complete)
- <=WM: (13808: I2 ^dir L)
- <=WM: (13807: I2 ^reward 1)
- <=WM: (13806: I2 ^see 0)
- =>WM: (13823: I2 ^level-1 L1-root)
- <=WM: (13809: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O1972 = -0.02155734064455064)
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1971 = 0.8155783412803204)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R990 ^value 1 +)
- (R1 ^reward R990 +)
- Firing propose*predict-yes
- -->
- (O1973 ^name predict-yes +)
- (S1 ^operator O1973 +)
- Firing propose*predict-no
- -->
- (O1974 ^name predict-no +)
- (S1 ^operator O1974 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1972 = 0.4476191987960876)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1971 = 0.1844091715509321)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1972 ^name predict-no +)
- (S1 ^operator O1972 +)
- Retracting propose*predict-yes
- -->
- (O1971 ^name predict-yes +)
- (S1 ^operator O1971 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R989 ^value 1 +)
- (R1 ^reward R989 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1972 = 0.3873369632550164)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1972 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1971 = 0.389539588642773)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1971 = 0.6104614609336363)
- =>WM: (13831: S1 ^operator O1974 +)
- =>WM: (13830: S1 ^operator O1973 +)
- =>WM: (13829: I3 ^dir R)
- =>WM: (13828: O1974 ^name predict-no)
- =>WM: (13827: O1973 ^name predict-yes)
- =>WM: (13826: R990 ^value 1)
- =>WM: (13825: R1 ^reward R990)
- =>WM: (13824: I3 ^see 1)
- <=WM: (13815: S1 ^operator O1971 +)
- <=WM: (13817: S1 ^operator O1971)
- <=WM: (13816: S1 ^operator O1972 +)
- <=WM: (13814: I3 ^dir L)
- <=WM: (13810: R1 ^reward R989)
- <=WM: (13742: I3 ^see 0)
- <=WM: (13813: O1972 ^name predict-no)
- <=WM: (13812: O1971 ^name predict-yes)
- <=WM: (13811: R989 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1973 = 0.1844091715509321)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1973 = 0.8155783412803204)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1974 = 0.4476191987960876)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O1974 = -0.02155734064455064)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1972 = 0.4476191987960876)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O1972 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1971 = 0.1844091715509321)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1971 = 0.8155783412803204)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.389539(R,m,v=1,0.890244,0.0983091)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*44 0.288049 0.322413 0.610461 -> 0.288049 0.322412 0.610461(R,m,v=1,1,0)
- =>WM: (13832: S1 ^operator O1973)
- 987: O: O1973 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N987 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N986 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13833: I3 ^predict-yes N987)
- <=WM: (13819: N986 ^status complete)
- <=WM: (13818: I3 ^predict-yes N986)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- /|--- Input Phase ---
- =>WM: (13837: I2 ^dir R)
- =>WM: (13836: I2 ^reward 1)
- =>WM: (13835: I2 ^see 1)
- =>WM: (13834: N987 ^status complete)
- <=WM: (13822: I2 ^dir R)
- <=WM: (13821: I2 ^reward 1)
- <=WM: (13820: I2 ^see 1)
- =>WM: (13838: I2 ^level-1 R1-root)
- <=WM: (13823: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1973 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1974 = 0.552382282966651)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R991 ^value 1 +)
- (R1 ^reward R991 +)
- Firing propose*predict-yes
- -->
- (O1975 ^name predict-yes +)
- (S1 ^operator O1975 +)
- Firing propose*predict-no
- -->
- (O1976 ^name predict-no +)
- (S1 ^operator O1976 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1974 = 0.4476191987960876)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1973 = 0.1844091715509321)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1974 ^name predict-no +)
- (S1 ^operator O1974 +)
- Retracting propose*predict-yes
- -->
- (O1973 ^name predict-yes +)
- (S1 ^operator O1973 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R990 ^value 1 +)
- (R1 ^reward R990 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O1974 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1974 = 0.4476191987960876)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1973 = 0.8155783412803204)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1973 = 0.1844091715509321)
- =>WM: (13844: S1 ^operator O1976 +)
- =>WM: (13843: S1 ^operator O1975 +)
- =>WM: (13842: O1976 ^name predict-no)
- =>WM: (13841: O1975 ^name predict-yes)
- =>WM: (13840: R991 ^value 1)
- =>WM: (13839: R1 ^reward R991)
- <=WM: (13830: S1 ^operator O1973 +)
- <=WM: (13832: S1 ^operator O1973)
- <=WM: (13831: S1 ^operator O1974 +)
- <=WM: (13825: R1 ^reward R990)
- <=WM: (13828: O1974 ^name predict-no)
- <=WM: (13827: O1973 ^name predict-yes)
- <=WM: (13826: R990 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1975 = 0.1844091715509321)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1975 = 0.1398795999120246)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1976 = 0.4476191987960876)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1976 = 0.552382282966651)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1974 = 0.4476191987960876)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1974 = 0.552382282966651)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1973 = 0.1844091715509321)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1973 = 0.1398795999120246)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675411 -0.491002 0.184409 -> 0.675414 -0.491003 0.184411(R,m,v=1,0.898204,0.0919847)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*46 0.324573 0.491006 0.815578 -> 0.324575 0.491005 0.81558(R,m,v=1,1,0)
- =>WM: (13845: S1 ^operator O1976)
- 988: O: O1976 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N988 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N987 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13846: I3 ^predict-no N988)
- <=WM: (13834: N987 ^status complete)
- <=WM: (13833: I3 ^predict-yes N987)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- \---- Input Phase ---
- =>WM: (13850: I2 ^dir R)
- =>WM: (13849: I2 ^reward 1)
- =>WM: (13848: I2 ^see 0)
- =>WM: (13847: N988 ^status complete)
- <=WM: (13837: I2 ^dir R)
- <=WM: (13836: I2 ^reward 1)
- <=WM: (13835: I2 ^see 1)
- =>WM: (13851: I2 ^level-1 R0-root)
- <=WM: (13838: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O1975 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O1976 = 0.5523787454722251)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R992 ^value 1 +)
- (R1 ^reward R992 +)
- Firing propose*predict-yes
- -->
- (O1977 ^name predict-yes +)
- (S1 ^operator O1977 +)
- Firing propose*predict-no
- -->
- (O1978 ^name predict-no +)
- (S1 ^operator O1978 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1976 = 0.4476191987960876)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1975 = 0.1844110446262441)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1976 ^name predict-no +)
- (S1 ^operator O1976 +)
- Retracting propose*predict-yes
- -->
- (O1975 ^name predict-yes +)
- (S1 ^operator O1975 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R991 ^value 1 +)
- (R1 ^reward R991 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1976 = 0.552382282966651)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1976 = 0.4476191987960876)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1975 = 0.1398795999120246)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1975 = 0.1844110446262441)
- =>WM: (13858: S1 ^operator O1978 +)
- =>WM: (13857: S1 ^operator O1977 +)
- =>WM: (13856: O1978 ^name predict-no)
- =>WM: (13855: O1977 ^name predict-yes)
- =>WM: (13854: R992 ^value 1)
- =>WM: (13853: R1 ^reward R992)
- =>WM: (13852: I3 ^see 0)
- <=WM: (13843: S1 ^operator O1975 +)
- <=WM: (13844: S1 ^operator O1976 +)
- <=WM: (13845: S1 ^operator O1976)
- <=WM: (13839: R1 ^reward R991)
- <=WM: (13824: I3 ^see 1)
- <=WM: (13842: O1976 ^name predict-no)
- <=WM: (13841: O1975 ^name predict-yes)
- <=WM: (13840: R991 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1977 = 0.1844110446262441)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O1977 = 0.1664311307472832)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1978 = 0.4476191987960876)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O1978 = 0.5523787454722251)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1976 = 0.4476191987960876)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O1976 = 0.5523787454722251)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1975 = 0.1844110446262441)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O1975 = 0.1664311307472832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.928571,0.0668571)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*39 0.377468 0.174914 0.552382 -> 0.377468 0.174914 0.552382(R,m,v=1,1,0)
- =>WM: (13859: S1 ^operator O1978)
- 989: O: O1978 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N989 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N988 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13860: I3 ^predict-no N989)
- <=WM: (13847: N988 ^status complete)
- <=WM: (13846: I3 ^predict-no N988)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (13864: I2 ^dir R)
- =>WM: (13863: I2 ^reward 1)
- =>WM: (13862: I2 ^see 0)
- =>WM: (13861: N989 ^status complete)
- <=WM: (13850: I2 ^dir R)
- <=WM: (13849: I2 ^reward 1)
- <=WM: (13848: I2 ^see 0)
- =>WM: (13865: I2 ^level-1 R0-root)
- <=WM: (13851: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O1977 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O1978 = 0.5523787454722251)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R993 ^value 1 +)
- (R1 ^reward R993 +)
- Firing propose*predict-yes
- -->
- (O1979 ^name predict-yes +)
- (S1 ^operator O1979 +)
- Firing propose*predict-no
- -->
- (O1980 ^name predict-no +)
- (S1 ^operator O1980 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1978 = 0.4476189765316768)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1977 = 0.1844110446262441)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1978 ^name predict-no +)
- (S1 ^operator O1978 +)
- Retracting propose*predict-yes
- -->
- (O1977 ^name predict-yes +)
- (S1 ^operator O1977 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R992 ^value 1 +)
- (R1 ^reward R992 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O1978 = 0.5523787454722251)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1978 = 0.4476189765316768)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O1977 = 0.1664311307472832)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1977 = 0.1844110446262441)
- =>WM: (13871: S1 ^operator O1980 +)
- =>WM: (13870: S1 ^operator O1979 +)
- =>WM: (13869: O1980 ^name predict-no)
- =>WM: (13868: O1979 ^name predict-yes)
- =>WM: (13867: R993 ^value 1)
- =>WM: (13866: R1 ^reward R993)
- <=WM: (13857: S1 ^operator O1977 +)
- <=WM: (13858: S1 ^operator O1978 +)
- <=WM: (13859: S1 ^operator O1978)
- <=WM: (13853: R1 ^reward R992)
- <=WM: (13856: O1978 ^name predict-no)
- <=WM: (13855: O1977 ^name predict-yes)
- <=WM: (13854: R992 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1979 = 0.1844110446262441)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O1979 = 0.1664311307472832)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1980 = 0.4476189765316768)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O1980 = 0.5523787454722251)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1978 = 0.4476189765316768)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O1978 = 0.5523787454722251)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1977 = 0.1844110446262441)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O1977 = 0.1664311307472832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.929134,0.0663667)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*41 0.377466 0.174913 0.552379 -> 0.377466 0.174913 0.552379(R,m,v=1,1,0)
- =>WM: (13872: S1 ^operator O1980)
- 990: O: O1980 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N990 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N989 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13873: I3 ^predict-no N990)
- <=WM: (13861: N989 ^status complete)
- <=WM: (13860: I3 ^predict-no N989)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- -/|\--- Input Phase ---
- =>WM: (13877: I2 ^dir L)
- =>WM: (13876: I2 ^reward 1)
- =>WM: (13875: I2 ^see 0)
- =>WM: (13874: N990 ^status complete)
- <=WM: (13864: I2 ^dir R)
- <=WM: (13863: I2 ^reward 1)
- <=WM: (13862: I2 ^see 0)
- =>WM: (13878: I2 ^level-1 R0-root)
- <=WM: (13865: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1979 = 0.6104613034971749)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1980 = 0.1063475139796038)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R994 ^value 1 +)
- (R1 ^reward R994 +)
- Firing propose*predict-yes
- -->
- (O1981 ^name predict-yes +)
- (S1 ^operator O1981 +)
- Firing propose*predict-no
- -->
- (O1982 ^name predict-no +)
- (S1 ^operator O1982 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1980 = 0.3873369632550164)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1979 = 0.3895394312063116)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1980 ^name predict-no +)
- (S1 ^operator O1980 +)
- Retracting propose*predict-yes
- -->
- (O1979 ^name predict-yes +)
- (S1 ^operator O1979 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R993 ^value 1 +)
- (R1 ^reward R993 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O1980 = 0.5523790871716397)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1980 = 0.4476193182310915)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O1979 = 0.1664311307472832)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1979 = 0.1844110446262441)
- =>WM: (13885: S1 ^operator O1982 +)
- =>WM: (13884: S1 ^operator O1981 +)
- =>WM: (13883: I3 ^dir L)
- =>WM: (13882: O1982 ^name predict-no)
- =>WM: (13881: O1981 ^name predict-yes)
- =>WM: (13880: R994 ^value 1)
- =>WM: (13879: R1 ^reward R994)
- <=WM: (13870: S1 ^operator O1979 +)
- <=WM: (13871: S1 ^operator O1980 +)
- <=WM: (13872: S1 ^operator O1980)
- <=WM: (13829: I3 ^dir R)
- <=WM: (13866: R1 ^reward R993)
- <=WM: (13869: O1980 ^name predict-no)
- <=WM: (13868: O1979 ^name predict-yes)
- <=WM: (13867: R993 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1981 = 0.6104613034971749)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1981 = 0.3895394312063116)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1982 = 0.1063475139796038)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1982 = 0.3873369632550164)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1980 = 0.3873369632550164)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1980 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1979 = 0.3895394312063116)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1979 = 0.6104613034971749)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174913 0.44762(R,m,v=1,0.929687,0.0658834)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*41 0.377466 0.174913 0.552379 -> 0.377466 0.174913 0.552379(R,m,v=1,1,0)
- =>WM: (13886: S1 ^operator O1981)
- 991: O: O1981 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N991 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N990 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13887: I3 ^predict-yes N991)
- <=WM: (13874: N990 ^status complete)
- <=WM: (13873: I3 ^predict-no N990)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- ---- Input Phase ---
- =>WM: (13891: I2 ^dir R)
- =>WM: (13890: I2 ^reward 1)
- =>WM: (13889: I2 ^see 1)
- =>WM: (13888: N991 ^status complete)
- <=WM: (13877: I2 ^dir L)
- <=WM: (13876: I2 ^reward 1)
- <=WM: (13875: I2 ^see 0)
- =>WM: (13892: I2 ^level-1 L1-root)
- <=WM: (13878: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O1982 = -0.02155734064455064)
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1981 = 0.8155802143556325)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R995 ^value 1 +)
- (R1 ^reward R995 +)
- Firing propose*predict-yes
- -->
- (O1983 ^name predict-yes +)
- (S1 ^operator O1983 +)
- Firing propose*predict-no
- -->
- (O1984 ^name predict-no +)
- (S1 ^operator O1984 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1982 = 0.4476195574206818)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1981 = 0.1844110446262441)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1982 ^name predict-no +)
- (S1 ^operator O1982 +)
- Retracting propose*predict-yes
- -->
- (O1981 ^name predict-yes +)
- (S1 ^operator O1981 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R994 ^value 1 +)
- (R1 ^reward R994 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1982 = 0.3873369632550164)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1982 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1981 = 0.3895394312063116)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1981 = 0.6104613034971749)
- =>WM: (13900: S1 ^operator O1984 +)
- =>WM: (13899: S1 ^operator O1983 +)
- =>WM: (13898: I3 ^dir R)
- =>WM: (13897: O1984 ^name predict-no)
- =>WM: (13896: O1983 ^name predict-yes)
- =>WM: (13895: R995 ^value 1)
- =>WM: (13894: R1 ^reward R995)
- =>WM: (13893: I3 ^see 1)
- <=WM: (13884: S1 ^operator O1981 +)
- <=WM: (13886: S1 ^operator O1981)
- <=WM: (13885: S1 ^operator O1982 +)
- <=WM: (13883: I3 ^dir L)
- <=WM: (13879: R1 ^reward R994)
- <=WM: (13852: I3 ^see 0)
- <=WM: (13882: O1982 ^name predict-no)
- <=WM: (13881: O1981 ^name predict-yes)
- <=WM: (13880: R994 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1983 = 0.1844110446262441)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1983 = 0.8155802143556325)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1984 = 0.4476195574206818)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O1984 = -0.02155734064455064)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1982 = 0.4476195574206818)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O1982 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1981 = 0.1844110446262441)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1981 = 0.8155802143556325)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.389539 -> 0.711951 -0.322412 0.389539(R,m,v=1,0.890909,0.0977827)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*44 0.288049 0.322412 0.610461 -> 0.288049 0.322412 0.610461(R,m,v=1,1,0)
- =>WM: (13901: S1 ^operator O1983)
- 992: O: O1983 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N992 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N991 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13902: I3 ^predict-yes N992)
- <=WM: (13888: N991 ^status complete)
- <=WM: (13887: I3 ^predict-yes N991)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- /|--- Input Phase ---
- =>WM: (13906: I2 ^dir L)
- =>WM: (13905: I2 ^reward 1)
- =>WM: (13904: I2 ^see 1)
- =>WM: (13903: N992 ^status complete)
- <=WM: (13891: I2 ^dir R)
- <=WM: (13890: I2 ^reward 1)
- <=WM: (13889: I2 ^see 1)
- =>WM: (13907: I2 ^level-1 R1-root)
- <=WM: (13892: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O1983 = 0.6104592422684716)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O1984 = 0.2714993082286609)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R996 ^value 1 +)
- (R1 ^reward R996 +)
- Firing propose*predict-yes
- -->
- (O1985 ^name predict-yes +)
- (S1 ^operator O1985 +)
- Firing propose*predict-no
- -->
- (O1986 ^name predict-no +)
- (S1 ^operator O1986 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1984 = 0.3873369632550164)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1983 = 0.3895393210007886)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1984 ^name predict-no +)
- (S1 ^operator O1984 +)
- Retracting propose*predict-yes
- -->
- (O1983 ^name predict-yes +)
- (S1 ^operator O1983 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R995 ^value 1 +)
- (R1 ^reward R995 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O1984 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1984 = 0.4476195574206818)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1983 = 0.8155802143556325)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1983 = 0.1844110446262441)
- =>WM: (13914: S1 ^operator O1986 +)
- =>WM: (13913: S1 ^operator O1985 +)
- =>WM: (13912: I3 ^dir L)
- =>WM: (13911: O1986 ^name predict-no)
- =>WM: (13910: O1985 ^name predict-yes)
- =>WM: (13909: R996 ^value 1)
- =>WM: (13908: R1 ^reward R996)
- <=WM: (13899: S1 ^operator O1983 +)
- <=WM: (13901: S1 ^operator O1983)
- <=WM: (13900: S1 ^operator O1984 +)
- <=WM: (13898: I3 ^dir R)
- <=WM: (13894: R1 ^reward R995)
- <=WM: (13897: O1984 ^name predict-no)
- <=WM: (13896: O1983 ^name predict-yes)
- <=WM: (13895: R995 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1985 = 0.3895393210007886)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O1985 = 0.6104592422684716)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1986 = 0.3873369632550164)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O1986 = 0.2714993082286609)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1984 = 0.3873369632550164)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O1984 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1983 = 0.3895393210007886)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O1983 = 0.6104592422684716)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675414 -0.491003 0.184411 -> 0.675415 -0.491003 0.184412(R,m,v=1,0.89881,0.0914956)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*46 0.324575 0.491005 0.81558 -> 0.324577 0.491005 0.815582(R,m,v=1,1,0)
- =>WM: (13915: S1 ^operator O1985)
- 993: O: O1985 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N993 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N992 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13916: I3 ^predict-yes N993)
- <=WM: (13903: N992 ^status complete)
- <=WM: (13902: I3 ^predict-yes N992)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- \--- Input Phase ---
- =>WM: (13920: I2 ^dir R)
- =>WM: (13919: I2 ^reward 1)
- =>WM: (13918: I2 ^see 1)
- =>WM: (13917: N993 ^status complete)
- <=WM: (13906: I2 ^dir L)
- <=WM: (13905: I2 ^reward 1)
- <=WM: (13904: I2 ^see 1)
- =>WM: (13921: I2 ^level-1 L1-root)
- <=WM: (13907: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O1986 = -0.02155734064455064)
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1985 = 0.8155815255083509)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R997 ^value 1 +)
- (R1 ^reward R997 +)
- Firing propose*predict-yes
- -->
- (O1987 ^name predict-yes +)
- (S1 ^operator O1987 +)
- Firing propose*predict-no
- -->
- (O1988 ^name predict-no +)
- (S1 ^operator O1988 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1986 = 0.4476195574206818)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1985 = 0.1844123557789626)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1986 ^name predict-no +)
- (S1 ^operator O1986 +)
- Retracting propose*predict-yes
- -->
- (O1985 ^name predict-yes +)
- (S1 ^operator O1985 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R996 ^value 1 +)
- (R1 ^reward R996 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O1986 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1986 = 0.3873369632550164)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O1985 = 0.6104592422684716)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1985 = 0.3895393210007886)
- =>WM: (13928: S1 ^operator O1988 +)
- =>WM: (13927: S1 ^operator O1987 +)
- =>WM: (13926: I3 ^dir R)
- =>WM: (13925: O1988 ^name predict-no)
- =>WM: (13924: O1987 ^name predict-yes)
- =>WM: (13923: R997 ^value 1)
- =>WM: (13922: R1 ^reward R997)
- <=WM: (13913: S1 ^operator O1985 +)
- <=WM: (13915: S1 ^operator O1985)
- <=WM: (13914: S1 ^operator O1986 +)
- <=WM: (13912: I3 ^dir L)
- <=WM: (13908: R1 ^reward R996)
- <=WM: (13911: O1986 ^name predict-no)
- <=WM: (13910: O1985 ^name predict-yes)
- <=WM: (13909: R996 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1987 = 0.1844123557789626)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1987 = 0.8155815255083509)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1988 = 0.4476195574206818)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O1988 = -0.02155734064455064)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1986 = 0.4476195574206818)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O1986 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1985 = 0.1844123557789626)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1985 = 0.8155815255083509)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.389539 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.891566,0.0972618)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*36 0.288049 0.322411 0.610459 -> 0.288049 0.322411 0.610459(R,m,v=1,1,0)
- =>WM: (13929: S1 ^operator O1987)
- 994: O: O1987 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N994 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N993 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13930: I3 ^predict-yes N994)
- <=WM: (13917: N993 ^status complete)
- <=WM: (13916: I3 ^predict-yes N993)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (13934: I2 ^dir L)
- =>WM: (13933: I2 ^reward 1)
- =>WM: (13932: I2 ^see 1)
- =>WM: (13931: N994 ^status complete)
- <=WM: (13920: I2 ^dir R)
- <=WM: (13919: I2 ^reward 1)
- <=WM: (13918: I2 ^see 1)
- =>WM: (13935: I2 ^level-1 R1-root)
- <=WM: (13921: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O1987 = 0.6104594577780825)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O1988 = 0.2714993082286609)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R998 ^value 1 +)
- (R1 ^reward R998 +)
- Firing propose*predict-yes
- -->
- (O1989 ^name predict-yes +)
- (S1 ^operator O1989 +)
- Firing propose*predict-no
- -->
- (O1990 ^name predict-no +)
- (S1 ^operator O1990 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1988 = 0.3873369632550164)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1987 = 0.3895395365103996)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1988 ^name predict-no +)
- (S1 ^operator O1988 +)
- Retracting propose*predict-yes
- -->
- (O1987 ^name predict-yes +)
- (S1 ^operator O1987 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R997 ^value 1 +)
- (R1 ^reward R997 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O1988 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1988 = 0.4476195574206818)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1987 = 0.8155815255083509)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1987 = 0.1844123557789626)
- =>WM: (13942: S1 ^operator O1990 +)
- =>WM: (13941: S1 ^operator O1989 +)
- =>WM: (13940: I3 ^dir L)
- =>WM: (13939: O1990 ^name predict-no)
- =>WM: (13938: O1989 ^name predict-yes)
- =>WM: (13937: R998 ^value 1)
- =>WM: (13936: R1 ^reward R998)
- <=WM: (13927: S1 ^operator O1987 +)
- <=WM: (13929: S1 ^operator O1987)
- <=WM: (13928: S1 ^operator O1988 +)
- <=WM: (13926: I3 ^dir R)
- <=WM: (13922: R1 ^reward R997)
- <=WM: (13925: O1988 ^name predict-no)
- <=WM: (13924: O1987 ^name predict-yes)
- <=WM: (13923: R997 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1989 = 0.3895395365103996)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O1989 = 0.6104594577780825)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1990 = 0.3873369632550164)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O1990 = 0.2714993082286609)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1988 = 0.3873369632550164)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O1988 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1987 = 0.3895395365103996)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O1987 = 0.6104594577780825)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675415 -0.491003 0.184412 -> 0.675417 -0.491003 0.184413(R,m,v=1,0.899408,0.0910116)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*46 0.324577 0.491005 0.815582 -> 0.324578 0.491005 0.815582(R,m,v=1,1,0)
- =>WM: (13943: S1 ^operator O1989)
- 995: O: O1989 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N995 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N994 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13944: I3 ^predict-yes N995)
- <=WM: (13931: N994 ^status complete)
- <=WM: (13930: I3 ^predict-yes N994)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- \---- Input Phase ---
- =>WM: (13948: I2 ^dir L)
- =>WM: (13947: I2 ^reward 1)
- =>WM: (13946: I2 ^see 1)
- =>WM: (13945: N995 ^status complete)
- <=WM: (13934: I2 ^dir L)
- <=WM: (13933: I2 ^reward 1)
- <=WM: (13932: I2 ^see 1)
- =>WM: (13949: I2 ^level-1 L1-root)
- <=WM: (13935: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1990 = 0.6126627481603084)
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1989 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R999 ^value 1 +)
- (R1 ^reward R999 +)
- Firing propose*predict-yes
- -->
- (O1991 ^name predict-yes +)
- (S1 ^operator O1991 +)
- Firing propose*predict-no
- -->
- (O1992 ^name predict-no +)
- (S1 ^operator O1992 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1990 = 0.3873369632550164)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1989 = 0.3895395365103996)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1990 ^name predict-no +)
- (S1 ^operator O1990 +)
- Retracting propose*predict-yes
- -->
- (O1989 ^name predict-yes +)
- (S1 ^operator O1989 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R998 ^value 1 +)
- (R1 ^reward R998 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O1990 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1990 = 0.3873369632550164)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O1989 = 0.6104594577780825)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1989 = 0.3895395365103996)
- =>WM: (13955: S1 ^operator O1992 +)
- =>WM: (13954: S1 ^operator O1991 +)
- =>WM: (13953: O1992 ^name predict-no)
- =>WM: (13952: O1991 ^name predict-yes)
- =>WM: (13951: R999 ^value 1)
- =>WM: (13950: R1 ^reward R999)
- <=WM: (13941: S1 ^operator O1989 +)
- <=WM: (13943: S1 ^operator O1989)
- <=WM: (13942: S1 ^operator O1990 +)
- <=WM: (13936: R1 ^reward R998)
- <=WM: (13939: O1990 ^name predict-no)
- <=WM: (13938: O1989 ^name predict-yes)
- <=WM: (13937: R998 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1991 = 0.3895395365103996)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1991 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1992 = 0.3873369632550164)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1992 = 0.6126627481603084)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1990 = 0.3873369632550164)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1990 = 0.6126627481603084)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1989 = 0.3895395365103996)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1989 = -0.02274740735326741)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.892216,0.0967463)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*36 0.288049 0.322411 0.610459 -> 0.288049 0.322411 0.61046(R,m,v=1,1,0)
- =>WM: (13956: S1 ^operator O1992)
- 996: O: O1992 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N996 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N995 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13957: I3 ^predict-no N996)
- <=WM: (13945: N995 ^status complete)
- <=WM: (13944: I3 ^predict-yes N995)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- /|\---- Input Phase ---
- =>WM: (13961: I2 ^dir R)
- =>WM: (13960: I2 ^reward 1)
- =>WM: (13959: I2 ^see 0)
- =>WM: (13958: N996 ^status complete)
- <=WM: (13948: I2 ^dir L)
- <=WM: (13947: I2 ^reward 1)
- <=WM: (13946: I2 ^see 1)
- =>WM: (13962: I2 ^level-1 L0-root)
- <=WM: (13949: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1991 = 0.8155947374398671)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1992 = -0.00558448899823713)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1000 ^value 1 +)
- (R1 ^reward R1000 +)
- Firing propose*predict-yes
- -->
- (O1993 ^name predict-yes +)
- (S1 ^operator O1993 +)
- Firing propose*predict-no
- -->
- (O1994 ^name predict-no +)
- (S1 ^operator O1994 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1992 = 0.4476195574206818)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1991 = 0.1844132735858656)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1992 ^name predict-no +)
- (S1 ^operator O1992 +)
- Retracting propose*predict-yes
- -->
- (O1991 ^name predict-yes +)
- (S1 ^operator O1991 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R999 ^value 1 +)
- (R1 ^reward R999 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O1992 = 0.6126627481603084)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1992 = 0.3873369632550164)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O1991 = -0.02274740735326741)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1991 = 0.3895396873671274)
- =>WM: (13970: S1 ^operator O1994 +)
- =>WM: (13969: S1 ^operator O1993 +)
- =>WM: (13968: I3 ^dir R)
- =>WM: (13967: O1994 ^name predict-no)
- =>WM: (13966: O1993 ^name predict-yes)
- =>WM: (13965: R1000 ^value 1)
- =>WM: (13964: R1 ^reward R1000)
- =>WM: (13963: I3 ^see 0)
- <=WM: (13954: S1 ^operator O1991 +)
- <=WM: (13955: S1 ^operator O1992 +)
- <=WM: (13956: S1 ^operator O1992)
- <=WM: (13940: I3 ^dir L)
- <=WM: (13950: R1 ^reward R999)
- <=WM: (13893: I3 ^see 1)
- <=WM: (13953: O1992 ^name predict-no)
- <=WM: (13952: O1991 ^name predict-yes)
- <=WM: (13951: R999 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1993 = 0.1844132735858656)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1993 = 0.8155947374398671)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1994 = 0.4476195574206818)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1994 = -0.00558448899823713)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1992 = 0.4476195574206818)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1992 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1991 = 0.1844132735858656)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1991 = 0.8155947374398671)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.719081 -0.331744 0.387337 -> 0.719081 -0.331744 0.387337(R,m,v=1,0.931818,0.0638961)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*32 0.280919 0.331744 0.612663 -> 0.280919 0.331744 0.612663(R,m,v=1,1,0)
- =>WM: (13971: S1 ^operator O1993)
- 997: O: O1993 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N997 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N996 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13972: I3 ^predict-yes N997)
- <=WM: (13958: N996 ^status complete)
- <=WM: (13957: I3 ^predict-no N996)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (13976: I2 ^dir R)
- =>WM: (13975: I2 ^reward 1)
- =>WM: (13974: I2 ^see 1)
- =>WM: (13973: N997 ^status complete)
- <=WM: (13961: I2 ^dir R)
- <=WM: (13960: I2 ^reward 1)
- <=WM: (13959: I2 ^see 0)
- =>WM: (13977: I2 ^level-1 R1-root)
- <=WM: (13962: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1993 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1994 = 0.5523820607022403)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1001 ^value 1 +)
- (R1 ^reward R1001 +)
- Firing propose*predict-yes
- -->
- (O1995 ^name predict-yes +)
- (S1 ^operator O1995 +)
- Firing propose*predict-no
- -->
- (O1996 ^name predict-no +)
- (S1 ^operator O1996 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1994 = 0.4476195574206818)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1993 = 0.1844132735858656)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1994 ^name predict-no +)
- (S1 ^operator O1994 +)
- Retracting propose*predict-yes
- -->
- (O1993 ^name predict-yes +)
- (S1 ^operator O1993 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1000 ^value 1 +)
- (R1 ^reward R1000 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O1994 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1994 = 0.4476195574206818)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O1993 = 0.8155947374398671)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1993 = 0.1844132735858656)
- =>WM: (13984: S1 ^operator O1996 +)
- =>WM: (13983: S1 ^operator O1995 +)
- =>WM: (13982: O1996 ^name predict-no)
- =>WM: (13981: O1995 ^name predict-yes)
- =>WM: (13980: R1001 ^value 1)
- =>WM: (13979: R1 ^reward R1001)
- =>WM: (13978: I3 ^see 1)
- <=WM: (13969: S1 ^operator O1993 +)
- <=WM: (13971: S1 ^operator O1993)
- <=WM: (13970: S1 ^operator O1994 +)
- <=WM: (13964: R1 ^reward R1000)
- <=WM: (13963: I3 ^see 0)
- <=WM: (13967: O1994 ^name predict-no)
- <=WM: (13966: O1993 ^name predict-yes)
- <=WM: (13965: R1000 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1995 = 0.1844132735858656)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1995 = 0.1398795999120246)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1996 = 0.4476195574206818)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1996 = 0.5523820607022403)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1994 = 0.4476195574206818)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1994 = 0.5523820607022403)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1993 = 0.1844132735858656)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1993 = 0.1398795999120246)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675417 -0.491003 0.184413 -> 0.675415 -0.491003 0.184412(R,m,v=1,0.9,0.0905325)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*34 0.324594 0.491001 0.815595 -> 0.324592 0.491001 0.815594(R,m,v=1,1,0)
- =>WM: (13985: S1 ^operator O1996)
- 998: O: O1996 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N998 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N997 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (13986: I3 ^predict-no N998)
- <=WM: (13973: N997 ^status complete)
- <=WM: (13972: I3 ^predict-yes N997)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (13990: I2 ^dir L)
- =>WM: (13989: I2 ^reward 1)
- =>WM: (13988: I2 ^see 0)
- =>WM: (13987: N998 ^status complete)
- <=WM: (13976: I2 ^dir R)
- <=WM: (13975: I2 ^reward 1)
- <=WM: (13974: I2 ^see 1)
- =>WM: (13991: I2 ^level-1 R0-root)
- <=WM: (13977: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1995 = 0.6104611932916519)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1996 = 0.1063475139796038)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1002 ^value 1 +)
- (R1 ^reward R1002 +)
- Firing propose*predict-yes
- -->
- (O1997 ^name predict-yes +)
- (S1 ^operator O1997 +)
- Firing propose*predict-no
- -->
- (O1998 ^name predict-no +)
- (S1 ^operator O1998 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1996 = 0.3873370065427176)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1995 = 0.3895396873671274)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O1996 ^name predict-no +)
- (S1 ^operator O1996 +)
- Retracting propose*predict-yes
- -->
- (O1995 ^name predict-yes +)
- (S1 ^operator O1995 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1001 ^value 1 +)
- (R1 ^reward R1001 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O1996 = 0.5523820607022403)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1996 = 0.4476195574206818)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O1995 = 0.1398795999120246)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1995 = 0.1844120719320057)
- =>WM: (13999: S1 ^operator O1998 +)
- =>WM: (13998: S1 ^operator O1997 +)
- =>WM: (13997: I3 ^dir L)
- =>WM: (13996: O1998 ^name predict-no)
- =>WM: (13995: O1997 ^name predict-yes)
- =>WM: (13994: R1002 ^value 1)
- =>WM: (13993: R1 ^reward R1002)
- =>WM: (13992: I3 ^see 0)
- <=WM: (13983: S1 ^operator O1995 +)
- <=WM: (13984: S1 ^operator O1996 +)
- <=WM: (13985: S1 ^operator O1996)
- <=WM: (13968: I3 ^dir R)
- <=WM: (13979: R1 ^reward R1001)
- <=WM: (13978: I3 ^see 1)
- <=WM: (13982: O1996 ^name predict-no)
- <=WM: (13981: O1995 ^name predict-yes)
- <=WM: (13980: R1001 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1997 = 0.3895396873671274)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1997 = 0.6104611932916519)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1998 = 0.3873370065427176)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1998 = 0.1063475139796038)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1996 = 0.3873370065427176)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1996 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1995 = 0.3895396873671274)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1995 = 0.6104611932916519)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174913 0.44762 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.930233,0.065407)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*39 0.377468 0.174914 0.552382 -> 0.377468 0.174914 0.552382(R,m,v=1,1,0)
- =>WM: (14000: S1 ^operator O1997)
- 999: O: O1997 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N999 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N998 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14001: I3 ^predict-yes N999)
- <=WM: (13987: N998 ^status complete)
- <=WM: (13986: I3 ^predict-no N998)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (14005: I2 ^dir R)
- =>WM: (14004: I2 ^reward 1)
- =>WM: (14003: I2 ^see 1)
- =>WM: (14002: N999 ^status complete)
- <=WM: (13990: I2 ^dir L)
- <=WM: (13989: I2 ^reward 1)
- <=WM: (13988: I2 ^see 0)
- =>WM: (14006: I2 ^level-1 L1-root)
- <=WM: (13991: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O1998 = -0.02155734064455064)
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1997 = 0.815582443315254)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1003 ^value 1 +)
- (R1 ^reward R1003 +)
- Firing propose*predict-yes
- -->
- (O1999 ^name predict-yes +)
- (S1 ^operator O1999 +)
- Firing propose*predict-no
- -->
- (O2000 ^name predict-no +)
- (S1 ^operator O2000 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1998 = 0.4476193147022436)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1997 = 0.1844120719320057)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O1998 ^name predict-no +)
- (S1 ^operator O1998 +)
- Retracting propose*predict-yes
- -->
- (O1997 ^name predict-yes +)
- (S1 ^operator O1997 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1002 ^value 1 +)
- (R1 ^reward R1002 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O1998 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O1998 = 0.3873370065427176)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O1997 = 0.6104611932916519)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O1997 = 0.3895396873671274)
- =>WM: (14014: S1 ^operator O2000 +)
- =>WM: (14013: S1 ^operator O1999 +)
- =>WM: (14012: I3 ^dir R)
- =>WM: (14011: O2000 ^name predict-no)
- =>WM: (14010: O1999 ^name predict-yes)
- =>WM: (14009: R1003 ^value 1)
- =>WM: (14008: R1 ^reward R1003)
- =>WM: (14007: I3 ^see 1)
- <=WM: (13998: S1 ^operator O1997 +)
- <=WM: (14000: S1 ^operator O1997)
- <=WM: (13999: S1 ^operator O1998 +)
- <=WM: (13997: I3 ^dir L)
- <=WM: (13993: R1 ^reward R1002)
- <=WM: (13992: I3 ^see 0)
- <=WM: (13996: O1998 ^name predict-no)
- <=WM: (13995: O1997 ^name predict-yes)
- <=WM: (13994: R1002 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1999 = 0.1844120719320057)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1999 = 0.815582443315254)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2000 = 0.4476193147022436)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2000 = -0.02155734064455064)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O1998 = 0.4476193147022436)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O1998 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1997 = 0.1844120719320057)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1997 = 0.815582443315254)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.892857,0.0962361)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*44 0.288049 0.322412 0.610461 -> 0.288049 0.322412 0.610461(R,m,v=1,1,0)
- =>WM: (14015: S1 ^operator O1999)
- 1000: O: O1999 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1000 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N999 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14016: I3 ^predict-yes N1000)
- <=WM: (14002: N999 ^status complete)
- <=WM: (14001: I3 ^predict-yes N999)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- |\-/|\-/|\---- Input Phase ---
- =>WM: (14020: I2 ^dir U)
- =>WM: (14019: I2 ^reward 1)
- =>WM: (14018: I2 ^see 1)
- =>WM: (14017: N1000 ^status complete)
- <=WM: (14005: I2 ^dir R)
- <=WM: (14004: I2 ^reward 1)
- <=WM: (14003: I2 ^see 1)
- =>WM: (14021: I2 ^level-1 R1-root)
- <=WM: (14006: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1004 ^value 1 +)
- (R1 ^reward R1004 +)
- Firing propose*predict-yes
- -->
- (O2001 ^name predict-yes +)
- (S1 ^operator O2001 +)
- Firing propose*predict-no
- -->
- (O2002 ^name predict-no +)
- (S1 ^operator O2002 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2000 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1999 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2000 ^name predict-no +)
- (S1 ^operator O2000 +)
- Retracting propose*predict-yes
- -->
- (O1999 ^name predict-yes +)
- (S1 ^operator O1999 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1003 ^value 1 +)
- (R1 ^reward R1003 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2000 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2000 = 0.4476193147022436)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O1999 = 0.815582443315254)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O1999 = 0.1844120719320057)
- =>WM: (14028: S1 ^operator O2002 +)
- =>WM: (14027: S1 ^operator O2001 +)
- =>WM: (14026: I3 ^dir U)
- =>WM: (14025: O2002 ^name predict-no)
- =>WM: (14024: O2001 ^name predict-yes)
- =>WM: (14023: R1004 ^value 1)
- =>WM: (14022: R1 ^reward R1004)
- <=WM: (14013: S1 ^operator O1999 +)
- <=WM: (14015: S1 ^operator O1999)
- <=WM: (14014: S1 ^operator O2000 +)
- <=WM: (14012: I3 ^dir R)
- <=WM: (14008: R1 ^reward R1003)
- <=WM: (14011: O2000 ^name predict-no)
- <=WM: (14010: O1999 ^name predict-yes)
- <=WM: (14009: R1003 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2001 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2002 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2000 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O1999 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675415 -0.491003 0.184412 -> 0.675416 -0.491003 0.184413(R,m,v=1,0.900585,0.0900585)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*46 0.324578 0.491005 0.815582 -> 0.324579 0.491004 0.815583(R,m,v=1,1,0)
- =>WM: (14029: S1 ^operator O2002)
- 1001: O: O2002 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1001 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1000 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14030: I3 ^predict-no N1001)
- <=WM: (14017: N1000 ^status complete)
- <=WM: (14016: I3 ^predict-yes N1000)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- /--- Input Phase ---
- =>WM: (14034: I2 ^dir L)
- =>WM: (14033: I2 ^reward 1)
- =>WM: (14032: I2 ^see 0)
- =>WM: (14031: N1001 ^status complete)
- <=WM: (14020: I2 ^dir U)
- <=WM: (14019: I2 ^reward 1)
- <=WM: (14018: I2 ^see 1)
- =>WM: (14035: I2 ^level-1 R1-root)
- <=WM: (14021: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2001 = 0.6104596086348102)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2002 = 0.2714993082286609)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1005 ^value 1 +)
- (R1 ^reward R1005 +)
- Firing propose*predict-yes
- -->
- (O2003 ^name predict-yes +)
- (S1 ^operator O2003 +)
- Firing propose*predict-no
- -->
- (O2004 ^name predict-no +)
- (S1 ^operator O2004 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2002 = 0.3873370065427176)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2001 = 0.3895395552683104)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2002 ^name predict-no +)
- (S1 ^operator O2002 +)
- Retracting propose*predict-yes
- -->
- (O2001 ^name predict-yes +)
- (S1 ^operator O2001 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1004 ^value 1 +)
- (R1 ^reward R1004 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2002 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2001 = 0.)
- =>WM: (14043: S1 ^operator O2004 +)
- =>WM: (14042: S1 ^operator O2003 +)
- =>WM: (14041: I3 ^dir L)
- =>WM: (14040: O2004 ^name predict-no)
- =>WM: (14039: O2003 ^name predict-yes)
- =>WM: (14038: R1005 ^value 1)
- =>WM: (14037: R1 ^reward R1005)
- =>WM: (14036: I3 ^see 0)
- <=WM: (14027: S1 ^operator O2001 +)
- <=WM: (14028: S1 ^operator O2002 +)
- <=WM: (14029: S1 ^operator O2002)
- <=WM: (14026: I3 ^dir U)
- <=WM: (14022: R1 ^reward R1004)
- <=WM: (14007: I3 ^see 1)
- <=WM: (14025: O2002 ^name predict-no)
- <=WM: (14024: O2001 ^name predict-yes)
- <=WM: (14023: R1004 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2003 = 0.6104596086348102)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2003 = 0.3895395552683104)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2004 = 0.2714993082286609)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2004 = 0.3873370065427176)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2002 = 0.3873370065427176)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2002 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2001 = 0.3895395552683104)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2001 = 0.6104596086348102)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14044: S1 ^operator O2003)
- 1002: O: O2003 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1002 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1001 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14045: I3 ^predict-yes N1002)
- <=WM: (14031: N1001 ^status complete)
- <=WM: (14030: I3 ^predict-no N1001)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- |\-/--- Input Phase ---
- =>WM: (14049: I2 ^dir L)
- =>WM: (14048: I2 ^reward 1)
- =>WM: (14047: I2 ^see 1)
- =>WM: (14046: N1002 ^status complete)
- <=WM: (14034: I2 ^dir L)
- <=WM: (14033: I2 ^reward 1)
- <=WM: (14032: I2 ^see 0)
- =>WM: (14050: I2 ^level-1 L1-root)
- <=WM: (14035: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2004 = 0.6126627914480096)
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2003 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1006 ^value 1 +)
- (R1 ^reward R1006 +)
- Firing propose*predict-yes
- -->
- (O2005 ^name predict-yes +)
- (S1 ^operator O2005 +)
- Firing propose*predict-no
- -->
- (O2006 ^name predict-no +)
- (S1 ^operator O2006 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2004 = 0.3873370065427176)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2003 = 0.3895395552683104)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2004 ^name predict-no +)
- (S1 ^operator O2004 +)
- Retracting propose*predict-yes
- -->
- (O2003 ^name predict-yes +)
- (S1 ^operator O2003 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1005 ^value 1 +)
- (R1 ^reward R1005 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2004 = 0.3873370065427176)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2004 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2003 = 0.3895395552683104)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2003 = 0.6104596086348102)
- =>WM: (14057: S1 ^operator O2006 +)
- =>WM: (14056: S1 ^operator O2005 +)
- =>WM: (14055: O2006 ^name predict-no)
- =>WM: (14054: O2005 ^name predict-yes)
- =>WM: (14053: R1006 ^value 1)
- =>WM: (14052: R1 ^reward R1006)
- =>WM: (14051: I3 ^see 1)
- <=WM: (14042: S1 ^operator O2003 +)
- <=WM: (14044: S1 ^operator O2003)
- <=WM: (14043: S1 ^operator O2004 +)
- <=WM: (14037: R1 ^reward R1005)
- <=WM: (14036: I3 ^see 0)
- <=WM: (14040: O2004 ^name predict-no)
- <=WM: (14039: O2003 ^name predict-yes)
- <=WM: (14038: R1005 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2005 = 0.3895395552683104)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2005 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2006 = 0.3873370065427176)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2006 = 0.6126627914480096)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2004 = 0.3873370065427176)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2004 = 0.6126627914480096)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2003 = 0.3895395552683104)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2003 = -0.02274740735326741)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.893491,0.0957312)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*36 0.288049 0.322411 0.61046 -> 0.288049 0.322411 0.61046(R,m,v=1,1,0)
- =>WM: (14058: S1 ^operator O2006)
- 1003: O: O2006 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1003 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1002 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14059: I3 ^predict-no N1003)
- <=WM: (14046: N1002 ^status complete)
- <=WM: (14045: I3 ^predict-yes N1002)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (14063: I2 ^dir R)
- =>WM: (14062: I2 ^reward 1)
- =>WM: (14061: I2 ^see 0)
- =>WM: (14060: N1003 ^status complete)
- <=WM: (14049: I2 ^dir L)
- <=WM: (14048: I2 ^reward 1)
- <=WM: (14047: I2 ^see 1)
- =>WM: (14064: I2 ^level-1 L0-root)
- <=WM: (14050: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2005 = 0.8155935357860071)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2006 = -0.00558448899823713)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1007 ^value 1 +)
- (R1 ^reward R1007 +)
- Firing propose*predict-yes
- -->
- (O2007 ^name predict-yes +)
- (S1 ^operator O2007 +)
- Firing propose*predict-no
- -->
- (O2008 ^name predict-no +)
- (S1 ^operator O2008 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2006 = 0.4476193147022436)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2005 = 0.1844128946449167)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2006 ^name predict-no +)
- (S1 ^operator O2006 +)
- Retracting propose*predict-yes
- -->
- (O2005 ^name predict-yes +)
- (S1 ^operator O2005 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1006 ^value 1 +)
- (R1 ^reward R1006 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2006 = 0.6126627914480096)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2006 = 0.3873370065427176)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2005 = -0.02274740735326741)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2005 = 0.3895396806828423)
- =>WM: (14072: S1 ^operator O2008 +)
- =>WM: (14071: S1 ^operator O2007 +)
- =>WM: (14070: I3 ^dir R)
- =>WM: (14069: O2008 ^name predict-no)
- =>WM: (14068: O2007 ^name predict-yes)
- =>WM: (14067: R1007 ^value 1)
- =>WM: (14066: R1 ^reward R1007)
- =>WM: (14065: I3 ^see 0)
- <=WM: (14056: S1 ^operator O2005 +)
- <=WM: (14057: S1 ^operator O2006 +)
- <=WM: (14058: S1 ^operator O2006)
- <=WM: (14041: I3 ^dir L)
- <=WM: (14052: R1 ^reward R1006)
- <=WM: (14051: I3 ^see 1)
- <=WM: (14055: O2006 ^name predict-no)
- <=WM: (14054: O2005 ^name predict-yes)
- <=WM: (14053: R1006 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2007 = 0.1844128946449167)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2007 = 0.8155935357860071)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2008 = 0.4476193147022436)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2008 = -0.00558448899823713)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2006 = 0.4476193147022436)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2006 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2005 = 0.1844128946449167)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2005 = 0.8155935357860071)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.719081 -0.331744 0.387337 -> 0.719081 -0.331744 0.387337(R,m,v=1,0.932203,0.0635593)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*32 0.280919 0.331744 0.612663 -> 0.280919 0.331744 0.612663(R,m,v=1,1,0)
- =>WM: (14073: S1 ^operator O2007)
- 1004: O: O2007 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1004 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1003 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14074: I3 ^predict-yes N1004)
- <=WM: (14060: N1003 ^status complete)
- <=WM: (14059: I3 ^predict-no N1003)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- /|--- Input Phase ---
- =>WM: (14078: I2 ^dir U)
- =>WM: (14077: I2 ^reward 1)
- =>WM: (14076: I2 ^see 1)
- =>WM: (14075: N1004 ^status complete)
- <=WM: (14063: I2 ^dir R)
- <=WM: (14062: I2 ^reward 1)
- <=WM: (14061: I2 ^see 0)
- =>WM: (14079: I2 ^level-1 R1-root)
- <=WM: (14064: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1008 ^value 1 +)
- (R1 ^reward R1008 +)
- Firing propose*predict-yes
- -->
- (O2009 ^name predict-yes +)
- (S1 ^operator O2009 +)
- Firing propose*predict-no
- -->
- (O2010 ^name predict-no +)
- (S1 ^operator O2010 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2008 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2007 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2008 ^name predict-no +)
- (S1 ^operator O2008 +)
- Retracting propose*predict-yes
- -->
- (O2007 ^name predict-yes +)
- (S1 ^operator O2007 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1007 ^value 1 +)
- (R1 ^reward R1007 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2008 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2008 = 0.4476193147022436)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2007 = 0.8155935357860071)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2007 = 0.1844128946449167)
- =>WM: (14087: S1 ^operator O2010 +)
- =>WM: (14086: S1 ^operator O2009 +)
- =>WM: (14085: I3 ^dir U)
- =>WM: (14084: O2010 ^name predict-no)
- =>WM: (14083: O2009 ^name predict-yes)
- =>WM: (14082: R1008 ^value 1)
- =>WM: (14081: R1 ^reward R1008)
- =>WM: (14080: I3 ^see 1)
- <=WM: (14071: S1 ^operator O2007 +)
- <=WM: (14073: S1 ^operator O2007)
- <=WM: (14072: S1 ^operator O2008 +)
- <=WM: (14070: I3 ^dir R)
- <=WM: (14066: R1 ^reward R1007)
- <=WM: (14065: I3 ^see 0)
- <=WM: (14069: O2008 ^name predict-no)
- <=WM: (14068: O2007 ^name predict-yes)
- <=WM: (14067: R1007 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2009 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2010 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2008 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2007 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675416 -0.491003 0.184413 -> 0.675415 -0.491003 0.184412(R,m,v=1,0.901163,0.0895893)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*34 0.324592 0.491001 0.815594 -> 0.324591 0.491001 0.815593(R,m,v=1,1,0)
- =>WM: (14088: S1 ^operator O2010)
- 1005: O: O2010 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1005 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1004 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14089: I3 ^predict-no N1005)
- <=WM: (14075: N1004 ^status complete)
- <=WM: (14074: I3 ^predict-yes N1004)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- \-/|--- Input Phase ---
- =>WM: (14093: I2 ^dir L)
- =>WM: (14092: I2 ^reward 1)
- =>WM: (14091: I2 ^see 0)
- =>WM: (14090: N1005 ^status complete)
- <=WM: (14078: I2 ^dir U)
- <=WM: (14077: I2 ^reward 1)
- <=WM: (14076: I2 ^see 1)
- =>WM: (14094: I2 ^level-1 R1-root)
- <=WM: (14079: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2009 = 0.6104597340493421)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2010 = 0.2714993082286609)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1009 ^value 1 +)
- (R1 ^reward R1009 +)
- Firing propose*predict-yes
- -->
- (O2011 ^name predict-yes +)
- (S1 ^operator O2011 +)
- Firing propose*predict-no
- -->
- (O2012 ^name predict-no +)
- (S1 ^operator O2012 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2010 = 0.3873370368441085)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2009 = 0.3895396806828423)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2010 ^name predict-no +)
- (S1 ^operator O2010 +)
- Retracting propose*predict-yes
- -->
- (O2009 ^name predict-yes +)
- (S1 ^operator O2009 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1008 ^value 1 +)
- (R1 ^reward R1008 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2010 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2009 = 0.)
- =>WM: (14102: S1 ^operator O2012 +)
- =>WM: (14101: S1 ^operator O2011 +)
- =>WM: (14100: I3 ^dir L)
- =>WM: (14099: O2012 ^name predict-no)
- =>WM: (14098: O2011 ^name predict-yes)
- =>WM: (14097: R1009 ^value 1)
- =>WM: (14096: R1 ^reward R1009)
- =>WM: (14095: I3 ^see 0)
- <=WM: (14086: S1 ^operator O2009 +)
- <=WM: (14087: S1 ^operator O2010 +)
- <=WM: (14088: S1 ^operator O2010)
- <=WM: (14085: I3 ^dir U)
- <=WM: (14081: R1 ^reward R1008)
- <=WM: (14080: I3 ^see 1)
- <=WM: (14084: O2010 ^name predict-no)
- <=WM: (14083: O2009 ^name predict-yes)
- <=WM: (14082: R1008 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2011 = 0.6104597340493421)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2011 = 0.3895396806828423)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2012 = 0.2714993082286609)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2012 = 0.3873370368441085)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2010 = 0.3873370368441085)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2010 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2009 = 0.3895396806828423)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2009 = 0.6104597340493421)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14103: S1 ^operator O2011)
- 1006: O: O2011 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1006 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1005 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14104: I3 ^predict-yes N1006)
- <=WM: (14090: N1005 ^status complete)
- <=WM: (14089: I3 ^predict-no N1005)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (14108: I2 ^dir U)
- =>WM: (14107: I2 ^reward 1)
- =>WM: (14106: I2 ^see 1)
- =>WM: (14105: N1006 ^status complete)
- <=WM: (14093: I2 ^dir L)
- <=WM: (14092: I2 ^reward 1)
- <=WM: (14091: I2 ^see 0)
- =>WM: (14109: I2 ^level-1 L1-root)
- <=WM: (14094: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1010 ^value 1 +)
- (R1 ^reward R1010 +)
- Firing propose*predict-yes
- -->
- (O2013 ^name predict-yes +)
- (S1 ^operator O2013 +)
- Firing propose*predict-no
- -->
- (O2014 ^name predict-no +)
- (S1 ^operator O2014 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2012 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2011 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2012 ^name predict-no +)
- (S1 ^operator O2012 +)
- Retracting propose*predict-yes
- -->
- (O2011 ^name predict-yes +)
- (S1 ^operator O2011 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1009 ^value 1 +)
- (R1 ^reward R1009 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2012 = 0.3873370368441085)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2012 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2011 = 0.3895396806828423)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2011 = 0.6104597340493421)
- =>WM: (14117: S1 ^operator O2014 +)
- =>WM: (14116: S1 ^operator O2013 +)
- =>WM: (14115: I3 ^dir U)
- =>WM: (14114: O2014 ^name predict-no)
- =>WM: (14113: O2013 ^name predict-yes)
- =>WM: (14112: R1010 ^value 1)
- =>WM: (14111: R1 ^reward R1010)
- =>WM: (14110: I3 ^see 1)
- <=WM: (14101: S1 ^operator O2011 +)
- <=WM: (14103: S1 ^operator O2011)
- <=WM: (14102: S1 ^operator O2012 +)
- <=WM: (14100: I3 ^dir L)
- <=WM: (14096: R1 ^reward R1009)
- <=WM: (14095: I3 ^see 0)
- <=WM: (14099: O2012 ^name predict-no)
- <=WM: (14098: O2011 ^name predict-yes)
- <=WM: (14097: R1009 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2013 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2014 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2012 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2011 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.894118,0.0952315)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*36 0.288049 0.322411 0.61046 -> 0.288049 0.322411 0.61046(R,m,v=1,1,0)
- =>WM: (14118: S1 ^operator O2014)
- 1007: O: O2014 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1007 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1006 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14119: I3 ^predict-no N1007)
- <=WM: (14105: N1006 ^status complete)
- <=WM: (14104: I3 ^predict-yes N1006)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (14123: I2 ^dir L)
- =>WM: (14122: I2 ^reward 1)
- =>WM: (14121: I2 ^see 0)
- =>WM: (14120: N1007 ^status complete)
- <=WM: (14108: I2 ^dir U)
- <=WM: (14107: I2 ^reward 1)
- <=WM: (14106: I2 ^see 1)
- =>WM: (14124: I2 ^level-1 L1-root)
- <=WM: (14109: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2014 = 0.6126628217494006)
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2013 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1011 ^value 1 +)
- (R1 ^reward R1011 +)
- Firing propose*predict-yes
- -->
- (O2015 ^name predict-yes +)
- (S1 ^operator O2015 +)
- Firing propose*predict-no
- -->
- (O2016 ^name predict-no +)
- (S1 ^operator O2016 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2014 = 0.3873370368441085)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2013 = 0.3895397684730147)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2014 ^name predict-no +)
- (S1 ^operator O2014 +)
- Retracting propose*predict-yes
- -->
- (O2013 ^name predict-yes +)
- (S1 ^operator O2013 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1010 ^value 1 +)
- (R1 ^reward R1010 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2014 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2013 = 0.)
- =>WM: (14132: S1 ^operator O2016 +)
- =>WM: (14131: S1 ^operator O2015 +)
- =>WM: (14130: I3 ^dir L)
- =>WM: (14129: O2016 ^name predict-no)
- =>WM: (14128: O2015 ^name predict-yes)
- =>WM: (14127: R1011 ^value 1)
- =>WM: (14126: R1 ^reward R1011)
- =>WM: (14125: I3 ^see 0)
- <=WM: (14116: S1 ^operator O2013 +)
- <=WM: (14117: S1 ^operator O2014 +)
- <=WM: (14118: S1 ^operator O2014)
- <=WM: (14115: I3 ^dir U)
- <=WM: (14111: R1 ^reward R1010)
- <=WM: (14110: I3 ^see 1)
- <=WM: (14114: O2014 ^name predict-no)
- <=WM: (14113: O2013 ^name predict-yes)
- <=WM: (14112: R1010 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2015 = -0.02274740735326741)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2015 = 0.3895397684730147)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2016 = 0.6126628217494006)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2016 = 0.3873370368441085)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2014 = 0.3873370368441085)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2014 = 0.6126628217494006)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2013 = 0.3895397684730147)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2013 = -0.02274740735326741)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14133: S1 ^operator O2016)
- 1008: O: O2016 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1008 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1007 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14134: I3 ^predict-no N1008)
- <=WM: (14120: N1007 ^status complete)
- <=WM: (14119: I3 ^predict-no N1007)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (14138: I2 ^dir U)
- =>WM: (14137: I2 ^reward 1)
- =>WM: (14136: I2 ^see 0)
- =>WM: (14135: N1008 ^status complete)
- <=WM: (14123: I2 ^dir L)
- <=WM: (14122: I2 ^reward 1)
- <=WM: (14121: I2 ^see 0)
- =>WM: (14139: I2 ^level-1 L0-root)
- <=WM: (14124: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1012 ^value 1 +)
- (R1 ^reward R1012 +)
- Firing propose*predict-yes
- -->
- (O2017 ^name predict-yes +)
- (S1 ^operator O2017 +)
- Firing propose*predict-no
- -->
- (O2018 ^name predict-no +)
- (S1 ^operator O2018 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2016 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2015 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2016 ^name predict-no +)
- (S1 ^operator O2016 +)
- Retracting propose*predict-yes
- -->
- (O2015 ^name predict-yes +)
- (S1 ^operator O2015 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1011 ^value 1 +)
- (R1 ^reward R1011 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2016 = 0.3873370368441085)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2016 = 0.6126628217494006)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2015 = 0.3895397684730147)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2015 = -0.02274740735326741)
- =>WM: (14146: S1 ^operator O2018 +)
- =>WM: (14145: S1 ^operator O2017 +)
- =>WM: (14144: I3 ^dir U)
- =>WM: (14143: O2018 ^name predict-no)
- =>WM: (14142: O2017 ^name predict-yes)
- =>WM: (14141: R1012 ^value 1)
- =>WM: (14140: R1 ^reward R1012)
- <=WM: (14131: S1 ^operator O2015 +)
- <=WM: (14132: S1 ^operator O2016 +)
- <=WM: (14133: S1 ^operator O2016)
- <=WM: (14130: I3 ^dir L)
- <=WM: (14126: R1 ^reward R1011)
- <=WM: (14129: O2016 ^name predict-no)
- <=WM: (14128: O2015 ^name predict-yes)
- <=WM: (14127: R1011 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2017 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2018 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2016 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2015 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.719081 -0.331744 0.387337 -> 0.719081 -0.331744 0.387337(R,m,v=1,0.932584,0.0632261)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*32 0.280919 0.331744 0.612663 -> 0.280919 0.331744 0.612663(R,m,v=1,1,0)
- =>WM: (14147: S1 ^operator O2018)
- 1009: O: O2018 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1009 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1008 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14148: I3 ^predict-no N1009)
- <=WM: (14135: N1008 ^status complete)
- <=WM: (14134: I3 ^predict-no N1008)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- -/|\--- Input Phase ---
- =>WM: (14152: I2 ^dir L)
- =>WM: (14151: I2 ^reward 1)
- =>WM: (14150: I2 ^see 0)
- =>WM: (14149: N1009 ^status complete)
- <=WM: (14138: I2 ^dir U)
- <=WM: (14137: I2 ^reward 1)
- <=WM: (14136: I2 ^see 0)
- =>WM: (14153: I2 ^level-1 L0-root)
- <=WM: (14139: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2017 = 0.1599599085218832)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2018 = 0.6126679931585133)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1013 ^value 1 +)
- (R1 ^reward R1013 +)
- Firing propose*predict-yes
- -->
- (O2019 ^name predict-yes +)
- (S1 ^operator O2019 +)
- Firing propose*predict-no
- -->
- (O2020 ^name predict-no +)
- (S1 ^operator O2020 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2018 = 0.3873370580550821)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2017 = 0.3895397684730147)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2018 ^name predict-no +)
- (S1 ^operator O2018 +)
- Retracting propose*predict-yes
- -->
- (O2017 ^name predict-yes +)
- (S1 ^operator O2017 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1012 ^value 1 +)
- (R1 ^reward R1012 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2018 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2017 = 0.)
- =>WM: (14160: S1 ^operator O2020 +)
- =>WM: (14159: S1 ^operator O2019 +)
- =>WM: (14158: I3 ^dir L)
- =>WM: (14157: O2020 ^name predict-no)
- =>WM: (14156: O2019 ^name predict-yes)
- =>WM: (14155: R1013 ^value 1)
- =>WM: (14154: R1 ^reward R1013)
- <=WM: (14145: S1 ^operator O2017 +)
- <=WM: (14146: S1 ^operator O2018 +)
- <=WM: (14147: S1 ^operator O2018)
- <=WM: (14144: I3 ^dir U)
- <=WM: (14140: R1 ^reward R1012)
- <=WM: (14143: O2018 ^name predict-no)
- <=WM: (14142: O2017 ^name predict-yes)
- <=WM: (14141: R1012 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2019 = 0.1599599085218832)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2019 = 0.3895397684730147)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2020 = 0.6126679931585133)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2020 = 0.3873370580550821)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2018 = 0.3873370580550821)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2018 = 0.6126679931585133)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2017 = 0.3895397684730147)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2017 = 0.1599599085218832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14161: S1 ^operator O2020)
- 1010: O: O2020 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1010 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1009 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14162: I3 ^predict-no N1010)
- <=WM: (14149: N1009 ^status complete)
- <=WM: (14148: I3 ^predict-no N1009)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (14166: I2 ^dir U)
- =>WM: (14165: I2 ^reward 1)
- =>WM: (14164: I2 ^see 0)
- =>WM: (14163: N1010 ^status complete)
- <=WM: (14152: I2 ^dir L)
- <=WM: (14151: I2 ^reward 1)
- <=WM: (14150: I2 ^see 0)
- =>WM: (14167: I2 ^level-1 L0-root)
- <=WM: (14153: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1014 ^value 1 +)
- (R1 ^reward R1014 +)
- Firing propose*predict-yes
- -->
- (O2021 ^name predict-yes +)
- (S1 ^operator O2021 +)
- Firing propose*predict-no
- -->
- (O2022 ^name predict-no +)
- (S1 ^operator O2022 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2020 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2019 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2020 ^name predict-no +)
- (S1 ^operator O2020 +)
- Retracting propose*predict-yes
- -->
- (O2019 ^name predict-yes +)
- (S1 ^operator O2019 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1013 ^value 1 +)
- (R1 ^reward R1013 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2020 = 0.3873370580550821)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2020 = 0.6126679931585133)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2019 = 0.3895397684730147)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2019 = 0.1599599085218832)
- =>WM: (14174: S1 ^operator O2022 +)
- =>WM: (14173: S1 ^operator O2021 +)
- =>WM: (14172: I3 ^dir U)
- =>WM: (14171: O2022 ^name predict-no)
- =>WM: (14170: O2021 ^name predict-yes)
- =>WM: (14169: R1014 ^value 1)
- =>WM: (14168: R1 ^reward R1014)
- <=WM: (14159: S1 ^operator O2019 +)
- <=WM: (14160: S1 ^operator O2020 +)
- <=WM: (14161: S1 ^operator O2020)
- <=WM: (14158: I3 ^dir L)
- <=WM: (14154: R1 ^reward R1013)
- <=WM: (14157: O2020 ^name predict-no)
- <=WM: (14156: O2019 ^name predict-yes)
- <=WM: (14155: R1013 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2021 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2022 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2020 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2019 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.719081 -0.331744 0.387337 -> 0.71908 -0.331744 0.387336(R,m,v=1,0.932961,0.0628962)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*37 0.280926 0.331742 0.612668 -> 0.280925 0.331742 0.612667(R,m,v=1,1,0)
- =>WM: (14175: S1 ^operator O2022)
- 1011: O: O2022 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1011 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1010 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14176: I3 ^predict-no N1011)
- <=WM: (14163: N1010 ^status complete)
- <=WM: (14162: I3 ^predict-no N1010)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- \--- Input Phase ---
- =>WM: (14180: I2 ^dir L)
- =>WM: (14179: I2 ^reward 1)
- =>WM: (14178: I2 ^see 0)
- =>WM: (14177: N1011 ^status complete)
- <=WM: (14166: I2 ^dir U)
- <=WM: (14165: I2 ^reward 1)
- <=WM: (14164: I2 ^see 0)
- =>WM: (14181: I2 ^level-1 L0-root)
- <=WM: (14167: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2021 = 0.1599599085218832)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2022 = 0.6126672354764739)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1015 ^value 1 +)
- (R1 ^reward R1015 +)
- Firing propose*predict-yes
- -->
- (O2023 ^name predict-yes +)
- (S1 ^operator O2023 +)
- Firing propose*predict-no
- -->
- (O2024 ^name predict-no +)
- (S1 ^operator O2024 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2022 = 0.3873363003730427)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2021 = 0.3895397684730147)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2022 ^name predict-no +)
- (S1 ^operator O2022 +)
- Retracting propose*predict-yes
- -->
- (O2021 ^name predict-yes +)
- (S1 ^operator O2021 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1014 ^value 1 +)
- (R1 ^reward R1014 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2022 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2021 = 0.)
- =>WM: (14188: S1 ^operator O2024 +)
- =>WM: (14187: S1 ^operator O2023 +)
- =>WM: (14186: I3 ^dir L)
- =>WM: (14185: O2024 ^name predict-no)
- =>WM: (14184: O2023 ^name predict-yes)
- =>WM: (14183: R1015 ^value 1)
- =>WM: (14182: R1 ^reward R1015)
- <=WM: (14173: S1 ^operator O2021 +)
- <=WM: (14174: S1 ^operator O2022 +)
- <=WM: (14175: S1 ^operator O2022)
- <=WM: (14172: I3 ^dir U)
- <=WM: (14168: R1 ^reward R1014)
- <=WM: (14171: O2022 ^name predict-no)
- <=WM: (14170: O2021 ^name predict-yes)
- <=WM: (14169: R1014 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2023 = 0.1599599085218832)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2023 = 0.3895397684730147)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2024 = 0.6126672354764739)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2024 = 0.3873363003730427)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2022 = 0.3873363003730427)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2022 = 0.6126672354764739)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2021 = 0.3895397684730147)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2021 = 0.1599599085218832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14189: S1 ^operator O2024)
- 1012: O: O2024 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1012 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1011 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14190: I3 ^predict-no N1012)
- <=WM: (14177: N1011 ^status complete)
- <=WM: (14176: I3 ^predict-no N1011)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (14194: I2 ^dir R)
- =>WM: (14193: I2 ^reward 1)
- =>WM: (14192: I2 ^see 0)
- =>WM: (14191: N1012 ^status complete)
- <=WM: (14180: I2 ^dir L)
- <=WM: (14179: I2 ^reward 1)
- <=WM: (14178: I2 ^see 0)
- =>WM: (14195: I2 ^level-1 L0-root)
- <=WM: (14181: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2023 = 0.8155925712213685)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2024 = -0.00558448899823713)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1016 ^value 1 +)
- (R1 ^reward R1016 +)
- Firing propose*predict-yes
- -->
- (O2025 ^name predict-yes +)
- (S1 ^operator O2025 +)
- Firing propose*predict-no
- -->
- (O2026 ^name predict-no +)
- (S1 ^operator O2026 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2024 = 0.4476193147022436)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2023 = 0.1844119300802781)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2024 ^name predict-no +)
- (S1 ^operator O2024 +)
- Retracting propose*predict-yes
- -->
- (O2023 ^name predict-yes +)
- (S1 ^operator O2023 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1015 ^value 1 +)
- (R1 ^reward R1015 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2024 = 0.3873363003730427)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2024 = 0.6126672354764739)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2023 = 0.3895397684730147)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2023 = 0.1599599085218832)
- =>WM: (14202: S1 ^operator O2026 +)
- =>WM: (14201: S1 ^operator O2025 +)
- =>WM: (14200: I3 ^dir R)
- =>WM: (14199: O2026 ^name predict-no)
- =>WM: (14198: O2025 ^name predict-yes)
- =>WM: (14197: R1016 ^value 1)
- =>WM: (14196: R1 ^reward R1016)
- <=WM: (14187: S1 ^operator O2023 +)
- <=WM: (14188: S1 ^operator O2024 +)
- <=WM: (14189: S1 ^operator O2024)
- <=WM: (14186: I3 ^dir L)
- <=WM: (14182: R1 ^reward R1015)
- <=WM: (14185: O2024 ^name predict-no)
- <=WM: (14184: O2023 ^name predict-yes)
- <=WM: (14183: R1015 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2025 = 0.8155925712213685)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2025 = 0.1844119300802781)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2026 = -0.00558448899823713)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2026 = 0.4476193147022436)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2024 = 0.4476193147022436)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2024 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2023 = 0.1844119300802781)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2023 = 0.8155925712213685)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.71908 -0.331744 0.387336 -> 0.719079 -0.331743 0.387336(R,m,v=1,0.933333,0.0625698)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*37 0.280925 0.331742 0.612667 -> 0.280924 0.331742 0.612667(R,m,v=1,1,0)
- =>WM: (14203: S1 ^operator O2025)
- 1013: O: O2025 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1013 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1012 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14204: I3 ^predict-yes N1013)
- <=WM: (14191: N1012 ^status complete)
- <=WM: (14190: I3 ^predict-no N1012)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (14208: I2 ^dir L)
- =>WM: (14207: I2 ^reward 1)
- =>WM: (14206: I2 ^see 1)
- =>WM: (14205: N1013 ^status complete)
- <=WM: (14194: I2 ^dir R)
- <=WM: (14193: I2 ^reward 1)
- <=WM: (14192: I2 ^see 0)
- =>WM: (14209: I2 ^level-1 R1-root)
- <=WM: (14195: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2025 = 0.6104598218395145)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2026 = 0.2714993082286609)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1017 ^value 1 +)
- (R1 ^reward R1017 +)
- Firing propose*predict-yes
- -->
- (O2027 ^name predict-yes +)
- (S1 ^operator O2027 +)
- Firing propose*predict-no
- -->
- (O2028 ^name predict-no +)
- (S1 ^operator O2028 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2026 = 0.3873357699956153)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2025 = 0.3895397684730147)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2026 ^name predict-no +)
- (S1 ^operator O2026 +)
- Retracting propose*predict-yes
- -->
- (O2025 ^name predict-yes +)
- (S1 ^operator O2025 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1016 ^value 1 +)
- (R1 ^reward R1016 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2026 = 0.4476193147022436)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2026 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2025 = 0.1844119300802781)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2025 = 0.8155925712213685)
- =>WM: (14217: S1 ^operator O2028 +)
- =>WM: (14216: S1 ^operator O2027 +)
- =>WM: (14215: I3 ^dir L)
- =>WM: (14214: O2028 ^name predict-no)
- =>WM: (14213: O2027 ^name predict-yes)
- =>WM: (14212: R1017 ^value 1)
- =>WM: (14211: R1 ^reward R1017)
- =>WM: (14210: I3 ^see 1)
- <=WM: (14201: S1 ^operator O2025 +)
- <=WM: (14203: S1 ^operator O2025)
- <=WM: (14202: S1 ^operator O2026 +)
- <=WM: (14200: I3 ^dir R)
- <=WM: (14196: R1 ^reward R1016)
- <=WM: (14125: I3 ^see 0)
- <=WM: (14199: O2026 ^name predict-no)
- <=WM: (14198: O2025 ^name predict-yes)
- <=WM: (14197: R1016 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2027 = 0.3895397684730147)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2027 = 0.6104598218395145)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2028 = 0.3873357699956153)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2028 = 0.2714993082286609)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2026 = 0.3873357699956153)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2026 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2025 = 0.3895397684730147)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2025 = 0.6104598218395145)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675415 -0.491003 0.184412 -> 0.675414 -0.491003 0.184411(R,m,v=1,0.901734,0.0891249)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*34 0.324591 0.491001 0.815593 -> 0.32459 0.491002 0.815592(R,m,v=1,1,0)
- =>WM: (14218: S1 ^operator O2027)
- 1014: O: O2027 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1014 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1013 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14219: I3 ^predict-yes N1014)
- <=WM: (14205: N1013 ^status complete)
- <=WM: (14204: I3 ^predict-yes N1013)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (14223: I2 ^dir U)
- =>WM: (14222: I2 ^reward 1)
- =>WM: (14221: I2 ^see 1)
- =>WM: (14220: N1014 ^status complete)
- <=WM: (14208: I2 ^dir L)
- <=WM: (14207: I2 ^reward 1)
- <=WM: (14206: I2 ^see 1)
- =>WM: (14224: I2 ^level-1 L1-root)
- <=WM: (14209: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1018 ^value 1 +)
- (R1 ^reward R1018 +)
- Firing propose*predict-yes
- -->
- (O2029 ^name predict-yes +)
- (S1 ^operator O2029 +)
- Firing propose*predict-no
- -->
- (O2030 ^name predict-no +)
- (S1 ^operator O2030 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2028 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2027 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2028 ^name predict-no +)
- (S1 ^operator O2028 +)
- Retracting propose*predict-yes
- -->
- (O2027 ^name predict-yes +)
- (S1 ^operator O2027 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1017 ^value 1 +)
- (R1 ^reward R1017 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2028 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2028 = 0.3873357699956153)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2027 = 0.6104598218395145)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2027 = 0.3895397684730147)
- =>WM: (14231: S1 ^operator O2030 +)
- =>WM: (14230: S1 ^operator O2029 +)
- =>WM: (14229: I3 ^dir U)
- =>WM: (14228: O2030 ^name predict-no)
- =>WM: (14227: O2029 ^name predict-yes)
- =>WM: (14226: R1018 ^value 1)
- =>WM: (14225: R1 ^reward R1018)
- <=WM: (14216: S1 ^operator O2027 +)
- <=WM: (14218: S1 ^operator O2027)
- <=WM: (14217: S1 ^operator O2028 +)
- <=WM: (14215: I3 ^dir L)
- <=WM: (14211: R1 ^reward R1017)
- <=WM: (14214: O2028 ^name predict-no)
- <=WM: (14213: O2027 ^name predict-yes)
- <=WM: (14212: R1017 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2029 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2030 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2028 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2027 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322411 0.38954(R,m,v=1,0.894737,0.0947368)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*36 0.288049 0.322411 0.61046 -> 0.288049 0.322411 0.61046(R,m,v=1,1,0)
- =>WM: (14232: S1 ^operator O2030)
- 1015: O: O2030 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1015 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1014 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14233: I3 ^predict-no N1015)
- <=WM: (14220: N1014 ^status complete)
- <=WM: (14219: I3 ^predict-yes N1014)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (14237: I2 ^dir U)
- =>WM: (14236: I2 ^reward 1)
- =>WM: (14235: I2 ^see 0)
- =>WM: (14234: N1015 ^status complete)
- <=WM: (14223: I2 ^dir U)
- <=WM: (14222: I2 ^reward 1)
- <=WM: (14221: I2 ^see 1)
- =>WM: (14238: I2 ^level-1 L1-root)
- <=WM: (14224: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1019 ^value 1 +)
- (R1 ^reward R1019 +)
- Firing propose*predict-yes
- -->
- (O2031 ^name predict-yes +)
- (S1 ^operator O2031 +)
- Firing propose*predict-no
- -->
- (O2032 ^name predict-no +)
- (S1 ^operator O2032 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2030 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2029 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2030 ^name predict-no +)
- (S1 ^operator O2030 +)
- Retracting propose*predict-yes
- -->
- (O2029 ^name predict-yes +)
- (S1 ^operator O2029 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1018 ^value 1 +)
- (R1 ^reward R1018 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2030 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2029 = 0.)
- =>WM: (14245: S1 ^operator O2032 +)
- =>WM: (14244: S1 ^operator O2031 +)
- =>WM: (14243: O2032 ^name predict-no)
- =>WM: (14242: O2031 ^name predict-yes)
- =>WM: (14241: R1019 ^value 1)
- =>WM: (14240: R1 ^reward R1019)
- =>WM: (14239: I3 ^see 0)
- <=WM: (14230: S1 ^operator O2029 +)
- <=WM: (14231: S1 ^operator O2030 +)
- <=WM: (14232: S1 ^operator O2030)
- <=WM: (14225: R1 ^reward R1018)
- <=WM: (14210: I3 ^see 1)
- <=WM: (14228: O2030 ^name predict-no)
- <=WM: (14227: O2029 ^name predict-yes)
- <=WM: (14226: R1018 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2031 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2032 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2030 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2029 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14246: S1 ^operator O2032)
- 1016: O: O2032 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1016 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1015 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14247: I3 ^predict-no N1016)
- <=WM: (14234: N1015 ^status complete)
- <=WM: (14233: I3 ^predict-no N1015)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (14251: I2 ^dir L)
- =>WM: (14250: I2 ^reward 1)
- =>WM: (14249: I2 ^see 0)
- =>WM: (14248: N1016 ^status complete)
- <=WM: (14237: I2 ^dir U)
- <=WM: (14236: I2 ^reward 1)
- <=WM: (14235: I2 ^see 0)
- =>WM: (14252: I2 ^level-1 L1-root)
- <=WM: (14238: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2032 = 0.6126628429603742)
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2031 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1020 ^value 1 +)
- (R1 ^reward R1020 +)
- Firing propose*predict-yes
- -->
- (O2033 ^name predict-yes +)
- (S1 ^operator O2033 +)
- Firing propose*predict-no
- -->
- (O2034 ^name predict-no +)
- (S1 ^operator O2034 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2032 = 0.3873357699956153)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2031 = 0.3895398299261354)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2032 ^name predict-no +)
- (S1 ^operator O2032 +)
- Retracting propose*predict-yes
- -->
- (O2031 ^name predict-yes +)
- (S1 ^operator O2031 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1019 ^value 1 +)
- (R1 ^reward R1019 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2032 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2031 = 0.)
- =>WM: (14259: S1 ^operator O2034 +)
- =>WM: (14258: S1 ^operator O2033 +)
- =>WM: (14257: I3 ^dir L)
- =>WM: (14256: O2034 ^name predict-no)
- =>WM: (14255: O2033 ^name predict-yes)
- =>WM: (14254: R1020 ^value 1)
- =>WM: (14253: R1 ^reward R1020)
- <=WM: (14244: S1 ^operator O2031 +)
- <=WM: (14245: S1 ^operator O2032 +)
- <=WM: (14246: S1 ^operator O2032)
- <=WM: (14229: I3 ^dir U)
- <=WM: (14240: R1 ^reward R1019)
- <=WM: (14243: O2032 ^name predict-no)
- <=WM: (14242: O2031 ^name predict-yes)
- <=WM: (14241: R1019 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2033 = -0.02274740735326741)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2033 = 0.3895398299261354)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2034 = 0.6126628429603742)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2034 = 0.3873357699956153)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2032 = 0.3873357699956153)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2032 = 0.6126628429603742)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2031 = 0.3895398299261354)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2031 = -0.02274740735326741)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14260: S1 ^operator O2034)
- 1017: O: O2034 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1017 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1016 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14261: I3 ^predict-no N1017)
- <=WM: (14248: N1016 ^status complete)
- <=WM: (14247: I3 ^predict-no N1016)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (14265: I2 ^dir L)
- =>WM: (14264: I2 ^reward 1)
- =>WM: (14263: I2 ^see 0)
- =>WM: (14262: N1017 ^status complete)
- <=WM: (14251: I2 ^dir L)
- <=WM: (14250: I2 ^reward 1)
- <=WM: (14249: I2 ^see 0)
- =>WM: (14266: I2 ^level-1 L0-root)
- <=WM: (14252: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2033 = 0.1599599085218832)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2034 = 0.6126667050990464)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1021 ^value 1 +)
- (R1 ^reward R1021 +)
- Firing propose*predict-yes
- -->
- (O2035 ^name predict-yes +)
- (S1 ^operator O2035 +)
- Firing propose*predict-no
- -->
- (O2036 ^name predict-no +)
- (S1 ^operator O2036 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2034 = 0.3873357699956153)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2033 = 0.3895398299261354)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2034 ^name predict-no +)
- (S1 ^operator O2034 +)
- Retracting propose*predict-yes
- -->
- (O2033 ^name predict-yes +)
- (S1 ^operator O2033 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1020 ^value 1 +)
- (R1 ^reward R1020 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2034 = 0.3873357699956153)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2034 = 0.6126628429603742)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2033 = 0.3895398299261354)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2033 = -0.02274740735326741)
- =>WM: (14272: S1 ^operator O2036 +)
- =>WM: (14271: S1 ^operator O2035 +)
- =>WM: (14270: O2036 ^name predict-no)
- =>WM: (14269: O2035 ^name predict-yes)
- =>WM: (14268: R1021 ^value 1)
- =>WM: (14267: R1 ^reward R1021)
- <=WM: (14258: S1 ^operator O2033 +)
- <=WM: (14259: S1 ^operator O2034 +)
- <=WM: (14260: S1 ^operator O2034)
- <=WM: (14253: R1 ^reward R1020)
- <=WM: (14256: O2034 ^name predict-no)
- <=WM: (14255: O2033 ^name predict-yes)
- <=WM: (14254: R1020 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2035 = 0.3895398299261354)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2035 = 0.1599599085218832)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2036 = 0.3873357699956153)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2036 = 0.6126667050990464)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2034 = 0.3873357699956153)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2034 = 0.6126667050990464)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2033 = 0.3895398299261354)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2033 = 0.1599599085218832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.719079 -0.331743 0.387336 -> 0.71908 -0.331744 0.387336(R,m,v=1,0.933702,0.0622468)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*32 0.280919 0.331744 0.612663 -> 0.280919 0.331744 0.612663(R,m,v=1,1,0)
- =>WM: (14273: S1 ^operator O2036)
- 1018: O: O2036 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1018 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1017 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14274: I3 ^predict-no N1018)
- <=WM: (14262: N1017 ^status complete)
- <=WM: (14261: I3 ^predict-no N1017)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- |\--- Input Phase ---
- =>WM: (14278: I2 ^dir U)
- =>WM: (14277: I2 ^reward 1)
- =>WM: (14276: I2 ^see 0)
- =>WM: (14275: N1018 ^status complete)
- <=WM: (14265: I2 ^dir L)
- <=WM: (14264: I2 ^reward 1)
- <=WM: (14263: I2 ^see 0)
- =>WM: (14279: I2 ^level-1 L0-root)
- <=WM: (14266: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1022 ^value 1 +)
- (R1 ^reward R1022 +)
- Firing propose*predict-yes
- -->
- (O2037 ^name predict-yes +)
- (S1 ^operator O2037 +)
- Firing propose*predict-no
- -->
- (O2038 ^name predict-no +)
- (S1 ^operator O2038 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2036 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2035 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2036 ^name predict-no +)
- (S1 ^operator O2036 +)
- Retracting propose*predict-yes
- -->
- (O2035 ^name predict-yes +)
- (S1 ^operator O2035 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1021 ^value 1 +)
- (R1 ^reward R1021 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2036 = 0.6126667050990464)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2036 = 0.3873359780522169)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2035 = 0.1599599085218832)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2035 = 0.3895398299261354)
- =>WM: (14286: S1 ^operator O2038 +)
- =>WM: (14285: S1 ^operator O2037 +)
- =>WM: (14284: I3 ^dir U)
- =>WM: (14283: O2038 ^name predict-no)
- =>WM: (14282: O2037 ^name predict-yes)
- =>WM: (14281: R1022 ^value 1)
- =>WM: (14280: R1 ^reward R1022)
- <=WM: (14271: S1 ^operator O2035 +)
- <=WM: (14272: S1 ^operator O2036 +)
- <=WM: (14273: S1 ^operator O2036)
- <=WM: (14257: I3 ^dir L)
- <=WM: (14267: R1 ^reward R1021)
- <=WM: (14270: O2036 ^name predict-no)
- <=WM: (14269: O2035 ^name predict-yes)
- <=WM: (14268: R1021 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2037 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2038 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2036 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2035 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.71908 -0.331744 0.387336 -> 0.719079 -0.331743 0.387336(R,m,v=1,0.934066,0.061927)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*37 0.280924 0.331742 0.612667 -> 0.280924 0.331742 0.612666(R,m,v=1,1,0)
- =>WM: (14287: S1 ^operator O2038)
- 1019: O: O2038 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1019 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1018 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14288: I3 ^predict-no N1019)
- <=WM: (14275: N1018 ^status complete)
- <=WM: (14274: I3 ^predict-no N1018)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (14292: I2 ^dir R)
- =>WM: (14291: I2 ^reward 1)
- =>WM: (14290: I2 ^see 0)
- =>WM: (14289: N1019 ^status complete)
- <=WM: (14278: I2 ^dir U)
- <=WM: (14277: I2 ^reward 1)
- <=WM: (14276: I2 ^see 0)
- =>WM: (14293: I2 ^level-1 L0-root)
- <=WM: (14279: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2037 = 0.8155918960261216)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2038 = -0.00558448899823713)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1023 ^value 1 +)
- (R1 ^reward R1023 +)
- Firing propose*predict-yes
- -->
- (O2039 ^name predict-yes +)
- (S1 ^operator O2039 +)
- Firing propose*predict-no
- -->
- (O2040 ^name predict-no +)
- (S1 ^operator O2040 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2038 = 0.4476193147022436)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2037 = 0.1844112548850312)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2038 ^name predict-no +)
- (S1 ^operator O2038 +)
- Retracting propose*predict-yes
- -->
- (O2037 ^name predict-yes +)
- (S1 ^operator O2037 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1022 ^value 1 +)
- (R1 ^reward R1022 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2038 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2037 = 0.)
- =>WM: (14300: S1 ^operator O2040 +)
- =>WM: (14299: S1 ^operator O2039 +)
- =>WM: (14298: I3 ^dir R)
- =>WM: (14297: O2040 ^name predict-no)
- =>WM: (14296: O2039 ^name predict-yes)
- =>WM: (14295: R1023 ^value 1)
- =>WM: (14294: R1 ^reward R1023)
- <=WM: (14285: S1 ^operator O2037 +)
- <=WM: (14286: S1 ^operator O2038 +)
- <=WM: (14287: S1 ^operator O2038)
- <=WM: (14284: I3 ^dir U)
- <=WM: (14280: R1 ^reward R1022)
- <=WM: (14283: O2038 ^name predict-no)
- <=WM: (14282: O2037 ^name predict-yes)
- <=WM: (14281: R1022 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2039 = 0.8155918960261216)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2039 = 0.1844112548850312)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2040 = -0.00558448899823713)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2040 = 0.4476193147022436)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2038 = 0.4476193147022436)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2038 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2037 = 0.1844112548850312)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2037 = 0.8155918960261216)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14301: S1 ^operator O2039)
- 1020: O: O2039 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1020 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1019 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14302: I3 ^predict-yes N1020)
- <=WM: (14289: N1019 ^status complete)
- <=WM: (14288: I3 ^predict-no N1019)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- \---- Input Phase ---
- =>WM: (14306: I2 ^dir R)
- =>WM: (14305: I2 ^reward 1)
- =>WM: (14304: I2 ^see 1)
- =>WM: (14303: N1020 ^status complete)
- <=WM: (14292: I2 ^dir R)
- <=WM: (14291: I2 ^reward 1)
- <=WM: (14290: I2 ^see 0)
- =>WM: (14307: I2 ^level-1 R1-root)
- <=WM: (14293: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2039 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2040 = 0.5523818179838019)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1024 ^value 1 +)
- (R1 ^reward R1024 +)
- Firing propose*predict-yes
- -->
- (O2041 ^name predict-yes +)
- (S1 ^operator O2041 +)
- Firing propose*predict-no
- -->
- (O2042 ^name predict-no +)
- (S1 ^operator O2042 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2040 = 0.4476193147022436)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2039 = 0.1844112548850312)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2040 ^name predict-no +)
- (S1 ^operator O2040 +)
- Retracting propose*predict-yes
- -->
- (O2039 ^name predict-yes +)
- (S1 ^operator O2039 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1023 ^value 1 +)
- (R1 ^reward R1023 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2040 = 0.4476193147022436)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2040 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2039 = 0.1844112548850312)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2039 = 0.8155918960261216)
- =>WM: (14314: S1 ^operator O2042 +)
- =>WM: (14313: S1 ^operator O2041 +)
- =>WM: (14312: O2042 ^name predict-no)
- =>WM: (14311: O2041 ^name predict-yes)
- =>WM: (14310: R1024 ^value 1)
- =>WM: (14309: R1 ^reward R1024)
- =>WM: (14308: I3 ^see 1)
- <=WM: (14299: S1 ^operator O2039 +)
- <=WM: (14301: S1 ^operator O2039)
- <=WM: (14300: S1 ^operator O2040 +)
- <=WM: (14294: R1 ^reward R1023)
- <=WM: (14239: I3 ^see 0)
- <=WM: (14297: O2040 ^name predict-no)
- <=WM: (14296: O2039 ^name predict-yes)
- <=WM: (14295: R1023 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2041 = 0.1844112548850312)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2041 = 0.1398795999120246)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2042 = 0.4476193147022436)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2042 = 0.5523818179838019)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2040 = 0.4476193147022436)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2040 = 0.5523818179838019)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2039 = 0.1844112548850312)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2039 = 0.1398795999120246)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675414 -0.491003 0.184411 -> 0.675413 -0.491002 0.184411(R,m,v=1,0.902299,0.0886652)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*34 0.32459 0.491002 0.815592 -> 0.32459 0.491002 0.815591(R,m,v=1,1,0)
- =>WM: (14315: S1 ^operator O2042)
- 1021: O: O2042 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1021 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1020 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14316: I3 ^predict-no N1021)
- <=WM: (14303: N1020 ^status complete)
- <=WM: (14302: I3 ^predict-yes N1020)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- /--- Input Phase ---
- =>WM: (14320: I2 ^dir R)
- =>WM: (14319: I2 ^reward 1)
- =>WM: (14318: I2 ^see 0)
- =>WM: (14317: N1021 ^status complete)
- <=WM: (14306: I2 ^dir R)
- <=WM: (14305: I2 ^reward 1)
- <=WM: (14304: I2 ^see 1)
- =>WM: (14321: I2 ^level-1 R0-root)
- <=WM: (14307: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2041 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2042 = 0.5523793263612301)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1025 ^value 1 +)
- (R1 ^reward R1025 +)
- Firing propose*predict-yes
- -->
- (O2043 ^name predict-yes +)
- (S1 ^operator O2043 +)
- Firing propose*predict-no
- -->
- (O2044 ^name predict-no +)
- (S1 ^operator O2044 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2042 = 0.4476193147022436)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2041 = 0.1844107822483583)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2042 ^name predict-no +)
- (S1 ^operator O2042 +)
- Retracting propose*predict-yes
- -->
- (O2041 ^name predict-yes +)
- (S1 ^operator O2041 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1024 ^value 1 +)
- (R1 ^reward R1024 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2042 = 0.5523818179838019)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2042 = 0.4476193147022436)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2041 = 0.1398795999120246)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2041 = 0.1844107822483583)
- =>WM: (14328: S1 ^operator O2044 +)
- =>WM: (14327: S1 ^operator O2043 +)
- =>WM: (14326: O2044 ^name predict-no)
- =>WM: (14325: O2043 ^name predict-yes)
- =>WM: (14324: R1025 ^value 1)
- =>WM: (14323: R1 ^reward R1025)
- =>WM: (14322: I3 ^see 0)
- <=WM: (14313: S1 ^operator O2041 +)
- <=WM: (14314: S1 ^operator O2042 +)
- <=WM: (14315: S1 ^operator O2042)
- <=WM: (14309: R1 ^reward R1024)
- <=WM: (14308: I3 ^see 1)
- <=WM: (14312: O2042 ^name predict-no)
- <=WM: (14311: O2041 ^name predict-yes)
- <=WM: (14310: R1024 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2043 = 0.1844107822483583)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2043 = 0.1664311307472832)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2044 = 0.4476193147022436)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2044 = 0.5523793263612301)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2042 = 0.4476193147022436)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2042 = 0.5523793263612301)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2041 = 0.1844107822483583)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2041 = 0.1664311307472832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.930769,0.0649374)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*39 0.377468 0.174914 0.552382 -> 0.377468 0.174914 0.552382(R,m,v=1,1,0)
- =>WM: (14329: S1 ^operator O2044)
- 1022: O: O2044 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1022 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1021 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14330: I3 ^predict-no N1022)
- <=WM: (14317: N1021 ^status complete)
- <=WM: (14316: I3 ^predict-no N1021)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- |\-/--- Input Phase ---
- =>WM: (14334: I2 ^dir R)
- =>WM: (14333: I2 ^reward 1)
- =>WM: (14332: I2 ^see 0)
- =>WM: (14331: N1022 ^status complete)
- <=WM: (14320: I2 ^dir R)
- <=WM: (14319: I2 ^reward 1)
- <=WM: (14318: I2 ^see 0)
- =>WM: (14335: I2 ^level-1 R0-root)
- <=WM: (14321: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2043 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2044 = 0.5523793263612301)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1026 ^value 1 +)
- (R1 ^reward R1026 +)
- Firing propose*predict-yes
- -->
- (O2045 ^name predict-yes +)
- (S1 ^operator O2045 +)
- Firing propose*predict-no
- -->
- (O2046 ^name predict-no +)
- (S1 ^operator O2046 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2044 = 0.4476191447993367)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2043 = 0.1844107822483583)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2044 ^name predict-no +)
- (S1 ^operator O2044 +)
- Retracting propose*predict-yes
- -->
- (O2043 ^name predict-yes +)
- (S1 ^operator O2043 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1025 ^value 1 +)
- (R1 ^reward R1025 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2044 = 0.5523793263612301)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2044 = 0.4476191447993367)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2043 = 0.1664311307472832)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2043 = 0.1844107822483583)
- =>WM: (14341: S1 ^operator O2046 +)
- =>WM: (14340: S1 ^operator O2045 +)
- =>WM: (14339: O2046 ^name predict-no)
- =>WM: (14338: O2045 ^name predict-yes)
- =>WM: (14337: R1026 ^value 1)
- =>WM: (14336: R1 ^reward R1026)
- <=WM: (14327: S1 ^operator O2043 +)
- <=WM: (14328: S1 ^operator O2044 +)
- <=WM: (14329: S1 ^operator O2044)
- <=WM: (14323: R1 ^reward R1025)
- <=WM: (14326: O2044 ^name predict-no)
- <=WM: (14325: O2043 ^name predict-yes)
- <=WM: (14324: R1025 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2045 = 0.1844107822483583)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2045 = 0.1664311307472832)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2046 = 0.4476191447993367)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2046 = 0.5523793263612301)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2044 = 0.4476191447993367)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2044 = 0.5523793263612301)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2043 = 0.1844107822483583)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2043 = 0.1664311307472832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.931298,0.0644745)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*41 0.377466 0.174913 0.552379 -> 0.377466 0.174913 0.55238(R,m,v=1,1,0)
- =>WM: (14342: S1 ^operator O2046)
- 1023: O: O2046 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1023 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1022 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14343: I3 ^predict-no N1023)
- <=WM: (14331: N1022 ^status complete)
- <=WM: (14330: I3 ^predict-no N1022)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (14347: I2 ^dir R)
- =>WM: (14346: I2 ^reward 1)
- =>WM: (14345: I2 ^see 0)
- =>WM: (14344: N1023 ^status complete)
- <=WM: (14334: I2 ^dir R)
- <=WM: (14333: I2 ^reward 1)
- <=WM: (14332: I2 ^see 0)
- =>WM: (14348: I2 ^level-1 R0-root)
- <=WM: (14335: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2045 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2046 = 0.552379555687145)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1027 ^value 1 +)
- (R1 ^reward R1027 +)
- Firing propose*predict-yes
- -->
- (O2047 ^name predict-yes +)
- (S1 ^operator O2047 +)
- Firing propose*predict-no
- -->
- (O2048 ^name predict-no +)
- (S1 ^operator O2048 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2046 = 0.4476193741252518)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2045 = 0.1844107822483583)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2046 ^name predict-no +)
- (S1 ^operator O2046 +)
- Retracting propose*predict-yes
- -->
- (O2045 ^name predict-yes +)
- (S1 ^operator O2045 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1026 ^value 1 +)
- (R1 ^reward R1026 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2046 = 0.552379555687145)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2046 = 0.4476193741252518)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2045 = 0.1664311307472832)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2045 = 0.1844107822483583)
- =>WM: (14354: S1 ^operator O2048 +)
- =>WM: (14353: S1 ^operator O2047 +)
- =>WM: (14352: O2048 ^name predict-no)
- =>WM: (14351: O2047 ^name predict-yes)
- =>WM: (14350: R1027 ^value 1)
- =>WM: (14349: R1 ^reward R1027)
- <=WM: (14340: S1 ^operator O2045 +)
- <=WM: (14341: S1 ^operator O2046 +)
- <=WM: (14342: S1 ^operator O2046)
- <=WM: (14336: R1 ^reward R1026)
- <=WM: (14339: O2046 ^name predict-no)
- <=WM: (14338: O2045 ^name predict-yes)
- <=WM: (14337: R1026 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2047 = 0.1844107822483583)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2047 = 0.1664311307472832)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2048 = 0.4476193741252518)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2048 = 0.552379555687145)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2046 = 0.4476193741252518)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2046 = 0.552379555687145)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2045 = 0.1844107822483583)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2045 = 0.1664311307472832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.44762(R,m,v=1,0.931818,0.064018)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*41 0.377466 0.174913 0.55238 -> 0.377466 0.174913 0.55238(R,m,v=1,1,0)
- =>WM: (14355: S1 ^operator O2048)
- 1024: O: O2048 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1024 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1023 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14356: I3 ^predict-no N1024)
- <=WM: (14344: N1023 ^status complete)
- <=WM: (14343: I3 ^predict-no N1023)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- /|\---- Input Phase ---
- =>WM: (14360: I2 ^dir R)
- =>WM: (14359: I2 ^reward 1)
- =>WM: (14358: I2 ^see 0)
- =>WM: (14357: N1024 ^status complete)
- <=WM: (14347: I2 ^dir R)
- <=WM: (14346: I2 ^reward 1)
- <=WM: (14345: I2 ^see 0)
- =>WM: (14361: I2 ^level-1 R0-root)
- <=WM: (14348: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2047 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2048 = 0.5523797162152855)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1028 ^value 1 +)
- (R1 ^reward R1028 +)
- Firing propose*predict-yes
- -->
- (O2049 ^name predict-yes +)
- (S1 ^operator O2049 +)
- Firing propose*predict-no
- -->
- (O2050 ^name predict-no +)
- (S1 ^operator O2050 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2048 = 0.4476195346533923)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2047 = 0.1844107822483583)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2048 ^name predict-no +)
- (S1 ^operator O2048 +)
- Retracting propose*predict-yes
- -->
- (O2047 ^name predict-yes +)
- (S1 ^operator O2047 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1027 ^value 1 +)
- (R1 ^reward R1027 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2048 = 0.5523797162152855)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2048 = 0.4476195346533923)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2047 = 0.1664311307472832)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2047 = 0.1844107822483583)
- =>WM: (14367: S1 ^operator O2050 +)
- =>WM: (14366: S1 ^operator O2049 +)
- =>WM: (14365: O2050 ^name predict-no)
- =>WM: (14364: O2049 ^name predict-yes)
- =>WM: (14363: R1028 ^value 1)
- =>WM: (14362: R1 ^reward R1028)
- <=WM: (14353: S1 ^operator O2047 +)
- <=WM: (14354: S1 ^operator O2048 +)
- <=WM: (14355: S1 ^operator O2048)
- <=WM: (14349: R1 ^reward R1027)
- <=WM: (14352: O2048 ^name predict-no)
- <=WM: (14351: O2047 ^name predict-yes)
- <=WM: (14350: R1027 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2049 = 0.1844107822483583)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2049 = 0.1664311307472832)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2050 = 0.4476195346533923)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2050 = 0.5523797162152855)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2048 = 0.4476195346533923)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2048 = 0.5523797162152855)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2047 = 0.1844107822483583)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2047 = 0.1664311307472832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.44762 -> 0.622533 -0.174913 0.44762(R,m,v=1,0.932331,0.063568)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*41 0.377466 0.174913 0.55238 -> 0.377466 0.174913 0.55238(R,m,v=1,1,0)
- =>WM: (14368: S1 ^operator O2050)
- 1025: O: O2050 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1025 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1024 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14369: I3 ^predict-no N1025)
- <=WM: (14357: N1024 ^status complete)
- <=WM: (14356: I3 ^predict-no N1024)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- /|--- Input Phase ---
- =>WM: (14373: I2 ^dir L)
- =>WM: (14372: I2 ^reward 1)
- =>WM: (14371: I2 ^see 0)
- =>WM: (14370: N1025 ^status complete)
- <=WM: (14360: I2 ^dir R)
- <=WM: (14359: I2 ^reward 1)
- <=WM: (14358: I2 ^see 0)
- =>WM: (14374: I2 ^level-1 R0-root)
- <=WM: (14361: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2049 = 0.6104610611928351)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2050 = 0.1063475139796038)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1029 ^value 1 +)
- (R1 ^reward R1029 +)
- Firing propose*predict-yes
- -->
- (O2051 ^name predict-yes +)
- (S1 ^operator O2051 +)
- Firing propose*predict-no
- -->
- (O2052 ^name predict-no +)
- (S1 ^operator O2052 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2050 = 0.3873355755795274)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2049 = 0.3895398299261354)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2050 ^name predict-no +)
- (S1 ^operator O2050 +)
- Retracting propose*predict-yes
- -->
- (O2049 ^name predict-yes +)
- (S1 ^operator O2049 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1028 ^value 1 +)
- (R1 ^reward R1028 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2050 = 0.5523798285849839)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2050 = 0.4476196470230906)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2049 = 0.1664311307472832)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2049 = 0.1844107822483583)
- =>WM: (14381: S1 ^operator O2052 +)
- =>WM: (14380: S1 ^operator O2051 +)
- =>WM: (14379: I3 ^dir L)
- =>WM: (14378: O2052 ^name predict-no)
- =>WM: (14377: O2051 ^name predict-yes)
- =>WM: (14376: R1029 ^value 1)
- =>WM: (14375: R1 ^reward R1029)
- <=WM: (14366: S1 ^operator O2049 +)
- <=WM: (14367: S1 ^operator O2050 +)
- <=WM: (14368: S1 ^operator O2050)
- <=WM: (14298: I3 ^dir R)
- <=WM: (14362: R1 ^reward R1028)
- <=WM: (14365: O2050 ^name predict-no)
- <=WM: (14364: O2049 ^name predict-yes)
- <=WM: (14363: R1028 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2051 = 0.6104610611928351)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2051 = 0.3895398299261354)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2052 = 0.1063475139796038)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2052 = 0.3873355755795274)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2050 = 0.3873355755795274)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2050 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2049 = 0.3895398299261354)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2049 = 0.6104610611928351)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174913 0.44762 -> 0.622533 -0.174913 0.44762(R,m,v=1,0.932836,0.0631242)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*41 0.377466 0.174913 0.55238 -> 0.377466 0.174913 0.55238(R,m,v=1,1,0)
- =>WM: (14382: S1 ^operator O2051)
- 1026: O: O2051 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1026 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1025 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14383: I3 ^predict-yes N1026)
- <=WM: (14370: N1025 ^status complete)
- <=WM: (14369: I3 ^predict-no N1025)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (14387: I2 ^dir R)
- =>WM: (14386: I2 ^reward 1)
- =>WM: (14385: I2 ^see 1)
- =>WM: (14384: N1026 ^status complete)
- <=WM: (14373: I2 ^dir L)
- <=WM: (14372: I2 ^reward 1)
- <=WM: (14371: I2 ^see 0)
- =>WM: (14388: I2 ^level-1 L1-root)
- <=WM: (14374: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2052 = -0.02155734064455064)
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2051 = 0.815583266028165)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1030 ^value 1 +)
- (R1 ^reward R1030 +)
- Firing propose*predict-yes
- -->
- (O2053 ^name predict-yes +)
- (S1 ^operator O2053 +)
- Firing propose*predict-no
- -->
- (O2054 ^name predict-no +)
- (S1 ^operator O2054 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2052 = 0.4476197256818795)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2051 = 0.1844107822483583)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2052 ^name predict-no +)
- (S1 ^operator O2052 +)
- Retracting propose*predict-yes
- -->
- (O2051 ^name predict-yes +)
- (S1 ^operator O2051 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1029 ^value 1 +)
- (R1 ^reward R1029 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2052 = 0.3873355755795274)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2052 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2051 = 0.3895398299261354)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2051 = 0.6104610611928351)
- =>WM: (14396: S1 ^operator O2054 +)
- =>WM: (14395: S1 ^operator O2053 +)
- =>WM: (14394: I3 ^dir R)
- =>WM: (14393: O2054 ^name predict-no)
- =>WM: (14392: O2053 ^name predict-yes)
- =>WM: (14391: R1030 ^value 1)
- =>WM: (14390: R1 ^reward R1030)
- =>WM: (14389: I3 ^see 1)
- <=WM: (14380: S1 ^operator O2051 +)
- <=WM: (14382: S1 ^operator O2051)
- <=WM: (14381: S1 ^operator O2052 +)
- <=WM: (14379: I3 ^dir L)
- <=WM: (14375: R1 ^reward R1029)
- <=WM: (14322: I3 ^see 0)
- <=WM: (14378: O2052 ^name predict-no)
- <=WM: (14377: O2051 ^name predict-yes)
- <=WM: (14376: R1029 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2053 = 0.1844107822483583)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2053 = 0.815583266028165)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2054 = 0.4476197256818795)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2054 = -0.02155734064455064)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2052 = 0.4476197256818795)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2052 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2051 = 0.1844107822483583)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2051 = 0.815583266028165)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322411 0.38954 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.895349,0.0942472)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*44 0.288049 0.322412 0.610461 -> 0.288049 0.322412 0.610461(R,m,v=1,1,0)
- =>WM: (14397: S1 ^operator O2053)
- 1027: O: O2053 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1027 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1026 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14398: I3 ^predict-yes N1027)
- <=WM: (14384: N1026 ^status complete)
- <=WM: (14383: I3 ^predict-yes N1026)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (14402: I2 ^dir U)
- =>WM: (14401: I2 ^reward 1)
- =>WM: (14400: I2 ^see 1)
- =>WM: (14399: N1027 ^status complete)
- <=WM: (14387: I2 ^dir R)
- <=WM: (14386: I2 ^reward 1)
- <=WM: (14385: I2 ^see 1)
- =>WM: (14403: I2 ^level-1 R1-root)
- <=WM: (14388: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1031 ^value 1 +)
- (R1 ^reward R1031 +)
- Firing propose*predict-yes
- -->
- (O2055 ^name predict-yes +)
- (S1 ^operator O2055 +)
- Firing propose*predict-no
- -->
- (O2056 ^name predict-no +)
- (S1 ^operator O2056 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2054 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2053 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2054 ^name predict-no +)
- (S1 ^operator O2054 +)
- Retracting propose*predict-yes
- -->
- (O2053 ^name predict-yes +)
- (S1 ^operator O2053 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1030 ^value 1 +)
- (R1 ^reward R1030 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2054 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2054 = 0.4476197256818795)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2053 = 0.815583266028165)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2053 = 0.1844107822483583)
- =>WM: (14410: S1 ^operator O2056 +)
- =>WM: (14409: S1 ^operator O2055 +)
- =>WM: (14408: I3 ^dir U)
- =>WM: (14407: O2056 ^name predict-no)
- =>WM: (14406: O2055 ^name predict-yes)
- =>WM: (14405: R1031 ^value 1)
- =>WM: (14404: R1 ^reward R1031)
- <=WM: (14395: S1 ^operator O2053 +)
- <=WM: (14397: S1 ^operator O2053)
- <=WM: (14396: S1 ^operator O2054 +)
- <=WM: (14394: I3 ^dir R)
- <=WM: (14390: R1 ^reward R1030)
- <=WM: (14393: O2054 ^name predict-no)
- <=WM: (14392: O2053 ^name predict-yes)
- <=WM: (14391: R1030 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2055 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2056 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2054 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2053 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675413 -0.491002 0.184411 -> 0.675414 -0.491003 0.184412(R,m,v=1,0.902857,0.0882102)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*46 0.324579 0.491004 0.815583 -> 0.32458 0.491004 0.815584(R,m,v=1,1,0)
- =>WM: (14411: S1 ^operator O2056)
- 1028: O: O2056 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1028 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1027 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14412: I3 ^predict-no N1028)
- <=WM: (14399: N1027 ^status complete)
- <=WM: (14398: I3 ^predict-yes N1027)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (14416: I2 ^dir U)
- =>WM: (14415: I2 ^reward 1)
- =>WM: (14414: I2 ^see 0)
- =>WM: (14413: N1028 ^status complete)
- <=WM: (14402: I2 ^dir U)
- <=WM: (14401: I2 ^reward 1)
- <=WM: (14400: I2 ^see 1)
- =>WM: (14417: I2 ^level-1 R1-root)
- <=WM: (14403: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1032 ^value 1 +)
- (R1 ^reward R1032 +)
- Firing propose*predict-yes
- -->
- (O2057 ^name predict-yes +)
- (S1 ^operator O2057 +)
- Firing propose*predict-no
- -->
- (O2058 ^name predict-no +)
- (S1 ^operator O2058 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2056 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2055 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2056 ^name predict-no +)
- (S1 ^operator O2056 +)
- Retracting propose*predict-yes
- -->
- (O2055 ^name predict-yes +)
- (S1 ^operator O2055 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1031 ^value 1 +)
- (R1 ^reward R1031 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2056 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2055 = 0.)
- =>WM: (14424: S1 ^operator O2058 +)
- =>WM: (14423: S1 ^operator O2057 +)
- =>WM: (14422: O2058 ^name predict-no)
- =>WM: (14421: O2057 ^name predict-yes)
- =>WM: (14420: R1032 ^value 1)
- =>WM: (14419: R1 ^reward R1032)
- =>WM: (14418: I3 ^see 0)
- <=WM: (14409: S1 ^operator O2055 +)
- <=WM: (14410: S1 ^operator O2056 +)
- <=WM: (14411: S1 ^operator O2056)
- <=WM: (14404: R1 ^reward R1031)
- <=WM: (14389: I3 ^see 1)
- <=WM: (14407: O2056 ^name predict-no)
- <=WM: (14406: O2055 ^name predict-yes)
- <=WM: (14405: R1031 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2057 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2058 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2056 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2055 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14425: S1 ^operator O2058)
- 1029: O: O2058 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1029 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1028 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14426: I3 ^predict-no N1029)
- <=WM: (14413: N1028 ^status complete)
- <=WM: (14412: I3 ^predict-no N1028)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (14430: I2 ^dir U)
- =>WM: (14429: I2 ^reward 1)
- =>WM: (14428: I2 ^see 0)
- =>WM: (14427: N1029 ^status complete)
- <=WM: (14416: I2 ^dir U)
- <=WM: (14415: I2 ^reward 1)
- <=WM: (14414: I2 ^see 0)
- =>WM: (14431: I2 ^level-1 R1-root)
- <=WM: (14417: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1033 ^value 1 +)
- (R1 ^reward R1033 +)
- Firing propose*predict-yes
- -->
- (O2059 ^name predict-yes +)
- (S1 ^operator O2059 +)
- Firing propose*predict-no
- -->
- (O2060 ^name predict-no +)
- (S1 ^operator O2060 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2058 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2057 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2058 ^name predict-no +)
- (S1 ^operator O2058 +)
- Retracting propose*predict-yes
- -->
- (O2057 ^name predict-yes +)
- (S1 ^operator O2057 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1032 ^value 1 +)
- (R1 ^reward R1032 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2058 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2057 = 0.)
- =>WM: (14437: S1 ^operator O2060 +)
- =>WM: (14436: S1 ^operator O2059 +)
- =>WM: (14435: O2060 ^name predict-no)
- =>WM: (14434: O2059 ^name predict-yes)
- =>WM: (14433: R1033 ^value 1)
- =>WM: (14432: R1 ^reward R1033)
- <=WM: (14423: S1 ^operator O2057 +)
- <=WM: (14424: S1 ^operator O2058 +)
- <=WM: (14425: S1 ^operator O2058)
- <=WM: (14419: R1 ^reward R1032)
- <=WM: (14422: O2058 ^name predict-no)
- <=WM: (14421: O2057 ^name predict-yes)
- <=WM: (14420: R1032 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2059 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2060 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2058 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2057 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14438: S1 ^operator O2060)
- 1030: O: O2060 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1030 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1029 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14439: I3 ^predict-no N1030)
- <=WM: (14427: N1029 ^status complete)
- <=WM: (14426: I3 ^predict-no N1029)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- \---- Input Phase ---
- =>WM: (14443: I2 ^dir U)
- =>WM: (14442: I2 ^reward 1)
- =>WM: (14441: I2 ^see 0)
- =>WM: (14440: N1030 ^status complete)
- <=WM: (14430: I2 ^dir U)
- <=WM: (14429: I2 ^reward 1)
- <=WM: (14428: I2 ^see 0)
- =>WM: (14444: I2 ^level-1 R1-root)
- <=WM: (14431: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1034 ^value 1 +)
- (R1 ^reward R1034 +)
- Firing propose*predict-yes
- -->
- (O2061 ^name predict-yes +)
- (S1 ^operator O2061 +)
- Firing propose*predict-no
- -->
- (O2062 ^name predict-no +)
- (S1 ^operator O2062 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2060 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2059 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2060 ^name predict-no +)
- (S1 ^operator O2060 +)
- Retracting propose*predict-yes
- -->
- (O2059 ^name predict-yes +)
- (S1 ^operator O2059 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1033 ^value 1 +)
- (R1 ^reward R1033 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2060 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2059 = 0.)
- =>WM: (14450: S1 ^operator O2062 +)
- =>WM: (14449: S1 ^operator O2061 +)
- =>WM: (14448: O2062 ^name predict-no)
- =>WM: (14447: O2061 ^name predict-yes)
- =>WM: (14446: R1034 ^value 1)
- =>WM: (14445: R1 ^reward R1034)
- <=WM: (14436: S1 ^operator O2059 +)
- <=WM: (14437: S1 ^operator O2060 +)
- <=WM: (14438: S1 ^operator O2060)
- <=WM: (14432: R1 ^reward R1033)
- <=WM: (14435: O2060 ^name predict-no)
- <=WM: (14434: O2059 ^name predict-yes)
- <=WM: (14433: R1033 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2061 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2062 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2060 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2059 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14451: S1 ^operator O2062)
- 1031: O: O2062 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1031 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1030 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14452: I3 ^predict-no N1031)
- <=WM: (14440: N1030 ^status complete)
- <=WM: (14439: I3 ^predict-no N1030)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- /--- Input Phase ---
- =>WM: (14456: I2 ^dir R)
- =>WM: (14455: I2 ^reward 1)
- =>WM: (14454: I2 ^see 0)
- =>WM: (14453: N1031 ^status complete)
- <=WM: (14443: I2 ^dir U)
- <=WM: (14442: I2 ^reward 1)
- <=WM: (14441: I2 ^see 0)
- =>WM: (14457: I2 ^level-1 R1-root)
- <=WM: (14444: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2061 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2062 = 0.5523816480808952)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1035 ^value 1 +)
- (R1 ^reward R1035 +)
- Firing propose*predict-yes
- -->
- (O2063 ^name predict-yes +)
- (S1 ^operator O2063 +)
- Firing propose*predict-no
- -->
- (O2064 ^name predict-no +)
- (S1 ^operator O2064 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2062 = 0.4476197256818795)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2061 = 0.1844116750068798)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2062 ^name predict-no +)
- (S1 ^operator O2062 +)
- Retracting propose*predict-yes
- -->
- (O2061 ^name predict-yes +)
- (S1 ^operator O2061 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1034 ^value 1 +)
- (R1 ^reward R1034 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2062 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2061 = 0.)
- =>WM: (14464: S1 ^operator O2064 +)
- =>WM: (14463: S1 ^operator O2063 +)
- =>WM: (14462: I3 ^dir R)
- =>WM: (14461: O2064 ^name predict-no)
- =>WM: (14460: O2063 ^name predict-yes)
- =>WM: (14459: R1035 ^value 1)
- =>WM: (14458: R1 ^reward R1035)
- <=WM: (14449: S1 ^operator O2061 +)
- <=WM: (14450: S1 ^operator O2062 +)
- <=WM: (14451: S1 ^operator O2062)
- <=WM: (14408: I3 ^dir U)
- <=WM: (14445: R1 ^reward R1034)
- <=WM: (14448: O2062 ^name predict-no)
- <=WM: (14447: O2061 ^name predict-yes)
- <=WM: (14446: R1034 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2063 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2063 = 0.1844116750068798)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2064 = 0.5523816480808952)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2064 = 0.4476197256818795)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2062 = 0.4476197256818795)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2062 = 0.5523816480808952)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2061 = 0.1844116750068798)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2061 = 0.1398795999120246)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14465: S1 ^operator O2064)
- 1032: O: O2064 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1032 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1031 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14466: I3 ^predict-no N1032)
- <=WM: (14453: N1031 ^status complete)
- <=WM: (14452: I3 ^predict-no N1031)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (14470: I2 ^dir R)
- =>WM: (14469: I2 ^reward 1)
- =>WM: (14468: I2 ^see 0)
- =>WM: (14467: N1032 ^status complete)
- <=WM: (14456: I2 ^dir R)
- <=WM: (14455: I2 ^reward 1)
- <=WM: (14454: I2 ^see 0)
- =>WM: (14471: I2 ^level-1 R0-root)
- <=WM: (14457: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2063 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2064 = 0.5523799072437727)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1036 ^value 1 +)
- (R1 ^reward R1036 +)
- Firing propose*predict-yes
- -->
- (O2065 ^name predict-yes +)
- (S1 ^operator O2065 +)
- Firing propose*predict-no
- -->
- (O2066 ^name predict-no +)
- (S1 ^operator O2066 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2064 = 0.4476197256818795)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2063 = 0.1844116750068798)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2064 ^name predict-no +)
- (S1 ^operator O2064 +)
- Retracting propose*predict-yes
- -->
- (O2063 ^name predict-yes +)
- (S1 ^operator O2063 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1035 ^value 1 +)
- (R1 ^reward R1035 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2064 = 0.4476197256818795)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2064 = 0.5523816480808952)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2063 = 0.1844116750068798)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2063 = 0.1398795999120246)
- =>WM: (14477: S1 ^operator O2066 +)
- =>WM: (14476: S1 ^operator O2065 +)
- =>WM: (14475: O2066 ^name predict-no)
- =>WM: (14474: O2065 ^name predict-yes)
- =>WM: (14473: R1036 ^value 1)
- =>WM: (14472: R1 ^reward R1036)
- <=WM: (14463: S1 ^operator O2063 +)
- <=WM: (14464: S1 ^operator O2064 +)
- <=WM: (14465: S1 ^operator O2064)
- <=WM: (14458: R1 ^reward R1035)
- <=WM: (14461: O2064 ^name predict-no)
- <=WM: (14460: O2063 ^name predict-yes)
- <=WM: (14459: R1035 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2065 = 0.1844116750068798)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2065 = 0.1664311307472832)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2066 = 0.4476197256818795)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2066 = 0.5523799072437727)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2064 = 0.4476197256818795)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2064 = 0.5523799072437727)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2063 = 0.1844116750068798)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2063 = 0.1664311307472832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174913 0.44762 -> 0.622533 -0.174914 0.44762(R,m,v=1,0.933333,0.0626866)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*39 0.377468 0.174914 0.552382 -> 0.377468 0.174914 0.552381(R,m,v=1,1,0)
- =>WM: (14478: S1 ^operator O2066)
- 1033: O: O2066 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1033 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1032 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14479: I3 ^predict-no N1033)
- <=WM: (14467: N1032 ^status complete)
- <=WM: (14466: I3 ^predict-no N1032)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (14483: I2 ^dir U)
- =>WM: (14482: I2 ^reward 1)
- =>WM: (14481: I2 ^see 0)
- =>WM: (14480: N1033 ^status complete)
- <=WM: (14470: I2 ^dir R)
- <=WM: (14469: I2 ^reward 1)
- <=WM: (14468: I2 ^see 0)
- =>WM: (14484: I2 ^level-1 R0-root)
- <=WM: (14471: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1037 ^value 1 +)
- (R1 ^reward R1037 +)
- Firing propose*predict-yes
- -->
- (O2067 ^name predict-yes +)
- (S1 ^operator O2067 +)
- Firing propose*predict-no
- -->
- (O2068 ^name predict-no +)
- (S1 ^operator O2068 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2066 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2065 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2066 ^name predict-no +)
- (S1 ^operator O2066 +)
- Retracting propose*predict-yes
- -->
- (O2065 ^name predict-yes +)
- (S1 ^operator O2065 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1036 ^value 1 +)
- (R1 ^reward R1036 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2066 = 0.5523799072437727)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2066 = 0.4476195196174632)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2065 = 0.1664311307472832)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2065 = 0.1844116750068798)
- =>WM: (14491: S1 ^operator O2068 +)
- =>WM: (14490: S1 ^operator O2067 +)
- =>WM: (14489: I3 ^dir U)
- =>WM: (14488: O2068 ^name predict-no)
- =>WM: (14487: O2067 ^name predict-yes)
- =>WM: (14486: R1037 ^value 1)
- =>WM: (14485: R1 ^reward R1037)
- <=WM: (14476: S1 ^operator O2065 +)
- <=WM: (14477: S1 ^operator O2066 +)
- <=WM: (14478: S1 ^operator O2066)
- <=WM: (14462: I3 ^dir R)
- <=WM: (14472: R1 ^reward R1036)
- <=WM: (14475: O2066 ^name predict-no)
- <=WM: (14474: O2065 ^name predict-yes)
- <=WM: (14473: R1036 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2067 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2068 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2066 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2065 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.44762 -> 0.622533 -0.174913 0.44762(R,m,v=1,0.933824,0.0622549)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*41 0.377466 0.174913 0.55238 -> 0.377467 0.174913 0.55238(R,m,v=1,1,0)
- =>WM: (14492: S1 ^operator O2068)
- 1034: O: O2068 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1034 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1033 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14493: I3 ^predict-no N1034)
- <=WM: (14480: N1033 ^status complete)
- <=WM: (14479: I3 ^predict-no N1033)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- -/|\--- Input Phase ---
- =>WM: (14497: I2 ^dir L)
- =>WM: (14496: I2 ^reward 1)
- =>WM: (14495: I2 ^see 0)
- =>WM: (14494: N1034 ^status complete)
- <=WM: (14483: I2 ^dir U)
- <=WM: (14482: I2 ^reward 1)
- <=WM: (14481: I2 ^see 0)
- =>WM: (14498: I2 ^level-1 R0-root)
- <=WM: (14484: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2067 = 0.6104609275249895)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2068 = 0.1063475139796038)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1038 ^value 1 +)
- (R1 ^reward R1038 +)
- Firing propose*predict-yes
- -->
- (O2069 ^name predict-yes +)
- (S1 ^operator O2069 +)
- Firing propose*predict-no
- -->
- (O2070 ^name predict-no +)
- (S1 ^operator O2070 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2068 = 0.3873355755795274)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2067 = 0.3895396962582899)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2068 ^name predict-no +)
- (S1 ^operator O2068 +)
- Retracting propose*predict-yes
- -->
- (O2067 ^name predict-yes +)
- (S1 ^operator O2067 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1037 ^value 1 +)
- (R1 ^reward R1037 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2068 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2067 = 0.)
- =>WM: (14505: S1 ^operator O2070 +)
- =>WM: (14504: S1 ^operator O2069 +)
- =>WM: (14503: I3 ^dir L)
- =>WM: (14502: O2070 ^name predict-no)
- =>WM: (14501: O2069 ^name predict-yes)
- =>WM: (14500: R1038 ^value 1)
- =>WM: (14499: R1 ^reward R1038)
- <=WM: (14490: S1 ^operator O2067 +)
- <=WM: (14491: S1 ^operator O2068 +)
- <=WM: (14492: S1 ^operator O2068)
- <=WM: (14489: I3 ^dir U)
- <=WM: (14485: R1 ^reward R1037)
- <=WM: (14488: O2068 ^name predict-no)
- <=WM: (14487: O2067 ^name predict-yes)
- <=WM: (14486: R1037 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2069 = 0.6104609275249895)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2069 = 0.3895396962582899)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2070 = 0.1063475139796038)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2070 = 0.3873355755795274)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2068 = 0.3873355755795274)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2068 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2067 = 0.3895396962582899)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2067 = 0.6104609275249895)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14506: S1 ^operator O2069)
- 1035: O: O2069 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1035 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1034 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14507: I3 ^predict-yes N1035)
- <=WM: (14494: N1034 ^status complete)
- <=WM: (14493: I3 ^predict-no N1034)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (14511: I2 ^dir U)
- =>WM: (14510: I2 ^reward 1)
- =>WM: (14509: I2 ^see 1)
- =>WM: (14508: N1035 ^status complete)
- <=WM: (14497: I2 ^dir L)
- <=WM: (14496: I2 ^reward 1)
- <=WM: (14495: I2 ^see 0)
- =>WM: (14512: I2 ^level-1 L1-root)
- <=WM: (14498: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1039 ^value 1 +)
- (R1 ^reward R1039 +)
- Firing propose*predict-yes
- -->
- (O2071 ^name predict-yes +)
- (S1 ^operator O2071 +)
- Firing propose*predict-no
- -->
- (O2072 ^name predict-no +)
- (S1 ^operator O2072 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2070 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2069 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2070 ^name predict-no +)
- (S1 ^operator O2070 +)
- Retracting propose*predict-yes
- -->
- (O2069 ^name predict-yes +)
- (S1 ^operator O2069 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1038 ^value 1 +)
- (R1 ^reward R1038 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2070 = 0.3873355755795274)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2070 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2069 = 0.3895396962582899)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2069 = 0.6104609275249895)
- =>WM: (14520: S1 ^operator O2072 +)
- =>WM: (14519: S1 ^operator O2071 +)
- =>WM: (14518: I3 ^dir U)
- =>WM: (14517: O2072 ^name predict-no)
- =>WM: (14516: O2071 ^name predict-yes)
- =>WM: (14515: R1039 ^value 1)
- =>WM: (14514: R1 ^reward R1039)
- =>WM: (14513: I3 ^see 1)
- <=WM: (14504: S1 ^operator O2069 +)
- <=WM: (14506: S1 ^operator O2069)
- <=WM: (14505: S1 ^operator O2070 +)
- <=WM: (14503: I3 ^dir L)
- <=WM: (14499: R1 ^reward R1038)
- <=WM: (14418: I3 ^see 0)
- <=WM: (14502: O2070 ^name predict-no)
- <=WM: (14501: O2069 ^name predict-yes)
- <=WM: (14500: R1038 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2071 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2072 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2070 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2069 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.895954,0.0937626)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*44 0.288049 0.322412 0.610461 -> 0.288049 0.322412 0.610461(R,m,v=1,1,0)
- =>WM: (14521: S1 ^operator O2072)
- 1036: O: O2072 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1036 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1035 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14522: I3 ^predict-no N1036)
- <=WM: (14508: N1035 ^status complete)
- <=WM: (14507: I3 ^predict-yes N1035)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (14526: I2 ^dir L)
- =>WM: (14525: I2 ^reward 1)
- =>WM: (14524: I2 ^see 0)
- =>WM: (14523: N1036 ^status complete)
- <=WM: (14511: I2 ^dir U)
- <=WM: (14510: I2 ^reward 1)
- <=WM: (14509: I2 ^see 1)
- =>WM: (14527: I2 ^level-1 L1-root)
- <=WM: (14512: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2072 = 0.6126630510169757)
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2071 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1040 ^value 1 +)
- (R1 ^reward R1040 +)
- Firing propose*predict-yes
- -->
- (O2073 ^name predict-yes +)
- (S1 ^operator O2073 +)
- Firing propose*predict-no
- -->
- (O2074 ^name predict-no +)
- (S1 ^operator O2074 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2072 = 0.3873355755795274)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2071 = 0.389539602690798)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2072 ^name predict-no +)
- (S1 ^operator O2072 +)
- Retracting propose*predict-yes
- -->
- (O2071 ^name predict-yes +)
- (S1 ^operator O2071 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1039 ^value 1 +)
- (R1 ^reward R1039 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2072 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2071 = 0.)
- =>WM: (14535: S1 ^operator O2074 +)
- =>WM: (14534: S1 ^operator O2073 +)
- =>WM: (14533: I3 ^dir L)
- =>WM: (14532: O2074 ^name predict-no)
- =>WM: (14531: O2073 ^name predict-yes)
- =>WM: (14530: R1040 ^value 1)
- =>WM: (14529: R1 ^reward R1040)
- =>WM: (14528: I3 ^see 0)
- <=WM: (14519: S1 ^operator O2071 +)
- <=WM: (14520: S1 ^operator O2072 +)
- <=WM: (14521: S1 ^operator O2072)
- <=WM: (14518: I3 ^dir U)
- <=WM: (14514: R1 ^reward R1039)
- <=WM: (14513: I3 ^see 1)
- <=WM: (14517: O2072 ^name predict-no)
- <=WM: (14516: O2071 ^name predict-yes)
- <=WM: (14515: R1039 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2073 = -0.02274740735326741)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2073 = 0.389539602690798)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2074 = 0.6126630510169757)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2074 = 0.3873355755795274)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2072 = 0.3873355755795274)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2072 = 0.6126630510169757)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2071 = 0.389539602690798)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2071 = -0.02274740735326741)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14536: S1 ^operator O2074)
- 1037: O: O2074 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1037 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1036 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14537: I3 ^predict-no N1037)
- <=WM: (14523: N1036 ^status complete)
- <=WM: (14522: I3 ^predict-no N1036)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (14541: I2 ^dir R)
- =>WM: (14540: I2 ^reward 1)
- =>WM: (14539: I2 ^see 0)
- =>WM: (14538: N1037 ^status complete)
- <=WM: (14526: I2 ^dir L)
- <=WM: (14525: I2 ^reward 1)
- <=WM: (14524: I2 ^see 0)
- =>WM: (14542: I2 ^level-1 L0-root)
- <=WM: (14527: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2073 = 0.8155914233894487)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2074 = -0.00558448899823713)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1041 ^value 1 +)
- (R1 ^reward R1041 +)
- Firing propose*predict-yes
- -->
- (O2075 ^name predict-yes +)
- (S1 ^operator O2075 +)
- Firing propose*predict-no
- -->
- (O2076 ^name predict-no +)
- (S1 ^operator O2076 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2074 = 0.4476196055882778)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2073 = 0.1844116750068798)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2074 ^name predict-no +)
- (S1 ^operator O2074 +)
- Retracting propose*predict-yes
- -->
- (O2073 ^name predict-yes +)
- (S1 ^operator O2073 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1040 ^value 1 +)
- (R1 ^reward R1040 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2074 = 0.3873355755795274)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2074 = 0.6126630510169757)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2073 = 0.389539602690798)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2073 = -0.02274740735326741)
- =>WM: (14549: S1 ^operator O2076 +)
- =>WM: (14548: S1 ^operator O2075 +)
- =>WM: (14547: I3 ^dir R)
- =>WM: (14546: O2076 ^name predict-no)
- =>WM: (14545: O2075 ^name predict-yes)
- =>WM: (14544: R1041 ^value 1)
- =>WM: (14543: R1 ^reward R1041)
- <=WM: (14534: S1 ^operator O2073 +)
- <=WM: (14535: S1 ^operator O2074 +)
- <=WM: (14536: S1 ^operator O2074)
- <=WM: (14533: I3 ^dir L)
- <=WM: (14529: R1 ^reward R1040)
- <=WM: (14532: O2074 ^name predict-no)
- <=WM: (14531: O2073 ^name predict-yes)
- <=WM: (14530: R1040 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2075 = 0.1844116750068798)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2075 = 0.8155914233894487)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2076 = 0.4476196055882778)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2076 = -0.00558448899823713)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2074 = 0.4476196055882778)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2074 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2073 = 0.1844116750068798)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2073 = 0.8155914233894487)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.719079 -0.331743 0.387336 -> 0.719079 -0.331743 0.387336(R,m,v=1,0.934426,0.0616105)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*32 0.280919 0.331744 0.612663 -> 0.280919 0.331744 0.612663(R,m,v=1,1,0)
- =>WM: (14550: S1 ^operator O2075)
- 1038: O: O2075 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1038 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1037 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14551: I3 ^predict-yes N1038)
- <=WM: (14538: N1037 ^status complete)
- <=WM: (14537: I3 ^predict-no N1037)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (14555: I2 ^dir R)
- =>WM: (14554: I2 ^reward 1)
- =>WM: (14553: I2 ^see 1)
- =>WM: (14552: N1038 ^status complete)
- <=WM: (14541: I2 ^dir R)
- <=WM: (14540: I2 ^reward 1)
- <=WM: (14539: I2 ^see 0)
- =>WM: (14556: I2 ^level-1 R1-root)
- <=WM: (14542: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2075 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2076 = 0.552381442016479)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1042 ^value 1 +)
- (R1 ^reward R1042 +)
- Firing propose*predict-yes
- -->
- (O2077 ^name predict-yes +)
- (S1 ^operator O2077 +)
- Firing propose*predict-no
- -->
- (O2078 ^name predict-no +)
- (S1 ^operator O2078 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2076 = 0.4476196055882778)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2075 = 0.1844116750068798)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2076 ^name predict-no +)
- (S1 ^operator O2076 +)
- Retracting propose*predict-yes
- -->
- (O2075 ^name predict-yes +)
- (S1 ^operator O2075 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1041 ^value 1 +)
- (R1 ^reward R1041 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2076 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2076 = 0.4476196055882778)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2075 = 0.8155914233894487)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2075 = 0.1844116750068798)
- =>WM: (14563: S1 ^operator O2078 +)
- =>WM: (14562: S1 ^operator O2077 +)
- =>WM: (14561: O2078 ^name predict-no)
- =>WM: (14560: O2077 ^name predict-yes)
- =>WM: (14559: R1042 ^value 1)
- =>WM: (14558: R1 ^reward R1042)
- =>WM: (14557: I3 ^see 1)
- <=WM: (14548: S1 ^operator O2075 +)
- <=WM: (14550: S1 ^operator O2075)
- <=WM: (14549: S1 ^operator O2076 +)
- <=WM: (14543: R1 ^reward R1041)
- <=WM: (14528: I3 ^see 0)
- <=WM: (14546: O2076 ^name predict-no)
- <=WM: (14545: O2075 ^name predict-yes)
- <=WM: (14544: R1041 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2077 = 0.1844116750068798)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2077 = 0.1398795999120246)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2078 = 0.4476196055882778)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2078 = 0.552381442016479)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2076 = 0.4476196055882778)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2076 = 0.552381442016479)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2075 = 0.1844116750068798)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2075 = 0.1398795999120246)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675414 -0.491003 0.184412 -> 0.675414 -0.491003 0.184411(R,m,v=1,0.903409,0.0877597)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*34 0.32459 0.491002 0.815591 -> 0.324589 0.491002 0.815591(R,m,v=1,1,0)
- =>WM: (14564: S1 ^operator O2078)
- 1039: O: O2078 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1039 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1038 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14565: I3 ^predict-no N1039)
- <=WM: (14552: N1038 ^status complete)
- <=WM: (14551: I3 ^predict-yes N1038)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- ---- Input Phase ---
- =>WM: (14569: I2 ^dir U)
- =>WM: (14568: I2 ^reward 1)
- =>WM: (14567: I2 ^see 0)
- =>WM: (14566: N1039 ^status complete)
- <=WM: (14555: I2 ^dir R)
- <=WM: (14554: I2 ^reward 1)
- <=WM: (14553: I2 ^see 1)
- =>WM: (14570: I2 ^level-1 R0-root)
- <=WM: (14556: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1043 ^value 1 +)
- (R1 ^reward R1043 +)
- Firing propose*predict-yes
- -->
- (O2079 ^name predict-yes +)
- (S1 ^operator O2079 +)
- Firing propose*predict-no
- -->
- (O2080 ^name predict-no +)
- (S1 ^operator O2080 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2078 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2077 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2078 ^name predict-no +)
- (S1 ^operator O2078 +)
- Retracting propose*predict-yes
- -->
- (O2077 ^name predict-yes +)
- (S1 ^operator O2077 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1042 ^value 1 +)
- (R1 ^reward R1042 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2078 = 0.552381442016479)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2078 = 0.4476196055882778)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2077 = 0.1398795999120246)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2077 = 0.1844112102474305)
- =>WM: (14578: S1 ^operator O2080 +)
- =>WM: (14577: S1 ^operator O2079 +)
- =>WM: (14576: I3 ^dir U)
- =>WM: (14575: O2080 ^name predict-no)
- =>WM: (14574: O2079 ^name predict-yes)
- =>WM: (14573: R1043 ^value 1)
- =>WM: (14572: R1 ^reward R1043)
- =>WM: (14571: I3 ^see 0)
- <=WM: (14562: S1 ^operator O2077 +)
- <=WM: (14563: S1 ^operator O2078 +)
- <=WM: (14564: S1 ^operator O2078)
- <=WM: (14547: I3 ^dir R)
- <=WM: (14558: R1 ^reward R1042)
- <=WM: (14557: I3 ^see 1)
- <=WM: (14561: O2078 ^name predict-no)
- <=WM: (14560: O2077 ^name predict-yes)
- <=WM: (14559: R1042 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2079 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2080 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2078 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2077 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174913 0.44762 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.934307,0.0618291)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*39 0.377468 0.174914 0.552381 -> 0.377468 0.174914 0.552381(R,m,v=1,1,0)
- =>WM: (14579: S1 ^operator O2080)
- 1040: O: O2080 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1040 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1039 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14580: I3 ^predict-no N1040)
- <=WM: (14566: N1039 ^status complete)
- <=WM: (14565: I3 ^predict-no N1039)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (14584: I2 ^dir L)
- =>WM: (14583: I2 ^reward 1)
- =>WM: (14582: I2 ^see 0)
- =>WM: (14581: N1040 ^status complete)
- <=WM: (14569: I2 ^dir U)
- <=WM: (14568: I2 ^reward 1)
- <=WM: (14567: I2 ^see 0)
- =>WM: (14585: I2 ^level-1 R0-root)
- <=WM: (14570: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2079 = 0.6104608339574975)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2080 = 0.1063475139796038)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1044 ^value 1 +)
- (R1 ^reward R1044 +)
- Firing propose*predict-yes
- -->
- (O2081 ^name predict-yes +)
- (S1 ^operator O2081 +)
- Firing propose*predict-no
- -->
- (O2082 ^name predict-no +)
- (S1 ^operator O2082 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2080 = 0.3873357815900519)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2079 = 0.389539602690798)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2080 ^name predict-no +)
- (S1 ^operator O2080 +)
- Retracting propose*predict-yes
- -->
- (O2079 ^name predict-yes +)
- (S1 ^operator O2079 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1043 ^value 1 +)
- (R1 ^reward R1043 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2080 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2079 = 0.)
- =>WM: (14592: S1 ^operator O2082 +)
- =>WM: (14591: S1 ^operator O2081 +)
- =>WM: (14590: I3 ^dir L)
- =>WM: (14589: O2082 ^name predict-no)
- =>WM: (14588: O2081 ^name predict-yes)
- =>WM: (14587: R1044 ^value 1)
- =>WM: (14586: R1 ^reward R1044)
- <=WM: (14577: S1 ^operator O2079 +)
- <=WM: (14578: S1 ^operator O2080 +)
- <=WM: (14579: S1 ^operator O2080)
- <=WM: (14576: I3 ^dir U)
- <=WM: (14572: R1 ^reward R1043)
- <=WM: (14575: O2080 ^name predict-no)
- <=WM: (14574: O2079 ^name predict-yes)
- <=WM: (14573: R1043 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2081 = 0.6104608339574975)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2081 = 0.389539602690798)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2082 = 0.1063475139796038)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2082 = 0.3873357815900519)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2080 = 0.3873357815900519)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2080 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2079 = 0.389539602690798)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2079 = 0.6104608339574975)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14593: S1 ^operator O2081)
- 1041: O: O2081 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1041 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1040 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14594: I3 ^predict-yes N1041)
- <=WM: (14581: N1040 ^status complete)
- <=WM: (14580: I3 ^predict-no N1040)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- ---- Input Phase ---
- =>WM: (14598: I2 ^dir U)
- =>WM: (14597: I2 ^reward 1)
- =>WM: (14596: I2 ^see 1)
- =>WM: (14595: N1041 ^status complete)
- <=WM: (14584: I2 ^dir L)
- <=WM: (14583: I2 ^reward 1)
- <=WM: (14582: I2 ^see 0)
- =>WM: (14599: I2 ^level-1 L1-root)
- <=WM: (14585: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1045 ^value 1 +)
- (R1 ^reward R1045 +)
- Firing propose*predict-yes
- -->
- (O2083 ^name predict-yes +)
- (S1 ^operator O2083 +)
- Firing propose*predict-no
- -->
- (O2084 ^name predict-no +)
- (S1 ^operator O2084 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2082 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2081 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2082 ^name predict-no +)
- (S1 ^operator O2082 +)
- Retracting propose*predict-yes
- -->
- (O2081 ^name predict-yes +)
- (S1 ^operator O2081 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1044 ^value 1 +)
- (R1 ^reward R1044 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2082 = 0.3873357815900519)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2082 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2081 = 0.389539602690798)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2081 = 0.6104608339574975)
- =>WM: (14607: S1 ^operator O2084 +)
- =>WM: (14606: S1 ^operator O2083 +)
- =>WM: (14605: I3 ^dir U)
- =>WM: (14604: O2084 ^name predict-no)
- =>WM: (14603: O2083 ^name predict-yes)
- =>WM: (14602: R1045 ^value 1)
- =>WM: (14601: R1 ^reward R1045)
- =>WM: (14600: I3 ^see 1)
- <=WM: (14591: S1 ^operator O2081 +)
- <=WM: (14593: S1 ^operator O2081)
- <=WM: (14592: S1 ^operator O2082 +)
- <=WM: (14590: I3 ^dir L)
- <=WM: (14586: R1 ^reward R1044)
- <=WM: (14571: I3 ^see 0)
- <=WM: (14589: O2082 ^name predict-no)
- <=WM: (14588: O2081 ^name predict-yes)
- <=WM: (14587: R1044 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2083 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2084 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2082 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2081 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.896552,0.0932828)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*44 0.288049 0.322412 0.610461 -> 0.288049 0.322412 0.610461(R,m,v=1,1,0)
- =>WM: (14608: S1 ^operator O2084)
- 1042: O: O2084 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1042 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1041 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14609: I3 ^predict-no N1042)
- <=WM: (14595: N1041 ^status complete)
- <=WM: (14594: I3 ^predict-yes N1041)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- /|\---- Input Phase ---
- =>WM: (14613: I2 ^dir U)
- =>WM: (14612: I2 ^reward 1)
- =>WM: (14611: I2 ^see 0)
- =>WM: (14610: N1042 ^status complete)
- <=WM: (14598: I2 ^dir U)
- <=WM: (14597: I2 ^reward 1)
- <=WM: (14596: I2 ^see 1)
- =>WM: (14614: I2 ^level-1 L1-root)
- <=WM: (14599: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1046 ^value 1 +)
- (R1 ^reward R1046 +)
- Firing propose*predict-yes
- -->
- (O2085 ^name predict-yes +)
- (S1 ^operator O2085 +)
- Firing propose*predict-no
- -->
- (O2086 ^name predict-no +)
- (S1 ^operator O2086 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2084 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2083 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2084 ^name predict-no +)
- (S1 ^operator O2084 +)
- Retracting propose*predict-yes
- -->
- (O2083 ^name predict-yes +)
- (S1 ^operator O2083 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1045 ^value 1 +)
- (R1 ^reward R1045 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2084 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2083 = 0.)
- =>WM: (14621: S1 ^operator O2086 +)
- =>WM: (14620: S1 ^operator O2085 +)
- =>WM: (14619: O2086 ^name predict-no)
- =>WM: (14618: O2085 ^name predict-yes)
- =>WM: (14617: R1046 ^value 1)
- =>WM: (14616: R1 ^reward R1046)
- =>WM: (14615: I3 ^see 0)
- <=WM: (14606: S1 ^operator O2083 +)
- <=WM: (14607: S1 ^operator O2084 +)
- <=WM: (14608: S1 ^operator O2084)
- <=WM: (14601: R1 ^reward R1045)
- <=WM: (14600: I3 ^see 1)
- <=WM: (14604: O2084 ^name predict-no)
- <=WM: (14603: O2083 ^name predict-yes)
- <=WM: (14602: R1045 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2085 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2086 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2084 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2083 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14622: S1 ^operator O2086)
- 1043: O: O2086 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1043 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1042 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14623: I3 ^predict-no N1043)
- <=WM: (14610: N1042 ^status complete)
- <=WM: (14609: I3 ^predict-no N1042)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (14627: I2 ^dir L)
- =>WM: (14626: I2 ^reward 1)
- =>WM: (14625: I2 ^see 0)
- =>WM: (14624: N1043 ^status complete)
- <=WM: (14613: I2 ^dir U)
- <=WM: (14612: I2 ^reward 1)
- <=WM: (14611: I2 ^see 0)
- =>WM: (14628: I2 ^level-1 L1-root)
- <=WM: (14614: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2086 = 0.6126632570275004)
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2085 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1047 ^value 1 +)
- (R1 ^reward R1047 +)
- Firing propose*predict-yes
- -->
- (O2087 ^name predict-yes +)
- (S1 ^operator O2087 +)
- Firing propose*predict-no
- -->
- (O2088 ^name predict-no +)
- (S1 ^operator O2088 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2086 = 0.3873357815900519)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2085 = 0.3895395371935536)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2086 ^name predict-no +)
- (S1 ^operator O2086 +)
- Retracting propose*predict-yes
- -->
- (O2085 ^name predict-yes +)
- (S1 ^operator O2085 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1046 ^value 1 +)
- (R1 ^reward R1046 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2086 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2085 = 0.)
- =>WM: (14635: S1 ^operator O2088 +)
- =>WM: (14634: S1 ^operator O2087 +)
- =>WM: (14633: I3 ^dir L)
- =>WM: (14632: O2088 ^name predict-no)
- =>WM: (14631: O2087 ^name predict-yes)
- =>WM: (14630: R1047 ^value 1)
- =>WM: (14629: R1 ^reward R1047)
- <=WM: (14620: S1 ^operator O2085 +)
- <=WM: (14621: S1 ^operator O2086 +)
- <=WM: (14622: S1 ^operator O2086)
- <=WM: (14605: I3 ^dir U)
- <=WM: (14616: R1 ^reward R1046)
- <=WM: (14619: O2086 ^name predict-no)
- <=WM: (14618: O2085 ^name predict-yes)
- <=WM: (14617: R1046 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2087 = -0.02274740735326741)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2087 = 0.3895395371935536)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2088 = 0.6126632570275004)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2088 = 0.3873357815900519)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2086 = 0.3873357815900519)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2086 = 0.6126632570275004)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2085 = 0.3895395371935536)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2085 = -0.02274740735326741)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14636: S1 ^operator O2088)
- 1044: O: O2088 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1044 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1043 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14637: I3 ^predict-no N1044)
- <=WM: (14624: N1043 ^status complete)
- <=WM: (14623: I3 ^predict-no N1043)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (14641: I2 ^dir U)
- =>WM: (14640: I2 ^reward 1)
- =>WM: (14639: I2 ^see 0)
- =>WM: (14638: N1044 ^status complete)
- <=WM: (14627: I2 ^dir L)
- <=WM: (14626: I2 ^reward 1)
- <=WM: (14625: I2 ^see 0)
- =>WM: (14642: I2 ^level-1 L0-root)
- <=WM: (14628: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1048 ^value 1 +)
- (R1 ^reward R1048 +)
- Firing propose*predict-yes
- -->
- (O2089 ^name predict-yes +)
- (S1 ^operator O2089 +)
- Firing propose*predict-no
- -->
- (O2090 ^name predict-no +)
- (S1 ^operator O2090 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2088 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2087 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2088 ^name predict-no +)
- (S1 ^operator O2088 +)
- Retracting propose*predict-yes
- -->
- (O2087 ^name predict-yes +)
- (S1 ^operator O2087 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1047 ^value 1 +)
- (R1 ^reward R1047 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2088 = 0.3873357815900519)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2088 = 0.6126632570275004)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2087 = 0.3895395371935536)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2087 = -0.02274740735326741)
- =>WM: (14649: S1 ^operator O2090 +)
- =>WM: (14648: S1 ^operator O2089 +)
- =>WM: (14647: I3 ^dir U)
- =>WM: (14646: O2090 ^name predict-no)
- =>WM: (14645: O2089 ^name predict-yes)
- =>WM: (14644: R1048 ^value 1)
- =>WM: (14643: R1 ^reward R1048)
- <=WM: (14634: S1 ^operator O2087 +)
- <=WM: (14635: S1 ^operator O2088 +)
- <=WM: (14636: S1 ^operator O2088)
- <=WM: (14633: I3 ^dir L)
- <=WM: (14629: R1 ^reward R1047)
- <=WM: (14632: O2088 ^name predict-no)
- <=WM: (14631: O2087 ^name predict-yes)
- <=WM: (14630: R1047 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2089 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2090 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2088 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2087 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.719079 -0.331743 0.387336 -> 0.719079 -0.331744 0.387336(R,m,v=1,0.934783,0.0612972)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*32 0.280919 0.331744 0.612663 -> 0.28092 0.331744 0.612663(R,m,v=1,1,0)
- =>WM: (14650: S1 ^operator O2090)
- 1045: O: O2090 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1045 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1044 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14651: I3 ^predict-no N1045)
- <=WM: (14638: N1044 ^status complete)
- <=WM: (14637: I3 ^predict-no N1044)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (14655: I2 ^dir L)
- =>WM: (14654: I2 ^reward 1)
- =>WM: (14653: I2 ^see 0)
- =>WM: (14652: N1045 ^status complete)
- <=WM: (14641: I2 ^dir U)
- <=WM: (14640: I2 ^reward 1)
- <=WM: (14639: I2 ^see 0)
- =>WM: (14656: I2 ^level-1 L0-root)
- <=WM: (14642: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2089 = 0.1599599085218832)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2090 = 0.6126663026263569)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1049 ^value 1 +)
- (R1 ^reward R1049 +)
- Firing propose*predict-yes
- -->
- (O2091 ^name predict-yes +)
- (S1 ^operator O2091 +)
- Firing propose*predict-no
- -->
- (O2092 ^name predict-no +)
- (S1 ^operator O2092 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2090 = 0.3873359257974192)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2089 = 0.3895395371935536)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2090 ^name predict-no +)
- (S1 ^operator O2090 +)
- Retracting propose*predict-yes
- -->
- (O2089 ^name predict-yes +)
- (S1 ^operator O2089 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1048 ^value 1 +)
- (R1 ^reward R1048 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2090 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2089 = 0.)
- =>WM: (14663: S1 ^operator O2092 +)
- =>WM: (14662: S1 ^operator O2091 +)
- =>WM: (14661: I3 ^dir L)
- =>WM: (14660: O2092 ^name predict-no)
- =>WM: (14659: O2091 ^name predict-yes)
- =>WM: (14658: R1049 ^value 1)
- =>WM: (14657: R1 ^reward R1049)
- <=WM: (14648: S1 ^operator O2089 +)
- <=WM: (14649: S1 ^operator O2090 +)
- <=WM: (14650: S1 ^operator O2090)
- <=WM: (14647: I3 ^dir U)
- <=WM: (14643: R1 ^reward R1048)
- <=WM: (14646: O2090 ^name predict-no)
- <=WM: (14645: O2089 ^name predict-yes)
- <=WM: (14644: R1048 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2091 = 0.1599599085218832)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2091 = 0.3895395371935536)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2092 = 0.6126663026263569)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2092 = 0.3873359257974192)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2090 = 0.3873359257974192)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2090 = 0.6126663026263569)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2089 = 0.3895395371935536)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2089 = 0.1599599085218832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14664: S1 ^operator O2092)
- 1046: O: O2092 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1046 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1045 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14665: I3 ^predict-no N1046)
- <=WM: (14652: N1045 ^status complete)
- <=WM: (14651: I3 ^predict-no N1045)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (14669: I2 ^dir U)
- =>WM: (14668: I2 ^reward 1)
- =>WM: (14667: I2 ^see 0)
- =>WM: (14666: N1046 ^status complete)
- <=WM: (14655: I2 ^dir L)
- <=WM: (14654: I2 ^reward 1)
- <=WM: (14653: I2 ^see 0)
- =>WM: (14670: I2 ^level-1 L0-root)
- <=WM: (14656: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1050 ^value 1 +)
- (R1 ^reward R1050 +)
- Firing propose*predict-yes
- -->
- (O2093 ^name predict-yes +)
- (S1 ^operator O2093 +)
- Firing propose*predict-no
- -->
- (O2094 ^name predict-no +)
- (S1 ^operator O2094 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2092 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2091 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2092 ^name predict-no +)
- (S1 ^operator O2092 +)
- Retracting propose*predict-yes
- -->
- (O2091 ^name predict-yes +)
- (S1 ^operator O2091 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1049 ^value 1 +)
- (R1 ^reward R1049 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2092 = 0.3873359257974192)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2092 = 0.6126663026263569)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2091 = 0.3895395371935536)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2091 = 0.1599599085218832)
- =>WM: (14677: S1 ^operator O2094 +)
- =>WM: (14676: S1 ^operator O2093 +)
- =>WM: (14675: I3 ^dir U)
- =>WM: (14674: O2094 ^name predict-no)
- =>WM: (14673: O2093 ^name predict-yes)
- =>WM: (14672: R1050 ^value 1)
- =>WM: (14671: R1 ^reward R1050)
- <=WM: (14662: S1 ^operator O2091 +)
- <=WM: (14663: S1 ^operator O2092 +)
- <=WM: (14664: S1 ^operator O2092)
- <=WM: (14661: I3 ^dir L)
- <=WM: (14657: R1 ^reward R1049)
- <=WM: (14660: O2092 ^name predict-no)
- <=WM: (14659: O2091 ^name predict-yes)
- <=WM: (14658: R1049 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2093 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2094 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2092 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2091 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.719079 -0.331744 0.387336 -> 0.719079 -0.331743 0.387336(R,m,v=1,0.935135,0.0609871)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*37 0.280924 0.331742 0.612666 -> 0.280923 0.331743 0.612666(R,m,v=1,1,0)
- =>WM: (14678: S1 ^operator O2094)
- 1047: O: O2094 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1047 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1046 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14679: I3 ^predict-no N1047)
- <=WM: (14666: N1046 ^status complete)
- <=WM: (14665: I3 ^predict-no N1046)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- /|\---- Input Phase ---
- =>WM: (14683: I2 ^dir U)
- =>WM: (14682: I2 ^reward 1)
- =>WM: (14681: I2 ^see 0)
- =>WM: (14680: N1047 ^status complete)
- <=WM: (14669: I2 ^dir U)
- <=WM: (14668: I2 ^reward 1)
- <=WM: (14667: I2 ^see 0)
- =>WM: (14684: I2 ^level-1 L0-root)
- <=WM: (14670: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1051 ^value 1 +)
- (R1 ^reward R1051 +)
- Firing propose*predict-yes
- -->
- (O2095 ^name predict-yes +)
- (S1 ^operator O2095 +)
- Firing propose*predict-no
- -->
- (O2096 ^name predict-no +)
- (S1 ^operator O2096 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2094 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2093 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2094 ^name predict-no +)
- (S1 ^operator O2094 +)
- Retracting propose*predict-yes
- -->
- (O2093 ^name predict-yes +)
- (S1 ^operator O2093 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1050 ^value 1 +)
- (R1 ^reward R1050 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2094 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2093 = 0.)
- =>WM: (14690: S1 ^operator O2096 +)
- =>WM: (14689: S1 ^operator O2095 +)
- =>WM: (14688: O2096 ^name predict-no)
- =>WM: (14687: O2095 ^name predict-yes)
- =>WM: (14686: R1051 ^value 1)
- =>WM: (14685: R1 ^reward R1051)
- <=WM: (14676: S1 ^operator O2093 +)
- <=WM: (14677: S1 ^operator O2094 +)
- <=WM: (14678: S1 ^operator O2094)
- <=WM: (14671: R1 ^reward R1050)
- <=WM: (14674: O2094 ^name predict-no)
- <=WM: (14673: O2093 ^name predict-yes)
- <=WM: (14672: R1050 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2095 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2096 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2094 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2093 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14691: S1 ^operator O2096)
- 1048: O: O2096 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1048 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1047 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14692: I3 ^predict-no N1048)
- <=WM: (14680: N1047 ^status complete)
- <=WM: (14679: I3 ^predict-no N1047)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- /|\---- Input Phase ---
- =>WM: (14696: I2 ^dir L)
- =>WM: (14695: I2 ^reward 1)
- =>WM: (14694: I2 ^see 0)
- =>WM: (14693: N1048 ^status complete)
- <=WM: (14683: I2 ^dir U)
- <=WM: (14682: I2 ^reward 1)
- <=WM: (14681: I2 ^see 0)
- =>WM: (14697: I2 ^level-1 L0-root)
- <=WM: (14684: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2095 = 0.1599599085218832)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2096 = 0.6126659683627904)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1052 ^value 1 +)
- (R1 ^reward R1052 +)
- Firing propose*predict-yes
- -->
- (O2097 ^name predict-yes +)
- (S1 ^operator O2097 +)
- Firing propose*predict-no
- -->
- (O2098 ^name predict-no +)
- (S1 ^operator O2098 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2096 = 0.3873355915338527)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2095 = 0.3895395371935536)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2096 ^name predict-no +)
- (S1 ^operator O2096 +)
- Retracting propose*predict-yes
- -->
- (O2095 ^name predict-yes +)
- (S1 ^operator O2095 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1051 ^value 1 +)
- (R1 ^reward R1051 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2096 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2095 = 0.)
- =>WM: (14704: S1 ^operator O2098 +)
- =>WM: (14703: S1 ^operator O2097 +)
- =>WM: (14702: I3 ^dir L)
- =>WM: (14701: O2098 ^name predict-no)
- =>WM: (14700: O2097 ^name predict-yes)
- =>WM: (14699: R1052 ^value 1)
- =>WM: (14698: R1 ^reward R1052)
- <=WM: (14689: S1 ^operator O2095 +)
- <=WM: (14690: S1 ^operator O2096 +)
- <=WM: (14691: S1 ^operator O2096)
- <=WM: (14675: I3 ^dir U)
- <=WM: (14685: R1 ^reward R1051)
- <=WM: (14688: O2096 ^name predict-no)
- <=WM: (14687: O2095 ^name predict-yes)
- <=WM: (14686: R1051 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2097 = 0.1599599085218832)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2097 = 0.3895395371935536)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2098 = 0.6126659683627904)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2098 = 0.3873355915338527)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2096 = 0.3873355915338527)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2096 = 0.6126659683627904)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2095 = 0.3895395371935536)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2095 = 0.1599599085218832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14705: S1 ^operator O2098)
- 1049: O: O2098 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1049 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1048 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14706: I3 ^predict-no N1049)
- <=WM: (14693: N1048 ^status complete)
- <=WM: (14692: I3 ^predict-no N1048)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (14710: I2 ^dir R)
- =>WM: (14709: I2 ^reward 1)
- =>WM: (14708: I2 ^see 0)
- =>WM: (14707: N1049 ^status complete)
- <=WM: (14696: I2 ^dir L)
- <=WM: (14695: I2 ^reward 1)
- <=WM: (14694: I2 ^see 0)
- =>WM: (14711: I2 ^level-1 L0-root)
- <=WM: (14697: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2097 = 0.8155909586299994)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2098 = -0.00558448899823713)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1053 ^value 1 +)
- (R1 ^reward R1053 +)
- Firing propose*predict-yes
- -->
- (O2099 ^name predict-yes +)
- (S1 ^operator O2099 +)
- Firing propose*predict-no
- -->
- (O2100 ^name predict-no +)
- (S1 ^operator O2100 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2098 = 0.4476194484475643)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2097 = 0.1844112102474305)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2098 ^name predict-no +)
- (S1 ^operator O2098 +)
- Retracting propose*predict-yes
- -->
- (O2097 ^name predict-yes +)
- (S1 ^operator O2097 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1052 ^value 1 +)
- (R1 ^reward R1052 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2098 = 0.3873355915338527)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2098 = 0.6126659683627904)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2097 = 0.3895395371935536)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2097 = 0.1599599085218832)
- =>WM: (14718: S1 ^operator O2100 +)
- =>WM: (14717: S1 ^operator O2099 +)
- =>WM: (14716: I3 ^dir R)
- =>WM: (14715: O2100 ^name predict-no)
- =>WM: (14714: O2099 ^name predict-yes)
- =>WM: (14713: R1053 ^value 1)
- =>WM: (14712: R1 ^reward R1053)
- <=WM: (14703: S1 ^operator O2097 +)
- <=WM: (14704: S1 ^operator O2098 +)
- <=WM: (14705: S1 ^operator O2098)
- <=WM: (14702: I3 ^dir L)
- <=WM: (14698: R1 ^reward R1052)
- <=WM: (14701: O2098 ^name predict-no)
- <=WM: (14700: O2097 ^name predict-yes)
- <=WM: (14699: R1052 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2099 = 0.8155909586299994)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2099 = 0.1844112102474305)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2100 = -0.00558448899823713)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2100 = 0.4476194484475643)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2098 = 0.4476194484475643)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2098 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2097 = 0.1844112102474305)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2097 = 0.8155909586299994)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.719079 -0.331743 0.387336 -> 0.719079 -0.331743 0.387335(R,m,v=1,0.935484,0.06068)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*37 0.280923 0.331743 0.612666 -> 0.280923 0.331743 0.612666(R,m,v=1,1,0)
- =>WM: (14719: S1 ^operator O2099)
- 1050: O: O2099 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1050 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1049 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14720: I3 ^predict-yes N1050)
- <=WM: (14707: N1049 ^status complete)
- <=WM: (14706: I3 ^predict-no N1049)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- -/--- Input Phase ---
- =>WM: (14724: I2 ^dir R)
- =>WM: (14723: I2 ^reward 1)
- =>WM: (14722: I2 ^see 1)
- =>WM: (14721: N1050 ^status complete)
- <=WM: (14710: I2 ^dir R)
- <=WM: (14709: I2 ^reward 1)
- <=WM: (14708: I2 ^see 0)
- =>WM: (14725: I2 ^level-1 R1-root)
- <=WM: (14711: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2099 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2100 = 0.5523812848757654)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1054 ^value 1 +)
- (R1 ^reward R1054 +)
- Firing propose*predict-yes
- -->
- (O2101 ^name predict-yes +)
- (S1 ^operator O2101 +)
- Firing propose*predict-no
- -->
- (O2102 ^name predict-no +)
- (S1 ^operator O2102 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2100 = 0.4476194484475643)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2099 = 0.1844112102474305)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2100 ^name predict-no +)
- (S1 ^operator O2100 +)
- Retracting propose*predict-yes
- -->
- (O2099 ^name predict-yes +)
- (S1 ^operator O2099 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1053 ^value 1 +)
- (R1 ^reward R1053 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2100 = 0.4476194484475643)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2100 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2099 = 0.1844112102474305)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2099 = 0.8155909586299994)
- =>WM: (14732: S1 ^operator O2102 +)
- =>WM: (14731: S1 ^operator O2101 +)
- =>WM: (14730: O2102 ^name predict-no)
- =>WM: (14729: O2101 ^name predict-yes)
- =>WM: (14728: R1054 ^value 1)
- =>WM: (14727: R1 ^reward R1054)
- =>WM: (14726: I3 ^see 1)
- <=WM: (14717: S1 ^operator O2099 +)
- <=WM: (14719: S1 ^operator O2099)
- <=WM: (14718: S1 ^operator O2100 +)
- <=WM: (14712: R1 ^reward R1053)
- <=WM: (14615: I3 ^see 0)
- <=WM: (14715: O2100 ^name predict-no)
- <=WM: (14714: O2099 ^name predict-yes)
- <=WM: (14713: R1053 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2101 = 0.1844112102474305)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2101 = 0.1398795999120246)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2102 = 0.4476194484475643)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2102 = 0.5523812848757654)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2100 = 0.4476194484475643)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2100 = 0.5523812848757654)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2099 = 0.1844112102474305)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2099 = 0.1398795999120246)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675414 -0.491003 0.184411 -> 0.675413 -0.491003 0.184411(R,m,v=1,0.903955,0.0873138)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*34 0.324589 0.491002 0.815591 -> 0.324589 0.491002 0.815591(R,m,v=1,1,0)
- =>WM: (14733: S1 ^operator O2102)
- 1051: O: O2102 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1051 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1050 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14734: I3 ^predict-no N1051)
- <=WM: (14721: N1050 ^status complete)
- <=WM: (14720: I3 ^predict-yes N1050)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- |--- Input Phase ---
- =>WM: (14738: I2 ^dir U)
- =>WM: (14737: I2 ^reward 1)
- =>WM: (14736: I2 ^see 0)
- =>WM: (14735: N1051 ^status complete)
- <=WM: (14724: I2 ^dir R)
- <=WM: (14723: I2 ^reward 1)
- <=WM: (14722: I2 ^see 1)
- =>WM: (14739: I2 ^level-1 R0-root)
- <=WM: (14725: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1055 ^value 1 +)
- (R1 ^reward R1055 +)
- Firing propose*predict-yes
- -->
- (O2103 ^name predict-yes +)
- (S1 ^operator O2103 +)
- Firing propose*predict-no
- -->
- (O2104 ^name predict-no +)
- (S1 ^operator O2104 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2102 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2101 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2102 ^name predict-no +)
- (S1 ^operator O2102 +)
- Retracting propose*predict-yes
- -->
- (O2101 ^name predict-yes +)
- (S1 ^operator O2101 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1054 ^value 1 +)
- (R1 ^reward R1054 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2102 = 0.5523812848757654)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2102 = 0.4476194484475643)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2101 = 0.1398795999120246)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2101 = 0.1844108849158159)
- =>WM: (14747: S1 ^operator O2104 +)
- =>WM: (14746: S1 ^operator O2103 +)
- =>WM: (14745: I3 ^dir U)
- =>WM: (14744: O2104 ^name predict-no)
- =>WM: (14743: O2103 ^name predict-yes)
- =>WM: (14742: R1055 ^value 1)
- =>WM: (14741: R1 ^reward R1055)
- =>WM: (14740: I3 ^see 0)
- <=WM: (14731: S1 ^operator O2101 +)
- <=WM: (14732: S1 ^operator O2102 +)
- <=WM: (14733: S1 ^operator O2102)
- <=WM: (14716: I3 ^dir R)
- <=WM: (14727: R1 ^reward R1054)
- <=WM: (14726: I3 ^see 1)
- <=WM: (14730: O2102 ^name predict-no)
- <=WM: (14729: O2101 ^name predict-yes)
- <=WM: (14728: R1054 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2103 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2104 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2102 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2101 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.934783,0.0614091)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*39 0.377468 0.174914 0.552381 -> 0.377468 0.174914 0.552381(R,m,v=1,1,0)
- =>WM: (14748: S1 ^operator O2104)
- 1052: O: O2104 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1052 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1051 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14749: I3 ^predict-no N1052)
- <=WM: (14735: N1051 ^status complete)
- <=WM: (14734: I3 ^predict-no N1051)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (14753: I2 ^dir U)
- =>WM: (14752: I2 ^reward 1)
- =>WM: (14751: I2 ^see 0)
- =>WM: (14750: N1052 ^status complete)
- <=WM: (14738: I2 ^dir U)
- <=WM: (14737: I2 ^reward 1)
- <=WM: (14736: I2 ^see 0)
- =>WM: (14754: I2 ^level-1 R0-root)
- <=WM: (14739: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1056 ^value 1 +)
- (R1 ^reward R1056 +)
- Firing propose*predict-yes
- -->
- (O2105 ^name predict-yes +)
- (S1 ^operator O2105 +)
- Firing propose*predict-no
- -->
- (O2106 ^name predict-no +)
- (S1 ^operator O2106 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2104 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2103 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2104 ^name predict-no +)
- (S1 ^operator O2104 +)
- Retracting propose*predict-yes
- -->
- (O2103 ^name predict-yes +)
- (S1 ^operator O2103 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1055 ^value 1 +)
- (R1 ^reward R1055 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2104 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2103 = 0.)
- =>WM: (14760: S1 ^operator O2106 +)
- =>WM: (14759: S1 ^operator O2105 +)
- =>WM: (14758: O2106 ^name predict-no)
- =>WM: (14757: O2105 ^name predict-yes)
- =>WM: (14756: R1056 ^value 1)
- =>WM: (14755: R1 ^reward R1056)
- <=WM: (14746: S1 ^operator O2103 +)
- <=WM: (14747: S1 ^operator O2104 +)
- <=WM: (14748: S1 ^operator O2104)
- <=WM: (14741: R1 ^reward R1055)
- <=WM: (14744: O2104 ^name predict-no)
- <=WM: (14743: O2103 ^name predict-yes)
- <=WM: (14742: R1055 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2105 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2106 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2104 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2103 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14761: S1 ^operator O2106)
- 1053: O: O2106 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1053 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1052 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14762: I3 ^predict-no N1053)
- <=WM: (14750: N1052 ^status complete)
- <=WM: (14749: I3 ^predict-no N1052)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (14766: I2 ^dir L)
- =>WM: (14765: I2 ^reward 1)
- =>WM: (14764: I2 ^see 0)
- =>WM: (14763: N1053 ^status complete)
- <=WM: (14753: I2 ^dir U)
- <=WM: (14752: I2 ^reward 1)
- <=WM: (14751: I2 ^see 0)
- =>WM: (14767: I2 ^level-1 R0-root)
- <=WM: (14754: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2105 = 0.6104607684602532)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2106 = 0.1063475139796038)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1057 ^value 1 +)
- (R1 ^reward R1057 +)
- Firing propose*predict-yes
- -->
- (O2107 ^name predict-yes +)
- (S1 ^operator O2107 +)
- Firing propose*predict-no
- -->
- (O2108 ^name predict-no +)
- (S1 ^operator O2108 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2106 = 0.3873353575493562)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2105 = 0.3895395371935536)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2106 ^name predict-no +)
- (S1 ^operator O2106 +)
- Retracting propose*predict-yes
- -->
- (O2105 ^name predict-yes +)
- (S1 ^operator O2105 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1056 ^value 1 +)
- (R1 ^reward R1056 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2106 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2105 = 0.)
- =>WM: (14774: S1 ^operator O2108 +)
- =>WM: (14773: S1 ^operator O2107 +)
- =>WM: (14772: I3 ^dir L)
- =>WM: (14771: O2108 ^name predict-no)
- =>WM: (14770: O2107 ^name predict-yes)
- =>WM: (14769: R1057 ^value 1)
- =>WM: (14768: R1 ^reward R1057)
- <=WM: (14759: S1 ^operator O2105 +)
- <=WM: (14760: S1 ^operator O2106 +)
- <=WM: (14761: S1 ^operator O2106)
- <=WM: (14745: I3 ^dir U)
- <=WM: (14755: R1 ^reward R1056)
- <=WM: (14758: O2106 ^name predict-no)
- <=WM: (14757: O2105 ^name predict-yes)
- <=WM: (14756: R1056 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2107 = 0.6104607684602532)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2107 = 0.3895395371935536)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2108 = 0.1063475139796038)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2108 = 0.3873353575493562)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2106 = 0.3873353575493562)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2106 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2105 = 0.3895395371935536)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2105 = 0.6104607684602532)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14775: S1 ^operator O2107)
- 1054: O: O2107 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1054 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1053 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14776: I3 ^predict-yes N1054)
- <=WM: (14763: N1053 ^status complete)
- <=WM: (14762: I3 ^predict-no N1053)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (14780: I2 ^dir L)
- =>WM: (14779: I2 ^reward 1)
- =>WM: (14778: I2 ^see 1)
- =>WM: (14777: N1054 ^status complete)
- <=WM: (14766: I2 ^dir L)
- <=WM: (14765: I2 ^reward 1)
- <=WM: (14764: I2 ^see 0)
- =>WM: (14781: I2 ^level-1 L1-root)
- <=WM: (14767: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2108 = 0.6126634012348675)
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2107 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1058 ^value 1 +)
- (R1 ^reward R1058 +)
- Firing propose*predict-yes
- -->
- (O2109 ^name predict-yes +)
- (S1 ^operator O2109 +)
- Firing propose*predict-no
- -->
- (O2110 ^name predict-no +)
- (S1 ^operator O2110 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2108 = 0.3873353575493562)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2107 = 0.3895395371935536)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2108 ^name predict-no +)
- (S1 ^operator O2108 +)
- Retracting propose*predict-yes
- -->
- (O2107 ^name predict-yes +)
- (S1 ^operator O2107 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1057 ^value 1 +)
- (R1 ^reward R1057 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2108 = 0.3873353575493562)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2108 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2107 = 0.3895395371935536)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2107 = 0.6104607684602532)
- =>WM: (14788: S1 ^operator O2110 +)
- =>WM: (14787: S1 ^operator O2109 +)
- =>WM: (14786: O2110 ^name predict-no)
- =>WM: (14785: O2109 ^name predict-yes)
- =>WM: (14784: R1058 ^value 1)
- =>WM: (14783: R1 ^reward R1058)
- =>WM: (14782: I3 ^see 1)
- <=WM: (14773: S1 ^operator O2107 +)
- <=WM: (14775: S1 ^operator O2107)
- <=WM: (14774: S1 ^operator O2108 +)
- <=WM: (14768: R1 ^reward R1057)
- <=WM: (14740: I3 ^see 0)
- <=WM: (14771: O2108 ^name predict-no)
- <=WM: (14770: O2107 ^name predict-yes)
- <=WM: (14769: R1057 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2109 = 0.3895395371935536)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2109 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2110 = 0.3873353575493562)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2110 = 0.6126634012348675)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2108 = 0.3873353575493562)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2108 = 0.6126634012348675)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2107 = 0.3895395371935536)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2107 = -0.02274740735326741)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.389539(R,m,v=1,0.897143,0.0928079)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*44 0.288049 0.322412 0.610461 -> 0.288049 0.322412 0.610461(R,m,v=1,1,0)
- =>WM: (14789: S1 ^operator O2110)
- 1055: O: O2110 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1055 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1054 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14790: I3 ^predict-no N1055)
- <=WM: (14777: N1054 ^status complete)
- <=WM: (14776: I3 ^predict-yes N1054)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- -/--- Input Phase ---
- =>WM: (14794: I2 ^dir R)
- =>WM: (14793: I2 ^reward 1)
- =>WM: (14792: I2 ^see 0)
- =>WM: (14791: N1055 ^status complete)
- <=WM: (14780: I2 ^dir L)
- <=WM: (14779: I2 ^reward 1)
- <=WM: (14778: I2 ^see 1)
- =>WM: (14795: I2 ^level-1 L0-root)
- <=WM: (14781: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2109 = 0.8155906332983849)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2110 = -0.00558448899823713)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1059 ^value 1 +)
- (R1 ^reward R1059 +)
- Firing propose*predict-yes
- -->
- (O2111 ^name predict-yes +)
- (S1 ^operator O2111 +)
- Firing propose*predict-no
- -->
- (O2112 ^name predict-no +)
- (S1 ^operator O2112 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2110 = 0.4476193384490649)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2109 = 0.1844108849158159)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2110 ^name predict-no +)
- (S1 ^operator O2110 +)
- Retracting propose*predict-yes
- -->
- (O2109 ^name predict-yes +)
- (S1 ^operator O2109 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1058 ^value 1 +)
- (R1 ^reward R1058 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2110 = 0.6126634012348675)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2110 = 0.3873353575493562)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2109 = -0.02274740735326741)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2109 = 0.3895394913454826)
- =>WM: (14803: S1 ^operator O2112 +)
- =>WM: (14802: S1 ^operator O2111 +)
- =>WM: (14801: I3 ^dir R)
- =>WM: (14800: O2112 ^name predict-no)
- =>WM: (14799: O2111 ^name predict-yes)
- =>WM: (14798: R1059 ^value 1)
- =>WM: (14797: R1 ^reward R1059)
- =>WM: (14796: I3 ^see 0)
- <=WM: (14787: S1 ^operator O2109 +)
- <=WM: (14788: S1 ^operator O2110 +)
- <=WM: (14789: S1 ^operator O2110)
- <=WM: (14772: I3 ^dir L)
- <=WM: (14783: R1 ^reward R1058)
- <=WM: (14782: I3 ^see 1)
- <=WM: (14786: O2110 ^name predict-no)
- <=WM: (14785: O2109 ^name predict-yes)
- <=WM: (14784: R1058 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2111 = 0.1844108849158159)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2111 = 0.8155906332983849)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2112 = 0.4476193384490649)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2112 = -0.00558448899823713)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2110 = 0.4476193384490649)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2110 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2109 = 0.1844108849158159)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2109 = 0.8155906332983849)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.719079 -0.331743 0.387335 -> 0.719079 -0.331743 0.387336(R,m,v=1,0.935829,0.0603761)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*32 0.28092 0.331744 0.612663 -> 0.28092 0.331744 0.612664(R,m,v=1,1,0)
- =>WM: (14804: S1 ^operator O2111)
- 1056: O: O2111 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1056 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1055 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14805: I3 ^predict-yes N1056)
- <=WM: (14791: N1055 ^status complete)
- <=WM: (14790: I3 ^predict-no N1055)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (14809: I2 ^dir U)
- =>WM: (14808: I2 ^reward 1)
- =>WM: (14807: I2 ^see 1)
- =>WM: (14806: N1056 ^status complete)
- <=WM: (14794: I2 ^dir R)
- <=WM: (14793: I2 ^reward 1)
- <=WM: (14792: I2 ^see 0)
- =>WM: (14810: I2 ^level-1 R1-root)
- <=WM: (14795: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1060 ^value 1 +)
- (R1 ^reward R1060 +)
- Firing propose*predict-yes
- -->
- (O2113 ^name predict-yes +)
- (S1 ^operator O2113 +)
- Firing propose*predict-no
- -->
- (O2114 ^name predict-no +)
- (S1 ^operator O2114 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2112 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2111 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2112 ^name predict-no +)
- (S1 ^operator O2112 +)
- Retracting propose*predict-yes
- -->
- (O2111 ^name predict-yes +)
- (S1 ^operator O2111 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1059 ^value 1 +)
- (R1 ^reward R1059 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2112 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2112 = 0.4476193384490649)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2111 = 0.8155906332983849)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2111 = 0.1844108849158159)
- =>WM: (14818: S1 ^operator O2114 +)
- =>WM: (14817: S1 ^operator O2113 +)
- =>WM: (14816: I3 ^dir U)
- =>WM: (14815: O2114 ^name predict-no)
- =>WM: (14814: O2113 ^name predict-yes)
- =>WM: (14813: R1060 ^value 1)
- =>WM: (14812: R1 ^reward R1060)
- =>WM: (14811: I3 ^see 1)
- <=WM: (14802: S1 ^operator O2111 +)
- <=WM: (14804: S1 ^operator O2111)
- <=WM: (14803: S1 ^operator O2112 +)
- <=WM: (14801: I3 ^dir R)
- <=WM: (14797: R1 ^reward R1059)
- <=WM: (14796: I3 ^see 0)
- <=WM: (14800: O2112 ^name predict-no)
- <=WM: (14799: O2111 ^name predict-yes)
- <=WM: (14798: R1059 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2113 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2114 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2112 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2111 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675413 -0.491003 0.184411 -> 0.675413 -0.491002 0.184411(R,m,v=1,0.904494,0.0868723)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*34 0.324589 0.491002 0.815591 -> 0.324588 0.491002 0.81559(R,m,v=1,1,0)
- =>WM: (14819: S1 ^operator O2114)
- 1057: O: O2114 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1057 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1056 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14820: I3 ^predict-no N1057)
- <=WM: (14806: N1056 ^status complete)
- <=WM: (14805: I3 ^predict-yes N1056)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (14824: I2 ^dir U)
- =>WM: (14823: I2 ^reward 1)
- =>WM: (14822: I2 ^see 0)
- =>WM: (14821: N1057 ^status complete)
- <=WM: (14809: I2 ^dir U)
- <=WM: (14808: I2 ^reward 1)
- <=WM: (14807: I2 ^see 1)
- =>WM: (14825: I2 ^level-1 R1-root)
- <=WM: (14810: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1061 ^value 1 +)
- (R1 ^reward R1061 +)
- Firing propose*predict-yes
- -->
- (O2115 ^name predict-yes +)
- (S1 ^operator O2115 +)
- Firing propose*predict-no
- -->
- (O2116 ^name predict-no +)
- (S1 ^operator O2116 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2114 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2113 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2114 ^name predict-no +)
- (S1 ^operator O2114 +)
- Retracting propose*predict-yes
- -->
- (O2113 ^name predict-yes +)
- (S1 ^operator O2113 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1060 ^value 1 +)
- (R1 ^reward R1060 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2114 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2113 = 0.)
- =>WM: (14832: S1 ^operator O2116 +)
- =>WM: (14831: S1 ^operator O2115 +)
- =>WM: (14830: O2116 ^name predict-no)
- =>WM: (14829: O2115 ^name predict-yes)
- =>WM: (14828: R1061 ^value 1)
- =>WM: (14827: R1 ^reward R1061)
- =>WM: (14826: I3 ^see 0)
- <=WM: (14817: S1 ^operator O2113 +)
- <=WM: (14818: S1 ^operator O2114 +)
- <=WM: (14819: S1 ^operator O2114)
- <=WM: (14812: R1 ^reward R1060)
- <=WM: (14811: I3 ^see 1)
- <=WM: (14815: O2114 ^name predict-no)
- <=WM: (14814: O2113 ^name predict-yes)
- <=WM: (14813: R1060 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2115 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2116 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2114 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2113 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14833: S1 ^operator O2116)
- 1058: O: O2116 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1058 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1057 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14834: I3 ^predict-no N1058)
- <=WM: (14821: N1057 ^status complete)
- <=WM: (14820: I3 ^predict-no N1057)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- -/|\--- Input Phase ---
- =>WM: (14838: I2 ^dir R)
- =>WM: (14837: I2 ^reward 1)
- =>WM: (14836: I2 ^see 0)
- =>WM: (14835: N1058 ^status complete)
- <=WM: (14824: I2 ^dir U)
- <=WM: (14823: I2 ^reward 1)
- <=WM: (14822: I2 ^see 0)
- =>WM: (14839: I2 ^level-1 R1-root)
- <=WM: (14825: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2115 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2116 = 0.552381174877266)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1062 ^value 1 +)
- (R1 ^reward R1062 +)
- Firing propose*predict-yes
- -->
- (O2117 ^name predict-yes +)
- (S1 ^operator O2117 +)
- Firing propose*predict-no
- -->
- (O2118 ^name predict-no +)
- (S1 ^operator O2118 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2116 = 0.4476193384490649)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2115 = 0.1844106571836858)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2116 ^name predict-no +)
- (S1 ^operator O2116 +)
- Retracting propose*predict-yes
- -->
- (O2115 ^name predict-yes +)
- (S1 ^operator O2115 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1061 ^value 1 +)
- (R1 ^reward R1061 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2116 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2115 = 0.)
- =>WM: (14846: S1 ^operator O2118 +)
- =>WM: (14845: S1 ^operator O2117 +)
- =>WM: (14844: I3 ^dir R)
- =>WM: (14843: O2118 ^name predict-no)
- =>WM: (14842: O2117 ^name predict-yes)
- =>WM: (14841: R1062 ^value 1)
- =>WM: (14840: R1 ^reward R1062)
- <=WM: (14831: S1 ^operator O2115 +)
- <=WM: (14832: S1 ^operator O2116 +)
- <=WM: (14833: S1 ^operator O2116)
- <=WM: (14816: I3 ^dir U)
- <=WM: (14827: R1 ^reward R1061)
- <=WM: (14830: O2116 ^name predict-no)
- <=WM: (14829: O2115 ^name predict-yes)
- <=WM: (14828: R1061 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2117 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2117 = 0.1844106571836858)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2118 = 0.552381174877266)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2118 = 0.4476193384490649)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2116 = 0.4476193384490649)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2116 = 0.552381174877266)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2115 = 0.1844106571836858)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2115 = 0.1398795999120246)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14847: S1 ^operator O2118)
- 1059: O: O2118 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1059 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1058 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14848: I3 ^predict-no N1059)
- <=WM: (14835: N1058 ^status complete)
- <=WM: (14834: I3 ^predict-no N1058)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- -/--- Input Phase ---
- =>WM: (14852: I2 ^dir U)
- =>WM: (14851: I2 ^reward 1)
- =>WM: (14850: I2 ^see 0)
- =>WM: (14849: N1059 ^status complete)
- <=WM: (14838: I2 ^dir R)
- <=WM: (14837: I2 ^reward 1)
- <=WM: (14836: I2 ^see 0)
- =>WM: (14853: I2 ^level-1 R0-root)
- <=WM: (14839: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1063 ^value 1 +)
- (R1 ^reward R1063 +)
- Firing propose*predict-yes
- -->
- (O2119 ^name predict-yes +)
- (S1 ^operator O2119 +)
- Firing propose*predict-no
- -->
- (O2120 ^name predict-no +)
- (S1 ^operator O2120 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2118 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2117 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2118 ^name predict-no +)
- (S1 ^operator O2118 +)
- Retracting propose*predict-yes
- -->
- (O2117 ^name predict-yes +)
- (S1 ^operator O2117 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1062 ^value 1 +)
- (R1 ^reward R1062 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2118 = 0.4476193384490649)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2118 = 0.552381174877266)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2117 = 0.1844106571836858)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2117 = 0.1398795999120246)
- =>WM: (14860: S1 ^operator O2120 +)
- =>WM: (14859: S1 ^operator O2119 +)
- =>WM: (14858: I3 ^dir U)
- =>WM: (14857: O2120 ^name predict-no)
- =>WM: (14856: O2119 ^name predict-yes)
- =>WM: (14855: R1063 ^value 1)
- =>WM: (14854: R1 ^reward R1063)
- <=WM: (14845: S1 ^operator O2117 +)
- <=WM: (14846: S1 ^operator O2118 +)
- <=WM: (14847: S1 ^operator O2118)
- <=WM: (14844: I3 ^dir R)
- <=WM: (14840: R1 ^reward R1062)
- <=WM: (14843: O2118 ^name predict-no)
- <=WM: (14842: O2117 ^name predict-yes)
- <=WM: (14841: R1062 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2119 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2120 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2118 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2117 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.935252,0.0609947)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*39 0.377468 0.174914 0.552381 -> 0.377467 0.174914 0.552381(R,m,v=1,1,0)
- =>WM: (14861: S1 ^operator O2120)
- 1060: O: O2120 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1060 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1059 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14862: I3 ^predict-no N1060)
- <=WM: (14849: N1059 ^status complete)
- <=WM: (14848: I3 ^predict-no N1059)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (14866: I2 ^dir U)
- =>WM: (14865: I2 ^reward 1)
- =>WM: (14864: I2 ^see 0)
- =>WM: (14863: N1060 ^status complete)
- <=WM: (14852: I2 ^dir U)
- <=WM: (14851: I2 ^reward 1)
- <=WM: (14850: I2 ^see 0)
- =>WM: (14867: I2 ^level-1 R0-root)
- <=WM: (14853: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1064 ^value 1 +)
- (R1 ^reward R1064 +)
- Firing propose*predict-yes
- -->
- (O2121 ^name predict-yes +)
- (S1 ^operator O2121 +)
- Firing propose*predict-no
- -->
- (O2122 ^name predict-no +)
- (S1 ^operator O2122 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2120 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2119 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2120 ^name predict-no +)
- (S1 ^operator O2120 +)
- Retracting propose*predict-yes
- -->
- (O2119 ^name predict-yes +)
- (S1 ^operator O2119 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1063 ^value 1 +)
- (R1 ^reward R1063 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2120 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2119 = 0.)
- =>WM: (14873: S1 ^operator O2122 +)
- =>WM: (14872: S1 ^operator O2121 +)
- =>WM: (14871: O2122 ^name predict-no)
- =>WM: (14870: O2121 ^name predict-yes)
- =>WM: (14869: R1064 ^value 1)
- =>WM: (14868: R1 ^reward R1064)
- <=WM: (14859: S1 ^operator O2119 +)
- <=WM: (14860: S1 ^operator O2120 +)
- <=WM: (14861: S1 ^operator O2120)
- <=WM: (14854: R1 ^reward R1063)
- <=WM: (14857: O2120 ^name predict-no)
- <=WM: (14856: O2119 ^name predict-yes)
- <=WM: (14855: R1063 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2121 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2122 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2120 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2119 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14874: S1 ^operator O2122)
- 1061: O: O2122 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1061 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1060 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14875: I3 ^predict-no N1061)
- <=WM: (14863: N1060 ^status complete)
- <=WM: (14862: I3 ^predict-no N1060)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- /--- Input Phase ---
- =>WM: (14879: I2 ^dir U)
- =>WM: (14878: I2 ^reward 1)
- =>WM: (14877: I2 ^see 0)
- =>WM: (14876: N1061 ^status complete)
- <=WM: (14866: I2 ^dir U)
- <=WM: (14865: I2 ^reward 1)
- <=WM: (14864: I2 ^see 0)
- =>WM: (14880: I2 ^level-1 R0-root)
- <=WM: (14867: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1065 ^value 1 +)
- (R1 ^reward R1065 +)
- Firing propose*predict-yes
- -->
- (O2123 ^name predict-yes +)
- (S1 ^operator O2123 +)
- Firing propose*predict-no
- -->
- (O2124 ^name predict-no +)
- (S1 ^operator O2124 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2122 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2121 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2122 ^name predict-no +)
- (S1 ^operator O2122 +)
- Retracting propose*predict-yes
- -->
- (O2121 ^name predict-yes +)
- (S1 ^operator O2121 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1064 ^value 1 +)
- (R1 ^reward R1064 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2122 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2121 = 0.)
- =>WM: (14886: S1 ^operator O2124 +)
- =>WM: (14885: S1 ^operator O2123 +)
- =>WM: (14884: O2124 ^name predict-no)
- =>WM: (14883: O2123 ^name predict-yes)
- =>WM: (14882: R1065 ^value 1)
- =>WM: (14881: R1 ^reward R1065)
- <=WM: (14872: S1 ^operator O2121 +)
- <=WM: (14873: S1 ^operator O2122 +)
- <=WM: (14874: S1 ^operator O2122)
- <=WM: (14868: R1 ^reward R1064)
- <=WM: (14871: O2122 ^name predict-no)
- <=WM: (14870: O2121 ^name predict-yes)
- <=WM: (14869: R1064 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2123 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2124 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2122 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2121 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14887: S1 ^operator O2124)
- 1062: O: O2124 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1062 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1061 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14888: I3 ^predict-no N1062)
- <=WM: (14876: N1061 ^status complete)
- <=WM: (14875: I3 ^predict-no N1061)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (14892: I2 ^dir R)
- =>WM: (14891: I2 ^reward 1)
- =>WM: (14890: I2 ^see 0)
- =>WM: (14889: N1062 ^status complete)
- <=WM: (14879: I2 ^dir U)
- <=WM: (14878: I2 ^reward 1)
- <=WM: (14877: I2 ^see 0)
- =>WM: (14893: I2 ^level-1 R0-root)
- <=WM: (14880: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2123 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2124 = 0.5523799932145873)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1066 ^value 1 +)
- (R1 ^reward R1066 +)
- Firing propose*predict-yes
- -->
- (O2125 ^name predict-yes +)
- (S1 ^operator O2125 +)
- Firing propose*predict-no
- -->
- (O2126 ^name predict-no +)
- (S1 ^operator O2126 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2124 = 0.4476192614501152)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2123 = 0.1844106571836858)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2124 ^name predict-no +)
- (S1 ^operator O2124 +)
- Retracting propose*predict-yes
- -->
- (O2123 ^name predict-yes +)
- (S1 ^operator O2123 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1065 ^value 1 +)
- (R1 ^reward R1065 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2124 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2123 = 0.)
- =>WM: (14900: S1 ^operator O2126 +)
- =>WM: (14899: S1 ^operator O2125 +)
- =>WM: (14898: I3 ^dir R)
- =>WM: (14897: O2126 ^name predict-no)
- =>WM: (14896: O2125 ^name predict-yes)
- =>WM: (14895: R1066 ^value 1)
- =>WM: (14894: R1 ^reward R1066)
- <=WM: (14885: S1 ^operator O2123 +)
- <=WM: (14886: S1 ^operator O2124 +)
- <=WM: (14887: S1 ^operator O2124)
- <=WM: (14858: I3 ^dir U)
- <=WM: (14881: R1 ^reward R1065)
- <=WM: (14884: O2124 ^name predict-no)
- <=WM: (14883: O2123 ^name predict-yes)
- <=WM: (14882: R1065 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2125 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2125 = 0.1844106571836858)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2126 = 0.5523799932145873)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2126 = 0.4476192614501152)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2124 = 0.4476192614501152)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2124 = 0.5523799932145873)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2123 = 0.1844106571836858)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2123 = 0.1664311307472832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14901: S1 ^operator O2126)
- 1063: O: O2126 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1063 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1062 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14902: I3 ^predict-no N1063)
- <=WM: (14889: N1062 ^status complete)
- <=WM: (14888: I3 ^predict-no N1062)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (14906: I2 ^dir L)
- =>WM: (14905: I2 ^reward 1)
- =>WM: (14904: I2 ^see 0)
- =>WM: (14903: N1063 ^status complete)
- <=WM: (14892: I2 ^dir R)
- <=WM: (14891: I2 ^reward 1)
- <=WM: (14890: I2 ^see 0)
- =>WM: (14907: I2 ^level-1 R0-root)
- <=WM: (14893: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2125 = 0.6104607226121822)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2126 = 0.1063475139796038)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1067 ^value 1 +)
- (R1 ^reward R1067 +)
- Firing propose*predict-yes
- -->
- (O2127 ^name predict-yes +)
- (S1 ^operator O2127 +)
- Firing propose*predict-no
- -->
- (O2128 ^name predict-no +)
- (S1 ^operator O2128 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2126 = 0.3873355437317227)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2125 = 0.3895394913454826)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2126 ^name predict-no +)
- (S1 ^operator O2126 +)
- Retracting propose*predict-yes
- -->
- (O2125 ^name predict-yes +)
- (S1 ^operator O2125 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1066 ^value 1 +)
- (R1 ^reward R1066 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2126 = 0.4476192614501152)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2126 = 0.5523799932145873)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2125 = 0.1844106571836858)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2125 = 0.1664311307472832)
- =>WM: (14914: S1 ^operator O2128 +)
- =>WM: (14913: S1 ^operator O2127 +)
- =>WM: (14912: I3 ^dir L)
- =>WM: (14911: O2128 ^name predict-no)
- =>WM: (14910: O2127 ^name predict-yes)
- =>WM: (14909: R1067 ^value 1)
- =>WM: (14908: R1 ^reward R1067)
- <=WM: (14899: S1 ^operator O2125 +)
- <=WM: (14900: S1 ^operator O2126 +)
- <=WM: (14901: S1 ^operator O2126)
- <=WM: (14898: I3 ^dir R)
- <=WM: (14894: R1 ^reward R1066)
- <=WM: (14897: O2126 ^name predict-no)
- <=WM: (14896: O2125 ^name predict-yes)
- <=WM: (14895: R1066 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2127 = 0.6104607226121822)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2127 = 0.3895394913454826)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2128 = 0.1063475139796038)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2128 = 0.3873355437317227)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2126 = 0.3873355437317227)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2126 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2125 = 0.3895394913454826)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2125 = 0.6104607226121822)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.935714,0.0605858)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*41 0.377467 0.174913 0.55238 -> 0.377467 0.174913 0.55238(R,m,v=1,1,0)
- =>WM: (14915: S1 ^operator O2127)
- 1064: O: O2127 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1064 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1063 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14916: I3 ^predict-yes N1064)
- <=WM: (14903: N1063 ^status complete)
- <=WM: (14902: I3 ^predict-no N1063)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- -/--- Input Phase ---
- =>WM: (14920: I2 ^dir U)
- =>WM: (14919: I2 ^reward 1)
- =>WM: (14918: I2 ^see 1)
- =>WM: (14917: N1064 ^status complete)
- <=WM: (14906: I2 ^dir L)
- <=WM: (14905: I2 ^reward 1)
- <=WM: (14904: I2 ^see 0)
- =>WM: (14921: I2 ^level-1 L1-root)
- <=WM: (14907: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1068 ^value 1 +)
- (R1 ^reward R1068 +)
- Firing propose*predict-yes
- -->
- (O2129 ^name predict-yes +)
- (S1 ^operator O2129 +)
- Firing propose*predict-no
- -->
- (O2130 ^name predict-no +)
- (S1 ^operator O2130 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2128 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2127 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2128 ^name predict-no +)
- (S1 ^operator O2128 +)
- Retracting propose*predict-yes
- -->
- (O2127 ^name predict-yes +)
- (S1 ^operator O2127 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1067 ^value 1 +)
- (R1 ^reward R1067 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2128 = 0.3873355437317227)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2128 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2127 = 0.3895394913454826)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2127 = 0.6104607226121822)
- =>WM: (14929: S1 ^operator O2130 +)
- =>WM: (14928: S1 ^operator O2129 +)
- =>WM: (14927: I3 ^dir U)
- =>WM: (14926: O2130 ^name predict-no)
- =>WM: (14925: O2129 ^name predict-yes)
- =>WM: (14924: R1068 ^value 1)
- =>WM: (14923: R1 ^reward R1068)
- =>WM: (14922: I3 ^see 1)
- <=WM: (14913: S1 ^operator O2127 +)
- <=WM: (14915: S1 ^operator O2127)
- <=WM: (14914: S1 ^operator O2128 +)
- <=WM: (14912: I3 ^dir L)
- <=WM: (14908: R1 ^reward R1067)
- <=WM: (14826: I3 ^see 0)
- <=WM: (14911: O2128 ^name predict-no)
- <=WM: (14910: O2127 ^name predict-yes)
- <=WM: (14909: R1067 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2129 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2130 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2128 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2127 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.389539 -> 0.711951 -0.322412 0.389539(R,m,v=1,0.897727,0.0923377)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*44 0.288049 0.322412 0.610461 -> 0.288049 0.322412 0.610461(R,m,v=1,1,0)
- =>WM: (14930: S1 ^operator O2130)
- 1065: O: O2130 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1065 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1064 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14931: I3 ^predict-no N1065)
- <=WM: (14917: N1064 ^status complete)
- <=WM: (14916: I3 ^predict-yes N1064)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- |\--- Input Phase ---
- =>WM: (14935: I2 ^dir U)
- =>WM: (14934: I2 ^reward 1)
- =>WM: (14933: I2 ^see 0)
- =>WM: (14932: N1065 ^status complete)
- <=WM: (14920: I2 ^dir U)
- <=WM: (14919: I2 ^reward 1)
- <=WM: (14918: I2 ^see 1)
- =>WM: (14936: I2 ^level-1 L1-root)
- <=WM: (14921: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1069 ^value 1 +)
- (R1 ^reward R1069 +)
- Firing propose*predict-yes
- -->
- (O2131 ^name predict-yes +)
- (S1 ^operator O2131 +)
- Firing propose*predict-no
- -->
- (O2132 ^name predict-no +)
- (S1 ^operator O2132 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2130 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2129 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2130 ^name predict-no +)
- (S1 ^operator O2130 +)
- Retracting propose*predict-yes
- -->
- (O2129 ^name predict-yes +)
- (S1 ^operator O2129 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1068 ^value 1 +)
- (R1 ^reward R1068 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2130 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2129 = 0.)
- =>WM: (14943: S1 ^operator O2132 +)
- =>WM: (14942: S1 ^operator O2131 +)
- =>WM: (14941: O2132 ^name predict-no)
- =>WM: (14940: O2131 ^name predict-yes)
- =>WM: (14939: R1069 ^value 1)
- =>WM: (14938: R1 ^reward R1069)
- =>WM: (14937: I3 ^see 0)
- <=WM: (14928: S1 ^operator O2129 +)
- <=WM: (14929: S1 ^operator O2130 +)
- <=WM: (14930: S1 ^operator O2130)
- <=WM: (14923: R1 ^reward R1068)
- <=WM: (14922: I3 ^see 1)
- <=WM: (14926: O2130 ^name predict-no)
- <=WM: (14925: O2129 ^name predict-yes)
- <=WM: (14924: R1068 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2131 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2132 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2130 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2129 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14944: S1 ^operator O2132)
- 1066: O: O2132 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1066 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1065 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14945: I3 ^predict-no N1066)
- <=WM: (14932: N1065 ^status complete)
- <=WM: (14931: I3 ^predict-no N1065)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- -/--- Input Phase ---
- =>WM: (14949: I2 ^dir U)
- =>WM: (14948: I2 ^reward 1)
- =>WM: (14947: I2 ^see 0)
- =>WM: (14946: N1066 ^status complete)
- <=WM: (14935: I2 ^dir U)
- <=WM: (14934: I2 ^reward 1)
- <=WM: (14933: I2 ^see 0)
- =>WM: (14950: I2 ^level-1 L1-root)
- <=WM: (14936: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1070 ^value 1 +)
- (R1 ^reward R1070 +)
- Firing propose*predict-yes
- -->
- (O2133 ^name predict-yes +)
- (S1 ^operator O2133 +)
- Firing propose*predict-no
- -->
- (O2134 ^name predict-no +)
- (S1 ^operator O2134 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2132 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2131 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2132 ^name predict-no +)
- (S1 ^operator O2132 +)
- Retracting propose*predict-yes
- -->
- (O2131 ^name predict-yes +)
- (S1 ^operator O2131 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1069 ^value 1 +)
- (R1 ^reward R1069 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2132 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2131 = 0.)
- =>WM: (14956: S1 ^operator O2134 +)
- =>WM: (14955: S1 ^operator O2133 +)
- =>WM: (14954: O2134 ^name predict-no)
- =>WM: (14953: O2133 ^name predict-yes)
- =>WM: (14952: R1070 ^value 1)
- =>WM: (14951: R1 ^reward R1070)
- <=WM: (14942: S1 ^operator O2131 +)
- <=WM: (14943: S1 ^operator O2132 +)
- <=WM: (14944: S1 ^operator O2132)
- <=WM: (14938: R1 ^reward R1069)
- <=WM: (14941: O2132 ^name predict-no)
- <=WM: (14940: O2131 ^name predict-yes)
- <=WM: (14939: R1069 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2133 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2134 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2132 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2131 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14957: S1 ^operator O2134)
- 1067: O: O2134 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1067 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1066 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14958: I3 ^predict-no N1067)
- <=WM: (14946: N1066 ^status complete)
- <=WM: (14945: I3 ^predict-no N1066)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (14962: I2 ^dir L)
- =>WM: (14961: I2 ^reward 1)
- =>WM: (14960: I2 ^see 0)
- =>WM: (14959: N1067 ^status complete)
- <=WM: (14949: I2 ^dir U)
- <=WM: (14948: I2 ^reward 1)
- <=WM: (14947: I2 ^see 0)
- =>WM: (14963: I2 ^level-1 L1-root)
- <=WM: (14950: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2134 = 0.6126635874172339)
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2133 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1071 ^value 1 +)
- (R1 ^reward R1071 +)
- Firing propose*predict-yes
- -->
- (O2135 ^name predict-yes +)
- (S1 ^operator O2135 +)
- Firing propose*predict-no
- -->
- (O2136 ^name predict-no +)
- (S1 ^operator O2136 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2134 = 0.3873355437317227)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2133 = 0.3895394592518329)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2134 ^name predict-no +)
- (S1 ^operator O2134 +)
- Retracting propose*predict-yes
- -->
- (O2133 ^name predict-yes +)
- (S1 ^operator O2133 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1070 ^value 1 +)
- (R1 ^reward R1070 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2134 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2133 = 0.)
- =>WM: (14970: S1 ^operator O2136 +)
- =>WM: (14969: S1 ^operator O2135 +)
- =>WM: (14968: I3 ^dir L)
- =>WM: (14967: O2136 ^name predict-no)
- =>WM: (14966: O2135 ^name predict-yes)
- =>WM: (14965: R1071 ^value 1)
- =>WM: (14964: R1 ^reward R1071)
- <=WM: (14955: S1 ^operator O2133 +)
- <=WM: (14956: S1 ^operator O2134 +)
- <=WM: (14957: S1 ^operator O2134)
- <=WM: (14927: I3 ^dir U)
- <=WM: (14951: R1 ^reward R1070)
- <=WM: (14954: O2134 ^name predict-no)
- <=WM: (14953: O2133 ^name predict-yes)
- <=WM: (14952: R1070 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2135 = -0.02274740735326741)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2135 = 0.3895394592518329)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2136 = 0.6126635874172339)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2136 = 0.3873355437317227)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2134 = 0.3873355437317227)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2134 = 0.6126635874172339)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2133 = 0.3895394592518329)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2133 = -0.02274740735326741)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (14971: S1 ^operator O2136)
- 1068: O: O2136 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1068 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1067 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14972: I3 ^predict-no N1068)
- <=WM: (14959: N1067 ^status complete)
- <=WM: (14958: I3 ^predict-no N1067)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- /|\---- Input Phase ---
- =>WM: (14976: I2 ^dir L)
- =>WM: (14975: I2 ^reward 1)
- =>WM: (14974: I2 ^see 0)
- =>WM: (14973: N1068 ^status complete)
- <=WM: (14962: I2 ^dir L)
- <=WM: (14961: I2 ^reward 1)
- <=WM: (14960: I2 ^see 0)
- =>WM: (14977: I2 ^level-1 L0-root)
- <=WM: (14963: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2135 = 0.1599599085218832)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2136 = 0.612665734378294)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1072 ^value 1 +)
- (R1 ^reward R1072 +)
- Firing propose*predict-yes
- -->
- (O2137 ^name predict-yes +)
- (S1 ^operator O2137 +)
- Firing propose*predict-no
- -->
- (O2138 ^name predict-no +)
- (S1 ^operator O2138 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2136 = 0.3873355437317227)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2135 = 0.3895394592518329)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2136 ^name predict-no +)
- (S1 ^operator O2136 +)
- Retracting propose*predict-yes
- -->
- (O2135 ^name predict-yes +)
- (S1 ^operator O2135 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1071 ^value 1 +)
- (R1 ^reward R1071 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2136 = 0.3873355437317227)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2136 = 0.6126635874172339)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2135 = 0.3895394592518329)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2135 = -0.02274740735326741)
- =>WM: (14983: S1 ^operator O2138 +)
- =>WM: (14982: S1 ^operator O2137 +)
- =>WM: (14981: O2138 ^name predict-no)
- =>WM: (14980: O2137 ^name predict-yes)
- =>WM: (14979: R1072 ^value 1)
- =>WM: (14978: R1 ^reward R1072)
- <=WM: (14969: S1 ^operator O2135 +)
- <=WM: (14970: S1 ^operator O2136 +)
- <=WM: (14971: S1 ^operator O2136)
- <=WM: (14964: R1 ^reward R1071)
- <=WM: (14967: O2136 ^name predict-no)
- <=WM: (14966: O2135 ^name predict-yes)
- <=WM: (14965: R1071 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2137 = 0.3895394592518329)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2137 = 0.1599599085218832)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2138 = 0.3873355437317227)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2138 = 0.612665734378294)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2136 = 0.3873355437317227)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2136 = 0.612665734378294)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2135 = 0.3895394592518329)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2135 = 0.1599599085218832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.719079 -0.331743 0.387336 -> 0.719079 -0.331743 0.387336(R,m,v=1,0.93617,0.0600751)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*32 0.28092 0.331744 0.612664 -> 0.28092 0.331744 0.612664(R,m,v=1,1,0)
- =>WM: (14984: S1 ^operator O2138)
- 1069: O: O2138 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1069 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1068 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14985: I3 ^predict-no N1069)
- <=WM: (14973: N1068 ^status complete)
- <=WM: (14972: I3 ^predict-no N1068)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (14989: I2 ^dir R)
- =>WM: (14988: I2 ^reward 1)
- =>WM: (14987: I2 ^see 0)
- =>WM: (14986: N1069 ^status complete)
- <=WM: (14976: I2 ^dir L)
- <=WM: (14975: I2 ^reward 1)
- <=WM: (14974: I2 ^see 0)
- =>WM: (14990: I2 ^level-1 L0-root)
- <=WM: (14977: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2137 = 0.8155904055662546)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2138 = -0.00558448899823713)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1073 ^value 1 +)
- (R1 ^reward R1073 +)
- Firing propose*predict-yes
- -->
- (O2139 ^name predict-yes +)
- (S1 ^operator O2139 +)
- Firing propose*predict-no
- -->
- (O2140 ^name predict-no +)
- (S1 ^operator O2140 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2138 = 0.4476193732504098)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2137 = 0.1844106571836858)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2138 ^name predict-no +)
- (S1 ^operator O2138 +)
- Retracting propose*predict-yes
- -->
- (O2137 ^name predict-yes +)
- (S1 ^operator O2137 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1072 ^value 1 +)
- (R1 ^reward R1072 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2138 = 0.612665734378294)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2138 = 0.3873356740593792)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2137 = 0.1599599085218832)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2137 = 0.3895394592518329)
- =>WM: (14997: S1 ^operator O2140 +)
- =>WM: (14996: S1 ^operator O2139 +)
- =>WM: (14995: I3 ^dir R)
- =>WM: (14994: O2140 ^name predict-no)
- =>WM: (14993: O2139 ^name predict-yes)
- =>WM: (14992: R1073 ^value 1)
- =>WM: (14991: R1 ^reward R1073)
- <=WM: (14982: S1 ^operator O2137 +)
- <=WM: (14983: S1 ^operator O2138 +)
- <=WM: (14984: S1 ^operator O2138)
- <=WM: (14968: I3 ^dir L)
- <=WM: (14978: R1 ^reward R1072)
- <=WM: (14981: O2138 ^name predict-no)
- <=WM: (14980: O2137 ^name predict-yes)
- <=WM: (14979: R1072 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2139 = 0.8155904055662546)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2139 = 0.1844106571836858)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2140 = -0.00558448899823713)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2140 = 0.4476193732504098)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2138 = 0.4476193732504098)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2138 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2137 = 0.1844106571836858)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2137 = 0.8155904055662546)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.719079 -0.331743 0.387336 -> 0.719079 -0.331743 0.387335(R,m,v=1,0.936508,0.0597771)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*37 0.280923 0.331743 0.612666 -> 0.280923 0.331743 0.612666(R,m,v=1,1,0)
- =>WM: (14998: S1 ^operator O2139)
- 1070: O: O2139 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1070 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1069 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (14999: I3 ^predict-yes N1070)
- <=WM: (14986: N1069 ^status complete)
- <=WM: (14985: I3 ^predict-no N1069)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (15003: I2 ^dir L)
- =>WM: (15002: I2 ^reward 1)
- =>WM: (15001: I2 ^see 1)
- =>WM: (15000: N1070 ^status complete)
- <=WM: (14989: I2 ^dir R)
- <=WM: (14988: I2 ^reward 1)
- <=WM: (14987: I2 ^see 0)
- =>WM: (15004: I2 ^level-1 R1-root)
- <=WM: (14990: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2139 = 0.6104598832926351)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2140 = 0.2714993082286609)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1074 ^value 1 +)
- (R1 ^reward R1074 +)
- Firing propose*predict-yes
- -->
- (O2141 ^name predict-yes +)
- (S1 ^operator O2141 +)
- Firing propose*predict-no
- -->
- (O2142 ^name predict-no +)
- (S1 ^operator O2142 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2140 = 0.3873354627937282)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2139 = 0.3895394592518329)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2140 ^name predict-no +)
- (S1 ^operator O2140 +)
- Retracting propose*predict-yes
- -->
- (O2139 ^name predict-yes +)
- (S1 ^operator O2139 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1073 ^value 1 +)
- (R1 ^reward R1073 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2140 = 0.4476193732504098)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2140 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2139 = 0.1844106571836858)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2139 = 0.8155904055662546)
- =>WM: (15012: S1 ^operator O2142 +)
- =>WM: (15011: S1 ^operator O2141 +)
- =>WM: (15010: I3 ^dir L)
- =>WM: (15009: O2142 ^name predict-no)
- =>WM: (15008: O2141 ^name predict-yes)
- =>WM: (15007: R1074 ^value 1)
- =>WM: (15006: R1 ^reward R1074)
- =>WM: (15005: I3 ^see 1)
- <=WM: (14996: S1 ^operator O2139 +)
- <=WM: (14998: S1 ^operator O2139)
- <=WM: (14997: S1 ^operator O2140 +)
- <=WM: (14995: I3 ^dir R)
- <=WM: (14991: R1 ^reward R1073)
- <=WM: (14937: I3 ^see 0)
- <=WM: (14994: O2140 ^name predict-no)
- <=WM: (14993: O2139 ^name predict-yes)
- <=WM: (14992: R1073 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2141 = 0.3895394592518329)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2141 = 0.6104598832926351)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2142 = 0.3873354627937282)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2142 = 0.2714993082286609)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2140 = 0.3873354627937282)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2140 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2139 = 0.3895394592518329)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2139 = 0.6104598832926351)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675413 -0.491002 0.184411 -> 0.675413 -0.491002 0.18441(R,m,v=1,0.905028,0.0864353)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*34 0.324588 0.491002 0.81559 -> 0.324588 0.491002 0.81559(R,m,v=1,1,0)
- =>WM: (15013: S1 ^operator O2141)
- 1071: O: O2141 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1071 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1070 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15014: I3 ^predict-yes N1071)
- <=WM: (15000: N1070 ^status complete)
- <=WM: (14999: I3 ^predict-yes N1070)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- \--- Input Phase ---
- =>WM: (15018: I2 ^dir R)
- =>WM: (15017: I2 ^reward 1)
- =>WM: (15016: I2 ^see 1)
- =>WM: (15015: N1071 ^status complete)
- <=WM: (15003: I2 ^dir L)
- <=WM: (15002: I2 ^reward 1)
- <=WM: (15001: I2 ^see 1)
- =>WM: (15019: I2 ^level-1 L1-root)
- <=WM: (15004: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2142 = -0.02155734064455064)
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2141 = 0.8155841587866866)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1075 ^value 1 +)
- (R1 ^reward R1075 +)
- Firing propose*predict-yes
- -->
- (O2143 ^name predict-yes +)
- (S1 ^operator O2143 +)
- Firing propose*predict-no
- -->
- (O2144 ^name predict-no +)
- (S1 ^operator O2144 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2142 = 0.4476193732504098)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2141 = 0.1844104977711947)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2142 ^name predict-no +)
- (S1 ^operator O2142 +)
- Retracting propose*predict-yes
- -->
- (O2141 ^name predict-yes +)
- (S1 ^operator O2141 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1074 ^value 1 +)
- (R1 ^reward R1074 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2142 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2142 = 0.3873354627937282)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2141 = 0.6104598832926351)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2141 = 0.3895394592518329)
- =>WM: (15026: S1 ^operator O2144 +)
- =>WM: (15025: S1 ^operator O2143 +)
- =>WM: (15024: I3 ^dir R)
- =>WM: (15023: O2144 ^name predict-no)
- =>WM: (15022: O2143 ^name predict-yes)
- =>WM: (15021: R1075 ^value 1)
- =>WM: (15020: R1 ^reward R1075)
- <=WM: (15011: S1 ^operator O2141 +)
- <=WM: (15013: S1 ^operator O2141)
- <=WM: (15012: S1 ^operator O2142 +)
- <=WM: (15010: I3 ^dir L)
- <=WM: (15006: R1 ^reward R1074)
- <=WM: (15009: O2142 ^name predict-no)
- <=WM: (15008: O2141 ^name predict-yes)
- <=WM: (15007: R1074 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2143 = 0.1844104977711947)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2143 = 0.8155841587866866)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2144 = 0.4476193732504098)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2144 = -0.02155734064455064)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2142 = 0.4476193732504098)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2142 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2141 = 0.1844104977711947)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2141 = 0.8155841587866866)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.389539 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.898305,0.0918721)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*36 0.288049 0.322411 0.61046 -> 0.288049 0.322411 0.61046(R,m,v=1,1,0)
- =>WM: (15027: S1 ^operator O2143)
- 1072: O: O2143 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1072 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1071 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15028: I3 ^predict-yes N1072)
- <=WM: (15015: N1071 ^status complete)
- <=WM: (15014: I3 ^predict-yes N1071)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (15032: I2 ^dir L)
- =>WM: (15031: I2 ^reward 1)
- =>WM: (15030: I2 ^see 1)
- =>WM: (15029: N1072 ^status complete)
- <=WM: (15018: I2 ^dir R)
- <=WM: (15017: I2 ^reward 1)
- <=WM: (15016: I2 ^see 1)
- =>WM: (15033: I2 ^level-1 R1-root)
- <=WM: (15019: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2143 = 0.6104599819109648)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2144 = 0.2714993082286609)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1076 ^value 1 +)
- (R1 ^reward R1076 +)
- Firing propose*predict-yes
- -->
- (O2145 ^name predict-yes +)
- (S1 ^operator O2145 +)
- Firing propose*predict-no
- -->
- (O2146 ^name predict-no +)
- (S1 ^operator O2146 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2144 = 0.3873354627937282)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2143 = 0.3895395578701628)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2144 ^name predict-no +)
- (S1 ^operator O2144 +)
- Retracting propose*predict-yes
- -->
- (O2143 ^name predict-yes +)
- (S1 ^operator O2143 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1075 ^value 1 +)
- (R1 ^reward R1075 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2144 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2144 = 0.4476193732504098)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2143 = 0.8155841587866866)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2143 = 0.1844104977711947)
- =>WM: (15040: S1 ^operator O2146 +)
- =>WM: (15039: S1 ^operator O2145 +)
- =>WM: (15038: I3 ^dir L)
- =>WM: (15037: O2146 ^name predict-no)
- =>WM: (15036: O2145 ^name predict-yes)
- =>WM: (15035: R1076 ^value 1)
- =>WM: (15034: R1 ^reward R1076)
- <=WM: (15025: S1 ^operator O2143 +)
- <=WM: (15027: S1 ^operator O2143)
- <=WM: (15026: S1 ^operator O2144 +)
- <=WM: (15024: I3 ^dir R)
- <=WM: (15020: R1 ^reward R1075)
- <=WM: (15023: O2144 ^name predict-no)
- <=WM: (15022: O2143 ^name predict-yes)
- <=WM: (15021: R1075 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2145 = 0.3895395578701628)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2145 = 0.6104599819109648)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2146 = 0.3873354627937282)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2146 = 0.2714993082286609)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2144 = 0.3873354627937282)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2144 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2143 = 0.3895395578701628)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2143 = 0.6104599819109648)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675413 -0.491002 0.18441 -> 0.675414 -0.491003 0.184411(R,m,v=1,0.905556,0.0860025)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*46 0.32458 0.491004 0.815584 -> 0.324581 0.491004 0.815585(R,m,v=1,1,0)
- =>WM: (15041: S1 ^operator O2145)
- 1073: O: O2145 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1073 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1072 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15042: I3 ^predict-yes N1073)
- <=WM: (15029: N1072 ^status complete)
- <=WM: (15028: I3 ^predict-yes N1072)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (15046: I2 ^dir R)
- =>WM: (15045: I2 ^reward 1)
- =>WM: (15044: I2 ^see 1)
- =>WM: (15043: N1073 ^status complete)
- <=WM: (15032: I2 ^dir L)
- <=WM: (15031: I2 ^reward 1)
- <=WM: (15030: I2 ^see 1)
- =>WM: (15047: I2 ^level-1 L1-root)
- <=WM: (15033: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2146 = -0.02155734064455064)
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2145 = 0.8155849603030043)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1077 ^value 1 +)
- (R1 ^reward R1077 +)
- Firing propose*predict-yes
- -->
- (O2147 ^name predict-yes +)
- (S1 ^operator O2147 +)
- Firing propose*predict-no
- -->
- (O2148 ^name predict-no +)
- (S1 ^operator O2148 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2146 = 0.4476193732504098)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2145 = 0.1844112992875125)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2146 ^name predict-no +)
- (S1 ^operator O2146 +)
- Retracting propose*predict-yes
- -->
- (O2145 ^name predict-yes +)
- (S1 ^operator O2145 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1076 ^value 1 +)
- (R1 ^reward R1076 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2146 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2146 = 0.3873354627937282)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2145 = 0.6104599819109648)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2145 = 0.3895395578701628)
- =>WM: (15054: S1 ^operator O2148 +)
- =>WM: (15053: S1 ^operator O2147 +)
- =>WM: (15052: I3 ^dir R)
- =>WM: (15051: O2148 ^name predict-no)
- =>WM: (15050: O2147 ^name predict-yes)
- =>WM: (15049: R1077 ^value 1)
- =>WM: (15048: R1 ^reward R1077)
- <=WM: (15039: S1 ^operator O2145 +)
- <=WM: (15041: S1 ^operator O2145)
- <=WM: (15040: S1 ^operator O2146 +)
- <=WM: (15038: I3 ^dir L)
- <=WM: (15034: R1 ^reward R1076)
- <=WM: (15037: O2146 ^name predict-no)
- <=WM: (15036: O2145 ^name predict-yes)
- <=WM: (15035: R1076 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2147 = 0.1844112992875125)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2147 = 0.8155849603030043)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2148 = 0.4476193732504098)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2148 = -0.02155734064455064)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2146 = 0.4476193732504098)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2146 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2145 = 0.1844112992875125)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2145 = 0.8155849603030043)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.898876,0.0914112)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*36 0.288049 0.322411 0.61046 -> 0.288049 0.322411 0.61046(R,m,v=1,1,0)
- =>WM: (15055: S1 ^operator O2147)
- 1074: O: O2147 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1074 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1073 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15056: I3 ^predict-yes N1074)
- <=WM: (15043: N1073 ^status complete)
- <=WM: (15042: I3 ^predict-yes N1073)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (15060: I2 ^dir R)
- =>WM: (15059: I2 ^reward 1)
- =>WM: (15058: I2 ^see 1)
- =>WM: (15057: N1074 ^status complete)
- <=WM: (15046: I2 ^dir R)
- <=WM: (15045: I2 ^reward 1)
- <=WM: (15044: I2 ^see 1)
- =>WM: (15061: I2 ^level-1 R1-root)
- <=WM: (15047: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2147 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2148 = 0.5523810978783164)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1078 ^value 1 +)
- (R1 ^reward R1078 +)
- Firing propose*predict-yes
- -->
- (O2149 ^name predict-yes +)
- (S1 ^operator O2149 +)
- Firing propose*predict-no
- -->
- (O2150 ^name predict-no +)
- (S1 ^operator O2150 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2148 = 0.4476193732504098)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2147 = 0.1844112992875125)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2148 ^name predict-no +)
- (S1 ^operator O2148 +)
- Retracting propose*predict-yes
- -->
- (O2147 ^name predict-yes +)
- (S1 ^operator O2147 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1077 ^value 1 +)
- (R1 ^reward R1077 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2148 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2148 = 0.4476193732504098)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2147 = 0.8155849603030043)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2147 = 0.1844112992875125)
- =>WM: (15067: S1 ^operator O2150 +)
- =>WM: (15066: S1 ^operator O2149 +)
- =>WM: (15065: O2150 ^name predict-no)
- =>WM: (15064: O2149 ^name predict-yes)
- =>WM: (15063: R1078 ^value 1)
- =>WM: (15062: R1 ^reward R1078)
- <=WM: (15053: S1 ^operator O2147 +)
- <=WM: (15055: S1 ^operator O2147)
- <=WM: (15054: S1 ^operator O2148 +)
- <=WM: (15048: R1 ^reward R1077)
- <=WM: (15051: O2148 ^name predict-no)
- <=WM: (15050: O2147 ^name predict-yes)
- <=WM: (15049: R1077 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2149 = 0.1844112992875125)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2149 = 0.1398795999120246)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2150 = 0.4476193732504098)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2150 = 0.5523810978783164)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2148 = 0.4476193732504098)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2148 = 0.5523810978783164)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2147 = 0.1844112992875125)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2147 = 0.1398795999120246)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675414 -0.491003 0.184411 -> 0.675415 -0.491003 0.184412(R,m,v=1,0.906077,0.085574)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*46 0.324581 0.491004 0.815585 -> 0.324582 0.491004 0.815586(R,m,v=1,1,0)
- =>WM: (15068: S1 ^operator O2150)
- 1075: O: O2150 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1075 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1074 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15069: I3 ^predict-no N1075)
- <=WM: (15057: N1074 ^status complete)
- <=WM: (15056: I3 ^predict-yes N1074)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (15073: I2 ^dir L)
- =>WM: (15072: I2 ^reward 1)
- =>WM: (15071: I2 ^see 0)
- =>WM: (15070: N1075 ^status complete)
- <=WM: (15060: I2 ^dir R)
- <=WM: (15059: I2 ^reward 1)
- <=WM: (15058: I2 ^see 1)
- =>WM: (15074: I2 ^level-1 R0-root)
- <=WM: (15061: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2149 = 0.6104606905185325)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2150 = 0.1063475139796038)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1079 ^value 1 +)
- (R1 ^reward R1079 +)
- Firing propose*predict-yes
- -->
- (O2151 ^name predict-yes +)
- (S1 ^operator O2151 +)
- Firing propose*predict-no
- -->
- (O2152 ^name predict-no +)
- (S1 ^operator O2152 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2150 = 0.3873354627937282)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2149 = 0.3895396269029936)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2150 ^name predict-no +)
- (S1 ^operator O2150 +)
- Retracting propose*predict-yes
- -->
- (O2149 ^name predict-yes +)
- (S1 ^operator O2149 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1078 ^value 1 +)
- (R1 ^reward R1078 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2150 = 0.5523810978783164)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2150 = 0.4476193732504098)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2149 = 0.1398795999120246)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2149 = 0.1844118603489351)
- =>WM: (15082: S1 ^operator O2152 +)
- =>WM: (15081: S1 ^operator O2151 +)
- =>WM: (15080: I3 ^dir L)
- =>WM: (15079: O2152 ^name predict-no)
- =>WM: (15078: O2151 ^name predict-yes)
- =>WM: (15077: R1079 ^value 1)
- =>WM: (15076: R1 ^reward R1079)
- =>WM: (15075: I3 ^see 0)
- <=WM: (15066: S1 ^operator O2149 +)
- <=WM: (15067: S1 ^operator O2150 +)
- <=WM: (15068: S1 ^operator O2150)
- <=WM: (15052: I3 ^dir R)
- <=WM: (15062: R1 ^reward R1078)
- <=WM: (15005: I3 ^see 1)
- <=WM: (15065: O2150 ^name predict-no)
- <=WM: (15064: O2149 ^name predict-yes)
- <=WM: (15063: R1078 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2151 = 0.3895396269029936)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2151 = 0.6104606905185325)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2152 = 0.3873354627937282)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2152 = 0.1063475139796038)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2150 = 0.3873354627937282)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2150 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2149 = 0.3895396269029936)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2149 = 0.6104606905185325)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.93617,0.0601824)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*39 0.377467 0.174914 0.552381 -> 0.377467 0.174914 0.552381(R,m,v=1,1,0)
- =>WM: (15083: S1 ^operator O2151)
- 1076: O: O2151 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1076 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1075 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15084: I3 ^predict-yes N1076)
- <=WM: (15070: N1075 ^status complete)
- <=WM: (15069: I3 ^predict-no N1075)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (15088: I2 ^dir L)
- =>WM: (15087: I2 ^reward 1)
- =>WM: (15086: I2 ^see 1)
- =>WM: (15085: N1076 ^status complete)
- <=WM: (15073: I2 ^dir L)
- <=WM: (15072: I2 ^reward 1)
- <=WM: (15071: I2 ^see 0)
- =>WM: (15089: I2 ^level-1 L1-root)
- <=WM: (15074: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2152 = 0.6126637177448905)
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2151 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1080 ^value 1 +)
- (R1 ^reward R1080 +)
- Firing propose*predict-yes
- -->
- (O2153 ^name predict-yes +)
- (S1 ^operator O2153 +)
- Firing propose*predict-no
- -->
- (O2154 ^name predict-no +)
- (S1 ^operator O2154 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2152 = 0.3873354627937282)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2151 = 0.3895396269029936)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2152 ^name predict-no +)
- (S1 ^operator O2152 +)
- Retracting propose*predict-yes
- -->
- (O2151 ^name predict-yes +)
- (S1 ^operator O2151 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1079 ^value 1 +)
- (R1 ^reward R1079 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2152 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2152 = 0.3873354627937282)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2151 = 0.6104606905185325)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2151 = 0.3895396269029936)
- =>WM: (15096: S1 ^operator O2154 +)
- =>WM: (15095: S1 ^operator O2153 +)
- =>WM: (15094: O2154 ^name predict-no)
- =>WM: (15093: O2153 ^name predict-yes)
- =>WM: (15092: R1080 ^value 1)
- =>WM: (15091: R1 ^reward R1080)
- =>WM: (15090: I3 ^see 1)
- <=WM: (15081: S1 ^operator O2151 +)
- <=WM: (15083: S1 ^operator O2151)
- <=WM: (15082: S1 ^operator O2152 +)
- <=WM: (15076: R1 ^reward R1079)
- <=WM: (15075: I3 ^see 0)
- <=WM: (15079: O2152 ^name predict-no)
- <=WM: (15078: O2151 ^name predict-yes)
- <=WM: (15077: R1079 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2153 = 0.3895396269029936)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2153 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2154 = 0.3873354627937282)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2154 = 0.6126637177448905)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2152 = 0.3873354627937282)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2152 = 0.6126637177448905)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2151 = 0.3895396269029936)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2151 = -0.02274740735326741)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.899441,0.0909547)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*44 0.288049 0.322412 0.610461 -> 0.288049 0.322412 0.610461(R,m,v=1,1,0)
- =>WM: (15097: S1 ^operator O2154)
- 1077: O: O2154 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1077 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1076 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15098: I3 ^predict-no N1077)
- <=WM: (15085: N1076 ^status complete)
- <=WM: (15084: I3 ^predict-yes N1076)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- \---- Input Phase ---
- =>WM: (15102: I2 ^dir R)
- =>WM: (15101: I2 ^reward 1)
- =>WM: (15100: I2 ^see 0)
- =>WM: (15099: N1077 ^status complete)
- <=WM: (15088: I2 ^dir L)
- <=WM: (15087: I2 ^reward 1)
- <=WM: (15086: I2 ^see 1)
- =>WM: (15103: I2 ^level-1 L0-root)
- <=WM: (15089: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2153 = 0.8155902461537636)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2154 = -0.00558448899823713)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1081 ^value 1 +)
- (R1 ^reward R1081 +)
- Firing propose*predict-yes
- -->
- (O2155 ^name predict-yes +)
- (S1 ^operator O2155 +)
- Firing propose*predict-no
- -->
- (O2156 ^name predict-no +)
- (S1 ^operator O2156 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2154 = 0.4476193025811009)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2153 = 0.1844118603489351)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2154 ^name predict-no +)
- (S1 ^operator O2154 +)
- Retracting propose*predict-yes
- -->
- (O2153 ^name predict-yes +)
- (S1 ^operator O2153 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1080 ^value 1 +)
- (R1 ^reward R1080 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2154 = 0.6126637177448905)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2154 = 0.3873354627937282)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2153 = -0.02274740735326741)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2153 = 0.3895395792897647)
- =>WM: (15111: S1 ^operator O2156 +)
- =>WM: (15110: S1 ^operator O2155 +)
- =>WM: (15109: I3 ^dir R)
- =>WM: (15108: O2156 ^name predict-no)
- =>WM: (15107: O2155 ^name predict-yes)
- =>WM: (15106: R1081 ^value 1)
- =>WM: (15105: R1 ^reward R1081)
- =>WM: (15104: I3 ^see 0)
- <=WM: (15095: S1 ^operator O2153 +)
- <=WM: (15096: S1 ^operator O2154 +)
- <=WM: (15097: S1 ^operator O2154)
- <=WM: (15080: I3 ^dir L)
- <=WM: (15091: R1 ^reward R1080)
- <=WM: (15090: I3 ^see 1)
- <=WM: (15094: O2154 ^name predict-no)
- <=WM: (15093: O2153 ^name predict-yes)
- <=WM: (15092: R1080 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2155 = 0.1844118603489351)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2155 = 0.8155902461537636)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2156 = 0.4476193025811009)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2156 = -0.00558448899823713)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2154 = 0.4476193025811009)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2154 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2153 = 0.1844118603489351)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2153 = 0.8155902461537636)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.719079 -0.331743 0.387335 -> 0.719079 -0.331743 0.387336(R,m,v=1,0.936842,0.059482)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*32 0.28092 0.331744 0.612664 -> 0.28092 0.331744 0.612664(R,m,v=1,1,0)
- =>WM: (15112: S1 ^operator O2155)
- 1078: O: O2155 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1078 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1077 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15113: I3 ^predict-yes N1078)
- <=WM: (15099: N1077 ^status complete)
- <=WM: (15098: I3 ^predict-no N1077)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- /|--- Input Phase ---
- =>WM: (15117: I2 ^dir L)
- =>WM: (15116: I2 ^reward 1)
- =>WM: (15115: I2 ^see 1)
- =>WM: (15114: N1078 ^status complete)
- <=WM: (15102: I2 ^dir R)
- <=WM: (15101: I2 ^reward 1)
- <=WM: (15100: I2 ^see 0)
- =>WM: (15118: I2 ^level-1 R1-root)
- <=WM: (15103: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2155 = 0.6104600509437957)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2156 = 0.2714993082286609)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1082 ^value 1 +)
- (R1 ^reward R1082 +)
- Firing propose*predict-yes
- -->
- (O2157 ^name predict-yes +)
- (S1 ^operator O2157 +)
- Firing propose*predict-no
- -->
- (O2158 ^name predict-no +)
- (S1 ^operator O2158 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2156 = 0.3873355857129354)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2155 = 0.3895395792897647)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2156 ^name predict-no +)
- (S1 ^operator O2156 +)
- Retracting propose*predict-yes
- -->
- (O2155 ^name predict-yes +)
- (S1 ^operator O2155 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1081 ^value 1 +)
- (R1 ^reward R1081 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2156 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2156 = 0.4476193025811009)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2155 = 0.8155902461537636)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2155 = 0.1844118603489351)
- =>WM: (15126: S1 ^operator O2158 +)
- =>WM: (15125: S1 ^operator O2157 +)
- =>WM: (15124: I3 ^dir L)
- =>WM: (15123: O2158 ^name predict-no)
- =>WM: (15122: O2157 ^name predict-yes)
- =>WM: (15121: R1082 ^value 1)
- =>WM: (15120: R1 ^reward R1082)
- =>WM: (15119: I3 ^see 1)
- <=WM: (15110: S1 ^operator O2155 +)
- <=WM: (15112: S1 ^operator O2155)
- <=WM: (15111: S1 ^operator O2156 +)
- <=WM: (15109: I3 ^dir R)
- <=WM: (15105: R1 ^reward R1081)
- <=WM: (15104: I3 ^see 0)
- <=WM: (15108: O2156 ^name predict-no)
- <=WM: (15107: O2155 ^name predict-yes)
- <=WM: (15106: R1081 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2157 = 0.3895395792897647)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2157 = 0.6104600509437957)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2158 = 0.3873355857129354)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2158 = 0.2714993082286609)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2156 = 0.3873355857129354)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2156 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2155 = 0.3895395792897647)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2155 = 0.6104600509437957)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675415 -0.491003 0.184412 -> 0.675414 -0.491003 0.184412(R,m,v=1,0.906593,0.0851497)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*34 0.324588 0.491002 0.81559 -> 0.324588 0.491002 0.81559(R,m,v=1,1,0)
- =>WM: (15127: S1 ^operator O2157)
- 1079: O: O2157 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1079 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1078 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15128: I3 ^predict-yes N1079)
- <=WM: (15114: N1078 ^status complete)
- <=WM: (15113: I3 ^predict-yes N1078)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (15132: I2 ^dir U)
- =>WM: (15131: I2 ^reward 1)
- =>WM: (15130: I2 ^see 1)
- =>WM: (15129: N1079 ^status complete)
- <=WM: (15117: I2 ^dir L)
- <=WM: (15116: I2 ^reward 1)
- <=WM: (15115: I2 ^see 1)
- =>WM: (15133: I2 ^level-1 L1-root)
- <=WM: (15118: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1083 ^value 1 +)
- (R1 ^reward R1083 +)
- Firing propose*predict-yes
- -->
- (O2159 ^name predict-yes +)
- (S1 ^operator O2159 +)
- Firing propose*predict-no
- -->
- (O2160 ^name predict-no +)
- (S1 ^operator O2160 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2158 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2157 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2158 ^name predict-no +)
- (S1 ^operator O2158 +)
- Retracting propose*predict-yes
- -->
- (O2157 ^name predict-yes +)
- (S1 ^operator O2157 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1082 ^value 1 +)
- (R1 ^reward R1082 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2158 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2158 = 0.3873355857129354)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2157 = 0.6104600509437957)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2157 = 0.3895395792897647)
- =>WM: (15140: S1 ^operator O2160 +)
- =>WM: (15139: S1 ^operator O2159 +)
- =>WM: (15138: I3 ^dir U)
- =>WM: (15137: O2160 ^name predict-no)
- =>WM: (15136: O2159 ^name predict-yes)
- =>WM: (15135: R1083 ^value 1)
- =>WM: (15134: R1 ^reward R1083)
- <=WM: (15125: S1 ^operator O2157 +)
- <=WM: (15127: S1 ^operator O2157)
- <=WM: (15126: S1 ^operator O2158 +)
- <=WM: (15124: I3 ^dir L)
- <=WM: (15120: R1 ^reward R1082)
- <=WM: (15123: O2158 ^name predict-no)
- <=WM: (15122: O2157 ^name predict-yes)
- <=WM: (15121: R1082 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2159 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2160 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2158 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2157 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.9,0.0905028)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*36 0.288049 0.322411 0.61046 -> 0.288049 0.322411 0.61046(R,m,v=1,1,0)
- =>WM: (15141: S1 ^operator O2160)
- 1080: O: O2160 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1080 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1079 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15142: I3 ^predict-no N1080)
- <=WM: (15129: N1079 ^status complete)
- <=WM: (15128: I3 ^predict-yes N1079)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (15146: I2 ^dir L)
- =>WM: (15145: I2 ^reward 1)
- =>WM: (15144: I2 ^see 0)
- =>WM: (15143: N1080 ^status complete)
- <=WM: (15132: I2 ^dir U)
- <=WM: (15131: I2 ^reward 1)
- <=WM: (15130: I2 ^see 1)
- =>WM: (15147: I2 ^level-1 L1-root)
- <=WM: (15133: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2160 = 0.6126638406640976)
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2159 = -0.02274740735326741)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1084 ^value 1 +)
- (R1 ^reward R1084 +)
- Firing propose*predict-yes
- -->
- (O2161 ^name predict-yes +)
- (S1 ^operator O2161 +)
- Firing propose*predict-no
- -->
- (O2162 ^name predict-no +)
- (S1 ^operator O2162 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2160 = 0.3873355857129354)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2159 = 0.3895396347547306)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2160 ^name predict-no +)
- (S1 ^operator O2160 +)
- Retracting propose*predict-yes
- -->
- (O2159 ^name predict-yes +)
- (S1 ^operator O2159 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1083 ^value 1 +)
- (R1 ^reward R1083 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2160 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2159 = 0.)
- =>WM: (15155: S1 ^operator O2162 +)
- =>WM: (15154: S1 ^operator O2161 +)
- =>WM: (15153: I3 ^dir L)
- =>WM: (15152: O2162 ^name predict-no)
- =>WM: (15151: O2161 ^name predict-yes)
- =>WM: (15150: R1084 ^value 1)
- =>WM: (15149: R1 ^reward R1084)
- =>WM: (15148: I3 ^see 0)
- <=WM: (15139: S1 ^operator O2159 +)
- <=WM: (15140: S1 ^operator O2160 +)
- <=WM: (15141: S1 ^operator O2160)
- <=WM: (15138: I3 ^dir U)
- <=WM: (15134: R1 ^reward R1083)
- <=WM: (15119: I3 ^see 1)
- <=WM: (15137: O2160 ^name predict-no)
- <=WM: (15136: O2159 ^name predict-yes)
- <=WM: (15135: R1083 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2161 = -0.02274740735326741)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2161 = 0.3895396347547306)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2162 = 0.6126638406640976)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2162 = 0.3873355857129354)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2160 = 0.3873355857129354)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2160 = 0.6126638406640976)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2159 = 0.3895396347547306)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2159 = -0.02274740735326741)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (15156: S1 ^operator O2162)
- 1081: O: O2162 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1081 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1080 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15157: I3 ^predict-no N1081)
- <=WM: (15143: N1080 ^status complete)
- <=WM: (15142: I3 ^predict-no N1080)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- /--- Input Phase ---
- =>WM: (15161: I2 ^dir U)
- =>WM: (15160: I2 ^reward 1)
- =>WM: (15159: I2 ^see 0)
- =>WM: (15158: N1081 ^status complete)
- <=WM: (15146: I2 ^dir L)
- <=WM: (15145: I2 ^reward 1)
- <=WM: (15144: I2 ^see 0)
- =>WM: (15162: I2 ^level-1 L0-root)
- <=WM: (15147: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1085 ^value 1 +)
- (R1 ^reward R1085 +)
- Firing propose*predict-yes
- -->
- (O2163 ^name predict-yes +)
- (S1 ^operator O2163 +)
- Firing propose*predict-no
- -->
- (O2164 ^name predict-no +)
- (S1 ^operator O2164 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2162 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2161 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2162 ^name predict-no +)
- (S1 ^operator O2162 +)
- Retracting propose*predict-yes
- -->
- (O2161 ^name predict-yes +)
- (S1 ^operator O2161 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1084 ^value 1 +)
- (R1 ^reward R1084 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2162 = 0.3873355857129354)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*32
- -->
- (S1 ^operator O2162 = 0.6126638406640976)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2161 = 0.3895396347547306)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*31
- -->
- (S1 ^operator O2161 = -0.02274740735326741)
- =>WM: (15169: S1 ^operator O2164 +)
- =>WM: (15168: S1 ^operator O2163 +)
- =>WM: (15167: I3 ^dir U)
- =>WM: (15166: O2164 ^name predict-no)
- =>WM: (15165: O2163 ^name predict-yes)
- =>WM: (15164: R1085 ^value 1)
- =>WM: (15163: R1 ^reward R1085)
- <=WM: (15154: S1 ^operator O2161 +)
- <=WM: (15155: S1 ^operator O2162 +)
- <=WM: (15156: S1 ^operator O2162)
- <=WM: (15153: I3 ^dir L)
- <=WM: (15149: R1 ^reward R1084)
- <=WM: (15152: O2162 ^name predict-no)
- <=WM: (15151: O2161 ^name predict-yes)
- <=WM: (15150: R1084 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2163 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2164 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2162 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2161 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.719079 -0.331743 0.387336 -> 0.719079 -0.331743 0.387336(R,m,v=1,0.937173,0.0591899)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*32 0.28092 0.331744 0.612664 -> 0.28092 0.331744 0.612664(R,m,v=1,1,0)
- =>WM: (15170: S1 ^operator O2164)
- 1082: O: O2164 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1082 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1081 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15171: I3 ^predict-no N1082)
- <=WM: (15158: N1081 ^status complete)
- <=WM: (15157: I3 ^predict-no N1081)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- |\-/--- Input Phase ---
- =>WM: (15175: I2 ^dir L)
- =>WM: (15174: I2 ^reward 1)
- =>WM: (15173: I2 ^see 0)
- =>WM: (15172: N1082 ^status complete)
- <=WM: (15161: I2 ^dir U)
- <=WM: (15160: I2 ^reward 1)
- <=WM: (15159: I2 ^see 0)
- =>WM: (15176: I2 ^level-1 L0-root)
- <=WM: (15162: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2163 = 0.1599599085218832)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2164 = 0.612665523112643)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1086 ^value 1 +)
- (R1 ^reward R1086 +)
- Firing propose*predict-yes
- -->
- (O2165 ^name predict-yes +)
- (S1 ^operator O2165 +)
- Firing propose*predict-no
- -->
- (O2166 ^name predict-no +)
- (S1 ^operator O2166 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2164 = 0.3873356717563805)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2163 = 0.3895396347547306)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2164 ^name predict-no +)
- (S1 ^operator O2164 +)
- Retracting propose*predict-yes
- -->
- (O2163 ^name predict-yes +)
- (S1 ^operator O2163 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1085 ^value 1 +)
- (R1 ^reward R1085 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2164 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2163 = 0.)
- =>WM: (15183: S1 ^operator O2166 +)
- =>WM: (15182: S1 ^operator O2165 +)
- =>WM: (15181: I3 ^dir L)
- =>WM: (15180: O2166 ^name predict-no)
- =>WM: (15179: O2165 ^name predict-yes)
- =>WM: (15178: R1086 ^value 1)
- =>WM: (15177: R1 ^reward R1086)
- <=WM: (15168: S1 ^operator O2163 +)
- <=WM: (15169: S1 ^operator O2164 +)
- <=WM: (15170: S1 ^operator O2164)
- <=WM: (15167: I3 ^dir U)
- <=WM: (15163: R1 ^reward R1085)
- <=WM: (15166: O2164 ^name predict-no)
- <=WM: (15165: O2163 ^name predict-yes)
- <=WM: (15164: R1085 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2165 = 0.1599599085218832)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2165 = 0.3895396347547306)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2166 = 0.612665523112643)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2166 = 0.3873356717563805)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2164 = 0.3873356717563805)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2164 = 0.612665523112643)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2163 = 0.3895396347547306)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2163 = 0.1599599085218832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (15184: S1 ^operator O2166)
- 1083: O: O2166 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1083 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1082 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15185: I3 ^predict-no N1083)
- <=WM: (15172: N1082 ^status complete)
- <=WM: (15171: I3 ^predict-no N1082)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction L in state State-A
- In State-A moving L
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (15189: I2 ^dir R)
- =>WM: (15188: I2 ^reward 1)
- =>WM: (15187: I2 ^see 0)
- =>WM: (15186: N1083 ^status complete)
- <=WM: (15175: I2 ^dir L)
- <=WM: (15174: I2 ^reward 1)
- <=WM: (15173: I2 ^see 0)
- =>WM: (15190: I2 ^level-1 L0-root)
- <=WM: (15176: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2165 = 0.8155899301783588)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2166 = -0.00558448899823713)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1087 ^value 1 +)
- (R1 ^reward R1087 +)
- Firing propose*predict-yes
- -->
- (O2167 ^name predict-yes +)
- (S1 ^operator O2167 +)
- Firing propose*predict-no
- -->
- (O2168 ^name predict-no +)
- (S1 ^operator O2168 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2166 = 0.4476193025811009)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2165 = 0.1844115443735304)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2166 ^name predict-no +)
- (S1 ^operator O2166 +)
- Retracting propose*predict-yes
- -->
- (O2165 ^name predict-yes +)
- (S1 ^operator O2165 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1086 ^value 1 +)
- (R1 ^reward R1086 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2166 = 0.3873356717563805)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*37
- -->
- (S1 ^operator O2166 = 0.612665523112643)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2165 = 0.3895396347547306)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*38
- -->
- (S1 ^operator O2165 = 0.1599599085218832)
- =>WM: (15197: S1 ^operator O2168 +)
- =>WM: (15196: S1 ^operator O2167 +)
- =>WM: (15195: I3 ^dir R)
- =>WM: (15194: O2168 ^name predict-no)
- =>WM: (15193: O2167 ^name predict-yes)
- =>WM: (15192: R1087 ^value 1)
- =>WM: (15191: R1 ^reward R1087)
- <=WM: (15182: S1 ^operator O2165 +)
- <=WM: (15183: S1 ^operator O2166 +)
- <=WM: (15184: S1 ^operator O2166)
- <=WM: (15181: I3 ^dir L)
- <=WM: (15177: R1 ^reward R1086)
- <=WM: (15180: O2166 ^name predict-no)
- <=WM: (15179: O2165 ^name predict-yes)
- <=WM: (15178: R1086 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2167 = 0.8155899301783588)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2167 = 0.1844115443735304)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2168 = -0.00558448899823713)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2168 = 0.4476193025811009)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2166 = 0.4476193025811009)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2166 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2165 = 0.1844115443735304)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2165 = 0.8155899301783588)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*2 0.719079 -0.331743 0.387336 -> 0.719079 -0.331743 0.387335(R,m,v=1,0.9375,0.0589005)
- RL update rl*prefer*rvt*predict-no*H0*2*v1*H1*37 0.280923 0.331743 0.612666 -> 0.280922 0.331743 0.612665(R,m,v=1,1,0)
- =>WM: (15198: S1 ^operator O2167)
- 1084: O: O2167 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1084 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1083 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15199: I3 ^predict-yes N1084)
- <=WM: (15186: N1083 ^status complete)
- <=WM: (15185: I3 ^predict-no N1083)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (15203: I2 ^dir U)
- =>WM: (15202: I2 ^reward 1)
- =>WM: (15201: I2 ^see 1)
- =>WM: (15200: N1084 ^status complete)
- <=WM: (15189: I2 ^dir R)
- <=WM: (15188: I2 ^reward 1)
- <=WM: (15187: I2 ^see 0)
- =>WM: (15204: I2 ^level-1 R1-root)
- <=WM: (15190: I2 ^level-1 L0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1088 ^value 1 +)
- (R1 ^reward R1088 +)
- Firing propose*predict-yes
- -->
- (O2169 ^name predict-yes +)
- (S1 ^operator O2169 +)
- Firing propose*predict-no
- -->
- (O2170 ^name predict-no +)
- (S1 ^operator O2170 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2168 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2167 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2168 ^name predict-no +)
- (S1 ^operator O2168 +)
- Retracting propose*predict-yes
- -->
- (O2167 ^name predict-yes +)
- (S1 ^operator O2167 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1087 ^value 1 +)
- (R1 ^reward R1087 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2168 = 0.4476193025811009)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*33
- -->
- (S1 ^operator O2168 = -0.00558448899823713)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2167 = 0.1844115443735304)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*34
- -->
- (S1 ^operator O2167 = 0.8155899301783588)
- =>WM: (15212: S1 ^operator O2170 +)
- =>WM: (15211: S1 ^operator O2169 +)
- =>WM: (15210: I3 ^dir U)
- =>WM: (15209: O2170 ^name predict-no)
- =>WM: (15208: O2169 ^name predict-yes)
- =>WM: (15207: R1088 ^value 1)
- =>WM: (15206: R1 ^reward R1088)
- =>WM: (15205: I3 ^see 1)
- <=WM: (15196: S1 ^operator O2167 +)
- <=WM: (15198: S1 ^operator O2167)
- <=WM: (15197: S1 ^operator O2168 +)
- <=WM: (15195: I3 ^dir R)
- <=WM: (15191: R1 ^reward R1087)
- <=WM: (15148: I3 ^see 0)
- <=WM: (15194: O2168 ^name predict-no)
- <=WM: (15193: O2167 ^name predict-yes)
- <=WM: (15192: R1087 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2169 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2170 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2168 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2167 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675414 -0.491003 0.184412 -> 0.675414 -0.491003 0.184411(R,m,v=1,0.907104,0.0847295)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*34 0.324588 0.491002 0.81559 -> 0.324587 0.491002 0.81559(R,m,v=1,1,0)
- =>WM: (15213: S1 ^operator O2170)
- 1085: O: O2170 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1085 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1084 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15214: I3 ^predict-no N1085)
- <=WM: (15200: N1084 ^status complete)
- <=WM: (15199: I3 ^predict-yes N1084)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (15218: I2 ^dir R)
- =>WM: (15217: I2 ^reward 1)
- =>WM: (15216: I2 ^see 0)
- =>WM: (15215: N1085 ^status complete)
- <=WM: (15203: I2 ^dir U)
- <=WM: (15202: I2 ^reward 1)
- <=WM: (15201: I2 ^see 1)
- =>WM: (15219: I2 ^level-1 R1-root)
- <=WM: (15204: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2169 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2170 = 0.5523810272090074)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1089 ^value 1 +)
- (R1 ^reward R1089 +)
- Firing propose*predict-yes
- -->
- (O2171 ^name predict-yes +)
- (S1 ^operator O2171 +)
- Firing propose*predict-no
- -->
- (O2172 ^name predict-no +)
- (S1 ^operator O2172 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2170 = 0.4476193025811009)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2169 = 0.1844113231907469)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2170 ^name predict-no +)
- (S1 ^operator O2170 +)
- Retracting propose*predict-yes
- -->
- (O2169 ^name predict-yes +)
- (S1 ^operator O2169 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1088 ^value 1 +)
- (R1 ^reward R1088 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2170 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2169 = 0.)
- =>WM: (15227: S1 ^operator O2172 +)
- =>WM: (15226: S1 ^operator O2171 +)
- =>WM: (15225: I3 ^dir R)
- =>WM: (15224: O2172 ^name predict-no)
- =>WM: (15223: O2171 ^name predict-yes)
- =>WM: (15222: R1089 ^value 1)
- =>WM: (15221: R1 ^reward R1089)
- =>WM: (15220: I3 ^see 0)
- <=WM: (15211: S1 ^operator O2169 +)
- <=WM: (15212: S1 ^operator O2170 +)
- <=WM: (15213: S1 ^operator O2170)
- <=WM: (15210: I3 ^dir U)
- <=WM: (15206: R1 ^reward R1088)
- <=WM: (15205: I3 ^see 1)
- <=WM: (15209: O2170 ^name predict-no)
- <=WM: (15208: O2169 ^name predict-yes)
- <=WM: (15207: R1088 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2171 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2171 = 0.1844113231907469)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2172 = 0.5523810272090074)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2172 = 0.4476193025811009)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2170 = 0.4476193025811009)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2170 = 0.5523810272090074)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2169 = 0.1844113231907469)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2169 = 0.1398795999120246)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (15228: S1 ^operator O2172)
- 1086: O: O2172 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1086 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1085 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15229: I3 ^predict-no N1086)
- <=WM: (15215: N1085 ^status complete)
- <=WM: (15214: I3 ^predict-no N1085)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (15233: I2 ^dir R)
- =>WM: (15232: I2 ^reward 1)
- =>WM: (15231: I2 ^see 0)
- =>WM: (15230: N1086 ^status complete)
- <=WM: (15218: I2 ^dir R)
- <=WM: (15217: I2 ^reward 1)
- <=WM: (15216: I2 ^see 0)
- =>WM: (15234: I2 ^level-1 R0-root)
- <=WM: (15219: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2171 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2172 = 0.552380105014882)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1090 ^value 1 +)
- (R1 ^reward R1090 +)
- Firing propose*predict-yes
- -->
- (O2173 ^name predict-yes +)
- (S1 ^operator O2173 +)
- Firing propose*predict-no
- -->
- (O2174 ^name predict-no +)
- (S1 ^operator O2174 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2172 = 0.4476193025811009)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2171 = 0.1844113231907469)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2172 ^name predict-no +)
- (S1 ^operator O2172 +)
- Retracting propose*predict-yes
- -->
- (O2171 ^name predict-yes +)
- (S1 ^operator O2171 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1089 ^value 1 +)
- (R1 ^reward R1089 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2172 = 0.4476193025811009)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2172 = 0.5523810272090074)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2171 = 0.1844113231907469)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2171 = 0.1398795999120246)
- =>WM: (15240: S1 ^operator O2174 +)
- =>WM: (15239: S1 ^operator O2173 +)
- =>WM: (15238: O2174 ^name predict-no)
- =>WM: (15237: O2173 ^name predict-yes)
- =>WM: (15236: R1090 ^value 1)
- =>WM: (15235: R1 ^reward R1090)
- <=WM: (15226: S1 ^operator O2171 +)
- <=WM: (15227: S1 ^operator O2172 +)
- <=WM: (15228: S1 ^operator O2172)
- <=WM: (15221: R1 ^reward R1089)
- <=WM: (15224: O2172 ^name predict-no)
- <=WM: (15223: O2171 ^name predict-yes)
- <=WM: (15222: R1089 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2173 = 0.1844113231907469)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2173 = 0.1664311307472832)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2174 = 0.4476193025811009)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2174 = 0.552380105014882)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2172 = 0.4476193025811009)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2172 = 0.552380105014882)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2171 = 0.1844113231907469)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2171 = 0.1664311307472832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.93662,0.0597842)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*39 0.377467 0.174914 0.552381 -> 0.377467 0.174914 0.552381(R,m,v=1,1,0)
- =>WM: (15241: S1 ^operator O2174)
- 1087: O: O2174 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1087 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1086 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15242: I3 ^predict-no N1087)
- <=WM: (15230: N1086 ^status complete)
- <=WM: (15229: I3 ^predict-no N1086)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- |\-/--- Input Phase ---
- =>WM: (15246: I2 ^dir U)
- =>WM: (15245: I2 ^reward 1)
- =>WM: (15244: I2 ^see 0)
- =>WM: (15243: N1087 ^status complete)
- <=WM: (15233: I2 ^dir R)
- <=WM: (15232: I2 ^reward 1)
- <=WM: (15231: I2 ^see 0)
- =>WM: (15247: I2 ^level-1 R0-root)
- <=WM: (15234: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1091 ^value 1 +)
- (R1 ^reward R1091 +)
- Firing propose*predict-yes
- -->
- (O2175 ^name predict-yes +)
- (S1 ^operator O2175 +)
- Firing propose*predict-no
- -->
- (O2176 ^name predict-no +)
- (S1 ^operator O2176 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2174 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2173 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2174 ^name predict-no +)
- (S1 ^operator O2174 +)
- Retracting propose*predict-yes
- -->
- (O2173 ^name predict-yes +)
- (S1 ^operator O2173 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1090 ^value 1 +)
- (R1 ^reward R1090 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2174 = 0.552380105014882)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2174 = 0.4476192531125847)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2173 = 0.1664311307472832)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2173 = 0.1844113231907469)
- =>WM: (15254: S1 ^operator O2176 +)
- =>WM: (15253: S1 ^operator O2175 +)
- =>WM: (15252: I3 ^dir U)
- =>WM: (15251: O2176 ^name predict-no)
- =>WM: (15250: O2175 ^name predict-yes)
- =>WM: (15249: R1091 ^value 1)
- =>WM: (15248: R1 ^reward R1091)
- <=WM: (15239: S1 ^operator O2173 +)
- <=WM: (15240: S1 ^operator O2174 +)
- <=WM: (15241: S1 ^operator O2174)
- <=WM: (15225: I3 ^dir R)
- <=WM: (15235: R1 ^reward R1090)
- <=WM: (15238: O2174 ^name predict-no)
- <=WM: (15237: O2173 ^name predict-yes)
- <=WM: (15236: R1090 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2175 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2176 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2174 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2173 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.937063,0.0593913)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*41 0.377467 0.174913 0.55238 -> 0.377467 0.174913 0.55238(R,m,v=1,1,0)
- =>WM: (15255: S1 ^operator O2176)
- 1088: O: O2176 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1088 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1087 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15256: I3 ^predict-no N1088)
- <=WM: (15243: N1087 ^status complete)
- <=WM: (15242: I3 ^predict-no N1087)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (15260: I2 ^dir L)
- =>WM: (15259: I2 ^reward 1)
- =>WM: (15258: I2 ^see 0)
- =>WM: (15257: N1088 ^status complete)
- <=WM: (15246: I2 ^dir U)
- <=WM: (15245: I2 ^reward 1)
- <=WM: (15244: I2 ^see 0)
- =>WM: (15261: I2 ^level-1 R0-root)
- <=WM: (15247: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2175 = 0.6104606429053037)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2176 = 0.1063475139796038)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1092 ^value 1 +)
- (R1 ^reward R1092 +)
- Firing propose*predict-yes
- -->
- (O2177 ^name predict-yes +)
- (S1 ^operator O2177 +)
- Firing propose*predict-no
- -->
- (O2178 ^name predict-no +)
- (S1 ^operator O2178 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2176 = 0.3873354925260269)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2175 = 0.3895396347547306)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2176 ^name predict-no +)
- (S1 ^operator O2176 +)
- Retracting propose*predict-yes
- -->
- (O2175 ^name predict-yes +)
- (S1 ^operator O2175 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1091 ^value 1 +)
- (R1 ^reward R1091 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2176 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2175 = 0.)
- =>WM: (15268: S1 ^operator O2178 +)
- =>WM: (15267: S1 ^operator O2177 +)
- =>WM: (15266: I3 ^dir L)
- =>WM: (15265: O2178 ^name predict-no)
- =>WM: (15264: O2177 ^name predict-yes)
- =>WM: (15263: R1092 ^value 1)
- =>WM: (15262: R1 ^reward R1092)
- <=WM: (15253: S1 ^operator O2175 +)
- <=WM: (15254: S1 ^operator O2176 +)
- <=WM: (15255: S1 ^operator O2176)
- <=WM: (15252: I3 ^dir U)
- <=WM: (15248: R1 ^reward R1091)
- <=WM: (15251: O2176 ^name predict-no)
- <=WM: (15250: O2175 ^name predict-yes)
- <=WM: (15249: R1091 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2177 = 0.6104606429053037)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2177 = 0.3895396347547306)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2178 = 0.1063475139796038)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2178 = 0.3873354925260269)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2176 = 0.3873354925260269)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2176 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2175 = 0.3895396347547306)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2175 = 0.6104606429053037)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (15269: S1 ^operator O2177)
- 1089: O: O2177 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1089 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1088 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15270: I3 ^predict-yes N1089)
- <=WM: (15257: N1088 ^status complete)
- <=WM: (15256: I3 ^predict-no N1088)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- /|\---- Input Phase ---
- =>WM: (15274: I2 ^dir U)
- =>WM: (15273: I2 ^reward 1)
- =>WM: (15272: I2 ^see 1)
- =>WM: (15271: N1089 ^status complete)
- <=WM: (15260: I2 ^dir L)
- <=WM: (15259: I2 ^reward 1)
- <=WM: (15258: I2 ^see 0)
- =>WM: (15275: I2 ^level-1 L1-root)
- <=WM: (15261: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1093 ^value 1 +)
- (R1 ^reward R1093 +)
- Firing propose*predict-yes
- -->
- (O2179 ^name predict-yes +)
- (S1 ^operator O2179 +)
- Firing propose*predict-no
- -->
- (O2180 ^name predict-no +)
- (S1 ^operator O2180 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2178 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2177 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2178 ^name predict-no +)
- (S1 ^operator O2178 +)
- Retracting propose*predict-yes
- -->
- (O2177 ^name predict-yes +)
- (S1 ^operator O2177 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1092 ^value 1 +)
- (R1 ^reward R1092 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2178 = 0.3873354925260269)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2178 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2177 = 0.3895396347547306)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2177 = 0.6104606429053037)
- =>WM: (15283: S1 ^operator O2180 +)
- =>WM: (15282: S1 ^operator O2179 +)
- =>WM: (15281: I3 ^dir U)
- =>WM: (15280: O2180 ^name predict-no)
- =>WM: (15279: O2179 ^name predict-yes)
- =>WM: (15278: R1093 ^value 1)
- =>WM: (15277: R1 ^reward R1093)
- =>WM: (15276: I3 ^see 1)
- <=WM: (15267: S1 ^operator O2177 +)
- <=WM: (15269: S1 ^operator O2177)
- <=WM: (15268: S1 ^operator O2178 +)
- <=WM: (15266: I3 ^dir L)
- <=WM: (15262: R1 ^reward R1092)
- <=WM: (15220: I3 ^see 0)
- <=WM: (15265: O2178 ^name predict-no)
- <=WM: (15264: O2177 ^name predict-yes)
- <=WM: (15263: R1092 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2179 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2180 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2178 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2177 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.900552,0.0900552)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*44 0.288049 0.322412 0.610461 -> 0.288049 0.322412 0.610461(R,m,v=1,1,0)
- =>WM: (15284: S1 ^operator O2180)
- 1090: O: O2180 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1090 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1089 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15285: I3 ^predict-no N1090)
- <=WM: (15271: N1089 ^status complete)
- <=WM: (15270: I3 ^predict-yes N1089)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-A
- In State-A moving U
- ENV: (next state, see, prediction correct?) = (State-A, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- /|\---- Input Phase ---
- =>WM: (15289: I2 ^dir R)
- =>WM: (15288: I2 ^reward 1)
- =>WM: (15287: I2 ^see 0)
- =>WM: (15286: N1090 ^status complete)
- <=WM: (15274: I2 ^dir U)
- <=WM: (15273: I2 ^reward 1)
- <=WM: (15272: I2 ^see 1)
- =>WM: (15290: I2 ^level-1 L1-root)
- <=WM: (15275: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2180 = -0.02155734064455064)
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2179 = 0.8155855213644267)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1094 ^value 1 +)
- (R1 ^reward R1094 +)
- Firing propose*predict-yes
- -->
- (O2181 ^name predict-yes +)
- (S1 ^operator O2181 +)
- Firing propose*predict-no
- -->
- (O2182 ^name predict-no +)
- (S1 ^operator O2182 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2180 = 0.4476193493934647)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2179 = 0.1844113231907469)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2180 ^name predict-no +)
- (S1 ^operator O2180 +)
- Retracting propose*predict-yes
- -->
- (O2179 ^name predict-yes +)
- (S1 ^operator O2179 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1093 ^value 1 +)
- (R1 ^reward R1093 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2180 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2179 = 0.)
- =>WM: (15298: S1 ^operator O2182 +)
- =>WM: (15297: S1 ^operator O2181 +)
- =>WM: (15296: I3 ^dir R)
- =>WM: (15295: O2182 ^name predict-no)
- =>WM: (15294: O2181 ^name predict-yes)
- =>WM: (15293: R1094 ^value 1)
- =>WM: (15292: R1 ^reward R1094)
- =>WM: (15291: I3 ^see 0)
- <=WM: (15282: S1 ^operator O2179 +)
- <=WM: (15283: S1 ^operator O2180 +)
- <=WM: (15284: S1 ^operator O2180)
- <=WM: (15281: I3 ^dir U)
- <=WM: (15277: R1 ^reward R1093)
- <=WM: (15276: I3 ^see 1)
- <=WM: (15280: O2180 ^name predict-no)
- <=WM: (15279: O2179 ^name predict-yes)
- <=WM: (15278: R1093 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2181 = 0.8155855213644267)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2181 = 0.1844113231907469)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2182 = -0.02155734064455064)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2182 = 0.4476193493934647)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2180 = 0.4476193493934647)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2180 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2179 = 0.1844113231907469)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2179 = 0.8155855213644267)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (15299: S1 ^operator O2181)
- 1091: O: O2181 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1091 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1090 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15300: I3 ^predict-yes N1091)
- <=WM: (15286: N1090 ^status complete)
- <=WM: (15285: I3 ^predict-no N1090)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- /--- Input Phase ---
- =>WM: (15304: I2 ^dir L)
- =>WM: (15303: I2 ^reward 1)
- =>WM: (15302: I2 ^see 1)
- =>WM: (15301: N1091 ^status complete)
- <=WM: (15289: I2 ^dir R)
- <=WM: (15288: I2 ^reward 1)
- <=WM: (15287: I2 ^see 0)
- =>WM: (15305: I2 ^level-1 R1-root)
- <=WM: (15290: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2181 = 0.6104601064087616)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2182 = 0.2714993082286609)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1095 ^value 1 +)
- (R1 ^reward R1095 +)
- Firing propose*predict-yes
- -->
- (O2183 ^name predict-yes +)
- (S1 ^operator O2183 +)
- Firing propose*predict-no
- -->
- (O2184 ^name predict-no +)
- (S1 ^operator O2184 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2182 = 0.3873354925260269)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2181 = 0.3895395931057254)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2182 ^name predict-no +)
- (S1 ^operator O2182 +)
- Retracting propose*predict-yes
- -->
- (O2181 ^name predict-yes +)
- (S1 ^operator O2181 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1094 ^value 1 +)
- (R1 ^reward R1094 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2182 = 0.4476193493934647)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2182 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2181 = 0.1844113231907469)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2181 = 0.8155855213644267)
- =>WM: (15313: S1 ^operator O2184 +)
- =>WM: (15312: S1 ^operator O2183 +)
- =>WM: (15311: I3 ^dir L)
- =>WM: (15310: O2184 ^name predict-no)
- =>WM: (15309: O2183 ^name predict-yes)
- =>WM: (15308: R1095 ^value 1)
- =>WM: (15307: R1 ^reward R1095)
- =>WM: (15306: I3 ^see 1)
- <=WM: (15297: S1 ^operator O2181 +)
- <=WM: (15299: S1 ^operator O2181)
- <=WM: (15298: S1 ^operator O2182 +)
- <=WM: (15296: I3 ^dir R)
- <=WM: (15292: R1 ^reward R1094)
- <=WM: (15291: I3 ^see 0)
- <=WM: (15295: O2182 ^name predict-no)
- <=WM: (15294: O2181 ^name predict-yes)
- <=WM: (15293: R1094 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2183 = 0.3895395931057254)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2183 = 0.6104601064087616)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2184 = 0.3873354925260269)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2184 = 0.2714993082286609)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2182 = 0.3873354925260269)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2182 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2181 = 0.3895395931057254)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2181 = 0.6104601064087616)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675414 -0.491003 0.184411 -> 0.675415 -0.491003 0.184412(R,m,v=1,0.907609,0.0843134)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*46 0.324582 0.491004 0.815586 -> 0.324583 0.491003 0.815586(R,m,v=1,1,0)
- =>WM: (15314: S1 ^operator O2183)
- 1092: O: O2183 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1092 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1091 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15315: I3 ^predict-yes N1092)
- <=WM: (15301: N1091 ^status complete)
- <=WM: (15300: I3 ^predict-yes N1091)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (15319: I2 ^dir R)
- =>WM: (15318: I2 ^reward 1)
- =>WM: (15317: I2 ^see 1)
- =>WM: (15316: N1092 ^status complete)
- <=WM: (15304: I2 ^dir L)
- <=WM: (15303: I2 ^reward 1)
- <=WM: (15302: I2 ^see 1)
- =>WM: (15320: I2 ^level-1 L1-root)
- <=WM: (15305: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2184 = -0.02155734064455064)
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2183 = 0.8155859946811508)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1096 ^value 1 +)
- (R1 ^reward R1096 +)
- Firing propose*predict-yes
- -->
- (O2185 ^name predict-yes +)
- (S1 ^operator O2185 +)
- Firing propose*predict-no
- -->
- (O2186 ^name predict-no +)
- (S1 ^operator O2186 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2184 = 0.4476193493934647)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2183 = 0.1844117965074709)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2184 ^name predict-no +)
- (S1 ^operator O2184 +)
- Retracting propose*predict-yes
- -->
- (O2183 ^name predict-yes +)
- (S1 ^operator O2183 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1095 ^value 1 +)
- (R1 ^reward R1095 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2184 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2184 = 0.3873354925260269)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2183 = 0.6104601064087616)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2183 = 0.3895395931057254)
- =>WM: (15327: S1 ^operator O2186 +)
- =>WM: (15326: S1 ^operator O2185 +)
- =>WM: (15325: I3 ^dir R)
- =>WM: (15324: O2186 ^name predict-no)
- =>WM: (15323: O2185 ^name predict-yes)
- =>WM: (15322: R1096 ^value 1)
- =>WM: (15321: R1 ^reward R1096)
- <=WM: (15312: S1 ^operator O2183 +)
- <=WM: (15314: S1 ^operator O2183)
- <=WM: (15313: S1 ^operator O2184 +)
- <=WM: (15311: I3 ^dir L)
- <=WM: (15307: R1 ^reward R1095)
- <=WM: (15310: O2184 ^name predict-no)
- <=WM: (15309: O2183 ^name predict-yes)
- <=WM: (15308: R1095 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2185 = 0.1844117965074709)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2185 = 0.8155859946811508)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2186 = 0.4476193493934647)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2186 = -0.02155734064455064)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2184 = 0.4476193493934647)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2184 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2183 = 0.1844117965074709)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2183 = 0.8155859946811508)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.901099,0.089612)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*36 0.288049 0.322411 0.61046 -> 0.288049 0.322411 0.61046(R,m,v=1,1,0)
- =>WM: (15328: S1 ^operator O2185)
- 1093: O: O2185 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1093 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1092 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15329: I3 ^predict-yes N1093)
- <=WM: (15316: N1092 ^status complete)
- <=WM: (15315: I3 ^predict-yes N1092)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (15333: I2 ^dir L)
- =>WM: (15332: I2 ^reward 1)
- =>WM: (15331: I2 ^see 1)
- =>WM: (15330: N1093 ^status complete)
- <=WM: (15319: I2 ^dir R)
- <=WM: (15318: I2 ^reward 1)
- <=WM: (15317: I2 ^see 1)
- =>WM: (15334: I2 ^level-1 R1-root)
- <=WM: (15320: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2185 = 0.6104601514815886)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2186 = 0.2714993082286609)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1097 ^value 1 +)
- (R1 ^reward R1097 +)
- Firing propose*predict-yes
- -->
- (O2187 ^name predict-yes +)
- (S1 ^operator O2187 +)
- Firing propose*predict-no
- -->
- (O2188 ^name predict-no +)
- (S1 ^operator O2188 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2186 = 0.3873354925260269)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2185 = 0.3895396381785524)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2186 ^name predict-no +)
- (S1 ^operator O2186 +)
- Retracting propose*predict-yes
- -->
- (O2185 ^name predict-yes +)
- (S1 ^operator O2185 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1096 ^value 1 +)
- (R1 ^reward R1096 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2186 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2186 = 0.4476193493934647)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2185 = 0.8155859946811508)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2185 = 0.1844117965074709)
- =>WM: (15341: S1 ^operator O2188 +)
- =>WM: (15340: S1 ^operator O2187 +)
- =>WM: (15339: I3 ^dir L)
- =>WM: (15338: O2188 ^name predict-no)
- =>WM: (15337: O2187 ^name predict-yes)
- =>WM: (15336: R1097 ^value 1)
- =>WM: (15335: R1 ^reward R1097)
- <=WM: (15326: S1 ^operator O2185 +)
- <=WM: (15328: S1 ^operator O2185)
- <=WM: (15327: S1 ^operator O2186 +)
- <=WM: (15325: I3 ^dir R)
- <=WM: (15321: R1 ^reward R1096)
- <=WM: (15324: O2186 ^name predict-no)
- <=WM: (15323: O2185 ^name predict-yes)
- <=WM: (15322: R1096 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2187 = 0.3895396381785524)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2187 = 0.6104601514815886)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2188 = 0.3873354925260269)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2188 = 0.2714993082286609)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2186 = 0.3873354925260269)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2186 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2185 = 0.3895396381785524)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2185 = 0.6104601514815886)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675415 -0.491003 0.184412 -> 0.675415 -0.491003 0.184412(R,m,v=1,0.908108,0.0839013)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*46 0.324583 0.491003 0.815586 -> 0.324583 0.491003 0.815586(R,m,v=1,1,0)
- =>WM: (15342: S1 ^operator O2187)
- 1094: O: O2187 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1094 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1093 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15343: I3 ^predict-yes N1094)
- <=WM: (15330: N1093 ^status complete)
- <=WM: (15329: I3 ^predict-yes N1093)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (15347: I2 ^dir R)
- =>WM: (15346: I2 ^reward 1)
- =>WM: (15345: I2 ^see 1)
- =>WM: (15344: N1094 ^status complete)
- <=WM: (15333: I2 ^dir L)
- <=WM: (15332: I2 ^reward 1)
- <=WM: (15331: I2 ^see 1)
- =>WM: (15348: I2 ^level-1 L1-root)
- <=WM: (15334: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2188 = -0.02155734064455064)
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2187 = 0.8155863260028575)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1098 ^value 1 +)
- (R1 ^reward R1098 +)
- Firing propose*predict-yes
- -->
- (O2189 ^name predict-yes +)
- (S1 ^operator O2189 +)
- Firing propose*predict-no
- -->
- (O2190 ^name predict-no +)
- (S1 ^operator O2190 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2188 = 0.4476193493934647)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2187 = 0.1844121278291776)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2188 ^name predict-no +)
- (S1 ^operator O2188 +)
- Retracting propose*predict-yes
- -->
- (O2187 ^name predict-yes +)
- (S1 ^operator O2187 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1097 ^value 1 +)
- (R1 ^reward R1097 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*35
- -->
- (S1 ^operator O2188 = 0.2714993082286609)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2188 = 0.3873354925260269)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*36
- -->
- (S1 ^operator O2187 = 0.6104601514815886)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2187 = 0.3895396381785524)
- =>WM: (15355: S1 ^operator O2190 +)
- =>WM: (15354: S1 ^operator O2189 +)
- =>WM: (15353: I3 ^dir R)
- =>WM: (15352: O2190 ^name predict-no)
- =>WM: (15351: O2189 ^name predict-yes)
- =>WM: (15350: R1098 ^value 1)
- =>WM: (15349: R1 ^reward R1098)
- <=WM: (15340: S1 ^operator O2187 +)
- <=WM: (15342: S1 ^operator O2187)
- <=WM: (15341: S1 ^operator O2188 +)
- <=WM: (15339: I3 ^dir L)
- <=WM: (15335: R1 ^reward R1097)
- <=WM: (15338: O2188 ^name predict-no)
- <=WM: (15337: O2187 ^name predict-yes)
- <=WM: (15336: R1097 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2189 = 0.1844121278291776)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2189 = 0.8155863260028575)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2190 = 0.4476193493934647)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2190 = -0.02155734064455064)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2188 = 0.4476193493934647)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2188 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2187 = 0.1844121278291776)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2187 = 0.8155863260028575)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.901639,0.0891731)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*36 0.288049 0.322411 0.61046 -> 0.288049 0.322411 0.61046(R,m,v=1,1,0)
- =>WM: (15356: S1 ^operator O2189)
- 1095: O: O2189 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1095 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1094 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15357: I3 ^predict-yes N1095)
- <=WM: (15344: N1094 ^status complete)
- <=WM: (15343: I3 ^predict-yes N1094)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (15361: I2 ^dir R)
- =>WM: (15360: I2 ^reward 1)
- =>WM: (15359: I2 ^see 1)
- =>WM: (15358: N1095 ^status complete)
- <=WM: (15347: I2 ^dir R)
- <=WM: (15346: I2 ^reward 1)
- <=WM: (15345: I2 ^see 1)
- =>WM: (15362: I2 ^level-1 R1-root)
- <=WM: (15348: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2189 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2190 = 0.5523809777404911)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1099 ^value 1 +)
- (R1 ^reward R1099 +)
- Firing propose*predict-yes
- -->
- (O2191 ^name predict-yes +)
- (S1 ^operator O2191 +)
- Firing propose*predict-no
- -->
- (O2192 ^name predict-no +)
- (S1 ^operator O2192 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2190 = 0.4476193493934647)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2189 = 0.1844121278291776)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2190 ^name predict-no +)
- (S1 ^operator O2190 +)
- Retracting propose*predict-yes
- -->
- (O2189 ^name predict-yes +)
- (S1 ^operator O2189 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1098 ^value 1 +)
- (R1 ^reward R1098 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2190 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2190 = 0.4476193493934647)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2189 = 0.8155863260028575)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2189 = 0.1844121278291776)
- =>WM: (15368: S1 ^operator O2192 +)
- =>WM: (15367: S1 ^operator O2191 +)
- =>WM: (15366: O2192 ^name predict-no)
- =>WM: (15365: O2191 ^name predict-yes)
- =>WM: (15364: R1099 ^value 1)
- =>WM: (15363: R1 ^reward R1099)
- <=WM: (15354: S1 ^operator O2189 +)
- <=WM: (15356: S1 ^operator O2189)
- <=WM: (15355: S1 ^operator O2190 +)
- <=WM: (15349: R1 ^reward R1098)
- <=WM: (15352: O2190 ^name predict-no)
- <=WM: (15351: O2189 ^name predict-yes)
- <=WM: (15350: R1098 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2191 = 0.1844121278291776)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2191 = 0.1398795999120246)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2192 = 0.4476193493934647)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2192 = 0.5523809777404911)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2190 = 0.4476193493934647)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2190 = 0.5523809777404911)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2189 = 0.1844121278291776)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2189 = 0.1398795999120246)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675415 -0.491003 0.184412 -> 0.675415 -0.491003 0.184412(R,m,v=1,0.908602,0.0834932)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*46 0.324583 0.491003 0.815586 -> 0.324583 0.491003 0.815587(R,m,v=1,1,0)
- =>WM: (15369: S1 ^operator O2192)
- 1096: O: O2192 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1096 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1095 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15370: I3 ^predict-no N1096)
- <=WM: (15358: N1095 ^status complete)
- <=WM: (15357: I3 ^predict-yes N1095)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- |\-/sleeping...
- |--- Input Phase ---
- =>WM: (15374: I2 ^dir R)
- =>WM: (15373: I2 ^reward 1)
- =>WM: (15372: I2 ^see 0)
- =>WM: (15371: N1096 ^status complete)
- <=WM: (15361: I2 ^dir R)
- <=WM: (15360: I2 ^reward 1)
- <=WM: (15359: I2 ^see 1)
- =>WM: (15375: I2 ^level-1 R0-root)
- <=WM: (15362: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2191 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2192 = 0.5523802012957619)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1100 ^value 1 +)
- (R1 ^reward R1100 +)
- Firing propose*predict-yes
- -->
- (O2193 ^name predict-yes +)
- (S1 ^operator O2193 +)
- Firing propose*predict-no
- -->
- (O2194 ^name predict-no +)
- (S1 ^operator O2194 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2192 = 0.4476193493934647)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2191 = 0.1844123597543724)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2192 ^name predict-no +)
- (S1 ^operator O2192 +)
- Retracting propose*predict-yes
- -->
- (O2191 ^name predict-yes +)
- (S1 ^operator O2191 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1099 ^value 1 +)
- (R1 ^reward R1099 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2192 = 0.5523809777404911)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2192 = 0.4476193493934647)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2191 = 0.1398795999120246)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2191 = 0.1844123597543724)
- =>WM: (15382: S1 ^operator O2194 +)
- =>WM: (15381: S1 ^operator O2193 +)
- =>WM: (15380: O2194 ^name predict-no)
- =>WM: (15379: O2193 ^name predict-yes)
- =>WM: (15378: R1100 ^value 1)
- =>WM: (15377: R1 ^reward R1100)
- =>WM: (15376: I3 ^see 0)
- <=WM: (15367: S1 ^operator O2191 +)
- <=WM: (15368: S1 ^operator O2192 +)
- <=WM: (15369: S1 ^operator O2192)
- <=WM: (15363: R1 ^reward R1099)
- <=WM: (15306: I3 ^see 1)
- <=WM: (15366: O2192 ^name predict-no)
- <=WM: (15365: O2191 ^name predict-yes)
- <=WM: (15364: R1099 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2193 = 0.1844123597543724)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2193 = 0.1664311307472832)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2194 = 0.4476193493934647)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2194 = 0.5523802012957619)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2192 = 0.4476193493934647)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2192 = 0.5523802012957619)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2191 = 0.1844123597543724)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2191 = 0.1664311307472832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.9375,0.0590035)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*39 0.377467 0.174914 0.552381 -> 0.377467 0.174914 0.552381(R,m,v=1,1,0)
- =>WM: (15383: S1 ^operator O2194)
- 1097: O: O2194 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1097 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1096 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15384: I3 ^predict-no N1097)
- <=WM: (15371: N1096 ^status complete)
- <=WM: (15370: I3 ^predict-no N1096)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (15388: I2 ^dir R)
- =>WM: (15387: I2 ^reward 1)
- =>WM: (15386: I2 ^see 0)
- =>WM: (15385: N1097 ^status complete)
- <=WM: (15374: I2 ^dir R)
- <=WM: (15373: I2 ^reward 1)
- <=WM: (15372: I2 ^see 0)
- =>WM: (15389: I2 ^level-1 R0-root)
- <=WM: (15375: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2193 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2194 = 0.5523802012957619)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1101 ^value 1 +)
- (R1 ^reward R1101 +)
- Firing propose*predict-yes
- -->
- (O2195 ^name predict-yes +)
- (S1 ^operator O2195 +)
- Firing propose*predict-no
- -->
- (O2196 ^name predict-no +)
- (S1 ^operator O2196 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2194 = 0.4476193003233713)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2193 = 0.1844123597543724)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2194 ^name predict-no +)
- (S1 ^operator O2194 +)
- Retracting propose*predict-yes
- -->
- (O2193 ^name predict-yes +)
- (S1 ^operator O2193 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1100 ^value 1 +)
- (R1 ^reward R1100 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2194 = 0.5523802012957619)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2194 = 0.4476193003233713)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2193 = 0.1664311307472832)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2193 = 0.1844123597543724)
- =>WM: (15395: S1 ^operator O2196 +)
- =>WM: (15394: S1 ^operator O2195 +)
- =>WM: (15393: O2196 ^name predict-no)
- =>WM: (15392: O2195 ^name predict-yes)
- =>WM: (15391: R1101 ^value 1)
- =>WM: (15390: R1 ^reward R1101)
- <=WM: (15381: S1 ^operator O2193 +)
- <=WM: (15382: S1 ^operator O2194 +)
- <=WM: (15383: S1 ^operator O2194)
- <=WM: (15377: R1 ^reward R1100)
- <=WM: (15380: O2194 ^name predict-no)
- <=WM: (15379: O2193 ^name predict-yes)
- <=WM: (15378: R1100 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2195 = 0.1844123597543724)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2195 = 0.1664311307472832)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2196 = 0.4476193003233713)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2196 = 0.5523802012957619)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2194 = 0.4476193003233713)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2194 = 0.5523802012957619)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2193 = 0.1844123597543724)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2193 = 0.1664311307472832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.937931,0.0586207)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*41 0.377467 0.174913 0.55238 -> 0.377467 0.174913 0.55238(R,m,v=1,1,0)
- =>WM: (15396: S1 ^operator O2196)
- 1098: O: O2196 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1098 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1097 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15397: I3 ^predict-no N1098)
- <=WM: (15385: N1097 ^status complete)
- <=WM: (15384: I3 ^predict-no N1097)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- |\--- Input Phase ---
- =>WM: (15401: I2 ^dir U)
- =>WM: (15400: I2 ^reward 1)
- =>WM: (15399: I2 ^see 0)
- =>WM: (15398: N1098 ^status complete)
- <=WM: (15388: I2 ^dir R)
- <=WM: (15387: I2 ^reward 1)
- <=WM: (15386: I2 ^see 0)
- =>WM: (15402: I2 ^level-1 R0-root)
- <=WM: (15389: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1102 ^value 1 +)
- (R1 ^reward R1102 +)
- Firing propose*predict-yes
- -->
- (O2197 ^name predict-yes +)
- (S1 ^operator O2197 +)
- Firing propose*predict-no
- -->
- (O2198 ^name predict-no +)
- (S1 ^operator O2198 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2196 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2195 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2196 ^name predict-no +)
- (S1 ^operator O2196 +)
- Retracting propose*predict-yes
- -->
- (O2195 ^name predict-yes +)
- (S1 ^operator O2195 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1101 ^value 1 +)
- (R1 ^reward R1101 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2196 = 0.552380276052892)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2196 = 0.4476193750805013)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2195 = 0.1664311307472832)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2195 = 0.1844123597543724)
- =>WM: (15409: S1 ^operator O2198 +)
- =>WM: (15408: S1 ^operator O2197 +)
- =>WM: (15407: I3 ^dir U)
- =>WM: (15406: O2198 ^name predict-no)
- =>WM: (15405: O2197 ^name predict-yes)
- =>WM: (15404: R1102 ^value 1)
- =>WM: (15403: R1 ^reward R1102)
- <=WM: (15394: S1 ^operator O2195 +)
- <=WM: (15395: S1 ^operator O2196 +)
- <=WM: (15396: S1 ^operator O2196)
- <=WM: (15353: I3 ^dir R)
- <=WM: (15390: R1 ^reward R1101)
- <=WM: (15393: O2196 ^name predict-no)
- <=WM: (15392: O2195 ^name predict-yes)
- <=WM: (15391: R1101 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2197 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2198 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2196 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2195 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.938356,0.0582428)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*41 0.377467 0.174913 0.55238 -> 0.377467 0.174913 0.55238(R,m,v=1,1,0)
- =>WM: (15410: S1 ^operator O2198)
- 1099: O: O2198 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1099 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1098 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15411: I3 ^predict-no N1099)
- <=WM: (15398: N1098 ^status complete)
- <=WM: (15397: I3 ^predict-no N1098)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- -/|\--- Input Phase ---
- =>WM: (15415: I2 ^dir U)
- =>WM: (15414: I2 ^reward 1)
- =>WM: (15413: I2 ^see 0)
- =>WM: (15412: N1099 ^status complete)
- <=WM: (15401: I2 ^dir U)
- <=WM: (15400: I2 ^reward 1)
- <=WM: (15399: I2 ^see 0)
- =>WM: (15416: I2 ^level-1 R0-root)
- <=WM: (15402: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1103 ^value 1 +)
- (R1 ^reward R1103 +)
- Firing propose*predict-yes
- -->
- (O2199 ^name predict-yes +)
- (S1 ^operator O2199 +)
- Firing propose*predict-no
- -->
- (O2200 ^name predict-no +)
- (S1 ^operator O2200 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2198 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2197 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2198 ^name predict-no +)
- (S1 ^operator O2198 +)
- Retracting propose*predict-yes
- -->
- (O2197 ^name predict-yes +)
- (S1 ^operator O2197 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1102 ^value 1 +)
- (R1 ^reward R1102 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2198 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2197 = 0.)
- =>WM: (15422: S1 ^operator O2200 +)
- =>WM: (15421: S1 ^operator O2199 +)
- =>WM: (15420: O2200 ^name predict-no)
- =>WM: (15419: O2199 ^name predict-yes)
- =>WM: (15418: R1103 ^value 1)
- =>WM: (15417: R1 ^reward R1103)
- <=WM: (15408: S1 ^operator O2197 +)
- <=WM: (15409: S1 ^operator O2198 +)
- <=WM: (15410: S1 ^operator O2198)
- <=WM: (15403: R1 ^reward R1102)
- <=WM: (15406: O2198 ^name predict-no)
- <=WM: (15405: O2197 ^name predict-yes)
- <=WM: (15404: R1102 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2199 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2200 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2198 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2197 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (15423: S1 ^operator O2200)
- 1100: O: O2200 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1100 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1099 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15424: I3 ^predict-no N1100)
- <=WM: (15412: N1099 ^status complete)
- <=WM: (15411: I3 ^predict-no N1099)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (15428: I2 ^dir L)
- =>WM: (15427: I2 ^reward 1)
- =>WM: (15426: I2 ^see 0)
- =>WM: (15425: N1100 ^status complete)
- <=WM: (15415: I2 ^dir U)
- <=WM: (15414: I2 ^reward 1)
- <=WM: (15413: I2 ^see 0)
- =>WM: (15429: I2 ^level-1 R0-root)
- <=WM: (15416: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2199 = 0.6104606012562985)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2200 = 0.1063475139796038)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1104 ^value 1 +)
- (R1 ^reward R1104 +)
- Firing propose*predict-yes
- -->
- (O2201 ^name predict-yes +)
- (S1 ^operator O2201 +)
- Firing propose*predict-no
- -->
- (O2202 ^name predict-no +)
- (S1 ^operator O2202 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2200 = 0.3873354925260269)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2199 = 0.3895396697295312)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2200 ^name predict-no +)
- (S1 ^operator O2200 +)
- Retracting propose*predict-yes
- -->
- (O2199 ^name predict-yes +)
- (S1 ^operator O2199 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1103 ^value 1 +)
- (R1 ^reward R1103 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2200 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2199 = 0.)
- =>WM: (15436: S1 ^operator O2202 +)
- =>WM: (15435: S1 ^operator O2201 +)
- =>WM: (15434: I3 ^dir L)
- =>WM: (15433: O2202 ^name predict-no)
- =>WM: (15432: O2201 ^name predict-yes)
- =>WM: (15431: R1104 ^value 1)
- =>WM: (15430: R1 ^reward R1104)
- <=WM: (15421: S1 ^operator O2199 +)
- <=WM: (15422: S1 ^operator O2200 +)
- <=WM: (15423: S1 ^operator O2200)
- <=WM: (15407: I3 ^dir U)
- <=WM: (15417: R1 ^reward R1103)
- <=WM: (15420: O2200 ^name predict-no)
- <=WM: (15419: O2199 ^name predict-yes)
- <=WM: (15418: R1103 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2201 = 0.6104606012562985)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2201 = 0.3895396697295312)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2202 = 0.1063475139796038)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2202 = 0.3873354925260269)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2200 = 0.3873354925260269)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2200 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2199 = 0.3895396697295312)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2199 = 0.6104606012562985)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (15437: S1 ^operator O2201)
- 1101: O: O2201 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1101 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1100 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15438: I3 ^predict-yes N1101)
- <=WM: (15425: N1100 ^status complete)
- <=WM: (15424: I3 ^predict-no N1100)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- \--- Input Phase ---
- =>WM: (15442: I2 ^dir R)
- =>WM: (15441: I2 ^reward 1)
- =>WM: (15440: I2 ^see 1)
- =>WM: (15439: N1101 ^status complete)
- <=WM: (15428: I2 ^dir L)
- <=WM: (15427: I2 ^reward 1)
- <=WM: (15426: I2 ^see 0)
- =>WM: (15443: I2 ^level-1 L1-root)
- <=WM: (15429: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2202 = -0.02155734064455064)
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2201 = 0.8155865579280523)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1105 ^value 1 +)
- (R1 ^reward R1105 +)
- Firing propose*predict-yes
- -->
- (O2203 ^name predict-yes +)
- (S1 ^operator O2203 +)
- Firing propose*predict-no
- -->
- (O2204 ^name predict-no +)
- (S1 ^operator O2204 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2202 = 0.4476194274104922)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2201 = 0.1844123597543724)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2202 ^name predict-no +)
- (S1 ^operator O2202 +)
- Retracting propose*predict-yes
- -->
- (O2201 ^name predict-yes +)
- (S1 ^operator O2201 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1104 ^value 1 +)
- (R1 ^reward R1104 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2202 = 0.3873354925260269)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2202 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2201 = 0.3895396697295312)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2201 = 0.6104606012562985)
- =>WM: (15451: S1 ^operator O2204 +)
- =>WM: (15450: S1 ^operator O2203 +)
- =>WM: (15449: I3 ^dir R)
- =>WM: (15448: O2204 ^name predict-no)
- =>WM: (15447: O2203 ^name predict-yes)
- =>WM: (15446: R1105 ^value 1)
- =>WM: (15445: R1 ^reward R1105)
- =>WM: (15444: I3 ^see 1)
- <=WM: (15435: S1 ^operator O2201 +)
- <=WM: (15437: S1 ^operator O2201)
- <=WM: (15436: S1 ^operator O2202 +)
- <=WM: (15434: I3 ^dir L)
- <=WM: (15430: R1 ^reward R1104)
- <=WM: (15376: I3 ^see 0)
- <=WM: (15433: O2202 ^name predict-no)
- <=WM: (15432: O2201 ^name predict-yes)
- <=WM: (15431: R1104 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2203 = 0.1844123597543724)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2203 = 0.8155865579280523)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2204 = 0.4476194274104922)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2204 = -0.02155734064455064)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2202 = 0.4476194274104922)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2202 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2201 = 0.1844123597543724)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2201 = 0.8155865579280523)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.902174,0.0887384)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*44 0.288049 0.322412 0.610461 -> 0.288049 0.322412 0.610461(R,m,v=1,1,0)
- =>WM: (15452: S1 ^operator O2203)
- 1102: O: O2203 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1102 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1101 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15453: I3 ^predict-yes N1102)
- <=WM: (15439: N1101 ^status complete)
- <=WM: (15438: I3 ^predict-yes N1101)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (15457: I2 ^dir U)
- =>WM: (15456: I2 ^reward 1)
- =>WM: (15455: I2 ^see 1)
- =>WM: (15454: N1102 ^status complete)
- <=WM: (15442: I2 ^dir R)
- <=WM: (15441: I2 ^reward 1)
- <=WM: (15440: I2 ^see 1)
- =>WM: (15458: I2 ^level-1 R1-root)
- <=WM: (15443: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1106 ^value 1 +)
- (R1 ^reward R1106 +)
- Firing propose*predict-yes
- -->
- (O2205 ^name predict-yes +)
- (S1 ^operator O2205 +)
- Firing propose*predict-no
- -->
- (O2206 ^name predict-no +)
- (S1 ^operator O2206 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2204 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2203 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2204 ^name predict-no +)
- (S1 ^operator O2204 +)
- Retracting propose*predict-yes
- -->
- (O2203 ^name predict-yes +)
- (S1 ^operator O2203 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1105 ^value 1 +)
- (R1 ^reward R1105 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2204 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2204 = 0.4476194274104922)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2203 = 0.8155865579280523)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2203 = 0.1844123597543724)
- =>WM: (15465: S1 ^operator O2206 +)
- =>WM: (15464: S1 ^operator O2205 +)
- =>WM: (15463: I3 ^dir U)
- =>WM: (15462: O2206 ^name predict-no)
- =>WM: (15461: O2205 ^name predict-yes)
- =>WM: (15460: R1106 ^value 1)
- =>WM: (15459: R1 ^reward R1106)
- <=WM: (15450: S1 ^operator O2203 +)
- <=WM: (15452: S1 ^operator O2203)
- <=WM: (15451: S1 ^operator O2204 +)
- <=WM: (15449: I3 ^dir R)
- <=WM: (15445: R1 ^reward R1105)
- <=WM: (15448: O2204 ^name predict-no)
- <=WM: (15447: O2203 ^name predict-yes)
- <=WM: (15446: R1105 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2205 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2206 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2204 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2203 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675415 -0.491003 0.184412 -> 0.675416 -0.491003 0.184413(R,m,v=1,0.909091,0.083089)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*46 0.324583 0.491003 0.815587 -> 0.324583 0.491003 0.815587(R,m,v=1,1,0)
- =>WM: (15466: S1 ^operator O2206)
- 1103: O: O2206 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1103 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1102 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15467: I3 ^predict-no N1103)
- <=WM: (15454: N1102 ^status complete)
- <=WM: (15453: I3 ^predict-yes N1102)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (15471: I2 ^dir U)
- =>WM: (15470: I2 ^reward 1)
- =>WM: (15469: I2 ^see 0)
- =>WM: (15468: N1103 ^status complete)
- <=WM: (15457: I2 ^dir U)
- <=WM: (15456: I2 ^reward 1)
- <=WM: (15455: I2 ^see 1)
- =>WM: (15472: I2 ^level-1 R1-root)
- <=WM: (15458: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1107 ^value 1 +)
- (R1 ^reward R1107 +)
- Firing propose*predict-yes
- -->
- (O2207 ^name predict-yes +)
- (S1 ^operator O2207 +)
- Firing propose*predict-no
- -->
- (O2208 ^name predict-no +)
- (S1 ^operator O2208 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2206 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2205 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2206 ^name predict-no +)
- (S1 ^operator O2206 +)
- Retracting propose*predict-yes
- -->
- (O2205 ^name predict-yes +)
- (S1 ^operator O2205 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1106 ^value 1 +)
- (R1 ^reward R1106 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2206 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2205 = 0.)
- =>WM: (15479: S1 ^operator O2208 +)
- =>WM: (15478: S1 ^operator O2207 +)
- =>WM: (15477: O2208 ^name predict-no)
- =>WM: (15476: O2207 ^name predict-yes)
- =>WM: (15475: R1107 ^value 1)
- =>WM: (15474: R1 ^reward R1107)
- =>WM: (15473: I3 ^see 0)
- <=WM: (15464: S1 ^operator O2205 +)
- <=WM: (15465: S1 ^operator O2206 +)
- <=WM: (15466: S1 ^operator O2206)
- <=WM: (15459: R1 ^reward R1106)
- <=WM: (15444: I3 ^see 1)
- <=WM: (15462: O2206 ^name predict-no)
- <=WM: (15461: O2205 ^name predict-yes)
- <=WM: (15460: R1106 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2207 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2208 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2206 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2205 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (15480: S1 ^operator O2208)
- 1104: O: O2208 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1104 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1103 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15481: I3 ^predict-no N1104)
- <=WM: (15468: N1103 ^status complete)
- <=WM: (15467: I3 ^predict-no N1103)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (15485: I2 ^dir R)
- =>WM: (15484: I2 ^reward 1)
- =>WM: (15483: I2 ^see 0)
- =>WM: (15482: N1104 ^status complete)
- <=WM: (15471: I2 ^dir U)
- <=WM: (15470: I2 ^reward 1)
- <=WM: (15469: I2 ^see 0)
- =>WM: (15486: I2 ^level-1 R1-root)
- <=WM: (15472: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2207 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2208 = 0.5523809286703978)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1108 ^value 1 +)
- (R1 ^reward R1108 +)
- Firing propose*predict-yes
- -->
- (O2209 ^name predict-yes +)
- (S1 ^operator O2209 +)
- Firing propose*predict-no
- -->
- (O2210 ^name predict-no +)
- (S1 ^operator O2210 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2208 = 0.4476194274104922)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2207 = 0.1844125221020087)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2208 ^name predict-no +)
- (S1 ^operator O2208 +)
- Retracting propose*predict-yes
- -->
- (O2207 ^name predict-yes +)
- (S1 ^operator O2207 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1107 ^value 1 +)
- (R1 ^reward R1107 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2208 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2207 = 0.)
- =>WM: (15493: S1 ^operator O2210 +)
- =>WM: (15492: S1 ^operator O2209 +)
- =>WM: (15491: I3 ^dir R)
- =>WM: (15490: O2210 ^name predict-no)
- =>WM: (15489: O2209 ^name predict-yes)
- =>WM: (15488: R1108 ^value 1)
- =>WM: (15487: R1 ^reward R1108)
- <=WM: (15478: S1 ^operator O2207 +)
- <=WM: (15479: S1 ^operator O2208 +)
- <=WM: (15480: S1 ^operator O2208)
- <=WM: (15463: I3 ^dir U)
- <=WM: (15474: R1 ^reward R1107)
- <=WM: (15477: O2208 ^name predict-no)
- <=WM: (15476: O2207 ^name predict-yes)
- <=WM: (15475: R1107 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2209 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2209 = 0.1844125221020087)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2210 = 0.5523809286703978)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2210 = 0.4476194274104922)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2208 = 0.4476194274104922)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2208 = 0.5523809286703978)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2207 = 0.1844125221020087)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2207 = 0.1398795999120246)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (15494: S1 ^operator O2210)
- 1105: O: O2210 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1105 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1104 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15495: I3 ^predict-no N1105)
- <=WM: (15482: N1104 ^status complete)
- <=WM: (15481: I3 ^predict-no N1104)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (15499: I2 ^dir R)
- =>WM: (15498: I2 ^reward 1)
- =>WM: (15497: I2 ^see 0)
- =>WM: (15496: N1105 ^status complete)
- <=WM: (15485: I2 ^dir R)
- <=WM: (15484: I2 ^reward 1)
- <=WM: (15483: I2 ^see 0)
- =>WM: (15500: I2 ^level-1 R0-root)
- <=WM: (15486: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2209 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2210 = 0.552380328382883)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1109 ^value 1 +)
- (R1 ^reward R1109 +)
- Firing propose*predict-yes
- -->
- (O2211 ^name predict-yes +)
- (S1 ^operator O2211 +)
- Firing propose*predict-no
- -->
- (O2212 ^name predict-no +)
- (S1 ^operator O2212 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2210 = 0.4476194274104922)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2209 = 0.1844125221020087)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2210 ^name predict-no +)
- (S1 ^operator O2210 +)
- Retracting propose*predict-yes
- -->
- (O2209 ^name predict-yes +)
- (S1 ^operator O2209 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1108 ^value 1 +)
- (R1 ^reward R1108 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2210 = 0.4476194274104922)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2210 = 0.5523809286703978)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2209 = 0.1844125221020087)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2209 = 0.1398795999120246)
- =>WM: (15506: S1 ^operator O2212 +)
- =>WM: (15505: S1 ^operator O2211 +)
- =>WM: (15504: O2212 ^name predict-no)
- =>WM: (15503: O2211 ^name predict-yes)
- =>WM: (15502: R1109 ^value 1)
- =>WM: (15501: R1 ^reward R1109)
- <=WM: (15492: S1 ^operator O2209 +)
- <=WM: (15493: S1 ^operator O2210 +)
- <=WM: (15494: S1 ^operator O2210)
- <=WM: (15487: R1 ^reward R1108)
- <=WM: (15490: O2210 ^name predict-no)
- <=WM: (15489: O2209 ^name predict-yes)
- <=WM: (15488: R1108 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2211 = 0.1844125221020087)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2211 = 0.1664311307472832)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2212 = 0.4476194274104922)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2212 = 0.552380328382883)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2210 = 0.4476194274104922)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2210 = 0.552380328382883)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2209 = 0.1844125221020087)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2209 = 0.1664311307472832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.938776,0.0578697)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*39 0.377467 0.174914 0.552381 -> 0.377467 0.174914 0.552381(R,m,v=1,1,0)
- =>WM: (15507: S1 ^operator O2212)
- 1106: O: O2212 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1106 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1105 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15508: I3 ^predict-no N1106)
- <=WM: (15496: N1105 ^status complete)
- <=WM: (15495: I3 ^predict-no N1105)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (15512: I2 ^dir U)
- =>WM: (15511: I2 ^reward 1)
- =>WM: (15510: I2 ^see 0)
- =>WM: (15509: N1106 ^status complete)
- <=WM: (15499: I2 ^dir R)
- <=WM: (15498: I2 ^reward 1)
- <=WM: (15497: I2 ^see 0)
- =>WM: (15513: I2 ^level-1 R0-root)
- <=WM: (15500: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1110 ^value 1 +)
- (R1 ^reward R1110 +)
- Firing propose*predict-yes
- -->
- (O2213 ^name predict-yes +)
- (S1 ^operator O2213 +)
- Firing propose*predict-no
- -->
- (O2214 ^name predict-no +)
- (S1 ^operator O2214 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2212 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2211 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2212 ^name predict-no +)
- (S1 ^operator O2212 +)
- Retracting propose*predict-yes
- -->
- (O2211 ^name predict-yes +)
- (S1 ^operator O2211 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1109 ^value 1 +)
- (R1 ^reward R1109 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2212 = 0.552380328382883)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2212 = 0.4476193739983587)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2211 = 0.1664311307472832)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2211 = 0.1844125221020087)
- =>WM: (15520: S1 ^operator O2214 +)
- =>WM: (15519: S1 ^operator O2213 +)
- =>WM: (15518: I3 ^dir U)
- =>WM: (15517: O2214 ^name predict-no)
- =>WM: (15516: O2213 ^name predict-yes)
- =>WM: (15515: R1110 ^value 1)
- =>WM: (15514: R1 ^reward R1110)
- <=WM: (15505: S1 ^operator O2211 +)
- <=WM: (15506: S1 ^operator O2212 +)
- <=WM: (15507: S1 ^operator O2212)
- <=WM: (15491: I3 ^dir R)
- <=WM: (15501: R1 ^reward R1109)
- <=WM: (15504: O2212 ^name predict-no)
- <=WM: (15503: O2211 ^name predict-yes)
- <=WM: (15502: R1109 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2213 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2214 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2212 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2211 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.939189,0.0575014)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*41 0.377467 0.174913 0.55238 -> 0.377467 0.174913 0.55238(R,m,v=1,1,0)
- =>WM: (15521: S1 ^operator O2214)
- 1107: O: O2214 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1107 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1106 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15522: I3 ^predict-no N1107)
- <=WM: (15509: N1106 ^status complete)
- <=WM: (15508: I3 ^predict-no N1106)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (15526: I2 ^dir R)
- =>WM: (15525: I2 ^reward 1)
- =>WM: (15524: I2 ^see 0)
- =>WM: (15523: N1107 ^status complete)
- <=WM: (15512: I2 ^dir U)
- <=WM: (15511: I2 ^reward 1)
- <=WM: (15510: I2 ^see 0)
- =>WM: (15527: I2 ^level-1 R0-root)
- <=WM: (15513: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2213 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2214 = 0.5523803730256968)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1111 ^value 1 +)
- (R1 ^reward R1111 +)
- Firing propose*predict-yes
- -->
- (O2215 ^name predict-yes +)
- (S1 ^operator O2215 +)
- Firing propose*predict-no
- -->
- (O2216 ^name predict-no +)
- (S1 ^operator O2216 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2214 = 0.4476194186411724)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2213 = 0.1844125221020087)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2214 ^name predict-no +)
- (S1 ^operator O2214 +)
- Retracting propose*predict-yes
- -->
- (O2213 ^name predict-yes +)
- (S1 ^operator O2213 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1110 ^value 1 +)
- (R1 ^reward R1110 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2214 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2213 = 0.)
- =>WM: (15534: S1 ^operator O2216 +)
- =>WM: (15533: S1 ^operator O2215 +)
- =>WM: (15532: I3 ^dir R)
- =>WM: (15531: O2216 ^name predict-no)
- =>WM: (15530: O2215 ^name predict-yes)
- =>WM: (15529: R1111 ^value 1)
- =>WM: (15528: R1 ^reward R1111)
- <=WM: (15519: S1 ^operator O2213 +)
- <=WM: (15520: S1 ^operator O2214 +)
- <=WM: (15521: S1 ^operator O2214)
- <=WM: (15518: I3 ^dir U)
- <=WM: (15514: R1 ^reward R1110)
- <=WM: (15517: O2214 ^name predict-no)
- <=WM: (15516: O2213 ^name predict-yes)
- <=WM: (15515: R1110 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2215 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2215 = 0.1844125221020087)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2216 = 0.5523803730256968)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2216 = 0.4476194186411724)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2214 = 0.4476194186411724)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2214 = 0.5523803730256968)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2213 = 0.1844125221020087)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2213 = 0.1664311307472832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (15535: S1 ^operator O2216)
- 1108: O: O2216 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1108 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1107 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15536: I3 ^predict-no N1108)
- <=WM: (15523: N1107 ^status complete)
- <=WM: (15522: I3 ^predict-no N1107)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (15540: I2 ^dir L)
- =>WM: (15539: I2 ^reward 1)
- =>WM: (15538: I2 ^see 0)
- =>WM: (15537: N1108 ^status complete)
- <=WM: (15526: I2 ^dir R)
- <=WM: (15525: I2 ^reward 1)
- <=WM: (15524: I2 ^see 0)
- =>WM: (15541: I2 ^level-1 R0-root)
- <=WM: (15527: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2215 = 0.610460560608424)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2216 = 0.1063475139796038)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1112 ^value 1 +)
- (R1 ^reward R1112 +)
- Firing propose*predict-yes
- -->
- (O2217 ^name predict-yes +)
- (S1 ^operator O2217 +)
- Firing propose*predict-no
- -->
- (O2218 ^name predict-no +)
- (S1 ^operator O2218 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2216 = 0.3873354925260269)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2215 = 0.3895396290816568)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2216 ^name predict-no +)
- (S1 ^operator O2216 +)
- Retracting propose*predict-yes
- -->
- (O2215 ^name predict-yes +)
- (S1 ^operator O2215 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1111 ^value 1 +)
- (R1 ^reward R1111 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2216 = 0.4476194186411724)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2216 = 0.5523803730256968)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2215 = 0.1844125221020087)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2215 = 0.1664311307472832)
- =>WM: (15548: S1 ^operator O2218 +)
- =>WM: (15547: S1 ^operator O2217 +)
- =>WM: (15546: I3 ^dir L)
- =>WM: (15545: O2218 ^name predict-no)
- =>WM: (15544: O2217 ^name predict-yes)
- =>WM: (15543: R1112 ^value 1)
- =>WM: (15542: R1 ^reward R1112)
- <=WM: (15533: S1 ^operator O2215 +)
- <=WM: (15534: S1 ^operator O2216 +)
- <=WM: (15535: S1 ^operator O2216)
- <=WM: (15532: I3 ^dir R)
- <=WM: (15528: R1 ^reward R1111)
- <=WM: (15531: O2216 ^name predict-no)
- <=WM: (15530: O2215 ^name predict-yes)
- <=WM: (15529: R1111 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2217 = 0.610460560608424)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2217 = 0.3895396290816568)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2218 = 0.1063475139796038)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2218 = 0.3873354925260269)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2216 = 0.3873354925260269)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2216 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2215 = 0.3895396290816568)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2215 = 0.610460560608424)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.939597,0.0571377)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*41 0.377467 0.174913 0.55238 -> 0.377467 0.174913 0.55238(R,m,v=1,1,0)
- =>WM: (15549: S1 ^operator O2217)
- 1109: O: O2217 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1109 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1108 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15550: I3 ^predict-yes N1109)
- <=WM: (15537: N1108 ^status complete)
- <=WM: (15536: I3 ^predict-no N1108)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- /|--- Input Phase ---
- =>WM: (15554: I2 ^dir R)
- =>WM: (15553: I2 ^reward 1)
- =>WM: (15552: I2 ^see 1)
- =>WM: (15551: N1109 ^status complete)
- <=WM: (15540: I2 ^dir L)
- <=WM: (15539: I2 ^reward 1)
- <=WM: (15538: I2 ^see 0)
- =>WM: (15555: I2 ^level-1 L1-root)
- <=WM: (15541: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2218 = -0.02155734064455064)
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2217 = 0.8155867202756886)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1113 ^value 1 +)
- (R1 ^reward R1113 +)
- Firing propose*predict-yes
- -->
- (O2219 ^name predict-yes +)
- (S1 ^operator O2219 +)
- Firing propose*predict-no
- -->
- (O2220 ^name predict-no +)
- (S1 ^operator O2220 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2218 = 0.447619449891142)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2217 = 0.1844125221020087)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2218 ^name predict-no +)
- (S1 ^operator O2218 +)
- Retracting propose*predict-yes
- -->
- (O2217 ^name predict-yes +)
- (S1 ^operator O2217 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1112 ^value 1 +)
- (R1 ^reward R1112 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2218 = 0.3873354925260269)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2218 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2217 = 0.3895396290816568)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2217 = 0.610460560608424)
- =>WM: (15563: S1 ^operator O2220 +)
- =>WM: (15562: S1 ^operator O2219 +)
- =>WM: (15561: I3 ^dir R)
- =>WM: (15560: O2220 ^name predict-no)
- =>WM: (15559: O2219 ^name predict-yes)
- =>WM: (15558: R1113 ^value 1)
- =>WM: (15557: R1 ^reward R1113)
- =>WM: (15556: I3 ^see 1)
- <=WM: (15547: S1 ^operator O2217 +)
- <=WM: (15549: S1 ^operator O2217)
- <=WM: (15548: S1 ^operator O2218 +)
- <=WM: (15546: I3 ^dir L)
- <=WM: (15542: R1 ^reward R1112)
- <=WM: (15473: I3 ^see 0)
- <=WM: (15545: O2218 ^name predict-no)
- <=WM: (15544: O2217 ^name predict-yes)
- <=WM: (15543: R1112 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2219 = 0.1844125221020087)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2219 = 0.8155867202756886)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2220 = 0.447619449891142)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2220 = -0.02155734064455064)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2218 = 0.447619449891142)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2218 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2217 = 0.1844125221020087)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2217 = 0.8155867202756886)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.902703,0.0883079)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*44 0.288049 0.322412 0.610461 -> 0.288049 0.322412 0.610461(R,m,v=1,1,0)
- =>WM: (15564: S1 ^operator O2219)
- 1110: O: O2219 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1110 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1109 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15565: I3 ^predict-yes N1110)
- <=WM: (15551: N1109 ^status complete)
- <=WM: (15550: I3 ^predict-yes N1109)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (15569: I2 ^dir R)
- =>WM: (15568: I2 ^reward 1)
- =>WM: (15567: I2 ^see 1)
- =>WM: (15566: N1110 ^status complete)
- <=WM: (15554: I2 ^dir R)
- <=WM: (15553: I2 ^reward 1)
- <=WM: (15552: I2 ^see 1)
- =>WM: (15570: I2 ^level-1 R1-root)
- <=WM: (15555: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2219 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2220 = 0.5523808752582643)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1114 ^value 1 +)
- (R1 ^reward R1114 +)
- Firing propose*predict-yes
- -->
- (O2221 ^name predict-yes +)
- (S1 ^operator O2221 +)
- Firing propose*predict-no
- -->
- (O2222 ^name predict-no +)
- (S1 ^operator O2222 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2220 = 0.447619449891142)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2219 = 0.1844125221020087)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2220 ^name predict-no +)
- (S1 ^operator O2220 +)
- Retracting propose*predict-yes
- -->
- (O2219 ^name predict-yes +)
- (S1 ^operator O2219 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1113 ^value 1 +)
- (R1 ^reward R1113 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2220 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2220 = 0.447619449891142)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2219 = 0.8155867202756886)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2219 = 0.1844125221020087)
- =>WM: (15576: S1 ^operator O2222 +)
- =>WM: (15575: S1 ^operator O2221 +)
- =>WM: (15574: O2222 ^name predict-no)
- =>WM: (15573: O2221 ^name predict-yes)
- =>WM: (15572: R1114 ^value 1)
- =>WM: (15571: R1 ^reward R1114)
- <=WM: (15562: S1 ^operator O2219 +)
- <=WM: (15564: S1 ^operator O2219)
- <=WM: (15563: S1 ^operator O2220 +)
- <=WM: (15557: R1 ^reward R1113)
- <=WM: (15560: O2220 ^name predict-no)
- <=WM: (15559: O2219 ^name predict-yes)
- <=WM: (15558: R1113 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2221 = 0.1844125221020087)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2221 = 0.1398795999120246)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2222 = 0.447619449891142)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2222 = 0.5523808752582643)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2220 = 0.447619449891142)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2220 = 0.5523808752582643)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2219 = 0.1844125221020087)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2219 = 0.1398795999120246)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675416 -0.491003 0.184413 -> 0.675416 -0.491003 0.184413(R,m,v=1,0.909574,0.0826886)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*46 0.324583 0.491003 0.815587 -> 0.324584 0.491003 0.815587(R,m,v=1,1,0)
- =>WM: (15577: S1 ^operator O2222)
- 1111: O: O2222 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1111 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1110 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15578: I3 ^predict-no N1111)
- <=WM: (15566: N1110 ^status complete)
- <=WM: (15565: I3 ^predict-yes N1110)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- |--- Input Phase ---
- =>WM: (15582: I2 ^dir U)
- =>WM: (15581: I2 ^reward 1)
- =>WM: (15580: I2 ^see 0)
- =>WM: (15579: N1111 ^status complete)
- <=WM: (15569: I2 ^dir R)
- <=WM: (15568: I2 ^reward 1)
- <=WM: (15567: I2 ^see 1)
- =>WM: (15583: I2 ^level-1 R0-root)
- <=WM: (15570: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1115 ^value 1 +)
- (R1 ^reward R1115 +)
- Firing propose*predict-yes
- -->
- (O2223 ^name predict-yes +)
- (S1 ^operator O2223 +)
- Firing propose*predict-no
- -->
- (O2224 ^name predict-no +)
- (S1 ^operator O2224 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2222 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2221 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2222 ^name predict-no +)
- (S1 ^operator O2222 +)
- Retracting propose*predict-yes
- -->
- (O2221 ^name predict-yes +)
- (S1 ^operator O2221 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1114 ^value 1 +)
- (R1 ^reward R1114 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2222 = 0.5523808752582643)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2222 = 0.447619449891142)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2221 = 0.1398795999120246)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2221 = 0.1844126357453541)
- =>WM: (15591: S1 ^operator O2224 +)
- =>WM: (15590: S1 ^operator O2223 +)
- =>WM: (15589: I3 ^dir U)
- =>WM: (15588: O2224 ^name predict-no)
- =>WM: (15587: O2223 ^name predict-yes)
- =>WM: (15586: R1115 ^value 1)
- =>WM: (15585: R1 ^reward R1115)
- =>WM: (15584: I3 ^see 0)
- <=WM: (15575: S1 ^operator O2221 +)
- <=WM: (15576: S1 ^operator O2222 +)
- <=WM: (15577: S1 ^operator O2222)
- <=WM: (15561: I3 ^dir R)
- <=WM: (15571: R1 ^reward R1114)
- <=WM: (15556: I3 ^see 1)
- <=WM: (15574: O2222 ^name predict-no)
- <=WM: (15573: O2221 ^name predict-yes)
- <=WM: (15572: R1114 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2223 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2224 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2222 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2221 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.94,0.0567785)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*39 0.377467 0.174914 0.552381 -> 0.377467 0.174914 0.552381(R,m,v=1,1,0)
- =>WM: (15592: S1 ^operator O2224)
- 1112: O: O2224 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1112 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1111 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15593: I3 ^predict-no N1112)
- <=WM: (15579: N1111 ^status complete)
- <=WM: (15578: I3 ^predict-no N1111)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (15597: I2 ^dir R)
- =>WM: (15596: I2 ^reward 1)
- =>WM: (15595: I2 ^see 0)
- =>WM: (15594: N1112 ^status complete)
- <=WM: (15582: I2 ^dir U)
- <=WM: (15581: I2 ^reward 1)
- <=WM: (15580: I2 ^see 0)
- =>WM: (15598: I2 ^level-1 R0-root)
- <=WM: (15583: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2223 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2224 = 0.5523804042756664)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1116 ^value 1 +)
- (R1 ^reward R1116 +)
- Firing propose*predict-yes
- -->
- (O2225 ^name predict-yes +)
- (S1 ^operator O2225 +)
- Firing propose*predict-no
- -->
- (O2226 ^name predict-no +)
- (S1 ^operator O2226 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2224 = 0.447619401118731)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2223 = 0.1844126357453541)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2224 ^name predict-no +)
- (S1 ^operator O2224 +)
- Retracting propose*predict-yes
- -->
- (O2223 ^name predict-yes +)
- (S1 ^operator O2223 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1115 ^value 1 +)
- (R1 ^reward R1115 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2224 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2223 = 0.)
- =>WM: (15605: S1 ^operator O2226 +)
- =>WM: (15604: S1 ^operator O2225 +)
- =>WM: (15603: I3 ^dir R)
- =>WM: (15602: O2226 ^name predict-no)
- =>WM: (15601: O2225 ^name predict-yes)
- =>WM: (15600: R1116 ^value 1)
- =>WM: (15599: R1 ^reward R1116)
- <=WM: (15590: S1 ^operator O2223 +)
- <=WM: (15591: S1 ^operator O2224 +)
- <=WM: (15592: S1 ^operator O2224)
- <=WM: (15589: I3 ^dir U)
- <=WM: (15585: R1 ^reward R1115)
- <=WM: (15588: O2224 ^name predict-no)
- <=WM: (15587: O2223 ^name predict-yes)
- <=WM: (15586: R1115 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2225 = 0.1664311307472832)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2225 = 0.1844126357453541)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2226 = 0.5523804042756664)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2226 = 0.447619401118731)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2224 = 0.447619401118731)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2224 = 0.5523804042756664)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2223 = 0.1844126357453541)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2223 = 0.1664311307472832)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*6 1 0 1 -> 1 0 1(R,m,v=1,1,0)
- =>WM: (15606: S1 ^operator O2226)
- 1113: O: O2226 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1113 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1112 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15607: I3 ^predict-no N1113)
- <=WM: (15594: N1112 ^status complete)
- <=WM: (15593: I3 ^predict-no N1112)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isL
- --- END Output Phase ---
- |\---- Input Phase ---
- =>WM: (15611: I2 ^dir L)
- =>WM: (15610: I2 ^reward 1)
- =>WM: (15609: I2 ^see 0)
- =>WM: (15608: N1113 ^status complete)
- <=WM: (15597: I2 ^dir R)
- <=WM: (15596: I2 ^reward 1)
- <=WM: (15595: I2 ^see 0)
- =>WM: (15612: I2 ^level-1 R0-root)
- <=WM: (15598: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2225 = 0.6104605321549119)
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2226 = 0.1063475139796038)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1117 ^value 1 +)
- (R1 ^reward R1117 +)
- Firing propose*predict-yes
- -->
- (O2227 ^name predict-yes +)
- (S1 ^operator O2227 +)
- Firing propose*predict-no
- -->
- (O2228 ^name predict-no +)
- (S1 ^operator O2228 +)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2226 = 0.3873354925260269)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2225 = 0.3895396006281447)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2226 ^name predict-no +)
- (S1 ^operator O2226 +)
- Retracting propose*predict-yes
- -->
- (O2225 ^name predict-yes +)
- (S1 ^operator O2225 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1116 ^value 1 +)
- (R1 ^reward R1116 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2226 = 0.447619401118731)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*41
- -->
- (S1 ^operator O2226 = 0.5523804042756664)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2225 = 0.1844126357453541)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*42
- -->
- (S1 ^operator O2225 = 0.1664311307472832)
- =>WM: (15619: S1 ^operator O2228 +)
- =>WM: (15618: S1 ^operator O2227 +)
- =>WM: (15617: I3 ^dir L)
- =>WM: (15616: O2228 ^name predict-no)
- =>WM: (15615: O2227 ^name predict-yes)
- =>WM: (15614: R1117 ^value 1)
- =>WM: (15613: R1 ^reward R1117)
- <=WM: (15604: S1 ^operator O2225 +)
- <=WM: (15605: S1 ^operator O2226 +)
- <=WM: (15606: S1 ^operator O2226)
- <=WM: (15603: I3 ^dir R)
- <=WM: (15599: R1 ^reward R1116)
- <=WM: (15602: O2226 ^name predict-no)
- <=WM: (15601: O2225 ^name predict-yes)
- <=WM: (15600: R1116 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2227 = 0.6104605321549119)
- Firing rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2227 = 0.3895396006281447)
- Firing prefer*rvt*predict-yes*H0*1*v1*H1
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2228 = 0.1063475139796038)
- Firing rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2228 = 0.3873354925260269)
- Firing prefer*rvt*predict-no*H0*2*v1*H1
- -->
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2226 = 0.3873354925260269)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2226 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2225 = 0.3895396006281447)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2225 = 0.6104605321549119)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.940397,0.0564238)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*41 0.377467 0.174913 0.55238 -> 0.377467 0.174913 0.55238(R,m,v=1,1,0)
- =>WM: (15620: S1 ^operator O2227)
- 1114: O: O2227 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1114 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1113 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15621: I3 ^predict-yes N1114)
- <=WM: (15608: N1113 ^status complete)
- <=WM: (15607: I3 ^predict-no N1113)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction L in state State-B
- In State-B moving L
- ENV: (next state, see, prediction correct?) = (State-A, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- /|\--- Input Phase ---
- =>WM: (15625: I2 ^dir R)
- =>WM: (15624: I2 ^reward 1)
- =>WM: (15623: I2 ^see 1)
- =>WM: (15622: N1114 ^status complete)
- <=WM: (15611: I2 ^dir L)
- <=WM: (15610: I2 ^reward 1)
- <=WM: (15609: I2 ^see 0)
- =>WM: (15626: I2 ^level-1 L1-root)
- <=WM: (15612: I2 ^level-1 R0-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2228 = -0.02155734064455064)
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2227 = 0.815586833919034)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1118 ^value 1 +)
- (R1 ^reward R1118 +)
- Firing propose*predict-yes
- -->
- (O2229 ^name predict-yes +)
- (S1 ^operator O2229 +)
- Firing propose*predict-no
- -->
- (O2230 ^name predict-no +)
- (S1 ^operator O2230 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2228 = 0.4476194303095714)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2227 = 0.1844126357453541)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Retracting propose*predict-no
- -->
- (O2228 ^name predict-no +)
- (S1 ^operator O2228 +)
- Retracting propose*predict-yes
- -->
- (O2227 ^name predict-yes +)
- (S1 ^operator O2227 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1117 ^value 1 +)
- (R1 ^reward R1117 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir L +)
- Retracting rl*prefer*rvt*predict-no*H0*2
- -->
- (S1 ^operator O2228 = 0.3873354925260269)
- Retracting rl*prefer*rvt*predict-no*H0*2*v1*H1*43
- -->
- (S1 ^operator O2228 = 0.1063475139796038)
- Retracting rl*prefer*rvt*predict-yes*H0*1
- -->
- (S1 ^operator O2227 = 0.3895396006281447)
- Retracting rl*prefer*rvt*predict-yes*H0*1*v1*H1*44
- -->
- (S1 ^operator O2227 = 0.6104605321549119)
- =>WM: (15634: S1 ^operator O2230 +)
- =>WM: (15633: S1 ^operator O2229 +)
- =>WM: (15632: I3 ^dir R)
- =>WM: (15631: O2230 ^name predict-no)
- =>WM: (15630: O2229 ^name predict-yes)
- =>WM: (15629: R1118 ^value 1)
- =>WM: (15628: R1 ^reward R1118)
- =>WM: (15627: I3 ^see 1)
- <=WM: (15618: S1 ^operator O2227 +)
- <=WM: (15620: S1 ^operator O2227)
- <=WM: (15619: S1 ^operator O2228 +)
- <=WM: (15617: I3 ^dir L)
- <=WM: (15613: R1 ^reward R1117)
- <=WM: (15584: I3 ^see 0)
- <=WM: (15616: O2228 ^name predict-no)
- <=WM: (15615: O2227 ^name predict-yes)
- <=WM: (15614: R1117 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2229 = 0.1844126357453541)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2229 = 0.815586833919034)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2230 = 0.4476194303095714)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2230 = -0.02155734064455064)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2228 = 0.4476194303095714)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2228 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2227 = 0.1844126357453541)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2227 = 0.815586833919034)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*1 0.711951 -0.322412 0.38954 -> 0.711951 -0.322412 0.38954(R,m,v=1,0.903226,0.0878814)
- RL update rl*prefer*rvt*predict-yes*H0*1*v1*H1*44 0.288049 0.322412 0.610461 -> 0.288049 0.322412 0.610461(R,m,v=1,1,0)
- =>WM: (15635: S1 ^operator O2229)
- 1115: O: O2229 (predict-yes)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-yes N1115 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1114 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15636: I3 ^predict-yes N1115)
- <=WM: (15622: N1114 ^status complete)
- <=WM: (15621: I3 ^predict-yes N1114)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-yes inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-yes for direction R in state State-A
- In State-A moving R
- ENV: (next state, see, prediction correct?) = (State-B, 1, True)
- predict error 0
- dir: dir isR
- --- END Output Phase ---
- -/|--- Input Phase ---
- =>WM: (15640: I2 ^dir R)
- =>WM: (15639: I2 ^reward 1)
- =>WM: (15638: I2 ^see 1)
- =>WM: (15637: N1115 ^status complete)
- <=WM: (15625: I2 ^dir R)
- <=WM: (15624: I2 ^reward 1)
- <=WM: (15623: I2 ^see 1)
- =>WM: (15641: I2 ^level-1 R1-root)
- <=WM: (15626: I2 ^level-1 L1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2229 = 0.1398795999120246)
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2230 = 0.5523808264858534)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1119 ^value 1 +)
- (R1 ^reward R1119 +)
- Firing propose*predict-yes
- -->
- (O2231 ^name predict-yes +)
- (S1 ^operator O2231 +)
- Firing propose*predict-no
- -->
- (O2232 ^name predict-no +)
- (S1 ^operator O2232 +)
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2230 = 0.4476194303095714)
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2229 = 0.1844126357453541)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2230 ^name predict-no +)
- (S1 ^operator O2230 +)
- Retracting propose*predict-yes
- -->
- (O2229 ^name predict-yes +)
- (S1 ^operator O2229 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1118 ^value 1 +)
- (R1 ^reward R1118 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*45
- -->
- (S1 ^operator O2230 = -0.02155734064455064)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2230 = 0.4476194303095714)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*46
- -->
- (S1 ^operator O2229 = 0.815586833919034)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2229 = 0.1844126357453541)
- =>WM: (15647: S1 ^operator O2232 +)
- =>WM: (15646: S1 ^operator O2231 +)
- =>WM: (15645: O2232 ^name predict-no)
- =>WM: (15644: O2231 ^name predict-yes)
- =>WM: (15643: R1119 ^value 1)
- =>WM: (15642: R1 ^reward R1119)
- <=WM: (15633: S1 ^operator O2229 +)
- <=WM: (15635: S1 ^operator O2229)
- <=WM: (15634: S1 ^operator O2230 +)
- <=WM: (15628: R1 ^reward R1118)
- <=WM: (15631: O2230 ^name predict-no)
- <=WM: (15630: O2229 ^name predict-yes)
- <=WM: (15629: R1118 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2231 = 0.1844126357453541)
- Firing prefer*rvt*predict-yes*H0*3*v1*H1
- -->
- Firing rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2231 = 0.1398795999120246)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2232 = 0.4476194303095714)
- Firing prefer*rvt*predict-no*H0*4*v1*H1
- -->
- Firing rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2232 = 0.5523808264858534)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2230 = 0.4476194303095714)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2230 = 0.5523808264858534)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2229 = 0.1844126357453541)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2229 = 0.1398795999120246)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-yes*H0*3 0.675416 -0.491003 0.184413 -> 0.675416 -0.491003 0.184413(R,m,v=1,0.910053,0.082292)
- RL update rl*prefer*rvt*predict-yes*H0*3*v1*H1*46 0.324584 0.491003 0.815587 -> 0.324584 0.491003 0.815587(R,m,v=1,1,0)
- =>WM: (15648: S1 ^operator O2232)
- 1116: O: O2232 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1116 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-yes N1115 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15649: I3 ^predict-no N1116)
- <=WM: (15637: N1115 ^status complete)
- <=WM: (15636: I3 ^predict-yes N1115)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction R in state State-B
- In State-B moving R
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict error 0
- dir: dir isU
- --- END Output Phase ---
- \-/--- Input Phase ---
- =>WM: (15653: I2 ^dir U)
- =>WM: (15652: I2 ^reward 1)
- =>WM: (15651: I2 ^see 0)
- =>WM: (15650: N1116 ^status complete)
- <=WM: (15640: I2 ^dir R)
- <=WM: (15639: I2 ^reward 1)
- <=WM: (15638: I2 ^see 1)
- =>WM: (15654: I2 ^level-1 R0-root)
- <=WM: (15641: I2 ^level-1 R1-root)
- --- END Input Phase ---
- --- Proposal Phase ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing elaborate*copy-see-to-output-link
- -->
- (I3 ^see 0 +)
- Firing elaborate*reward*based*on*reward
- -->
- (R1120 ^value 1 +)
- (R1 ^reward R1120 +)
- Firing propose*predict-yes
- -->
- (O2233 ^name predict-yes +)
- (S1 ^operator O2233 +)
- Firing propose*predict-no
- -->
- (O2234 ^name predict-no +)
- (S1 ^operator O2234 +)
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2232 = 0.9999999999999999)
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2231 = 0.)
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing prefer*rvt*predict-no*H0
- -->
- Firing elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir U +)
- inner elaboration loop at bottom goal.
- Retracting elaborate*copy-see-to-output-link
- -->
- (I3 ^see 1 +)
- Retracting propose*predict-no
- -->
- (O2232 ^name predict-no +)
- (S1 ^operator O2232 +)
- Retracting propose*predict-yes
- -->
- (O2231 ^name predict-yes +)
- (S1 ^operator O2231 +)
- Retracting elaborate*reward*based*on*reward
- -->
- (R1119 ^value 1 +)
- (R1 ^reward R1119 +)
- Retracting elaborate*copy-dir-to-output-link
- -->
- (I3 ^dir R +)
- Retracting rl*prefer*rvt*predict-no*H0*4*v1*H1*39
- -->
- (S1 ^operator O2232 = 0.5523808264858534)
- Retracting rl*prefer*rvt*predict-no*H0*4
- -->
- (S1 ^operator O2232 = 0.4476194303095714)
- Retracting rl*prefer*rvt*predict-yes*H0*3*v1*H1*40
- -->
- (S1 ^operator O2231 = 0.1398795999120246)
- Retracting rl*prefer*rvt*predict-yes*H0*3
- -->
- (S1 ^operator O2231 = 0.1844127152956959)
- =>WM: (15662: S1 ^operator O2234 +)
- =>WM: (15661: S1 ^operator O2233 +)
- =>WM: (15660: I3 ^dir U)
- =>WM: (15659: O2234 ^name predict-no)
- =>WM: (15658: O2233 ^name predict-yes)
- =>WM: (15657: R1120 ^value 1)
- =>WM: (15656: R1 ^reward R1120)
- =>WM: (15655: I3 ^see 0)
- <=WM: (15646: S1 ^operator O2231 +)
- <=WM: (15647: S1 ^operator O2232 +)
- <=WM: (15648: S1 ^operator O2232)
- <=WM: (15632: I3 ^dir R)
- <=WM: (15642: R1 ^reward R1119)
- <=WM: (15627: I3 ^see 1)
- <=WM: (15645: O2232 ^name predict-no)
- <=WM: (15644: O2231 ^name predict-yes)
- <=WM: (15643: R1119 ^value 1)
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing prefer*rvt*predict-yes*H0
- -->
- Firing rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2233 = 0.)
- Firing prefer*rvt*predict-no*H0
- -->
- Firing rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2234 = 0.9999999999999999)
- inner elaboration loop at bottom goal.
- Retracting rl*prefer*rvt*predict-no*H0*6
- -->
- (S1 ^operator O2232 = 0.9999999999999999)
- Retracting rl*prefer*rvt*predict-yes*H0*5
- -->
- (S1 ^operator O2231 = 0.)
- --- END Proposal Phase ---
- --- Decision Phase ---
- RL update rl*prefer*rvt*predict-no*H0*4 0.622533 -0.174914 0.447619 -> 0.622533 -0.174914 0.447619(R,m,v=1,0.940789,0.0560735)
- RL update rl*prefer*rvt*predict-no*H0*4*v1*H1*39 0.377467 0.174914 0.552381 -> 0.377467 0.174914 0.552381(R,m,v=1,1,0)
- =>WM: (15663: S1 ^operator O2234)
- 1117: O: O2234 (predict-no)
- --- END Decision Phase ---
- --- Application Phase ---
- --- Firing Productions (PE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing apply*operator
- -->
- (I3 ^predict-no N1117 + :O )
- Firing apply*operator*complete
- -->
- (I3 ^predict-no N1116 - :O )
- inner elaboration loop at bottom goal.
- --- Change Working Memory (PE) ---
- =>WM: (15664: I3 ^predict-no N1117)
- <=WM: (15650: N1116 ^status complete)
- <=WM: (15649: I3 ^predict-no N1116)
- --- Firing Productions (IE) For State At Depth 1 ---
- --- Inner Elaboration Phase, active level 1 (S1) ---
- Firing monitor*world
- -->
- I see 1 and I'm going to do: predict-no inner elaboration loop at bottom goal.
- --- Change Working Memory (IE) ---
- --- END Application Phase ---
- --- Output Phase ---
- ENV: Agent did: predict-no for direction U in state State-B
- In State-B moving U
- ENV: (next state, see, prediction correct?) = (State-B, 0, True)
- predict er