-
Notifications
You must be signed in to change notification settings - Fork 19
Description
I am having trouble understanding why for term 1, the sum is taken between the two computed entropy terms:
deep-active-inference-mc/src/tfmodel.py
Lines 343 to 344 in c40ef0d
| # E [ log Q(s|pi) - log Q(s|o,pi) ] | |
| term1 = - tf.reduce_sum(entropy_normal_from_logvar(ps1_logvar) + entropy_normal_from_logvar(qs1_logvar), axis=1) |
I understand this gives the sum of two entropy terms (where ps1 qs1
But in the paper, we see that term 1 is given by:
Why is there the discrepancy between the "+" and the "-"? Or where is my understanding breaking down? Am I simplifying the equations incorrectly? If so, can you explain how to correctly transform between the two?