We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Hello dear mr. Yuhang Song,
In the paper, it is mentioned that the rewards for action v are given by And the parameters θv are optimized according to the rule:
In the code,
DHP/envs.py
Lines 488 to 502 in 73ddec2
DHP/suppor_lib.py
Lines 154 to 159 in 73ddec2
DHP/a3c.py
Lines 238 to 239 in 73ddec2
Is there any particular reason why the direct sum of rewards is not calculated, and instead the above approach is considered?
The text was updated successfully, but these errors were encountered:
Bump!
Sorry, something went wrong.
No branches or pull requests
Hello dear mr. Yuhang Song,
In the paper, it is mentioned that the rewards for action v are given by
And the parameters θv are optimized according to the rule:
In the code,
DHP/envs.py
Lines 488 to 502 in 73ddec2
there seems to be no reward for v calculated, instead v_lable is estimated as a "weighted" target value (sum of subject_i_v * similarity),
DHP/suppor_lib.py
Lines 154 to 159 in 73ddec2
which then contributes another term (v-v_lable)^2 in the loss function:
DHP/a3c.py
Lines 238 to 239 in 73ddec2
Is there any particular reason why the direct sum of rewards is not calculated, and instead the above approach is considered?
The text was updated successfully, but these errors were encountered: