You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, I have a doubt about the reward function.
If in a specific state I get the action to close a LONG position, why do I calculate the reward starting from the closing of the next candle instead of the current one in which the action was indicated?
Hi, I have a doubt about the reward function.
If in a specific state I get the action to close a LONG position, why do I calculate the reward starting from the closing of the next candle instead of the current one in which the action was indicated?
in code: trading_env.py
def step(self, action):
self._truncated = False
self._current_tick += 1
why increase self._current_tick += 1 before reward?
shouldn't I reward him with the closing value of the candle in which this action was decided?
thank you!
The text was updated successfully, but these errors were encountered: