The Wayback Machine - https://web.archive.org/web/20220422081938/https://github.com/AI4Finance-Foundation/FinRL/issues/432
Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Mechanism of Repositioning: sell low valuable assets in exchange of high valuable assets (increase the turnover rate) #432

Open
BruceYanghy opened this issue Jan 16, 2022 · 4 comments
Labels
discussion good first issue help_wanted

Comments

@BruceYanghy
Copy link
Member

@BruceYanghy BruceYanghy commented Jan 16, 2022

We found that during the backtesting period of the StockTrading demo, the number of stock adjustments was very small. Most stocks established a large position in the early stage of trading, and then the position was hardly adjusted. Very little repositioning. As the time increases, our portfolio has no different than the market index.

This picture shows the stock price (left column) and position (right column) of the backtest range we are looking at. Different colored lines in the right column represent the results of multiple trials. x-axis represents the time. We observe that many stocks are rarely rebalanced (3-year backtest interval).

We are asking for discussion to make the RL agent increase the turnover rate/repositioning frequency during the trading process.

Thanks.

image

@BruceYanghy BruceYanghy added discussion good first issue help_wanted labels Jan 16, 2022
@BruceYanghy BruceYanghy changed the title Mechanism of Repositioning: sell low valuable assets in exchange of high valuable assets Mechanism of Repositioning: sell low valuable assets in exchange of high valuable assets (increase the turnover rate) Jan 16, 2022
@lcavalie
Copy link

@lcavalie lcavalie commented Jan 18, 2022

@A5230171
Copy link

@A5230171 A5230171 commented Jan 22, 2022

能否通过对 奖励函数(Reward Function)的定义来改变智能体(Agent)的行为(Action),因为总体上股价是呈上升趋势的,所以智能体只要一开始选择持有股票,到回合结束后也能获得不错的收益(奖励 Reward),因此如果在智能体【持有股票】的过程中会 持续的、定期的 扣取 管理费用,以此来减少智能体长期持有股票的累计奖励,激励智能体 能够 稍微频繁 交易,通过赚取少量的差价,来增大 在回合内 赚取的 钱呢? ------- 在训练环境中定义

通过训练环境中对 奖励函数 的定义,而测试环境中不受到 扣取 持有股票管理费的限制,能否在测试环境中,改变智能体的行为,智能体不会总是长期持有股票,而是会中间过程中,能在预测到股票价格下跌的情况下,会选择抛售股票,再等预测到股票价格上涨的时机时,再买入来赚取更多的差价呢?

以上是我的设想,还需要后续实验的验证

@EternalConfession
Copy link

@EternalConfession EternalConfession commented Mar 29, 2022

Why this issue is closed.

I also find that the agent will no longger trading after buy a large position in the early stage of trading.

I also tried to change the fee to 0, in case that the agent will be more willing to do more trading, but in vain. In this case, the agent seems not to learn any reasonable strategy.

@XiaoYangLiu-FinRL
Copy link
Member

@XiaoYangLiu-FinRL XiaoYangLiu-FinRL commented Mar 29, 2022

Now we have more discussions, since you joined. Let me reopen it. LoL.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
discussion good first issue help_wanted
5 participants