强化学习
夏普比率
计算机科学
交易策略
人工智能
股票市场
算法交易
投资策略
文件夹
投资组合优化
深度学习
机器学习
库存(枪支)
增强学习
市场流动性
计量经济学
经济
金融经济学
财务
古生物学
机械工程
工程类
马
生物
作者
Hongyang Yang,Xiao-Yang Liu,Shan Zhong,Anwar Walid
标识
DOI:10.1145/3383455.3422540
摘要
Stock trading strategies play a critical role in investment. However, it is challenging to design a profitable strategy in a complex and dynamic stock market. In this paper, we propose an ensemble strategy that employs deep reinforcement schemes to learn a stock trading strategy by maximizing investment return. We train a deep reinforcement learning agent and obtain an ensemble trading strategy using three actor-critic based algorithms: Proximal Policy Optimization (PPO), Advantage Actor Critic (A2C), and Deep Deterministic Policy Gradient (DDPG). The ensemble strategy inherits and integrates the best features of the three algorithms, thereby robustly adjusting to different market situations. In order to avoid the large memory consumption in training networks with continuous action space, we employ a load-on-demand technique for processing very large data. We test our algorithms on the 30 Dow Jones stocks that have adequate liquidity. The performance of the trading agent with different reinforcement learning algorithms is evaluated and compared with both the Dow Jones Industrial Average index and the traditional min-variance portfolio allocation strategy. The proposed deep ensemble strategy is shown to outperform the three individual algorithms and two baselines in terms of the risk-adjusted return measured by the Sharpe ratio.
科研通智能强力驱动
Strongly Powered by AbleSci AI