Modelling A.I. in Economics

A NEW ARCHITECTURE FOR GAME THEORETIC DEEP REINFORCEMENT LEARNING

 

ABSTRACT

Game theory can be applied to the stock market in several ways. One way is to analyze the strategic decision-making of market participants, such as traders, investors, and companies. For example, game theory can be used to understand how traders might make decisions about when to buy or sell a particular stock, or how companies might make decisions about when to issue new stock or buy back existing stock. Another way that game theory can be applied to the stock market is to analyze the overall market dynamics and how they might affect stock prices. For example, game theory can be used to understand how the actions of individual market participants might influence the overall market and how different market conditions, such as supply and demand, might affect stock prices. Game theory and neural networks can be used together in a variety of ways. One way to use game theory with neural networks is to apply game-theoretic concepts and techniques to the design and analysis of neural networks. We trained this model using Reinforcement Learning from decision functions (game theory). We trained an initial model using supervised fine-tuning to understand the strategic behavior of agents that are trained to interact with each other using neural networks. To create a reward model for reinforcement learning, we needed to collect test data, which consisted of two or more model responses statistically ranked by quality. To collect this data, we use best-response functions (represent the action that a player will take in response to the actions of the other players.) Results show that our method can significantly improve the accuracy of game theory based stock prediction models. The accuracy of our method is 92.54%, which is significantly higher than the accuracy of the baseline methods.


REFERENCES

  1. Jorgenson, D., Gollop, F.M. and Fraumeni, B., 2016. Productivity and US economic growth. Elsevier.
  2. Dulac-Arnold, G., Evans, R., van Hasselt, H., Sunehag, P., Lillicrap, T., Hunt, J., Mann, T., Weber, T., Degris, T. and Coppin, B., 2015. Deep reinforcement learning in large discrete action spaces. arXiv preprint arXiv:1512.07679.
  3. Sunehag, P., Evans, R., Dulac-Arnold, G., Zwols, Y., Visentin, D. and Coppin, B., 2015. Deep reinforcement learning with attention for slate markov decision processes with high-dimensional states and actions. arXiv preprint arXiv:1512.01124.
  4. Nair, A., Srinivasan, P., Blackwell, S., Alcicek, C., Fearon, R., De Maria, A., Panneershelvam, V., Suleyman, M., Beattie, C., Petersen, S. and Legg, S., 2015. Massively parallel methods for deep reinforcement learning. arXiv preprint arXiv:1507.04296.
  5. Mann, T.A., Penedones, H., Mannor, S. and Hester, T., 2016. Adaptive lambda least-squares temporal difference learning. arXiv preprint arXiv:1612.09465.
  6. Nair, A., McGrew, B., Andrychowicz, M., Zaremba, W. and Abbeel, P., 2018, May. Overcoming exploration in reinforcement learning with demonstrations. In 2018 IEEE international conference on robotics and automation (ICRA) (pp. 6292-6299). IEEE.
  7. Sunehag, P., Lever, G., Gruslys, A., Czarnecki, W.M., Zambaldi, V., Jaderberg, M., Lanctot, M., Sonnerat, N., Leibo, J.Z., Tuyls, K. and Graepel, T., 2017. Value-decomposition networks for cooperative multi-agent learning. arXiv preprint arXiv:1706.05296.
  8. Fedus, W., Rosca, M., Lakshminarayanan, B., Dai, A.M., Mohamed, S. and Goodfellow, I., 2017. Many paths to equilibrium: GANs do not need to decrease a divergence at every step. arXiv preprint arXiv:1710.08446.
  9. Ma, X., Xia, L., Zhou, Z., Yang, J. and Zhao, Q., 2020. Dsac: Distributional soft actor critic for risk-sensitive reinforcement learning. arXiv preprint arXiv:2004.14547.


Read Full Article

People also ask

⚐ What are the top stocks to invest in right now?
☵ What happens to stocks when they're delisted?
This project is licensed under the license; additional terms may apply.