Reinforcement Learning for Trading Practical Examples and Lessons Learned by Dr. Tom Starke

preview_player
Показать описание
This talk, titled, “Reinforcement Learning for Trading Practical Examples and Lessons Learned” was given by Dr. Tom Starke at QuantCon 2018.

Description:
Since AlphaGo beat the world Go champion, reinforcement learning has received considerable attention and seems like an attractive choice for completely autonomous trading systems. This talk shows practical aspects and examples of deep reinforcement learning applied to trading and discusses the pros and cons of this technology.

About the Speaker:
Dr. Tom Starke has a Ph.D. in Physics and works as an algorithmic trader at a proprietary trading company in Sydney. He has a keen interest in mathematical modeling and machine learning in the financial markets. He has previously lectured computer simulation at Oxford University and lead strategic research projects for Rolls-Royce Plc.

Tom is very active in the quantitative trading community, running workshops for Quantopian, teaching people quantitative analysis techniques, and organizing algorithmic trading meetup groups such as Cybertraders Syd.

Disclaimer
Quantopian provides this presentation to help people write trading algorithms - it is not intended to provide investment advice.

More specifically, the material is provided for informational purposes only and does not constitute an offer to sell, a solicitation to buy, or a recommendation or endorsement for any security or strategy, nor does it constitute an offer to provide investment advisory or other services by Quantopian.

In addition, the content neither constitutes investment advice nor offers any opinion with respect to the suitability of any security or any specific investment. Quantopian makes no guarantees as to accuracy or completeness of the views expressed in the website. The views are subject to change, and may have become unreliable for various reasons, including changes in market conditions or economic circumstances.
Рекомендации по теме
Комментарии
Автор

I'm immersed in this. I read a book with a similar theme, and I was completely immersed. "The Art of Saying No: Mastering Boundaries for a Fulfilling Life" by Samuel Dawn

Bill
Автор

I think neural networks start to truly outperform SVM’s as the decision boundary becomes more and more nonlinear. Kernel tricks would have to become more and more complex.

scalbylasusjim
Автор

And yet another improvement. You regularize a state data by substracting mean and dividing by standard deviation. It is a good thing to regularize but a state data is made of 141 values, including ohlc prices for M5, H1, D1 bars and different indicators ranging from -10 to 100. I do not think you can merge quantities of products with prices for products and then regularize, as they say, altogether.

Otvazhnii
Автор

Yet another improvement. You cannot rely on one attempt of training. Initial weights are random. You have to implement a multiprocessing logic and run 10 attempts at a time and print the 10 pictures of profit and loss curves at the end.

Otvazhnii
Автор

I wonder how he smoothes the data - perhaps "now" timestamp was already including partial info of the next data point. If it was smoothly only backwards then the next timestamp at exit might be completely off than the real exit price.

michelletadmor
Автор

30:14 you are updating state and applying the action. When we chose an action, first we need to apply than we need to update the state and get the reward. Let's say current price is 100.20. When agent decides to buy, it's has to buy from the price 100.20 (excluding spread/slippage and commission). In your example, it's buying with the next price. Am I wrong?

AtillaYurtseven
Автор

I come up with a number of improvements to the code. Firstly, the epsilon calculation runs out to zero after trade 5, while 99% of random numbers fall between 0, 1 and 0, 9. So no exploration after trade 5. Secondly, H1 and D1 bars are made from M5 bars by choosing only the 2 left and right M5 bars. This is correct for open and close prices, but not for high and low prices which move very noisily during every hour and still more so during the day. Thirdly, the way the code is built, it may take well over a month to run 11500 games (trades) as is indicated in your code. By converting the pandas data to numpy data and by then building a numpy array of states before training, you can speed up the code literally 10, 000 times. And finally, the Apple stock goes bluntly up at some point so your strategy which drops exploration after trade 5 and starts learning the replay memory of the last trades, does it not fit so nicely to the growing trend of the stock?

Otvazhnii
Автор

SVM has some type of geometry element which responds well to any freely traded market

blackprinze
Автор

Just as with any other AI algorithm, you need to clean your data before you give it to your reinforcement learner. But you can make a neural net that cleans that data for you, with relative succes. Noise is also an issue in other domains, not just finance. Ofcourse you are creating a feedback loop. When you buy/sell with succes your competitors will adapt, and so the problem shifts to a more difficult state...adding overal noise (randomness) to the system.

chrisminnoy
Автор

On 19:53 you mentions the Right Regime - 'does it end up choosing some training process ' It's easy to understand what is Mean Reversion process, but what does it mean 'training process' ?

AlexeyMatushevsky
Автор

3:32 guys please upload his talk too !

harendrasingh_
Автор

does this really work for stock trading? any trackrecord to check for last 5 years?

williamdad
Автор

but the technical indicators are from past prices also. isnt it better for the nn/rl interpret the prices themselves?

alrey
Автор

I am doing deep learning but now I'm thinking of integrating it with a reinforcement learning as an ensemble on the outside.
(with a money management system on the side)
Is there Anyone in California interested in my project?

alute
Автор

Why can't we give reward in every time step?

gogae
Автор

"LSTMs, they're somewhat new" they've been around since 1997 lol

norabelrose
Автор

Thank you sir for good explanation!
Please help me to solve this error - ImportError: cannot import name 'sgd' from 'keras.optimizers' am not able to fix this error and if anyone to fix this error please help me

joysahoo
Автор

BTC for $75K by end of this year& Control
of The Currency is already Decentralised And now the China disruption would simply
Decentralise the Mining setup for the better

monanica
Автор

95% of the trades are made by big money they can hire and build the most advance systems and programmers .. and they fail .
so it probably will not .. will not ever work ..

henrifritsmaarseveen
Автор

Siri was triggered at 1:40, hahaha. Time to rethink about ML?

oliverli