http://www.capitalico.com http://alpaca.ai What Technical Traders - - PowerPoint PPT Presentation

http capitalico com
SMART_READER_LITE
LIVE PREVIEW

http://www.capitalico.com http://alpaca.ai What Technical Traders - - PowerPoint PPT Presentation

1 Chart Pattern Matching in Financial Trading Using RNN Make you trade ideas into AI. Hitoshi Harada Start free. On mobile. CTO hitoshi@alpacadb.com http://www.capitalico.com http://alpaca.ai What Technical Traders Are Looking For 2 Entry


slide-1
SLIDE 1

1

Make you trade ideas into AI. Start free. On mobile.

Hitoshi Harada CTO hitoshi@alpacadb.com http://alpaca.ai

Chart Pattern Matching in Financial Trading Using RNN

http://www.capitalico.com

slide-2
SLIDE 2

Entry Point

What Technical Traders Are Looking For

2

slide-3
SLIDE 3

Diversity Of The Pattern - All Downtrend

3

slide-4
SLIDE 4
  • Fuzzy Pattern Recognition for everyone
  • Generalization (no hand crafted features)
  • Multiple time series (OHLC price + indicators)
  • Time scale, value scale, distortion

James N.K. Liu *, Raymond W.M. Kwong : Automatic extraction and identification of chart patterns towards financial forecast, 2006

Problem And Needs - Fuzzy Pattern Recognition

Zhe Zhang, Jian Jiang, Xiaoyan Liu, Ricky Lau, Huaiqing Wang, Rui Zhang: A Real Time Hybrid Pattern Matching Scheme for Stock Time Series, 2010

4

slide-5
SLIDE 5

How To Solve The Problem?

5

SPEECH RECOGNITION WITH DEEP RECURRENT NEURAL NETWORKS, Hinton, et al. 2013 Capitalico

“ah” “p” “down trend”

slide-6
SLIDE 6

Interactive Training Data Collection & Training

6

slide-7
SLIDE 7
  • Train by what you see & judge
  • No programming nor

conditional setting, but purely from charts like traders do


  • Multi-dimensional input
  • Not only the single time-

series data of price movement but also various indicators altogether

Our Approach - Fuzzy Pattern Recognition without Programming

7

slide-8
SLIDE 8
  • Network
  • Input:
  • N-dim Fully Connected Layer
  • LSTM Layer x 2 or 4 ( x250 units )
  • Fully Connected Layer ( x250 units )
  • Dropout
  • Sigmoid
  • Output:
  • 1-dim confidence level

  • Training
  • Align with fixed number of candles
  • Mean squared error for loss
  • AdaDelta for optimizer
  • BPTT through aligned length
  • Data
  • 1k+ samples collected by experts
  • about hundred instances for each strategy

Input LSTM LSTM Fully Connected Output Fully Connected

Time

Experiments Deep Learning Based Approach

8

Sigmoid Input LSTM LSTM Fully Connected Output Fully Connected Sigmoid Input LSTM LSTM Fully Connected Output Fully Connected Sigmoid

slide-9
SLIDE 9

x-axis: time (1.0=entry point) blue: training data / orange: testing data

Experiments Fitting Reasonably

y-axis: confidence

9

slide-10
SLIDE 10

Experiments Framework

10

slide-11
SLIDE 11

Dropout

  • Dropout vs # of training samples
  • Bigger Mini-Batches by looping samples
  • Made it Adaptive depending on importance

11

dropout enabled (x: iteration count, y: loss) dropout w/ bigger mini-batches (x: iteration count, y: loss)

slide-12
SLIDE 12

Forget Gate Bias (Learning To Forget: Continual Prediction With Lstm, Felix Et Al.)

12

slide-13
SLIDE 13

Trial And Error To Speed Up Training

13

  • Dynamic Dropout
  • Dynamic Batchsize
  • Multi-GPU Training
  • Other Frameworks like Keras
  • GRU
  • IRNN
  • Lot more…
slide-14
SLIDE 14

14

  • Previous studies have limitations to diffjculty of feature crafting.
  • LSTM based deep neural network fits well with individual patterns.
  • LSTM-variant doesn’t make much difgerence, but forget-gate bias,

normalization, preprocessing, and modeling etc. matter

  • Build better base model by pre-training
  • Reinforcement Learning using profit and risk preference
  • Visualize and rationalize LSTM decision making
  • Generative Model

Conclusion & Future Work

14

slide-15
SLIDE 15

QUESTIONS AND ANSWERS

http://alpaca.ai / info@alpacadb.com

http://www.capitalico.com

Make you trade ideas into AI. Start free. On mobile.

slide-16
SLIDE 16
  • Ken-ichi Kainijo and Tetsuji Tanigawa:


Stock Price Pattern Recognition - A Recurrent Neural Network Approach -, 1990

  • S Hochreiter, J Schmidhuber:


Long short-term memory, 1997

  • FA Gers, J Schmidhuber, F Cummins:


Learning to forget: Continual prediction with LSTM, 2000

  • James N.K. Liu *, Raymond W.M. Kwong:


Automatic extraction and identification of chart patterns towards financial forecast, 2006

  • X Guo, X Liang, X Li:


A stock pattern recognition algorithm based on neural networks, 2007

  • Z Zhang, J Jiang, X Liu, R Lau, H Wang:


A real time hybrid pattern matching scheme for stock time series, 2010

  • A Graves, A Mohamed, G Hinton:


Speech recognition with deep recurrent neural networks, 2013

  • A Graves, N Jaitly:


A Mohamed, Hybrid speech recognition with deep bidirectional LSTM, 2013

  • Tara N. Sainath, Oriol Vinyals, Andrew Senior, Has¸im Sak:


CONVOLUTIONAL, LONG SHORT-TERM MEMORY, FULLY CONNECTED DEEP NEURAL NETWORKS

References

16

slide-17
SLIDE 17

Need For Gpu And Distributed Computation

17

  • Model Training
  • Takes around 10 minutes on a single GPU core
  • Requires 2GB of GPU RAM
  • Backtesting
  • Calculate various metrics over the historical data
  • Livetesting
  • Thousands of models need to monitor live candles and update the state of LSTM
slide-18
SLIDE 18

Need For Distributed Computation

18

DB Postgresql Redis etcd Load Balancer WEB Flask

tesla k80

WORKER Live Market Watch Market Data Historical Real time Queue Celery Trading

Algos = ~10MB x1-10K