I have been working on Neural Networks for various purposes lately. I have had great success in digit recognition, XOR, and various other easy/hello world\'ish applications.
I think that you've got the basic idea: a "sliding window" approach where a network is trained to use the last k
values of a series (Tn-k ... Tn-1) to predict the current value (Tn).
There are a lot of ways you can do this, however. For example:
Often people end up figuring out the best way to learn from their particular data by trial and error.
There are a fair number of publicly-accessible papers out there about this stuff. Start with these, and look at their citations and papers that cite them via Google Scholar, and you should have plenty to read:
There is a kind of neural networks named recurrent neural networks (RNNs. One advantage of using these models is you do not have to define an sliding window for the input examples. A variant of RNNs known as Long-Short Term Memory (LSTM) can potentially take into account many instances in the previous time stamps and a "forget gate" is used to allow or disallow remembering the previous results from the previous time stamps.
Technically this is the same as your digit recognition - it is recognizing something and returning what it was...
Well - now your inputs are the previous steps (T-5 ... T-1) - and your output or outputs are the predicted steps (T0, T1...).
The mechanics in the ANN itself are the same - you will have to teach every layer for feature detection, correcting its reconstruction of the thing, so that it looks like what is actually going to happen.
(some more info about what do I mean: tech talk )