Due to recent updates, all users are required to create an Altair One account to login to the RapidMiner community. Click the Register button to create your account using the same email that you have previously used to login to the RapidMiner community. This will ensure that any previously created content will be synced to your Altair One account. Once you login, you will be asked to provide a username that identifies you to other Community users. Email us at Community with questions.
Generating Label Data for Financial Time Series
I'm working on creating a number of financial time series data with directional BUY HOLD SELL labels for Supervised learning. Can anyone suggest a method of generating the signals/labels for my training data other than to visually chart the data and find the tops and bottoms of the charts and manually code them in. My streams are not trying to predict say - the closing price of tomorrow - but to generate BUY HOLD SELL signals only. There must be a way to generate these training data signals automatically somehow... Any suggestions ?
0
Answers
Best regards,
Marius
I thought someone out there might know of some tools avail for this type of work. I actually own software called Trading Solutions (I would never recommend this software - very expensive and very buggy!!) and one of the few clever things it provides is the ability to provide a time series dataset and thenTrading Solutions will generate what they call an optimized signal including all the BUY HOLD SELL recommendations used for modeling / comparing out of sample data signals with It does a very nice jobs doing with - automatically finding all the peaks and valleys within the data. . Unfortunately, when exporting the Optimized Signal to Excel/CVS, it does not include/maintain these signals in the data making the software virtually useless to me..
I guess I just need graph the data first, find the corresponding date time for the peaks/valleys and manually generate my labels for training. I very slow and tedious process when dealing with large datasets (i.e. 5 - 10 years)
D