竞赛笔记

出题方将制作过程加密。根据前面帖子的专利文献里推测出来的字段含义,供各位同道参考
在这里插入图片描述
在这里插入图片描述
** 11楼,20楼有最新的推测

Stock Technical Analysis – Python Tutorial
http://www.andrewshamlet.net/

Stock Prediction with ML: Feature Engineering
https://alphascientist.com/feature_engineering.html

Alex_Scut(35th in this Competition)
I just train a single LSTM model with attention and mask for every stock.Features that i used are just the original data with shifting.

Trainer Brock
(8th in this Competition)
The LSTM was one of the ideas I had (one model for all the stocks) but I thought it was too uncertain (and I have no experience training LSTMs) so I tried other things instead. Too bad we cannot team up, I use just LightGBM, so the model variety would have been very nice for blending!
markmipt(19th in this Competition)

Hi! It is my first competition at Kaggle. I did not used qianqian’s kernel and any “lag” features, but have managed to get 0.76 score. I want say that 0.7 can be reached just by using market data + a few basic features which can be googled without few minutes + careful validation of catboost/lightgbm training. The rest scores came from some personal ideas, some of them were based on my knowledge (which is not very deep actually) of finance. My advice is you should clearly realize that this competition is mostly not about hyperparameters tuning, stacking or precise prediction of values which are usually in kaggle competitions, but about proper using of predicted values

https://www.kaggle.com/c/two-sigma-financial-news/discussion/75585
I have a score of 0.71515 (currently 69th place) - I’m not using news or any lag features – just these with a single instance of LGBM:

(‘returnsOpenPrevMktres1’, 72)
(‘returnsClosePrevMktres1’, 78)
(‘returnsClosePrevMktres10’, 112)
(‘returnsClosePrevRaw1’, 114)
(‘returnsClosePrevRaw10’, 164)
(‘returnsOpenPrevRaw1’, 200)
(‘returnsOpenPrevMktres10’, 207)
(‘returnsOpenPrevRaw10’, 409)
(‘returnsClosePrevRaw1avg’, 568) (‘returnsOpenPrevMktres10avg’, 617)
(‘returnsClosePrevMktres1avg’, 649) (‘closeopenratioavg’, 650)
(‘returnsClosePrevRaw10avg’, 652) (‘returnsOpenPrevMktres1avg’, 661)
(‘returnsClosePrevMktres10avg’, 673) (‘returnsOpenPrevRaw1avg’, 673)
(‘returnsOpenPrevRaw10_avg’, 721)

For me, the big win came from normalising the submission scores (and applying a slight bias)…
John Wakefield(85th in this Competition)
@jonnydossantos - correct, my averages are calculated by grouping on each training day.
@aditya1702 - yes, I’m simply subtracting the mean for each prediction day.

Don’t use public kernels without refactoring. Most of them contain a lot of issues in data preparation. It doesn’t mean they can not win (due to the core of this competition - market data) and we will see a lot of shakeups.
Code carefully (your code should be fast and memory efficient). It helps with making experiments.
Use sample data to test the code before real training.
Train/Validation sets are really important - try to think how you validate your model.
Try more features and their combinations.
Read more about market data and financial approaches. E.g. in this competition we are predicting ‘sharpe ratio’.
Check kernels for new ideas of features. Sometimes you may find them in comments but not in the code.
Public leaderboard is not a final target. 😃
UPDATES:
Answers about current score 0.9:

we don’t use fine tuning during test phase
we use lag features
there are none news features at the moment
parameters tuning - random search on cross-validation with manual selection
UPDATES 2:

it’s not easy to predict 2007-2008, but it’s possible to train model based on this period (so, there is a signal inside)

Bollinger Band, MACD, and EWMA

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值