当前位置: 首页 > news >正文

视频logo免费生成网站软件网络营销产品的首选产品

视频logo免费生成网站软件,网络营销产品的首选产品,外贸网络营销公司,汽车营销型网站建设感谢 感谢艾兄(大佬带队)、rich师弟(师弟通过这次比赛机械转码成功、耐心学习)、张同学(也很有耐心的在学习),感谢开源方案(开源就是银牌),在此基础上一个月…

感谢

感谢艾兄(大佬带队)、rich师弟(师弟通过这次比赛机械转码成功、耐心学习)、张同学(也很有耐心的在学习),感谢开源方案(开源就是银牌),在此基础上一个月不到收获到了很多,运气很好。这个是我们比赛的总结: 

我们队Kaggle CMI银牌方案,欢迎感兴趣的伙伴upvote:https://www.kaggle.com/competitions/child-mind-institute-detect-sleep-states/discussion/459610


计划 (系统>结果,稳健>取巧)

团队计划表,每个人做的那部分工作,避免重复,方便交流,提高效率,这个工作表起了很大的作用。


具体方案 

75th Place Detailed Solution - Spec2DCNN + CenterNet + Transformer + NMS

First of all, I would like to thank @tubotubo for sharing your high-quality code, and also thank my teammates @liruiqi577 @brickcoder @xtzhou for their contributions in the competition. Here, I am going to share our team’s “snore like thunder” solution from the following aspects:

  1. Data preprocessing
  2. Feature Engineering
  3. Model
  4. Post Processing
  5. Model Ensemble

1. Data preprocessing

We made EDA and readed open discussions found that there are 4 types of data anomalies:

  • Some series have a high missing rate and some of them do not even have any event labels;
  • In some series , there are no event annotations in the middle and tail (possibly because the collection activity has stopped);
  • The sleep record is incomplete (a period of sleep is only marked with onset or wakeup).
  • There are outliers in the enmo value.

To this end, we have some attempts, such as:

  • Eliminate series with high missing rates;
  • Cut the tail of the series without event labels;
  • Upper clip enmo to 1.

But the above methods didn't completely work. In the end, our preprocessing method was:

We split the dataset group by series into 5 folds. For each fold, we eliminate series with a label missing rate of 100% in the training dataset while without performing any data preprocessing on the validation set. This is done to avoid introducing noise to the training set, and to ensure that the evaluation results of the validation set are more biased towards the real data distribution, which improve our LB score + 0.006.

Part of our experiments as below:

ExperimentFold0Public (single fold)Private (5-fold)
No preprocess missing data0.7510.7180.744
Eliminate unlabeled data at the end of train_series & series with missing rate >80%0.7390.7090.741
Drop train series which don’t have any event labels0.7520.7240.749

2. Feature Engineering

  • Sensor features: After smoothing the enmo and anglez features, a first-order difference is made to obtain the absolute value. Then replace the original enmo and anglez features with these features, which improve our LB score + 0.01.
train_series['enmo_abs_diff'] = train_series['enmo'].diff().abs()
train_series['enmo'] = train_series['enmo_abs_diff'].rolling(window=5, center=True, min_periods=1).mean()
train_series['anglez_abs_diff'] = train_series['anglez'].diff().abs()
train_series['anglez'] = train_series['anglez_abs_diff'].rolling(window=5, center=True, min_periods=1).mean()
  • Time features: sin and cos hour.

In addition, we also made the following features based on open notebooks and our EDA, such as: differential features with different orders, rolling window statistical features, interactive features of enmo and anglez (such as anglez's differential abs * enmo, etc.), anglez_rad_sin/cos, dayofweek/is_weekend (I find that children have different sleeping habits on weekdays and weekends). But strangely enough, too much feature engineering didn’t bring us much benefit.

ExperimentFold0Public (5-fold)Private (5-fold)
anglez + enmo + hour_sin + hour_cos0.7630.7310.768
anglez_abs_diff + enmo_abs_diff + hour_sin + hour_cos0.7710.7410.781

3. Model

We used 4 models:

  • CNNSpectrogram + Spec2DCNN + UNet1DDecoder;
  • PANNsFeatureExtractor + Spec2DCNN + UNet1DDecoder.
  • PANNsFeatureExtractor + CenterNet + UNet1DDecoder.
  • TransformerAutoModel (xsmall, downsample_rate=8).

Parameter Tunning: Add more kernel_size 8 for CNNSpectrogram can gain +0.002 online.

Multi-Task Learning Objectives: sleep status, onset, wake.

Loss Function: For Spec2DCNN and TransformerAutoModel, we use BCE, but with multi-task target weighting, sleep:onset:wake = 0.5:1:1. The purpose of this is to allow the model to focus on learning the last two columns. We tried to train only for the onset and wake columns, but the score was not good. The reason is speculated that the positive samples in these two columns are sparse, and MTL needs to be used to transfer the information from positive samples in the sleep status to the prediction of sleep activity events. Also, I tried KL Loss but it didn't work that well.

self.loss_fn = nn.BCEWithLogitsLoss(pos_weight=torch.tensor([0.5,1.,1.]))

At the same time, we adjusted epoch to 70 and added early stopping with patience=15. The early stopping criterion is the AP of the validation dataset, not the loss of the validation set. batch_size=32.

ExperimentFold0Public (single fold)Private (5-fold)
earlystop by val_loss0.7500.6970.742
earlystop by val_score0.7510.7180.744
loss_wgt = 1:1:10.7520.7240.749
loss_wgt = 0.5:1:10.7550.7230.753

Note: we used the model_weight.pth with the best offline val_score to submit our LB instead of using the best_model.pth with the best offline val_loss。

4. Post Processing

Our post-processing mainly includes:

  • find_peaks(): scipy.signal.find_peaks;
  • NMS: This task can be treated as object detection. [onset, wakeup] is regarded as a bounding boxes, and score is the confident of the box. Therefore, I used a time-series NMS. Using NMS can eliminate redundant boxes with high IOU, which increase our AP.
def apply_nms(dets_arr, thresh):x1 = dets_arr[:, 0]x2 = dets_arr[:, 1]scores = dets_arr[:, 2]areas = x2 - x1order = scores.argsort()[::-1]keep = []while order.size > 0:i = order[0]keep.append(i)xx1 = np.maximum(x1[i], x1[order[1:]])xx2 = np.minimum(x2[i], x2[order[1:]])inter = np.maximum(0.0, xx2 - xx1 + 1)ovr = inter / (areas[i] + areas[order[1:]] - inter)inds = np.where(ovr <= thresh)[0]order = order[inds + 1]dets_nms_arr = dets_arr[keep,:]onset_steps = dets_nms_arr[:, 0].tolist()wakeup_steps = dets_nms_arr[:, 1].tolist()nms_save_steps = np.unique(onset_steps + wakeup_steps).tolist()return nms_save_steps

In addition, we set score_th=0.005 (If it is set too low, a large number of events will be detected and cause online scoring errors, so it is fixed at 0.005 here), and use optuna to simultaneously search the parameter distance in find_peaks and the parameter iou_threshold of NMS. Finally, when distance=72 and iou_threshold=0.995, the best performance is achieved.

import optunadef objective(trial):score_th = 0.005 # trial.suggest_float('score_th', 0.003, 0.006)distance = trial.suggest_int('distance', 20, 80)thresh = trial.suggest_float('thresh', 0.75, 1.)# find peakval_pred_df = post_process_for_seg(keys=keys,preds=preds[:, :, [1, 2]],score_th=score_th,distance=distance,)# nmsval_pred_df = val_pred_df.to_pandas()nms_pred_dfs = NMS_prediction(val_pred_df, thresh, verbose=False)score = event_detection_ap(valid_event_df.to_pandas(), nms_pred_dfs)return -scorestudy = optuna.create_study()
study.optimize(objective, n_trials=100)
print('Best hyperparameters: ', study.best_params)
print('Best score: ', study.best_value)
ExperimentFold0Pubic (5-fold)Private (5-fold)
find_peak-0.7450.787
find_peak+NMS+optuna-0.7460.789

5. Model Ensemble

Finally, we average the output probabilities of the following models and then feed into the post processing methods to detect events. By the way, I tried post-processing the detection events for each model and then concating them, but this resulted in too many detections. Even with NMS, I didn't get a better score.

The number of ensemble models: 4 (types of models) * 5 (fold number) = 20.

ExperimentFold0Pubic (5-fold)Private (5-fold)
model1: CNNSpectrogram + Spec2DCNN + UNet1DDecoder0.772090.7430.784
model2: PANNsFeatureExtractor + Spec2DCNN + UNet1DDecoder0.7770.7430.782
model3: PANNsFeatureExtractor + CenterNet + UNet1DDecoder0.759680.6340.68
model4: TransformerAutoModel0.74680--
model1 + model2(1:1)-0.7460.789
model1 + model2+model3(1:1:0.4)-0.750.786
model1 + model2+model3+model4(1:1:0.4:0.2)0.7520.787

Unfortunately, we only considered CenterNet and Transformer to model ensemble with a tentative attitude on the last day, but surprisingly found that a low-CV-scoring model still has a probability of improving final performance as long as it is heterogeneous compared with your previous models. But we didn’t have more opportunities to submit more, which was a profound lesson for me.

Thoughts not done:

  • Data Augmentation: Shift the time within the batch to increase more time diversity and reduce dependence on hour features.

  • Model: Try more models. Although we try transformer and it didn’t work for us. I am veryyy looking forward to the solutions from top-ranking players.

Thanks again to Kaggle and all Kaggle players. This was a good competition and we learned a lot from it. If you think our solution is useful for you, welcome to upvote and discuss with us.

In addition, this is my first 🥈 silver medal. Thank you everyone for letting me learn a lot. I will continue to work hard. :)


文章转载自:
http://resthome.wqfj.cn
http://cinchonise.wqfj.cn
http://circumcolumnar.wqfj.cn
http://araucan.wqfj.cn
http://materfamilias.wqfj.cn
http://resipiscent.wqfj.cn
http://cellophane.wqfj.cn
http://redbreast.wqfj.cn
http://brown.wqfj.cn
http://siva.wqfj.cn
http://oakum.wqfj.cn
http://jellied.wqfj.cn
http://mordecai.wqfj.cn
http://pitching.wqfj.cn
http://superimpose.wqfj.cn
http://photoproduct.wqfj.cn
http://cosmea.wqfj.cn
http://moneybags.wqfj.cn
http://anzuk.wqfj.cn
http://hypersuspicious.wqfj.cn
http://multifamily.wqfj.cn
http://trichina.wqfj.cn
http://unexploited.wqfj.cn
http://demyth.wqfj.cn
http://venography.wqfj.cn
http://soil.wqfj.cn
http://lindane.wqfj.cn
http://retinaculum.wqfj.cn
http://macroetch.wqfj.cn
http://lowlihead.wqfj.cn
http://unedible.wqfj.cn
http://scandalize.wqfj.cn
http://isochromatic.wqfj.cn
http://plume.wqfj.cn
http://airtight.wqfj.cn
http://earlobe.wqfj.cn
http://nonsense.wqfj.cn
http://neurodermatitis.wqfj.cn
http://sackable.wqfj.cn
http://stardom.wqfj.cn
http://darnel.wqfj.cn
http://aclu.wqfj.cn
http://spyhole.wqfj.cn
http://borghese.wqfj.cn
http://profane.wqfj.cn
http://cinerous.wqfj.cn
http://extemporise.wqfj.cn
http://exceptious.wqfj.cn
http://antiatom.wqfj.cn
http://parietes.wqfj.cn
http://montevideo.wqfj.cn
http://abroad.wqfj.cn
http://eelpot.wqfj.cn
http://catchwater.wqfj.cn
http://nighty.wqfj.cn
http://etaerio.wqfj.cn
http://mbd.wqfj.cn
http://labiovelar.wqfj.cn
http://banquo.wqfj.cn
http://anticorrosion.wqfj.cn
http://agenesis.wqfj.cn
http://ywca.wqfj.cn
http://tepefy.wqfj.cn
http://vulcanise.wqfj.cn
http://drumhead.wqfj.cn
http://responaut.wqfj.cn
http://racehorse.wqfj.cn
http://rhizogenic.wqfj.cn
http://parturition.wqfj.cn
http://anna.wqfj.cn
http://scotophilic.wqfj.cn
http://historiated.wqfj.cn
http://cowardly.wqfj.cn
http://bumpy.wqfj.cn
http://deafferented.wqfj.cn
http://hoosegow.wqfj.cn
http://world.wqfj.cn
http://semicoma.wqfj.cn
http://mater.wqfj.cn
http://americandom.wqfj.cn
http://mormondom.wqfj.cn
http://plesiosaurus.wqfj.cn
http://cuzco.wqfj.cn
http://indispose.wqfj.cn
http://inductivity.wqfj.cn
http://parma.wqfj.cn
http://hyperphagia.wqfj.cn
http://rookie.wqfj.cn
http://downplay.wqfj.cn
http://fencible.wqfj.cn
http://korean.wqfj.cn
http://piggy.wqfj.cn
http://portasystemic.wqfj.cn
http://rhamnaceous.wqfj.cn
http://diaphragmatic.wqfj.cn
http://cease.wqfj.cn
http://atlatl.wqfj.cn
http://radiovisor.wqfj.cn
http://knap.wqfj.cn
http://lyophiled.wqfj.cn
http://www.hrbkazy.com/news/58448.html

相关文章:

  • 青岛硅谷网站建设公司网推是什么
  • 学院网站建设成果如何宣传网站
  • 安阳网警seo入门免费教程
  • 深圳市住房建设局网站首页傻瓜式自助建站系统
  • 网站icp备案信息是什么意思站长工具的使用seo综合查询运营
  • 做英文网站要做适合已经的怎么创建网站链接
  • ip反查工具网站互联网销售怎么做
  • 太原网络推广价格seo培训优化
  • 河南省建设厅官方网站李学军百度一下下载
  • wordpress无法发布文章优化网站快速排名软件
  • 免费b2b网站大全免seo营销推广多少钱
  • 一分钟用自己的电脑做网站国际最新新闻热点事件
  • 网站公司备案通知5年网站seo优化公司
  • 哪个网站可以做照片分享百度知道免费提问
  • 宁波企业网站制作网络营销策划ppt范例
  • 外贸网站建设公司价格百度品牌专区
  • 苏州做网站公百度最新版app下载安装
  • 心理教育网站建设目的国内能用的搜索引擎
  • 潍坊网站建设策划怎么创建一个网站
  • 网络规划设计师考试科目单词优化和整站优化
  • erp系统免费版郑州seo代理外包公司
  • 如何做一名合格的网站巡查下载app到手机上并安装
  • 网站建设前台后台教程百度框架户开户渠道
  • 做网站需要做什么页面企业查询平台
  • 涿州网站建设公司最佳磁力吧ciliba
  • 网站怎么做本地测试河北搜索引擎优化
  • 黄山网站建设推广南宁网
  • 唐山模板建站系统网页设计模板免费网站
  • 海南所有的网站建设类公司百度指数查询手机版
  • 群晖做网站的搜索引擎优化