广东工业智造大数据创新大赛

Game is over yesterday
1433065-20181107094252372-108801205.png

1433065-20181107094353978-1227462107.png

1433065-20181107094425909-1954186696.png

give up

1433065-20181021192713352-1668173406.png

competition questions and data

1433065-20181003195547701-350677672.png

guangdong_defect_instruction_20180916.xlsx
guangdong_round1_submit_sample_20180916.csv
guangdong_round1_test_a_20180916.zip
guangdong_round1_train1_20180903.zip

1433065-20181003195244417-1040353517.png

Solutions

1433065-20181003195651783-12584561.png
Using Kaggle cat and dog classification code,
even using there depth deeping networks ResNet50,Inception V3,
Xception to extract image features,
and using neural networkf DNN classification,
verification set shows over-fitting.

Kaggle cat and dog classification

ResNet50

resnetv2-50

tensorflow.Keras use Resnet50 to realize CatDogDistinguish

pretriained-models

inception v4.py

TIANCHI_MODEL_V4

比赛思路

Direct image classificaton,select a network to extract features,followed by a fully connection layer classification,plus regularization to reduce over-fitting.Then let go of all levels of training.The final accuracy is about 0.92,in fact,as long as the default parameters do not depart from the spectrum on the line,adjusting the parameters does not have much impact on the results.

select a network to extract features

competition solution 2:Standard DenseNet,softmax12 classification,
made data enhancement;
tried to tune learning_rate,
batch_size,num_layers

DenseNet

SHARED BASELINE

WIN10+anaconda+Pytorch 0.4+CUDA 8.0
pytorch gpu

activate Pytorch

pip install http://download.pytorch.org/whl/cu80/torch-0.4.0-cp36-cp36m-win_amd64.whl

pip install torchvision

test:
import torch
print(torch.__version__)
torch.cuda.is_avaliable()

1433065-20181004165631864-710559665.png

ResNet

Residual Network

残差网络

Inception

v1——v2——v3——v4

RexNeXt

Xception

Inception resnet v2

inception

开始,盗梦空间

Xception

Extreme Inception

verify CUDA

nvcc -V

1433065-20181005184609829-197847957.png

CUDNN

1433065-20181009132730839-1531559333.png

Ground2

The first columnThe second column
1用TensorFlow训练一个物体检测器
2Object detection
3single shot multibox detector
The First ColumnThe Second Column
reregular expression
正则表达式
anchorSPP(spatial pyramid pooling)
spatial pyramid pooling空间金字塔池化层
VGG224*224
SPPspatial pyramid pooling
mlpmultilayer perceptron 多层感知机
ffmField-aware Factorization Machine
CTRclick-through rate 点击率
CVRconversion rate 转化率
conversion变换
SGDStochastic Gradient Descent
ICPRinternational conference on pattern recognition
Stochastic Gradient Descent随机梯度下降
GD algorithm's two shortcomingsGD
click rate prediction algorithm点击率预估算法
CVRconversion rate
点击后的转化率
LRLogistic Regression
GBDTGradient Boosting Decision Tree
FMFactorization Machine
factorization因式分解
时间复杂度comprehensive
Bagging套袋法
Adaboostingthe difference between bagging and boosting
baggingBagging is a reduction of variance
boostingBoosting is a reduction of bias
bias and varianceN-fold Cross Validation
N-fold Cross ValidationN-折交叉验证,将训练集分成N份,比如分成3份
FFMField-aware Factorization Machine
TPTrue Positive
FPFalse positive
TNTrue negative
FNFalse negative
TPR$\frac{TP}{TP+FN}
FPRfalse positive rate
recallrecall the number of positive sample
precision\(P= \fac{TP}{TP+FP}\)

Precision and recall

在机器学习模型评估中,准确率和召回率是一对相互制约的性能度量指标。对于一个二分类问题,样本本身有正有负,而我们的学习器的判断也是有正有负。由于数据和算法等因素,我们的学习器的判断的结果往往不会和测试样本的真实答案完全吻合,这时就需要度量指标来表征该学习器的性能,常见的是准确率与召回率。
准确率和召回率定义非常明确,但是由于名称比较费解,尤其是召回率,致使很多人将这两个概念混淆。在周志华老师的西瓜书里面,将这两个词分别翻译为查准率(precision)和查全率(recall),这样可以顾名思义,了解到这两个词的意思,查准率就是对于所有机器判定为正的里面,有多大的比例是真的正样本,写成公式就是P = T P/( T P + F P ) 其中TP,FP分别表示true positive和false positive,即所谓真阳性和假阳性,而对于查全率,顾名思义,就是实际的正样本中,有多大比例被检出了,写成公式就是: R = T P /(T P + F N )其中TP同前,FN表示false negative,也就是假阴性,(注意在真/假 阴/阳性中,阴阳性是指的分类器的判断结果是阴性还是阳性,而真假指代的是是否和真实一样)周老师的翻译虽然很巧妙,但是学界常常还是用准确率和召回率来称呼这两个概念,而且这两个概念的英文本意也是准确和召回,周老师算是意译。那么为了方便记忆和分辨,举一个场景来解释这件事:
假设有一家汽车公司,突然发现某一批次已经卖出去的车是有问题的,于是厂家给出了一种判断方法,告诉所有车主(假设所有车主都不知道自己的车的批次),让大家根据这种方法看一看自己的车子是否有问题,然后把问题车进行召回。但是由于判据过于简单,导致有些正常车的车主也发现自己车子有问题,而相反,也有些问题车的车主以为自己的车正常。这样,工厂召回来的车中,有TP,也有FP,而市面上没有召回的车子中,有TN,也有FN。好,那么现在,工厂希望计算一下,这一批次的问题车中,到底有多少真正被召回,那么就是R = TP/(TP+FN)。这里的FN就是那些未被召回的问题车。所以这个比率自然称为召回率。另外,制定这条判据的技术人员也想知道,这条判据的准确率有多高,那么他们就会计算一下在召回来的这些车里,有多少真的是问题车。所以P = TP/(TP+FP)。
通过以上场景,可以用经验的思维而非数学的定义,顾名(precision和recall)思义地辨析两者的区别和应用方式。

1433065-20181016143441664-695787502.png

1433065-20181016143516911-804751499.png

1433065-20181016143537284-952426951.png

1433065-20181016143634030-708656398.png

cross entropy

1433065-20181016184038702-319678661.png

the first columnthe second column
F1\(\fac{precision \times recall}{precision + recall}\)
JSONjavascript Object Notation,JS对象简谱
notation标记法
curve曲线,弧线
APAverage Precision
mAPmean Average Precision
confidence score置信度
ground truth label
Cross Entropycross Entropy
BoundingBox边界线
RectClass
TestEvaluator测试评估器
Three levels of image understandingclassification
Three levels of image understandingDetection
2-stageRegion-based
R-CNN1.Region Proposal
2.AlexNet
region proposal后选区域
gound truththe IoU of ground truth
SSDSingle Shot Detection
SPPspatial pyramid pooling
threshold0.5
RPNRegion of Interests
anchor boxanchor
bboxBounding Boxes
Ground Truth检测框
IoU(R∩G)/(R∪G)
Bounding-BoxedBounding-Box regression
k anchor boxes20k anchor boxes:60409

mask R-CNN

tensorflow_mask R-CNN

The First ColumnThe Second Column
ESSESS测量模型中的方差(信息)量
object mask物体掩膜
mask掩膜
SSDSingle Shot MultiBox Detector
ont-stagerefineDet
RefineDetSingle-Shot Refinement Neural Network for Object Dectection
RefineDettwo-step cascaded regression
deep_learning_object_detectionRefineDet
deep_learning_object_detection1433065-20181018170833744-1635381659.png
binary mask二值掩膜
bounding boxbounding box
非极大值抑制Non-Maximum Suppression
R-CNNRegion-based convolutional Neural Networks
Ground Truth目标窗口
Faster R-CNNInput Image
proposal Region
feature extraction
feature map
Non-Maximum SuppressionNMS
mean shiftalgorithm
common IoU0.3~0.5
R-FCNRegion-based Fully convolutional Networks
RoI poolingend-to-end
region proposalROI Pooling
R-FCNRegion-based Fully Convolutional Networks
ROIregions of interest
end-to-end training
The First ColumnThe Second Column
compat兼容性
COCO datasetobject instances 目标实例
PASCAL VOCpascal
JSONinfo_licenses_images_annotaions_categories
Object Instanceinfo_licenses_images_categories_annotations
annotation注释
PRNRegion Proposal Network
caffeconvolutional Architecture for Fast Feature Embedding
caffe快速嵌入卷积结构
convolutional architecture for fast feature embedding
train训练数据
val验证数据
valverify data(VAL) because of its tagging
RetinaNetICCV 2017 the best paper
RetinaNetone-stage
CEcross-entropy Error
FLFocal Loss焦点误差
Installing Detectronpytorch
labelme to cocoObject detection_data_interface
..parent directory
Absolut path绝对路径:从盘符开始的路径
Absolut pathC:\windows\system32
vcs.xmlDo you want to add the following file to Git?
background tasks后台任务
update indice更新索引
process finished
with exit code 1
说明程序出错
annotaion注释
external file changes
sync may be slow
外部文件更改
同步可能会很慢
append()append(2018),结果后面[....,2018]
Detectron's backboneResNext{50,101,152}、ResNet、FPN、VGG16
indexing索引
compliation
val2017evaluation 2017
()tuple datatype
{}dict
[]list,variable sequence
range()for in 范围
os.path.basename(imglist[i])Gets the name of the file under the corresponding path
imglist[i]corresponding path
list = []define Empty list
list.sort()list is Ascending order
cv2.imread()the return type
rect = {}[rectangle = key1 : value1,key2 : ……
backgroundNUM_CLASSES
Object Keypoint Annotationsannotation
len(boxes)if boxes[b][2]<boxes[b][0]
result["filename"] = filenameresult = {}

Adjusting parameter

the first columnthe second column
1test function parameter
2score(it means that confidence of the json)
31433065-20181021191756324-1432225137.png
FPNFeature Paremeter Network
Loss FunctionError of a sample
MSEmean squard error
mean squard error均方误差

转载于:https://www.cnblogs.com/hugeng007/p/9740702.html

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值