Toutiao signed more than 300 Zhihu big V’s in one breath this year, and JUST signed me, and it is to give money, the annual income is higher than ordinary white-collar workers. After signing all the content can not be sent zhihu. The quality of Zhihu has gone from bad to worse…
Zhang Liang, co-founder of Zhihu, said: “For at least the past year, I have been hoping for two people to leave Zhihu. One is the ‘evil dad’ and the other is the ‘Huo Master’.”
Zhihu mountain Cup Machine Learning Challenge
Let’s take a look back at the game.
Background introduction
Task description
Evaluation method
-
The five predicted hashtags were ranked in order of their predicted score.
-
Hashtags are not duplicated by default. When a topic tag is repeated, only the first occurrence is retained, and the subsequent tags are replaced. If there are less than five tags after deduplication, the remaining positions are -1 by default. -1 does not match any tag. Those with more than 5 hashtags will be ignored from the sixth place onwards.
-
Evaluation criteria:
Precision: If the predicted tag hits any one of the tagged tags, it is regarded as correct. The final accuracy is weighted by position for accuracy at each position. Math Precision = \sum_{pos \in \{1,2,3,4,5\}} \frac {Precision@pos} {log_{pos + 1}}
Recall: The predicted amount of coverage of the original labels in the Top 5 labels.
The final evaluation index was harmonic mean of Precision and Recall. That is:
The result of the match
a.com/competition/zh
ihu/leaderboard/
), and the final screening confirmation.
To verify the results, after the winning team submitted their method description and reproducible code and model data, the review team reviewed the winning team’s method and code one by one, randomly selected models submitted by some of the teams, and verified the results with another validation dataset. A total of 7 teams won prizes in this competition. The winners are as follows:
-
Koala’s team from Beijing University of Posts and Telecommunications;
-
YesOfCourse team from CAS, Google and Baidu;
-
NLPFakers from Microsoft and Peking University;
-
Gower Street & 81 Road from Wuhan University and University College London;
-
Ye team from Beijing University of Posts and Telecommunications;
-
Yin & Bird team from Zhengzhou Railway Bureau, Flush Company and Zhongshan University of Electronic Science and Technology.
Games excel
-
Init team: TextCNN + TextRNN + RCNN, shared Embedding learns jointly. In terms of model integration, Bagging method with multi-model and equal weight is used for integration. In data preprocessing, delete and Shuffle are used for data enhancement.
-
Koala team: FastText + TextCNN + TextRNN, used boosting idea to train neural network layer by layer, and weighted average bagging method was used among each network.
-
YesOfCourse team: TextCNN + LSTM/GRU + RCNN was used as the base model, and GBRank was used to fuse the outputs of multiple neural networks;
-
NLPFakers team: TextCNN + RNN + RCNN was used as the base model, and linear weighting was used for model integration; Attentional mechanism is used in neural network training.
-
Gower Street & 81 Road team: RNN neural network was used as the basic model, and the similarity of Query-TopicTitle was jointly trained with the neural network. Finally, Bagging with Ensemble Selection was adopted as the model integration strategy.
-
Ye team used TextCNN + BiGRU as the basic model and bagging with weight search as the model integration strategy.
-
Yin&Bird team: Using LSTM and Bayes methods as the base model and stacking methods for model integration.
In modeling the problem, all the teams translated the problem into a “text multiple classification” or “text label prediction” problem. During training, most teams use the Cross Entropy as a loss function. All teams applied the idea of integrated learning, using multiple models to complement each other to improve performance. At the same time, the players also carried out a lot of optimization for their understanding of the problem, and there were some very bright optimization methods.
Koala, the second place winner, used a layer-by-layer boosting method to improve the performance of a single neural network model in neural network training. According to its description, this optimization can improve the performance of multilayer neural networks by about 1.5 percentage points.
The YesOfCourse team, in third place, turned the Tag Precition process into a two-step Recall-Rarank problem; A large number of neural network models are used for recall, and the predicted score of neural network labels is taken as the characteristic input of GBRank, and the Pairwise method is used to optimize the sorting of labels, and the first 5 sorted labels are selected as the output of the model. From the descriptions submitted by the YesOfCourse team, we can see that the results obtained by using Recall + Rerank model are more than two thousandths better than those obtained by non-Linear NN Ensemble. At the same time, YesOfCourse also tried to use a variety of Loss functions and attention mechanisms to ensure the difference between models.
The Gower Street & R1 Road team, ranked fifth, made use of the topic title information provided by the data and used RNN + question-topic Similarity information to conduct joint training of the model. The result of single model was improved from 0.415 to 0.419, and the ensemble of 20 models was used to achieve a good result of 0.432.
The future of algorithmic contests
So whether these companies are trying to expand their influence or attract talent, it’s at least a good start.