终日所思,不如须臾之所学
古今之成大事业、大学问者,必经过三种之境界: “昨夜西风凋碧树,独上高楼,望尽天涯路。” 此第一境也。 “ 衣带渐宽终不悔,为伊消得人憔悴。” 此第二境也。 “ 众里寻他千百度,蓦然回首,那人却在,灯火阑珊处。”此第三境也。
1. medium文章Papers with Code 2020 : A Year in Review
列举了2020年度paperwithcode网站搜索前十的文章,库或者主题,其中前10的论文是:
-
EfficientDet: Scalable and Efficient Object Detection — Tan et al https://paperswithcode.com/paper/efficientdet-scalable-and-efficient-object
-
Fixing the train-test resolution discrepancy — Touvron et al https://paperswithcode.com/paper/fixing-the-train-test-resolution-discrepancy-2
-
ResNeSt: Split-Attention Networks — Zhang et al https://paperswithcode.com/paper/resnest-split-attention-networks
-
Big Transfer (BiT) — Kolesnikov et al https://paperswithcode.com/paper/large-scale-learning-of-general-visual
-
Object-Contextual Representations for Semantic Segmentation — Yuan et al https://paperswithcode.com/paper/object-contextual-representations-for
-
Self-training with Noisy Student improves ImageNet classification — Xie et al https://paperswithcode.com/paper/self-training-with-noisy-student-improves
-
YOLOv4: Optimal Speed and Accuracy of Object Detection — Bochkovskiy et al https://paperswithcode.com/paper/yolov4-optimal-speed-and-accuracy-of-object
-
An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale — Dosovitskiy et al https://paperswithcode.com/paper/an-image-is-worth-16x16-words-transformers-1
-
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer — Raffel et al https://paperswithcode.com/paper/exploring-the-limits-of-transfer-learning
-
Hierarchical Multi-Scale Attention for Semantic Segmentation — Tao et al https://paperswithcode.com/paper/hierarchical-multi-scale-attention-for