├── CVer学术交流群.png
├── ECCV2020-Papers-with-Code.md
├── ECCV2022-Papers-with-Code.md
└── README.md
/CVer学术交流群.png:
--------------------------------------------------------------------------------
https://raw.githubusercontent.com/amusi/ECCV2024-Papers-with-Code/b3864f2e54221f3badf1d30b20919b3d71eedc8a/CVer学术交流群.png
--------------------------------------------------------------------------------
/ECCV2020-Papers-with-Code.md:
--------------------------------------------------------------------------------
1 | # ECCV2020-Code
2 | ECCV 2020 论文开源项目合集,同时欢迎各位大佬提交issue,分享ECCV 2020开源项目
3 |
4 | 关于往年CV顶会论文(如CVPR 2020、ICCV 2019、ECCV 2018)以及其他优质CV论文和大盘点,详见: https://github.com/amusi/daily-paper-computer-vision
5 |
6 | - [CNN](#CNN)
7 | - [图像分类](#Image-Classification)
8 | - [2D目标检测](#Object-Detection)
9 | - [3D目标检测](#3D-Object-Detection)
10 | - [视频目标检测](#Video-Object-Detection)
11 | - [语义分割](#Semantic-Segmentation)
12 | - [实例分割](#Instance-Segmentation)
13 | - [全景分割](#Panoptic-Segmentation)
14 | - [视频目标分割](#VOS)
15 | - [单/多目标跟踪](#Object-Tracking)
16 | - [GAN](#GAN)
17 | - [NAS](#NAS)
18 | - [3D点云(分类/分割/配准/补全等)](#3D-PointCloud)
19 | - [人脸(检测/识别/解析等)](#Face)
20 | - [Re-ID](#Re-ID)
21 | - [显著性检测(SOD)](#Saliency)
22 | - [模型压缩(剪枝/知识蒸馏等)](#Model-Compression)
23 | - [视频理解/行为识别/行为检测](#Action-Recognition)
24 | - [场景文本检测](#Scene-Text-Detection)
25 | - [场景文本识别](#Scene-Text-Recognition)
26 | - [特征点检测/描述符/匹配](#Feature)
27 | - [姿态估计](#Pose-Estimation)
28 | - [深度估计](#Depth-Estimation)
29 | - [深度补全](#Depth-Completion)
30 | - [域泛化](#Domain-Generalization)
31 | - [超分辨率](#Super-Resolution)
32 | - [去模糊](#Deblurring)
33 | - [去雨](#Deraining)
34 | - [图像/视频恢复](#Image-Restoration)
35 | - [图像/视频修复(补全)](#Image-Video-Inpainting)
36 | - [风格迁移](#Style-Transfer)
37 | - [三维重建](#3D-Reconstruction)
38 | - [图像描述](#Image-Caption)
39 | - [图像检索](#Image-Retrieval)
40 | - [光流估计](#Optical-Flow-Estimation)
41 | - [视频插帧](#Video-Interpolation)
42 | - [车道线检测](#Lane-Detection)
43 | - [轨迹预测](#TP)
44 | - [线段检测](#Line-Detection)
45 | - [视线估计](#Gaze-Estimation)
46 | - [眼动追踪](#Eye-Tracking)
47 | - [对抗攻击](#Adversarial-Attack)
48 | - [数据集](#Datasets)
49 | - [其他](#Others)
50 | - [不确定中没中](#Not-Sure)
51 |
52 |
53 |
54 | # CNN
55 |
56 | **Beyond Fixed Grid: Learning Geometric Image Representation with a Deformable Grid**
57 |
58 | - 主页:http://www.cs.toronto.edu/~jungao/def-grid/
59 | - 论文:http://xxx.itp.ac.cn/abs/2008.09269
60 | - 代码:https://github.com/fidler-lab/deformable-grid-release
61 |
62 | **WeightNet: Revisiting the Design Space of Weight Networks**
63 |
64 | - 论文:https://arxiv.org/abs/2007.11823
65 | - 代码:https://github.com/megvii-model/WeightNet
66 |
67 | **Feature Pyramid Transformer**
68 |
69 | - 论文:https://arxiv.org/abs/2007.09451
70 |
71 | - 代码:https://github.com/ZHANGDONG-NJUST/FPT
72 |
73 | **Dynamic Group Convolution for Accelerating Convolutional Neural Networks**
74 |
75 | - 论文:https://arxiv.org/abs/2007.04242
76 | - 代码:https://github.com/zhuogege1943/dgc
77 |
78 | **Learning to Learn Parameterized Classification Networks for Scalable Input Images**
79 |
80 | - 论文:https://arxiv.org/abs/2007.06181
81 |
82 | - 代码:https://github.com/d-li14/SAN
83 |
84 | **Rethinking Bottleneck Structure for Efficient Mobile Network Design**
85 |
86 | - 论文:https://arxiv.org/abs/2007.02269
87 | - 代码:https://github.com/zhoudaquan/rethinking_bottleneck_design
88 |
89 | **MutualNet: Adaptive ConvNet via Mutual Learning from Network Width and Resolution**
90 |
91 | - 论文:Oral
92 | - 论文:https://arxiv.org/abs/1909.12978
93 | - 代码:https://github.com/taoyang1122/MutualNet
94 |
95 | **PSConv: Squeezing Feature Pyramid into One Compact Poly-Scale Convolutional Layer**
96 |
97 | - 论文:https://arxiv.org/abs/2007.06191
98 | - 代码:https://github.com/d-li14/PSConv
99 |
100 |
101 |
102 | # 图像分类
103 |
104 | **Learning to Learn Parameterized Classification Networks for Scalable Input Images**
105 |
106 | - 论文:暂无
107 |
108 | - 代码:https://github.com/d-li14/SAN
109 |
110 | **Learning To Classify Images Without Labels**
111 |
112 | - 论文:https://arxiv.org/abs/2005.12320
113 | - 代码:https://github.com/wvangansbeke/Unsupervised-Classification
114 |
115 |
116 |
117 | # 2D目标检测
118 |
119 | **Learning Data Augmentation Strategies for Object Detection**
120 |
121 | - 论文:https://www.ecva.net/papers/eccv_2020/papers_ECCV/html/5865_ECCV_2020_paper.php
122 | - 代码:https://github.com/tensorflow/tpu/tree/master/models/official/detection
123 |
124 | **AABO: Adaptive Anchor Box Optimization for Object Detection via Bayesian Sub-sampling**
125 |
126 | - 论文:https://www.ecva.net/papers/eccv_2020/papers_ECCV/html/3977_ECCV_2020_paper.php
127 | - 代码:https://github.com/wwdkl/AABO
128 |
129 | **Side-Aware Boundary Localization for More Precise Object Detection**
130 |
131 | - 论文:https://www.ecva.net/papers/eccv_2020/papers_ECCV/html/2272_ECCV_2020_paper.php
132 | - 代码:https://github.com/open-mmlab/mmdetection
133 |
134 | **TIDE: A General Toolbox for Identifying Object Detection Errors**
135 |
136 | - 主页:https://dbolya.github.io/tide/
137 |
138 | - 论文:https://arxiv.org/abs/2008.08115
139 |
140 | - 代码:https://github.com/dbolya/tide
141 |
142 | **Every Pixel Matters: Center-aware Feature Alignment for Domain Adaptive Object Detector**
143 |
144 | - 主页:https://chengchunhsu.github.io/EveryPixelMatters/
145 | - 论文:https://arxiv.org/abs/2008.08574
146 | - 代码:https://github.com/chengchunhsu/EveryPixelMatters
147 |
148 | **Dense RepPoints: Representing Visual Objects with Dense Point Sets**
149 |
150 | - 论文:https://arxiv.org/abs/1912.11473
151 | - 代码:https://github.com/justimyhxu/Dense-RepPoints
152 |
153 | **Corner Proposal Network for Anchor-free, Two-stage Object Detection**
154 |
155 | - 论文:https://arxiv.org/abs/2007.13816
156 |
157 | - 代码:https://github.com/Duankaiwen/CPNDet
158 |
159 | **BorderDet: Border Feature for Dense Object Detection**
160 |
161 | - 论文:https://arxiv.org/abs/2007.11056
162 |
163 | - 代码:https://github.com/Megvii-BaseDetection/BorderDet
164 | - 中文解读:https://zhuanlan.zhihu.com/p/163044323
165 |
166 | **Multi-Scale Positive Sample Refinement for Few-Shot Object Detection**
167 |
168 | - 论文:https://arxiv.org/abs/2007.09384
169 |
170 | - 代码:https://github.com/jiaxi-wu/MPSR
171 |
172 | **PIoU Loss: Towards Accurate Oriented Object Detection in Complex Environments**
173 |
174 | - 论文:https://arxiv.org/abs/2007.09584
175 |
176 | - 代码:https://github.com/clobotics/piou
177 |
178 | - 数据集:https://github.com/clobotics/piou
179 |
180 | **Probabilistic Anchor Assignment with IoU Prediction for Object Detection**
181 |
182 | - 论文:https://arxiv.org/abs/2007.08103
183 | - 代码:https://github.com/kkhoot/PAA
184 |
185 | **HoughNet: Integrating near and long-range evidence for bottom-up object detection**
186 |
187 | - 论文:https://arxiv.org/abs/2007.02355
188 | - 代码:https://github.com/nerminsamet/houghnet
189 |
190 | **OS2D: One-Stage One-Shot Object Detection by Matching Anchor Features**
191 |
192 | - 论文:https://arxiv.org/abs/2003.06800
193 |
194 | - 代码:https://github.com/aosokin/os2d
195 |
196 | **End-to-End Object Detection with Transformers**
197 |
198 | - Oral
199 |
200 | - 论文:https://ai.facebook.com/research/publications/end-to-end-object-detection-with-transformers
201 | - 代码:https://github.com/facebookresearch/detr
202 |
203 | **Dynamic R-CNN: Towards High Quality Object Detection via Dynamic Training**
204 |
205 | - 论文:https://arxiv.org/abs/2004.06002
206 | - 代码:https://github.com/hkzhang95/DynamicRCNN
207 |
208 | **OS2D: One-Stage One-Shot Object Detection by Matching Anchor Features**
209 |
210 | - 论文:https://www.ecva.net/papers/eccv_2020/papers_ECCV/html/2424_ECCV_2020_paper.php
211 | - 代码:https://github.com/aosokin/os2d
212 |
213 | **Object Detection with a Unified Label Space from Multiple Datasets**
214 |
215 | - 主页:http://www.nec-labs.com/~mas/UniDet/
216 | - 论文:https://arxiv.org/abs/2008.06614
217 | - 代码:暂无
218 | - 数据集:http://www.nec-labs.com/~mas/UniDet/resources/UOD_dataset_ECCV20.zip
219 |
220 | ### 弱监督目标检测
221 |
222 | **Enabling Deep Residual Networks for Weakly Supervised Object Detection**
223 |
224 | - 论文:https://www.ecva.net/papers/eccv_2020/papers_ECCV/html/479_ECCV_2020_paper.php
225 | - 代码:https://github.com/shenyunhang/DRN-WSOD
226 |
227 | **UFO²: A Unified Framework towards Omni-supervised Object Detection**
228 |
229 | - 论文:https://www.ecva.net/papers/eccv_2020/papers_ECCV/html/3205_ECCV_2020_paper.php
230 | - 代码:https://github.com/NVlabs/wetectron
231 |
232 | **Boosting Weakly Supervised Object Detection with Progressive Knowledge Transfer**
233 |
234 | - 论文:https://arxiv.org/abs/2007.07986
235 | - 代码:https://github.com/mikuhatsune/wsod_transfer
236 |
237 | ### 域自适应目标检测
238 |
239 | **Collaborative Training between Region Proposal Localization and Classification for Domain Adaptive Object Detection**
240 |
241 | - 论文:https://www.ecva.net/papers/eccv_2020/papers_ECCV/html/2920_ECCV_2020_paper.php
242 | - 代码:https://github.com/GanlongZhao/CST_DA_detection
243 |
244 | **Every Pixel Matters: Center-aware Feature Alignment for Domain Adaptive Object Detector**
245 |
246 | - 主页:https://chengchunhsu.github.io/EveryPixelMatters/
247 | - 论文:https://arxiv.org/abs/2008.08574
248 | - 代码:https://github.com/chengchunhsu/EveryPixelMatters
249 |
250 | ### Few-Shot 目标检测
251 |
252 | **Multi-Scale Positive Sample Refinement for Few-Shot Object Detection**
253 |
254 | - 论文:https://www.ecva.net/papers/eccv_2020/papers_ECCV/html/2581_ECCV_2020_paper.php
255 | - 代码:https://github.com/jiaxi-wu/MPSR
256 |
257 | ### 水下目标检测
258 |
259 | **Dual Refinement Underwater Object Detection Network**
260 |
261 | - 论文:https://www.ecva.net/papers/eccv_2020/papers_ECCV/html/3425_ECCV_2020_paper.php
262 | - 代码:https://github.com/Peterchen111/FERNet
263 |
264 | ## 遥感旋转目标检测
265 |
266 | **PIoU Loss: Towards Accurate Oriented Object Detection in Complex Environments**
267 |
268 | - 论文:https://www.ecva.net/papers/eccv_2020/papers_ECCV/html/3087_ECCV_2020_paper.php
269 | - 代码:https://github.com/clobotics/piou
270 | - 数据集:https://github.com/clobotics/piou
271 |
272 | **Arbitrary-Oriented Object Detection with Circular Smooth Label**
273 |
274 | - 论文:https://arxiv.org/abs/2003.05597
275 | - 代码:https://github.com/Thinklab-SJTU/CSL_RetinaNet_Tensorflow
276 |
277 |
278 |
279 | # 3D目标检测
280 |
281 | **Rethinking Pseudo-LiDAR Representation**
282 |
283 | - 论文:https://arxiv.org/abs/2008.04582
284 |
285 | - 代码:https://github.com/xinzhuma/patchnet
286 |
287 | **Pillar-based Object Detection for Autonomous Driving**
288 |
289 | - 论文:https://arxiv.org/abs/2007.10323
290 | - 代码:https://github.com/WangYueFt/pillar-od
291 |
292 | **EPNet: Enhancing Point Features with Image Semantics for 3D Object Detection**
293 |
294 | - 论文:https://arxiv.org/abs/2007.08856
295 | - 代码:https://github.com/happinesslz/EPNet
296 |
297 |
298 |
299 | # 视频目标检测
300 |
301 | **Mining Inter-Video Proposal Relations for Video Object Detection**
302 |
303 | - 论文:https://www.ecva.net/papers/eccv_2020/papers_ECCV/html/3764_ECCV_2020_paper.php
304 | - 代码:https://github.com/youthHan/HVRNet
305 |
306 | **Learning Where to Focus for Efficient Video Object Detection**
307 |
308 | - 主页:https://jiangzhengkai.github.io/LSTS/
309 | - 论文:https://www.ecva.net/papers/eccv_2020/papers_ECCV/papers/123610018.pdf
310 | - 代码:https://github.com/jiangzhengkai/LSTS
311 |
312 |
313 |
314 | # 语义分割
315 |
316 | **SNE-RoadSeg: Incorporating Surface Normal Information into Semantic Segmentation for Accurate Freespace Detection**
317 |
318 | - 主页:https://sites.google.com/view/sne-roadseg
319 | - 论文:https://arxiv.org/abs/2008.11351
320 | - 代码:https://github.com/hlwang1124/SNE-RoadSeg
321 | - 数据集:https://drive.google.com/file/d/1YnkqPmzxtjNfMi2B1gMy7LQa5Gnu-BsH/view
322 |
323 | **Tensor Low-Rank Reconstruction for Semantic Segmentation**
324 |
325 | - 论文:https://arxiv.org/abs/2008.00490
326 |
327 | - 代码:https://github.com/CWanli/RecoNet
328 |
329 | **Bi-directional Cross-Modality Feature Propagation with Separation-and-Aggregation Gate for RGB-D Semantic Segmentation**
330 |
331 | - 论文:https://arxiv.org/abs/2007.09183
332 |
333 | - 代码:https://github.com/charlesCXK/RGBD_Semantic_Segmentation_PyTorch
334 |
335 | **GMNet: Graph Matching Network for Large Scale Part Semantic Segmentation in the Wild**
336 |
337 | - 主页:https://lttm.dei.unipd.it/paper_data/GMNet/
338 | - 论文:https://arxiv.org/abs/2007.09073
339 | - 代码:https://github.com/LTTM/GMNet
340 |
341 | **SegFix: Model-Agnostic Boundary Refinement for Segmentation**
342 |
343 | - 论文:https://arxiv.org/abs/2007.04269
344 |
345 | - 代码:https://github.com/openseg-group/openseg.pytorch
346 |
347 | **Mining Cross-Image Semantics for Weakly Supervised Semantic Segmentation**
348 |
349 | - Oral
350 | - 论文:https://arxiv.org/abs/2007.01947
351 | - 代码:https://github.com/GuoleiSun/MCIS_wsss
352 |
353 | **Improving Semantic Segmentation via Decoupled Body and Edge Supervision**
354 |
355 | - 论文:https://arxiv.org/abs/2007.10035
356 |
357 | - 代码:https://github.com/lxtGH/DecoupleSegNets
358 |
359 |
360 |
361 | # 实例分割
362 |
363 | **SipMask: Spatial Information Preservation for Fast Image and Video Instance Segmentation**
364 |
365 | - 论文:https://arxiv.org/abs/2007.14772
366 |
367 | - 代码:https://github.com/JialeCao001/SipMask
368 |
369 | **Commonality-Parsing Network across Shape and Appearance for Partially Supervised Instance Segmentation**
370 |
371 | - 论文:https://arxiv.org/abs/2007.12387
372 |
373 | - 代码:https://github.com/fanq15/CPMask
374 |
375 | **Boundary-preserving Mask R-CNN**
376 |
377 | - 论文:https://arxiv.org/abs/2007.08921
378 |
379 | - 代码:https://github.com/hustvl/BMaskR-CNN
380 |
381 | **Conditional Convolutions for Instance Segmentation**
382 |
383 | - Oral
384 | - 论文:https://arxiv.org/abs/2003.05664
385 | - 代码:https://github.com/aim-uofa/AdelaiDet/blob/master/configs/CondInst/README.md
386 |
387 | **SOLO: Segmenting Objects by Locations**
388 |
389 | - 论文:https://arxiv.org/abs/1912.04488
390 | - 代码:https://github.com/WXinlong/SOLO
391 |
392 | - 知乎话题:https://www.zhihu.com/question/360594484
393 |
394 |
395 |
396 | # 全景分割
397 |
398 | **Axial-DeepLab: Stand-Alone Axial-Attention for Panoptic Segmentation**
399 |
400 | - 论文:https://arxiv.org/abs/2003.07853
401 |
402 | - 代码:https://github.com/csrhddlam/axial-deeplab
403 | - 视频:https://youtu.be/-iAXF-vibdE
404 |
405 |
406 |
407 | # 视频目标分割
408 |
409 | **Collaborative Video Object Segmentation by Foreground-Background Integration**
410 |
411 | - 论文:https://arxiv.org/abs/2003.08333
412 | - 代码:https://github.com/z-x-yang/CFBI
413 |
414 | **Video Object Segmentation with Episodic Graph Memory Networks**
415 |
416 | - 论文:https://arxiv.org/abs/2007.07020
417 |
418 | - 代码:https://github.com/carrierlxk/GraphMemVOS
419 |
420 |
421 |
422 | # 单/多目标跟踪
423 |
424 | **Ocean: Object-aware Anchor-Free Tracking**
425 |
426 | - 论文:https://arxiv.org/abs/2006.10721
427 |
428 | - 代码:https://github.com/researchmm/TracKit
429 |
430 | ## 多目标跟踪
431 |
432 | **Towards Real-Time Multi-Object Tracking**
433 |
434 | - 论文:暂无
435 | - 代码:https://github.com/Zhongdao/Towards-Realtime-MOT
436 |
437 | **Simultaneous Detection and Tracking with Motion Modelling for Multiple Object Tracking**
438 |
439 | - 论文:https://arxiv.org/abs/2008.08826
440 | - 代码:https://github.com/shijieS/DMMN
441 | - 数据集:https://github.com/shijieS/OmniMOTDataset
442 |
443 | **Chained-Tracker: Chaining Paired Attentive Regression Results for End-to-End Joint Multiple-Object Detection and Tracking**
444 |
445 | - 论文:https://arxiv.org/abs/2007.14557
446 |
447 | - 代码:https://github.com/pjl1995/CTracker
448 |
449 | **Ocean: Object-aware Anchor-Free Tracking**
450 |
451 | - 论文:https://arxiv.org/abs/2006.10721
452 |
453 | - 代码:https://github.com/researchmm/TracKit
454 |
455 | **TAO: A Large-Scale Benchmark for Tracking Any Object**
456 |
457 | - 主页:http://taodataset.org/
458 | - 论文:https://arxiv.org/abs/2005.10356
459 | - 代码:https://github.com/TAO-Dataset/tao
460 |
461 | **Segment as Points for Efficient Online Multi-Object Tracking and Segmentation**
462 |
463 | - Oral
464 | - 论文:https://arxiv.org/abs/2007.01550
465 | - 代码:https://github.com/detectRecog/PointTrack
466 | - 数据集:https://github.com/detectRecog/PointTrack
467 |
468 |
469 |
470 | # GAN
471 |
472 | **Rewriting a Deep Generative Model**
473 |
474 | - 论文:https://arxiv.org/abs/2007.15646
475 |
476 | - 代码:https://github.com/davidbau/rewriting
477 |
478 | **Contrastive Learning for Unpaired Image-to-Image Translation**
479 |
480 | - 论文:https://arxiv.org/abs/2007.15651
481 | - 代码:https://github.com/taesungp/contrastive-unpaired-translation
482 |
483 | **XingGAN for Person Image Generation**
484 |
485 | - 论文:暂无
486 | - 代码:https://github.com/Ha0Tang/XingGAN
487 |
488 |
489 |
490 | # NAS
491 |
492 | **Are Labels Necessary for Neural Architecture Search?**
493 |
494 | - 论文:https://arxiv.org/abs/2003.12056
495 |
496 | - 代码:https://github.com/facebookresearch/unnas
497 |
498 | **Rethinking Bottleneck Structure for Efficient Mobile Network Design**
499 |
500 | - 论文:https://arxiv.org/abs/2007.02269
501 | - 代码:https://github.com/zhoudaquan/rethinking_bottleneck_design
502 |
503 | **Fair DARTS: Eliminating Unfair Advantages in Differentiable Architecture Search**
504 |
505 | - 论文:https://arxiv.org/abs/1911.12126
506 | - 代码:https://github.com/xiaomi-automl/fairdarts
507 |
508 |
509 |
510 | # 3D点云(分类/分割/配准/补全等)
511 |
512 | **AdvPC: Transferable Adversarial Perturbations on 3D Point Clouds**
513 |
514 | - 论文:https://arxiv.org/abs/1912.00461
515 |
516 | - 代码:https://github.com/ajhamdi/AdvPC
517 |
518 | **A Closer Look at Local Aggregation Operators in Point Cloud Analysis**
519 |
520 | - 论文:https://arxiv.org/abs/2007.01294
521 | - 代码:https://github.com/zeliu98/CloserLook3D
522 |
523 | ## 3D点云补全
524 |
525 | **Multimodal Shape Completion via Conditional Generative Adversarial Networks**
526 |
527 | - 论文:https://arxiv.org/abs/2003.07717
528 | - 代码:https://github.com/ChrisWu1997/Multimodal-Shape-Completion
529 |
530 | **GRNet: Gridding Residual Network for Dense Point Cloud Completion**
531 |
532 | - 论文:https://arxiv.org/abs/2006.03761
533 | - 代码:https://github.com/hzxie/GRNet
534 |
535 | ## 3D点云生成
536 |
537 | **Progressive Point Cloud Deconvolution Generation Network**
538 |
539 | - 论文:https://arxiv.org/abs/2007.05361
540 |
541 | - 代码:https://github.com/fpthink/PDGN
542 |
543 |
544 |
545 | # 人脸(检测/识别/解析等)
546 |
547 | ## 人脸检测
548 |
549 | **ProgressFace: Scale-Aware Progressive Learning for Face Detection**
550 |
551 | - 论文:http://www.ecva.net/papers/eccv_2020/papers_ECCV/papers/123510341.pdf
552 | - 代码:https://github.com/jiashu-zhu/ProgressFace
553 |
554 | ## 人脸识别
555 |
556 | **Explainable Face Recognition**
557 |
558 | - 论文:https://arxiv.org/abs/2008.00916
559 |
560 | - 主页:https://stresearch.github.io/xfr/
561 | - 代码:https://github.com/stresearch/xfr
562 |
563 | ## 3D人脸重建
564 |
565 | **Self-Supervised Monocular 3D Face Reconstruction by Occlusion-Aware Multi-view Geometry Consistency**
566 |
567 | - 论文:https://arxiv.org/abs/2007.12494
568 | - 代码:https://github.com/jiaxiangshang/MGCNet
569 |
570 | ## 人脸活体检测
571 |
572 | **CelebA-Spoof: Large-Scale Face Anti-Spoofing Dataset with Rich Annotations**
573 |
574 | - 论文:https://arxiv.org/abs/2007.12342
575 |
576 | - 数据集:https://github.com/Davidzhangyuanhan/CelebA-Spoof
577 |
578 | ## 人脸解析
579 |
580 | **Edge-aware Graph Representation Learning and Reasoning for Face Parsing**
581 |
582 | - 论文:https://arxiv.org/abs/2007.11240
583 | - 代码:https://github.com/tegusi/EAGRNet
584 |
585 | ## DeepFakes
586 |
587 | **What makes fake images detectable? Understanding properties that generalize**
588 |
589 | - 主页:https://chail.github.io/patch-forensics/
590 | - 论文:http://xxx.itp.ac.cn/abs/2008.10588
591 | - 代码:https://github.com/chail/patch-forensics
592 |
593 | ## 其他
594 |
595 | **Lifespan Age Transformation Synthesis**
596 |
597 | - 论文:https://arxiv.org/abs/2003.09764
598 | - 代码:https://github.com/royorel/Lifespan_Age_Transformation_Synthesis
599 |
600 |
601 |
602 | # Re-ID
603 |
604 | ## 行人重识别
605 |
606 | **Rethinking the Distribution Gap of Person Re-identification with Camera-based Batch Normalization**
607 |
608 | - 论文:https://arxiv.org/abs/2001.08680
609 | - 代码:https://github.com/automan000/Camera-based-Person-ReID
610 |
611 | **Appearance-Preserving 3D Convolution for Video-based Person Re-identification**
612 |
613 | - Oral
614 |
615 | - 论文:https://arxiv.org/pdf/2007.08434
616 | - 代码:https://github.com/guxinqian/AP3D
617 |
618 | **Do Not Disturb Me: Person Re-identification Under the Interference of Other Pedestrians**
619 |
620 | - 论文:https://arxiv.org/abs/2008.06963
621 | - 代码:https://github.com/X-BrainLab/PI-ReID
622 |
623 | **Faster Person Re-Identification**
624 |
625 | - 论文:https://arxiv.org/abs/2008.06826
626 |
627 | - 代码:https://github.com/wangguanan/light-reid
628 |
629 | **Temporal Complementary Learning for Video Person Re-Identification**
630 |
631 | - 论文:https://arxiv.org/abs/2007.09357
632 |
633 | - 代码:https://github.com/blue-blue272/VideoReID-TCLNet
634 |
635 | **Joint Disentangling and Adaptation for Cross-Domain Person Re-Identification**
636 |
637 | - 论文:https://arxiv.org/abs/2007.10315
638 | - 代码:https://github.com/NVlabs/DG-Net-PP
639 |
640 | **Robust Re-Identification by Multiple Views Knowledge Distillation**
641 |
642 | - 论文:https://arxiv.org/abs/2007.04174
643 | - 代码:https://github.com/aimagelab/VKD
644 |
645 | **Multiple Expert Brainstorming for Domain Adaptive Person Re-identification**
646 |
647 | - 论文:https://arxiv.org/abs/2007.01546
648 | - 代码:https://github.com/YunpengZhai/MEB-Net
649 |
650 | ## 车辆重识别
651 |
652 | **Simulating Content Consistent Vehicle Datasets with Attribute Descent**
653 |
654 | - 论文:https://arxiv.org/abs/1912.08855
655 | - 代码:https://github.com/yorkeyao/VehicleX
656 | - 数据集:https://github.com/yorkeyao/VehicleX
657 |
658 | **Orientation-aware Vehicle Re-identification with Semantics-guided Part Attention Network**
659 |
660 | - 主页:http://media.ee.ntu.edu.tw/research/SPAN/
661 |
662 | - 论文:https://arxiv.org/abs/2008.11423
663 | - 代码:https://github.com/tsaishien-chen/SPAN
664 |
665 |
666 |
667 | # 显著性检测(SOD)
668 |
669 | **Progressively Guided Alternate Refinement Network for RGB-D Salient Object Detection**
670 |
671 | - 论文:http://xxx.itp.ac.cn/abs/2008.07064
672 |
673 | - 代码:https://github.com/ShuhanChen/PGAR_ECCV20
674 |
675 | **Suppress and Balance: A Simple Gated Network for Salient Object Detection**
676 |
677 | - Oral
678 |
679 | - 论文:https://arxiv.org/abs/2007.08074
680 | - 代码:https://github.com/Xiaoqi-Zhao-DLUT/GateNet-RGB-Saliency
681 |
682 | **Hierarchical Dynamic Filtering Network for RGB-D Salient Object Detection**
683 |
684 | - 论文:https://arxiv.org/abs/2007.06227
685 |
686 | - 代码:https://github.com/lartpang/HDFNet
687 |
688 | **A Single Stream Network for Robust and Real-time RGB-D Salient Object Detection**
689 |
690 | - 论文:https://arxiv.org/abs/2007.06811
691 | - 代码:https://github.com/Xiaoqi-Zhao-DLUT/DANet-RGBD-Saliency
692 |
693 | **Cross-Modal Weighting Network for RGB-D Salient Object Detection**
694 |
695 | - 论文:暂无
696 |
697 | - 代码:https://github.com/MathLee/CMWNet
698 |
699 | **BBS-Net: RGB-D Salient Object Detection with a Bifurcated Backbone Strategy Network**
700 |
701 | - 论文:暂无
702 | - 代码:https://github.com/DengPingFan/BBS-Net
703 |
704 | **Highly Efficient Salient Object Detection with 100K Parameters**
705 |
706 | - 论文:https://arxiv.org/abs/2003.05643
707 | - 代码:https://github.com/MCG-NKU/Sal100K
708 |
709 |
710 |
711 | # 模型压缩(剪枝/知识蒸馏等)
712 |
713 | **EagleEye: Fast Sub-net Evaluation for Efficient Neural Network Pruning**
714 |
715 | - 论文:https://arxiv.org/abs/2007.02491
716 | - 代码:https://github.com/anonymous47823493/EagleEye
717 |
718 |
719 |
720 | # 视频理解/行为识别/行为检测
721 |
722 | **AssembleNet++: Assembling Modality Representations via Attention Connections**
723 |
724 | - 论文:https://arxiv.org/abs/2008.08072
725 | - 代码:https://sites.google.com/corp/view/assemblenet/
726 |
727 | **LEMMA: A Multi-view Dataset for Learning Multi-agent Multi-task Activities**
728 |
729 | - 主页:https://sites.google.com/view/lemma-activity
730 |
731 | - 论文:https://arxiv.org/abs/2007.15781
732 |
733 | - 数据集:https://sites.google.com/view/lemma-activity/home/dataset
734 | - 代码:https://github.com/Buzz-Beater/LEMMA
735 |
736 | **AR-Net: Adaptive Frame Resolution for Efficient Action Recognition**
737 |
738 | - 主页:https://mengyuest.github.io/AR-Net/
739 | - 论文:https://arxiv.org/abs/2007.15796
740 | - 代码:https://github.com/mengyuest/AR-Net
741 |
742 | **Context-Aware RCNN: A Baseline for Action Detection in Videos**
743 |
744 | - 论文:https://arxiv.org/abs/2007.09861
745 |
746 | - 代码:https://github.com/MCG-NJU/CRCNN-Action
747 |
748 | **Actions as Moving Points**
749 |
750 | - 论文:https://arxiv.org/abs/2001.04608
751 | - 代码:https://github.com/MCG-NJU/MOC-Detector
752 |
753 | **SF-Net: Single-Frame Supervision for Temporal Action Localization**
754 |
755 | - 论文:https://arxiv.org/abs/2003.06845
756 | - 代码:https://github.com/Flowerfan/SF-Net
757 |
758 | **Asynchronous Interaction Aggregation for Action Detection**
759 |
760 | - 论文:https://arxiv.org/abs/2004.07485
761 |
762 | - 代码:https://github.com/MVIG-SJTU/AlphAction
763 |
764 |
765 |
766 | # 场景文本检测
767 |
768 | **Mask TextSpotter v3: Segmentation Proposal Network for Robust Scene Text Spotting**
769 |
770 | - 论文:https://arxiv.org/abs/2007.09482
771 |
772 | - 代码:https://github.com/MhLiao/MaskTextSpotterV3
773 |
774 |
775 |
776 | # 场景文本识别
777 |
778 | **Adaptive Text Recognition through Visual Matching**
779 |
780 | - 主页:http://www.robots.ox.ac.uk/~vgg/research/FontAdaptor20/
781 |
782 | - 论文:https://arxiv.org/abs/2009.06610
783 |
784 | - 代码:https://github.com/Chuhanxx/FontAdaptor
785 |
786 | **Mask TextSpotter v3: Segmentation Proposal Network for Robust Scene Text Spotting**
787 |
788 | - 论文:https://arxiv.org/abs/2007.09482
789 |
790 | - 代码:https://github.com/MhLiao/MaskTextSpotterV3
791 |
792 |
793 |
794 | # 特征点检测/描述符/匹配
795 |
796 | **Learning and aggregating deep local descriptors for instance-level recognition**
797 |
798 | - 论文:https://arxiv.org/abs/2007.13172
799 |
800 | - 代码:https://github.com/gtolias/how
801 |
802 | **Online Invariance Selection for Local Feature Descriptors**
803 |
804 | - Oral
805 | - 论文:https://arxiv.org/abs/2007.08988
806 | - 代码:https://github.com/rpautrat/LISRD
807 |
808 | **Single-Image Depth Prediction Makes Feature Matching Easier**
809 |
810 | - 论文:https://arxiv.org/abs/2008.09497
811 |
812 | - 代码:http://www.github.com/nianticlabs/rectified-features
813 |
814 |
815 |
816 | # 姿态估计
817 |
818 | **Pose2Mesh: Graph Convolutional Network for 3D Human Pose and Mesh Recovery from a 2D Human Pose**
819 |
820 | - 论文:https://arxiv.org/abs/2008.09047
821 | - 代码:https://github.com/hongsukchoi/Pose2Mesh_RELEASE
822 |
823 | **Key Frame Proposal Network for Efficient Pose Estimation in Videos**
824 |
825 | - 论文:https://arxiv.org/abs/2007.15217
826 | - 代码:https://github.com/Yuexiaoxi10/Key-Frame-Proposal-Network-for-Efficient-Pose-Estimation-in-Videos
827 |
828 | ## 3D人体姿态估计
829 |
830 | **DOPE: Distillation Of Part Experts for whole-body 3D pose estimation in the wild**
831 |
832 | - 主页:https://europe.naverlabs.com/blog/dope-distillation-of-part-experts-for-whole-body-3d-pose-estimation-in-the-wild/
833 |
834 | - 论文:https://arxiv.org/abs/2008.09457
835 |
836 | - 代码:https://github.com/naver/dope
837 |
838 |
839 | **SMAP: Single-Shot Multi-Person Absolute 3D Pose Estimation**
840 |
841 | - 主页:https://zju3dv.github.io/SMAP/
842 | - 论文:https://arxiv.org/abs/2008.11469
843 | - 代码:https://github.com/zju3dv/SMAP
844 |
845 | ## 6D位姿估计
846 |
847 | **CosyPose: Consistent multi-view multi-object 6D pose estimation**
848 |
849 | - 主页:https://www.di.ens.fr/willow/research/cosypose/
850 |
851 | - 论文:http://xxx.itp.ac.cn/abs/2008.08465
852 |
853 | - 代码:https://github.com/ylabbe/cosypose
854 |
855 |
856 |
857 | # 深度估计
858 |
859 | **Learning Stereo from Single Images**
860 |
861 | - 论文:https://arxiv.org/abs/2008.01484
862 | - 代码:https://github.com/nianticlabs/stereo-from-mono/
863 |
864 | ## 单目深度估计
865 |
866 | **P2Net: Patch-match and Plane-regularization for Unsupervised Indoor Depth Estimation**
867 |
868 | - 论文:https://arxiv.org/abs/2007.07696
869 | - 代码:https://github.com/svip-lab/Indoor-SfMLearner
870 |
871 | **Self-Supervised Monocular Depth Estimation: Solving the Dynamic Object Problem by Semantic Guidance**
872 |
873 | - 论文:https://arxiv.org/abs/2007.06936
874 |
875 | - 代码:https://github.com/ifnspaml/SGDepth
876 |
877 |
878 |
879 | # 深度补全
880 |
881 | **Non-Local Spatial Propagation Network for Depth Completion**
882 |
883 | - 论文:https://arxiv.org/abs/2007.10042
884 | - 代码:https://github.com/zzangjinsun/NLSPN_ECCV20
885 |
886 |
887 |
888 | # 域泛化
889 |
890 | **Learning from Extrinsic and Intrinsic Supervisions for Domain Generalization**
891 |
892 | - 论文:https://arxiv.org/abs/2007.09316
893 | - 代码:https://github.com/EmmaW8/EISNet
894 |
895 |
896 |
897 | # 超分辨率
898 |
899 | ## 图像超分辨率
900 |
901 | **Learning the Super-Resolution Space with Normalizing Flow**
902 |
903 | - 论文:https://arxiv.org/abs/2006.14200
904 | - 代码:https://github.com/andreas128/SRFlow
905 |
906 | **Deep Decomposition Learning for Inverse Imaging Problems**
907 |
908 | - 论文:https://arxiv.org/abs/1911.11028
909 | - 代码:https://github.com/edongdongchen/DDN
910 |
911 | **Component Divide-and-Conquer for Real-World Image Super-Resolution**
912 |
913 | - 论文:https://arxiv.org/abs/2008.01928
914 | - 代码:https://github.com/xiezw5/Component-Divide-and-Conquer-for-Real-World-Image-Super-Resolution
915 | - 数据集:https://github.com/xiezw5/Component-Divide-and-Conquer-for-Real-World-Image-Super-Resolution
916 |
917 | **Learning with Privileged Information for Efficient Image Super-Resolution**
918 |
919 | - 主页:https://cvlab.yonsei.ac.kr/projects/PISR/
920 | - 论文:https://arxiv.org/abs/2007.07524
921 | - 代码:https://github.com/cvlab-yonsei/PISR
922 |
923 | **Spatial-Angular Interaction for Light Field Image Super-Resolution**
924 |
925 | - 论文:https://arxiv.org/abs/1912.07849
926 | - 代码:https://github.com/YingqianWang/LF-InterNet
927 |
928 | **Invertible Image Rescaling**
929 |
930 | - 论文:https://arxiv.org/abs/2005.05650
931 | - 代码:https://github.com/pkuxmq/Invertible-Image-Rescaling
932 |
933 | ## 视频超分辨率
934 |
935 | **Video Super-Resolution with Recurrent Structure-Detail Network**
936 |
937 | - 论文:https://arxiv.org/abs/2008.00455
938 |
939 | - 代码:https://github.com/junpan19/RSDN
940 |
941 |
942 |
943 | # 去模糊
944 |
945 | ## 图像去模糊
946 |
947 | **End-to-end Interpretable Learning of Non-blind Image Deblurring**
948 |
949 | - 论文:https://arxiv.org/abs/2007.01769
950 | - 代码:暂无(即将出来)
951 |
952 | ## 视频去模糊
953 |
954 | **Efficient Spatio-Temporal Recurrent Neural Network for Video Deblurring**
955 |
956 | - 论文:https://www.ecva.net/papers/eccv_2020/papers_ECCV/html/5116_ECCV_2020_paper.php
957 | - 代码:https://github.com/zzh-tech/ESTRNN
958 |
959 |
960 |
961 | # 去雨
962 |
963 | **Rethinking Image Deraining via Rain Streaks and Vapors**
964 |
965 | - 论文:https://arxiv.org/abs/2008.00823
966 | - 代码:https://github.com/yluestc/derain
967 |
968 |
969 |
970 | # 图像/视频恢复
971 |
972 | **Learning Enriched Features for Real Image Restoration and Enhancement**
973 |
974 | - 论文:https://arxiv.org/abs/2003.06792
975 | - 代码:https://github.com/swz30/MIRNet
976 |
977 |
978 |
979 | # 图像/视频修复(补全)
980 |
981 | **NAS-DIP: Learning Deep Image Prior with Neural Architecture Search**
982 |
983 | - 主页:https://yunchunchen.github.io/NAS-DIP/
984 | - 论文:https://arxiv.org/abs/2008.11713
985 | - 代码:https://github.com/YunChunChen/NAS-DIP-pytorch
986 |
987 | **Learning Joint Spatial-Temporal Transformations for Video Inpainting**
988 |
989 | - 论文:https://arxiv.org/abs/2007.10247
990 |
991 | - 代码:https://github.com/researchmm/STTN
992 |
993 | **Rethinking Image Inpainting via a Mutual Encoder-Decoder with Feature Equalizations**
994 |
995 | - Oral
996 | - 论文:暂无
997 | - 代码:https://github.com/KumapowerLIU/ECCV2020oralRethinking-Image-Inpainting-via-a-Mutual-Encoder-Decoder-with-Feature-Equalizations
998 |
999 |
1000 |
1001 | # 风格迁移
1002 |
1003 | **Domain-Specific Mappings for Generative Adversarial Style Transfer**
1004 |
1005 | - 主页:https://acht7111020.github.io/DSMAP-demo/
1006 | - 论文:http://xxx.itp.ac.cn/abs/2008.02198
1007 | - 代码:https://github.com/acht7111020/DSMAP
1008 |
1009 |
1010 |
1011 | # 三维重建
1012 |
1013 | **Atlas: End-to-End 3D Scene Reconstruction from Posed Images**
1014 |
1015 | - 主页:http://zak.murez.com/atlas/
1016 | - 论文:https://arxiv.org/abs/2003.10432
1017 | - 代码:https://github.com/magicleap/Atlas
1018 | - 视频:https://youtu.be/9NOPcOGV6nU
1019 |
1020 | **3D Bird Reconstruction: a Dataset, Model, and Shape Recovery from a Single View**
1021 |
1022 | - 主页:https://marcbadger.github.io/avian-mesh/
1023 | - 论文:https://arxiv.org/abs/2008.06133
1024 | - 代码:https://github.com/marcbadger/avian-mesh
1025 | - 数据集:https://drive.google.com/file/d/1vyXYIJIo9jneIqC7lowB4GVi17rjztjn/view?usp=sharing
1026 |
1027 | **Stochastic Bundle Adjustment for Efficient and Scalable 3D Reconstruction**
1028 |
1029 | - 论文:http://xxx.itp.ac.cn/abs/2008.00446
1030 |
1031 | - 代码:https://github.com/zlthinker/STBA
1032 |
1033 |
1034 |
1035 | # 图像描述
1036 |
1037 | **Fashion Captioning: Towards Generating Accurate Descriptions with Semantic Rewards**
1038 |
1039 | - 论文:https://arxiv.org/abs/2008.02693
1040 |
1041 | - 代码: https://github.com/xuewyang/Fashion_Captioning
1042 | - 数据集:https://drive.google.com/drive/folders/1J6SZOt_WFwZToX1Jf7QiXzFVwt23lGwW?usp=sharing
1043 |
1044 |
1045 |
1046 | # 图像检索
1047 |
1048 | **SOLAR: Second-Order Loss and Attention for Image Retrieval**
1049 |
1050 | - 论文:https://arxiv.org/abs/2001.08972
1051 |
1052 | - 代码:https://github.com/tonyngjichun/SOLAR
1053 |
1054 | **Self-supervising Fine-grained Region Similarities for Large-scale Image Localization**
1055 |
1056 | - 主页:https://yxgeee.github.io/projects/sfrs
1057 | - 论文:https://arxiv.org/abs/2006.03926
1058 |
1059 | - 代码:https://github.com/yxgeee/SFRS
1060 |
1061 |
1062 |
1063 | # 光流估计
1064 |
1065 | **RAFT: Recurrent All-Pairs Field Transforms for Optical Flow**
1066 |
1067 | - 论文:https://arxiv.org/abs/2003.12039
1068 |
1069 | - 代码:https://github.com/princeton-vl/RAFT
1070 |
1071 | **LiteFlowNet3: Resolving Correspondence Ambiguity for More Accurate Optical Flow Estimation**
1072 |
1073 | - 论文:https://arxiv.org/abs/2007.09319
1074 | - 代码:https://github.com/twhui/LiteFlowNet3
1075 |
1076 |
1077 |
1078 | # 视频插帧
1079 |
1080 | **BMBC: Bilateral Motion Estimation with Bilateral Cost Volume for Video Interpolation**
1081 |
1082 | - 论文:https://arxiv.org/abs/2007.12622
1083 |
1084 | - 代码:https://github.com/JunHeum/BMBC
1085 |
1086 |
1087 |
1088 | # 车道线检测
1089 |
1090 | **CurveLane-NAS: Unifying Lane-Sensitive Architecture Search and Adaptive Point Blending**
1091 |
1092 | - 论文:https://arxiv.org/abs/2007.12147
1093 | - 数据集:https://github.com/xbjxh/curvelanes
1094 |
1095 | **Ultra Fast Structure-aware Deep Lane Detection**
1096 |
1097 | - 论文:https://arxiv.org/abs/2004.11757
1098 |
1099 | - 代码:https://github.com/cfzd/Ultra-Fast-Lane-Detection
1100 | - 论文解读:https://mp.weixin.qq.com/s/TYzDx8R1oUbVr0FxGnFspQ
1101 |
1102 | **Gen-LaneNet: a generalized and scalable approach for 3D lane detection**
1103 |
1104 | - 论文:https://arxiv.org/abs/2003.10656
1105 | - 代码:https://github.com/yuliangguo/Pytorch_Generalized_3D_Lane_Detection
1106 | - 数据集:https://github.com/yuliangguo/3D_Lane_Synthetic_Dataset
1107 |
1108 |
1109 |
1110 | # 轨迹预测
1111 |
1112 | **SimAug: Learning Robust Representations from 3D Simulation for Pedestrian Trajectory Prediction in Unseen Cameras**
1113 |
1114 | - 论文:https://arxiv.org/abs/2004.02022
1115 | - 代码:https://github.com/JunweiLiang/Multiverse
1116 |
1117 |
1118 |
1119 | # 线段检测
1120 |
1121 | **Deep Hough-Transform Line Priors**
1122 |
1123 | - 论文:https://arxiv.org/abs/2007.09493
1124 |
1125 | - 代码:https://github.com/yanconglin/Deep-Hough-Transform-Line-Priors
1126 |
1127 |
1128 |
1129 | # 视线估计
1130 |
1131 | **ETH-XGaze: A Large Scale Dataset for Gaze Estimation under Extreme Head Pose and Gaze Variation**
1132 |
1133 | - 主页:https://ait.ethz.ch/projects/2020/ETH-XGaze
1134 |
1135 | - 论文:https://arxiv.org/abs/2007.15837
1136 |
1137 |
1138 |
1139 | # 眼动追踪
1140 |
1141 | **Towards End-to-end Video-based Eye-Tracking**
1142 |
1143 | - 主页:https://ait.ethz.ch/projects/2020/EVE/
1144 | - 论文:https://arxiv.org/abs/2007.13120
1145 |
1146 |
1147 |
1148 | # 对抗攻击
1149 |
1150 | **Adversarial Ranking Attack and Defense**
1151 |
1152 | - 论文:https://arxiv.org/abs/2002.11293
1153 | - 代码:https://github.com/cdluminate/advrank
1154 |
1155 | **Square Attack: a query-efficient black-box adversarial attack via random search**
1156 |
1157 | - 论文:https://arxiv.org/abs/1912.00049
1158 | - 代码:https://github.com/max-andr/square-attack
1159 |
1160 |
1161 |
1162 | # 数据集
1163 |
1164 | **Long-term Human Motion Prediction with Scene Context**
1165 |
1166 | - 主页:https://people.eecs.berkeley.edu/~zhecao/hmp/index.html
1167 | - 论文:https://arxiv.org/abs/2007.03672
1168 |
1169 | - 数据集:https://github.com/ZheC/GTA-IM-Dataset
1170 |
1171 | **Object Detection with a Unified Label Space from Multiple Datasets**
1172 |
1173 | - 主页:http://www.nec-labs.com/~mas/UniDet/
1174 | - 论文:https://arxiv.org/abs/2008.06614
1175 | - 代码:暂无
1176 | - 数据集:http://www.nec-labs.com/~mas/UniDet/resources/UOD_dataset_ECCV20.zip
1177 |
1178 | **Simulating Content Consistent Vehicle Datasets with Attribute Descent**
1179 |
1180 | - 论文:https://arxiv.org/abs/1912.08855
1181 | - 代码:https://github.com/yorkeyao/VehicleX
1182 | - 数据集:https://github.com/yorkeyao/VehicleX
1183 |
1184 | **InterHand2.6M: A Dataset and Baseline for 3D Interacting Hand Pose Estimation from a Single RGB Image**
1185 |
1186 | - 主页:https://mks0601.github.io/InterHand2.6M/
1187 | - 论文:https://arxiv.org/abs/2008.09309
1188 | - 代码:https://github.com/facebookresearch/InterHand2.6M
1189 |
1190 |
1191 | **SNE-RoadSeg: Incorporating Surface Normal Information into Semantic Segmentation for Accurate Freespace Detection**
1192 |
1193 | - 主页:https://sites.google.com/view/sne-roadseg
1194 | - 论文:https://arxiv.org/abs/2008.11351
1195 | - 代码:https://github.com/hlwang1124/SNE-RoadSeg
1196 | - 数据集:https://drive.google.com/file/d/1YnkqPmzxtjNfMi2B1gMy7LQa5Gnu-BsH/view
1197 |
1198 | **CurveLane-NAS: Unifying Lane-Sensitive Architecture Search and Adaptive Point Blending**
1199 |
1200 | - 论文:https://arxiv.org/abs/2007.12147
1201 | - 数据集:https://github.com/xbjxh/curvelanes
1202 |
1203 | **Detecting natural disasters, damage, and incidents in the wild**
1204 |
1205 | - 主页:http://incidentsdataset.csail.mit.edu/
1206 | - 论文:https://arxiv.org/abs/2008.09188
1207 | - 数据集:https://github.com/ethanweber/IncidentsDataset
1208 |
1209 | **Simultaneous Detection and Tracking with Motion Modelling for Multiple Object Tracking**
1210 |
1211 | - 论文:https://arxiv.org/abs/2008.08826
1212 | - 代码:https://github.com/shijieS/DMMN
1213 | - 数据集:https://github.com/shijieS/OmniMOTDataset
1214 |
1215 | **3D Bird Reconstruction: a Dataset, Model, and Shape Recovery from a Single View**
1216 |
1217 | - 主页:https://marcbadger.github.io/avian-mesh/
1218 | - 论文:https://arxiv.org/abs/2008.06133
1219 | - 代码:https://github.com/marcbadger/avian-mesh
1220 | - 数据集:https://drive.google.com/file/d/1vyXYIJIo9jneIqC7lowB4GVi17rjztjn/view?usp=sharing
1221 |
1222 | **Fashion Captioning: Towards Generating Accurate Descriptions with Semantic Rewards**
1223 |
1224 | - 论文:https://arxiv.org/abs/2008.02693
1225 |
1226 | - 代码: https://github.com/xuewyang/Fashion_Captioning
1227 | - 数据集:https://drive.google.com/drive/folders/1J6SZOt_WFwZToX1Jf7QiXzFVwt23lGwW?usp=sharing
1228 |
1229 | **From Shadow Segmentation to Shadow Removal**
1230 |
1231 | - 论文:http://xxx.itp.ac.cn/abs/2008.00267
1232 |
1233 | - 数据集:https://www3.cs.stonybrook.edu/~cvl/projects/FSS2SR/index.html
1234 |
1235 | **LEMMA: A Multi-view Dataset for Learning Multi-agent Multi-task Activities**
1236 |
1237 | - 主页:https://sites.google.com/view/lemma-activity
1238 |
1239 | - 论文:https://arxiv.org/abs/2007.15781
1240 |
1241 | - 数据集:https://sites.google.com/view/lemma-activity/home/dataset
1242 | - 代码:https://github.com/Buzz-Beater/LEMMA
1243 |
1244 | **Component Divide-and-Conquer for Real-World Image Super-Resolution**
1245 |
1246 | - 论文:https://arxiv.org/abs/2008.01928
1247 |
1248 | - 代码和数据集:https://github.com/xiezw5/Component-Divide-and-Conquer-for-Real-World-Image-Super-Resolution
1249 |
1250 | **Towards End-to-end Video-based Eye-Tracking**
1251 |
1252 | - 主页:https://ait.ethz.ch/projects/2020/EVE/
1253 | - 论文:https://arxiv.org/abs/2007.13120
1254 |
1255 | **Reconstructing NBA Players**
1256 |
1257 | - 主页:http://grail.cs.washington.edu/projects/nba_players/
1258 |
1259 | - 论文:https://arxiv.org/abs/2007.13303
1260 |
1261 | **CelebA-Spoof: Large-Scale Face Anti-Spoofing Dataset with Rich Annotations**
1262 |
1263 | - 论文:https://arxiv.org/abs/2007.12342
1264 |
1265 | - 数据集:https://github.com/Davidzhangyuanhan/CelebA-Spoof
1266 |
1267 | **PIoU Loss: Towards Accurate Oriented Object Detection in Complex Environments**
1268 |
1269 | - 论文:https://arxiv.org/abs/2007.09584
1270 |
1271 | - 代码:https://github.com/clobotics/piou
1272 |
1273 | - 数据集:https://github.com/clobotics/piou
1274 |
1275 | **DanbooRegion: An Illustration Region Dataset**
1276 |
1277 | - 主页:https://lllyasviel.github.io/DanbooRegion/
1278 |
1279 | - 论文:https://lllyasviel.github.io/DanbooRegion/paper/paper.pdf
1280 |
1281 | - 数据集:https://github.com/lllyasviel/DanbooRegion
1282 |
1283 | **Segment as Points for Efficient Online Multi-Object Tracking and Segmentation**
1284 |
1285 | - Oral
1286 | - 论文:https://arxiv.org/abs/2007.01550
1287 | - 代码:https://github.com/detectRecog/PointTrack
1288 | - 数据集:https://github.com/detectRecog/PointTrack
1289 |
1290 | **Gen-LaneNet: a generalized and scalable approach for 3D lane detection**
1291 |
1292 | - 论文:https://arxiv.org/abs/2003.10656
1293 | - 代码:https://github.com/yuliangguo/Pytorch_Generalized_3D_Lane_Detection
1294 | - 数据集:https://github.com/yuliangguo/3D_Lane_Synthetic_Dataset
1295 |
1296 | **TAO: A Large-Scale Benchmark for Tracking Any Object**
1297 |
1298 | - 主页:http://taodataset.org/
1299 | - 论文:https://arxiv.org/abs/2005.10356
1300 | - 代码:https://github.com/TAO-Dataset/tao
1301 |
1302 | **Structured3D: A Large Photo-realistic Dataset for Structured 3D Modeling**
1303 |
1304 | - 主页:[http://structured3d-dataset.org](http://structured3d-dataset.org/)
1305 | - 论文:https://arxiv.org/abs/1908.00222
1306 | - 代码:https://github.com/bertjiazheng/Structured3D
1307 |
1308 | **AiR: Attention with Reasoning Capability**
1309 |
1310 | - 论文:暂无
1311 |
1312 | - 代码:https://github.com/szzexpoi/AiR
1313 | - 数据集:https://github.com/szzexpoi/AiR
1314 |
1315 |
1316 |
1317 | # 其他
1318 |
1319 | **Defocus Blur Detection via Depth Distillation**
1320 |
1321 | - 论文:https://arxiv.org/abs/2007.08113
1322 | - 代码:https://github.com/vinthony/depth-distillation
1323 |
1324 | **Pose Augmentation: Class-agnostic Object Pose Transformation for Object Recognition**
1325 |
1326 | - 论文:https://arxiv.org/abs/2003.08526
1327 |
1328 | - 代码:https://github.com/gyhandy/Pose-Augmentation
1329 |
1330 | **Improving Multispectral Pedestrian Detection by Addressing Modality Imbalance Problems**
1331 |
1332 | - 论文:https://arxiv.org/abs/2008.03043
1333 | - 代码:https://github.com/CalayZhou/MBNet
1334 | - Demo:https://www.bilibili.com/video/BV1Hi4y137aS
1335 |
1336 | **From Shadow Segmentation to Shadow Removal**
1337 |
1338 | 论文:http://xxx.itp.ac.cn/abs/2008.00267
1339 |
1340 | 代码和数据集:https://www3.cs.stonybrook.edu/~cvl/projects/FSS2SR/index.html
1341 |
1342 | **Funnel Activation for Visual Recognition**
1343 |
1344 | - 论文:https://arxiv.org/abs/2007.11824
1345 |
1346 | - 代码:https://github.com/megvii-model/FunnelAct
1347 |
1348 | **Open-Edit: Open-Domain Image Manipulation with Open-Vocabulary Instructions**
1349 |
1350 | - 论文:暂无
1351 | - 代码:https://github.com/xh-liu/Open-Edit
1352 | - Video:https://youtu.be/8E3bwvjCHYE
1353 |
1354 | **Consensus-Aware Visual-Semantic Embedding for Image-Text Matching**
1355 |
1356 | - 论文:https://arxiv.org/abs/2007.08883
1357 | - 代码:https://github.com/BruceW91/CVSE
1358 |
1359 | **Perceiving 3D Human-Object Spatial Arrangements from a Single Image in the Wild**
1360 |
1361 | - 主页:https://jasonyzhang.com/phosa/
1362 | - 论文下载:https://arxiv.org/abs/2007.15649
1363 | - 代码:https://github.com/jasonyzhang/phosa
1364 |
1365 | **AiR: Attention with Reasoning Capability**
1366 |
1367 | - Oral
1368 |
1369 | - 论文:https://arxiv.org/abs/2007.14419
1370 |
1371 | - 代码:https://github.com/szzexpoi/AiR
1372 |
1373 | **Distribution-Balanced Loss for Multi-Label Classification in Long-Tailed Datasets**
1374 |
1375 | - 论文:https://arxiv.org/abs/2007.09654
1376 | - 代码:https://github.com/wutong16/DistributionBalancedLoss
1377 |
1378 | **A Generic Visualization Approach for Convolutional Neural Networks**
1379 |
1380 | - 论文:https://arxiv.org/abs/2007.09748
1381 |
1382 | - 代码:https://github.com/ahmdtaha/constrained_attention_filter
1383 |
1384 | **Deep Plastic Surgery: Robust and Controllable Image Editing with Human-Drawn Sketches**
1385 |
1386 | - 主页:https://williamyang1991.github.io/projects/ECCV2020
1387 | - 论文:https://arxiv.org/abs/2001.02890
1388 | - 代码:https://github.com/TAMU-VITA/DeepPS
1389 |
1390 | **GIQA: Generated Image Quality Assessment**
1391 |
1392 | - 论文:https://arxiv.org/abs/2003.08932
1393 | - 代码:https://github.com/cientgu/GIQA
1394 |
1395 | **Structured3D: A Large Photo-realistic Dataset for Structured 3D Modeling**
1396 |
1397 | - 主页:[http://structured3d-dataset.org](http://structured3d-dataset.org/)
1398 | - 论文:https://arxiv.org/abs/1908.00222
1399 | - 代码:https://github.com/bertjiazheng/Structured3D
1400 |
1401 | **AiR: Attention with Reasoning Capability**
1402 |
1403 | - 论文:暂无
1404 |
1405 | - 代码:https://github.com/szzexpoi/AiR
1406 | - 数据集:https://github.com/szzexpoi/AiR
1407 |
1408 |
1409 |
1410 | # 不确定中没中
1411 |
1412 | **Relation Aware Panoptic Segmentation**
1413 |
1414 | - 论文:暂无
1415 | - 代码:https://github.com/RAPNet/RAP
1416 |
1417 | **Spatial-Angular Interaction for Light Field Image Super-Resolution**
1418 |
1419 | - 论文:暂无
1420 | - 代码:https://github.com/YingqianWang/LF-InterNet
1421 |
1422 | **TVR: A Large-Scale Dataset for Video-Subtitle Moment Retrieval**
1423 |
1424 | - 论文:https://arxiv.org/abs/2001.09099
1425 | - 代码:https://github.com/jayleicn/TVRetrieval
1426 | - 代码:https://github.com/jayleicn/TVCaption
1427 |
1428 | **Self-supervising Fine-grained Region Similarities for IBL**
1429 |
1430 | - 论文:暂无
1431 | - 代码: https://github.com/ID2191/ECCV2020
1432 |
1433 | https://github.com/lelechen63/eccv2020
1434 |
1435 |
--------------------------------------------------------------------------------
/README.md:
--------------------------------------------------------------------------------
1 | # ECCV 2024 论文和开源项目合集(Papers with Code)
2 |
3 | ECCV 2024 decisions are now available!
4 |
5 |
6 | > 注1:欢迎各位大佬提交issue,分享ECCV 2024论文和开源项目!
7 | >
8 | > 注2:关于往年CV顶会论文以及其他优质CV论文和大盘点,详见: https://github.com/amusi/daily-paper-computer-vision
9 | >
10 | > - [CVPR 2024](https://github.com/amusi/CVPR2024-Papers-with-Code)
11 | > - [ECCV 2022](ECCV2022-Papers-with-Code.md)
12 | > - [ECCV 2020](ECCV2020-Papers-with-Code.md)
13 |
14 | 想看ECCV 2024和最新最全的顶会工作,欢迎扫码加入【CVer学术交流群】,这是最大的计算机视觉AI知识星球!每日更新,第一时间分享最新最前沿的计算机视觉、深度学习、自动驾驶、医疗影像和AIGC等方向的学习资料,学起来!
15 |
16 | 
17 |
18 | # 【ECCV 2024 论文开源目录】
19 |
20 | - [3DGS(Gaussian Splatting)](#3DGS)
21 | - [Mamba / SSM)](#Mamba)
22 | - [Avatars](#Avatars)
23 | - [Backbone](#Backbone)
24 | - [CLIP](#CLIP)
25 | - [MAE](#MAE)
26 | - [Embodied AI](#Embodied-AI)
27 | - [GAN](#GAN)
28 | - [GNN](#GNN)
29 | - [多模态大语言模型(MLLM)](#MLLM)
30 | - [大语言模型(LLM)](#LLM)
31 | - [NAS](#NAS)
32 | - [OCR](#OCR)
33 | - [NeRF](#NeRF)
34 | - [DETR](#DETR)
35 | - [Prompt](#Prompt)
36 | - [扩散模型(Diffusion Models)](#Diffusion)
37 | - [ReID(重识别)](#ReID)
38 | - [长尾分布(Long-Tail)](#Long-Tail)
39 | - [Vision Transformer](#Vision-Transformer)
40 | - [视觉和语言(Vision-Language)](#VL)
41 | - [自监督学习(Self-supervised Learning)](#SSL)
42 | - [数据增强(Data Augmentation)](#DA)
43 | - [目标检测(Object Detection)](#Object-Detection)
44 | - [异常检测(Anomaly Detection)](#Anomaly-Detection)
45 | - [目标跟踪(Visual Tracking)](#VT)
46 | - [语义分割(Semantic Segmentation)](#Semantic-Segmentation)
47 | - [实例分割(Instance Segmentation)](#Instance-Segmentation)
48 | - [全景分割(Panoptic Segmentation)](#Panoptic-Segmentation)
49 | - [医学图像(Medical Image)](#MI)
50 | - [医学图像分割(Medical Image Segmentation)](#MIS)
51 | - [视频目标分割(Video Object Segmentation)](#VOS)
52 | - [视频实例分割(Video Instance Segmentation)](#VIS)
53 | - [参考图像分割(Referring Image Segmentation)](#RIS)
54 | - [图像抠图(Image Matting)](#Matting)
55 | - [图像编辑(Image Editing)](#Image-Editing)
56 | - [Low-level Vision](#LLV)
57 | - [超分辨率(Super-Resolution)](#SR)
58 | - [去噪(Denoising)](#Denoising)
59 | - [去模糊(Deblur)](#Deblur)
60 | - [自动驾驶(Autonomous Driving)](#Autonomous-Driving)
61 | - [3D点云(3D Point Cloud)](#3D-Point-Cloud)
62 | - [3D目标检测(3D Object Detection)](#3DOD)
63 | - [3D语义分割(3D Semantic Segmentation)](#3DSS)
64 | - [3D目标跟踪(3D Object Tracking)](#3D-Object-Tracking)
65 | - [3D语义场景补全(3D Semantic Scene Completion)](#3DSSC)
66 | - [3D配准(3D Registration)](#3D-Registration)
67 | - [3D人体姿态估计(3D Human Pose Estimation)](#3D-Human-Pose-Estimation)
68 | - [3D人体Mesh估计(3D Human Mesh Estimation)](#3D-Human-Pose-Estimation)
69 | - [医学图像(Medical Image)](#Medical-Image)
70 | - [图像生成(Image Generation)](#Image-Generation)
71 | - [视频生成(Video Generation)](#Video-Generation)
72 | - [3D生成(3D Generation)](#3D-Generation)
73 | - [视频理解(Video Understanding)](#Video-Understanding)
74 | - [行为识别(Action Recognition)](#Action-Recognition)
75 | - [行为检测(Action Detection)](#Action-Detection)
76 | - [文本检测(Text Detection)](#Text-Detection)
77 | - [知识蒸馏(Knowledge Distillation)](#KD)
78 | - [模型剪枝(Model Pruning)](#Pruning)
79 | - [图像压缩(Image Compression)](#IC)
80 | - [三维重建(3D Reconstruction)](#3D-Reconstruction)
81 | - [深度估计(Depth Estimation)](#Depth-Estimation)
82 | - [轨迹预测(Trajectory Prediction)](#TP)
83 | - [车道线检测(Lane Detection)](#Lane-Detection)
84 | - [图像描述(Image Captioning)](#Image-Captioning)
85 | - [视觉问答(Visual Question Answering)](#VQA)
86 | - [手语识别(Sign Language Recognition)](#SLR)
87 | - [视频预测(Video Prediction)](#Video-Prediction)
88 | - [新视点合成(Novel View Synthesis)](#NVS)
89 | - [Zero-Shot Learning(零样本学习)](#ZSL)
90 | - [立体匹配(Stereo Matching)](#Stereo-Matching)
91 | - [特征匹配(Feature Matching)](#Feature-Matching)
92 | - [场景图生成(Scene Graph Generation)](#SGG)
93 | - [计数(Counting)](#Counting)
94 | - [隐式神经表示(Implicit Neural Representations)](#INR)
95 | - [图像质量评价(Image Quality Assessment)](#IQA)
96 | - [视频质量评价(Video Quality Assessment)](#Video-Quality-Assessment)
97 | - [数据集(Datasets)](#Datasets)
98 | - [新任务(New Tasks)](#New-Tasks)
99 | - [其他(Others)](#Others)
100 |
101 |
102 |
103 | # 3DGS(Gaussian Splatting)
104 |
105 | **MVSplat: Efficient 3D Gaussian Splatting from Sparse Multi-View Images**
106 |
107 | - Project: https://donydchen.github.io/mvsplat
108 | - Paper: https://arxiv.org/abs/2403.14627
109 | - Code:https://github.com/donydchen/mvsplat
110 |
111 | **CityGaussian: Real-time High-quality Large-Scale Scene Rendering with Gaussians**
112 |
113 | - Paper: https://arxiv.org/abs/2404.01133
114 | - Code: https://github.com/DekuLiuTesla/CityGaussian
115 |
116 | **FSGS: Real-Time Few-shot View Synthesis using Gaussian Splatting**
117 |
118 | - Project: https://zehaozhu.github.io/FSGS/
119 | - Paper: https://arxiv.org/abs/2312.00451
120 | - Code: https://github.com/VITA-Group/FSGS
121 |
122 |
123 |
124 |
125 |
126 | # Mamba / SSM
127 |
128 | **VideoMamba: State Space Model for Efficient Video Understanding**
129 |
130 | - Paper: https://arxiv.org/abs/2403.06977
131 | - Code: https://github.com/OpenGVLab/VideoMamba
132 |
133 | **ZIGMA: A DiT-style Zigzag Mamba Diffusion Model**
134 |
135 | - Paper: https://arxiv.org/abs/2403.13802
136 | - Code: https://taohu.me/zigma/
137 |
138 |
139 |
140 | # Avatars
141 |
142 |
143 |
144 |
145 |
146 |
147 |
148 | # Backbone
149 |
150 |
151 |
152 |
153 |
154 | # CLIP
155 |
156 |
157 |
158 |
159 |
160 |
161 |
162 | # MAE
163 |
164 |
165 |
166 | # Embodied AI
167 |
168 |
169 |
170 |
171 |
172 | # GAN
173 |
174 |
175 |
176 | # OCR
177 |
178 | **Bridging Synthetic and Real Worlds for Pre-training Scene Text Detectors**
179 |
180 | - Paper: https://arxiv.org/pdf/2312.05286
181 |
182 | - Code: https://github.com/SJTU-DeepVisionLab/FreeReal
183 |
184 | **PosFormer: Recognizing Complex Handwritten Mathematical Expression with Position Forest Transformer**
185 |
186 | - Paper: https://arxiv.org/abs/2407.07764
187 | - Code: https://github.com/SJTU-DeepVisionLab/PosFormer
188 |
189 |
190 |
191 | # Occupancy
192 |
193 | **Fully Sparse 3D Occupancy Prediction**
194 |
195 | - Paper: https://arxiv.org/abs/2312.17118
196 | - Code: https://github.com/MCG-NJU/SparseOcc
197 |
198 |
199 |
200 |
201 |
202 | # NeRF
203 |
204 | **NeRF-MAE: Masked AutoEncoders for Self-Supervised 3D Representation Learning for Neural Radiance Fields**
205 |
206 | - Project: https://nerf-mae.github.io/
207 | - Paper: https://arxiv.org/pdf/2404.01300
208 | - Code: https://github.com/zubair-irshad/NeRF-MAE
209 |
210 |
211 |
212 | # DETR
213 |
214 |
215 |
216 |
217 |
218 | # Prompt
219 |
220 |
221 |
222 | # 多模态大语言模型(MLLM)
223 |
224 | **SQ-LLaVA: Self-Questioning for Large Vision-Language Assistant**
225 |
226 | - Paper: https://arxiv.org/abs/2403.11299
227 | - Code: https://github.com/heliossun/SQ-LLaVA
228 |
229 | **ControlCap: Controllable Region-level Captioning**
230 |
231 | - Paper: https://arxiv.org/abs/2401.17910
232 | - Code: https://github.com/callsys/ControlCap
233 |
234 |
235 |
236 | # 大语言模型(LLM)
237 |
238 |
239 |
240 |
241 |
242 | # NAS
243 |
244 |
245 |
246 | # ReID(重识别)
247 |
248 |
249 |
250 |
251 |
252 | # 扩散模型(Diffusion Models)
253 |
254 | **ZIGMA: A DiT-style Zigzag Mamba Diffusion Model**
255 |
256 | - Paper: https://arxiv.org/abs/2403.13802
257 | - Code: https://taohu.me/zigma/
258 |
259 | **Skews in the Phenomenon Space Hinder Generalization in Text-to-Image Generation**
260 |
261 | - Paper: https://arxiv.org/abs/2403.16394
262 | - Code: https://github.com/zdxdsw/skewed_relations_T2I
263 |
264 | **The Lottery Ticket Hypothesis in Denoising: Towards Semantic-Driven Initialization**
265 |
266 | - Project: https://ut-mao.github.io/noise.github.io/
267 | - Paper: https://arxiv.org/abs/2312.08872
268 | - Code: https://github.com/UT-Mao/Initial-Noise-Construction
269 |
270 |
271 |
272 | # Vision Transformer
273 |
274 | **GiT: Towards Generalist Vision Transformer through Universal Language Interface**
275 |
276 | - Paper: https://arxiv.org/abs/2403.09394
277 | - Code: https://github.com/Haiyang-W/GiT
278 |
279 |
280 |
281 | # 视觉和语言(Vision-Language)
282 |
283 | **GalLoP: Learning Global and Local Prompts for Vision-Language Models**
284 |
285 | - Paper:https://arxiv.org/abs/2407.01400
286 |
287 |
288 |
289 | # 目标检测(Object Detection)
290 |
291 | **Relation DETR: Exploring Explicit Position Relation Prior for Object Detection**
292 |
293 | - Paper: https://arxiv.org/abs/2407.11699v1
294 | - Code: https://github.com/xiuqhou/Relation-DETR
295 | - Dataset: https://huggingface.co/datasets/xiuqhou/SA-Det-100k
296 |
297 | **Cross-Domain Few-Shot Object Detection via Enhanced Open-Set Object Detector**
298 |
299 | - Project: http://yuqianfu.com/CDFSOD-benchmark/
300 | - Paper: https://arxiv.org/pdf/2402.03094
301 | - Code: https://github.com/lovelyqian/CDFSOD-benchmark
302 |
303 |
304 |
305 | # 异常检测(Anomaly Detection)
306 |
307 |
308 |
309 |
310 |
311 | # 目标跟踪(Object Tracking)
312 |
313 |
314 |
315 |
316 |
317 |
318 |
319 | # 语义分割(Semantic Segmentation)
320 |
321 | **Context-Guided Spatial Feature Reconstruction for Efficient Semantic Segmentation**
322 |
323 | - Paper: https://arxiv.org/abs/2405.06228
324 |
325 | - Code: https://github.com/nizhenliang/CGRSeg
326 |
327 |
328 |
329 | # 医学图像(Medical Image)
330 |
331 | **Brain-ID: Learning Contrast-agnostic Anatomical Representations for Brain Imaging**
332 |
333 | - Paper: https://arxiv.org/abs/2311.16914
334 | - Code: https://github.com/peirong26/Brain-ID
335 |
336 | **FairDomain: Achieving Fairness in Cross-Domain Medical Image Segmentation and Classification**
337 |
338 | - Project: https://ophai.hms.harvard.edu/datasets/harvard-fairdomain20k
339 | - Paper : https://arxiv.org/abs/2407.08813
340 | - Dataset: https://drive.google.com/drive/u/1/folders/1huH93JVeXMj9rK6p1OZRub868vv0UK0O
341 | - Code: https://github.com/Harvard-Ophthalmology-AI-Lab/FairDomain
342 |
343 |
344 |
345 | # 医学图像分割(Medical Image Segmentation)
346 |
347 | **ScribblePrompt: Fast and Flexible Interactive Segmentation for Any Biomedical Image**
348 |
349 | - Project: https://scribbleprompt.csail.mit.edu/
350 | - Paper: https://arxiv.org/abs/2312.07381
351 | - Code: https://github.com/halleewong/ScribblePrompt
352 |
353 | **AnatoMask: Enhancing Medical Image Segmentation with Reconstruction-guided Self-masking**
354 |
355 | - Paper: https://arxiv.org/abs/2407.06468
356 | - Code: https://github.com/ricklisz/AnatoMask
357 |
358 | **Representing Topological Self-Similarity Using Fractal Feature Maps for Accurate Segmentation of Tubular Structures**
359 |
360 | - Paper: https://arxiv.org/abs/2407.14754
361 | - Code: https://github.com/cbmi-group/FFM-Multi-Decoder-Network
362 |
363 |
364 |
365 | # 视频目标分割(Video Object Segmentation)
366 |
367 | **DVIS-DAQ: Improving Video Segmentation via Dynamic Anchor Queries**
368 |
369 | - Project: https://zhang-tao-whu.github.io/projects/DVIS_DAQ/
370 | - Paper: https://arxiv.org/abs/2404.00086
371 | - Code: https://github.com/zhang-tao-whu/DVIS_Plus
372 |
373 |
374 |
375 | # 自动驾驶(Autonomous Driving)
376 |
377 | **Fully Sparse 3D Occupancy Prediction**
378 |
379 | - Paper: https://arxiv.org/abs/2312.17118
380 | - Code: https://github.com/MCG-NJU/SparseOcc
381 |
382 | **milliFlow: Scene Flow Estimation on mmWave Radar Point Cloud for Human Motion Sensing**
383 |
384 | - Paper: https://arxiv.org/abs/2306.17010
385 | - Code: https://github.com/Toytiny/milliFlow/
386 |
387 | **4D Contrastive Superflows are Dense 3D Representation Learners**
388 |
389 | - Paper : https://arxiv.org/abs/2407.06190
390 | - Code: https://github.com/Xiangxu-0103/SuperFlow
391 |
392 |
393 |
394 | # 3D点云(3D-Point-Cloud)
395 |
396 |
397 |
398 |
399 |
400 | # 3D目标检测(3D Object Detection)
401 |
402 | **3D Small Object Detection with Dynamic Spatial Pruning**
403 |
404 | - Project: https://xuxw98.github.io/DSPDet3D/
405 | - Paper: https://arxiv.org/abs/2305.03716
406 | - Code: https://github.com/xuxw98/DSPDet3D
407 |
408 | **Ray Denoising: Depth-aware Hard Negative Sampling for Multi-view 3D Object Detection**
409 |
410 | - Paper: https://arxiv.org/abs/2402.03634
411 | - Code: https://github.com/LiewFeng/RayDN
412 |
413 |
414 |
415 | # 3D语义分割(3D Semantic Segmentation)
416 |
417 |
418 |
419 | # 图像编辑(Image Editing)
420 |
421 |
422 |
423 |
424 |
425 |
426 |
427 | # 图像补全/图像修复(Image Inpainting)
428 |
429 | **BrushNet: A Plug-and-Play Image Inpainting Model with Decomposed Dual-Branch Diffusion**
430 |
431 | - Project https://tencentarc.github.io/BrushNet/
432 | - Paper: https://arxiv.org/abs/2403.06976
433 | - Code: https://github.com/TencentARC/BrushNet
434 |
435 |
436 |
437 | # 视频编辑(Video Editing)
438 |
439 |
440 |
441 |
442 |
443 | # Low-level Vision
444 |
445 | **Restoring Images in Adverse Weather Conditions via Histogram Transformer**
446 |
447 | - Paper: https://arxiv.org/abs/2407.10172
448 | - Code: https://github.com/sunshangquan/Histoformer
449 |
450 | **OneRestore: A Universal Restoration Framework for Composite Degradation**
451 |
452 | - Project https://gy65896.github.io/projects/ECCV2024_OneRestore
453 | - Paper: https://arxiv.org/abs/2407.04621
454 | - Code: https://github.com/gy65896/OneRestore
455 |
456 | # 超分辨率(Super-Resolution)
457 |
458 |
459 |
460 |
461 |
462 | # 去噪(Denoising)
463 |
464 | ## 图像去噪(Image Denoising)
465 |
466 |
467 |
468 | # 3D人体姿态估计(3D Human Pose Estimation)
469 |
470 |
471 |
472 |
473 |
474 | # 图像生成(Image Generation)
475 |
476 | **Object-Conditioned Energy-Based Attention Map Alignment in Text-to-Image Diffusion Models**
477 |
478 | - Paper: https://arxiv.org/abs/2404.07389
479 | - Code: https://github.com/YasminZhang/EBAMA
480 |
481 | **Every Pixel Has its Moments: Ultra-High-Resolution Unpaired Image-to-Image Translation via Dense Normalization**
482 |
483 | - Project: https://kaminyou.com/Dense-Normalization/
484 | - Paper: https://arxiv.org/abs/2407.04245
485 | - Code: https://github.com/Kaminyou/Dense-Normalization
486 |
487 | **ZIGMA: A DiT-style Zigzag Mamba Diffusion Model**
488 |
489 | - Paper: https://arxiv.org/abs/2403.13802
490 | - Code: https://taohu.me/zigma/
491 |
492 | **Skews in the Phenomenon Space Hinder Generalization in Text-to-Image Generation**
493 |
494 | - Paper: https://arxiv.org/abs/2403.16394
495 | - Code: https://github.com/zdxdsw/skewed_relations_T2I
496 |
497 |
498 |
499 | # 视频生成(Video Generation)
500 |
501 | **VideoStudio: Generating Consistent-Content and Multi-Scene Videos**
502 |
503 | - Project: https://vidstudio.github.io/
504 | - Code: https://github.com/FuchenUSTC/VideoStudio
505 |
506 |
507 |
508 |
509 |
510 | # 3D生成
511 |
512 |
513 |
514 |
515 |
516 | # 视频理解(Video Understanding)
517 |
518 | **VideoMamba: State Space Model for Efficient Video Understanding**
519 |
520 | - Paper: https://arxiv.org/abs/2403.06977
521 | - Code: https://github.com/OpenGVLab/VideoMamba
522 |
523 | **C2C: Component-to-Composition Learning for Zero-Shot Compositional Action Recognition**
524 |
525 | - Paper: https://arxiv.org/abs/2407.06113
526 | - Code: https://github.com/RongchangLi/ZSCAR_C2C
527 |
528 |
529 |
530 | # 行为识别(Action Recognition)
531 |
532 | **SA-DVAE: Improving Zero-Shot Skeleton-Based Action Recognition by Disentangled Variational Autoencoders**
533 |
534 | - Paper: https://arxiv.org/abs/2407.13460
535 | - Code: https://github.com/pha123661/SA-DVAE
536 |
537 |
538 |
539 | # 知识蒸馏(Knowledge Distillation)
540 |
541 |
542 |
543 | # 图像压缩(Image Compression)
544 |
545 | **Image Compression for Machine and Human Vision With Spatial-Frequency Adaptation**
546 |
547 | - Code: https://github.com/qingshi9974/ECCV2024-AdpatICMH
548 | - Paper: http://arxiv.org/abs/2407.09853
549 |
550 |
551 |
552 | # 立体匹配(Stereo Matching)
553 |
554 |
555 |
556 |
557 |
558 | # 场景图生成(Scene Graph Generation)
559 |
560 |
561 |
562 |
563 |
564 | # 计数(Counting)
565 |
566 | **Zero-shot Object Counting with Good Exemplars**
567 |
568 | - Paper: https://arxiv.org/abs/2407.04948
569 | - Code: https://github.com/HopooLinZ/VA-Count
570 |
571 |
572 |
573 |
574 |
575 | # 视频质量评价(Video Quality Assessment)
576 |
577 |
578 |
579 | # 数据集(Datasets)
580 |
581 |
582 |
583 | # 其他(Others)
584 |
585 | **Multi-branch Collaborative Learning Network for 3D Visual Grounding**
586 |
587 | - Paper: https://arxiv.org/abs/2407.05363v2
588 | - Code: https://github.com/qzp2018/MCLN
589 |
590 | **PDiscoFormer: Relaxing Part Discovery Constraints with Vision Transformers**
591 |
592 | - Code: https://github.com/ananthu-aniraj/pdiscoformer
593 | - Paper: https://arxiv.org/abs/2407.04538
594 |
595 | **SPVLoc: Semantic Panoramic Viewport Matching for 6D Camera Localization in Unseen Environments**
596 |
597 | - Project: https://fraunhoferhhi.github.io/spvloc/
598 | - Paper: https://arxiv.org/abs/2404.10527
599 | - Code: https://github.com/fraunhoferhhi/spvloc
600 |
601 | **REFRAME: Reflective Surface Real-Time Rendering for Mobile Devices**
602 |
603 | - Project: https://xdimlab.github.io/REFRAME/
604 | - Paper: https://arxiv.org/abs/2403.16481
605 | - Code: https://github.com/MARVELOUSJI/REFRAME
--------------------------------------------------------------------------------