• Stars
    star
    199
  • Rank 194,916 (Top 4 %)
  • Language
  • Created over 4 years ago
  • Updated about 2 months ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

[CVPR 2020 & 2021 & 2022 & 2023] Agriculture-Vision Dataset, Prize Challenge and Workshop: A joint effort with many great collaborators to bring Agriculture and Computer Vision/AI communities together to benefit humanity!

Agriculture-Vision

Agriculture-Vision Dataset, Challenge and Workshop (CVPR 2020)

A joint effort with many great collaborators to bring Agriculture and Computer Vision / AI communities together to benefit humanity!

Update:

The 2nd Agriculture-Vision Prize Challenge has started and new deadline is June 5, 2021. Total Prize: 20,000$.

Content

Papers:

CVPR paper on Agriculture-Vision
ArXiv, CVPR 2020 open access

@article{chiu2020agriculture,
  title={Agriculture-Vision: A Large Aerial Image Database for Agricultural Pattern Analysis},
  author={Chiu, Mang Tik and Xu, Xingqian and Wei, Yunchao and Huang, Zilong and Schwing, Alexander and Brunner, Robert and Khachatrian, Hrant and Karapetyan, Hovnatan and Dozier, Ivan and Rose, Greg and others},
  journal={arXiv preprint arXiv:2001.01306},
  year={2020}
}

@InProceedings{Chiu_2020_CVPR,
author = {Chiu, Mang Tik and Xu, Xingqian and Wei, Yunchao and Huang, Zilong and Schwing, Alexander G. and Brunner, Robert and Khachatrian, Hrant and Karapetyan, Hovnatan and Dozier, Ivan and Rose, Greg and Wilson, David and Tudor, Adrian and Hovakimyan, Naira and Huang, Thomas S. and Shi, Honghui},
title = {Agriculture-Vision: A Large Aerial Image Database for Agricultural Pattern Analysis},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2020}
}

CVPR Workshop paper on the 1st Agriculture-Vision Challenge (challenge dataset, methods and results)
ArXiv, CVPR 2020 open access

@article{chiu20201st,
  title={The 1st Agriculture-Vision Challenge: Methods and Results},
  author={Chiu, Mang Tik and Xu, Xingqian and Wang, Kai and Hobbs, Jennifer and Hovakimyan, Naira and Huang, Thomas S. and Shi, Honghui and others},
  journal={arXiv preprint arXiv:2004.09754},
  year={2020}
}

@InProceedings{Chiu_2020_CVPR_Workshops,
author = {Chiu, Mang Tik and Xu, Xingqian and Wang, Kai and Hobbs, Jennifer and Hovakimyan, Naira and Huang, Thomas S. and Shi, Honghui},
title = {The 1st Agriculture-Vision Challenge: Methods and Results},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops},
month = {June},
year = {2020}
}

Dataset

Agriculture-Vision Challenge Dataset

The dataset used in this challenge is a subset of the Agriculture-Vision dataset. The challenge dataset contains 21,061 aerial farmland images captured throughout 2019 across the US. Each image consists of four 512x512 color channels, which are RGB and Near Infra-red (NIR). Each image also has a boundary map and a mask. The boundary map indicates the region of the farmland, and the mask indicates valid pixels in the image. Regions outside of either the boundary map or the mask are not evaluated.

This dataset contains six types of annotations: Cloud shadow, Double plant, Planter skip, Standing Water, Waterway and Weed cluster. These types of field anomalies have great impacts on the potential yield of farmlands, therefore it is extremely important to accurately locate them. In the Agriculture-Vision dataset, these six patterns are stored separately as binary masks due to potential overlaps between patterns. Users are free to decide how to use these annotations.

Each field image has a file name in the format of (field id)_(x1)-(y1)-(x2)-(y2).(jpg/png). Each field id uniquely identifies the farmland that the image is cropped from, and (x1, y1, x2, y2) is a 4-tuple indicating the position in which the image is cropped. Please refer to our paper for more details regarding how we construct the dataset.

Download

The challenge dataset contains images, boundaries and masks for train, val and test set. It also contains labels for the train and val set only. The 2021 new dataset .tar.gz file is around 20 GB. Please visit this page to get access. By downloading or using the dataset, user signifies its agreement to Agriculture-Vision Workshop Terms and Conditions.

Evaluation

Evaluation metric

We use mean Intersection-over-Union (mIoU) as our main quantitative evaluation metric, which is one of the most commonly used measures in semantic segmentation datasets. The mIoU is computed as:

Where c is the number of annotation types (c = 7 in our dataset, with 6 patterns + background), Pc and Tc are the predicted mask and ground truth mask of class c respectively.

Since our annotations may overlap, we modify the canonical mIoU metric to accommodate this property. For pixels with multiple labels, a prediction of either label will be counted as a correct pixel classification for that label, and a prediction that does not contain any ground truth labels will be counted as an incorrect classification for all ground truth labels.

Concretely, we construct the confusion matrix Mc×c with the following rules:

For each prediction x and label set Y:

  1. If x ∈ Y, then My,y = My,y + 1 for each y in Y
  2. Otherwise, Mx,y = Mx,y + 1 for each y in Y

The mIoU is finally computed by (true_positive) / (prediction + target - true_positive), averaged across all classes.

Submission

Update:

  • The 1st Agriculture-Vision Prize Challenge has ended on April 20, 2020 (10:00 AM PDT). The final results as of the end date of the challenge can be found in these links (png, csv).
  • The Codalab evaluation server will remain open to encourage further research in Agriculture-Vision. No registration is required.

Registration

We are now hosting our evaluation/competition on Codalab. The competition page can be found here. Each participating team is required to register for the challenge. To register your team, fill out the registration form here, then register on the competition page.

*Make sure your Codalab account email matches one of the member emails in the registration form. Each team can only register once.

Codalab submission

All registered teams can evaluate their results on Codalab and publish their results on the leaderboard. The submission file should be a compressed .zip file that contains all prediction images. All prediction images should be in png format and the file names and image sizes should match the input images exactly. The prediction images will be converted to a 2D numpy array with the following code:

numpy.array(PIL.Image.open(‘field-id_x1-y1-x2-y2.png’))

In the loaded numpy array, only 0-6 integer labels are allowed, and they represent the annotations in the following way:

0 - background
1 - cloud_shadow
2 - double_plant
3 - planter_skip
4 - standing_water
5 - waterway
6 - weed_cluster

This label order will be strictly followed during evaluation.

All teams can have 2 submissions per day and 20 submissions in total.

Challenge

We hosted the 1st Agriculture-Vision Prize Challenge with the following terms:

  • Model size will be limited below 150M parameters in total.
  • The mIoU derived from the "results/" folder in the final submission should match the mIoU on the leaderboard.
  • Predictions in "results/" in the final submission can be reproduced with the resources in "code/" and "models/".
  • The training process of the method can be reproduced and the retrained model should have a similar performance.
  • The test set is off-limits in training.
  • For fairness, teams need to specify what public datasets are used for training/pre-training their models in their challenge_ report.pdf. Results that are generated from models using private datasets, and results without such details will be excluded from prize evaluation. (Results using private datasets can still be included in the report.)
The prize award will be granted to the top 3 teams on the prize challenge leaderboard (closed on 4/20/2020) that provide a valid final submission.

Teams and Methods

Please refer to our workshop summary paper for more details regarding notable methods. Selected teams & methods:

Team Hyunseong: Hyunseong Park, Junhee Kim, Sungho Kim (Agency for Defense Development, South Korea)

  • Method: Residual DenseNet with Expert Network for Semantic Segmentation

Team SCG Vision: Qinghui Liu, Michael C. Kampffmeyer, Robert Jenssen, Arnt B. Salberg (Norwegian Computing Center, UiT The Arctic University of Norway)

Team AGR: Alexandre Barbosa, Rodrigo Trevisan (University of Illinois at Urbana Champaign)

  • Method: Techniques for Improving Aerial Agricultural Image Semantic Segmentation

Team TJU: Bingchen Zhao, Shaozuo Yu, Siwei Yang, Yin Wang (Tongji University)

Team Haossr: Hao Sheng, Xiao Chen, Jingyi Su, Ram Rajagopal, Andrew Ng (Stanford University, Chegg, Inc.)

  • Method: Effective Data Fusion with Generalized Vegetation Index: Evidence from Land Cover Segmentation in Agriculture

Team CNUPR_TH2L: Van Thong Huynh, Soo-Hyung Kim, In-Seop Na (Chonnam National University, Chosun University)

  • Method: Multi-spectra Attention in Aerial Agricultural Segmentation

Team TeamTiger: Ujjwal Baid, Shubham Innani, Prasad Dutande, Bhakti Baheti, Sanjay Talbar (SGGS Institute of Engineering and Technology)

  • Method: Eff-PN: A Novel Deep Learning Approach for Anomaly Pattern Segmentation in Aerial Agricultural Images

Challenge leaderboard

Team mIoU (%) Background Cloud shadow Double plant Planter skip Standing water Waterway Weed cluster
Hyunseong 63.9 80.6 56.0 57.9 57.5 75.0 63.7 56.9
seungjae 62.2 79.3 44.4 60.4 65.9 76.9 55.4 53.2
yjl912.2 61.5 80.1 53.7 46.1 48.6 76.8 71.5 53.6
ddcm 60.8 80.5 51.0 58.6 49.8 72.0 59.8 53.8
RodrigoTrevisan 60.5 80.2 43.8 57.5 51.6 75.3 66.2 49.2
SYDu 59.5 81.3 41.6 50.3 43.4 73.2 71.7 55.2
agri 59.2 78.2 55.8 42.9 42.0 77.5 64.7 53.2
Tennant 57.4 79.9 36.6 54.8 41.4 69.8 66.9 52.0
celery03.0 55.4 79.1 38.9 43.3 41.2 73.0 61.5 50.5
stevenwudi 55.0 77.4 42.0 54.4 20.1 69.5 67.7 53.8
PAII 55.0 79.9 38.6 47.6 26.2 74.6 62.1 55.7
agrichallenge1.2 54.6 80.9 50.9 39.3 29.2 73.4 57.8 50.5
hui 54.0 80.2 41.6 46.4 20.8 72.8 64.8 51.4
shenchen61.6 53.7 79.4 36.7 56.3 21.6 67.0 61.8 52.8
NTU 53.6 79.8 41.4 49.4 13.5 73.3 61.8 56.0
tpys 53.0 81.1 50.5 37.1 25.9 67.4 58.7 50.1
Simple 52.7 80.2 40.0 45.2 24.6 70.9 57.6 50.4
Ursus 52.3 78.9 36.3 37.8 34.4 69.3 57.1 52.3
liepieshov 52.1 77.2 40.2 46.0 16.0 71.3 62.9 51.1
Lunhao 49.4 79.5 40.4 38.8 10.5 69.4 58.3 49.1
tetelias-mipt 49.2 80.4 37.8 34.8 04.6 70.6 62.5 53.8
Dataloader 48.9 79.1 42.0 35.8 09.1 68.7 56.7 51.3
Hakjin 46.4 78.6 32.0 38.3 01.8 66.2 58.0 49.9
JianyuTANG 44.6 78.1 37.9 31.8 15.4 47.3 54.8 46.9
haossr 43.9 79.2 21.4 28.1 02.7 67.5 56.4 52.3
rpartsey 41.5 72.5 21.6 36.2 09.1 59.7 40.7 50.6
baidujjwal 40.8 75.2 26.1 40.1 09.9 48.0 37.1 49.5
Chaturlal 40.7 77.7 23.0 20.4 05.0 55.0 51.0 52.9
Sciforce 40.2 80.5 29.6 24.4 0.0 41.2 55.9 50.0
MustafaA 40.1 76.5 34.4 25.6 11.1 46.0 36.5 50.3
HaotianYan 36.8 77.1 21.9 25.1 13.7 57.5 24.3 37.9
gro 36.3 76.4 37.5 08.4 0.0 60.3 29.7 41.8
oscmansan 35.5 71.6 29.6 03.0 0.0 52.4 46.2 45.9
ThorstenC 33.6 72.3 22.3 10.0 02.0 40.8 40.1 47.8
ZHwang 33.5 76.5 32.4 12.9 0.0 57.2 15.9 39.9
fayzur2.0 22.1 65.4 21.8 02.2 00.2 23.3 13.4 28.7
gaslen.2 21.5 71.0 03.3 17.9 00.8 10.2 06.9 40.1
dvkhandelwal 16.3 71.5 0.0 0.0 0.0 42.6 0.0 0.0
ajeetsinghiitd 10.3 56.9 00.2 00.4 0.0 0.0 00.1 14.5

Workshop

Schedule

(Sunday, Jun 14 PDT)


08:30 – 08:40 Opening Remarks and Welcome

Humphrey Shi (University of Oregon)

Naira Hovakimyan (UIUC)


08:40 – 09:00 Invited Talk 1: Scaling Spatio-Temporal Analytics: A Case in Agricultural Insights (Talk video here)

Sharath Pankanti (IBM T.J. Watson Research Center)

09:00 – 09:20 Invited Talk 2: Multi-modality Remote Sensing in High Throughput Phenotyping: Opportunities for Machine Learning (Talk video here)

Melba M. Crawford, Edward J. Delp (Purdue)

09:20 – 09:40 Invited Talk 3: Bridging Application and Research. Is agriculture just another application? (Talk video here)

Jennifer Hobbs (IntelinAir Inc.)

09:40 – 10:00 Invited Talk 4: Improving Visual and Speech Recognition on Out-Domain Data (Talk video here)

Liangliang Cao (Google Inc. and Umass)

10:00 – 10:20 Invited Talk 5: Learning to Anticipate (Talk video here)

Alex Schwing (UIUC)


10:20 – 11:10 Poster Session 1


11:10 – 11:20 Oral 1: MSCG-Net with Adaptive Class Weighting Loss for Semantic Segmentation (Talk video here)
11:20 – 11:30 Oral 2: Finding Berries: Segmentation and Counting of Cranberries using Point Supervision and Shape Priors (Talk video here)
11:30 – 11:40 Oral 3: Visual 3D Reconstruction and Dynamic Simulation of Fruit Trees for Robotic Manipulation (Talk video here)
11:40 – 11:50 Oral 4: Cross-Regional Oil Palm Tree Detection (Talk video here)
11:50 – 11:00 Oral 5: Effective Data Fusion with Generalized Vegetation Index (Talk video here)
11:00 – 12:10 Oral 6: Weakly Supervised Learning Guided by Activation Mapping Applied to a Novel Citrus Pest Benchmark (Talk video here)
12:10 – 12:20 Oral 7: Fine-Grained Recognition in High-throughput Phenotyping (Talk video here)
12:20 – 12:30 Oral 8: Climate Adaptation: Reliably Predicting from Imbalanced Satellite Data (Talk video here)


12:30 – 13:30 Lunch Break


13:30 – 13:50 Invited Talk 6: Towards Understanding Agricultural Systems at Scale using Machine Learning (Talk video here)

Stefano Ermon (Stanford)

13:50 – 14:10 Invited Talk 7: Interactive Object Segmentation (Live*) (Talk video here)

Yunchao Wei (University of Technology Sydney)

14:10 – 14:30 Invited Talk 8: Project FarmBeats: Aerial Mapping forAgricultural Farms and Beyond (Live*) (Talk video here)

Ranveer Chandra, Sudipta Sinha (Microsoft)

14:30 – 14:50 Invited Talk 9: Enabling the African Farmer (Live*) (Talk video here)

Munther Dahleh (MIT)

14:50 – 15:10 Invited Talk 10: (Live*)

Zsolt Kira (Georgia Tech)


15:10 – 15:55 Poster Session 2


15:55 – 16:10 Agriculture-Vision Prize Challenge Results Summary (Talk video here)

16:10 – 16:20 Challenge Presentation 1: Residual DenseNet with Expert Network for Semantic Segmentation

16:20 – 16:30 Challenge Presentation 2: MSCG-Net Models for The 1st Agriculture-Vision Challenge


16:30 – 17:20 Panel Discussion: Challenge & Opportunities for Computer Vision in Agriculture (video here)

Host:
Naira Hovakimyan (UIUC)

Panelists:
Sharath Pankanti (IBM T. J. Watson Research Center)

Liangliang Cao (Google Research)

Ranveer Chandra (Microsoft Azure Global)

Sudipta Sinha (Microsoft)

Munther Dahleh (MIT)

Ed Delp (Purdue)

Melba Crawford (Purdue)

Jennifer Hobbs (Intelinair Inc.)

Yunchao Wei (University of Technology Sydney)

Alexander Schwing (UIUC)

Zsolt Kira (Georgia Tech)

David Wilson (Intelinair Inc.)

Jim Yuan (Google X)

Humphrey Shi (U of Oregon)


17:20 – 17:30 Closing Remarks

Humphrey Shi (University of Oregon)

Naira Hovakimyan (UIUC)


Paper Track

  • We accept papers related to Agriculture-Vision with a rigorous peer-review process with program committee members from multiple research communities including computer vision, machine learning, image processing, remote sensing, agriculture etc.
  • All accepted paper will be published in IEEE CVPR 2020 Workshop Proceedings.
  • Several top accepted papers will be given oral presentation opportunity at our workshop, the rest accepted papers will be given poster presentation slots during the workshop.

Topic Description

The Workshop will be open to the broader community addressing various technical and application aspects of challenges and opportunities in computer vision research for agriculture. We aim to provide a venue to both show current relevant efforts in interdisciplinary areas between computer vision and agriculture, and to encourage further research and conversations within the computer vision community to tackle impactful agriculture-vision problems. We will solicit papers from the following wide range of topics:

  • Computer vision research for agricultural imagery in general
  • Resources and benchmarks for agricultural imagery based pattern analysis
  • Farmland pattern classification and segmentation from LSHR aerial imagery
  • Efficient data sampling methods for effective training on agricultural imageries
  • Effective data fusion of multi-spectral image data such as RGB and Near-Infrared (RGB-NIR)
  • Robust and stable pattern recognition on sparse and imbalanced annotations
  • Other novel vision applications in agriculture

Accepted Papers

Accepted Posters

  • Scoring Root Necrosis in Cassava Using Semantic Segmentation,
    Joyce Nakatumba-Nabende, Jeremy Tusubira, Benjamin Akera, Ernest Mwebaze (Makerere University, Google)

  • Active Learning Sampling Strategies For Weed Segmentation,
    Andrei Polzounov, Anshul Tibrewal, Ben Cline, Olgert Denas, Chris Padwick, Jim Ostrowski (Blue River Technology)

  • Convolutional Neural Networks for Image-based Corn Kernel Detection and Counting,
    Saeed Khaki, Hieu Pham, Ye Han, Andy Kuhl, Wade Kent, Lizhi Wang (Iowa State University, Syngenta)

  • Eff-PN: A Novel Deep Learning Approach for Anomaly Pattern Segmentation in Aerial Agricultural Images,
    Shubham Innani, Prasad Dutande, Ujjwal R Baid, Bhakti Vilas Baheti, Sanjay Talbar (SGGSIET Nanded)

  • Deep Unsupervised Deblurring Approach for Improving Crops Disease Classification,
    Javed Ahmad, Fahad Shamshad, Junaid Maqbool, Ali Ahmed (Information Technology University)

  • NDVI Vegetation Index Estimation Through an\ Unsupervised Deep Learning based Approach,
    Patricia L Suarez, Angel Sappa, Boris Vintimilla (ESPOL POLYTECHNIC UNIVERSITY)

  • Plants Disease Detection in Low Data Regime,
    Shruti Jadon (Brown University (Rhode Island Hospital))

  • Semi-Supervised Crop Detection in UAV Orthophotographs,
    Jana Wieme, Sam Leroux, Simon Cool, Ruben Van De Vijver, Wouter H. Maes, Jan Pieters (Ghent University)

  • Counting Corn Stands using Drone Images from Cloud to Edge –A Deep Learning-based Object Detection Approach,
    Zhiyi Li, Kshitiz Dhakal, Song Li (Virginia Tech)

More Repositories

1

OneFormer

OneFormer: One Transformer to Rule Universal Image Segmentation, arxiv 2022 / CVPR 2023
Jupyter Notebook
1,391
star
2

Versatile-Diffusion

Versatile Diffusion: Text, Images and Variations All in One Diffusion Model, arXiv 2022 / ICCV 2023
Python
1,300
star
3

Neighborhood-Attention-Transformer

Neighborhood Attention Transformer, arxiv 2022 / CVPR 2023. Dilated Neighborhood Attention Transformer, arxiv 2022
Python
1,037
star
4

Prompt-Free-Diffusion

Prompt-Free Diffusion: Taking "Text" out of Text-to-Image Diffusion Models, arxiv 2023 / CVPR 2024
Python
719
star
5

Matting-Anything

Matting Anything Model (MAM), an efficient and versatile framework for estimating the alpha matte of any instance in an image with flexible and interactive visual or linguistic user prompt guidance.
Python
595
star
6

Compact-Transformers

Escaping the Big Data Paradigm with Compact Transformers, 2021 (Train your Vision Transformers in 30 mins on CIFAR-10 with a single GPU!)
Python
487
star
7

Cross-Scale-Non-Local-Attention

PyTorch code for our paper "Image Super-Resolution with Cross-Scale Non-Local Attention and Exhaustive Self-Exemplars Mining" (CVPR2020).
Python
401
star
8

Pyramid-Attention-Networks

[IJCV] Pyramid Attention Networks for Image Restoration: new SOTA results on multiple image restoration tasks: denoising, demosaicing, compression artifact reduction, super-resolution
Python
382
star
9

NATTEN

Neighborhood Attention Extension. Bringing attention to a neighborhood near you!
Cuda
333
star
10

Smooth-Diffusion

Smooth Diffusion: Crafting Smooth Latent Spaces in Diffusion Models arXiv 2023 / CVPR 2024
Python
287
star
11

VCoder

VCoder: Versatile Vision Encoders for Multimodal Large Language Models, arXiv 2023 / CVPR 2024
Python
245
star
12

Rethinking-Text-Segmentation

[CVPR 2021] Rethinking Text Segmentation: A Novel Dataset and A Text-Specific Refinement Approach
Python
241
star
13

Self-Similarity-Grouping

Self-similarity Grouping: A Simple Unsupervised Cross Domain Adaptation Approach for Person Re-identification (ICCV 2019, Oral)
Python
188
star
14

FcF-Inpainting

[WACV 2023] Keys to Better Image Inpainting: Structure and Texture Go Hand in Hand
Jupyter Notebook
169
star
15

Decoupled-Classification-Refinement

Revisiting RCNN: On Awakening the Classification Power of Faster RCNN (ECCV 2018)
Python
167
star
16

Convolutional-MLPs

[Preprint] ConvMLP: Hierarchical Convolutional MLPs for Vision, 2021
Python
163
star
17

3D-Point-Cloud-Learning

130
star
18

CuMo

CuMo: Scaling Multimodal LLM with Co-Upcycled Mixture-of-Experts
Python
130
star
19

Forget-Me-Not

Forget-Me-Not: Learning to Forget in Text-to-Image Diffusion Models, 2023
Python
107
star
20

VMFormer

[Preprint] VMFormer: End-to-End Video Matting with Transformer
Python
106
star
21

Semi-Supervised-Transfer-Learning

[CVPR 2021] Adaptive Consistency Regularization for Semi-Supervised Transfer Learning
Jupyter Notebook
101
star
22

SGL-Retinal-Vessel-Segmentation

[MICCAI 2021] Study Group Learning: Improving Retinal Vessel Segmentation Trained with Noisy Labels: New SOTA on both DRIVE and CHASE_DB1.
Jupyter Notebook
101
star
23

StyleNAT

New flexible and efficient image generation framework that sets new SOTA on FFHQ-256 with FID 2.05, 2022
Python
97
star
24

Unsupervised-Domain-Adaptation-with-Differential-Treatment

[CVPR 2020] Differential Treatment for Stuff and Things: A Simple Unsupervised Domain Adaptation Method for Semantic Segmentation
Python
88
star
25

Text2Video-Zero-sd-webui

Python
79
star
26

GFR-DSOD

Improving Object Detection from Scratch via Gated Feature Reuse (BMVC 2019)
Python
65
star
27

SH-GAN

[WACV 2023] Image Completion with Heterogeneously Filtered Spectral Hints
Python
62
star
28

VIM

Python
54
star
29

UltraSR-Arbitrary-Scale-Super-Resolution

[Preprint] UltraSR: Spatial Encoding is a Missing Key for Implicit Image Function-based Arbitrary-Scale Super-Resolution, 2021
53
star
30

Any-Precision-DNNs

Any-Precision Deep Neural Networks (AAAI 2021)
Python
44
star
31

Horizontal-Pyramid-Matching

Horizontal Pyramid Matching for Person Re-identification (AAAI 2019)
Python
39
star
32

Pseudo-IoU-for-Anchor-Free-Object-Detection

Pseudo-IoU: Improving Label Assignment in Anchor-Free Object Detection
Python
30
star
33

Human-Object-Interaction-Detection

25
star
34

CompFeat-for-Video-Instance-Segmentation

CompFeat: Comprehensive Feature Aggregation for Video Instance Segmentation (AAAI 2021)
19
star
35

OneFormer-Colab

[Colab Demo Code] OneFormer: One Transformer to Rule Universal Image Segmentation.
Python
13
star
36

DiSparse-Multitask-Model-Compression

[CVPR 2022] DiSparse: Disentangled Sparsification for Multitask Model Compression
Jupyter Notebook
13
star
37

Diffusion-Driven-Test-Time-Adaptation-via-Synthetic-Domain-Alignment

Everything to the Synthetic: Diffusion-driven Test-time Adaptation via Synthetic-Domain Alignment
Python
12
star
38

Interpretable-Visual-Reasoning

[ICCV 2021] Interpretable Visual Reasoning via Induced Symbolic Space
9
star
39

Mask-Selection-Networks

[CVPR 2021] Youtube-VIS 2021 3rd place, [CVPR 2020] winner DAVIS 2020. Code for mask selection based methods.
6
star
40

Activity-Recognition

5
star
41

Boosted-Dynamic-Networks

Boosted Dynamic Neural Networks, AAAI 2023
Python
4
star
42

Aneurysm-Segmentation-with-Multi-Teacher-Pseudo-Labels

1
star