Pytorch implementation of OCNet series and SegFix.

Overview

openseg.pytorch

PWC

PWC

PWC

PWC

PWC

News

  • 2021/09/14 MMSegmentation has supported our ISANet and refer to ISANet for more details.

  • 2021/08/13 We have released the implementation for HRFormer and the combination of HRFormer and OCR achieves better semantic segmentation performance.

  • 2021/03/12 The late ACCPET is finally here, our "OCNet: Object context network for scene parsing" has been accepted by IJCV-2021, which consists of two of our previous technical reports: OCNet and ISA. Congratulations to all the co-authors!

  • 2021/02/16 Support pytorch-1.7, mixed-precision, and distributed training. Based on the PaddleClas ImageNet pretrained weights, we achieve 83.22% on Cityscapes val, 59.62% on PASCAL-Context val (new SOTA), 45.20% on COCO-Stuff val (new SOTA), 58.21% on LIP val and 47.98% on ADE20K val. Please checkout branch pytorch-1.7 for more details.

  • 2020/12/07 PaddleSeg has supported our ISA and HRNet + OCR. Jittor also has supported our ResNet-101 + OCR.

  • 2020/08/16 MMSegmentation has supported our HRNet + OCR.

  • 2020/07/20 The researchers from AInnovation have achieved Rank#1 on ADE20K Leaderboard via training our HRNet + OCR with a semi-supervised learning scheme. More details are in their Technical Report.

  • 2020/07/09 OCR (Spotlight) and SegFix have been accepted by the ECCV-2020. Notably, the reseachers from Nvidia set a new state-of-the-art performance on Cityscapes leaderboard: 85.4% via combining our HRNet + OCR with a new hierarchical mult-scale attention scheme.

  • 2020/05/11 We have released the checkpoints/logs of "HRNet + OCR" on all the 5 benchmarks including Cityscapes, ADE20K, LIP, PASCAL-Context and COCO-Stuff in the Model Zoo. Please feel free to try our method on your own dataset.

  • 2020/04/18 We have released some of our checkpoints/logs of OCNet, ISA, OCR and SegFix. We highly recommend you to use our SegFix to improve your segmentation results as it is super easy & fast to use.

  • 2020/03/12 Our SegFix could be used to improve the performance of various SOTA methods on both semantic segmentation and instance segmentation, e.g., "PolyTransform + SegFix" achieves Rank#2 on Cityscapes leaderboard (instance segmentation track) with performance as 41.2%.

  • 2020/01/13 The source code for reproduced HRNet+OCR has been made public.

  • 2020/01/09 "HRNet + OCR + SegFix" achieves Rank#1 on Cityscapes leaderboard with mIoU as 84.5%.

  • 2019/09/25 We have released the paper OCR, which is method of our Rank#2 entry to the leaderboard of Cityscapes.

  • 2019/07/31 We have released the paper ISA, which is very easy to use and implement while being much more efficient than OCNet or DANet based on conventional self-attention.

  • 2019/07/23 We (HRNet + OCR w/ ASP) achieve Rank#1 on the leaderboard of Cityscapes (with a single model) on 3 of 4 metrics.

  • 2019/05/27 We achieve SOTA on 6 different semantic segmentation benchmarks including: Cityscapes, ADE20K, LIP, Pascal-Context, Pascal-VOC, COCO-Stuff. We provide the source code for our approach on all the six benchmarks.

Model Zoo and Baselines

We provide a set of baseline results and trained models available for download in the Model Zoo.

Introduction

This is the official code of OCR, OCNet, ISA and SegFix. OCR, OCNet, and ISA focus on better context aggregation mechanisms (in the semantic segmentation task) and ISA focuses on addressing the boundary errors (in both semantic segmentation and instance segmentation tasks). We highlight the overall framework of OCR and SegFix in the figures as shown below:

OCR

Fig.1 - Illustrating the pipeline of OCR. (i) form the soft object regions in the pink dashed box. (ii) estimate the object region representations in the purple dashed box. (iii) compute the object contextual representations and the augmented representations in the orange dashed box.

SegFix

Fig.2 - Illustrating the SegFix framework: In the training stage, we first send the input image into a backbone to predict a feature map. Then we apply a boundary branch to predict a binary boundary map and a direction branch to predict a direction map and mask it with the binary boundary map. We apply boundary loss and direction loss on the predicted boundary map and direction map separately. In the testing stage, we first convert the direction map to offset map and then refine the segmentation results of any existing methods according to the offset map.

Citation

Please consider citing our work if you find it helps you,

@article{YuanW18,
  title={Ocnet: Object context network for scene parsing},
  author={Yuhui Yuan and Jingdong Wang},
  journal={arXiv preprint arXiv:1809.00916},
  year={2018}
}

@article{HuangYGZCW19,
  title={Interlaced Sparse Self-Attention for Semantic Segmentation},
  author={Lang Huang and Yuhui Yuan and Jianyuan Guo and Chao Zhang and Xilin Chen and Jingdong Wang},
  journal={arXiv preprint arXiv:1907.12273},
  year={2019}
}

@article{YuanCW20,
  title={Object-Contextual Representations for Semantic Segmentation},
  author={Yuhui Yuan and Xilin Chen and Jingdong Wang},
  journal={arXiv preprint arXiv:1909.11065},
  year={2020}
}

@article{YuanXCW20,
  title={SegFix: Model-Agnostic Boundary Refinement for Segmentation},
  author={Yuhui Yuan and Jingyi Xie and Xilin Chen and Jingdong Wang},
  journal={arXiv preprint arXiv:2007.04269},
  year={2020}
}

@article{YuanFHZCW21,
  title={HRT: High-Resolution Transformer for Dense Prediction},
  author={Yuhui Yuan and Rao Fu and Lang Huang and Weihong Lin and Chao Zhang and Xilin Chen and Jingdong Wang},
  booktitle={arXiv preprint arXiv:2110.09408},
  year={2021}
}

Acknowledgment

This project is developed based on the segbox.pytorch and the author of segbox.pytorch donnyyou retains all the copyright of the reproduced Deeplabv3, PSPNet related code.

Comments
  • questions/issues on training segfix with own data

    questions/issues on training segfix with own data

    I was excited to try segfix training on my own data.

    I could produce the mat files for train and val data. Training works with run_h_48_d_4_segfix.sh and loss convergences. But on the validation the IoU is more or less random (I have 2 classes)

    2020-08-20 10:47:41,932 INFO [base.py, 32] Result for mask 2020-08-20 10:47:41,932 INFO [base.py, 48] Mean IOU: 0.7853758111568029 2020-08-20 10:47:41,933 INFO [base.py, 49] Pixel ACC: 0.9692584678389714 2020-08-20 10:47:41,933 INFO [base.py, 54] F1 Score: 0.7523384841507573 Precision: 0.7928424176432377 Recall: 0.7157718538603068 2020-08-20 10:47:41,933 INFO [base.py, 32] Result for dir (mask) 2020-08-20 10:47:41,933 INFO [base.py, 48] Mean IOU: 0.5390945167184129 2020-08-20 10:47:41,933 INFO [base.py, 49] Pixel ACC: 0.7248566725097775 2020-08-20 10:47:41,933 INFO [base.py, 32] Result for dir (GT) 2020-08-20 10:47:41,934 INFO [base.py, 48] Mean IOU: 0.41990305666871003 2020-08-20 10:47:41,934 INFO [base.py, 49] Pixel ACC: 0.6007717101395131

    to investigate the issue further I tried to analyse the predicted mat files with bash scripts/cityscapes/segfix/run_h_48_d_4_segfix.sh segfix_pred_val 1

    with "input_size": [640, 480] this exception happens: File "/home/rsa-key-20190908/openseg.pytorch/lib/datasets/tools/collate.py", line 108, in collate assert pad_height >= 0 and pad_width >= 0 after fixing it more or less, iv got similar results as val during training They were around 3Kb instead of ~70kb btw, it took "input_size": [640, 480] config from "test": { leave instead "val": {

    is it possible validation only works with "input_size": [2048, 1024],? Can you give me any hints how to manually verify the .mat files of there correctness? Currently I'm diving into 2007.04269.pdf and the code of dt_offset_generator.py to get an understanding.

    opened by marcok 18
  • How to prepare the Cityscapes data

    How to prepare the Cityscapes data

    Hello. I'm trying to reproduce your CityScapes results for our BMVC paper.

    after I followed the data directory format in the config.profile file and running bash ./scripts/cityscapes/hrnet/run_h_48_d_4_ocr.sh val 1 I get this error:

    ERROR: Found no prediction for ground truth /home/arash/openseg.pytorch/dataset/cityscapes/val/label/munster_000027_000019_gtFine_labelIds.png

    could you explain how did you prepare the data? Thanks

    opened by arashash 15
  • about json file, the input size and crop size should based on what

    about json file, the input size and crop size should based on what

    my dataset image size is 256*256,and i dont know how to modifiy the json file

    {
        "dataset": "BDCI",
        "method": "fcn_segmentor",
        "data": {
          "image_tool": "cv2",
          "input_mode": "BGR",
          "num_classes": 7,
          "label_list": [0, 1, 2, 3, 4, 5, 6, 255],
          "data_dir": "~/DataSet/BDCI",
          "workers": 8
        },
       "train": {
          "batch_size": 16,
          "data_transformer": {
            "size_mode": "fix_size",
            "input_size": [256, 256],
            "align_method": "only_pad",
            "pad_mode": "random"
          }
        },
        "val": {
          "batch_size": 4,
          "mode": "ss_test",
          "data_transformer": {
            "size_mode": "fix_size",
            "input_size": [256, 256],
            "align_method": "only_pad"
          }
        },
        "test": {
          "batch_size": 4,
          "mode": "ss_test",
          "out_dir": "~/DataSet/BDCI/seg_result/BDCI",
          "data_transformer": {
            "size_mode": "fix_size",
            "input_size": [256, 256],
            "align_method": "only_pad"
          }
        },
        "train_trans": {
          "trans_seq": ["random_resize", "random_crop", "random_hflip", "random_brightness"],
          "random_brightness": {
            "ratio": 1.0,
            "shift_value": 10
          },
          "random_hflip": {
            "ratio": 0.5,
            "swap_pair": []
          },
          "random_resize": {
            "ratio": 1.0,
            "method": "random",
            "scale_range": [0.5, 2.0],
            "aspect_range": [0.9, 1.1]
          },
          "random_crop":{
            "ratio": 1.0,
            "crop_size": [256, 256],
            "method": "random",
            "allow_outside_center": false
          }
        },
        "val_trans": {
          "trans_seq": []
        },
        "normalize": {
          "div_value": 255.0,
          "mean_value": [0.485, 0.456, 0.406],
          "mean": [0.485, 0.456, 0.406],
          "std": [0.229, 0.224, 0.225]
        },
        "checkpoints": {
          "checkpoints_name": "fs_baseocnet_BDCI_seg",
          "checkpoints_dir": "./checkpoints/BDCI",
          "save_iters": 500
        },
        "network":{
          "backbone": "deepbase_resnet101_dilated8",
          "multi_grid": [1, 1, 1],
          "model_name": "base_ocnet",
          "bn_type": "inplace_abn",
          "stride": 8,
          "factors": [[8, 8]],
          "loss_weights": {
            "corr_loss": 0.01,
            "aux_loss": 0.4,
            "seg_loss": 1.0
          }
        },
        "logging": {
          "logfile_level": "info",
          "stdout_level": "info",
          "log_file": "./log/BDCI/fs_baseocnet_BDCI_seg.log",
          "log_format": "%(asctime)s %(levelname)-7s %(message)s",
          "rewrite": true
        },
        "lr": {
          "base_lr": 0.01,
          "metric": "iters",
          "lr_policy": "lambda_poly",
          "step": {
            "gamma": 0.5,
            "step_size": 100
          }
        },
        "solver": {
          "display_iter": 10,
          "test_interval": 1000,
          "max_iters": 40000
        },
        "optim": {
          "optim_method": "sgd",
          "adam": {
            "betas": [0.9, 0.999],
            "eps": 1e-08,
            "weight_decay": 0.0001
          },
          "sgd": {
            "weight_decay": 0.0005,
            "momentum": 0.9,
            "nesterov": false
          }
        },
        "loss": {
          "loss_type": "fs_auxce_loss",
          "params": {
            "ce_weight": [0.8373, 0.9180, 0.8660, 1.0345, 1.0166, 0.9969, 0.9754,
                          1.0489, 0.8786, 1.0023, 0.9539, 0.9843, 1.1116, 0.9037,
                          1.0865, 1.0955, 1.0865, 1.1529, 1.0507],
            "ce_reduction": "elementwise_mean",
            "ce_ignore_index": -1,
            "ohem_minkeep": 100000,
            "ohem_thresh": 0.9
          }
        }
    }
    
    

    here is my json file, and when i try to train my dataset, there is such sizemisbatch error...like: image image and so on, environment should be satisfied: image

    this is my val error: image and the config.profile: image this is my log file screenshot: image image image image

    opened by ShiMinghao0208 12
  • problem occured in  hrnet_backbone.py

    problem occured in hrnet_backbone.py

    Dear Author,

    Thank you for your excellent work, but some errors are reported for backbones.

    checkpoint names:
    checkpoints/cityscapes/hrnet_w48_ocr_1_latest.pth
    
    
    commands:
    (for HRNet-W48:)
    python -u main.py --configs configs/cityscapes/H_48_D_4.json --drop_last y --backbone hrnet48 --model_name hrnet_w48_ocr --checkpoints_name hrnet_w48_ocr_1 --phase test --gpu 0 --resume ./checkpoints/cityscapes/hrnet_w48_ocr_1_latest.pth --loss_type fs_auxce_loss --test_dir input_images --out_dir output_images
    

    Error messages:

    2020-07-15 21:00:10,470 INFO [module_runner.py, 44] BN Type is inplace_abn. Traceback (most recent call last): File "main.py", line 214, in model = Tester(configer)
    File "/home/dai/code/semantic_segmentation/9/openseg.pytorch-master/segmentor/tester.py", line 69, in init self._init_model() File "/home/dai/code/semantic_segmentation/9/openseg.pytorch-master/segmentor/tester.py", line 72, in _init_model self.seg_net = self.model_manager.semantic_segmentor() File "/home/dai/code/semantic_segmentation/9/openseg.pytorch-master/lib/models/model_manager.py", line 81, in semantic_segmentor model = SEG_MODEL_DICTmodel_name File "/home/dai/code/semantic_segmentation/9/openseg.pytorch-master/lib/models/nets/hrnet.py", line 105, in init self.backbone = BackboneSelector(configer).get_backbone() File "/home/dai/code/semantic_segmentation/9/openseg.pytorch-master/lib/models/backbones/backbone_selector.py", line 34, in get_backbone model = HRNetBackbone(self.configer)(**params) File "/home/dai/code/semantic_segmentation/9/openseg.pytorch-master/lib/models/backbones/hrnet/hrnet_backbone.py", line 598, in call bn_momentum=0.1) File "/home/dai/code/semantic_segmentation/9/openseg.pytorch-master/lib/models/backbones/hrnet/hrnet_backbone.py", line 307, in init self.bn1 = ModuleHelper.BatchNorm2d(bn_type=bn_type)(64, momentum=bn_momentum) TypeError: 'NoneType' object is not callable

    Could you please tell me what is wrong? thank you.

    opened by daixiaolei623 12
  • Problem with OCR similarity map

    Problem with OCR similarity map

    Thanks for sharing this wonderful work with us!

    I have a problem with the computing of similarity map in the OCR module. In line 131 in lib/models/seg_hrnet_orc.py sim_map = (self.key_channels**-.5) * sim_map Why multiply a small value (self.key_channels**-.5) to sim_map before softmax?

    During validation, I have printed the final result of sim_map and I found all values in this map are very close to 0.0526 (equals to 1/19), which means the probabilities of a pixel i belong to different classesk are almost equal. Is this contradicting the assumption that the similarity map should represent the relation between the _i_th pixel and the _k_th object region?

    #######################

    Your former answer:

    • Multiplying the small value is following the original self-attention scheme. Please refer to the last paragraph of 3.2.1 in the paper "Attention Is All You Need". However, we find this small factor does not influence the segmentation performance.

    • As the final result of the sim_map, we do not understand why all the values are almost the same in your case. What checkpoints are you testing? How about the performance of the used checkpoint? Please provide more information so that we can help you.

    #########################

    Thanks a lot for your reply! I used the checkpoint posted on HRNet-OCR. The segmentation performance is good ad the mIoU is 81.6, too. Screenshot from 2020-06-15 12-07-11 In inference, I have printed 10 random rows in the sim_map like below: Screenshot from 2020-06-15 12-19-24 All values in this map are very close to 0.0526 (equals to 1/19).

    opened by Mayy1994 11
  • SegFix paper link

    SegFix paper link

    Hi!

    Thanks for your nice work. It is really impressive. I'm interested in the SegFix algorithm. Could you send a copy of the paper "SegFix: Model-Agnostic Boundary Refinement for Segmentation", since I cannot find it on arXiv.

    Best, David

    opened by davidblom603 8
  • The performance of renset101-ocr

    The performance of renset101-ocr

    Hi, I want to reproduce the results of ocr paper, specially for pascal context and ade20k. Should I use the HRNet-OCR repo or this repo? In fact, I follow the default settings of HRNet-OCR and just replace HRNet with resnet101, but I can not reproduce the results on pascal context (54.8%mIoU) and ade20k (45.3%mIoU).

    opened by ydhongHIT 7
  • Test sets results

    Test sets results

    For comparison in our paper, we are looking for the detailed test set results (class IoUs) of these prediction files that you shared: https://drive.google.com/drive/folders/156vMABydr7btdPDBU6b9J-e0jJHuPI73 Do you happen to have a snapshot of the submission results obtained with these predictions? Thank you for your consideration.

    opened by arashash 7
  • class-id mapping for mapillary dataset

    class-id mapping for mapillary dataset

    I cannot find any class-id mapping in README or the config file. Just like Road in the ground truth with a label of 0 and traffic light is 1, the unlabeled is 255, etc.

    could you provide the mapping for v1.2 of mapillary?

    opened by lingorX 6
  • when i use H_SEGFIX.json to train cityscapes datasets meet the error:

    when i use H_SEGFIX.json to train cityscapes datasets meet the error:

    In loss_heleper.py In the calculation of loss function, the input is two tensors[1,8,128,128] /[1,2,128,128], and the corresponding label of single is three tensors.[1,512,512],[1,512,512],[1,512,512]

    targets=targets_.clone().unsqueeze(1).float() AttributeError:'list' object has no attribute 'clone'

    opened by qingchengboy 6
  • How to draw pictures

    How to draw pictures

    01 02 03 Coarse Label Map,Offset Map,Refined Label Map,Distance Map, Direction Map and the last one,How to draw them。Which drawing software is used, which is a program, what is the name of the software, and can the program be open source?I want to apply Figure 2 and Figure 3 to my own grayscale map. If it can be open sourced, will it be possible in the near future?Thanks you very much. 您好,抱歉我的英语太渣了,想了解一下这3张图是如何制作的。哪些图用了画图软件,是什么软件,哪些用了程序,程序可以开源吗。我想把图2和图3应用到自己的灰度图上,如果可以开源,近期可以吗?谢谢各位大佬,万分感谢。

    opened by Klaviersonate 5
  • need *.mat when I want to train segfix on my own dataset

    need *.mat when I want to train segfix on my own dataset

    I want to train segfix on my own dataset with script "scripts/cityscapes/segfix/run_hx_20_d_2_segfix_trainval.sh", but it seems that it needs file like *.mat, how to solve this problem? thank you.

    image

    opened by jhyin12 0
  • preprocess scripts for LIP

    preprocess scripts for LIP

    Thanks for your work!

    I download LIP dataset from here, and get dataset folder structure as below:

    .
    |-- ATR
    |   `-- humanparsing
    |       |-- JPEGImages
    |       `-- SegmentationClassAug
    |-- CIHP
    |   `-- instance-level_human_parsing
    |       |-- Testing
    |       |   `-- Images
    |       |-- Training
    |       |   |-- Categories
    |       |   |-- Category_ids
    |       |   |-- Human
    |       |   |-- Human_ids
    |       |   |-- Images
    |       |   |-- Instance_ids
    |       |   `-- Instances
    |       `-- Validation
    |           |-- Categories
    |           |-- Category_ids
    |           |-- Human
    |           |-- Human_ids
    |           |-- Images
    |           |-- Instance_ids
    |           `-- Instances
    `-- LIP
    

    that is different from the structure you mentioned in GETTING_STARTED.md:

    
    ├── lip
    │   ├── atr
    │   │   ├── edge
    │   │   ├── image
    │   │   └── label
    │   ├── cihp
    │   │   ├── image
    │   │   └── label
    │   ├── train
    │   │   ├── edge
    │   │   ├── image
    │   │   └── label
    │   ├── val
    │   │   ├── edge
    │   │   ├── image
    │   │   └── label
    
    

    could you please provide the scripts to preprocess LIP dataset? Thanks for a lot!

    opened by shouyanxiang 0
  • Result of refinement by SegFix on HRNet / HRNet-Semantic-Segmentation open source

    Result of refinement by SegFix on HRNet / HRNet-Semantic-Segmentation open source

    From my understanding, the two open source (HRNet-Semantic-Segmentation & openseg.pytorch) doesn't differ greatly.

    So I applied SegFix to results generated from HRNet-Semantic-Segmentation. The original mIoU is like below.

    image

    Obviously, I assumed that the final mIoU after applying SegFix would increase. However, that's not the case. mIoU actually decreased to 80.29.

    I applied SegFix the way described in MODEL_ZOO.md (below)

    image

    Is this the correct way to apply SegFix? Or is there any other way to apply SegFix?

    opened by Jonnyboyyyy 0
  • question about flops

    question about flops

    how do you calculate the flops in the figure 4 if I want to calculate input size of 512 * 97 * 97? i use the underline formula but the result is much larger than expectation. 捕获

    opened by HaoGuo98 0
Owner
openseg-group
openseg-group
LTR_CrossEncoder: Legal Text Retrieval Zalo AI Challenge 2021

LTR_CrossEncoder: Legal Text Retrieval Zalo AI Challenge 2021 We propose a cross encoder model (LTR_CrossEncoder) for information retrieval, re-retrie

Xuan Hieu Duong 7 Jan 12, 2022
The Environment I built to study Reinforcement Learning + Pokemon Showdown

pokemon-showdown-rl-environment The Environment I built to study Reinforcement Learning + Pokemon Showdown Been a while since I ran this. Think it is

3 Jan 16, 2022
TensorFlow implementation of the paper "Hierarchical Attention Networks for Document Classification"

Hierarchical Attention Networks for Document Classification This is an implementation of the paper Hierarchical Attention Networks for Document Classi

Quoc-Tuan Truong 83 Dec 05, 2022
CNNs for Sentence Classification in PyTorch

Introduction This is the implementation of Kim's Convolutional Neural Networks for Sentence Classification paper in PyTorch. Kim's implementation of t

Shawn Ng 956 Dec 19, 2022
Animatable Neural Radiance Fields for Modeling Dynamic Human Bodies

To make the comparison with Animatable NeRF easier on the Human3.6M dataset, we save the quantitative results at here, which also contains the results of other methods, including Neural Body, D-NeRF,

ZJU3DV 359 Jan 08, 2023
Activity image-based video retrieval

Cross-modal-retrieval Our approach is focus on Activity Image-to-Video Retrieval (AIVR) task. The compared methods are state-of-the-art single modalit

BCMI 75 Oct 21, 2021
Evaluating saliency methods on artificial data with different background types

Evaluating saliency methods on artificial data with different background types This repository contains the relevant code for the MedNeurips 2021 subm

2 Jul 05, 2022
Visualizer for neural network, deep learning, and machine learning models

Netron is a viewer for neural network, deep learning and machine learning models. Netron supports ONNX (.onnx, .pb, .pbtxt), Keras (.h5, .keras), Tens

Lutz Roeder 21k Jan 06, 2023
The official implementation of the CVPR2021 paper: Decoupled Dynamic Filter Networks

Decoupled Dynamic Filter Networks This repo is the official implementation of CVPR2021 paper: "Decoupled Dynamic Filter Networks". Introduction DDF is

F.S.Fire 180 Dec 30, 2022
Learn other languages ​​using artificial intelligence with python.

The main idea of ​​the project is to facilitate the learning of other languages. We created a simple AI that will interact with you. Just ask questions that if she knows, she will answer.

Pedro Rodrigues 2 Jun 07, 2022
Learning to Stylize Novel Views

Learning to Stylize Novel Views [Project] [Paper] Contact: Hsin-Ping Huang ([ema

34 Nov 27, 2022
A Robust Non-IoU Alternative to Non-Maxima Suppression in Object Detection

Confluence: A Robust Non-IoU Alternative to Non-Maxima Suppression in Object Detection 1. 介绍 用以替代 NMS,在所有 bbox 中挑选出最优的集合。 NMS 仅考虑了 bbox 的得分,然后根据 IOU 来

44 Sep 15, 2022
Image data augmentation scheduler for albumentations transforms

albu_scheduler Scheduler for albumentations transforms based on PyTorch schedulers interface Usage TransformMultiStepScheduler import albumentations a

19 Aug 04, 2021
This code is 3d-CNN model that can predict environmental value

Predict-environmental-value-3dCNN This code is 3d-CNN model that can predict environmental value. Firstly, I built a model that can create a lot of bu

1 Jan 06, 2022
Keras-tensorflow implementation of Fully Convolutional Networks for Semantic Segmentation(Unfinished)

Keras-FCN Fully convolutional networks and semantic segmentation with Keras. Models Models are found in models.py, and include ResNet and DenseNet bas

645 Dec 29, 2022
[AAAI22] Reliable Propagation-Correction Modulation for Video Object Segmentation

Reliable Propagation-Correction Modulation for Video Object Segmentation (AAAI22) Preview version paper of this work is available at: https://arxiv.or

Xiaohao Xu 70 Dec 04, 2022
The source code of "SIDE: Center-based Stereo 3D Detector with Structure-aware Instance Depth Estimation", accepted to WACV 2022.

SIDE: Center-based Stereo 3D Detector with Structure-aware Instance Depth Estimation The source code of our work "SIDE: Center-based Stereo 3D Detecto

10 Dec 18, 2022
NR-GAN: Noise Robust Generative Adversarial Networks

Lexicon Enhanced Chinese Sequence Labeling Using BERT Adapter Code and checkpoints for the ACL2021 paper "Lexicon Enhanced Chinese Sequence Labelling

Takuhiro Kaneko 59 Dec 11, 2022
Pytorch implementation of "M-LSD: Towards Light-weight and Real-time Line Segment Detection"

M-LSD: Towards Light-weight and Real-time Line Segment Detection Pytorch implementation of "M-LSD: Towards Light-weight and Real-time Line Segment Det

123 Jan 04, 2023
Research code of ICCV 2021 paper "Mesh Graphormer"

MeshGraphormer ✨ ✨ This is our research code of Mesh Graphormer. Mesh Graphormer is a new transformer-based method for human pose and mesh reconsructi

Microsoft 251 Jan 08, 2023