YOLOv7 - Framework Beyond Detection

Overview

YOLOv7 - Framework Beyond Detection


DocumentationInstallation InstructionsDeploymentContributingReporting Issues


PyPI - Python Version PyPI version PyPI downloads Github downloads

codecov license Slack PRs Welcome


This is the first and only (for now) YOLO family variant with transformers! and more advanced YOLO with multi-tasking such as detect & segmentation at the same time!

🔥 🔥 🔥 Just another yolo variant implemented based on detectron2. Be note that YOLOv7 doesn't meant to be a successor of yolo family, 7 is just my magic and lucky number. In our humble opinion, a good opensource project must have these features:

  • It must be reproduceble;
  • It must be simple and understandable;
  • It must be build with the weapon of the edge;
  • It must have a good maintainance, listen to the voice from community;

However, we found many opensource detection framework such as YOLOv5, Efficientdet have their own weakness, for example, YOLOv5 is very good at reproduceable but really over-engineered, too many messy codes. What's more surprisingly, there were at least 20+ different version of re-implementation of YOLOv3-YOLOv4 in pytorch, 99.99% of them were totally wrong, either can u train your dataset nor make it mAP comparable with origin paper.(However, doesn't mean this work is totally right, use at your own risk.)

That's why we have this project! It's much more simpler to experiment different ARCH of YOLO build upon detectron2 with YOLOv7! Most importantly, more and more decent YOLO series model merged into this repo such as YOLOX (most decent in 2021). We also welcome any trick/experiment PR on YOLOv7, help us build it better and stronger!!. Please star it and fork it right now!.

The supported matrix in YOLOv7 are:

  • YOLOv4 contained with CSP-Darknet53;
  • YOLOv7 arch with resnets backbone;
  • YOLOv7 arch with resnet-vd backbone (likely as PP-YOLO), deformable conv, Mish etc;
  • GridMask augmentation from PP-YOLO included;
  • Mosiac transform supported with a custom datasetmapper;
  • YOLOv7 arch Swin-Transformer support (higher accuracy but lower speed);
  • YOLOv7 arch Efficientnet + BiFPN;
  • YOLOv5 style positive samples selection, new coordinates coding style;
  • RandomColorDistortion, RandomExpand, RandomCrop, RandomFlip;
  • CIoU loss (DIoU, GIoU) and label smoothing (from YOLOv5 & YOLOv4);
  • YOLOF also included;
  • YOLOv7 Res2net + FPN supported;
  • Pyramid Vision Transformer v2 (PVTv2) supported;
  • WBF (Weighted Box Fusion), this works better than NMS, link;
  • YOLOX like head design and anchor design, also training support;
  • YOLOX s,m,l backbone and PAFPN added, we have a new combination of YOLOX backbone and pafpn;
  • YOLOv7 with Res2Net-v1d backbone, we found res2net-v1d have a better accuracy then darknet53;
  • Added PPYOLOv2 PAN neck with SPP and dropblock;
  • YOLOX arch added, now you can train YOLOX model (anchor free yolo) as well;
  • DETR: transformer based detection model and onnx export supported, as well as TensorRT acceleration;
  • AnchorDETR: Faster converge version of detr, now supported!

what's more, there are some features awesome inside repo:

  • Almost all models can export to onnx;
  • Supports TensorRT deployment for DETR and other transformer models;
  • It will integrate with wanwu, a torch-free deploy framework run fastest on your target platform.

💁‍♂️ Results

YOLOv7 Instance Face & Detection

🤔 Features

Some highlights of YOLOv7 are:

  • A simple and standard training framework for any detection && instance segmentation tasks, based on detectron2;
  • Supports DETR and many transformer based detection framework out-of-box;
  • Supports easy to deploy pipeline thought onnx.
  • This is the only framework support YOLOv4 + InstanceSegmentation in single stage style;
  • Easily plugin into transformers based detector;

We are strongly recommend you send PR if you have any further development on this project, the only reason for opensource it is just for using community power to make it stronger and further. It's very welcome for anyone contribute on any features!

🙌 Tasks Want Community to Finish

  • train a hand detection model, with YOLOX;
  • add more instance segmentation models into YOLOv7;

😎 Rules

There are some rules you must follow to if you want train on your own dataset:

  • Rule No.1: Always set your own anchors on your dataset, using tools/compute_anchors.py, this applys to any other anchor-based detection methods as well (EfficientDet etc.);
  • Rule No.2: Keep a faith on your loss will goes down eventually, if not, dig deeper to find out why (but do not post issues repeated caused I might don't know either.).
  • Rule No.3: No one will tells u but it's real: do not change backbone easily, whole params coupled with your backbone, dont think its simple as you think it should be, also a Deeplearning engineer is not an easy work as you think, the whole knowledge like an ocean, and your knowledge is just a tiny drop of water...
  • Rule No.4: must using pretrain weights for transoformer based backbone, otherwise your loss will bump;

Make sure you have read rules before ask me any questions.

🆕 News!

  • 2021.09.16: First transformer based DETR model added, will explore more DETR series models;
  • 2021.08.02: YOLOX arch added, you can train YOLOX as well in this repo;
  • 2021.07.25: We found YOLOv7-Res2net50 beat res50 and darknet53 at same speed level! 5% AP boost on custom dataset;
  • 2021.07.04: Added YOLOF and we can have a anchor free support as well, YOLOF achieves a better trade off on speed and accuracy;
  • 2021.06.25: this project first started.
  • more

🧑‍🦯 Installation && Quick Start

😎 Train

For training, quit simple, same as detectron2:

python train_net.py --config-file configs/coco/darknet53.yaml --num-gpus 8

If you want train YOLOX, you can using config file configs/coco/yolox_s.yaml. All support arch are:

  • YOLOX: anchor free yolo;
  • YOLOv7: traditional yolo with some explorations, mainly focus on loss experiments;
  • YOLOv7P: traditional yolo merged with decent arch from YOLOX;
  • YOLOMask: arch do detection and segmentation at the same time (tbd);
  • YOLOInsSeg: instance segmentation based on YOLO detection (tbd);

🥰 Demo

Run a quick demo would be like:

python3 demo.py --config-file configs/wearmask/darknet53.yaml --input ./datasets/wearmask/images/val2017 --opts MODEL.WEIGHTS output/model_0009999.pth

an update based on detectron2 newly introduced LazyConfig system, run with a LazyConfig model using:

python3 demo_lazyconfig.py --config-file configs/new_baselines/panoptic_fpn_regnetx_0.4g.py --opts train.init_checkpoint=output/model_0004999.pth

Export ONNX && TensorRTT && TVM

  1. detr:
python export_onnx.py --config-file detr/config/file

this works has been done, inference script included inside tools.

  1. AnchorDETR:

anchorDETR also supported training and exporting to ONNX.

More Advanced YOLO

Here we show some highlights on multi-tasking:

Performance

Here is a dedicated performance compare with other packages.

Some Tiny Object Datasets supported

Detection Results

Some Exp Visualizations

  1. GridMask

    Our GridMask augmentation also supports 2 modes.

  2. Mosaic

    Our Mosaic support any size and any any image numbers!

    new: we merged another mosiac implementation from YOLOX, this version will do random pespective:

License

Code released under GPL license. Please pull request to this source repo before you make your changes public or commercial usage. All rights reserved by Lucas Jin.

Comments
  • AssertionError: Invalid REFERENCE_WORLD_SIZE in config!

    AssertionError: Invalid REFERENCE_WORLD_SIZE in config!

    File "//python3.7/site-packages/detectron2/engine/defaults.py", line 674, in auto_scale_workers ), "Invalid REFERENCE_WORLD_SIZE in config!" AssertionError: Invalid REFERENCE_WORLD_SIZE in config!

    opened by zkailinzhang 10
  • Error in demo.py (I think I know the solution)

    Error in demo.py (I think I know the solution)

    Hi, Thanks for this repo :-D

    When I run the demo.py I get the following error:

    python3 demo.py --output /home/ws/Desktop/ --config-file configs/coco-instance/yolomask.yaml --input /home/ws/data/dataset/nets_kinneret_only24_2/images/0_2.jpg --opts MODEL.WEIGHTS output/coco_yolomask/model_final.pth
    Install mish-cuda to speed up training and inference. More importantly, replace the naive Mish with MishCuda will give a ~1.5G memory saving during training.
    [05/08 10:45:50 detectron2]: Arguments: Namespace(confidence_threshold=0.21, config_file='configs/coco-instance/yolomask.yaml', input='/home/ws/data/dataset/nets_kinneret_only24_2/images/0_2.jpg', nms_threshold=0.6, opts=['MODEL.WEIGHTS', 'output/coco_yolomask/model_final.pth'], output='/home/ws/Desktop/', video_input=None, webcam=False)
    10:45:50 05.08 INFO yolomask.py:86]: YOLO.ANCHORS: [[142, 110], [192, 243], [459, 401], [36, 75], [76, 55], [72, 146], [12, 16], [19, 36], [40, 28]]
    10:45:50 05.08 INFO yolomask.py:90]: backboneshape: [64, 128, 256, 512], size_divisibility: 32
    [[142, 110], [192, 243], [459, 401], [36, 75], [76, 55], [72, 146], [12, 16], [19, 36], [40, 28]]
    /home/ws/.local/lib/python3.6/site-packages/torch/functional.py:445: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at  ../aten/src/ATen/native/TensorShape.cpp:2157.)
      return _VF.meshgrid(tensors, **kwargs)  # type: ignore[attr-defined]
    [05/08 10:45:52 fvcore.common.checkpoint]: [Checkpointer] Loading from output/coco_yolomask/model_final.pth ...
    [05/08 10:45:53 d2.checkpoint.c2_model_loading]: Following weights matched with model:
    | Names in Model                                    | Names in Checkpoint
                                        | Shapes                         |
    |:--------------------------------------------------|:------------------------------------------------------------------------------------------------------|:-------------------------------|
    | backbone.dark2.0.bn.*                             | backbone.dark2.0.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                        | (64,) () (64,) (64,) (64,)     |
    | backbone.dark2.0.conv.weight                      | backbone.dark2.0.conv.weight
                                        | (64, 32, 3, 3)                 |
    | backbone.dark2.1.conv1.bn.*                       | backbone.dark2.1.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                  | (32,) () (32,) (32,) (32,)     |
    | backbone.dark2.1.conv1.conv.weight                | backbone.dark2.1.conv1.conv.weight
                                        | (32, 64, 1, 1)                 |
    | backbone.dark2.1.conv2.bn.*                       | backbone.dark2.1.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                  | (32,) () (32,) (32,) (32,)     |
    | backbone.dark2.1.conv2.conv.weight                | backbone.dark2.1.conv2.conv.weight
                                        | (32, 64, 1, 1)                 |
    | backbone.dark2.1.conv3.bn.*                       | backbone.dark2.1.conv3.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                  | (64,) () (64,) (64,) (64,)     |
    | backbone.dark2.1.conv3.conv.weight                | backbone.dark2.1.conv3.conv.weight
                                        | (64, 64, 1, 1)                 |
    | backbone.dark2.1.m.0.conv1.bn.*                   | backbone.dark2.1.m.0.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}              | (32,) () (32,) (32,) (32,)     |
    | backbone.dark2.1.m.0.conv1.conv.weight            | backbone.dark2.1.m.0.conv1.conv.weight
                                        | (32, 32, 1, 1)                 |
    | backbone.dark2.1.m.0.conv2.bn.*                   | backbone.dark2.1.m.0.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}              | (32,) () (32,) (32,) (32,)     |
    | backbone.dark2.1.m.0.conv2.conv.weight            | backbone.dark2.1.m.0.conv2.conv.weight
                                        | (32, 32, 3, 3)                 |
    | backbone.dark3.0.bn.*                             | backbone.dark3.0.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                        | (128,) () (128,) (128,) (128,) |
    | backbone.dark3.0.conv.weight                      | backbone.dark3.0.conv.weight
                                        | (128, 64, 3, 3)                |
    | backbone.dark3.1.conv1.bn.*                       | backbone.dark3.1.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                  | (64,) () (64,) (64,) (64,)     |
    | backbone.dark3.1.conv1.conv.weight                | backbone.dark3.1.conv1.conv.weight
                                        | (64, 128, 1, 1)                |
    | backbone.dark3.1.conv2.bn.*                       | backbone.dark3.1.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                  | (64,) () (64,) (64,) (64,)     |
    | backbone.dark3.1.conv2.conv.weight                | backbone.dark3.1.conv2.conv.weight
                                        | (64, 128, 1, 1)                |
    | backbone.dark3.1.conv3.bn.*                       | backbone.dark3.1.conv3.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                  | (128,) () (128,) (128,) (128,) |
    | backbone.dark3.1.conv3.conv.weight                | backbone.dark3.1.conv3.conv.weight
                                        | (128, 128, 1, 1)               |
    | backbone.dark3.1.m.0.conv1.bn.*                   | backbone.dark3.1.m.0.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}              | (64,) () (64,) (64,) (64,)     |
    | backbone.dark3.1.m.0.conv1.conv.weight            | backbone.dark3.1.m.0.conv1.conv.weight
                                        | (64, 64, 1, 1)                 |
    | backbone.dark3.1.m.0.conv2.bn.*                   | backbone.dark3.1.m.0.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}              | (64,) () (64,) (64,) (64,)     |
    | backbone.dark3.1.m.0.conv2.conv.weight            | backbone.dark3.1.m.0.conv2.conv.weight
                                        | (64, 64, 3, 3)                 |
    | backbone.dark3.1.m.1.conv1.bn.*                   | backbone.dark3.1.m.1.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}              | (64,) () (64,) (64,) (64,)     |
    | backbone.dark3.1.m.1.conv1.conv.weight            | backbone.dark3.1.m.1.conv1.conv.weight
                                        | (64, 64, 1, 1)                 |
    | backbone.dark3.1.m.1.conv2.bn.*                   | backbone.dark3.1.m.1.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}              | (64,) () (64,) (64,) (64,)     |
    | backbone.dark3.1.m.1.conv2.conv.weight            | backbone.dark3.1.m.1.conv2.conv.weight
                                        | (64, 64, 3, 3)                 |
    | backbone.dark3.1.m.2.conv1.bn.*                   | backbone.dark3.1.m.2.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}              | (64,) () (64,) (64,) (64,)     |
    | backbone.dark3.1.m.2.conv1.conv.weight            | backbone.dark3.1.m.2.conv1.conv.weight
                                        | (64, 64, 1, 1)                 |
    | backbone.dark3.1.m.2.conv2.bn.*                   | backbone.dark3.1.m.2.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}              | (64,) () (64,) (64,) (64,)     |
    | backbone.dark3.1.m.2.conv2.conv.weight            | backbone.dark3.1.m.2.conv2.conv.weight
                                        | (64, 64, 3, 3)                 |
    | backbone.dark4.0.bn.*                             | backbone.dark4.0.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                        | (256,) () (256,) (256,) (256,) |
    | backbone.dark4.0.conv.weight                      | backbone.dark4.0.conv.weight
                                        | (256, 128, 3, 3)               |
    | backbone.dark4.1.conv1.bn.*                       | backbone.dark4.1.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                  | (128,) () (128,) (128,) (128,) |
    | backbone.dark4.1.conv1.conv.weight                | backbone.dark4.1.conv1.conv.weight
                                        | (128, 256, 1, 1)               |
    | backbone.dark4.1.conv2.bn.*                       | backbone.dark4.1.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                  | (128,) () (128,) (128,) (128,) |
    | backbone.dark4.1.conv2.conv.weight                | backbone.dark4.1.conv2.conv.weight
                                        | (128, 256, 1, 1)               |
    | backbone.dark4.1.conv3.bn.*                       | backbone.dark4.1.conv3.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                  | (256,) () (256,) (256,) (256,) |
    | backbone.dark4.1.conv3.conv.weight                | backbone.dark4.1.conv3.conv.weight
                                        | (256, 256, 1, 1)               |
    | backbone.dark4.1.m.0.conv1.bn.*                   | backbone.dark4.1.m.0.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}              | (128,) () (128,) (128,) (128,) |
    | backbone.dark4.1.m.0.conv1.conv.weight            | backbone.dark4.1.m.0.conv1.conv.weight
                                        | (128, 128, 1, 1)               |
    | backbone.dark4.1.m.0.conv2.bn.*                   | backbone.dark4.1.m.0.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}              | (128,) () (128,) (128,) (128,) |
    | backbone.dark4.1.m.0.conv2.conv.weight            | backbone.dark4.1.m.0.conv2.conv.weight
                                        | (128, 128, 3, 3)               |
    | backbone.dark4.1.m.1.conv1.bn.*                   | backbone.dark4.1.m.1.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}              | (128,) () (128,) (128,) (128,) |
    | backbone.dark4.1.m.1.conv1.conv.weight            | backbone.dark4.1.m.1.conv1.conv.weight
                                        | (128, 128, 1, 1)               |
    | backbone.dark4.1.m.1.conv2.bn.*                   | backbone.dark4.1.m.1.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}              | (128,) () (128,) (128,) (128,) |
    | backbone.dark4.1.m.1.conv2.conv.weight            | backbone.dark4.1.m.1.conv2.conv.weight
                                        | (128, 128, 3, 3)               |
    | backbone.dark4.1.m.2.conv1.bn.*                   | backbone.dark4.1.m.2.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}              | (128,) () (128,) (128,) (128,) |
    | backbone.dark4.1.m.2.conv1.conv.weight            | backbone.dark4.1.m.2.conv1.conv.weight
                                        | (128, 128, 1, 1)               |
    | backbone.dark4.1.m.2.conv2.bn.*                   | backbone.dark4.1.m.2.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}              | (128,) () (128,) (128,) (128,) |
    | backbone.dark4.1.m.2.conv2.conv.weight            | backbone.dark4.1.m.2.conv2.conv.weight
                                        | (128, 128, 3, 3)               |
    | backbone.dark5.0.bn.*                             | backbone.dark5.0.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                        | (512,) () (512,) (512,) (512,) |
    | backbone.dark5.0.conv.weight                      | backbone.dark5.0.conv.weight
                                        | (512, 256, 3, 3)               |
    | backbone.dark5.1.conv1.bn.*                       | backbone.dark5.1.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                  | (256,) () (256,) (256,) (256,) |
    | backbone.dark5.1.conv1.conv.weight                | backbone.dark5.1.conv1.conv.weight
                                        | (256, 512, 1, 1)               |
    | backbone.dark5.1.conv2.bn.*                       | backbone.dark5.1.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                  | (512,) () (512,) (512,) (512,) |
    | backbone.dark5.1.conv2.conv.weight                | backbone.dark5.1.conv2.conv.weight
                                        | (512, 1024, 1, 1)              |
    | backbone.dark5.2.conv1.bn.*                       | backbone.dark5.2.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                  | (256,) () (256,) (256,) (256,) |
    | backbone.dark5.2.conv1.conv.weight                | backbone.dark5.2.conv1.conv.weight
                                        | (256, 512, 1, 1)               |
    | backbone.dark5.2.conv2.bn.*                       | backbone.dark5.2.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                  | (256,) () (256,) (256,) (256,) |
    | backbone.dark5.2.conv2.conv.weight                | backbone.dark5.2.conv2.conv.weight
                                        | (256, 512, 1, 1)               |
    | backbone.dark5.2.conv3.bn.*                       | backbone.dark5.2.conv3.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                  | (512,) () (512,) (512,) (512,) |
    | backbone.dark5.2.conv3.conv.weight                | backbone.dark5.2.conv3.conv.weight
                                        | (512, 512, 1, 1)               |
    | backbone.dark5.2.m.0.conv1.bn.*                   | backbone.dark5.2.m.0.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}              | (256,) () (256,) (256,) (256,) |
    | backbone.dark5.2.m.0.conv1.conv.weight            | backbone.dark5.2.m.0.conv1.conv.weight
                                        | (256, 256, 1, 1)               |
    | backbone.dark5.2.m.0.conv2.bn.*                   | backbone.dark5.2.m.0.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}              | (256,) () (256,) (256,) (256,) |
    | backbone.dark5.2.m.0.conv2.conv.weight            | backbone.dark5.2.m.0.conv2.conv.weight
                                        | (256, 256, 3, 3)               |
    | backbone.stem.conv.bn.*                           | backbone.stem.conv.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                      | (32,) () (32,) (32,) (32,)     |
    | backbone.stem.conv.conv.weight                    | backbone.stem.conv.conv.weight
                                        | (32, 12, 3, 3)                 |
    | m.0.*                                             | m.0.{bias,weight}
                                        | (255,) (255,512,1,1)           |
    | m.1.*                                             | m.1.{bias,weight}
                                        | (255,) (255,256,1,1)           |
    | m.2.*                                             | m.2.{bias,weight}
                                        | (255,) (255,128,1,1)           |
    | neck.C3_n3.conv1.bn.*                             | neck.C3_n3.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                        | (128,) () (128,) (128,) (128,) |
    | neck.C3_n3.conv1.conv.weight                      | neck.C3_n3.conv1.conv.weight
                                        | (128, 256, 1, 1)               |
    | neck.C3_n3.conv2.bn.*                             | neck.C3_n3.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                        | (128,) () (128,) (128,) (128,) |
    | neck.C3_n3.conv2.conv.weight                      | neck.C3_n3.conv2.conv.weight
                                        | (128, 256, 1, 1)               |
    | neck.C3_n3.conv3.bn.*                             | neck.C3_n3.conv3.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                        | (256,) () (256,) (256,) (256,) |
    | neck.C3_n3.conv3.conv.weight                      | neck.C3_n3.conv3.conv.weight
                                        | (256, 256, 1, 1)               |
    | neck.C3_n3.m.0.conv1.bn.*                         | neck.C3_n3.m.0.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                    | (128,) () (128,) (128,) (128,) |
    | neck.C3_n3.m.0.conv1.conv.weight                  | neck.C3_n3.m.0.conv1.conv.weight
                                        | (128, 128, 1, 1)               |
    | neck.C3_n3.m.0.conv2.bn.*                         | neck.C3_n3.m.0.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                    | (128,) () (128,) (128,) (128,) |
    | neck.C3_n3.m.0.conv2.conv.weight                  | neck.C3_n3.m.0.conv2.conv.weight
                                        | (128, 128, 3, 3)               |
    | neck.C3_n4.conv1.bn.*                             | neck.C3_n4.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                        | (256,) () (256,) (256,) (256,) |
    | neck.C3_n4.conv1.conv.weight                      | neck.C3_n4.conv1.conv.weight
                                        | (256, 512, 1, 1)               |
    | neck.C3_n4.conv2.bn.*                             | neck.C3_n4.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                        | (256,) () (256,) (256,) (256,) |
    | neck.C3_n4.conv2.conv.weight                      | neck.C3_n4.conv2.conv.weight
                                        | (256, 512, 1, 1)               |
    | neck.C3_n4.conv3.bn.*                             | neck.C3_n4.conv3.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                        | (512,) () (512,) (512,) (512,) |
    | neck.C3_n4.conv3.conv.weight                      | neck.C3_n4.conv3.conv.weight
                                        | (512, 512, 1, 1)               |
    | neck.C3_n4.m.0.conv1.bn.*                         | neck.C3_n4.m.0.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                    | (256,) () (256,) (256,) (256,) |
    | neck.C3_n4.m.0.conv1.conv.weight                  | neck.C3_n4.m.0.conv1.conv.weight
                                        | (256, 256, 1, 1)               |
    | neck.C3_n4.m.0.conv2.bn.*                         | neck.C3_n4.m.0.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                    | (256,) () (256,) (256,) (256,) |
    | neck.C3_n4.m.0.conv2.conv.weight                  | neck.C3_n4.m.0.conv2.conv.weight
                                        | (256, 256, 3, 3)               |
    | neck.C3_p3.conv1.bn.*                             | neck.C3_p3.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                        | (64,) () (64,) (64,) (64,)     |
    | neck.C3_p3.conv1.conv.weight                      | neck.C3_p3.conv1.conv.weight
                                        | (64, 256, 1, 1)                |
    | neck.C3_p3.conv2.bn.*                             | neck.C3_p3.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                        | (64,) () (64,) (64,) (64,)     |
    | neck.C3_p3.conv2.conv.weight                      | neck.C3_p3.conv2.conv.weight
                                        | (64, 256, 1, 1)                |
    | neck.C3_p3.conv3.bn.*                             | neck.C3_p3.conv3.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                        | (128,) () (128,) (128,) (128,) |
    | neck.C3_p3.conv3.conv.weight                      | neck.C3_p3.conv3.conv.weight
                                        | (128, 128, 1, 1)               |
    | neck.C3_p3.m.0.conv1.bn.*                         | neck.C3_p3.m.0.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                    | (64,) () (64,) (64,) (64,)     |
    | neck.C3_p3.m.0.conv1.conv.weight                  | neck.C3_p3.m.0.conv1.conv.weight
                                        | (64, 64, 1, 1)                 |
    | neck.C3_p3.m.0.conv2.bn.*                         | neck.C3_p3.m.0.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                    | (64,) () (64,) (64,) (64,)     |
    | neck.C3_p3.m.0.conv2.conv.weight                  | neck.C3_p3.m.0.conv2.conv.weight
                                        | (64, 64, 3, 3)                 |
    | neck.C3_p4.conv1.bn.*                             | neck.C3_p4.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                        | (128,) () (128,) (128,) (128,) |
    | neck.C3_p4.conv1.conv.weight                      | neck.C3_p4.conv1.conv.weight
                                        | (128, 512, 1, 1)               |
    | neck.C3_p4.conv2.bn.*                             | neck.C3_p4.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                        | (128,) () (128,) (128,) (128,) |
    | neck.C3_p4.conv2.conv.weight                      | neck.C3_p4.conv2.conv.weight
                                        | (128, 512, 1, 1)               |
    | neck.C3_p4.conv3.bn.*                             | neck.C3_p4.conv3.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                        | (256,) () (256,) (256,) (256,) |
    | neck.C3_p4.conv3.conv.weight                      | neck.C3_p4.conv3.conv.weight
                                        | (256, 256, 1, 1)               |
    | neck.C3_p4.m.0.conv1.bn.*                         | neck.C3_p4.m.0.conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                    | (128,) () (128,) (128,) (128,) |
    | neck.C3_p4.m.0.conv1.conv.weight                  | neck.C3_p4.m.0.conv1.conv.weight
                                        | (128, 128, 1, 1)               |
    | neck.C3_p4.m.0.conv2.bn.*                         | neck.C3_p4.m.0.conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                    | (128,) () (128,) (128,) (128,) |
    | neck.C3_p4.m.0.conv2.conv.weight                  | neck.C3_p4.m.0.conv2.conv.weight
                                        | (128, 128, 3, 3)               |
    | neck.bu_conv1.bn.*                                | neck.bu_conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                           | (256,) () (256,) (256,) (256,) |
    | neck.bu_conv1.conv.weight                         | neck.bu_conv1.conv.weight
                                        | (256, 256, 3, 3)               |
    | neck.bu_conv2.bn.*                                | neck.bu_conv2.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                           | (128,) () (128,) (128,) (128,) |
    | neck.bu_conv2.conv.weight                         | neck.bu_conv2.conv.weight
                                        | (128, 128, 3, 3)               |
    | neck.lateral_conv0.bn.*                           | neck.lateral_conv0.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                      | (256,) () (256,) (256,) (256,) |
    | neck.lateral_conv0.conv.weight                    | neck.lateral_conv0.conv.weight
                                        | (256, 512, 1, 1)               |
    | neck.reduce_conv1.bn.*                            | neck.reduce_conv1.bn.{bias,num_batches_tracked,running_mean,running_var,weight}                       | (128,) () (128,) (128,) (128,) |
    | neck.reduce_conv1.conv.weight                     | neck.reduce_conv1.conv.weight
                                        | (128, 256, 1, 1)               |
    | orien_head.neck_orien.0.conv_block.0.weight       | orien_head.neck_orien.0.conv_block.0.weight
                                        | (128, 256, 1, 1)               |
    | orien_head.neck_orien.0.conv_block.1.*            | orien_head.neck_orien.0.conv_block.1.{bias,num_batches_tracked,running_mean,running_var,weight}       | (128,) () (128,) (128,) (128,) |
    | orien_head.neck_orien.1.conv_block.0.weight       | orien_head.neck_orien.1.conv_block.0.weight
                                        | (256, 128, 3, 3)               |
    | orien_head.neck_orien.1.conv_block.1.*            | orien_head.neck_orien.1.conv_block.1.{bias,num_batches_tracked,running_mean,running_var,weight}       | (256,) () (256,) (256,) (256,) |
    | orien_head.neck_orien.2.conv_block.0.weight       | orien_head.neck_orien.2.conv_block.0.weight
                                        | (128, 256, 1, 1)               |
    | orien_head.neck_orien.2.conv_block.1.*            | orien_head.neck_orien.2.conv_block.1.{bias,num_batches_tracked,running_mean,running_var,weight}       | (128,) () (128,) (128,) (128,) |
    | orien_head.neck_orien.3.conv_block.0.weight       | orien_head.neck_orien.3.conv_block.0.weight
                                        | (256, 128, 3, 3)               |
    | orien_head.neck_orien.3.conv_block.1.*            | orien_head.neck_orien.3.conv_block.1.{bias,num_batches_tracked,running_mean,running_var,weight}       | (256,) () (256,) (256,) (256,) |
    | orien_head.neck_orien.4.conv_block.0.weight       | orien_head.neck_orien.4.conv_block.0.weight
                                        | (128, 256, 1, 1)               |
    | orien_head.neck_orien.4.conv_block.1.*            | orien_head.neck_orien.4.conv_block.1.{bias,num_batches_tracked,running_mean,running_var,weight}       | (128,) () (128,) (128,) (128,) |
    | orien_head.orien_m.0.conv_block.0.weight          | orien_head.orien_m.0.conv_block.0.weight
                                        | (256, 128, 3, 3)               |
    | orien_head.orien_m.0.conv_block.1.*               | orien_head.orien_m.0.conv_block.1.{bias,num_batches_tracked,running_mean,running_var,weight}          | (256,) () (256,) (256,) (256,) |
    | orien_head.orien_m.1.conv_block.0.weight          | orien_head.orien_m.1.conv_block.0.weight
                                        | (128, 256, 1, 1)               |
    | orien_head.orien_m.1.conv_block.1.*               | orien_head.orien_m.1.conv_block.1.{bias,num_batches_tracked,running_mean,running_var,weight}          | (128,) () (128,) (128,) (128,) |
    | orien_head.orien_m.2.conv_block.0.weight          | orien_head.orien_m.2.conv_block.0.weight
                                        | (256, 128, 3, 3)               |
    | orien_head.orien_m.2.conv_block.1.*               | orien_head.orien_m.2.conv_block.1.{bias,num_batches_tracked,running_mean,running_var,weight}          | (256,) () (256,) (256,) (256,) |
    | orien_head.orien_m.3.conv_block.0.weight          | orien_head.orien_m.3.conv_block.0.weight
                                        | (128, 256, 1, 1)               |
    | orien_head.orien_m.3.conv_block.1.*               | orien_head.orien_m.3.conv_block.1.{bias,num_batches_tracked,running_mean,running_var,weight}          | (128,) () (128,) (128,) (128,) |
    | orien_head.orien_m.4.conv_block.0.weight          | orien_head.orien_m.4.conv_block.0.weight
                                        | (256, 128, 3, 3)               |
    | orien_head.orien_m.4.conv_block.1.*               | orien_head.orien_m.4.conv_block.1.{bias,num_batches_tracked,running_mean,running_var,weight}          | (256,) () (256,) (256,) (256,) |
    | orien_head.orien_m.5.*                            | orien_head.orien_m.5.{bias,weight}
                                        | (18,) (18,256,1,1)             |
    | orien_head.up_levels_2to5.0.conv_block.0.weight   | orien_head.up_levels_2to5.0.conv_block.0.weight
                                        | (64, 64, 1, 1)                 |
    | orien_head.up_levels_2to5.0.conv_block.1.*        | orien_head.up_levels_2to5.0.conv_block.1.{bias,num_batches_tracked,running_mean,running_var,weight}   | (64,) () (64,) (64,) (64,)     |
    | orien_head.up_levels_2to5.1.0.conv_block.0.weight | orien_head.up_levels_2to5.1.0.conv_block.0.weight
                                        | (64, 128, 1, 1)                |
    | orien_head.up_levels_2to5.1.0.conv_block.1.*      | orien_head.up_levels_2to5.1.0.conv_block.1.{bias,num_batches_tracked,running_mean,running_var,weight} | (64,) () (64,) (64,) (64,)     |
    | orien_head.up_levels_2to5.2.0.conv_block.0.weight | orien_head.up_levels_2to5.2.0.conv_block.0.weight
                                        | (64, 256, 1, 1)                |
    | orien_head.up_levels_2to5.2.0.conv_block.1.*      | orien_head.up_levels_2to5.2.0.conv_block.1.{bias,num_batches_tracked,running_mean,running_var,weight} | (64,) () (64,) (64,) (64,)     |
    | orien_head.up_levels_2to5.3.0.conv_block.0.weight | orien_head.up_levels_2to5.3.0.conv_block.0.weight
                                        | (64, 512, 1, 1)                |
    | orien_head.up_levels_2to5.3.0.conv_block.1.*      | orien_head.up_levels_2to5.3.0.conv_block.1.{bias,num_batches_tracked,running_mean,running_var,weight} | (64,) () (64,) (64,) (64,)     |
    416 416 600
    confidence thresh:  0.21
    image after transform:  (416, 416, 3)
    cost: 0.02672123908996582, fps: 37.42341425983922
    Traceback (most recent call last):
      File "demo.py", line 211, in <module>
        res = vis_res_fast(res, img, class_names, colors, conf_thresh)
      File "demo.py", line 146, in vis_res_fast
        img, clss, bit_masks, force_colors=None, draw_contours=False
      File "/home/ws/.local/lib/python3.6/site-packages/alfred/vis/image/mask.py", line 302, in vis_bitmasks_with_classes
        txt = f'{class_names[classes[i]]}'
    TypeError: list indices must be integers or slices, not numpy.float32
    

    I think this is happened because this line: scores = ins.scores.cpu().numpy() clss = ins.pred_classes.cpu().numpy()

    This cuse clss to by a numpy array and not a list of int. Am I right. How it's works for you?

    Also, in line (157): if bboxes: bboxes is not bool Maybe it's should by: if bboxes is not None:

    Thanks

    opened by sdimantsd 10
  • KeyError: 'dark3'

    KeyError: 'dark3'

    i do traning on cutstom dataset and i got this error

    `[07/11 22:24:10 detectron2]: Running with full config was omitted.
    [07/11 22:24:10 detectron2]: Full config saved to ./output/config.yaml
    [07/11 22:24:10 d2.utils.env]: Using a generated random seed 10422596
    [07/11 22:24:10 d2.engine.defaults]: Auto-scaling the config to batch_size=2, learning_rate=0.0025, max_iter=720000, warmup=8000.
    Traceback (most recent call last):
      File "train_ahmed.py", line 61, in <module>
        args=(args,),
      File "/content/detectron2/detectron2/engine/launch.py", line 82, in launch
        main_func(*args)
      File "train_ahmed.py", line 48, in main
        trainer = Trainer(cfg)
      File "/content/detectron2/detectron2/engine/defaults.py", line 376, in __init__
        model = self.build_model(cfg)
      File "/content/yolov7/train_det.py", line 37, in build_model
        model = build_model(cfg)
      File "/content/detectron2/detectron2/modeling/meta_arch/build.py", line 22, in build_model
        model = META_ARCH_REGISTRY.get(meta_arch)(cfg)
      File "/content/yolov7/yolov7/modeling/meta_arch/yolov7.py", line 95, in __init__
        backbone_shape = [backbone_shape[i].channels for i in self.in_features]
      File "/content/yolov7/yolov7/modeling/meta_arch/yolov7.py", line 95, in <listcomp>
        backbone_shape = [backbone_shape[i].channels for i in self.in_features]
    KeyError: 'dark3'`
    
    opened by totoadel 9
  • Inference on SparseInst onnxruntime gives bad output data

    Inference on SparseInst onnxruntime gives bad output data

    Hello!

    Has anyone successfully exported the sparseinst model to onnx and run inference on it using onnxruntime? I'm struggling getting reasonable output when running inference in C++ onnxruntime. I'm running the same frame through the network pre-export and the output looks well. But post-export the model outputs scores size (1, 50) with all zeros but a few scores ~0.04. For comparison the model pre-export output two person labels and one sports ball all over 80% confidence.

    I exported the model using the provided export_onnx.py as per readme instruction. The onnx-model looks, from what I can tell, fine when looking at it Netron.

    Just wondering if anyone has successfully used the exported model and could share some examples? =)

    opened by mattiasbax 9
  • SparseInst Model exportation to ONNX

    SparseInst Model exportation to ONNX

    Thank you so much for your work @jinfagang. I have tested yolov7 and I realized that SparseInst models cannot be converted to ONNX. Is the export onnx code compatible with exporting SparseInst models?

    opened by ayoolaolafenwa 9
  • SparseInst expecting pred_boxes

    SparseInst expecting pred_boxes

    While trying to train a SparseInst on a custom dataset for instance segmentation, I came across the following error:

    AttributeError: Cannot find field 'pred_boxes' in the given Instances!

    The error is thrown from within the detectron2library. When I print one Instance, there are indeed no pred_boxesin there, only pred_masks. But as far as I understand, SparseInst does not output bounding boxes, it predicts directly seg. masks, as they highlight in their paper: https://arxiv.org/abs/2203.12827 Then I am wondering why the code is looking for predicted boxes in this case... o_0

    opened by TeodorChiaburu 8
  • SparseInst export to ONNX input data format

    SparseInst export to ONNX input data format

    Hi, I tried exporting the weights of SparseInst (GIAM) that are in this repository to ONNX format using export.py with the following command (I assume I need to use this command? The documentation reads 'use export_onnx.py', but there's no export_onnx.py in the current branch of this repository).

    python export.py --config-file configs/coco/sparseinst/sparse_inst_r50_giam_aug.yaml --opts MODEL.WEIGHTS weights/base_giam.pth INPUT.MIN_SIZE_TEST 512

    This leads to the following issue:

      File ".../yolov7/yolov7/modeling/meta_arch/sparseinst.py", line 95, in <listcomp>
        images = [x["image"].to(self.device) for x in batched_inputs]
    IndexError: too many indices for tensor of dimension 3
    

    Any ideas on how to fix this? Thank you.

    opened by LukasMahieuArinti 6
  • sparseinst no results

    sparseinst no results

    I try the demo of sparseinst like this

    python demo.py --config-file configs/coco/sparseinst/sparse_inst_r50vd_giam_aug.yaml --video-input ~/Movies/Videos/86277963_nb2-1-80.flv -c 0.4 --opts MODEL.WEIGHTS weights/sparse_inst_r50vd_giam_aug_8bc5b3.pth
    

    but I get error

    Traceback (most recent call last):
      File "/home/dreamdeck/Documents/MJJ/code/Seg/yolov7-main/demo.py", line 230, in <module>
        res = vis_res_fast(res, frame, class_names, colors, conf_thresh)
      File "/home/dreamdeck/Documents/MJJ/code/Seg/yolov7-main/demo.py", line 155, in vis_res_fast
        img, clss, bit_masks, force_colors=None, draw_contours=False
      File "/home/dreamdeck/anaconda3/envs/detectron2/lib/python3.6/site-packages/alfred/vis/image/mask.py", line 285, in vis_bitmasks_with_classes
        thickness=-1, lineType=cv2.LINE_AA)
    cv2.error: OpenCV(4.5.5) /io/opencv/modules/imgproc/src/drawing.cpp:2599: error: (-215:Assertion failed) reader.ptr != NULL in function 'cvDrawContours'
    

    I debuged and found that there is no masks when run predictions = self.model([inputs]). All values in pred_masks are False.

    Do I need modify the config file or download other files?( I have dolwnloaded the sparse_inst_r50vd_giam_aug_8bc5b3.pth) What is the full steps to test demo?

    opened by MolianWH 6
  • SparseInst onnx2trt error

    SparseInst onnx2trt error

    @jinfagang ,hello, thank you for your work! when i convert SparseInst onnx model to tensorrt model using trtexec , the error is In node -1(importRange):UNSUPPORTED_NODE:Assertion failed:inputs.at(0).isInt32() && "For range operator with dynamic inputs,this version og TensorRT only supports INT32!

    tensorrt version:7.2.3 onnx version : v7 pytorch version:1.10

    opened by taxuezcy 6
  • Log train metrics into Weights & Biases for Object Detection task

    Log train metrics into Weights & Biases for Object Detection task

    This PR adds the ability to log training metrics for object detection into Weights & Biases. Here's an example dashboard of fine-tuning the faster_rcnn_R_101_FPN_3x detection model from detectron2 model zoo.

    opened by parambharat 5
  • 评估时报错cuda out of memory

    评估时报错cuda out of memory

    您好,很感谢您提供代码,我在使用代码时遇到两个问题,希望您能帮忙解答一下

    1. 当我使用yolomask_8gpu.yaml在coco上做evaluation时,到第27张图片进入model得到输出后显存会增加300多M,到第47张图片又会同样增加,然后就报错cuda out of memory。

    2. coco-instance中包含两个config,其中yolomask.yaml中IMS_PER_BATCH:3,但是默认的REFERENCE_WORLD_SIZE: 8,运行时会报错提示IMS_PER_BATCH是无效的,查看源码提示IMS_PER_BATCH%REFERENCE_WORLD_SIZE==0

    opened by erpingzi 5
  • unable to load DarkNet53 backbone for yolov7

    unable to load DarkNet53 backbone for yolov7

    Hi, @jinfagang

    According to darknet53.yaml, I downloaded yolov3.pt as the backbone for initializing the yolov7.

    However, I met the following problem, do you have any suggestions to provide accurate parameters for yolov7?

    RuntimeError: Error(s) in loading state_dict for Darknet: Missing key(s) in state_dict: "stem.conv.weight", "stem.bn.weight", "stem.bn.bias", "stem.bn.running_mean", "stem.bn.running_var", "dark1.0.conv.weight", "dark1.0.bn.weight", "dark1.0.bn.bias", "dark1.0.bn.running_mean", "dark1.0.bn.running_var", "dark1.1.layer1.conv.weight", "dark1.1.layer1.bn.weight", "dark1.1.layer1.bn.bias", "dark1.1.layer1.bn.running_mean", "dark1.1.layer1.bn.running_var", "dark1.1.layer2.conv.weight", "dark1.1.layer2.bn.weight", ....... .......

    opened by GrassBro 0
  • Backbone of yolo v7

    Backbone of yolo v7

    Hello, I appreciate your novel work. I have a question about your implementation. From official yolo v7 github, there are no public backbone network application around yolo v7 variants. Therefore, I can't find the certain name of yolov7 backbone or head network. Then, Could you give me an exact combination for each yolov7 variant from your implementation? For example, darknet53 + yolofpn + yoloxhead = yolo_v7_xxx.

    Best regards.

    opened by chey0313 1
  • ONNX export failed: Couldn't export Python operator _DeformConv

    ONNX export failed: Couldn't export Python operator _DeformConv

    Hi,

    • in docker container from pytorch/pytorch:1.11.0-cuda11.3-cudnn8-devel
    • torch 1.11.0
    • torchvision 0.12.0
    • latest detectron2 main branch
    • python3 export.py failed while exporting model sparse_inst_r50vd_dcn_giam_aug.yaml

    Error log:

    Traceback (most recent call last):
      File "export.py", line 285, in <module>
        torch.onnx.export(
      File "/opt/conda/lib/python3.8/site-packages/torch/onnx/__init__.py", line 305, in export
        return utils.export(model, args, f, export_params, verbose, training,
      File "/opt/conda/lib/python3.8/site-packages/torch/onnx/utils.py", line 118, in export
        _export(model, args, f, export_params, verbose, training, input_names, output_names,
      File "/opt/conda/lib/python3.8/site-packages/torch/onnx/utils.py", line 738, in _export
        proto, export_map, val_use_external_data_format = graph._export_onnx(
    RuntimeError: ONNX export failed: Couldn't export Python operator _DeformConv
    
    
    Defined at:
    /yolov7_d2/detectron2/detectron2/layers/deform_conv.py(394): forward
    /opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py(1098): _slow_forward
    /opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py(1110): _call_impl
    /yolov7_d2/yolov7/modeling/backbone/resnetvd.py(105): forward
    /opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py(1098): _slow_forward
    /opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py(1110): _call_impl
    /opt/conda/lib/python3.8/site-packages/torch/nn/modules/container.py(141): forward
    /opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py(1098): _slow_forward
    /opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py(1110): _call_impl
    /yolov7_d2/yolov7/modeling/backbone/resnetvd.py(509): forward
    /opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py(1098): _slow_forward
    /opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py(1110): _call_impl
    /yolov7_d2/yolov7/modeling/meta_arch/sparseinst.py(147): forward
    /opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py(1098): _slow_forward
    /opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py(1110): _call_impl
    /opt/conda/lib/python3.8/site-packages/torch/jit/_trace.py(118): wrapper
    /opt/conda/lib/python3.8/site-packages/torch/jit/_trace.py(127): forward
    /opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py(1110): _call_impl
    /opt/conda/lib/python3.8/site-packages/torch/jit/_trace.py(1166): _get_trace_graph
    /opt/conda/lib/python3.8/site-packages/torch/onnx/utils.py(391): _trace_and_get_graph_from_model
    /opt/conda/lib/python3.8/site-packages/torch/onnx/utils.py(440): _create_jit_graph
    /opt/conda/lib/python3.8/site-packages/torch/onnx/utils.py(499): _model_to_graph
    /opt/conda/lib/python3.8/site-packages/torch/onnx/utils.py(719): _export
    /opt/conda/lib/python3.8/site-packages/torch/onnx/utils.py(118): export
    /opt/conda/lib/python3.8/site-packages/torch/onnx/__init__.py(305): export
    export.py(285): <module>
    
    opened by XYudong 3
Releases(v1.0)
  • v1.0(Jun 20, 2022)

  • v.0.0.3(May 10, 2022)

    Since YOLOv7 not just yolo, it aims at a mature modeling tool based on detectron2, so I will continue add more models into it. It now also received a lot of PR from community. Let's build YOLOv7 togather and make a lot of experiment on it.

    Current is a Beta release, it becoming close to a Stable release version, we are keep on doing on that.

    • YOLOX-Pose model added, now you should able to training YOLOx with Pose head;
    • SOLOv2 training support, now you can using YOLOv7 training SOLOv2;
    • Sparseinst training support, now you can train sparseinst, we also support ONNX exportation;
    • Some bugs fixes in sparseinst;

    Feature you will see in the next release:

    • Int8 quantization on Sparseinst, CPU Realtime instance segmentation, just from yolov7!
    • TensorRT on SOLOv2 and Sparseinst;
    • Pretrain model of YOLOX-Pose.

    Just start and fork it! Join the community!

    Source code(tar.gz)
    Source code(zip)
  • v0.0.2(Mar 25, 2022)

    updates on:

    • remove dependencies of mobilecv, which is now optional;
    • remove nbnb deps now is optional;
    • add setup.py for install, usage can be outscope anywhere and any other project.
    Source code(tar.gz)
    Source code(zip)
Owner
JinTian
You know who I am.
JinTian
Session-aware Item-combination Recommendation with Transformer Network

Session-aware Item-combination Recommendation with Transformer Network 2nd place (0.39224) code and report for IEEE BigData Cup 2021 Track1 Report EDA

Tzu-Heng Lin 6 Mar 10, 2022
Code for our CVPR2021 paper coordinate attention

Coordinate Attention for Efficient Mobile Network Design (preprint) This repository is a PyTorch implementation of our coordinate attention (will appe

Qibin (Andrew) Hou 726 Jan 05, 2023
A python script to dump all the challenges locally of a CTFd-based Capture the Flag.

A python script to dump all the challenges locally of a CTFd-based Capture the Flag. Features Connects and logins to a remote CTFd instance. Dumps all

Podalirius 77 Dec 07, 2022
Learn the Deep Learning for Computer Vision in three steps: theory from base to SotA, code in PyTorch, and space-repetition with Anki

DeepCourse: Deep Learning for Computer Vision arthurdouillard.com/deepcourse/ This is a course I'm giving to the French engineering school EPITA each

Arthur Douillard 113 Nov 29, 2022
Sematic-Segmantation - Semantic Segmentation on MIT ADE20K dataset in PyTorch

Semantic Segmentation on MIT ADE20K dataset in PyTorch This is a PyTorch impleme

Berat Eren Terzioğlu 4 Mar 22, 2022
Python scripts for performing 3D human pose estimation using the Mobile Human Pose model in ONNX.

Python scripts for performing 3D human pose estimation using the Mobile Human Pose model in ONNX.

Ibai Gorordo 99 Dec 31, 2022
Paddle implementation for "Cross-Lingual Word Embedding Refinement by ℓ1 Norm Optimisation" (NAACL 2021)

L1-Refinement Paddle implementation for "Cross-Lingual Word Embedding Refinement by ℓ1 Norm Optimisation" (NAACL 2021) 🙈 A more detailed readme is co

Lincedo Lab 4 Jun 09, 2021
Hard cater examples from Hopper ICLR paper

CATER-h Honglu Zhou*, Asim Kadav, Farley Lai, Alexandru Niculescu-Mizil, Martin Renqiang Min, Mubbasir Kapadia, Hans Peter Graf (*Contact: honglu.zhou

NECLA ML Group 6 May 11, 2021
Adjust Decision Boundary for Class Imbalanced Learning

Adjusting Decision Boundary for Class Imbalanced Learning This repository is the official PyTorch implementation of WVN-RS, introduced in Adjusting De

Peyton Byungju Kim 16 Jan 04, 2023
Aesara is a Python library that allows one to define, optimize, and efficiently evaluate mathematical expressions involving multi-dimensional arrays.

Aesara is a Python library that allows one to define, optimize, and efficiently evaluate mathematical expressions involving multi-dimensional arrays.

Aesara 898 Jan 07, 2023
[AAAI 2022] Separate Contrastive Learning for Organs-at-Risk and Gross-Tumor-Volume Segmentation with Limited Annotation

A paper Introduction This is an official release of the paper Separate Contrastive Learning for Organs-at-Risk and Gross-Tumor-Volume Segmentation wit

Jiacheng Wang 14 Dec 08, 2022
Related resources for our EMNLP 2021 paper

Plan-then-Generate: Controlled Data-to-Text Generation via Planning Authors: Yixuan Su, David Vandyke, Sihui Wang, Yimai Fang, and Nigel Collier Code

Yixuan Su 61 Jan 03, 2023
Code for the paper "Balancing Training for Multilingual Neural Machine Translation, ACL 2020"

Balancing Training for Multilingual Neural Machine Translation Implementation of the paper Balancing Training for Multilingual Neural Machine Translat

Xinyi Wang 21 May 18, 2022
Gradient Step Denoiser for convergent Plug-and-Play

Source code for the paper "Gradient Step Denoiser for convergent Plug-and-Play"

Samuel Hurault 11 Sep 17, 2022
Improving Factual Consistency of Abstractive Text Summarization

Improving Factual Consistency of Abstractive Text Summarization We provide the code for the papers: "Entity-level Factual Consistency of Abstractive T

61 Nov 27, 2022
TensorFlow Tutorial and Examples for Beginners (support TF v1 & v2)

TensorFlow Examples This tutorial was designed for easily diving into TensorFlow, through examples. For readability, it includes both notebooks and so

Aymeric Damien 42.5k Jan 08, 2023
Pytorch reimplementation of PSM-Net: "Pyramid Stereo Matching Network"

This is a Pytorch Lightning version PSMNet which is based on JiaRenChang/PSMNet. use python main.py to start training. PSM-Net Pytorch reimplementatio

XIAOTIAN LIU 1 Nov 25, 2021
PyTorch EO aims to make Deep Learning for Earth Observation data easy and accessible to real-world cases and research alike.

Pytorch EO Deep Learning for Earth Observation applications and research. 🚧 This project is in early development, so bugs and breaking changes are ex

earthpulse 28 Aug 25, 2022
Posterior temperature optimized Bayesian models for inverse problems in medical imaging

Posterior temperature optimized Bayesian models for inverse problems in medical imaging Max-Heinrich Laves*, Malte Tölle*, Alexander Schlaefer, Sandy

Artificial Intelligence in Cardiovascular Medicine (AICM) 6 Sep 19, 2022
《LXMERT: Learning Cross-Modality Encoder Representations from Transformers》(EMNLP 2020)

The Most Important Thing. Our code is developed based on: LXMERT: Learning Cross-Modality Encoder Representations from Transformers

53 Dec 16, 2022