pith. machine review for the scientific record. sign in

hub

MMDetection: Open mmlab detection toolbox and benchmark

21 Pith papers cite this work. Polarity classification is still indexing.

21 Pith papers citing it
abstract

We present MMDetection, an object detection toolbox that contains a rich set of object detection and instance segmentation methods as well as related components and modules. The toolbox started from a codebase of MMDet team who won the detection track of COCO Challenge 2018. It gradually evolves into a unified platform that covers many popular detection methods and contemporary modules. It not only includes training and inference codes, but also provides weights for more than 200 network models. We believe this toolbox is by far the most complete detection toolbox. In this paper, we introduce the various features of this toolbox. In addition, we also conduct a benchmarking study on different methods, components, and their hyper-parameters. We wish that the toolbox and benchmark could serve the growing research community by providing a flexible toolkit to reimplement existing methods and develop their own new detectors. Code and models are available at https://github.com/open-mmlab/mmdetection. The project is under active development and we will keep this document updated.

hub tools

citation-role summary

background 1 method 1

citation-polarity summary

clear filters

representative citing papers

SignDATA: Data Pipeline for Sign Language Translation

cs.CV · 2026-04-22 · unverdicted · novelty 6.0

SignDATA provides a reproducible, config-driven preprocessing toolkit that converts heterogeneous sign language corpora into standardized pose or video outputs using interchangeable backends and privacy-aware options.

Portable Active Learning for Object Detection

cs.CV · 2026-05-11 · unverdicted · novelty 5.0

PAL is a portable active learning method for object detection that uses class-specific logistic classifiers for uncertainty and image-level diversity to select annotation batches, showing better label efficiency than baselines on COCO, VOC, and BDD100K.

Seed1.5-VL Technical Report

cs.CV · 2025-05-11 · unverdicted · novelty 4.0

Seed1.5-VL is a compact multimodal model that sets new records on dozens of vision-language benchmarks and outperforms prior systems on agent-style tasks.

citing papers explorer

Showing 1 of 1 citing paper after filters.

  • Seed1.5-VL Technical Report cs.CV · 2025-05-11 · unverdicted · none · ref 14

    Seed1.5-VL is a compact multimodal model that sets new records on dozens of vision-language benchmarks and outperforms prior systems on agent-style tasks.