YOLO – You only look 10647 times

01/16/2022
by   Christian Limberg, et al.
0

With this work we are explaining the "You Only Look Once" (YOLO) single-stage object detection approach as a parallel classification of 10647 fixed region proposals. We support this view by showing that each of YOLOs output pixel is attentive to a specific sub-region of previous layers, comparable to a local region proposal. This understanding reduces the conceptual gap between YOLO-like single-stage object detection models, RCNN-like two-stage region proposal based models, and ResNet-like image classification models. In addition, we created interactive exploration tools for a better visual understanding of the YOLO information processing streams: https://limchr.github.io/yolo_visualization

READ FULL TEXT

page 3

page 4

page 5

page 6

research
06/04/2015

Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks

State-of-the-art object detection networks depend on region proposal alg...
research
01/24/2019

Object Detection based on Region Decomposition and Assembly

Region-based object detection infers object regions for one or more cate...
research
02/12/2019

You Only Look & Listen Once: Towards Fast and Accurate Visual Grounding

Visual Grounding (VG) aims to locate the most relevant region in an imag...
research
05/22/2020

KL-Divergence-Based Region Proposal Network for Object Detection

The learning of the region proposal in object detection using the deep n...
research
04/12/2016

CRAFT Objects from Images

Object detection is a fundamental problem in image understanding. One po...
research
08/29/2016

PVANET: Deep but Lightweight Neural Networks for Real-time Object Detection

This paper presents how we can achieve the state-of-the-art accuracy in ...
research
01/07/2022

Extending One-Stage Detection with Open-World Proposals

In many applications, such as autonomous driving, hand manipulation, or ...

Please sign up or login with your details

Forgot password? Click here to reset