YOLOv5 Is Here: State-of-the-Art Object Detection at 140 FPS

栏目: IT技术 · 发布时间: 3年前

内容简介:Less than 50 days after the release YOLOv4, YOLOv5 improves state-of-the-art for realtime object detection.Realtime object detection is improving quickly. TheOn March 18, Google

Less than 50 days after the release YOLOv4, YOLOv5 improves state-of-the-art for realtime object detection.

Realtime object detection is improving quickly. The rate of improvement is improving even more quickly. The results are stunning.

On March 18, Google open sourced their implementation ofEfficientDet, a fast-training model with various sizes, one of which offers realtime output. On April 23, Alexey Bochoviskiy et al. open sourced YOLOv4. On June 9, Glenn Jocher open sourced an implementation ofYOLOv5.

Just Looking to Train YOLOv5?

Skip this info post and jump straight to ourYOLOv5 tutorial. You'll have a trained YOLOv5 model on your custom data in minutes.

The Evolution of YOLO Models

YOLO (You Only Look Once) is a family of models that PJ Reddie originally coined with a 2016 publication . YOLO models are infamous for being highly performant yet incredibly small – making them ideal candidates for realtime conditions and on-device deployment environments.

PJ Reddie's research team is responsible for subsequently introducing YOLOv2 andYOLOv3, both of which made continued improvement in both model performance and model speed. In February 2020, PJ Reddie noted he would discontinue research in computer vision.

In April 2020, Alexey Bochkovskiy, Chien-Yao Wang, and Hong-Yuan Mark Liao introducedYOLOv4, demonstrating impressive gains.

YOLOv5 Is Here: State-of-the-Art Object Detection at 140 FPS
Image via the YOLOv4 paper .

Notably, many of YOLOv4's improvements came from improveddata augmentation as much as model architecture. (We've written abreakdown on YOLOv4 as well as how to train a YOLOv4 model on custom objects .)

YOLOv5: The Leader in Realtime Object Detection

Glenn Jocher releasedYOLOv5 with a number of differences and improvements. (Notably, Glenn is the creator of mosaic augmentation, which is an included technique in what improved YOLOv4.) The release of YOLOv5 includes five different models sizes: YOLOv5s (smallest), YOLOv5m, YOLOv5l, YOLOv5x (largest).

Let's breakdown YOLOv5. How does YOLOv5 compare?

First, this is the first native release of models in the YOLO family to be written in PyTorch first rather than PJ Reddie's Darknet. Darknet is an incredibly flexible research framework, but it is not built with production environments in mind. It has a smaller community of users. Taken together, this results in Darknet being more challenging to configure and less production-ready.

Because YOLOv5 is implemented in PyTorch initially, it benefits from the established PyTorch ecosystem: support is simpler, and deployment is easier. Moreover as a more widely known research framework, iterating on YOLOv5 may be easier for the broader research community. This also makes deploying to mobile devices simpler as the model can be compiled to ONNX and CoreML with ease.

Second, YOLOv5 is fast – blazingly fast. In a YOLOv5 Colab notebook , running a Tesla P100, we saw inference times up to 0.007 seconds per image, meaning 140 frames per second (FPS) ! By contrast, YOLOv4 achieved 50 FPS after having been converted to the same Ultralytics PyTorch library.

Third, YOLOv5 is accurate. In our tests on the blood cell count and detection (BCCD) dataset , we achieved roughly 0.895 mean average precision (mAP) after training for just 100 epochs. Admittedly, we saw comparable performance from EfficientDet and YOLOv4, but it is rare to see such across-the-board performance improvements without any loss in accuracy.

Fourth, YOLOv5 is small. Specifically, a weights file for YOLOv5 is 27 megabytes. Our weights file for YOLOv4 (with Darknet architecture) is 244 megabytes. YOLOv5 is nearly 90 percent smaller than YOLOv4. This means YOLOv5 can be deployed to embedded devices much more easily.

Many of these changes are well-summarized in YOLOv5's graphic measuring performance.

YOLOv5 Is Here: State-of-the-Art Object Detection at 140 FPS
YOLO is more accurate and faster than EfficientDet. Credit: Glenn Jocher

Get Started with YOLOv5

We're eager to see what you are able to build with new state-of-the-art detectors.

To that end, we've published a guide on how to train YOLOv5 on a custom dataset , making it quick and easy. If you would like to use standard COCO weights, see this notebook .

You can always visit YOLOv5.com for more resources as they become available.

Stay tuned for additional deeper dives on YOLOv5, and good luck building!

Want to be the first to know about new computer vision tutorials and content like our synthetic dataset creation guide ? Subscribe to our updates :mailbox_with_mail: .

Roboflow accelerates your computer vision workflow through automated annotation quality assurance, universal annotation format conversion (like PASCAL VOC XML to COCO JSON andcreating TFRecords ), team sharing and versioning, and easy integration with popular open source computer vision models . Getting started with your first 1000 images are completely free.


以上就是本文的全部内容,希望本文的内容对大家的学习或者工作能带来一定的帮助,也希望大家多多支持 码农网

查看所有标签

猜你喜欢:

本站部分资源来源于网络,本站转载出于传递更多信息之目的,版权归原作者或者来源机构所有,如转载稿涉及版权问题,请联系我们

ASP.NET 2.0开发指南

ASP.NET 2.0开发指南

郝刚 / 人民邮电出版社 / 2006 / 88.0

本书紧紧围绕ASP.NET 2.0技术精髓展开深入讲解,全书分为6个部分,共18章。第1部分介绍基础知识,包括ASP.NET 2.0概述、Visual Studio 2005集成开发环境、创建ASP.NET应用程序和C# 2.0程序设计基础。第2部分讲解用户界面方面的特性,包括母版页、主题和皮肤、站点导航控件和其他新增服务器控件。第3部分探讨了数据访问方面的内容,包括数据访问技术概述、数据源控件、......一起来看看 《ASP.NET 2.0开发指南》 这本书的介绍吧!

图片转BASE64编码
图片转BASE64编码

在线图片转Base64编码工具

html转js在线工具
html转js在线工具

html转js在线工具

RGB CMYK 转换工具
RGB CMYK 转换工具

RGB CMYK 互转工具