Skip to content

Commit 16d2caf

Browse files
authored
Merge pull request #574 from microsoft/staging
Small edits to readmes (#573)
2 parents 109f42d + 9ada614 commit 16d2caf

File tree

2 files changed

+7
-7
lines changed

2 files changed

+7
-7
lines changed

README.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
<img src="scenarios/media/logo_cvbp.png" align="right" alt="" width="300"/>
22

33
```diff
4-
+ Update June 24: Added action recognition as new core scenario.
4+
+ Update June 24: Added action recognition as new core scenario.
55
+ Object tracking coming soon (in 2-4 weeks).
66
```
77

@@ -37,7 +37,7 @@ Our target audience for this repository includes data scientists and machine lea
3737
To get started, navigate to the [Setup Guide](SETUP.md), which lists
3838
instructions on how to setup the compute environment and dependencies needed to run the
3939
notebooks in this repo. Once your environment is setup, navigate to the
40-
[Scenarios](scenarios) folder and start exploring the notebooks.
40+
[Scenarios](scenarios) folder and start exploring the notebooks. We recommend to start with the *image classification* notebooks, since this introduces concepts which are also used by the other scenarios (e.g. pre-training on ImageNet).
4141

4242
Alternatively, we support Binder
4343
[![Binder](https://mybinder.org/badge_logo.svg)](https://mybinder.org/v2/gh/PatrickBue/computervision-recipes/master?filepath=scenarios%2Fclassification%2F01_training_introduction_BINDER.ipynb)

scenarios/tracking/README.md

+5-5
Original file line numberDiff line numberDiff line change
@@ -1,23 +1,23 @@
11
# Multi-Object Tracking
22

33
```diff
4-
+ June 2020: This work is ongoing.
4+
+ June 2020: All notebooks/code in this directory is work-in-progress and might not fully execute.
55
```
66

77
This directory provides examples and best practices for building multi-object tracking systems. Our goal is to enable the users to bring their own datasets and train a high-accuracytracking model easily. While there are many open-source trackers available, we have implemented the [FairMOT tracker](https://github.com/ifzhang/FairMOT) specifically, as its algorithm has shown competitive tracking performance in recent MOT benchmarking challenges, at fast inference speed.
88

99
## Technology
10-
Multi-object-tracking (MOT) is one of the hot research topics in Computer Vision, due to its wide applications in autonomous driving, traffic surveillance, etc. It builds on object detection technology, in order to detect and track all objects in a dynamic scene over time. Inferring target trajectories correctly across successive image frames remains challenging: occlusion happens when objects overlap; the number of and appearance of objects can change. Compared to object detection algorithms, which aim to output rectangular bounding boxes around the objects, MOT algorithms additionally associated an ID number to each box to identify that specific object across the image frames.
10+
Multi-object-tracking (MOT) is one of the hot research topics in Computer Vision, due to its wide applications in autonomous driving, traffic surveillance, etc. It builds on object detection technology, in order to detect and track all objects in a dynamic scene over time. Inferring target trajectories correctly across successive image frames remains challenging: occlusion happens when objects overlap; the number of and appearance of objects can change. Compared to object detection algorithms, which aim to output rectangular bounding boxes around the objects, MOT algorithms additionally associated an ID number to each box to identify that specific object across the image frames.
1111

1212
As seen in the figure below ([Ciaparrone, 2019](https://arxiv.org/pdf/1907.12740.pdf)), a typical multi-object-tracking algorithm performs part or all of the following steps:
1313
* Detection: Given the input raw image frames (step 1), the detector identifies object(s) on each image frame as bounding box(es) (step 2).
14-
* Feature extraction/motion prediction: For every detected object, visual appearance and motion features are extracted (step 3). Sometimes, a motion predictor (e.g. Kalman Filter) is also added to predict the next position of each tracked target.
15-
* Affinity: The feature and motion predictions are used to calculate similarity/distance scores between pairs of detections and/or tracklets, or the probabilities of detections belonging to a given target or tracklet (step 4).
14+
* Feature extraction/motion prediction: For every detected object, visual appearance and motion features are extracted (step 3). Sometimes, a motion predictor (e.g. Kalman Filter) is also added to predict the next position of each tracked target.
15+
* Affinity: The feature and motion predictions are used to calculate similarity/distance scores between pairs of detections and/or tracklets, or the probabilities of detections belonging to a given target or tracklet (step 4).
1616
* Association: Based on these scores/probabilities, a specific numerical ID is assigned to each detected object as it is tracked across successive image frames (step 5).
1717

1818
<p align="center">
1919
<img src="./media/figure_motmodules2.jpg" width="700" align="center"/>
20-
</p>
20+
</p>
2121

2222

2323
## State-of-the-art (SoTA)

0 commit comments

Comments
 (0)