Pytorch implementation for "Learning Open-World Object Proposals without Learning to Classify" (RA-L and ICRA 2022)
Dahun Kim, Tsung-Yi Lin, Anelia Angelova, In So Kweon, and Weicheng Kuo.
@article{kim2021oln,
title={Learning Open-World Object Proposals without Learning to Classify},
author={Kim, Dahun and Lin, Tsung-Yi and Angelova, Anelia and Kweon, In So and Kuo, Weicheng},
journal={IEEE Robotics and Automation Letters (RA-L)},
year={2022}
}
Humans can recognize novel objects in this image despite having never seen them before. “Is it possible to learn open-world (novel) object proposals?” In this paper we propose Object Localization Network (OLN) that learns localization cues instead of foreground vs background classification. Only trained on COCO, OLN is able to propose many novel objects (top) missed by Mask R-CNN (bottom) on an out-of-sample frame in an ego-centric video.
We train OLN on COCO VOC categories, and test on non-VOC categories. Note our AR@k evaluation does not count those proposals on the 'seen' classes into the budget (k), to avoid evaluating recall on see-class objects.
Method | AUC | AR@10 | AR@30 | AR@100 | AR@300 | AR@1000 | Download |
---|---|---|---|---|---|---|---|
OLN-Box | 24.8 | 18.0 | 26.4 | 33.4 | 39.0 | 45.0 | model |
This repo is tested under Python 3.7, PyTorch 1.7.0, Cuda 11.0, and mmcv==1.2.5.
This repo is built based on mmdetection.
You can use following commands to create conda env with related dependencies.
conda create -n oln python=3.7 -y
conda activate oln
conda install pytorch=1.7.0 torchvision cudatoolkit=11.0 -c pytorch -y
pip install mmcv-full==1.2.7
pip install -r requirements.txt
pip install -v -e .
Please also refer to get_started.md for more details of installation.
COCO dataset is available from official websites. It is recommended to download and extract the dataset somewhere outside the project directory and symlink the dataset root to $OLN/data as below.
object_localization_network
├── mmdet
├── tools
├── configs
├── data
│ ├── coco
│ │ ├── annotations
│ │ ├── train2017
│ │ ├── val2017
│ │ ├── test2017
Our trained models are available for download here. Place it under trained_weights/latest.pth
and run the following commands to test OLN on COCO dataset.
# Multi-GPU distributed testing
bash tools/dist_test_bbox.sh configs/oln_box/oln_box.py \
trained_weights/latest.pth ${NUM_GPUS}
# OR
python tools/test.py configs/oln_box/oln_box.py work_dirs/oln_box/latest.pth --eval bbox
# Multi-GPU distributed training
bash tools/dist_train.sh configs/oln_box/oln_box.py ${NUM_GPUS}
If you have any questions regarding the repo, please contact Dahun Kim ([email protected]) or create an issue.