Skip to content

Latest commit

 

History

History
155 lines (132 loc) · 5.49 KB

README.md

File metadata and controls

155 lines (132 loc) · 5.49 KB

Semantic Room Wireframe Detection

This repo contains code to generate Semantic Room Wireframe (SRW) annotations from Structured3D and LSUN. It also holds the implementation of SRW-Net. See the preprint of our paper at arXiv or the published ICPR2022 paper for more details.

Junction annotation explained

Annotated image illustrating false and proper junctions together with semantic line labels.

Clone repo and get submodules

git clone --recurse-submodules git@github.com:DavidGillsjo/SRW-Net.git

alternatively

git clone git@github.com:DavidGillsjo/SRW-Net.git
git submodule init
git submodule update

Docker

We supply a Dockerfile to build a docker image which can run the code. First, modify line 7 so that gpu_arch matches your GPU architecture. See for example this blog post to find your arch code.

Then build and run:

cd docker
./build.sh
./run.sh

You will find your HOME directory mounted to /host_home.

Fix Python path

Add parsing folder to python path for correct imports.

source init_env.sh

Build

To run the network, some C-code needs compiling.

./build.sh

Download Structured3D

To train the network you need access to the fully furnished perspective images from Structured3D. See the Structured3D website for instructions. Then download Structured3D_perspective_full and place in data/Structured3D. You may use the python script download_structured3D.py together with the Chrome extension CurlWget.

Download Annotations and Model Weights

Here you find the annotations together with the pre-trained models to reproduce the result from the paper. We also include the pretrained HAWP model which we used as initialization for the training.

Unzip data.zip to the data folder. You may also put the model weights in the data folder, the rest of this README will assume you did.

Generate Annotations

If you prefer to generate the annotations, then run

python3 preprocessing/structured3D2wireframe.py --help

for instructions.

Inference

There are a number of ways to run inference, see python3 scripts/test.py --help for details.

Run inference on test set

To run on the test set, do

cd scripts
python3 test.py \
--config-file ../config-files/layout-SRW-S3D.yaml \
CHECKPOINT ../data/model_proposal_s3d.pth \
GNN_CHECKPOINT ../data/model_gnn_s3d.pth \
OUTPUT_DIR ../runs/test

To run on the validation data, add the flag --val.

Run inference on your own images

To run on a set of images

cd scripts
python3 test.py \
--config-file ../config-files/layout-SRW-S3D.yaml \
--img-folder <my-image-folder> \
CHECKPOINT ../data/model_proposal_s3d.pth \
GNN_CHECKPOINT ../data/model_gnn_s3d.pth \
OUTPUT_DIR <my-output-folder>

and the result will be placed in <my-output-folder>, see layout-SRW-S3D.yaml for default value.

Train

The Predictor and the Refinement module are trained separately in the following steps

Train Predictor from the HAWP model weights

cd scripts
python3 train.py \
--config-file ../config-files/Pred-SRW-S3D.yaml \
CHECKPOINT ../data/model_hawp.pth \
OUTPUT_DIR ../runs/predictor \
TRANSFER_LEARN True

Model checkpoints will be in runs/predictor/<datetime>/model_<epoch>.pth.

To monitor the training you may start a tensorboard instance (also a docker container)

./start_tensorboard.sh

Then open your browser and go to http://localhost:6006.

Generate intermediate dataset

Now we use the trained model to generate a dataset for the Refinement module.

cd preprocessing
python3 generate_gnn_dataset.py \
--config-file ../config-files/Pred-SRW-S3D.yaml \
--output-dir ../data/Structured3D_wf_open_doors_1mm/gnn_npz \
CHECKPOINT ../runs/predictor/<datetime>/model_39.pth \
IMS_PER_BATCH 1

The data we used is available for download.

Train the Refinement Module

Finally, we train the Refinement GCN.

cd scripts
python3 train.py \
--config-file ../config-files/GNN-SRW-S3D.yaml \
OUTPUT_DIR ../runs/refinement

and you will find the model weights at ../runs/refinement/<datetime>/model_09.pth.

Citation

If you use it in your research, please cite

@INPROCEEDINGS{srw-net,
  author={Gillsjö, David and Flood, Gabrielle and Åström, Kalle},
  booktitle={2022 26th International Conference on Pattern Recognition (ICPR)}, 
  title={Semantic Room Wireframe Detection from a Single View}, 
  year={2022},
  pages={1886-1893},
  doi={10.1109/ICPR56361.2022.9956252}
}