-
Notifications
You must be signed in to change notification settings - Fork 1
2016 05 22 First meeting: research results
Arthur Silber edited this page May 23, 2017
·
8 revisions
Found Datasets (created wiki page here):
- Oxford Building Dataset: 5k images, semi-sorted, no BBs
- Oxford Paris Dataset: 500-1k images, sorted by landmark, no BBs
- LabelMe: has per-pixel labels (-> BB extraction via script possible)
- MIT Scene Parsing Benchmark: semi-useful per-pixel labels (Buildings are often too big / cropped on images)
- TODO Fabian: Fassaden-dataset with per-pixel labels
Discussion:
- Restrict images to dailight only? (many photos found on google images, e.g. Brandenburger Tor, are taken at night)
- Image synthesis - generate test images manually --> use for pre-training to aid model training --> Arthur: 3D renderings / Martin: photogammetry
- Fabian: found a thesis: memory efficient database for mobile image search --> research how its done, anything we can use?
Network Training approach:
- General training e.g. imgnet
- Train especially on "is there a building"
- specific training e.g. for berlin (hand annotated images?)
- Reserach mxnet SSD in detail
- How to build image dataset:
- --> 6a) test segmentation - background versus building / search class "building" - test with images
- how to generate BBs?
- image synthesis
- goal: pretraining
- MaskRCNN?
Who:
- Adrian: Labels to BB (LabelMe dataset)
- Arthur: test SSD with pictures