-
Notifications
You must be signed in to change notification settings - Fork 55
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
What is the structure of the split-#-#-#-minsc-2.0.h5 file? #28
Comments
Going back over the paper again, I think it's possible that this is the section that the above code covers, but I'd still appreciate confirmation on this point if anyone can provide it. |
Sorry, we cannot support that part for now. |
I'm not asking for any kind of code support. At this point I've made some progress and I believe I have all intermediate h5 training files including the split-kp-minsc file correctly generated from incorporating my own code, interpreting the documentation provided, and reading through the paper (both of which were very helpful). My request is for just enough information regarding the h5 file compositions to make sure that I am building the network from the same kind of inputs in the same way the paper describes, in order to validate the training process and to avoid misrepresenting what the network is and isn't capable of learning. It would assist myself and others in validating your results if there were details about the h5 file structures generated during training which could be provided somewhere. A brief set of key:value descriptions for the file structures would likely be sufficient. Once I'm certain I've got the training pipeline correctly implemented (which I think I already do) I would happily contribute the code I have for building custom training datasets as a pull request. |
@etrulls do you have any of these things left? e.g. for a small example? |
Hi, I am also planning to implement LIFT on Kitti datasets, since my project is about traffic objects; Regards, |
@4xle Hello, have you solved this problem? I also have this problem. Can I discuss it with you? |
I think I have it figured out but I'm still waiting to hear back from @etrulls to confirm that I've got it right. It's possible I have something very similar but not exactly the same which would give me different results. Once I'm certain I've got it working properly I plan to submit a detailed PR. |
Hi, Did you figure out training with your own datasets? If so, maybe we can discuss a bit on this problem. Regards, |
Sorry, I missed this (feel free to send an e-mail directly next time). It's just a list with the indices of the points used for training, validation, and testing.
Example: https://www.dropbox.com/s/uynuszoatimsbe1/split-60-20-20-minsc-2.0-fix.h5?dl=0 |
Hi @4xle, do you still plan to submit a PR about the training pipeline? |
Excuse me, @4xle |
@punisher220 Certainly, though if we can keep the discussion on here it may help other people. @hudsonmartins I did intend to submit a PR, but thought I lost the work due to a bad hard drive. However, past me appears to have had a better backup policy than I remembered implementing, and I have found a copy at least some of the changes that I made for tooling, though they may be no longer up to date. Assuming I can figure out what I did and document the changes, I will make that PR. |
PR #42 has been made. |
@4xle Thank you for your hard work to the PR. |
Excuse me, @4xle I still can not solve the training data construction problem. Then as the eccv.py displays, at Line 410 the code will turn into "get_scale_hist(train_data_dir, param)" in helper.py I noticed that from Line 836 to Line 840 in helper.py you wrote comment to use readAsciiSiftFile function Could you please help me with more hints? Thank you. |
Excuse me, @kmyid At present, with the help of PR from 4xle and the example folder uploaded by etrulls, I come closer to the LIFT training data construction. 1.In your helper.py(in datasets/eccv2016 folder) you set a function named get_scale_hist(train_data_dir, param) to generate the scales-histogram-minsc-2.0.h5 file. (1)If so, I have tried to load my ".mat" files as you did at Line 812 with scipy.io.loadmat(kp_file_name)['TFeatures'], but an error occurred:
Then I noticed that my ".mat" file may be ASCII files then I tried to load them with readAsciiSiftFile(filepath) from 4xle's PR helper.py.
Now I got no ideas about how to load the ".mat" files(also the ".sift" files). This is the main problem for me. (2)If not, how did you get those "_p.mat" files which can be loaded by scipy.io.loadmat function. I even doubt that we are using different Visual SFM software.
|
I'll check through my notes to see if I forgot to include any steps in the PR Readme and see if I have the data backed up for testing. However, I do not have a Windows machine I can test this on currently (which is where it was originally written).
Bradley Koskowich
Research Assistant
Spatial {Query} Lab
MANTIS Lab
…________________________________
From: punisher220 <[email protected]>
Sent: Thursday, April 16, 2020 8:00:03 AM
To: cvlab-epfl/tf-lift <[email protected]>
Cc: Koskowich, Bradley <[email protected]>; Mention <[email protected]>
Subject: Re: [cvlab-epfl/tf-lift] What is the structure of the split-#-#-#-minsc-2.0.h5 file? (#28)
Excuse me, @4xle<https://github.com/4xle>
I am sorry to bother you after you uploaded your PR files.
I still can not solve the training data construction problem.
I got stuck in creating files with name "kpfilename" at Line 841 in helper.py with string as: "-kp-minsc-2.0.h5"
I put images files(including .jpg, .mat and .sift file for each image) into train_data_dir,
I am not sure whether they are enough for the training data h5 files construction.
I need to confirm whether I have missed some important files in Visual SFM pipeline.
It is my first question.
Then as the eccv.py displays, at Line 410 the code will turn into "get_scale_hist(train_data_dir, param)" in helper.py
because I did not generate the "hist_file_path" file
Next, at Line 843 in helper.py the code requires the "kpiflename" (as "r") readable to continue with hist file generation.
My second question is: How to obtain the "kpfilename" files from the files in train_data_dir?
I did not find out any other places to edit those "kpfilename" files with "w".
I noticed that from Line 836 to Line 840 in helper.py you wrote comment to use readAsciiSiftFile function
to load the sift file. But I still could not work out the approach to generate the "-kp-minsc-2.0.h5" files.
Could you please help me with more hints? Thank you.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub<#28 (comment)>, or unsubscribe<https://github.com/notifications/unsubscribe-auth/AFTYPVCJCHHGEV2RFCTEINTRM36NHANCNFSM4GFAXORQ>.
|
@punisher220 The _P.mat files should be the same as the .mat files. The ReadAsciiSiftFile was for reading the .sift files, but I think in the end that function is not called anywhere (it was primarily for debugging/development). @kmyid I've found some additional notes & scripts for pre-processing data prior to training, I'll clean them up and add them to the PR, then finish cleaning up the PR. |
Hello, @4xle @kmyid The matching for every image pairs(generating .mat files and .sift files) Even with the help of gpu acceleration, it took me nearly 30 hours to finish the matching step for 1700-image dataset in Visual SFM. What a desperate wait. I wonder whether there are some other approaches to speedup the matching process in VSFM settings. |
@punisher220 You can tweak the parameters of VSFM uses for matching to reduce the GPU computation load, though note that may affect the quality of the final results. If your input is video, you can use the sequence match option to reduce the number of match operations greatly. VSFM is also more affected by available RAM than VRAM as it does the keypoint identification on the GPU, but the matching on CPU if I recall correctly. A bigger GPU will get the identification done faster and be able to keep more images for identifying against in VRAM, but then the CPU is slammed with a matching backlog by comparison, especially if the images are large and have many keypoints. You can call VSFM on the command line without the GUI to save on some overhead that incurs, or you can write scripts which compute the keypoints more efficiently yourself and then call those directly from VSFM or run them against your dataset, importing the results with the images to then do the matching in VSFM, or do the matching yourself and saving it to then load in VSFM. However, VSFM is pretty fast as it is for what it does, so unless you have access to parallelized resources YMMV. On a separate note, this is the second time that this project has killed a hard drive of mine through excessive use, it seems. I was rebuilding the Piccadilly dataset in VSFM to then test the code with however that will be on hold until I can figure out how I'm going to get the drive I need replaced as I was using a remote machine which is not currently physically accessible due to COVID. |
I am very sorry to hear that. Yes, the code is nowhere near perfect :( Could be much better... |
I don't think the code is responsible for it, its just the nature of the work. Lots of repeated tiny accesses are a quick way to wear out a non-SSD disk. I also do a lot of this kind of work and we have plenty of spares to accommodate the heavier projects, it's just a case of not being able to physically get at the machine currently being the biggest hurdle.
|
Hello, I have a small question about Visual SFM models information in nvm file. I tried Visual SFM with several datasets(just hundreds of images) I want to confirm: Thank you. And, @4xle Thanks for your reply. I am sorry to hear that you are in hurdle due to COVID. |
We probably just used the largest. As long as you have enough data you'll be fine. You probably want more variation in terms of different scenes anyway. |
The nvm file saves all models in a format which has identifiers for model starts, IIRC there's no difference between a file with one model or more except for that.
Bradley Koskowich
Research Assistant
Spatial {Query} Lab
MANTIS Lab
…________________________________
From: Eduard Trulls <[email protected]>
Sent: Wednesday, April 29, 2020 8:08:52 AM
To: cvlab-epfl/tf-lift <[email protected]>
Cc: Koskowich, Bradley <[email protected]>; Mention <[email protected]>
Subject: Re: [cvlab-epfl/tf-lift] What is the structure of the split-#-#-#-minsc-2.0.h5 file? (#28)
We probably just used the largest. As long as you have enough data you'll be fine. You probably want more variation in terms of different scenes anyway.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub<#28 (comment)>, or unsubscribe<https://github.com/notifications/unsubscribe-auth/AFTYPVFCIC23BACPZP3HI5DRPARGJANCNFSM4GFAXORQ>.
|
Either one should technically work, but you'll probably get better results training from the multi-model as there is more data.
Bradley Koskowich
Research Assistant
Spatial {Query} Lab
MANTIS Lab
…________________________________
From: punisher220 <[email protected]>
Sent: Wednesday, April 29, 2020 8:49:06 AM
To: cvlab-epfl/tf-lift <[email protected]>
Cc: Koskowich, Bradley <[email protected]>; Mention <[email protected]>
Subject: Re: [cvlab-epfl/tf-lift] What is the structure of the split-#-#-#-minsc-2.0.h5 file? (#28)
I am sure about: When you finish 3D reconstruction in VSFM, the nvm file will contain all models information. No matter which .ply model files you choose to generate from VSFM, the nvm file will not change.
I did VSFM with small dataset and the reconstruction result was a single-model nvm file.
But when I tried VSFM with Travel-image datasets the nvm file contains many models.
Based on that, I want to confirm whether it is proper to use the multi-models nvm file directly or pick up the part data of largest one from the nvm file and go on.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub<#28 (comment)>, or unsubscribe<https://github.com/notifications/unsubscribe-auth/AFTYPVFS7CY6OV23DECN5KDRPAV5FANCNFSM4GFAXORQ>.
|
I am sure about: When you finish 3D reconstruction in VSFM, the nvm file will contain all models information. No matter which .ply model files you choose to generate from VSFM, the nvm file will not change. I did VSFM with small dataset and the reconstruction result was a single-model nvm file. Based on that, I want to confirm whether it is proper to use the multi-models nvm file directly or pick up the part data of largest one from the nvm file and go on. However, Either way I choose, I still can not work out the training data required h5 files generation. |
Hello, I did test part of the tf-lift repo with the model provided and the result is impressive. I am trying to complete training with a small dataset to make sure the necessary steps for the LIFT training. How to make the complex process done? I admit the question may be hard to solve but I still appreciate for more help sincerely. |
Hi, @4xle |
Hi, @etrulls @kmyid @4xle I feel sorry to bother you.
In my opinion, this part described you did feature points mark with VSFM as: Question 1: Question 2: Question 3: |
I'm working on a created a documented training dataset pipeline for this network to see how different datasets might induce different learning outcomes, and I've been able to follow the process for generating training data using VisualSFM fairly well with some educated guesses combined with the descriptions in the paper and code documentation, which is pretty good. I am attempting to recreate the training pipeline as closely as possible to the original work with some additional documentation which I would not mind turning into a PR once it is finished and I know that I've got it right.
However, there is a hole regarding this section:
This collection of valid indices and the concept of mutually excluding data is not referred to anywhere in the paper or elsewhere in the code, or if it is I cannot find it. I can tell by the name it is supposed to be some collection of ids for point features (my guess would be for SfM pointids?), but given its location in the code it is unclear how those values would be determined. Also, the access section for the dump files:
before the comparison is done has no obvious correlation with the "valid_keypoints" and "other_keypoints" keys in the kp-minsc files, so I'm drawing a blank as to what those values could be.
Could someone who is familiar with these files provide:
Thanks,
4xle
The text was updated successfully, but these errors were encountered: