-
Notifications
You must be signed in to change notification settings - Fork 28
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Supposed file corruption during Pixel Clustering step 2.2: Assign pixel SOM clusters #688
Comments
@sventrx I'm assuming no errors were encountered during the After deleting your current
|
@sventrx I had another user encounter an error like this, their problem ended up being a multithreading issue with setting the number of cores. I'll link the issue I just opened here: #695. If you need to test a fix out immediately, you'll need to modify |
I have encountered a similar error, however neither of the solutions mentioned was able to solve this problem. I am running it in a conda environment and it works successfully with 10 markers, but upon increasing the number of markers for the pixel clustering, it is now stuck on this step. One of the marker files was potentially corrupt, but upon removing it still did not work, and it seems unlikely that many of the single channel tiff files are corrupt. Do you happen to have any other suggests about what could cause this error in this case? |
@jonhsussman we recently updated our repo to include more detailed error messages in case of FOV processing errors, can you ensure you have the most recent changes by running |
I ran it using the new files, and the current error message is below. Of note, perhaps I should mention that the preceding step, I initially received an error of: Error in Please let me know if you have any other thoughts here! |
@jonhsussman for the The |
Thanks for you reply here! Regarding the SOM function, I was not using Docker, but instead using conda, and thus referencing a native version of R/4.2.1, on which I have installed all the necessary R dependencies. Did you modify the SOM function, and if so, are you able to provide the most up-to-date version of the SOM function for implementation onnative R installation.
Thanks also for for your suggestion on the fovStatuses error. I will try this shortly, but wanted to see if you have some information about the correct SOM function for native R. But about this error, although in the previous version I wasn't running into that exact same error message it was still giving me an error at that same step but only the message simply changed.
…On Tue, Sep 20, 2022, 4:03 PM alex-l-kong ***@***.***> wrote:
@jonhsussman
<https://urldefense.com/v3/__https://github.com/jonhsussman__;!!LIr3w8kk_Xxm!rU78-YwxBk5HU9Tt4BdG-ghqm8L2sojiUFmNTnuaVjRlr2BY1GLPWHHtz8TvU-4Ax6s-I8P7oLzvYMLbD_fMYd0S$>
for the SOM error, I would say wait a bit until we get a new Docker image
pushed. It seems like you may be using an older image which didn't
implement the map = FALSE argument in the SOM function.
The fovStatuses[i, 'status'] == 1 error is surprising, since that seems
to imply the fovStatuses data frame being returned on your end doesn't
contain the column 'status' or is placing a NULL value there. I'm
assuming you didn't run into this issue prior to the updates, can you try
printing out fovStatuses and seeing what happening?
—
Reply to this email directly, view it on GitHub
<https://urldefense.com/v3/__https://github.com/angelolab/ark-analysis/issues/688*issuecomment-1252844996__;Iw!!LIr3w8kk_Xxm!rU78-YwxBk5HU9Tt4BdG-ghqm8L2sojiUFmNTnuaVjRlr2BY1GLPWHHtz8TvU-4Ax6s-I8P7oLzvYMLbD8xOJB68$>,
or unsubscribe
<https://urldefense.com/v3/__https://github.com/notifications/unsubscribe-auth/AKIGYMWYID7OI7TDRI47OQDV7IKADANCNFSM6AAAAAAQE7D2VA__;!!LIr3w8kk_Xxm!rU78-YwxBk5HU9Tt4BdG-ghqm8L2sojiUFmNTnuaVjRlr2BY1GLPWHHtz8TvU-4Ax6s-I8P7oLzvYMLbD-605TZm$>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
Hey @jonhsussman, if you look in the docker file you'll see that we're importing a forked version of Flowsom. Given the complexity of the dependencies for |
I managed to solve this problem by processing this step in batches, I am linking to the solution here so others can see if they encounter the same problem: |
I wanted to get back to this to expand on it and list the direct solution here so it can be more easily referenced. It is clear that on large fovs, the existing code (Docker and conda alike) will crash at this
The block_size can be adjusted, I found that 100 works well, taking an hour or so to complete which seemed reasonable. It probably can go up to 1,000,000 and might be much faster with a very large block, but I didn't have a good reason to test that. It might need some experimentation to optimize this. Perhaps if the image is above a certain size, it could use this approach, or if it's fast enough in all cases, simply to use this in general to avoid other possible errors that may arise |
Hey @jonhsussman, thanks for following up! Can you confirm (if you haven't already) that this is only an issue, even on your end, with large FOVs? Maybe just taking a small crop (around 2k x 2k, which is what our default is) from one of your larger images and making sure it runs with our current pipeline? I just want to make sure there isn't some other error you've found with the pipeline in general, and that it really is specific to large images |
@ngreenwald That is correct, in fact this error never came up at all with one of our images that was only slightly smaller (but still very massive). And also ran just fine with the example images provided. |
Okay awesome, thanks! |
Hey @jonhsussman, just had our weekly meeting to discuss. Not sure what your bandwidth is, but if you're interested in contributing to the repository, we always welcome pull requests from collaborators! If not, no worries, we've added this to our todo list and we'll get the change merged in. |
I would be very happy to contribute to the repository! My bandwidth has
been very much reduced the last couple of weeks (which also explains my
delay here as I've had to turn away my attention), but I will get back to
this quite soon and start testing some of the solutions and contributing
pull requests.
…On Tue, Sep 27, 2022 at 5:26 PM Noah F. Greenwald ***@***.***> wrote:
Hey @jonhsussman
<https://urldefense.com/v3/__https://github.com/jonhsussman__;!!LIr3w8kk_Xxm!uJd95REklNYTbpHz2wYMggwmXxTTiI0rJc77xwqxDA4NQ5FRUVR1gDGVZ5HkJcYVjAVKqNAvtMVoesPLLvvriWuG$>,
just had our weekly meeting to discuss. Not sure what your bandwidth is,
but if you're interested in contributing to the repository, we always
welcome pull requests from collaborators! If not, no worries, we've added
this to our todo list and we'll get the change merged in.
—
Reply to this email directly, view it on GitHub
<https://urldefense.com/v3/__https://github.com/angelolab/ark-analysis/issues/688*issuecomment-1260071510__;Iw!!LIr3w8kk_Xxm!uJd95REklNYTbpHz2wYMggwmXxTTiI0rJc77xwqxDA4NQ5FRUVR1gDGVZ5HkJcYVjAVKqNAvtMVoesPLLjVqrxA-$>,
or unsubscribe
<https://urldefense.com/v3/__https://github.com/notifications/unsubscribe-auth/AKIGYMXVJS73H73S6JHQV6DWANQ7DANCNFSM6AAAAAAQE7D2VA__;!!LIr3w8kk_Xxm!uJd95REklNYTbpHz2wYMggwmXxTTiI0rJc77xwqxDA4NQ5FRUVR1gDGVZ5HkJcYVjAVKqNAvtMVoesPLLr7G8haI$>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
That's sounds great! No rush, totally understand that you have other responsibilities. We have some general guidelines here: https://ark-analysis.readthedocs.io/en/latest/_rtd/contributing.html This one is pretty straightforward, and you already have code solution which is great. The main thing is just to confirm that you get the exact same results for your new implementation compared to the base version. Generating some random vectors and passing them to both versions of the function, for example. We don't need to test the whole pixie pipeline, just that specific flowsom call. |
Thanks for the link and inclusion to the codes! Might take a week or
longer, but will keep you posted with updates herein
…On Mon, Oct 10, 2022 at 9:16 PM Noah F. Greenwald ***@***.***> wrote:
That's sounds great! No rush, totally understand that you have other
responsibilities. We have some general guidelines here:
https://ark-analysis.readthedocs.io/en/latest/_rtd/contributing.html
<https://urldefense.com/v3/__https://ark-analysis.readthedocs.io/en/latest/_rtd/contributing.html__;!!LIr3w8kk_Xxm!pbfQqs7TJ8mxvhVXzpGkjqh-mDJn487SdxBUqn8wz1Qp6rjyj2iyUraHjMWOSVn5T0Nl0FgF1lG6AEdqw-yNYs3m$>
This one is pretty straightforward, and you already have code solution
which is great. The main thing is just to confirm that you get the exact
same results for your new implementation compared to the base version.
Generating some random vectors and passing them to both versions of the
function, for example. We don't need to test the whole pixie pipeline, just
that specific flowsom call.
—
Reply to this email directly, view it on GitHub
<https://urldefense.com/v3/__https://github.com/angelolab/ark-analysis/issues/688*issuecomment-1273967244__;Iw!!LIr3w8kk_Xxm!pbfQqs7TJ8mxvhVXzpGkjqh-mDJn487SdxBUqn8wz1Qp6rjyj2iyUraHjMWOSVn5T0Nl0FgF1lG6AEdqw7nkeYtg$>,
or unsubscribe
<https://urldefense.com/v3/__https://github.com/notifications/unsubscribe-auth/AKIGYMTT5XRCJWKFKBFOQZ3WCS5WNANCNFSM6AAAAAAQE7D2VA__;!!LIr3w8kk_Xxm!pbfQqs7TJ8mxvhVXzpGkjqh-mDJn487SdxBUqn8wz1Qp6rjyj2iyUraHjMWOSVn5T0Nl0FgF1lG6AEdqw4L6LY2S$>
.
You are receiving this because you were assigned.Message ID:
***@***.***>
|
Dear ark team,
I was trying out the pixel clustering notebook out today and ran into an issue I don't know how to resolve.
After successfully training the pixel SOM (step 2.1), trying to run the next cell first gives the output "The data for FOV has been corrupted, removing" for all FOVs in the fovs list, then proceeds to throw the key error "pixel_som_cluster", as there are no FOVs left to process.
I tried reinstalling the codebase and updating docker, rerunning the notebook, unfortunately to no avail.
I also checked all involved paths and couldn't find any issues there.
Do you guys have an idea where the issue may lie?
Thanks already for your help!
Best wishes,
Sven
The text was updated successfully, but these errors were encountered: