-
Notifications
You must be signed in to change notification settings - Fork 7
Parsing metadata from Yokogawa experiments #25
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Thanks for the clear description @gusqgm! Otherwise, let's make a test dataset and mention it here, so we know what to test this parsing on. |
Hey @jluethi , all of the data from FMI currently contains metadata associated with it within the folders. |
I added a new task metadata_parsing: b98278648227709e1bdda74c8e17392bf28c79fc & 04266345e1dff319901819013e6a240a727e5ec2 The
The table is double-indexed, by well & field_id (= field of view, e.g. field_id 1 is the file that contains F001 in its filename). It contains 3 types of information:
b) The scale information about pixel sizes in x, y & z in micrometers. In the case of the table above, the value for each site would be
The parsing performs quite a few checks for consistencies, e.g. bit_depth could be 8 or 16, but needs to be the same within all images of a well (likely always will be 16 bit for all the yokogawa images). x & y positions need to be the same for all images of a given site (otherwise something went seriously wrong), pixel sizes need to be consistent etc. Even with all this, the parsing still runs in under a minute to process the metadata for my 23 well example that consists of 165792 image records (slightly fewer images in the test case, long story). So looks to me like it should scale pretty well for our use-cases :) |
@tcompa Let's talk about how we best integrate the parsing functionality into the OME-Zarr creation workflow. Probably easiest to have a quick Zoom call on this, e.g. on Monday or as part of the Fractal call on Tuesday morning. A few notes here from my side:
Thus, if we integrate the aggregation pandas table with the site metadata into the workflow and write a little wrapper for the simple case, we should support all of the core use cases of Yokogawa data acquisition we're currently targeting for parsing into OME-Zarr: Simple, full wells, sparsely recorded fields on a grid as well as "search-first, gridless" images arranged at random positions in a well (see here: https://github.com/fractal-analytics-platform/mwe_fractal/issues/23) Also, this kind of intermediate pandas dataframe may make it easier to integrate data from other microscopes (file names also change, but we have a target of how we need to prepare the metadata for it to be read in) |
I also added |
Note: this issue is currently on hold, waiting for the single-FOV vs multi-FOV decision (https://github.com/fractal-analytics-platform/mwe_fractal/issues/74). |
Any experiments coming from the Yokogawa microscope generate a couple of metadata files (with
.mrf
and.mlf
extensions).You can find information on how this is being parsed by Drogon here. The corresponding empty dataframes for each meta file are created just before this line.
Proper parsing of metadata is important not only for being able to potentially save it (within?/alongisde?) .zarr output for further reference but also for processing steps such as #8 where we need the positional information of each recorded field.
The text was updated successfully, but these errors were encountered: