-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
normalize-data add_alpha step is failing #802
Comments
re-running |
before copying that input file from the failed will do some testing with the successful |
In doing some manual testing yesterday on the copy I made of the file that ran into this issue, I saw the following output from
The
A random stack overflow post suggests that A seemingly relevant doc suggests using
Another possibly useful few links: Also worth noting: I monitored system resource usage using |
This all sounds good -- thanks for the testing! I wonder if it will work with the additional option? I guess it'll take even longer lol :-) |
Using the change in 17b104e, I was able to retry the gory detail about what motivated this new approach... I did more manual testing with First, I tried: Unfortunately I wasn't tracking timing closely, but at what felt like about an hour into that operation, the output file with the alpha channel was 47 GB, the progress meter indicated that it wasn't yet 50% done, and was clearly unlikely to fit in the available space (62 GB!). So I cancelled execution and deleted the output. Then I retried the original (successful on kg552qb0295) command just for good measure ( Given the above, I thought I'd see if anyone else had run into the surprising behavior of much much larger files being generated when the compress option is given to I settled on piping the output from one command to the other using GDAL's virtual file system, e.g. (when manually testing yesterday) That new TIFF is 3.1 GB. The |
one other observation about memory usage: both given that there are a max of 3 robot threads likely to be running at any given time, and given that these normalization operations seem to consume about 500 MB of memory each over the 1 GB baseline usage, i'd guess that 5.8 GB total should be fine for now? i.e., no need to up the RAM on the gis-robot-suite VMs just yet, i don't think. |
Unfortunately the two step approach outlined above During a huddle today (@jmartin-sul, @kimdurante, @edsu and @aaron-collier) we decided to back out the addition of the alpha channel since it is not having the desired effect. This effectively closes this ticket, at least for now. |
As you can see in druid:kg552qb0295 that the
normalize_data
step is throwing an error (see the related honeybadger alert):It appears that the input file being given to
gdalwarp
doesn't exist, but perhaps the temporary directory has already been cleaned up?To replicate the problem:
add Workflow
gisAssessionWF
Note: this may be related to #658 since the item includes a COG GeoTIFF?
The text was updated successfully, but these errors were encountered: