-
Notifications
You must be signed in to change notification settings - Fork 31
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Ingest NHD High Resolution Streams, Add Tile Layer #3417
Conversation
The previous version was failing on my machine.
@@ -92,6 +92,90 @@ | |||
} | |||
} | |||
|
|||
#nhdflowlinehr { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
These styles are identical to those for #nhdflowline
An update. I set up a new set of vagrant VMs on an external drive and increased the memory and CPU as suggested. Loading the high res data has been chugging on ALTER TABLE for a while, a lot more than the suggested "30-60 minutes"
The disk on my services VM is up to about 33GB. I will let it go overnight. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I was unable to see the data load complete in a reasonable amount of time when attempting to run the VM from my external spinning hard disk. The process did appear to be working, and the changes in the PR are mostly a reuse of existing functionality. 👍
Thanks for taking a look. I'll merge this and we can get this on staging and evaluate there. (Although we may have to wait for #3416 before staging deployments work again.) |
Overview
The full ingest process is documented in #3415 (comment). This process took several days, not just to figure out the data shape, but to process it, because of the volume. This is by far the largest volume of data we've added to MMW, and may have non-trivial consequences for performance and hosting costs that will reveal themselves as this moves to staging and production.
The work here adds the NHD High Resolution stream data to the MMW database, and wires up the tiler to render it on the map. It does not switch our analyses and models to use the new High Resolution data, which still use Medium Resolution. That will be done in future cards.
I was unable to provision the tiler VM on my local, which could only be resolved by upgrading the NPM version on that VM. It seemingly works well enough, and is isolated from the app VM, so should be fine.
Connects #3415
Demo
Notes
The compressed data is ~17GB, and in the database it is ~26GB:
There's a total of ~24.5M rows:
I attempted to drop the extraneous columns in the table to see if that would reduce the size much, but it didn't. I imagine that most of the data is the geometry of the streams, which we cannot elide.
About 2% of the streams do not have a value for
stream_order
orslope
, which were obtained by joining with the Value Added Attributes table:Testing Instructions
Before you begin, ensure you have ~80+ GB of free space on your host computer.
$ vagrant ssh app -c 'cd /vagrant && ./scripts/aws/setupdb.sh -S'
$ vagrant reload --provision tiler