A small demo using NYC taxi data. Experimental, the GE core team uses this to test out things. No guarantee that this actually works consistently. The full demo-able and maintained webinar material is in the superconductive/ge_tutorials repo.
The postgres database actually lives in the superconductive/ge_tutorials repo in the getting started tutorial. You can run the docker container with that PG database and add the creds to run the expectations and checkpoints configured here. Or just use the getting_started tutorial repo content itself.
- Get credentials for the RDS instance.
- Add these to your
great_expectations
configuration. - To run the pipeline, also set these two environment variables:
PGPASSWORD=<SECRETS!> DEMO_USER=<SECRETS!> DEMO_HOST=<SECRETS!>
- Verify your Great Expectations install with
great_expectations suite list
psql -h $DEMO_HOST -d demo -U $DEMO_USER -f location_frequency.sql
- You should see output like
DROP TABLE SELECT 10000 DROP TABLE SELECT 28