The Tuva Claims Demo provides a quick and easy way to run the Tuva Project with synthetic demo data. By default, the project will run with built-in sample data of 100 patients loaded as dbt seeds. If you want to run the demo project with the full demo data from a data share (available for free), please follow the instructions below under "Running the Project with Full Demo Data".
To set up the Tuva Project with your own claims data or to better understand what the Tuva Project does, please review the ReadMe in The Tuva Project package for a detailed walkthrough and setup.
For information on data models and to view the entire DAG check out our dbt Docs.
- BigQuery
- Redshift
- Snowflake
- You have dbt installed and configured (i.e. connected to your data warehouse). If you have not installed dbt, here are instructions for doing so.
- You have created a database for the output of this project to be written in your data warehouse.
Complete the following steps to configure the project to run in your environment.
- Clone this repo to your local machine or environment.
- Update the
dbt_project.yml
file:- Add the dbt profile connected to your data warehouse.
- Update the variable
tuva_database
to use the new database you created for this project.
- Run
dbt deps
to install the package. - Run
dbt build
to run the entire project with the built-in sample data.
Complete the following additional steps to configure the project to run with the full demo data from one of the supported data shares.
The full demo data is available as a Redshift data product in the AWS Data Exchange. You can learn more about how to access the Data Exchange here.
- Make sure you are in the US East (N. Virginia) region.
- Go to the Tuva Project Claims Demo product on the AWS Data Exchange.
- Select Continue to Subscribe. Once you subscribe, it will take a few minutes for AWS to process the request.
- Go to the Redshift Console and select your cluster.
- Go to the Datashare tab and select any database to view your subscriptions.
- Under Subscriptions to AWS Data Exchange datashares, select
tuva_project_claims_demo
, and choose Create database from datashare. - Specify a database name for the data in the listing.
- Update the
dbt_project.yml
file:- Update the variable
tuva_database
to the database specified in step 7. - Set the variable
full_data_override
to 'true'.
- Update the variable
- Run
dbt deps
to install the Tuva package (you only need to do this once per local environment). - Run
dbt build
to run the entire project.
The full demo data is available as a public dataset in BigQuery and can be queried directly.
- Update the
dbt_project.yml
file:- Update the variable
tuva_database
to your Google Cloud project id. - Set the variable
full_data_override
to 'true'. (Note: This project includes logic to use the public dataset automatically when this variable is set to true.)
- Update the variable
- Run
dbt deps
to install the Tuva package (you only need to do this once per local environment). - Run
dbt build
to run the entire project.
The full demo data is available in the Snowflake Marketplace. You can learn more about how to access Snowflake Marketplace listings here.
- Go to the Tuva Project Claims Demo listing on the Snowflake Marketplace.
- Select Get.
- Specify a database name for the data in the listing.
- (Optional) Add roles to grant access to the database created from the listing.
- Update the
dbt_project.yml
file:- Update the variable
tuva_database
to the database specified in step 3. - Set the variable
full_data_override
to 'true'.
- Update the variable
- Run
dbt deps
to install the Tuva package (you only need to do this once per local environment). - Run
dbt build
to run the entire project.
The Tuva Project team maintaining this project only maintains the latest version of the project. We highly recommend you stay consistent with the latest version.
Have an opinion on the mappings? Notice any bugs when installing and running the project? If so, we highly encourage and welcome feedback! While we work on a formal process in Github, we can be easily reached on our Slack community.
Join our growing community of healthcare data practitioners on Slack!