Skip to content

A collection of tools for extracting FHIR resources and analytics services on top of that data.

License

Apache-2.0, Unknown licenses found

Licenses found

Apache-2.0
LICENSE
Unknown
license-header.txt
Notifications You must be signed in to change notification settings

google/fhir-data-pipes

Folders and files

NameName
Last commit message
Last commit date
Mar 5, 2024
May 21, 2024
Mar 12, 2024
Jan 26, 2023
May 30, 2024
Oct 6, 2023
May 29, 2024
May 30, 2024
Jan 16, 2024
May 29, 2024
Oct 26, 2023
Jan 31, 2023
Apr 25, 2023
Jul 13, 2020
Mar 5, 2024
Jun 30, 2020
Mar 8, 2024
May 30, 2024
Feb 13, 2024
Sep 19, 2022
Apr 16, 2024

Build Status codecov

What is this?

This repository includes pipelines to transform data from a FHIR server (like HAPI, GCP FHIR store, or even OpenMRS) using the FHIR format into a data warehouse based on Apache Parquet files, or another FHIR server. There is also a query library in Python to make working with FHIR-based data warehouses simpler.

These tools are intended to be generic and eventually work with any FHIR-based data source and data warehouse. Here is the list of main directories with a brief description of their content:

  • pipelines/ *START HERE*: Batch and streaming pipelines to transform data from a FHIR-based source to an analytics-friendly data warehouse or another FHIR store.

  • docker/: Docker configurations for various servers/pipelines.

  • doc/: Documentation for project contributors. See the pipelines README and wiki for usage documentation.

  • utils/: Various artifacts for setting up an initial database, running pipelines, etc.

  • dwh/: Query library for working with distributed FHIR-based data warehouses.

  • bunsen/: A fork of a subset of the Bunsen project which is used to transform FHIR JSON resources to Avro records with SQL-on-FHIR schema.

  • e2e-tests/: Scripts for testing pipelines end-to-end.

NOTE: This was originally started as a collaboration between Google and the OpenMRS community.