Skip to content

andrem8/surf_dash

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Surfline Dashboard

Architecture

Surfline App Architecture

Overview

The pipeline collects data from the surfline API and exports a csv file to S3. Then the most recent file in S3 is downloaded to be ingested into the Postgres datawarehouse. A temp table is created and then the unique rows are inserted into the data tables. Airflow is used for orchestration and hosted locally with docker-compose and mysql. Postgres is also running locally in a docker container. The data dashboard is run locally with ploty.

ETL

image

Data Warehouse - Postgres

image

Data Dashboard

image

Learning Resources

Airflow Basics:

Airflow DAG: Coding your first DAG for Beginners

Running Airflow 2.0 with Docker in 5 mins

S3 Basics:

Setting Up Airflow Tasks To Connect Postgres And S3

How to Upload files to AWS S3 using Python and Boto3

Download files from S3

Docker Basics:

Docker Tutorial for Beginners

Docker and PostgreSQL

Build your first pipeline DAG | Apache airflow for beginners

Run Airflow 2.0 via Docker | Minimal Setup | Apache airflow for beginners

Docker Network Bridge

Docker Curriculum

Docker Compose - Airflow

Plotly:

Introduction to Plotly

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages