Astro SDK Python allows rapid and clean development of {Extract, Load, Transform} workflows using Python. It helps DAG authors to achieve more with less code. It is powered by Apache Airflow and maintained by Astronomer.
⚠️ Disclaimer This project is in a preview release state. In other words, it is not production-ready yet. The interfaces may change. We welcome users to try out the interfaces and provide us with feedback.
Astro SDK Python is available at PyPI. Use the standard Python installation tools.
To install a cloud-agnostic version of Astro SDK Python, run:
pip install astro-sdk-python
If using cloud providers, install using the optional dependencies of interest:
pip install astro-sdk-python[amazon,google,snowflake,postgres]
After installing Astro, copy the following example dag calculate_popular_movies.py
to a local directory named dags
:
from datetime import datetime
from airflow import DAG
from astro import sql as aql
from astro.files import File
from astro.sql.table import Table
@aql.transform()
def top_five_animations(input_table: Table):
return """
SELECT Title, Rating
FROM {{input_table}}
WHERE Genre1=='Animation'
ORDER BY Rating desc
LIMIT 5;
"""
with DAG(
"calculate_popular_movies",
schedule_interval=None,
start_date=datetime(2000, 1, 1),
catchup=False,
) as dag:
imdb_movies = aql.load_file(
File("https://raw.githubusercontent.com/astronomer/astro-sdk/main/tests/data/imdb.csv"),
output_table=Table(
name="imdb_movies", conn_id="sqlite_default"
),
)
top_five_animations(
input_table=imdb_movies,
output_table=Table(
name="top_animation"
),
)
Set up a local instance of Airflow by running:
export AIRFLOW_HOME=`pwd`
export AIRFLOW__CORE__ENABLE_XCOM_PICKLING=True
airflow db init
Create an SQLite database for the example to run with and run the DAG:
# The sqlite_default connection has different host for MAC vs. Linux
export SQL_TABLE_NAME=`airflow connections get sqlite_default -o yaml | grep host | awk '{print $2}'`
sqlite3 "$SQL_TABLE_NAME" "VACUUM;"
airflow dags test calculate_popular_movies `date -Iseconds`
Check the top five animations calculated by your first Astro DAG by running:
sqlite3 "$SQL_TABLE_NAME" "select * from top_animation;" ".exit"
You should see the following output:
$ sqlite3 "$SQL_TABLE_NAME" "select * from top_animation;" ".exit"
Toy Story 3 (2010)|8.3
Inside Out (2015)|8.2
How to Train Your Dragon (2010)|8.1
Zootopia (2016)|8.1
How to Train Your Dragon 2 (2014)|7.9
Astro SDK Python depends on Apache Airflow >= 2.1.0.
Databases |
---|
Google BigQuery |
Postgres |
Snowflake |
SQLite |
File types |
---|
CSV |
JSON |
NDJSON |
Parquet |
File stores |
---|
Amazon S3 |
Filesystem |
Google GCS |
A summary of the currently available operations in Astro SDK Python.
load_file
: load a given file into a SQL tabletransform
: applies a SQL select statement to a source table and saves the result to a destination tabletruncate
: remove all records from a SQL tablerun_raw_sql
: run any SQL statement without handling its outputappend
: insert rows from the source SQL table into the destination SQL table, if there are no conflictsmerge
: insert rows from the source SQL table into the destination SQL table, depending on conflicts:- ignore: do not add rows that already exist
- update: replace existing rows with new ones
export_file
: export SQL table rows into a destination filedataframe
: export given SQL table into in-memory Pandas data-frame
The documentation is a work in progress--we aim to follow the Diátaxis system:
- Getting Started: a hands-on introduction to Astro SDK Python
- How-to guides: simple step-by-step user guides to accomplish specific tasks
- Reference guide: commands, modules, classes and methods
- Explanation: Clarification and discussion of key decisions when designing the project.
We follow Semantic Versioning for releases. Check the changelog for the latest changes.
To learn more about our release philosophy and steps, check here
All contributions, bug reports, bug fixes, documentation improvements, enhancements, and ideas are welcome.
Read the Contribution Guideline for a detailed overview on how to contribute.
As contributors and maintainers to this project, you should abide by the Contributor Code of Conduct.