Mage AI

๐Ÿง™ A modern replacement for Airflow.

Deploy Mage AI

Postgres

railwayapp-templates/postgres-ssl:16

Just deployed

/var/lib/postgresql/data

mage-ai

botwayorg/mage-ai-docker

Just deployed

/home/src

Mage

๐Ÿง™ A modern replacement for Airflow.

Give your data team magical powers

Integrate and synchronize data from 3rd party sources

Build real-time and batch pipelines to transform data using Python, SQL, and R

Run, monitor, and orchestrate thousands of pipelines without losing sleep


1๏ธโƒฃ ๐Ÿ—๏ธ

Build

Have you met anyone who said they loved developing in Airflow?
Thatโ€™s why we designed an easy developer experience that youโ€™ll enjoy.

Easy developer experience
Start developing locally with a single command or launch a dev environment in your cloud using Terraform.

Language of choice
Write code in Python, SQL, or R in the same data pipeline for ultimate flexibility.

Engineering best practices built-in
Each step in your pipeline is a standalone file containing modular code thatโ€™s reusable and testable with data validations. No more DAGs with spaghetti code.

โ†“

2๏ธโƒฃ ๐Ÿ”ฎ

Preview

Stop wasting time waiting around for your DAGs to finish testing.
Get instant feedback from your code each time you run it.

Interactive code
Immediately see results from your codeโ€™s output with an interactive notebook UI.

Data is a first-class citizen
Each block of code in your pipeline produces data that can be versioned, partitioned, and cataloged for future use.

Collaborate on cloud
Develop collaboratively on cloud resources, version control with Git, and test pipelines without waiting for an available shared staging environment.

โ†“

3๏ธโƒฃ ๐Ÿš€

Launch

Donโ€™t have a large team dedicated to Airflow?
Mage makes it easy for a single developer or small team to scale up and manage thousands of pipelines.

Fast deploy
Deploy Mage to AWS, GCP, or Azure with only 2 commands using maintained Terraform templates.

Scaling made simple
Transform very large datasets directly in your data warehouse or through a native integration with Spark.

Observability
Operationalize your pipelines with built-in monitoring, alerting, and observability through an intuitive UI.

๐Ÿง™ Intro

Mage is an open-source data pipeline tool for transforming and integrating data.

  1. Install
  2. Demo
  3. Tutorials
  4. Documentation
  5. Features
  6. Core design principles
  7. Core abstractions
  8. Contributing

Looking for help? The fastest way to get started is by checking out our documentation here.

Looking for quick examples? Open a demo project right in your browser or check out our guides.

๐ŸŽฎ Demo

Live demo

Build and run a data pipeline with our demo app.

> WARNING > > The live demo is public to everyone, please donโ€™t save anything sensitive (e.g. passwords, secrets, etc).

Demo video (5 min)

Mage quick start demo

Click the image to play video


๐Ÿ‘ฉโ€๐Ÿซ Tutorials

Fire mage

๐Ÿ”ฎ Features

๐ŸŽถOrchestrationSchedule and manage data pipelines with observability.
๐Ÿ““NotebookInteractive Python, SQL, & R editor for coding data pipelines.
๐Ÿ—๏ธData integrationsSynchronize data from 3rd party sources to your internal destinations.
๐ŸšฐStreaming pipelinesIngest and transform real-time data.
โŽdbtBuild, run, and manage your dbt models with Mage.

A sample data pipeline defined across 3 files โž

  1. Load data โž
    @data_loader
    def load_csv_from_file():
        return pd.read_csv('default_repo/titanic.csv')
    
  2. Transform data โž
    @transformer
    def select_columns_from_df(df, *args):
        return df[['Age', 'Fare', 'Survived']]
    
  3. Export data โž
    @data_exporter
    def export_titanic_data_to_disk(df) -> None:
        df.to_csv('default_repo/titanic_transformed.csv')
    

What the data pipeline looks like in the UI โž

data pipeline overview

New? We recommend reading about blocks and learning from a hands-on tutorial.

Ask us questions on Slack


๐Ÿ”๏ธ Core design principles

Every user experience and technical design decision adheres to these principles.

๐Ÿ’ปEasy developer experienceOpen-source engine that comes with a custom notebook UI for building data pipelines.
๐ŸšขEngineering best practices built-inBuild and deploy data pipelines using modular code. No more writing throwaway code or trying to turn notebooks into scripts.
๐Ÿ’ณData is a first-class citizenDesigned from the ground up specifically for running data-intensive workflows.
๐ŸชScaling is made simpleAnalyze and process large data quickly for rapid iteration.

๐Ÿ›ธ Core abstractions

These are the fundamental concepts that Mage uses to operate.

ProjectLike a repository on GitHub; this is where you write all your code.
PipelineContains references to all the blocks of code you want to run, charts for visualizing data, and organizes the dependency between each block of code.
BlockA file with code that can be executed independently or within a pipeline.
Data productEvery block produces data after it's been executed. These are called data products in Mage.
TriggerA set of instructions that determine when or how a pipeline should run.
RunStores information about when it was started, its status, when it was completed, any runtime variables used in the execution of the pipeline or block, etc.

๐Ÿค” Frequently Asked Questions (FAQs)

Check out our FAQ page to find answers to some of our most asked questions.


๐Ÿชช License

See the LICENSE file for licensing information.

Water mage casting spell



Template Content

More templates in this category

View Template

Chat Chat

Chat Chat, your own unified chat and search to AI platform.


View Template

openui

Deploy OpenUI: AI-powered UI generation with GitHub OAuth and OpenAI API.


View Template

firecrawl

firecrawl api server + worker without auth, works with dify