Ready to dive into the lake?

lakeFS is currently only
available on desktop.

For an optimal experience, provide your email below and one of our lifeguards will send you a link to start swimming in the lake!

lakeFS Community

Tutorials

Tutorials

Authorization (RBAC) in lakeFS: Step-by-Step Configuration Tutorial

Amit Kesarwani
March 14, 2023

Introduction Last month, the lakeFS team decided to move from the decoupled security authentication and access control features to enable you to plug your own authentication and security mechanism. Consequently, the team decided to change the architecture to a pluggable one which enables you to choose your preference without being dependent on the lakeFS solution. …

Authorization (RBAC) in lakeFS: Step-by-Step Configuration Tutorial Read More »

Tutorials

The Airflow and lakeFS Integration: Step-by-Step Configuration Tutorial

Amit Kesarwani
March 14, 2023

Introduction lakeFS makes creating isolated environments for data ingestion instantaneous so you can run data ingestion jobs without impacting your production data and merge ingested data atomically to your production data instantaneously. This frees you from spending time on environment maintenance and makes it possible to create as many environments as needed. If ingested data …

The Airflow and lakeFS Integration: Step-by-Step Configuration Tutorial Read More »

Integrations Tutorials

Databricks and lakeFS Integration: Step-by-Step Configuration Tutorial

Iddo Avneri
March 14, 2023

Introduction This tutorial will review all steps needed to configure lakeFS on Databricks.  This tutorial assumes that lakeFS is already set up and running against your storage (in this example AWS s3), and is focused on setting up the Databricks and lakeFS integration. Prerequisites Step 1 – Acquire lakeFS Key and Secret In this step, …

Databricks and lakeFS Integration: Step-by-Step Configuration Tutorial Read More »

Tutorials Use Cases

How to Build an Isolated Testing Environment for Data with lakeFS

Barak Amar
March 14, 2023

Overview Our routine work with data includes developing code, choosing and upgrading compute infrastructure, and testing new and changed data pipelines. Usually, this requires running our tested pipelines in parallel to production, in order to test the changes we wish to apply. Every data engineer knows that this convoluted process requires copying data, manually updating …

How to Build an Isolated Testing Environment for Data with lakeFS Read More »

Git for Data – lakeFS

  • Get Started
    Get Started