Lakevision is a tool that provides insights into your Data Lakehouse based on the Apache Iceberg table format.
It lists every namespace and table in your Lakehouse—along with each table’s schema, properties, snapshots, partitions, sort-orders, references/tags, and sample data—and supports nested namespaces. This helps you quickly understand data layout, file locations, and change history.
Lakevision is built with pyiceberg, a FastAPI backend, and a SvelteKit frontend, keeping other dependencies to a minimum.
https://github.com/user-attachments/assets/b6b2eef5-9f27-40ca-a80d-27b88d4a8cfd
Before running Lakevision, you'll need to create and configure your local .env file:
cp my.env .env
Then edit .env to provide values for:
Your Iceberg catalog configuration (URI, warehouse path, etc.)
🧪 Don’t have a catalog yet? You can start with a sample one. See make make sample-catalog in the Makefile section.
Authentication details (e.g., token or credentials)
Optional cloud settings (S3, GCP, etc.)
This avoids modifying my.env, which is version-controlled and serves as a template.
The easiest way to run Lakevision is with Docker.
Clone the repository and cd into the project root.
Build the image
docker build -t lakevision:1.0 .
Run the container
Make sure you’ve completed the Environment Setup step first.
docker run --env-file .env -p 8081:8081 lakevision:1.0 /app/start.sh
Run the health worker container
If the health functionality is enabled you need to start the container for the health worker.
docker run --env-file .env lakevision:1.0 /app/worker.sh
Once started, the backend listens on port 8000 and Nginx runs on port 8081. Visit http://localhost:8081 to explore the UI.
✅ Tested on Linux and macOS with the Iceberg REST catalog. Other PyIceberg-compatible catalogs should work too.
To build the image with the sample in-memory Iceberg catalog included:
docker build --build-arg ENABLE_SAMPLE_CATALOG=true -t lakevision:1.0 .
.env, comment out the default catalog settings and uncomment the sample catalog lines.Make sure you’ve completed the Environment Setup step first.
You can use the Makefile to automate common setup steps:
make init-be # Set up Python backend
make sample-catalog # Populate a local Iceberg catalog with sample data
make init-fe # Install frontend dependencies
make run-be # Start backend (FastAPI)
make run-fe # Start frontend (SvelteKit)
make help # List all Makefile commands
Once running, visit http://localhost:8081 to use the app.
Make sure you’ve completed the Environment Setup step first.
💡 Frontend note: All environment variables that begin with
PUBLIC_must be available in a separate.envfile inside the/fefolder. You can do this manually, or by running:
make prepare-fe-env
This ensures the frontend build system (Vite) can access the variables during development.
cd be
python -m venv .venv
source .venv/bin/activate
pip install -r requirements.txt
set -a; source ../.env; set +a
PYTHONPATH=app uvicorn app.api:app --reload --port 8000
cd ../fe
npm install
npm run dev -- --port 8081
Implement your custom implementation module in the backend, must follow app/be/authz.py
Configure the following properties in your environment.
and run the be. E.g. make run-be
In case you need to run the frontend with https you can follow this simple steps:
Install a compatible plugin-basic-ssl to the vite version in the fe.
Add "@vitejs/plugin-basic-ssl": "^1.2.0" under devDependency in the package.json and install dependencies. Refers to: Running Locally section.
Update the vite config (vite.config.js):
...
import basicSsl from '@vitejs/plugin-basic-ssl';
export default defineConfig({
plugins: [
sveltekit(),
// Optimize CSS from `carbon-components-svelte` when building for production.
optimizeCss(),
basicSsl()
],
...
This auto-generates a self-signed cert for dev. You’ll get a warning page you can bypass.
Run the frontend. E.g.: make run-fe
Want to deploy Lakevision on Kubernetes or OpenShift?
Sample manifests are provided in k8s/, including example Deployment, Service, ConfigMap, and Secret YAMLs for running the unified (backend + frontend) container.
k8s/README.md for quickstart instructions and customization notes.The Lakehouse Health feature provides a system for running, scheduling, and monitoring data quality and health checks across your lakehouse.
When enabled, it adds two main UI components:
This feature is disabled by default and operates as a small, services-oriented system. It relies on a central database and two independent background processes to function.
api.py): This is the main web server. It serves the frontend UI, handles user-triggered actions (e.g., "Run Health Check Now"), and reads from the database to display results.scheduler.py): This is a lightweight, separate background process. Its only job is to run periodically (e.g., every 10 minutes), check for any scheduled jobs that are due, and enqueue them as tasks in the database.worker.py): This is the heavy-lifting background process. It constantly polls the database task queue. When it finds a new task, it executes the actual health check against the Iceberg table, generates the results, and writes them back to the database. This would ideally run in another container, so that you can scale and have multiple workers active.This separation ensures that a long-running health check (e.g., on a huge table) does not block or slow down the main API server.
To enable this feature, you must set two environment variables.
PUBLIC_LAKEVISION_HEALTH_ENABLED
true: Enables the feature in both the frontend and backend.false (or not set): Disables the feature entirely./api/insights, /api/jobs) will not be loaded.scheduler.py and worker.py scripts will exit immediately if you try to run them.LAKEVISION_DATABASE_URL
PUBLIC_LAKEVISION_HEALTH_ENABLED is true.When the health feature is enabled, you must run three separate processes for it to function correctly. Beside the main backend, you need to run 2 additional processes:
# 1. The scheduler process
scheduler: python -m app.scheduler
# 2. The worker process
worker: python -m app.worker
## 🧭 Roadmap
* Chat with Lakehouse capability using an LLM
* Table-level reports (most snapshots, partitions, columns, size, etc.)
* Optimization recommendations
* Limited SQL capabilities ✅
* Partition details (name, file count, records, size) ✅
* Sample data by partition ✅
* Table-level insights
* Time-travel queries
## 🤝 Contributing
Contributions are welcome!
1. **Fork** the repository and clone it locally.
2. **Create** a branch for your change, referencing an issue if one exists.
3. **Add tests** for new functionality where appropriate.
4. **Open a pull request** with a clear description of the changes.