Data AI in a Day · A closed workshop for mixed data teams

AI - Full Data Project Lifecycle in 8h

PMs, DPOs, data engineers, and analysts run the full AI-augmented data lifecycle together — on a real dataset, in your stack. No slides. Your team ships an ADO backlog, a medallion dbt model, a Streamlit data agent, and a live ticket-to-PR loop by 5pm.

  • Five modules. PMs, DPOs, data engineers, and analysts each lead the module that fits them — the rest review. Mixed-team learning by design.
  • Built on Claude Code, MCP servers, dbt, DuckDB, and a pre-loaded NordicMart dataset. All licenses provided for the day.
  • From a 3-page stakeholder transcript through medallion engineering to an agentic ticket-to-PR demo — end-to-end, hands-on.
Book a 20-minute scoping call Download the workshop brief (PDF) Capacity: we deliver 2 workshops per month. Q3 2026 has 3 slots remaining.
fasterCycle time on the work AI augments well
~€345kAnnualized capacity recovered, 8-person team*
2 sprintsTypical payback after the workshop
From €9kUp to 12 participants per cohort
The why

Two costs your business pays every week.

AI changed what a competent data team can produce in a week, and what a leadership team can ask of its data without waiting two weeks for an answer. Most organizations are still operating at pre-AI speed. The workshop closes that gap in one day, with the AI Governance scaffolding to keep it defensible as you scale.

Cost #1 — Productivity

Your team is doing in five days what AI-equipped teams do in two.

Read the spec. Find the right sources. Write the model, the tests, the docs. Open the PR. Handle review. Merge. Deploy. On a real week that takes four to five days. With AI agents wired into the workflow — conventions, tests, warehouse access — the same loop closes in a half day.

Cost #2 — Action

Time spent on dashboards is time not spent driving decisions.

Most analytics teams spend their hours maintaining dashboards and answering ad-hoc data requests. With AI agents handling the plumbing, your team has the bandwidth to engage with business owners, hear that complaints from key customers are spiking, and ship a chatbot that surfaces those complaints the same morning.

And yes — that quarterly close where three departments send three different answers? The workshop also covers AI Governance: the patterns, controls, and audit trail your team needs so AI agents stay legible to your CTO, your auditors, and your board as you roll them out across the org.
What this workshop is

One day. Five modules. Four roles. Real artefacts on a real stack.

We run Data AI in a Day inside a self-contained environment we ship as docker-compose — Claude Code, MCP servers, dbt, DuckDB, Azure DevOps, and Streamlit pre-wired and pre-loaded. Your team starts working at 09:00 after the install half-day is already done.

The day mirrors the AI-augmented data lifecycle end-to-end: requirement, engineering, analytics, agentic loop. PMs, DPOs, data engineers, and analysts each lead their module — the rest review and learn how the work hands off.

01

A purpose-built environment, ready from minute one

Modern data stack — warehouse, dbt, BI, Python orchestration, Jira-equivalent, Git — pre-configured with realistic data, agents, and integrations. NordicMart Q2 RCA dataset loaded by default.

02

An end-to-end use case, from problem to action

Your team carries one realistic case through every step: stakeholder transcript, backlog, medallion model, talking dashboard, agentic PR. Each step shows where AI agents accelerate the work and where human judgment stays load-bearing.

03

A playbook your team carries home

Markdown knowledge base, prompt library, ticket-writer skill, one-page runbook, Plan/Code/Ask reference card, and a measured before-and-after — the artefacts you need to bring the pattern back into your own stack.

The day, in five modules

Five modules. One mixed team. Eight hours from requirement to production.

Each module is one stage of the AI-augmented data lifecycle. Each module feeds the next. The full team is in the room all day; the lead chair rotates to match the work — and every module ships concrete artefacts.

1

The AI-Native Data Team

09:00 — 10:15 · All roles

You leave with

  • Claude Code repo set up — shared team zone and individual workspaces, anchored in markdown
  • Skills introduction + IDE walkthrough for the markdown knowledge layer
  • MCP servers (filesystem, DuckDB, ADO) and core skills wired up
  • Shared mental model: markdown KB vs RAG, Plan / Code / Ask modes
2

Requirements + Inception

10:15 — 11:30 · PM & DPO lead

You leave with

  • Project board populated with epics and user stories derived from the stakeholder transcript
  • Source-to-target map (raw data → KPIs)
  • “Ticket-writer” skill introduced and applied across the backlog
  • Sprint-1 cut line + stakeholder confirmation message
3

AI-Powered Engineering

11:45 — 13:45 · Data Engineer leads

You leave with

  • LLM-assisted medallion in dbt — complete bronze, silver, and gold model
  • Legacy margin script refactored into clean medallion SQL
  • Tests, docs, and lineage graph generated for every model
  • Dedup pattern with reviewed sample matches and near-misses
4

Intelligent Analytics

14:30 — 15:45 · Analyst leads

You leave with

  • Streamlit “talking dashboard” + HTML dashboards built live with the team
  • Data analytics with Claude — semantic EDA in plain English
  • Three guardrails: 30s query timeout, SELECT-only, PII auto-mask
  • “Save as finding” pattern writing to findings.md
5

The Grand Finale

15:45 — 17:00 · All roles

You leave with

  • Pull request auto-generated from an ADO ticket via MCP (ticket → plan → code → test → PR)
  • Governance + cost-discipline playbook for production agentic workflows

Coffee 11:30–11:45 · Lunch 13:45–14:30 · Hour-by-hour schedule available in the workshop brief PDF.

Our take

Markdown knowledge base. Not RAG. Here’s why.

For a project-scale knowledge base — 10–20 cross-linked markdown files capturing decisions, requirements, findings — we don’t use vector indexes. We use Claude Code reading the files it needs, when it needs them, with full fidelity. This isn’t dogma. It’s leverage.

Our approach Markdown knowledge base, read by Claude on demand

  • 10 files fit comfortably in a 200K-token context window. No chunking, no retrieval-quality problems, no embedding drift.
  • Markdown is reviewable, diff-able, and version-controlled. A markdown diff is auditable. A vector index drifts silently.
  • Works with Obsidian, your IDE, git, and any teammate who doesn’t use an LLM. One artefact, many readers.
  • When Claude misses a file, you point at it directly. With RAG, you can’t.

The default RAG approach Vector index + retrieval pipeline

  • Exists because chat models can’t fit project knowledge in their context. Claude Code can.
  • Adds chunking + embedding + retrieval as moving parts to maintain and tune.
  • Black box: when retrieval picks the wrong chunk, debugging is opaque.
  • Heavyweight for project-scale knowledge. Right tool for enterprise-scale full-corpus search — wrong tool for one team’s working memory.

RAG has its place — at company-wide corpus scale. For the workshop, and for most data team projects after, the markdown approach wins on speed, transparency, and audit trail.

The toolbox

The full AI-native data stack, ready on every participant's laptop in 60 minutes.

A modern, opinionated stack pre-loaded with realistic data. The team works with the same tools we use at our clients — no “toy” environments, no shortcuts.

Agent
Claude Code

CLI agent that operates on your filesystem, runs commands, and lives inside the repo. The driver for the day.

Integration
MCP servers

Filesystem, DuckDB, and Azure DevOps — three MCP servers wired to Claude Code, your portable interface to the stack.

Warehouse
DuckDB

Embedded analytical engine. The medallion runs locally on the participant’s laptop — no cloud bill, no security review.

Modelling
dbt

The transformation layer. Bronze, silver, gold with tests and docs generated as part of the build, not after.

Project mgmt
Azure DevOps

The backlog target. Epics, stories, and the live PR loop at the end of the day. Jira swap on request.

Orchestration
Airflow

For the production-shape DAG patterns. Reviewed in context, not built from scratch in one day.

Frontend
Streamlit

The “talking dashboard” front-end. PII guardrails, visible SQL, save-as-finding pattern baked in.

Container
Docker

The whole stack as docker-compose. One make up brings it live. Reproducible, self-contained, off the laptop in one command.

Bring your own? Default stack is the one above. Common adaptations: Jira instead of ADO, GitHub Copilot alongside Claude Code, Snowflake / BigQuery instead of DuckDB, Power BI / Looker for the final dashboard. We confirm tool swaps in the scoping call.
What you walk away with

Two lists. One for the team in the room. One for the CFO who signed off.

For your team The day in the room

  • A configured Claude Code repo with shared team zone, individual workspaces, and pre-loaded skills
  • ADO backlog, medallion dbt model, Streamlit talking dashboard, and a live PR — all built end-to-end during the day
  • Markdown knowledge base + prompt library scoped to your data domain, in a portable format
  • One-page runbook covering agent setup, MCP scopes, Plan/Code/Ask modes, and rollback patterns
  • 30-day Teams support channel with our engineers for Monday-morning questions

For the business The week after

  • A measurable cycle-time baseline (workshop pace) to benchmark against your team’s pace at home
  • A written scope for the next 90 days of AI-augmented data work
  • A defensible answer to “how are we using AI in data” for your next board meeting
  • An AI Governance starter pack: DPIA-lite template, scope boundaries, controls, audit hooks, escalation patterns
Fit check

Built for mixed data teams who want to work better together.

This is built for you if:

  • You’re a mixed data team — PM, DPO, data engineer, analyst — looking for a working pattern that survives the handoffs between you
  • You run a data or analytics team of 4 to 20 people inside a company doing €50M to €5B in revenue
  • You’re under pressure to increase team productivity and reduce cost in the data and analytics function
  • Your team ships in a modern stack — dbt with a warehouse, Power BI / Looker / Tableau dashboards, Python pipelines — on Jira or Azure DevOps and GitHub/GitLab
  • Your leadership team is asking “what is AI actually changing for us” and waiting on a straight answer
  • Your engineers and analysts are experimenting with Claude, Copilot, or ChatGPT and looking for a standard way to work as a team
  • You want a real case study to spark imagination and a working pattern by next sprint
The CFO math

The math your CFO will run.

We don’t ask anyone to take “AI makes teams faster” on faith. Run the numbers. If they don’t come in above 4× ROI on your business, we tell you not to buy the workshop.

Worked example: 8-person data team

Industry data shows roughly 40% of a data team’s time goes to model authoring, tests, and PR drafting. We see ~60% reduction on that work after the pattern is wired in. At a loaded cost of €180,000 per engineer per year, an 8-person team recovers around €345,000 in annualized capacity. The workshop pays back on the second sprint.

Team size (data engineers + analysts) 8
Loaded cost per person per year €180,000
Share of time on authoring & PR work 40%
Cycle-time reduction on that work 60%
Annualized capacity recovered ~€345k
MD
— The person in the room with your team

Michał Dębski. Astral Forest, lead.

There’s a moment every leader knows. You ask a simple question — “what are our margins this quarter?” — and two weeks later three departments send three different answers.

I’ve spent 15 years inside complex organizations across Europe and beyond, and watched the same loop repeat: data scattered, numbers inconsistent, AI pilots that never reach production. So I co-founded Astral Forest with one mission — make data work for the people running the business, not the other way around.

One client used to spend 2,000 person-days a year producing a quarterly financial report. After working with us, every employee gets the answer they need in 15 seconds. That’s what transformation looks like — and the workshop is how we get teams there in days, not quarters.
  • 15 years building production data & AI inside fintech, manufacturing, and professional services
  • 80+ shipped BI, data engineering, and AI projects across 14 years of Astral Forest delivery
  • Lecturer at Warsaw University of Technology (Politechnika Warszawska) — academic practice alongside the consulting work
  • Hands-on across GCP, Azure, dbt, Power BI, Anthropic Claude, and Vertex AI
  • Delivered in English, Polish, and French
LinkedIn profile Book a call with Michał
Pricing & how we run it

One workshop. One fixed fee. One scoping call to decide.

We don’t run open enrolments. Every workshop is private, purchased by one company, scoped to your stack and your data before any commitment.

Engagements start at
€9,000

Up to 12 participants from one team. Fixed fee, fixed scope, fixed date.

  • 1-day workshop, on-site at your office (EU and UK; US East Coast on request)
  • Pre-workshop scoping: case mirror chosen, environment tuned for your context
  • Access to the workshop environment maintained by Astral Forest
  • 30-day Microsoft Teams support window with our engineers
  • Standard MSA available; we sign yours if you have one
Capacity: we deliver 2 workshops per month. Q3 2026 has 3 slots remaining. Travel and one prep day billed at cost outside the EU and UK.

How it works

  1. Scoping call20 minutes

    We walk your stack, talk through the work your team actually does, and agree on the case mirror to load into the workshop environment.

  2. Proposal & SOWWithin 48 hours

    Fixed fee, fixed scope, fixed date. Standard MSA available; we also sign yours.

  3. Pre-workshop prep1 week before

    We tune the workshop environment to your case, calibrate the agents, and send your team a 15-minute pre-read.

  4. The day itself09:00 — 17:00 in your timezone

    Five modules. Pattern running by 5pm. Runbook in your team’s hands.

Security posture. The workshop runs inside an environment maintained by Astral Forest, with access scoped to workshop participants for the day plus a 7-day window for handover. Realistic data is synthetic or fully anonymized — your production data and source code stay outside the engagement. Foundation models are configured in enterprise mode, with inputs and outputs excluded from training. A security one-pager is available on request.
After the workshop

Three paths from here.

The workshop is a contained engagement. About 60% of teams continue with us into one of two delivery shapes; 40% run the pattern themselves. We’re explicit about all three because pretending otherwise insults your procurement team.

Path 1

Run it yourself

Your team takes the runbook and the prompt library and runs the pattern internally. We stay available over Teams for 30 days. About 40% of clients choose this.

No additional cost · 30-day Teams support included
Path 2

90-day implementation sprint

We embed with your team for 90 days, migrate your top-priority data domain onto agent-augmented workflows, and hand you a measured before-and-after.

Fixed fee · Scoped against your stack and team
Path 3

Analytics agent subscription

An ongoing capacity pack: a named Astral Forest engineer runs your agent-augmented pipeline work alongside your team. Scales up and down with your backlog.

Monthly · Month-to-month commitment

All three paths start with the same scoping call.

Honest answers

Straight answers to the questions procurement will ask.

About the workshop

Why one day and not three?

Because the behavior change happens in one day and the rest is practice. We’ve run longer formats and seen the same outcome. If your team needs more structured support afterwards, we sell that as a 90-day implementation sprint.

Can you run it for more than 12 people?

No. Above 12, each participant stops getting hands-on time with the agent on a real build, and the mechanism breaks down. If you need to train more than 12, we run the workshop twice.

Remote or on-site?

On-site only — at your office. Co-locating the leadership and data team in one room is part of how the day works. Travel and one prep day are billed at cost outside the EU and UK.

What languages do you deliver in?

English, Polish, or French. Michał delivers in all three.

Do we need to bring our own AI or dev tool licenses?

No. For the workshop day, Astral Forest provides every license you need — Claude Code, MCP servers, dbt, the DuckDB warehouse, ADO / Jira sandbox, and Streamlit. Your team focuses on the work, not on procurement. After the workshop, we hand you a sourcing one-pager for replicating the stack at home.

Can you adapt the workshop to our tools?

Yes. Default stack is Claude Code + Azure DevOps + DuckDB + dbt + Streamlit. Common adaptations: Jira instead of ADO, GitHub Copilot alongside Claude Code, Snowflake / BigQuery instead of DuckDB, Power BI / Looker for the final dashboard. We confirm tool swaps in the scoping call.

Can you use our dataset?

Yes — we encourage it. Bring an anonymized or synthetic mirror of your real data and we shape the modules around it. The transcript, the backlog, the dbt models, the talking dashboard — all built on your case. Sharper learning, faster transfer back home. We default to NordicMart Q2 RCA when you’d rather not share data; this also works.

About security & IP

Does our production code or data go anywhere?

No. The workshop runs inside an environment maintained by Astral Forest, loaded with synthetic or fully anonymized data. Your production code and data stay outside the engagement. The artifacts your team takes home — agent templates, prompts, runbook — apply back into your stack on your terms, with your security and platform teams in the loop.

Who owns the artifacts produced during the workshop?

You do. The configured agent templates, prompts, conventions documentation, and runbook are yours, transferred under the engagement.

Are inputs or outputs used to train models?

No. Foundation models from Anthropic and OpenAI both offer enterprise modes where inputs and outputs aren’t used for training. We configure agents to use those modes by default.

Do you have a security one-pager?

Yes — sent on request, signed under NDA. Covers data handling, access model, agent scoping, and our subprocessor list.

About what happens after

What if our team forgets everything in two weeks?

That’s what the runbook and the 30-day Teams channel are for. If the pattern hasn’t stuck after 30 days, we’ll say so — and we’ll usually recommend the 90-day implementation sprint rather than another workshop.

Is this just a lead magnet for a bigger project?

About 60% of clients continue into a 90-day sprint or an analytics agent subscription. The other 40% run the pattern themselves. We’re upfront about both, and we walk away clean if neither is the right fit.

Do you offer a guarantee?

We write the success criteria into the SOW: by end of day, the team will have shipped at least two agent-drafted artifacts to production. If we don’t hit that, we deliver a second session at no additional cost.

Can we get a reference call?

Yes. We set one up after the scoping call, matched to your industry and stack.

Next step

Your team can ship the first AI-augmented PR on Monday.

20-minute call. We walk your stack, agree on the case mirror to load into the workshop environment, and send a fixed-fee proposal within 48 hours.