# Building a Streaming Session Analytics Pipeline with Kafka, Postgres, and dbt

ayoabass777
How I built an end-to-end clickstream pipeline with exactly-once delivery guarantees When I set out to build Pulse, I had a specific goal: demonstrate that I could work with streaming data, not just batch. My first portfolio project (Ballistics) was a batch pipeline — API calls on a schedule, Airflow orchestration, daily refreshes. That's the bread and butter of most data engineering work, but it's only half the picture. Pulse is the other half. Real-time events flowing through Kafka, landing in