Posts

Lessons from Building Data Pipelines on DLT

Couple of months ago, I shared my initial experience with Data Load Tool (DLT). Since then, I have built a few projects with the help of their documentation and an incredibly supportive Slack community. I was immediately drawn towards DLT for its Python-first approach when it comes to building data pipelines. Over the past two months, I’ve deepened my appreciation for idiomatic Python while exploring and applying several of DLT’s features, which I am eager to leverage in real-world projects.

Evolution of Data Lakehouse Architecture

The first time I have heard about data lakehouse was 2.5 years ago at a conference. Back then, I was still at the university and most of the content of the day went over my head. Fast forward to 2024, I have since graduated and have been working as a data engineer for nearly 1.5 years now. Just over a month ago, I came across data lakehouse again as I was starting to get myself ready for some upcoming projects on it.

Exploring the Capabilties of Data Load Tool (DLT)

By and large, there is no shortage of data products coming into the market. Below is a snapshot of 2024 Machine Learning, AI and Data landscape, which has 2011 logos in total. It is incredibly difficult to pinpoint what is good amongst all of this. However, every now and then, there is a team behind a product that attempts to do something different. This article is about one such tool that I believe has a strong case to make an impact in its category.