News

Marco Bonzanini discusses the process of building data pipelines, e.g. extraction, cleaning, integration, pre-processing of data; in general, all the steps necessary to prepare data for a data ...
Databricks today announced the general availability (GA) of Delta Live Tables (DLT), a new offering designed to simplify the building and maintenance of data pipelines for extract, transform, and load ...
The platform is based on Chronon, an open-source data management engine developed by Zipline AI co-founders Varant Zanoyan ...
Notably, the new Lakeflow Declarative Pipelines capabilities allow data engineers to build end-to-end production pipelines in SQL or Python without having to manage infrastructure.
Struggling to integrate your Python enrichment services effectively into Scala data processing pipelines? Roi Yarden, Senior Software Engineer at ZipRecruiter, shares how we sewed it all together ...
MemSQL, provider of real-time databases for transactions and analytics, today announced the latest version of MemSQL Ops which accelerates the use of Spark with Spark SQL pushdowns, allows for ...
In recent years, the shortage of data engineers has at times exceeded the shortage of data scientists. To help close the gap, a Silicon Valley startup called Prophecy today unveiled a low-code data ...