<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Lakehouse &#8211; Prologika</title>
	<atom:link href="https://prologika.com/tag/lakehouse/feed/" rel="self" type="application/rss+xml" />
	<link>https://prologika.com</link>
	<description>Business Intelligence Consulting and Training in Atlanta</description>
	<lastBuildDate>Sun, 08 Mar 2026 21:31:52 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.3</generator>
	<item>
		<title>Prologika Newsletter Spring 2026</title>
		<link>https://prologika.com/prologika-newsletter-spring-2026/</link>
					<comments>https://prologika.com/prologika-newsletter-spring-2026/#respond</comments>
		
		<dc:creator><![CDATA[Prologika - Teo Lachev]]></dc:creator>
		<pubDate>Sun, 08 Mar 2026 21:31:52 +0000</pubDate>
				<category><![CDATA[Newsletter]]></category>
		<category><![CDATA[Data Virtualization]]></category>
		<category><![CDATA[Fabric]]></category>
		<category><![CDATA[Lakehouse]]></category>
		<category><![CDATA[Power BI]]></category>
		<guid isPermaLink="false">https://prologika.com/?p=9575</guid>

					<description><![CDATA[At Ignite in November, 2025, Microsoft introduced Fabric IQ. I noted to go beyond the marketing hype and check if Fabric IQ makes any sense. The next thing I know, [&#8230;]]]></description>
										<content:encoded><![CDATA[<p><img decoding="async" class="wp-image-9535" style="padding: 0px 10px;" src="https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1.png" width="120" height="120" align="left" srcset="https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1.png 1024w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-300x300.png 300w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-80x80.png 80w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-768x768.png 768w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-36x36.png 36w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-180x180.png 180w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-705x705.png 705w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-120x120.png 120w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-450x450.png 450w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-50x50.png 50w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-100x100.png 100w" sizes="(max-width: 120px) 100vw, 120px" /></p>
<p>At Ignite in November, 2025, Microsoft <a href="https://youtu.be/RjU0slwcZGs?si=QTRsZMg-jFQ5wsB0">introduced</a> <a href="https://youtu.be/RjU0slwcZGs?si=QTRsZMg-jFQ5wsB0">Fabric IQ</a>. I noted to go beyond the marketing hype and check if Fabric IQ makes any sense. The next thing I know, around the holidays I’m talking to an enterprise strategy manager from an airline company and McKinsey consultant about ontologies. In this newsletter I share my thoughts of FabricIQ based on my initial evaluation. Let&#8217;s start with what ontology is and how Fabric IQ uses it to integrate your enterprise data.</p>
<p><em>Ontology – A branch of philosophy, ontology is the study of being that investigates the nature of existence, the features all entities have in common, and how they are divided into basic categories of being. In computer science and AI, ontology refers to a set of concepts and categories in a subject area or domain that shows their properties and the relations between them.</em></p>
<h2>What is Fabric IQ?</h2>
<p>According to Microsoft, Fabric IQ is “a unified intelligence platform developed by Microsoft that enhances data management and decision-making through semantic understanding and AI capabilities.” Clear enough? If not, if you view Fabric as Microsoft’s answer to Palantir’s Foundry, then Fabric IQ is the Microsoft equivalent of Palantir’s Foundry Ontology, whose success apparently inspired Microsoft.</p>
<blockquote><p>Therefore, my unassuming layman definition of Fabric IQ is a metadata layer on top of data in Fabric that defines entities and their relationships so that AI can make sense of and relate the underlying data.</p></blockquote>
<p>For example, you may have an organizational semantic model built on top of an enterprise data warehouse (EDW) that spans several subject areas. And then you might have some data that isn’t in EDW and therefore outside the semantic model, such as HR file extracts in a lakehouse. You can use Fabric IQ as a glue that bridges that data together. And so, when the user asks the agent “correlate revenue by employee with hours they worked”, the agent knows where to go for answers. This screenshot shows how you can define such relationships between two entities.</p>
<p><img decoding="async" src="https://learn.microsoft.com/en-us/fabric/iq/ontology/media/tutorial-1-create-ontology/semantic-model/all-entity-types.png" alt="Screenshot of the renamed entity types." /></p>
<p>Following this line of thinking, Microsoft BI practitioners may view Fabric IQ as a Power BI composite semantic model on steroids. The big difference is that a composite model can only reference other semantic models while Fabric IQ can span data in multiple formats.</p>
<h2>The Good</h2>
<p>Palantir had a head start of a decade or so compared to Microsoft Fabric, but yet even in its preview stage, I like a thing or two about Fabric IQ from what I’ve seen so far:</p>
<ul>
<li>Its oncology can span Power BI semantic models (with caveats explained in the next section), powered by best-in-class technology. As I mentioned before, this allows you to bridge all the business logic and calculations you carefully crafted in a semantic model to the rest of your Fabric data estate.</li>
<li>Fabric IQ integrates with other Microsoft technologies, such as real-time intelligence (eventhouses), Copilot Studio, Graph. This tight integration turns Fabric into a true &#8220;intelligence platform,&#8221; reducing duplicated logic, one-off models, and maintenance while enabling multi-hop reasoning and real-time operational agents.</li>
<li>Democratized and no-code friendly &#8211; Visual tools allow business users to build and evolve the ontology, lowering barriers compared to more engineering-heavy alternatives. Making it easy to use has always been a Microsoft strength.</li>
<li>Groundbreaking semantics for AI Agents: Fabric IQ elevates AI from pattern-matching to true business understanding, allowing agents to reason over cascading effects, constraints, and objectives—leading to more reliable, auditable decisions and automation.</li>
<li>Compared to Palantir, I also like that Fabric OneLake has standardized on an open Delta Parquet format and embraced data movements tools Microsoft BI pros and business users are already familiar with, such as Dataflows and pipelines, to bring data in Fabric and therefore Fabric IQ.</li>
</ul>
<h2>The Bad</h2>
<p>I hope some of these limitations will be lifted after the preview but:</p>
<ul>
<li>Only DirectLake semantic models <a href="https://learn.microsoft.com/en-us/fabric/iq/ontology/concepts-generate">are accessible</a> to AI agents. Import and DirectQuery models are not currently supported for entity and relationships binding. Not only does this limitation rule out pretty much 99.9% of the existing semantic models, but it also prevents useful business scenarios, such as accessing the data where it is with DirectQuery instead of duplicating the data in OneLake.</li>
<li>No automatic ontology building – It requires cross-functional agreement on business definitions, workshops, and governance—labor-intensive for organizations without mature semantic models. I hope Microsoft will simplify this process like how Purview has automated scans.</li>
<li>Risk of overhype vs. delivery gap – We’ve seen this before when new products got unveiled with a lot of fanfare, only to be abandoned later.</li>
</ul>
<h2>The Ugly</h2>
<p>OneLake-centric dependency. Except for shortcuts to Delta Parquet files which can be kept external, your data must be in OneLake. What about these enterprises with investments in Google BigQuery, Teradata, Snowflake, and even SQL Server or Azure SQL DB? Gotta bring that data over to OneLake. Even shortcut transformations to CSV, Parquet, JSON files in OneLake, S3, Google Cloud Storage, will copy the data to OneLake. By contrast, Palantir has limited support for virtual tables to some popular file formats, such as Parquet, Iceberg, Delta, etc.</p>
<p>What happened to all the investments in data virtualization and logical warehouses that Microsoft has made over years, such as <a href="https://prologika.com/prologika-newsletter-winter-2021/">PolyBase</a> and the deprecated <a href="https://prologika.com/synapse-serverless-the-good-the-bad-and-the-ugly/">Polaris in Synapse Serverless</a>? What’s this fascination with copying data and having all the data in OneLake? Why can’t we build Fabric IQ on top of true data virtualization?</p>
<p>Which is where I was thinking that semantic models with DirectQuery can be used as a workaround to avoid copying data over from supported data sources, but alas Fabric IQ doesn’t like them yet.</p>
<h2>Summary</h2>
<p>Microsoft Fabric IQ is a metadata layer on top of Fabric data to build ontologies and expose relevant data to AI reasoning. It will be undoubtedly appealing to enterprise customers with complex data estates and existing investments in Power BI and Fabric. However, as it stands, Fabric IQ is OneLake-centric. Expect Microsoft to invest heavily in Fabric and Fabric IQ to compete better with Palantir.</p>
<p><img decoding="async" src="https://prologika.com/wp-content/uploads/2017/06/060417_1725_PrologikaNe2.png" alt="" /><br />
Teo Lachev<br />
Prologika, LLC | Making Sense of Data<br />
<a href="https://prologika.com/wp-content/uploads/2016/01/logo.png" rel="attachment wp-att-12"><img decoding="async" class="alignnone size-full wp-image-12" src="https://prologika.com/wp-content/uploads/2016/01/logo.png" alt="logo" width="165" height="45" /></a></p>
]]></content:encoded>
					
					<wfw:commentRss>https://prologika.com/prologika-newsletter-spring-2026/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>First Look at Fabric IQ: The Good, The Bad, and The Ugly</title>
		<link>https://prologika.com/first-look-at-fabric-iq-the-good-the-bad-and-the-ugly/</link>
					<comments>https://prologika.com/first-look-at-fabric-iq-the-good-the-bad-and-the-ugly/#respond</comments>
		
		<dc:creator><![CDATA[Prologika - Teo Lachev]]></dc:creator>
		<pubDate>Sat, 27 Dec 2025 21:09:38 +0000</pubDate>
				<category><![CDATA[Blog]]></category>
		<category><![CDATA[Data Virtualization]]></category>
		<category><![CDATA[Fabric]]></category>
		<category><![CDATA[Lakehouse]]></category>
		<category><![CDATA[Power BI]]></category>
		<guid isPermaLink="false">https://prologika.com/?p=9533</guid>

					<description><![CDATA[Telegraph sang a song about the world outside Telegraph road got so deep and so wide Like a rolling river… The Telegraph Road, Dire Straits At Ignite in November, 2025, [&#8230;]]]></description>
										<content:encoded><![CDATA[<p><em>Telegraph sang a song about the world outside<br />
Telegraph road got so deep and so wide<br />
Like a rolling river…</em></p>
<p><em>The Telegraph Road, Dire Straits</em></p>
<p>At Ignite in November, 2025, Microsoft <a href="https://youtu.be/RjU0slwcZGs?si=QTRsZMg-jFQ5wsB0">introduced</a> <a href="https://youtu.be/RjU0slwcZGs?si=QTRsZMg-jFQ5wsB0">Fabric IQ</a>. I noted to go beyond the marketing hype and check if Fabric IQ makes any sense. The next thing I know, around the holidays I’m talking to an enterprise strategy manager from an airline company and McKinsey consultant about ontologies.</p>
<p><em>Ontology – A branch of philosophy, ontology is the study of being that investigates the nature of existence, the features all entities have in common, and how they are divided into basic categories of being. In computer science and AI, ontology refers to a set of concepts and categories in a subject area or domain that shows their properties and the relations between them.</em></p>
<p>So, what better way to spend the holidays than to play with new shaky software?</p>
<h2>What is Fabric IQ?</h2>
<p>According to Microsoft, Fabric IQ is “a unified intelligence platform developed by Microsoft that enhances data management and decision-making through semantic understanding and AI capabilities.” Clear enough? If not, if you view Fabric as Microsoft’s answer to Palantir’s Foundry, then Fabric IQ is the Microsoft equivalent of Palantir’s Foundry Ontology, whose success apparently inspired Microsoft.</p>
<blockquote><p>Therefore, my unassuming layman definition of Fabric IQ is a metadata layer on top of data in Fabric that defines entities and their relationships so that AI can make sense of and relate the underlying data.</p></blockquote>
<p>For example, you may have an organizational semantic model built on top of an enterprise data warehouse (EDW) that spans several subject areas. And then you might have some data that isn’t in EDW and therefore outside the semantic model, such as HR file extracts in a lakehouse. You can use Fabric IQ as a glue that bridges that data together. And so, when the user asks the agent “correlate revenue by employee with hours they worked”, the agent knows where to go for answers.</p>
<p>Following this line of thinking, Microsoft BI practitioners may view Fabric IQ as a Power BI composite semantic model on steroids. The big difference is that a composite model can only reference other semantic models while Fabric IQ can span data in multiple formats.</p>
<h2>The Good</h2>
<p>Palantir had a head start of a decade or so compared to Microsoft Fabric, but yet even in its preview stage, I like a thing or two about Fabric IQ from what I’ve seen so far:</p>
<ul>
<li>Its oncology can span Power BI semantic models (with caveats explained in the next section), powered by best-in-class technology. As I mentioned before, this allows you to bridge all the business logic and calculations you carefully crafted in a semantic model to the rest of your Fabric data estate.</li>
<li>Fabric IQ integrates with other Microsoft technologies, such as real-time intelligence (eventhouses), Copilot Studio, Graph. This tight integration turns Fabric into a true &#8220;intelligence platform,&#8221; reducing duplicated logic, one-off models, and maintenance while enabling multi-hop reasoning and real-time operational agents.</li>
<li>Democratized and no-code friendly &#8211; Visual tools allow business users to build and evolve the ontology, lowering barriers compared to more engineering-heavy alternatives. Making it easy to use has always been a Microsoft strength.</li>
<li>Groundbreaking semantics for AI Agents: Fabric IQ elevates AI from pattern-matching to true business understanding, allowing agents to reason over cascading effects, constraints, and objectives—leading to more reliable, auditable decisions and automation.</li>
<li>Compared to Palantir, I also like that Fabric OneLake has standardized on an open Delta Parquet format and embraced data movements tools Microsoft BI pros and business users are already familiar with, such as Dataflows and pipelines, to bring data in Fabric and therefore Fabric IQ.</li>
</ul>
<h2>The Bad</h2>
<p>I hope some of these limitations will be lifted after the preview but:</p>
<ul>
<li>Only DirectLake semantic models <a href="https://learn.microsoft.com/en-us/fabric/iq/ontology/concepts-generate">are accessible</a> to AI agents. Import and DirectQuery models are not currently supported for entity and relationships binding. Not only does this limitation rule out pretty much 99.9% of the existing semantic models, but it also prevents useful business scenarios, such as accessing the data where it is with DirectQuery instead of duplicating the data in OneLake.</li>
<li>No automatic ontology building – It requires cross-functional agreement on business definitions, workshops, and governance—labor-intensive for organizations without mature semantic models. I hope Microsoft will simplify this process like how Purview has automated scans.</li>
<li>Risk of overhype vs. delivery gap – We’ve seen this before when new products got unveiled with a lot of fanfare, only to be abandoned later.</li>
</ul>
<h2>The Ugly</h2>
<p>OneLake-centric dependency. Except for shortcuts to Delta Parquet files which can be kept external, your data must be in OneLake. What about these enterprises with investments in Google BigQuery, Teradata, Snowflake, and even SQL Server or Azure SQL DB? Gotta bring that data over to OneLake. Even shortcut transformations to CSV, Parquet, JSON files in OneLake, S3, Google Cloud Storage, will copy the data to OneLake. By contrast, Palantir has limited support for virtual tables to some popular file formats, such as Parquet, Iceberg, Delta, etc.</p>
<p>What happened to all the investments in data virtualization and logical warehouses that Microsoft has made over years, such as <a href="https://prologika.com/prologika-newsletter-winter-2021/">PolyBase</a> and the deprecated <a href="https://prologika.com/synapse-serverless-the-good-the-bad-and-the-ugly/">Polaris in Synapse Serverless</a>? What’s this fascination with copying data and having all the data in OneLake? Why can’t we build Fabric IQ on top of true data virtualization?</p>
<p>Which is where I was thinking that semantic models with DirectQuery can be used as a workaround to avoid copying data over from supported data sources, but alas Fabric IQ doesn’t like them yet.</p>
<h2>Summary</h2>
<p>Microsoft Fabric IQ is a metadata layer on top of Fabric data to build ontologies and expose relevant data to AI reasoning. It will be undoubtedly appealing to enterprise customers with complex data estates and existing investments in Power BI and Fabric. However, as it stands, Fabric IQ is OneLake-centric. Expect Microsoft to invest heavily in Fabric and Fabric IQ to compete better with Palantir.</p>
<p><img decoding="async" class="wp-image-9535" src="https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1.png" width="202" height="202" srcset="https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1.png 1024w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-300x300.png 300w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-80x80.png 80w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-768x768.png 768w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-36x36.png 36w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-180x180.png 180w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-705x705.png 705w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-120x120.png 120w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-450x450.png 450w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-50x50.png 50w, https://prologika.com/wp-content/uploads/2025/12/word-image-9533-1-100x100.png 100w" sizes="(max-width: 202px) 100vw, 202px" /></p>
]]></content:encoded>
					
					<wfw:commentRss>https://prologika.com/first-look-at-fabric-iq-the-good-the-bad-and-the-ugly/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Prologika Newsletter Winter 2025</title>
		<link>https://prologika.com/prologika-newsletter-winter-2025/</link>
					<comments>https://prologika.com/prologika-newsletter-winter-2025/#respond</comments>
		
		<dc:creator><![CDATA[Prologika - Teo Lachev]]></dc:creator>
		<pubDate>Sat, 20 Dec 2025 00:05:32 +0000</pubDate>
				<category><![CDATA[Newsletter]]></category>
		<category><![CDATA[ETL]]></category>
		<category><![CDATA[Fabric]]></category>
		<category><![CDATA[Lakehouse]]></category>
		<guid isPermaLink="false">https://prologika.com/?p=9522</guid>

					<description><![CDATA[If Microsoft Fabric is in your future, you need to come up with a strategy to get your data in Fabric OneLake. That’s because the holy grail of Fabric is [&#8230;]]]></description>
										<content:encoded><![CDATA[<p><img loading="lazy" decoding="async" class="wp-image-9478" style="padding: 0px 10px;" src="https://prologika.com/wp-content/uploads/2025/12/word-image-9522-1.png" alt="Diogenes holding a lantern" width="176" height="102" align="left" /></p>
<p>If Microsoft Fabric is in your future, you need to come up with a strategy to get your data in Fabric OneLake. That’s because the holy grail of Fabric is the Delta Parquet file format. The good news is that all Fabric data ingestion options (Dataflows Gen 2, pipelines, Copy Job and notebooks) support this format and the Microsoft <a href="https://learn.microsoft.com/en-us/fabric/data-warehouse/v-order">V-Order extension</a> that’s important for Direct Lake performance. Fabric also supports mirroring data from a growing list of data sources. This could be useful if your data is outside Fabric, such as EDW hosted in Google BigQuery, which is the scenario discussed in this newsletter.</p>
<h2>Avoiding mirroring issues</h2>
<p>A recent engagement required replicating some DW tables from Google BigQuery to a Fabric Lakehouse. We considered the <a href="https://learn.microsoft.com/en-us/fabric/mirroring/google-bigquery">Fabric mirroring feature for Google BigQuery</a> (back then in private preview, now in public preview) and learned some lessons along the way:</p>
<p>1. 400 Error during replication configuration – Caused by attempting to use a read-only GBQ dataset that is linked to another GBQ dataset, but the link was broken.</p>
<p>2. Internal System Error – Again caused by GBQ linked datasets which are read-only. Fabric mirroring requires GBQ change history to be enabled on tables so that it can track changes and only mirror incremental changes after first initial load.</p>
<p>3. (Showstopper for this project) The two permissions that raised security red flags are bigquery.datasets.create and bigquery.jobs.create. To grant those permissions, you must assign one of these BigQuery roles:</p>
<p>• BigQuery Admin</p>
<p>• BigQuery Data Editor</p>
<p>• BigQuery Data Owner</p>
<p>• BigQuery Studio Admin</p>
<p>• BigQuery User</p>
<p>All these roles grant other permissions, and the client was cautious about data security. At the end, we end up using a nightly Fabric Copy Job to replicate the data.</p>
<h2>Fabric Copy Job Pros and Cons</h2>
<p>The client was overall pleased with the Fabric Copy Job.</p>
<p><strong>Pros</strong></p>
<ul>
<li>250 million rows replicated in 30-40 seconds!</li>
<li>You can have only one job to replicate all tables in Overwrite mode.</li>
<li>In the simplest case, you don’t need to create pipelines.</li>
</ul>
<p><strong>Cons </strong></p>
<p>The Copy Job is work in progress and subject to various limitations.</p>
<ul>
<li>No incremental extraction</li>
<li>You can’t mix different load options (Append and Overwrite) so you must split tables in separate jobs</li>
<li>No custom SQL SELECT when copying multiple tables</li>
<li>(Bug) Lost explicit column bindings when making changes</li>
<li>Cannot change the job’s JSON file</li>
<li>The user interface is clunky and it’s difficult to work with</li>
<li>No failure notification mechanism. As a workaround: add Copy Job to data pipeline or call it via REST API</li>
</ul>
<h2>Summary</h2>
<p>In summary, the Fabric Google BigQuery built-in mirroring could be useful for real-time data replication. However, it relies on GBQ change history which requires certain permissions. Kudos to Microsoft for their excellent support during the private preview.</p>
<p><img decoding="async" src="https://prologika.com/wp-content/uploads/2017/06/060417_1725_PrologikaNe2.png" alt="" /><br />
Teo Lachev<br />
Prologika, LLC | Making Sense of Data<br />
<a href="https://prologika.com/wp-content/uploads/2016/01/logo.png" rel="attachment wp-att-12"><img loading="lazy" decoding="async" class="alignnone size-full wp-image-12" src="https://prologika.com/wp-content/uploads/2016/01/logo.png" alt="logo" width="165" height="45" /></a></p>
]]></content:encoded>
					
					<wfw:commentRss>https://prologika.com/prologika-newsletter-winter-2025/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Prologika Newsletter Summer 2024</title>
		<link>https://prologika.com/prologika-newsletter-summer-2024/</link>
					<comments>https://prologika.com/prologika-newsletter-summer-2024/#respond</comments>
		
		<dc:creator><![CDATA[Prologika - Teo Lachev]]></dc:creator>
		<pubDate>Sun, 16 Jun 2024 15:11:21 +0000</pubDate>
				<category><![CDATA[Newsletter]]></category>
		<category><![CDATA[Data Warehousing]]></category>
		<category><![CDATA[Fabric]]></category>
		<category><![CDATA[Lakehouse]]></category>
		<guid isPermaLink="false">https://prologika.com/?p=9234</guid>

					<description><![CDATA[I’ve written in the past about the dangers of blindly following “modern” data architectures (see the “Are you modern yet?” and “Data Lakehouse: The Good, the Bad, and the Ugly”) [&#8230;]]]></description>
										<content:encoded><![CDATA[<p><img loading="lazy" decoding="async" class="" style="padding: 0px 10px;" src="https://prologika.com/wp-content/uploads/2024/06/lake.png" alt="" width="112" height="99" align="left" />I’ve written in the past about the dangers of blindly following “modern” data architectures (see the “<a href="https://prologika.com/are-you-modern-yet/">Are you modern yet?</a>” and “<a href="https://prologika.com/data-lakehouse-the-good-the-bad-and-the-ugly/">Data Lakehouse: The Good, the Bad, and the Ugly</a>”) but a recent assessment inspired to me write about this topic again. This newsletter advocates a hybrid and cautionary approach for data integration to avoid overdoing data lakes and warns about pitfalls of over-staging source data to files. It recommends instead following the &#8220;Discipline at the core, flexibility at the edge&#8221; methodology with emphasis on implementing enterprise data warehouse and organizational semantic models.</p>
<h1>Data Lake Overstaging</h1>
<p>How did the large vendor attempt to solve these horrible issues? Modern Data Warehouse (MDM) architecture of course. Nothing wrong with it except that EDW and organizational semantic model(s) are missing and that most of the effort went into implementing the data lake medallion architecture where all the incoming data ended up staged as Parquet files. It didn’t matter that 99% of the data came from relational databases. Further, to solve a data change tracking requirement, the vendor decided to create a new file each time ETL runs. So even if nothing has changed in the source feed, the data is duplicated should one day the user wants to go back in time and see what the data looked like then. There are of course better ways to handle this that doesn’t even require ETL, such as SQL Server temporal tables, but I digress.</p>
<p>At least some cool heads prevailed and the Silver layer got implemented as a relational ODS to serve the needs of home-grown applications, so the apps didn’t have to deal with files. What about EDW and organizational semantic models? Not there because the project ran out of budget and time. I bet if that vendor got hired today, they would have gone straight for Fabric Lakehouse and Fabric premium pricing (nowadays Microsoft treats partners as an extension to its salesforce and requires them to meet certain revenue targets as I explain in “<a href="https://prologika.com/dissolving-partnerships/">Dissolving Partnerships</a>“), which alone would have produced the same outcome.</p>
<blockquote><p>What did the vendor accomplish? Not much. Nor only didn’t the implementation address the main challenges, but it introduced new, such as overcomplicated ETL and redundant data staging. Although there might be good reasons for file staging (see the second blog above), in most cases I consider it a lunacy to stage perfect relational data to files, along the way losing metadata, complicating ETL, ending up serverless, and then reloading the same data into a relational database (ODS in this case).</p></blockquote>
<p>I’ve heard that the vendor justified the lake effort by empowering data scientists to do ML one day. I’d argue that if that day ever comes, the likelihood (pun not intended) of data scientists working directly on the source schema would be infinitely small since more than likely they would require the input datasets to be shaped in a different way which would probably require another ETL pipeline altogether.</p>
<h1>Better Data Staging</h1>
<p>I don’t subject my clients to excessive file staging. My file staging litmus test is what’s the source data format. If I can connect to a server and get in a tabular (relational) format, I stage it directly to a relational database (ODS or DW). However, if it’s provided as files (downloaded or pushed, reference data, or unstructured data), then obviously there is no other way. That’s why we have lakes.</p>
<p><img loading="lazy" decoding="async" class="alignnone wp-image-9215" src="https://prologika.com/wp-content/uploads/2024/06/a-diagram-of-a-computer-data-processing-descripti-4.jpeg" alt="A diagram of a computer data processing Description automatically generated" width="468" height="444" srcset="https://prologika.com/wp-content/uploads/2024/06/a-diagram-of-a-computer-data-processing-descripti-4.jpeg 946w, https://prologika.com/wp-content/uploads/2024/06/a-diagram-of-a-computer-data-processing-descripti-4-300x284.jpeg 300w, https://prologika.com/wp-content/uploads/2024/06/a-diagram-of-a-computer-data-processing-descripti-4-768x728.jpeg 768w, https://prologika.com/wp-content/uploads/2024/06/a-diagram-of-a-computer-data-processing-descripti-4-705x668.jpeg 705w, https://prologika.com/wp-content/uploads/2024/06/a-diagram-of-a-computer-data-processing-descripti-4-450x427.jpeg 450w" sizes="auto, (max-width: 468px) 100vw, 468px" /></p>
<p>Fast forward a few years, and your humble correspondent got hired to assess the damage and come up with a strategy. Data lakes won’t do it. Lakehouses and Delta Parquet (a poor attempt to recreate and replace relational databases) won’t do it. Fabric won’t do it and it’s too bad that Microsoft pushes Lakehouse while the main focus should have been on Fabric Data Warehouse, which unfortunately is not ready for prime time (fortunately, we have plenty of other options).</p>
<blockquote><p>What will do it? Going back to the basics and embracing the “<a href="https://learn.microsoft.com/en-us/power-bi/guidance/center-of-excellence-microsoft-business-intelligence-transformation">Discipline at the core, flexibility at edge</a>” ideology (kudos to Microsoft for publishing their lessons learned). From a technology standpoint, the critical pieces are EDW and organizational semantic models. If you don’t have these, I’m sorry but you are not modern yet. In fact, you aren’t even classic, considering that they have been around for long, long time.</p></blockquote>
<p><img decoding="async" src="https://prologika.com/wp-content/uploads/2017/06/060417_1725_PrologikaNe2.png" alt="" /><br />
Teo Lachev<br />
Prologika, LLC | Making Sense of Data<br />
<a href="https://prologika.com/wp-content/uploads/2016/01/logo.png" rel="attachment wp-att-12"><img loading="lazy" decoding="async" class="alignnone size-full wp-image-12" src="https://prologika.com/wp-content/uploads/2016/01/logo.png" alt="logo" width="165" height="45" /></a></p>
<p>&nbsp;</p>
<p>&nbsp;</p>
]]></content:encoded>
					
					<wfw:commentRss>https://prologika.com/prologika-newsletter-summer-2024/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Modern Data Warehouse (MDM) Reloaded</title>
		<link>https://prologika.com/modern-data-warehouse-mdm-reloaded/</link>
					<comments>https://prologika.com/modern-data-warehouse-mdm-reloaded/#respond</comments>
		
		<dc:creator><![CDATA[Prologika - Teo Lachev]]></dc:creator>
		<pubDate>Sat, 01 Jun 2024 21:35:29 +0000</pubDate>
				<category><![CDATA[Blog]]></category>
		<category><![CDATA[Azure Data Lake]]></category>
		<category><![CDATA[Data Warehousing]]></category>
		<category><![CDATA[General]]></category>
		<category><![CDATA[Lakehouse]]></category>
		<guid isPermaLink="false">https://prologika.com/?p=9209</guid>

					<description><![CDATA[&#8220;Where are the prophets, where are the visionaries, where are the poets To breach the dawn of the sentimental mercenary&#8221; &#8220;Fugazi&#8221;, Marillion I’ve written in the past about the dangers [&#8230;]]]></description>
										<content:encoded><![CDATA[<p><em>&#8220;Where are the prophets, where are the visionaries, where are the poets</em><br />
<em>To breach the dawn of the sentimental mercenary&#8221;<br />
</em><em>&#8220;Fugazi&#8221;, Marillion</em></p>
<p>I’ve written in the past about the dangers of blindly following “modern” data architectures (see my posts “<a href="https://prologika.com/are-you-modern-yet/">Are you modern yet?</a>” and “<a href="https://prologika.com/data-lakehouse-the-good-the-bad-and-the-ugly/">Data Lakehouse: The Good, the Bad, and the Ugly</a>”) but here we go again. Once upon a time, a large company hired a large Microsoft partner to solve a common and pervasive challenge. Left on their own, departments have set up their own data servers, thus creating data silos leading to data duplication and inconsistent results.</p>
<p>How did the large vendor attempt to solve these horrible issues? Modern Data Warehouse (MDM) architecture of course. Nothing wrong with it except that EDW and organizational semantic model(s) are missing and that most of the effort went into implementing the data lake medallion architecture where all the incoming data ended up staged as Parquet files. It didn’t matter that 99% of the data came from relational databases. Further, to solve a data change tracking requirement, the vendor decided to create a new file each time ETL runs. So even if nothing has changed in the source feed, the data is duplicated should one day the user wants to go back in time and see what the data looked like then. There are of course better ways to handle this that doesn’t even require ETL, such as SQL Server temporal tables, but I digress.</p>
<p>At least some cool heads prevailed and the Silver layer got implemented as a relational ODS to serve the needs of home-grown applications, so the apps didn’t have to deal with files. What about EDW and organizational semantic models? Not there because the project ran out of budget and time. I bet if that vendor got hired today, they would have gone straight for Fabric Lakehouse and Fabric premium pricing (nowadays Microsoft treats partners as an extension to its salesforce and requires them to meet certain revenue targets as I explain in &#8220;<a href="https://prologika.com/dissolving-partnerships/">Dissolving Partnerships</a>&#8220;), which alone would have produced the same outcome.</p>
<blockquote><p>What did the vendor accomplish? Not much. Nor only didn’t the implementation address the main challenges, but it introduced new, such as overcomplicated ETL and redundant data staging. Although there might be good reasons for file staging (see the second blog above), in most cases I consider it a lunacy to stage perfect relational data to files, along the way losing metadata, complicating ETL, ending up serverless, and then reloading the same data into a relational database (ODS in this case).</p></blockquote>
<p>I’ve heard that the vendor justified the lake effort by empowering data scientists to do ML one day. I’d argue that if that day ever comes, the likelihood (pun not intended) of data scientists working directly on the source schema would be infinitely small since more than likely they would require the input datasets to be shaped in a different way which would probably require another ETL pipeline altogether.</p>
<p>I don’t subject my clients to excessive file staging. My file staging litmus test is what’s the source data format. If I can connect to a server and get in a tabular (relational) format, I stage it directly to a relational database (ODS or DW). However, if it’s provided as files (downloaded or pushed, reference data, or unstructured data), then obviously there is no other way. That’s why we have lakes.</p>
<p>Fast forward a few years, and your humble correspondent got hired to assess the damage and come up with remediation. Data lakes won’t do it. Lakehouses and Delta Parquet (a poor attempt to recreate and replace relational databases) won’t do it. Fabric won’t do it and it’s too bad that Microsoft pushes Lakehouse while the main focus should be Fabric Data Warehouse, which unfortunately is not ready for prime time (but we have plenty of other options).</p>
<blockquote><p>What will do it? Going back to the basics and embracing the “<a href="https://learn.microsoft.com/en-us/power-bi/guidance/center-of-excellence-microsoft-business-intelligence-transformation">Discipline at the core, flexibility at edge</a>” ideology (kudos to Microsoft for publishing their lessons learned). From a technology standpoint, the critical pieces are EDW and organizational semantic models. If you don’t have these, I’m sorry but you are not modern yet. In fact, you aren&#8217;t even classic, considering that they have been around for long, long time.</p></blockquote>
<p>So, keep on implementing these medallion lakes to keep my consulting pipeline full. Drop a comment how they work for you.</p>
<p><img loading="lazy" decoding="async" class="wp-image-9215" src="https://prologika.com/wp-content/uploads/2024/06/a-diagram-of-a-computer-data-processing-descripti-4.jpeg" alt="A diagram of a computer data processing Description automatically generated" width="393" height="373" srcset="https://prologika.com/wp-content/uploads/2024/06/a-diagram-of-a-computer-data-processing-descripti-4.jpeg 946w, https://prologika.com/wp-content/uploads/2024/06/a-diagram-of-a-computer-data-processing-descripti-4-300x284.jpeg 300w, https://prologika.com/wp-content/uploads/2024/06/a-diagram-of-a-computer-data-processing-descripti-4-768x728.jpeg 768w, https://prologika.com/wp-content/uploads/2024/06/a-diagram-of-a-computer-data-processing-descripti-4-705x668.jpeg 705w, https://prologika.com/wp-content/uploads/2024/06/a-diagram-of-a-computer-data-processing-descripti-4-450x427.jpeg 450w" sizes="auto, (max-width: 393px) 100vw, 393px" /></p>
<p>&nbsp;</p>
]]></content:encoded>
					
					<wfw:commentRss>https://prologika.com/modern-data-warehouse-mdm-reloaded/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>What Can Fabric Do For My Lake?</title>
		<link>https://prologika.com/what-can-fabric-do-for-my-lake/</link>
					<comments>https://prologika.com/what-can-fabric-do-for-my-lake/#respond</comments>
		
		<dc:creator><![CDATA[Prologika - Teo Lachev]]></dc:creator>
		<pubDate>Wed, 13 Mar 2024 18:19:36 +0000</pubDate>
				<category><![CDATA[Blog]]></category>
		<category><![CDATA[Architecture]]></category>
		<category><![CDATA[Fabric]]></category>
		<category><![CDATA[Lakehouse]]></category>
		<guid isPermaLink="false">https://prologika.com/?p=9163</guid>

					<description><![CDATA[Previously, I discussed the pros and cons of Microsoft Fabric OneLake and Lakehouse. But what if you have a data lake already? Will Fabric add any value, especially if your [&#8230;]]]></description>
										<content:encoded><![CDATA[<p>Previously, I discussed the pros and cons of Microsoft Fabric <a href="https://prologika.com/fabric-onelake-the-good-the-bad-and-the-ugly/">OneLake</a> and <a href="https://prologika.com/fabric-lakehouse-the-good-the-bad-and-the-ugly/">Lakehouse</a>. But what if you have a data lake already? Will Fabric add any value, especially if your organization is on Power BI Premium and you get Fabric features for free (that is, assuming you are not overloading your capacity resources)? Well, it depends.</p>
<h1>Managed Area</h1>
<p>A Fabric lakehouse defines two areas: managed and unmanaged. The managed area (Tables folder) is exclusively for Delta/Parquet tables. If you have your own data lake with Delta/Parquet files, such as Databricks delta lake, you can create shortcuts to these files or folders located in ADLS Gen 2 or Amazon S3. Consequently, the Fabric lakehouse would automatically register these shortcuts as tables.</p>
<p><img loading="lazy" decoding="async" class="alignnone wp-image-9164" src="https://prologika.com/wp-content/uploads/2024/03/lakehouse_folders.jpg" alt="" width="325" height="143" srcset="https://prologika.com/wp-content/uploads/2024/03/lakehouse_folders.jpg 541w, https://prologika.com/wp-content/uploads/2024/03/lakehouse_folders-300x132.jpg 300w, https://prologika.com/wp-content/uploads/2024/03/lakehouse_folders-450x198.jpg 450w" sizes="auto, (max-width: 325px) 100vw, 325px" /></p>
<p>Life is good in the managed area. Shortcuts to Delta/Parquet tables open interesting possibilities for data virtualization, such as:</p>
<ol>
<li>Your users can use the Lakehouse SQL Analytics endpoint to join tables using SQL. This is useful for ad-hoc analysis. Joins could also be useful so users can shape the data they need before importing it in Power BI Desktop as opposed to connecting to individual files and using Power Query to join the tables. Not only could this reduce the size of the ingested data, but it could also improve refresh performance.</li>
<li>Users can decide not to import the data at all but build semantic models in <a href="https://learn.microsoft.com/en-us/power-bi/enterprise/directlake-overview">Direct Lake mode</a>. This could be very useful to reduce latency or avoid caching large volumes of data.</li>
</ol>
<h1>Unmanaged Area</h1>
<p>Very few organizations would have lakes with Delta Parquet files. Most data lakes contain heterogeneous files, such as text, Excel, or regular Parquet files. While a Fabric lakehouse can create shortcuts to any file, non Delta/Parquet shortcuts will go to the unmanaged area (Files folder).</p>
<p>Life is miserable in the unmanaged area. None of the cool stuff you see in demos happens here because the analytical endpoint and direct lake modes are not available. A weak case can still be made for data virtualization that shortcuts bring data readily available to where business users collaborate in Power BI: the Power BI workspace.</p>
<p>But what can the user do with these unmanaged shortcuts? Not much really. Power BI Desktop doesn’t even expose them when you connect to the lakehouse. Power BI dataflows Gen2 do give the user access to the Files folder so potentially users can create dataflows and transform data from these files.</p>
<p><img loading="lazy" decoding="async" class="alignnone wp-image-9165" src="https://prologika.com/wp-content/uploads/2024/03/dataflowgen2.jpg" alt="" width="414" height="261" srcset="https://prologika.com/wp-content/uploads/2024/03/dataflowgen2.jpg 827w, https://prologika.com/wp-content/uploads/2024/03/dataflowgen2-300x189.jpg 300w, https://prologika.com/wp-content/uploads/2024/03/dataflowgen2-768x484.jpg 768w, https://prologika.com/wp-content/uploads/2024/03/dataflowgen2-705x444.jpg 705w, https://prologika.com/wp-content/uploads/2024/03/dataflowgen2-450x283.jpg 450w" sizes="auto, (max-width: 414px) 100vw, 414px" /></p>
<p>Of course, the tradeoff here is that you are adding dependencies to OneLake which could be a problem should one day you decide to part ways. Another issue could be that you are layering Power BI security on top of your data lake security.</p>
<p>Oh yes, users can also load Parquet and CSV files to Delta tables by right-clicking a folder or a file in the Unmanaged area, and then selecting Load to Tables (New or Existing). Unfortunately, as it stands, this is a manual process that must be repeated when the source data changes.</p>
<h1>Imagined Unmanaged Data Virtualization</h1>
<blockquote><p>This brings me to the two things that I believe Microsoft can do to greatly increase the value proposition of “unmanaged” data virtualization:</p></blockquote>
<ol>
<li>Extend load to table to the most popular file formats, such as JSON, XML, and Excel. Or, at least the ones that <a href="https://learn.microsoft.com/en-us/sql/t-sql/statements/create-external-file-format-transact-sql">Polybase has been supporting</a> for years. Not sure why we have to obsess with Delta Parquet and nothing else if Microsoft is serious about data virtualization.</li>
<li>Implement automatic synchronization to update the corresponding Delta table when the source file changes.</li>
</ol>
<p>If these features are added, throwing Fabric to the mix could become more appealing.</p>
<p>In summary, Microsoft Fabric has embraced Delta Parquet as its native storage file format and has added various features that targets it. Unfortunately none of these features extend to other file formats. You must evaluate pros and cons when adopting Fabric with existing data lakes. As it stands, Fabric probably wouldn’t add much business value for data virtualization over file formats other than Delta Paquet files. As Fabric matures, new scenarios might be feasible to justify Fabric integration and dependency.</p>
<p>&nbsp;</p>
<p>&nbsp;</p>
]]></content:encoded>
					
					<wfw:commentRss>https://prologika.com/what-can-fabric-do-for-my-lake/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Atlanta Microsoft BI Group Meeting on September 11th (Introducing Lakehouse in Microsoft Fabric)</title>
		<link>https://prologika.com/atlanta-microsoft-bi-group-meeting-202309/</link>
					<comments>https://prologika.com/atlanta-microsoft-bi-group-meeting-202309/#respond</comments>
		
		<dc:creator><![CDATA[Prologika - Teo Lachev]]></dc:creator>
		<pubDate>Tue, 05 Sep 2023 21:01:18 +0000</pubDate>
				<category><![CDATA[Blog]]></category>
		<category><![CDATA[Events]]></category>
		<category><![CDATA[Atlanta.MBI]]></category>
		<category><![CDATA[Fabric]]></category>
		<category><![CDATA[Lakehouse]]></category>
		<category><![CDATA[Power BI]]></category>
		<guid isPermaLink="false">https://prologika.com/?p=8991</guid>

					<description><![CDATA[Atlanta BI fans, please join us for the next meeting on Monday, September 11th, at 6:30 PM ET.  Shabnam Waston (BI Consultant and Microsoft MVP) will introduce us to the [&#8230;]]]></description>
										<content:encoded><![CDATA[<p>Atlanta BI fans, please join us for the next meeting on Monday, September 11th, at 6:30 PM ET.  Shabnam Waston (BI Consultant and Microsoft MVP) will introduce us to the Lakehouse engine in Microsoft Fabric. Shabnam will also sponsor the meeting. Your humble correspondent will help you catch up on Microsoft BI latest. For more details and sign up, visit our <a href="https://www.meetup.com/Atlanta-Microsoft-Business-Intelligence-Users/">group page</a>.</p>
<p>PLEASE NOTE A CHANGE TO OUR MEETING POLICY. WE HAVE DISCONTINUED ONLINE MEETINGS VIA TEAMS. THIS GROUP MEETS ONLY IN PERSON. WE WON’T RECORD MEETINGS ANYMORE. THEREFORE, AS DURING THE PRE-PANDEMIC TIMES, PLEASE RSVP AND ATTEND IN PERSON IF YOU ARE INTERESTED IN THIS MEETING.</p>
<p><strong>Presentation:</strong> Introducing Lakehouse in Microsoft Fabric</p>
<p><strong>Delivery:</strong> Onsite</p>
<p><strong>Date: </strong>September 11<sup>th</sup></p>
<p><strong>Time: </strong>18:30 – 20:30 ET</p>
<p><strong>Level</strong>: Beginner to Intermediate</p>
<p><strong>Food</strong>: Sponsor wanted</p>
<p>&nbsp;</p>
<p><strong>Agenda: </strong></p>
<p>18:15-18:30 Registration and networking</p>
<p>18:30-19:00 Organizer and sponsor time (events, Power BI latest, sponsor marketing)</p>
<p>19:00-20:15 Main presentation</p>
<p>20:15-20:30 Q&amp;A</p>
<p><strong> </strong></p>
<p><strong>VENUE</strong></p>
<p>Improving Office<br />
11675 Rainwater Dr<br />
Suite #100<br />
Alpharetta, GA 30009</p>
<p><strong>Overview:</strong> Join this session to learn about Lakehouse architecture in Microsoft Fabric. Microsoft Fabric is an end-to-end big data analytics platform that offers many capabilities including data integration, data engineering, data science, data lake, data warehouse, and many more, all in one unified SaaS model. In this session, you will learn how to create a lakehouse in Microsoft Fabric, load it with sample data using Notebooks/Pipelines, and work with its built-in SQL Endpoint as well as its default Power BI dataset which uses a brand-new storage mode called Direct Lake.</p>
<p><strong>Speaker: </strong>Shabnam Watson is a Business Intelligence consultant, speaker, blogger, and Microsoft Data Platform MVP with 20+ years of experience developing Data Warehouse and Business Intelligence solutions. Her work focus within the Microsoft BI Stack has been on Analysis Services and Power BI and most recently on Azure Synapse Analytics. She has worked across several industries including Supply Chain, Finance, Retail, Insurance, and Health Care. Her areas of interest include Power BI, Analysis Services, Performance Tuning, PowerShell, DevOps, Azure, Natural Language Processing, and AI. She is a regular speaker and volunteer at national and local user groups and conferences. She holds a bachelor’s degree in computer engineering and a master’s degree in computer science.</p>
<p>Sponsor: Shabnam Watson</p>
<p><a href="https://prologika.com/wp-content/uploads/2019/10/PowerBILogo.png" rel="attachment wp-att-6368"><img loading="lazy" decoding="async" class="alignnone size-full wp-image-6368" src="https://prologika.com/wp-content/uploads/2019/10/PowerBILogo.png" alt="PowerBILogo" width="410" height="109" srcset="https://prologika.com/wp-content/uploads/2019/10/PowerBILogo.png 410w, https://prologika.com/wp-content/uploads/2019/10/PowerBILogo-300x80.png 300w" sizes="auto, (max-width: 410px) 100vw, 410px" /></a></p>
]]></content:encoded>
					
					<wfw:commentRss>https://prologika.com/atlanta-microsoft-bi-group-meeting-202309/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Prologika Newsletter Spring 2023</title>
		<link>https://prologika.com/prologika-newsletter-spring-2023/</link>
					<comments>https://prologika.com/prologika-newsletter-spring-2023/#respond</comments>
		
		<dc:creator><![CDATA[Prologika - Teo Lachev]]></dc:creator>
		<pubDate>Sat, 04 Mar 2023 22:26:23 +0000</pubDate>
				<category><![CDATA[Newsletter]]></category>
		<category><![CDATA[Azure Data Lake]]></category>
		<category><![CDATA[Lakehouse]]></category>
		<category><![CDATA[Synapse]]></category>
		<guid isPermaLink="false">https://prologika.com/?p=8817</guid>

					<description><![CDATA[There has been a lot of noise surrounding a data lakehouse nowadays, so I felt the urge to chime in. In fact, the famous guy in cube, Patrick LeBlanc, gave [&#8230;]]]></description>
										<content:encoded><![CDATA[<p><img decoding="async" style="padding: 0px 10px;" src="https://prologika.com/wp-content/uploads/2023/03/lakehouse2.png" alt="" height="141" align="left" />There has been a lot of noise surrounding a data lakehouse nowadays, so I felt the urge to chime in. In fact, the famous guy in cube, Patrick LeBlanc, gave a great presentation on this subject to our <a href="https://www.meetup.com/atlanta-microsoft-business-intelligence-users/events/290184694/">Atlanta Power BI Group</a> and you can find the recording <a href="https://youtu.be/UVpIDE_NeIM">here</a> (I have to admit we could have done better job with the recording quality, but we are still learning in the post-COVID era).</p>
<h4><strong>What&#8217;s Data Lakehouse<br />
</strong></h4>
<p>According to Databricks which are credited with this term, a data <a href="https://www.databricks.com/glossary/data-lakehouse">lakehouse</a> is “a new, open data management architecture that combines the flexibility, cost-efficiency, and scale of data lakes with the data management and ACID transactions of data warehouses, enabling business intelligence (BI) and machine learning (ML) on all data.” It other words, it’s a hybrid between a relational data warehouse and a data lake. Sounds great, right? Visualizing this in Microsoft parlor, the last incarnation of the lakehouse architecture that I came across looks like this:</p>
<p><img decoding="async" class="lazy-loaded" src="https://prologika.com/wp-content/uploads/2023/02/020823_2159_DataLakehou1.png" alt="" data-lazy-type="image" data-src="https://prologika.com/wp-content/uploads/2023/02/020823_2159_DataLakehou1.png" /></p>
<h3>The Good</h3>
<p>I’m sure that many large companies or companies with complex data integration needs could benefit from a similar architecture. As I said many times, staging data to a lake is a good thing when you must deal with files. For example, some cloud vendor that hasn’t matured enough to give direct access to your data, could decide to push files instead (I described a similar scenario in this <a href="https://prologika.com/uploading-files-to-adls-gen2-with-python-and-service-principal-authentication/">blog</a>). A “network share” on steroids, the data lake is the best place to store files. A good question here and the one I personally struggled with would be “what if the data comes from relational databases or from REST APIs?” Should you stage that data in a data lake as files before it flows into the data warehouse? A wise consultant’s answer here would be “it depends”. Here are some good reasons when this might make sense.</p>
<ol>
<li><strong>Stage data first </strong>– For some, a large ISV company (see related newsletter <a href="https://prologika.com/prologika-newsletter-fall-2021/">here</a>), had to integrate data from many databases with similar but not the same schema. They preferred to stage the data to a data lake and figure out the integration “mess” caused my schema discrepancies and data quality later.</li>
<li><strong>A glorified archive </strong>– For example, in case you want to reload the data, you can do it from the lake in the case where the source systems truncate data. However, my personal preference to address this scenario would be to stage the data into a relational Operational Data Store (ODS), especially in the case where changes must be tracked. In a nutshell, if I’m given a choice between a file or relational database, I’d go with the latter.</li>
<li><strong>Synapse </strong>– If you decide to host your data warehouse in a Synapse dedicated SQL pool and use Azure Data Factory (ADF) to load the data, ADF will stage the data to Azure Data Lake Service (ADLS) anyway to load it faster into Synapse. Another good thing for Synapse here is that you can use Synapse Serverless to query that data using SQL which might come handy (I share some “serverless” lessons learned <a href="https://prologika.com/serverless-lessons-learned/">here</a>).</li>
<li><strong>Data science </strong>– There are some good reasons why data scientists prefer files instead of loading the data from a relational database. Or so I was told (I’m not a data scientist).</li>
<li><strong>Uniformity </strong>– If your organization prefers a uniform data flow path despite the additional effort, inconvenience, and redundancy, then this might make sense. Then despite the source data type (structured or unstructured), all data follows the same ingestion pipeline. Just make sure to hire more ETL developers.</li>
</ol>
<blockquote><p>Outside these considerations, when you can connect directly to the data source, staging data to files is probably overkill as files are notoriously difficult to deal with.</p></blockquote>
<h3>The Bad</h3>
<p>Now let’s look at the so-called zones in the lake: raw, enriched and curated, sometimes also referenced as bronze, silver, and gold. The idea here is to enrich the staged data. So, the raw zone has the staged data 1:1 as in the source. Then let’s say a data scientist needs some enrichment, and we spin more ETL to add a bunch of columns to some file. And then Business needs to reference the data that might require more enrichment. So, into the ETL rabbit hole we go again.</p>
<p>The problem is that many people take this architecture verbatum, whether it makes sense or not. A question came from the audience during Patrick’s presentation “What data do we add to these zones?” How do we know when it’s time to move to the next zone? And the answer here is that these zones are just a recommendation that someone has come up with. A large organization might benefit from them. But in most cases in my opinion spinning more and more ETL and moving data around just so that you follow some vendor’s best practices, makes no sense. And should you stage the data 1:1 from the source? In some cases, like the Get Data First aforementioned scenario, it might make sense. But in most cases, it would be much more efficient to stage the data in the shape you need it, which may necessitate joining multiple tables at the source (by the way, a relational server is the best place to handle joins).</p>
<p>The omni-presence of Synapse in such architectural diagrams is questionable at least. As I stated in another <a href="https://prologika.com/prologika-newsletter-winter-2022/">newsletter</a>, like a red giant star, Synapse seems to engulf everything in its path in order to increase its value potential. But Synapse shouldn’t be a default choice for most organizations. It’s rather expensive and has limitations, such as lacking important T-SQL features.</p>
<p>Finally, Spark/Databricks that orchestrates the data preparation with Python or some other custom code since all the toolset you get is a notebook with a blinking cursor. What happened to low code, no code approach? More ETL developers to the rescue…</p>
<h3>The Ugly</h3>
<p>The omnipresence of the <strong>delta lake</strong> regardless if it makes sense or not. I’m sure that some scenarios for staging changing data into a lake, such as IoT streaming, will benefit greatly from a delta lake. But it shouldn’t be a default recommendation. The moment we introduce a delta lake, our tool choice becomes rather restricted because of the file format. On ETL side of things, for example, you must use data flows with Azure Data Factory (I’d personally favor ELT over data flows). And to read the data, you must provision either a Spark cluster or Synapse Serverless. So, complexity increases together with cost while data accessibility decreases.</p>
<blockquote><p>UPDATE 05/28/2023 Microsoft Fabric embraced the delta format as the its native storage but provides more options, including Power Query dataflows and Azure Data Factory copy activity, to load data. All Fabric services save and read data from the delta lake and you don&#8217;t have to provision anything.</p></blockquote>
<p>And if you go with Databricks (credited for inventing the delta lake too), they are far more ambitious . They want to replace RDBMs for OLAP (OLTP won’t work with a delta lake for performance reasons). We’ve seen similar claims before and how they ended. Another question came from the audience during the presentation was if a lakehouse can deliver the same performance as a relational database. One house must be redundant, right? True, after rewriting their software, Databricks can deliver some decent performance (they even <a href="https://www.bing.com/aclk?ld=e8sEMEV61UblyUrZrSPjK7IjVUCUwzjq3RWvdDzuaT-l5a2-IerxQTe-Di_z1Mfczg5fNKJSwOe3p5vzACle-WdmKeEP2asXTv248bVnkRIXmz3NGsGJbYGA13FaccZIy6807LTyLzGzSSZFBzqNZPu4Jhhb2HeN_4k3alT-XOkoSXYVxK&amp;u=aHR0cHMlM2ElMmYlMmZ3d3cuZGF0YWJyaWNrcy5jb20lMmZyZXNvdXJjZXMlMmZ3ZWJpbmFyJTJmbGVhcm4tZGF0YWJyaWNrcy1zcWwtZnJvbS10aGUtZXhwZXJ0cyUzZnV0bV9tZWRpdW0lM2RwYWlkJTJic2VhcmNoJTI2dXRtX3NvdXJjZSUzZGJpbmclMjZ1dG1fY2FtcGFpZ24lM2Q0Mjk1NDIwNjAlMjZ1dG1fYWRncm91cCUzZDEzMTYxMTY4ODIyOTYzMzklMjZ1dG1fY29udGVudCUzZG9kJTJid2ViaW5hciUyNnV0bV9vZmZlciUzZGxlYXJuLWRhdGFicmlja3Mtc3FsLWZyb20tdGhlLWV4cGVydHMlMjZ1dG1fYWQlM2QlMjZ1dG1fdGVybSUzZGRhdGFicmlja3MlMjUyMHNxbCUyNm1zY2xraWQlM2QwNWIwZmJhYjE1NjUxM2ZmODliNDk4NDFhOTk4MDg2Nw&amp;rlid=05b0fbab156513ff89b49841a9980867&amp;ntb=1">claim</a> to be the world’s fastest “data warehouse” although only one other vendor submitted results to that specific benchmark). James Serra (Data &amp; AI Solution Architect at Microsoft), whose excellent <a href="https://www.jamesserra.com/">blog</a> discusses these topics in detail, recently gave our group a presentation and said that anyone he knows of that has tried replacing a relational data warehouse with a data lake, has failed. Enough said.</p>
<blockquote><p>What’s a best practice? A best practice to me is adopting the most efficient way to achieve something without sacrificing too much flexibility for what might be thrown at you in the future. To me, a lakehouse as a replacement for a relational data warehouse or as a default staging area is as big of a hype as Big Data was, with all the vendor propaganda surrounding it to buy stuff you don’t need. Large organizations with complex integration needs might benefit from the lakehouse architecture shown above. However, most companies could save a lot of implementation, maintenance, and licensing costs by simplifying it and judicially introducing pieces when it makes sense.</p></blockquote>
<p><img decoding="async" src="https://prologika.com/wp-content/uploads/2017/06/060417_1725_PrologikaNe2.png" alt="" /><br />
Teo Lachev<br />
Prologika, LLC | Making Sense of Data<br />
<a href="https://prologika.com/wp-content/uploads/2016/01/logo.png" rel="attachment wp-att-12"><img loading="lazy" decoding="async" class="alignnone size-full wp-image-12" src="https://prologika.com/wp-content/uploads/2016/01/logo.png" alt="logo" width="165" height="45" /></a></p>
]]></content:encoded>
					
					<wfw:commentRss>https://prologika.com/prologika-newsletter-spring-2023/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Atlanta MS BI and Power BI Group Meeting on March 6th (The Semantic Lakehouse: Power BI and Databricks)</title>
		<link>https://prologika.com/atlanta-ms-bi-and-power-bi-group-meeting-202303/</link>
					<comments>https://prologika.com/atlanta-ms-bi-and-power-bi-group-meeting-202303/#respond</comments>
		
		<dc:creator><![CDATA[Prologika - Teo Lachev]]></dc:creator>
		<pubDate>Fri, 03 Mar 2023 20:37:02 +0000</pubDate>
				<category><![CDATA[Blog]]></category>
		<category><![CDATA[Events]]></category>
		<category><![CDATA[Atlanta.MBI]]></category>
		<category><![CDATA[Azure Data Lake]]></category>
		<category><![CDATA[Lakehouse]]></category>
		<category><![CDATA[Power BI]]></category>
		<category><![CDATA[Synapse]]></category>
		<guid isPermaLink="false">https://prologika.com/?p=8814</guid>

					<description><![CDATA[Please join us for the next meeting on Monday, March 6th, at 6:30 PM ET.  Leo Furlong (Senior Solutions Architect at Databricks) will share their point of view on why [&#8230;]]]></description>
										<content:encoded><![CDATA[<p>Please join us for the next meeting on Monday, March 6th, at 6:30 PM ET.  Leo Furlong (Senior Solutions Architect at Databricks) will share their point of view on why “the best data warehouse is a lakehouse.” For more details and sign up, visit our <a href="https://www.meetup.com/Atlanta-Microsoft-Business-Intelligence-Users/">group page</a>.</p>
<p>PLEASE NOTE THAT OUR IN-PERSON MEETING LOCATION HAS CHANGED! WE STRONGLY ENCOURAGE YOU TO ATTEND THE EVENT IN PERSON FOR BEST EXPERIENCE. ALTERNATIVELY, YOU CAN JOIN OUR MEETINGS ONLINE VIA MS TEAMS. WHEN POSSIBLE, WE WILL RECORD THE MEETINGS AND MAKE RECORDINGS AVAILABLE AT <a href="HTTPS://BIT.LY/ATLANTABIRECS">HTTPS://BIT.LY/ATLANTABIRECS</a>. PLEASE <strong>RSVP</strong> ONLY IF COMING TO OUR IN-PERSON MEETING.</p>
<p><strong>Presentation:</strong> The Semantic Lakehouse: Power BI and Databricks</p>
<p><strong>Date: </strong>March 6th</p>
<p><strong>Time: </strong>18:30 – 20:30 ET</p>
<p><strong>Place: </strong>Onsite and online</p>
<p><strong>Level</strong>: Intermediate</p>
<p><strong>Food</strong>: Food and drinks will be available for this meeting</p>
<p>&nbsp;</p>
<p><strong>Agenda: </strong></p>
<p>18:15-18:30 Registration and networking</p>
<p>18:30-19:00 Organizer and sponsor time (events, Power BI latest, sponsor marketing)</p>
<p>19:00-20:15 Main presentation</p>
<p>20:15-20:30 Q&amp;A</p>
<p><strong> </strong></p>
<p><strong>ONSITE (RECOMMENDED)</strong></p>
<p>Improving Office</p>
<p>11675 Rainwater Dr.</p>
<p>Suite #100</p>
<p>Alpharetta, GA 30009</p>
<p>&nbsp;</p>
<p><strong>ONLINE</strong></p>
<p><a href="https://teams.microsoft.com/l/meetup-join/19%3ameeting_ZDYzMjVmMzAtZmZmYS00NzJhLThjYmUtYWJlZGRiYzgxNGVk%40thread.v2/0?context=%7b%22Tid%22%3a%22e7b81d0a-a949-4103-83dc-feff6277c109%22%2c%22Oid%22%3a%228b6a39f9-03a0-4118-b057-cfd26416a35c%22%7d">Click here to join the meeting</a></p>
<p><strong>Overview:</strong> The team from Databricks will come and share their point of view on why “the best data warehouse is a lakehouse.” We’ll go over lakehouse 101, when you might (or might not!) need a lakehouse, some best practices for operating a BI solution with Databricks, and walk through a demo highlighting how PowerBI’s and Databricks’ SQL capabilities complement each other.</p>
<p><strong>Speaker: </strong>Leo Furlong, Senior Solutions Architect at Databricks Leo is a seasoned data and analytics professional with 15 years of consulting experience building Data Warehousing and BI solutions using SQL Server, Power BI, and Azure technologies prior to joining Databricks in 2021. As an Atlanta native, Leo is a Georgia Tech and Georgia State grad and lives in the Smyrna/Vinings area with his 4 kids and 4 dogs.</p>
<p><strong>Sponsor: </strong>Databricks</p>
<p><strong>Prototypes with Pizza: </strong>Power BI Latest with Teo Lachev</p>
<p><a href="https://prologika.com/wp-content/uploads/2019/10/PowerBILogo.png" rel="attachment wp-att-6368"><img loading="lazy" decoding="async" class="alignnone size-full wp-image-6368" src="https://prologika.com/wp-content/uploads/2019/10/PowerBILogo.png" alt="PowerBILogo" width="410" height="109" srcset="https://prologika.com/wp-content/uploads/2019/10/PowerBILogo.png 410w, https://prologika.com/wp-content/uploads/2019/10/PowerBILogo-300x80.png 300w" sizes="auto, (max-width: 410px) 100vw, 410px" /></a></p>
]]></content:encoded>
					
					<wfw:commentRss>https://prologika.com/atlanta-ms-bi-and-power-bi-group-meeting-202303/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
	</channel>
</rss>
