Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

See when key Fabric features will launch and what’s already live, all in one place and always up to date. Explore the new Fabric roadmap

Reply
smpa01
Super User
Super User

Gen2 df external table

[Edited by admin for unnecessary tagging without context]

A widely preferred pattern for data engineering with lakehouse for us has been creation of External Delta table. This is only possible for data sources that can be consumed from a notebook.

 

However, there are data sources that exist beyond that and the only alternative is gen2 df for them. But gen 2 df only inserts into a lakehouse table, is there any way to insert into a chosen lakehouse subfolder instead of a table.

I don’t think it is doable now. If that is the case, if it is on cards?

Did I answer your question? Mark my post as a solution!
Proud to be a Super User!
My custom visualization projects
Plotting Live Sound: Viz1
Beautiful News:Viz1, Viz2, Viz3
Visual Capitalist: Working Hrs
1 ACCEPTED SOLUTION
miguel
Community Admin
Community Admin

At the moment, Dataflow Gen2 only loads data to tables. Do feel free to suggest new destinations (and formats) in the Fabric Ideas portal (https://aka.ms/FabricIdeas)

 

An alternative is to leverage the copy activity or a copy job. Especially as the bronze layer is typically used for the files in its raw state, so no transformation should be performed at that layer and instead a simple copy activity should be good enough. If a connector is missing from the copy job / copy activity, then would you mind letting us know what the source is? you can also post a new idea for such connector in the Ideas Portal.

View solution in original post

5 REPLIES 5
miguel
Community Admin
Community Admin

At the moment, Dataflow Gen2 only loads data to tables. Do feel free to suggest new destinations (and formats) in the Fabric Ideas portal (https://aka.ms/FabricIdeas)

 

An alternative is to leverage the copy activity or a copy job. Especially as the bronze layer is typically used for the files in its raw state, so no transformation should be performed at that layer and instead a simple copy activity should be good enough. If a connector is missing from the copy job / copy activity, then would you mind letting us know what the source is? you can also post a new idea for such connector in the Ideas Portal.

frithjof_v
Super User
Super User

I'm curious, what are the benefits of writing to files instead of just appending to a lakehouse bronze delta table?

Why External Tables are Ideal for the Bronze Layer in Production Data Lakes (according to my practical experience of data engineering and servicing BI)

In a well-architected Data Lake, data flows through three layers:

  • Bronze (Raw Ingestion),
  • Silver (Cleaned & Enriched),
  • Gold (Curated Business Data with Semantic Models).

The Bronze Layer is where raw data from various sources like on-prem SQL, SharePoint, Azure SQL, Oracle, APIs, and Databricks is ingested. Using external tables for this layer is highly advantageous for the following reasons:


1. Data Persists Beyond Table Lifetime

  • External tables store data separately from the metadata, so dropping the table does not delete the data.
  • This ensures raw ingested data is always available for reprocessing or auditing.

2. Easy Table Rebuilds Without Re-ingestion

  • Since the data remains in the storage layer, you can recreate the table schema at any time without fetching the source data again.
  • This is crucial for schema adjustments or optimization without risking data loss.

3. Multiple Silver/Gold Views from the Same Data

  • External tables allow you to build multiple transformations (Silver/Gold) from the same Bronze data.
  • This eliminates redundancy and maintains a single source of truth for different business units like Finance, Procurement, Leasing, and Engineering.

4. Flexible Backfills and Schema Evolutions

  • Adding new columns, adjusting schemas, or historical backfills are seamless.
  • You can introduce new attributes for all past, present, and future data without re-ingesting or dropping the table.

5. Enhanced Audit Traceability

  • Every row can be traced back to its original source file or API batch.
  • This provides clear visibility into when and where data was ingested — critical for regulatory compliance and debugging.

Conclusion

External tables in the Bronze layer offer:

  • Data safety beyond table lifecycle
  • Rebuild flexibility without re-fetching data
  • Multi-view capability for different business requirements
  • Smooth schema evolution and backfills
  • Full audit traceability for compliance and debugging

This design pattern forms the backbone of a resilient, scalable, and auditable Data Lake architecture.

Did I answer your question? Mark my post as a solution!
Proud to be a Super User!
My custom visualization projects
Plotting Live Sound: Viz1
Beautiful News:Viz1, Viz2, Viz3
Visual Capitalist: Working Hrs
nilendraFabric
Community Champion
Community Champion

Hi @smpa01 

 

You arr right,Dataflow Gen2  currently supports writing data only to Lakehouse tables, not specific subfolders.

 

one thing which you can try is Use Dataflow Gen2 to land data in a staging table and then read this staging table in notebook and write to your desired location 

That's too much honestly to maintain. pre_bronze->brone->silver so on and so forth.

 

Dataflows have an advantage over notebooks when it comes to connecting to certain sources that don't have equivalent connectors available in notebooks — for example, on-premises SQL Server, SharePoint, etc. In such cases, there is no alternative but to use a dataflow.

 

Currently, dataflows remain relevant largely because of this limitation. So, for writing to destination, it only makes sense that df gen2 provides same options as notebook. 

 

To keep up with the norm, df gen2 must give the ability to write to subfolders. After all, any bronze should land in files for audit trailing.

Did I answer your question? Mark my post as a solution!
Proud to be a Super User!
My custom visualization projects
Plotting Live Sound: Viz1
Beautiful News:Viz1, Viz2, Viz3
Visual Capitalist: Working Hrs

Helpful resources

Announcements
May FBC25 Carousel

Fabric Monthly Update - May 2025

Check out the May 2025 Fabric update to learn about new features.

May 2025 Monthly Update

Fabric Community Update - May 2025

Find out what's new and trending in the Fabric community.