Delta Live Tables || change data capture (CDC) in DLT || SCD1 and SCD 2 || Apply Changes DLT

Sdílet
Vložit
  • čas přidán 6. 09. 2024
  • Delta Live Tables (DLT) Introduction
    Introduction to Lakehouse Architecture
    Challenges with Lakehouse Architecture
    Procedural ETL vs Declarative ETL
    DLT is Declarative ETL
    Features present in DLT
    #DLT
    #StreamingTable
    #MeterializeView
    #views
    #lineage2
    #pipeline
    #DeclarativeFramework
    #ELTFramework
    #ETL
    #databrickstesting
    #dataengineers
    #dataengineering
    #Databricks
    #StreamingETL
    #BatchETL
    #DataQuality
    #DataIntegration
    #MergeExpectations
    #ELT
    #DataProcessing
    #BigData
    #DataManagement
    #SmartContracts
    #DataGovernance
    #DataAnalytics
    #DataScience
    #DataEngineering
    #ETLProcess
    #TechInnovation

Komentáře • 10

  • @nandakishormutalikdesai3408
    @nandakishormutalikdesai3408 Před 4 měsíci +1

    Crisp and Clear explanation, Thank you.

  • @joseduarte5663
    @joseduarte5663 Před 5 měsíci +1

    Hey man, awesome video and I loved how you got straight to the point, thanks for uploading this!

  • @saharamanson1970
    @saharamanson1970 Před 5 měsíci +1

    thank you

  • @letsunderstand3824
    @letsunderstand3824 Před 2 měsíci

    How to track lineage as after Apply changes drops the Lineage

  • @user-nv9fv2up5d
    @user-nv9fv2up5d Před 5 měsíci

    Quick Question : If a record is dropped from Source table i.e hard delete how does apply_changes handle it .

    • @softwaredevelopmentenginee5650
      @softwaredevelopmentenginee5650  Před 5 měsíci +1

      we should not drop any record from the source (or bronze layer), ideally we should do deduplication at silver layer . so that if we need source data again we can process it from bronze.

  • @girishrao5816
    @girishrao5816 Před 4 měsíci

    If we need to implement any transformation on silver table how to do it

  • @ArpitJain-pm3cy
    @ArpitJain-pm3cy Před 5 měsíci

    Hi , I was implementing SCD1 SCD2 in DLT . I have batch data in source and daily new file will come from source.
    After day1 success run , I modified day 2 file , so that some new records will insert and some will update.
    But i am getting below error :-
    Flow '' has FAILED fatally. An error occurred because we detected an update or delete to one or more rows in the source table. Streaming tables may only use append-only streaming sources. If you expect to delete or update rows to the source table in the future, please convert table to a live table instead of a streaming live table. To resolve this issue, perform a Full Refresh to table . A Full Refresh will attempt to clear all data from table and then load all data from the streaming source.

    • @softwaredevelopmentenginee5650
      @softwaredevelopmentenginee5650  Před 5 měsíci

      few questions ->
      - if you have a batch data why are you using DLT
      - even if you are using may i know which function u are using for reading the file like read or readStream or autoloader ?
      - if you are using read then you should work with some kind of folder strucure based on date since you are getting 1 file each day
      -if you are using readStream make sure you deleted the existing the check points and then restart your pipeline
      - if you are using the autoloader i guess you should not face any issue just try fresh load