Micro partitions in Snowflake

Sdílet
Vložit
  • čas přidán 6. 08. 2024
  • All data in Snowflake tables is automatically divided into micro-partitions with each partition containing between 50 MB and 500 MB of uncompressed data.
    Each micro-partition maps to a group of rows and is organized in a columnar fashion. This size and structure allows for both optimization and efficiency in query processing. Micro-partitions allow users to perform extremely efficient DML and fine-grained pruning on very large tables, which can be composed of millions, or even hundreds of millions, of micro-partitions.
    In simpler words, if a query specifies a filter predicate on a range of values that accesses 30% of the values in the range, it should ideally only scan the portion of the micro-partitions that contain the data, in this case, 30% of the values. Micro-partitions are derived automatically as data is ingested in Snowflake; they don’t need to be explicitly defined up-front or maintained by users.
    Unlike tables in traditional data warehouses that typically have a relatively small number of partitions (for example, 1 partition per day and per product), a table in Snowflake can easily have millions of partitions.

Komentáře • 3

  • @alexanderheath8973
    @alexanderheath8973 Před 20 dny

    Immediately understand what is going on. Great explanation. Cheers fella

  • @rajrajabhathor2996
    @rajrajabhathor2996 Před měsícem

    Totally agree ! Very clear and simple explanations ! - all of this Gentleman's Video series...

  • @danielejiofor3493
    @danielejiofor3493 Před rokem

    Best Explanation I have seen!!!!