Which file format is MOST performant in Snowflake for data loading?
Which file format is MOST performant in Snowflake for data loading?
Parquet is considered to be one of the most performant file formats for data loading in Snowflake. It is a columnar storage file format that is optimized for querying and analytics workloads, thus offering efficient data compression and encoding schemes. These characteristics enable faster data loading and querying compared to row-based file formats like CSV. Additionally, its structure aligns well with Snowflake's architecture, further enhancing performance.
Loading from Gzipped CSV is several times faster than loading from ORC and Parquet at an impressive 15 TB/Hour. While 5-6 TB/hour is decent if your data is originally in ORC or Parquet, don’t go out of your way to CREATE ORC or Parquet files from CSV in the hope that it will load Snowflake faster. https://community.snowflake.com/s/article/How-to-Load-Terabytes-Into-Snowflake-Speeds-Feeds-and-Techniques#:~:text=Loading%20data%20into%20Snowflake%20is,into%20fully%20structured%20Snowflake%20tables.
B. CSV
Correct