SequenceFile format files, read in Snowflake as external table - snowflake-cloud-data-platform

Is there a method to read "SequenceFile" format in Snowflake? (https://dwgeek.com/hive-different-file-formats-text-sequence-rc-avro-orc-parquet-file.html)

Related

CSV to Snowflake - Converts some data to Uppercase Automatically

Hello Techies,
Thanks for your time in reading this and trying to help.
I have a peculiar situation where I have a CSV file with data (separated by commas) in an external stage. I have snowflake stored procedure, that accesses, reads this file, loads the data into a snowflake table.
Everything is good except, Some of the data from the file gets automatically converted into Uppercase.
My T-SQL in the procedure does not do any Transformation or Conversion of data to upper case.
The Optional parameters in the Snowflake File Format for CSV file type does not have any parameters to preserve the proper case or current case of data in the file.
Please help
Thanks in advance.
Regards,
Sathya

Loading CSV file from S3 to external stage

i am loading csv file from s3 to external stage .
some of the fields has "value" , how can i avoid " while loading csv file.
also i am getting below error :
Can't parse '"2021-03-03 16:43:31"' as timestamp with format 'AUTO'
what should be done to avoid this error?
Snowflake supports automatic detection of most common date, time, and timestamp formats; however, some formats might produce ambiguous results, which can cause Snowflake to apply an incorrect format when using AUTO for data loading.
To guarantee correct loading of data, Snowflake strongly recommends explicitly setting the file format options for data loading/unloading etc
https://docs.snowflake.com/en/sql-reference/parameters.html#label-timestamp-input-format

Parsing CEF files in snowflake

We have staged the log files in external stage s3.The staged log files are in CEF file format.How to parse CEF files from stage to move the data to snowflake?
If the files have a fixed format (i.e. there are record and field delimiters and each record has the same number of columns) then you can just treat it as a text file and create an appropriate file format.
If the file has a semi-structured format then you should be able to load it into a variant column - whether you can create multiple rows per file or only one depends in the file structure. If you can only create one record per file then you may run into issues with file size as a variant column has a maximum file size.
Once the data is in a variant column you should be able to process it to extract usable data from it. If there is a structure Snowflake can process (e.g. xml or json) then you can use the native capabilities. If there is no recognisable structure then you'd have to write your own parsing logic in a stored procedure.
Alternatively, you could try and find another tool that will convert your files to an xml/json format and then Snowflake can easily process those files.

How to convert a CSV file into bcp formatted file?

I'm into a task of importing a CSV file to SQL server table. I'm using bcp tool as my data can be large. The issue im facing with bcp is that the table where I'm gonna import CSV into can have a mix of data types like date, int, etc and if I use bcp using native mode (-n), I will need bcp file as the input but I have CSV file.
Is there any way to convert CSV file into bcp file? or
How can I import a CSV file into SQL server table given that my table columns can have any data type and not just character types?
Had it been that all columns are of character type, i would have used bcp tool with -c option.
Actually... the safest thing to do when importing data, especially when it ins bulk like this, is to import it into a staging table first. In this case where all of the fields are string/varchars. That then allows you to scrub/validate the data and make sure it's safe for consumption. Then once you've verified it, move/copy it to your production tables converting it to the proper type as you go. That's typically what I do when dealing with import data.
a CSV file is just a text file that is delimited by commas. With regard to importing text files, there is no such thing as a 'BCP' file. BCP has an option to work with native SQL data (unreadable to the human eye with a text editor), but the default is to just work with text the same as what you have in your CSV file. There is no conversion needed, with using textual data, there is no such thing as a "BCP file". It's just a ascii text file.
Whoever created the text file has already completed a conversion from their natural datatypes into text. As others have suggested, you will save yourself some pain later if you just load the textual CSV data file you have into a "load" table of all "VARCHAR" fields. Then from that load table you can manipulate the data into whatever datatypes you require in your final destination table. Better to do this than to make SQL do implied conversions by having BCP insert data directly into the final destination table.

skip first column from data field and first column from data base table using format file

I have situation to loading quoted delimiter file into table using format file and bcp
file format:
"Col1"|"Col2"
"Col1"|"Col2"
Data base table columns dLoadDate, col1, col2
Can any one provide the format file for above scenario?
It is actually easy, you can create a format file using the bcp utility (msdn.microsoft.com). The link has an easy example on how to do this. For specifying field delimeters, this link has an example using BULK INSERT on how to import the data with delimeters. HTH

Resources