how to configure the Penthao 8.2 bulk insert with snowflake? - snowflake-cloud-data-platform

how to configure the Penthao 8.2 PDI bulk insert with snowflake?
Or is there any plug in available to bulk insert?

I'd recommend upgrading to Pentaho 8.3, as they have released a lot of new capabilities with their connection to Snowflake. Read here:
https://help.pentaho.com/Documentation/8.3/What%27s_new_in_Pentaho_8.3
More details about bulk loading features here:
https://help.pentaho.com/Documentation/8.3/Products/Bulk_load_into_Snowflake#GUID-C28ACDBB-F9B7-4866-BF52-C29A0415BA56

Related

Symmetricds support spatialite database?

How can I configure the symmetric db engine to replicate a SpatiaLite database ?
If there are no way what's the steps to add a new platform to symmetric db?
No, there’s no support
Fork the source code from developer’s repository, find a support, try to replicate it and after testing submit a patch back
Have tried using the SQL Lite dialect? It appears the SpatiaLite database is just an extension to SQL Lite for spatial data. SymmetricDS works with Oracle and Postgres spatial data out of the box. So treating it as a SQL Lite node may be sufficient. If there are additional libraries that need added you can put them in the lib folder of SymmetricDS and restart.

HSQLDB Junit tes failing for DB2 and SQL Server NOLOCK queries

for our sprinboot application for junit we have used HSQL and actual database that we are using is DB2 or MS SQL Server, unfortunately in the DB2 queries we are supposed to use NOLOCK which is failing my dao related junits as HSQLDB does not support nolock.
My HSQLDB version is 2.3.3 and I guess there is some provision for ignoring no lock if so how to go about it .Or is there any work around like using some #primary bean where I can put my queries so that junit reads no lock related queries.
I am using sprinbootest runner for my junits. Aplogies as I am unable to paste my code.
HSQLDB supports the DB2 syntax for no lock. You need to use the latest version 2.4.1. The NOLOCK keyword is for SQL Server. DB2 uses UR
SELECT * FROM atable WITH UR

How to load data from oracle and sql server to HAWQ using Spring XD

Hi I have tables in Oracle and SQL Server. I need to load data from oracle and sql server into Pivotal HAWQ using Spring XD. Couldn't find in documentation.
You need to integrate sqoop jobs with Spring XD. See link below for sqoop jobs with springxd
https://github.com/tzolov/spring-xd-sqoop-job
You can use jdbchdfs job to load the data in HDFS as CSV or any PXF supported format. Then you can map the loaded data to HAWQ tables using PXF External tables support. If you need to load this data to Native HAWQ tables then you can do a SELECT INSERT from there, or have SELECT INSERT configured as another batch job that loads the data from PXF External table to HAWQ native.
Outsourcer is another open source solution that was originally designed to load data from Oracle and SQL Server into Greenplum but was enhanced some time ago to also support HAWQ.
All of the documentation and downloads are on http://www.pivotalguru.com/
And if you are interested in seeing the source code, here it is: https://github.com/pivotalguru/outsourcer

Redshift with SSIS/SSDT

Has anyone been successful using Amazon Redshift as a source or destination ODBC component in SQL Server Data Tools 2012?
I've installed the PostgreSQL drivers provided by Amazon and have successfully tested a connection in the Windows ODBC driver administrator but keep running into arcane error messages when I choose my saved DSN and try to pull a table listing.
Redshift is based on quite an old version of Postgres (8.0). Postgres has changed quite a bit since then and the Postgres tools have changed with it. When downloading any tools to use with Redshift you will probably need to use previous versions from several years ago.
The table listing problem is particularly annoying but I have yet to find a version of psql that can properly list Redshift tables. As an alternative you can use the INFORMATION_SCHEMA tables to find this kind of info, and in my opinion this is what SSIS/SSDT should be doing by default.
I would not expect SSIS to be able to load data into Redshift reliably, i.e. create a Redshift destination. This is because Redshift does not really support INSERT INTO as a way to load data. If you use INSERT INTO you will only be able to load ~10 rows per second. Redshift can only load data quickly from S3 or DynamoDB using the COPY command.
It's a similar story for all other ETL tools I've tried, notably the open source tools Pentaho PDI (aka Kettle) and Talend Open Studio. This is particularly annoying in Talend's case as they have Redshift components but they actually try to use INSERT INTO for loading. Even Amazon's own ETL tool Data Pipeline does not yet have support for Redshift as 'node'.
I have been successful. Try installing both the 32-bit and 64-bit versions of the PostgreSQL ODBC drivers.
Also, in your Project Properties under 'Configuration Properties' > 'Debugging', set 'Run64BitRuntime' to False.
You can also try specifying the connection string in Connection Manager. For example:
Driver={PostgreSQL ANSI};
server=redshiftdb.d113klxjd4ac.us-west-2.redshift.amazonaws.com;uid=;database=;port=5432

Unable to insert data into database

I am using SQLite.
When I try to insert data into the tables I get the following:
Mixed mode assembly is built against
version 'v2.0.50727' of the runtime
and cannot be loaded in the 4.0
runtime without additional
configuration information.
I am using ADO.NET 2.0 Provider for SQLite
Here
Here
Here
are links to previous questions that will help.

Resources