Ingesting data from Snowflake to Azure Data Catalog - snowflake-cloud-data-platform

I have a project to ingest metadata from the Snowflake data warehouse into the Azure Data Catalog (ADC). The ADC does not natively support this, so I must use the ADC API or ODBC.
Before I proceed on implementing the API solution, I thought it wise to ask if anyone has suggestions on this. A Google search returned nothing of note.

Unfortunately ODBC will not work either. I am moving onto a solution to load via API.

The API or ODBC are the best choices.

Related

Sync API table to SQL Database

I'm exploring options on how to one-way sync from a table available via API to an SQL database. Does anyone have any suggestions on how to achieve this?
The data from the "Source" is often updated and should be copied to the "Destination" as the changes happen (live).
Source
Read Only table from an ERP available via an API. Webhooks on the source are not possible. Entries to this table may be created, updated or deleted. There would be approximately 150,000 entries in the table with about 1000 changes per day.
Destination
Azure MS SQL database which I have full control over.
I'm looking for best practice or any ideas on how to achieve this. There seems to be very few articles that I can find with anything helpful.
I'm open to using any tool on Azure including Logic Apps and Azure Functions but want to stay away from using 3rd party tools.
If you are trying to achieving this through logic apps, Below is the flow that you can follow.
Note: Make sure you preprocess the data before sending the data to SQL database using appropriate actions based on the type of data that you are receiving.

Is there an option to move data directly to Snowflake without going through S3 and other cloud storage

We choosed Snowflake as our DWH and we would like to connect different data sources like (Salesforce, Hubspot and Zendesk).
Is there a way to extract data from these sources and store them in Snowflake in a staging schema without having to store the data in cloud storage like S3 then reading the data into Snowflake?
Many thanks in advance.
You can use any of the connectors Snowflake provide (odbc, jdbc, python, etc) and any tool that can use one of these connectors. However they wont perform well compared to the COPY INTO approach that is optimised for bulk loading.
There are ETL tools, such as Matillion, that use the stage/copy into approach but do it in the background so that it appears that you are loading directly into Snowflake.

Hive performance to create Dashboard using Tableau?

We are planning to implement a project in Azure cloud where data storage will be Azure Data lake for now and in future HDP will be implemented and ADLS will be the extended datanode. From ADLS we want to expose data for Dashboard creation using Tableau. Initial plan was to use Hive and Tableau will connect to Data through Hive. But here comes the performance issue as:
There will be multiple users who will have access to Data through Tableau(100+)
We will also have to expose Data to different portal with API calls.
Which means multiple connectivity will be established at the same time which will hit hive . My question is:
Can hive serve the purpose with minimal time?
How can i measure the performance?
I dont want to let my users to sit back after running a query in tableau and wait for a long time to see the dashboard.
Would you please share your experiences in this design issue? Should we use Hive or should We use some other tools which have better performance to work with tableau and HDFS storage. Someone suggested me to use Azure SQL Server and connect Tableau to SQL server. But its again the old fashion and also matter of cost as price is related with the execution of each query.
If you have any better solution experience please share , would be greatly appreciated.
Thanks in advance.
Hive LLAP could work, if you can get it installed.
Otherwise, at my work, we've had good experience with PrestoDB and Tableau on S3 data.
Some teams use Spark SQL, and you can setup a Spark Thrift Server, that should be compatible with the Hive JDBC/ODBC drivers

MSSQL Server performance metrics in asp.net mvc

I am working on a project where i need to display the database mssql server's performance metrics for example memory consumed/free, storage free space etc. I have researched for this purpose and one thing came up was DOGSTATSD.
Datadog provides the library for .net project to get custom metrics but that was not the solution for me because the metrics works on datadog website. I have to display the all (in graph or whatever suited) data, received from MSSQL SERVER. There will be multiple servers/instances.
Is there a way to do that, our WebApp connected with multiple databases and we receive/display information.
I cannot use already available tools for the insights.
You can easily get all needed data via querying dmv and other resources inside SQL Server. Good start is here.

How do I get data from Dynamics CRM online in to my SQL server using SQL?

I can't believe that I have not yet found the answer to this question.
All I want to know is where to start in terms of using SQL tools (queries or SSIS I assume) to query CRM online (2013) and bring data into a table.
Is this even possible?
I want to download data, then I want to transform then I want to put it back. I'm sure I can work out that once I know to start!!!
Many thanks, Warren
Unfortunately, CRM Online doesn't give you direct access to the database layer via SSMS.
There are some good samples in the CRM SDK that show how to retrieve data if you are up for writing some C#. This is their intro sample for how to connect to your CRM instance, gather some data, display it, and then update the entities. http://msdn.microsoft.com/en-us/library/hh675400.aspx
Once you manage to connect to your CRM instance, you will have access to all of your CRM data objects.
What I can suggest you based on my experience is you can get Data From CRM Online to your SQL using SSIS and you can use SSIS Integration Toolkit for CRM for it.

Resources