Anyone have experience integrating datadog monitoring with Snowflake? - snowflake-cloud-data-platform

Does anyone know if Datadog agent works on snowflake? We want to use Datadog to collect snowflake metrics, traces and logs and create dashboards, graphs, and monitors.

No, you cannot install the Datadog agent on a Snowflake host.
We use our separate job scheduling system to monitor Snowflake by running queries (e.g. checks on SYSTEM$CLUSTERING_DEPTH, aggregate queries against the QUERY_HISTORY for timing, etc) via the JDBC connector then relaying the results to our monitoring stack (similar to how Datadog agent would work.)

Related

Is there a way to check the DB2 sql log for actual SQL operations executed into it?, ie how many rows were fetched etc?

I am using a DB2 v10.5, and I am pushing messages into the database I created using a gateway. Is there a way to check the DB2 sql logs for actual SQL operation executed?, ie how many rows were fetched etc? While googling, I find these logs inside the DB2 server in the DIAGPATH /db2/db2inst1/sqllib/db2dump/ but I don't see any SQL messages in there.
I have been checking DB2 guides as well but any ideas to help me on this is greatly appreciated. Thank you.
Activity event monitoring
Briefly:
It acts like "logger" for executed statements. The information is written to the tables of such an event monitor for sessions with such a "logging" enabled.
There is also the package cache. This holds aggregate metrics for all executions of a statement that are still in the package cache (entries get evicted from the cache as newer statement arrive). MON_GET_PKG_CACHE_STMT
You can also use the Db2 Database Management Console which is
A new browser-based console that helps you administer, monitor, manage and optimize the performance of IBM Db2 for Linux, UNIX and Windows databases.
and which itself collects data via functions such as MON_GET_PKG_CACHE_STMT and Activity Event Monitors

How to monitor oracle 11gr2 performace metrics using java or REST API

How to monitor oracle 11gr2 performance metrics and do we have any rest api to get the perfomance metrics.
How do we monitor performance metrics in 11gr2 and 12c?
Do we have REST API to query performance metric such as cpu utilization,memory usage etc.
Can some help me to understand how to query oracle DB performance metrics in java
You can monitor the REST service by starting the JVM running your service with debug switches. When that is done you can use something like the Visual VM profiler in Netbeans 8.2 or 9 to also monitor the Queries. You will need another tool to monitor the db itself.

comparing MapReduce to cloud database services

Are the databases offered by cloud services such as Windows Azure SQL Database or AWS Big Data capable of distributed computing, in the sense that the query optimizer divides the work across servers which compute in parallel, similar to how MapReduce distributes computation across nodes?
I haven't found anything about any such query optimization in the Azure documentation, although PDW seems like it may do this.
AWS has EMR (Elastic Map-Reduce) which is Hadoop provisioned by AWS.
Azure has HDInsights which is Hortonwork's data platform (Hadoop) installed on Windows VMs.
Microsoft's PDW (parallel data warehouse) doesn't support map-reduce right as far as I know but they are working on it (http://www.zdnet.com/microsofts-polybase-mashes-up-sql-server-and-hadoop-7000007424/) - PDW is essentially a few SQL Server machines with a central management layer that allows partitioning and distribution of the data between the different nodes - it can and will break a query between the PDW nodes if the data resides on more than one but the parallelism is not map-reduce in nature.

Import data on HDFS to SQL Server or export data on HDFS to SQL Server

I had been trying to figure out on which is the best approach for porting data from HDFS to SQL Server.
Do I import data from Cloudera Hadoop using sqoop Hadoop Connector for SQL Server 2008 R2 or
Do I export data from Cloudera Hadoop using sqoop into SQL Server
I am sure that both are possible based on the bunch of links I read through
http://www.cloudera.com/blog/2011/10/apache-sqoop-overview/
http://www.microsoft.com/en-in/download/details.aspx?id=27584
But when I am looking for possible issues that could rise at level of configuration and maintenance I don't have proper answers.
I strongly feel that I should go for import, but I am not comfortable in troubleshooting and maintaining the issues that could come up every now and then.
Can someone share their thoughts on what could be the best?
Both of your options use the same method: Apache Sqoop's Export utility. Using the licensed Microsoft connector/driver jar should expectedly yield more performance for the task than using a generic connector offered by Apache Sqoop.
In terms of maintenance, there should be none once you have it working fine. So long as the version of SQL Server in use is supported by the driver jar, it should continue to work as normally expected of it.
In terms of configuration, you may initially have to manually tune to find the best -m value for parallelism of your Export MapReduce job launched by the export tool. Using a too high value would cause problems on the DB side, while using a too low value would not give you ideal performance. Some trial and error is required here to arrive at the right -m value, along with knowledge of the load periods of your DB, in order to set the parallelism right.
The Apache Sqoop (v1) doc page for users of the export tool also lists down a set of common reasons for the failure of the export job. You may want to view those here.
On the MapReduce side, you may also want to dedicate a defined scheduler pool or queue for such external-writing jobs as they may be business critical, and schedulers like FairScheduler and CapacityScheduler help define SLA guarantees on each pool or queue such that the jobs get adequate resources to run when they're launched.

How to get replication status from code

I have done replication on my SQL Server 2005 database.
Now i want to show the status from my GUI.(Can be C# or else)
Is there is any method or API by which i can monitor the status of replication.
This is for client confirmation that replication is working.
Thanks
Way something like this.
http://www.sqlmonster.com/Uwe/Forum.aspx/sql-server-replication/11698/How-to-get-the-replication-status-using-RMO-in
A great option is this: How to Programmatically Monitor Replication (via T-SQL):
http://msdn.microsoft.com/en-us/library/ms147874(SQL.90).aspx
And here's the SQL Server 2008 R2 version for anyone looking at this as well:
http://msdn.microsoft.com/en-us/library/ms147874(SQL.100).aspx
Using these special procedures can be a tiny bit of a pain from within T-SQL (as you'll commonly want to export their output into temp tables or table variables to 'filter it' an additional bit or what-not), but most devs will find this kind of interaction MUCH easier to deal with than using RMOs.
I like to use tracer tokens posted at regular intervals and then monitoring their status as they flow from publisher to distributor to subscriber. This can be done programatically with sp_posttracertoken at the publisher and checking the tracer token tables in the distributor. Hope this helps.
Monitoring Replication with Replication Monitor
Microsoft SQL Server Replication
Monitor is a graphical tool that
allows you to monitor the overall
health of a replication topology.
Replication Monitor provides detailed
information on the status and
performance of publications and
subscriptions...
How to: Start Replication Monitor (Replication Monitor)

Resources