Is it possible to export or download Datadog dashboards via Datadog REST API?
Export and update of Datadog Monitors works fine. I need the same functionality for dashboards.
UPDATED ANSWER:
Still yes.
Docs for new Dashboard endpoint here.
ORIGINAL ANSWER:
Yes.
Docs for screenboards here.
Docs for timeboards here.
Related
Is there any way to submit a PyFlink job to a cluster using Rest API?
I check out this link https://nightlies.apache.org/flink/flink-docs-release-1.14/docs/ops/rest_api/ but did not find any API related to python files.
Thanks!
As far as I know, this is not possible at the moment. I also don't see a feature request for this in the Flink Jira tickets, but you can create one of course.
I've setup a React Amplify project. I have successfully got Auth working using Cognito User Pools but can't seem to figure out DataStore/API.
I currently use DataStore to save data locally but it doesn't seem to update in the backend. I do have the aws_appsync_graphqlEndpoint in the aws-exports.js.
Not sure how to enable the Sync Engine in this guide.
You must import Amplify from #aws-amplify/core and NOT from aws-amplify
The following code example worked well in App.js, at the BOTTOM of all your imports:
import Amplify from "#aws-amplify/core";
import { DataStore, Predicates } from "#aws-amplify/datastore";
import { Post, PostStatus } from "./models";
//Use next two lines only if syncing with the cloud
import awsconfig from "./aws-exports";
Amplify.configure(awsconfig);
Source: https://docs.amplify.aws/lib/datastore/examples/q/platform/js
You can use DataStore.query to query the data.
DataStore.save to save the record etc.
Below example should be good starting point.
https://github.com/dabit3/amplify-datastore-example
If I understand the question correctly everything is set up with DataStore and working locally. The problem is that it doesn't sync to the cloud. One of the answers suggested using Apollo. You don't need Apollo. By default, and the suggested development mode of using DataStore, is local only. When you are ready to deploy to the cloud you use the command
amplify push
which provisions all of your cloud resources for the app. For more details refer to DataStore docs.
We want to somehow consume a Chatbot Analytics so we can create our Own Analytics site for our Clients.
Is that any Possible?
Are there Any tools that will help?
We don't want tools, we want to consume their Data and Present them in our own site on behalf of our Clients.
Conside creating Chatbot via Chatfuel, API.ai or something.
You can use chatbotproxy.com API to fetch app and page specific metrics.
Currently, it collects 10 metrics, Note: if there is no data; then API does not return 0, it skips keys with 0 count. ChatbotProxy Metrics
We don't want tools, we want to consume their Data and Present them in our own site on behalf of our Clients.
By assuming you are referring consumer response as Data, yes that is possible.
To gather that data you should use AWS cloud services; you can use AWS Lex and AWS Lambda to build chatbot. In AWS ecosystem,you build the skeleton with Lex and provide functionality using Lambda function which will be triggered on catching an intent.
Considering you want to do some custom analysis on your consumer's responses AWS provides the best solution. AWS implementations are more flexible, transparent and their SDKs are available for a diverse set of platforms.
If your bot is created using api.ai, unfortunately, there's no way to consume analytics data via API calls, instead, they have developed & announced Analytics dashboard in api.ai console. Here, you can review statistics relevant to the specific agent. The solution to your problem can be logging everything via webhooks and write your own analytics service, but you'd probably know that already.
I need to see in the query explorer (google analytics) a query for use in dashboard with active users rt:activeUsers for use in dashboard klipfolio and
I do not know how because only see ga: not rt:; any idea is appreciated
tks a lot
R.
RT is only available in Google Analytics Realtime API, it's using a different API than what Klipfolio natively plugs into.
You need an intermediate stage, where you fetch the API token and then query the Google Analytics RT API.
We have a large dataset from an appengine app in our datastore. Now I want to do some ETL on them to push them to bigquery, and I thought of using a Dataflow batch job.
All examples I find are using this class to query the Datastore:
import com.google.api.services.datastore.DatastoreV1.Query;
And that does work. However, I'm not familiar wit this DatastoreV1 API and would like to use the API provided with the appengine SDK, like this:
import com.google.appengine.api.datastore.Query;
The problem is that the DatastoreIO doesn't accept these queries:
PCollection<Entity> projects = p.apply(Read.from(DatastoreIO.source().withQuery(q).withDataset(DATASET_ID)));
It will only take DatastoreV1.Query objects. Is there any way to use the app engine provided API's? I'm much more familiar with those calls. Better yet, if we could use Objectify, that would be awesome :)
Thanks!
This isn't possible with the current implementation of the API. We can look at adding as a feature, and would gladly accept a pull request to expand the current functionality. The AppEngine team is also actively working on increasing interoperability between their SDK and the Datastore API.