Azure Data Factory API call to Kimble - sql-server

I'm trying to configure the API connection from Kimble to Azure Data Factory, but I need some support in understanding what Base URL and Relative URL to add, so that I can get access to the data tables in Kimble, in order to populate an Azure SQL database instance. Has anyone tried to achieve this before and can anyone share any guidance on the implementation?
Ideally, I'd achieve as much as possible with the UI functionality opposed to heavy coding in JSON, but if needed, this can be included too.

Related

It is possible to create an API call to SQL database?

Well, probably this is a stupid question, but after to surfing the web the last week...this is my last shot.
Questions:
It is possible to create an API to call a SQL database to retrieve records?..In a secure way..?
What apps/tools/methods can I use? hopefully...VS, Python, PowerShell...etc
Can I encrypt that call?and create a method to decipher the data?
I'm not the app owner, what questions should I make to the app developer?
FYI Currently I cannot use cloud vendors or third party tools to do the transformation :(. But if there is a strong tool/vendor maybe I can negotiate it.
I have consulted:
API to database: API to Database?
How to quickly create a simple REST API for SQL Server database - https://medium.com/voobans-tech-stories/how-to-quickly-create-a-simple-rest-api-for-sql-server-database-7ddb595f751a
Any ideas?
Try rolling out redash with the bitnami stack. It's free, as in speech.
Bitnami will take care of setting up https etc if you do it on AWS. You need a t2-small I think. There's how-tos for let's encrypt etc also.
You can dish out links to "widgets" with an embedded API key. The widgets can be json or csv. They will only provide access to predefined data queries, but you can get that data from anywhere, not just your SQL server.

exporting data for analytics use in SaaS

We are a SaaS product and we would like to be able have per-user data exports that will be used with various analytical (BI) tools like Tableau or PowerBI. Instead of just managing all those exports manually, we thought of using some cloud database such as AWS Redshift (which will be part of our service). But then, it is not clear how is user will access those databases naturally, unless we do some kind of SSO integration with AWS.
So - what is the best practice for exporting data for analytics use in SaaS products?
In this case you can build your security in to your backend API layer.
First you can set up processes to load your data to Redshift, then make sure that only your backend API server/cluster has access to redshift (e.g. through a vpc with no external ip access to redshift)
Now you have your data, you can validate your user as usual through your backend service, then when a user requests a download through the backend API, the backend can create a query to extract from redshift only the correct data based upon the users security role. In order to make this possible you may need to build some kind of security column into your redshift data model.
I am assuming getting data to redshift is not a problem.
What you are looking for, if I understand correctly is a OEM solutions.
The problem is how does one mimic the security model you have in place for your SaaS offering.
That depends on how complex is your security model.
If it is as simple as just authenticate the user and he has access to all tenant data or the data can be easily filtered for user. Things are simple for you. Trusted authentication will allow you to authenticate that user and user filtering will allow you to show him all that he has access to.
But here is the kicker, if your security is really complex , then it can become really difficult to mimic it within these products.
Here for integrating tableau this link will help:-
https://tableau.github.io/embedding-playbook/#
Power BI, this product am not a fan off. I tried to embed a view in one my applications and data refresh was a big issue.
Its almost like they want you to be a azure shop for real time reporting.( I like GCP more )
If you create the api's and populate datasets then they have crazy restrictions like 1MB/sec etc.
On the other instances datasets can be refreshed only 8 times.
I gave up on them.
Very recently I got a call from Sisense and they seemed promising as well from a OEM perspective. You might was to try them.

Client-accessible noSql database?

Im building a simple angular application and there is a small administrator panel for updating the content (a .json document). I'm looking for a way to edit the json document from the administrator panel.
I can manipulate the memory-loaded json but I can't save it. Is there a way to put the json file in some kind of cloud database and connect to it without setting up a server or backend for my application?
I want my application to be easily deployable on any ftp so I can't setup a nodeserver or install something like couchdb.
Any ideas are appreciated.
You could use a provider like Parse. It's free (up to a limit of requests/month), has a nice JavaScript SDK that would get you up and running quickly. https://parse.com/
Also, check out this query builder to aid in retrieving your data from Parse. It's built as an Angular service for easy integration. https://github.com/dpollot/parse-query
EDIT
Parse also offers hosting, for free.

Backend for iOS app

My question is how do i get information from a server to my iphone app. let's assume I have completed my current project I'm working on that only needs data to be uploaded to my application.
I understand there is a database or server I must create but how do I go about creating or modifying one for my needs.
I mainly want to store login information from one user and allow users to search for people who have entered login information (name) to add to a friends lists within the current app.
i think in your case you can use Django-tastypie for backend will be good choice.since using django you can develop it in quick time and the tastypie has api services which can used easily for retrieval and sending data
you can go through this
http://django-tastypie.readthedocs.org/en/latest/
Take a look at services like Stackmob or Parse. These types of service could make it really easy for you to get the server side part of your application up and running. These services would act as your database and also provide an easy api for you to access the server side pieces.

Visualforce and VPN

I'm looking to integrate my Salesforce implementation with an external database. I know that in most circumstances I would use Visualforce with an Apex controller/extension to access the data, however the external database will require a VPN connection. Since Visualforce and any controllers or extensions are processed server-side, is there any way to do this through a VPN?
The simplest solution that occurs to me would be to expose the data via a web service and then write an Apex webservice class to call the data. I've done this several times when data was behind a firewall.
For more complex integrations, where you actually wanted to store the external data in Salesforce with more complex logic it could be useful to use an integration tool like Jitterbit or Cast Iron to do this for you.

Resources