I am looking for an easy and maintainable solution for connection-string storage and retrieval on Pivotal Cloud Foundry (PCF) running on-premise. I have explored CredHub, ConfigServer, Vault and I am looking for anybody who may have faced this dilemma in the past. I would appreciate any input.
Technical Specifications
.NET Core 2.2
Prefer to use existing SteelToe functionality.
C#
Related
I have an SQL Server on Google Compute's VM. I would like to use Extensible Key Management. The requirement is to have everything in GCP. Ideally I would like Google Cloud Platform to be the provider of EKM. Otherwise I will have to install a third party EKM in GCP.
I looked into Google KMS but I cannot see any integration with SQL Server.
I cannot find any info about other Google services being able to manage keys and to integrate with MS SQL Server.
I am struggling to find any answers on Google.
Is there a GCP service that could do that?
UPDATE:
MS SQL Server has multiple databases. I would like to have each database encrypted with a different key and allow for a regular change of keys.
https://cloud.google.com/compute/docs/disks/customer-managed-encryption may meet your needs. This page describes how to use CMEK keys from KMS on a Compute VM and which parts of the VM are protected.
To simulate a device I want to use a Logic App to fill a Azure Database for PostgreSQL with test data.
I found the two connectors PostgreSQL and SQL server but unfortunately both don't work for me. The PostgreSQL connector does only support onpremise databases and neither the autodetection nor the manually input of the connectionstring works with SQL Server.
Would great if somebody could give me a hint how to write data into a Azure Database for PostgreSQL from a Logic app.
Unfortunately, there is no connector available for this as of today.
The current alternatives would be
Use Azure Functions (which can be called from Logic Apps if required)
Create Custom APIs and a Custom Logic App Connector
Also, there is a feature request on UserVoice that you may upvote to gain traction.
I am building a backend for an application with Google App Engine and Cloud SQL.
I do have:
A webserver as a proxy in front of my API server which handles sessions (using Cloud SQL and memcache) and calls the API
An API server which has access to the resource in the Cloud SQL instance
oAuth server which also needs Cloud SQL and memcache for tokens etc.
So my questions: Do I need three Cloud SQL project, which all have their own replica? Or is it ok to have one Cloud SQL project and all three App Engine projects access this Cloud SQL instance through the Cloud SQL proxy?
All projects will be located in the us-central region.
Would love to hear some thoughts.
Thanks!
I’m adding this information as a formal answer for the community. All credit goes to Dan Cornilescu.
You do not need to create 3 different projects. You can have 3 Google App Engine services running and a single Cloud SQL instance in the same project. That seems to be the best option for your situation. Using multiple services within a single project has its advantages one of them being increased performance.
Note that you could also have multiple Cloud SQL instances running in the same project. You can follow this document that talks more about creating a Cloud SQL instance:
Creating Instances
In case you need more information about Google App Engine services, this is a good resource:
Microservices Architecture on Google App Engine
I would like to use the application identifier access control method for allowing my app engine service connect to a Cloud SQL database. The app engine service is written using dart with sqljocky.
In the Developers Console, Storage, Cloud SQL, db instance, Overview tab, there is a link titled "How to connect to your Cloud SQL instance". There, I only see sample code for doing this in PHP, Java and Python. I've tried to find an analogous implementation for dart with no luck.
Can anyone suggest an approach for doing this?
A client of mine has 2 SAP local servers and he would like to recover his data on the cloud using GCP.
The 2 SAP servers are working with a shared MS SQL database with clustering methodology, one server is active while the other is passive. His reason of recovering the data is that whenever his SAP servers are down he could still access his database with the same functionality he has locally.
I kept on trying to search for a solution to do so using the Google App Engine but as far as I understood there will be no full functionality.
What I thought of is to use the CloudSQL database as a mirror to his SQL database and to get the GAE as a passive server that only works when the systems are down. I found this link which was useful on connecting SAP with GAE but still not with much help on running GAE as an active server.
Any ideas on any other work around?
Thank you.