How can I use a CF-defined destination during local development? - sap-cloud-sdk

For CAP/FaaS, can the SAP Cloud SDK be used with a destination defined on SCP CF, by means of a proxy through that destination? This would allow a single destination setup to be used for both local development as eventual cloud runtime.
https://sap.github.io/cloud-sdk/docs/js/features/connectivity/destination-js-sdk/#service-instance
I would expect there to be an example of how to provide the credentials in VCAP_SERVICES so that the Cloud SDK could access the destination instance which would provide access to the destination. However, that is not twat is being described in that section.

First of all, the link is broken and here you can find some relevant documentation regarding destination service.
In general, when running your application locally, it is recommended to use a convenient function mockDestinationsEnv to set destination information. You can find more details here.
Of course, you can copy the VCAP environment variables from the CF and save it to your local VCAP_SERVICES to mock the runtime. But this is not recommended for some reasons. First of all, VCAP variables might contain lots of sensitive information from different services bound to your application which should be treated carefully. Secondly, it will not work for some authentication types (e.g., principal propagation related) which are using JWT. Therefore, you should never use this approach unless you are using e.g., BasicAuth so that no JWT is needed.

Related

How to simple edit local config file throuth API

In all servers we got some .env files, which sets configs for server (Node.JS) on start.
Now I want to edit this files from admin pane (another web-service, working with main server through API).
Is there any best practices or just good ideas how can I realize that?
First idea - create another web-server on instance, which will have only two API endpoints (read, write) and which will restart server after editing configs. This idea looking too heavy.
Second idea is to create bash script, which will send requests to admin servers to take actual configs and rewrite local .env file if find some changes, but here will be a lot unnecessary requests. (Request every minute, but configs will change 1 time per month).
What do you think? Any ideas?
You have a couple of options and it depends primarily on your deployment strategy..
If you have a distributed environment and/or your configuration changes often (i.e.: running multiple docker containers, rotating keys, etc.) I'd highly recommend using a K/V store and reading configuration(s) dynamically during application start. Check out HashiCorp Vault, etcd or even mongodb.
If your configuration contains sensitive data definitely use something like HashiCorp Vault. If you use a configuration tool like ansible, it has ansible-vault which will encrypt your secret(s) at rest and decrypt them during deployment.
I would highly advise against storing (even potentially) sensitive data such as api keys, tokens, etc. in version control. This is a pretty big attack vector and will lead you down a dark road.
Worst case scenario use environment variables. Almost all CI/CD tooling supports these and you can maintain separation of concerns.

Centralised Environment variables

I am looking for a tool which can help to maintain the environment variables and their values in a centralised location and used by many applications from the centralised location.
For example, I have a environment variable called API_KEY and the value of API_KEY in UAT environment for specific service will be same and I want the application to fetch the value of API_KEY from a tool.
Please share your thoughts if you have come across similar tools.
Thanks and Regards!!!
Ananth Francis.
Consider using ConfigMaps.
ConfigMaps bind configuration files, command-line arguments,
environment variables, port numbers, and other configuration artifacts
to your Pods' containers and system components at runtime.
Here is an official documentation describing how to use it.
ConfigMaps are useful for storing and sharing non-sensitive,
unencrypted configuration information. To use sensitive information in
your clusters, you must use Secrets.
Analogically here you can find more info about Secrets.
Please let me know if that helped.

Can Azure Functions written in Node.js access Connection Strings?

The App Settings for an Azure Function App contain values for database connection strings that can be set in the portal. In C# they can be accessed using
ConfigurationManager.ConnectionStrings["ConnectionString"].ConnectionString
For an Azure Function written in JavaScript, is there an equivalent construction that has access to the connection strings? I understand that they can be stored in the application settings, but since there is a section on the portal devoted to connection strings, I am asking if this has any application to Node.js functions.
The distinction between app settings and connection strings makes sense in .NET, but not as much in Node. When using Node, the suggestion is to use app settings for all your secrets and connection strings. You can them access them using process.env.YourAppSetting.
And to answer your question directly, there is no easy way to access connection strings in Node, unless you start making assumptions on prefixes that are not guaranteed to work forever.
try
process.env['YOUR_APP_SETTING'];
Microsoft Documentation
I ran into same issue and I was under the same impression that we cannot access connection strings in NodeJs application through Runtime Environment variable.
But after thorough checking in the Microsoft documentation, got to know that we can access the connection strings in Node JS, PHP, etc., which are of custom type as below
CUSTOMCONNSTR_<connection_string>
Ref:
https://learn.microsoft.com/en-us/azure/app-service/web-sites-configure#connection-strings

Right way to store sensitive credentials for web app

I have a Java web app running on EC2 under Tomcat (a WAR) that requires various sensitive configuration parameters - for example, the credentials associated with various other AWS services. I had been setting these as environment variables, but then discovered that running Tomcat as a service removes almost all environment variables. So currently I use a simple configuration file to store these values.
I don't believe this is a wise choice going forward, however, and would like to find an alternative. What is the right way to handle this kind of sensitive information?
IAM Roles are going to be your best friend here. The official docs here will point you in the right direction. There's also a post on the AWS security blog about it here.

GAE: best practices for storing secret keys?

Are there any non-terrible ways of storing secret keys for Google App Engine? Or, at least, less terrible than checking them into source control?
In the meantime, Google added a Key Management Service: https://cloud.google.com/kms/
You could use it to encrypt your secrets before storing them in a database, or store them in source control encrypted. Only people with both 'decrypt' access to KMS and to your secrets would be able to use them.
The fact remains that people who can deploy code will always be able to get to your secrets (assuming your GAE app needs to be able to use the secrets), but there's no way around that as far as I can think of.
Not exactly an answer:
If you keep keys in the model, anyone who can deploy can read the keys from the model, and deploy again to cover their tracks. While Google lets you download code (unless you disable this feature), I think it only keeps the latest copy of each numbered version.
If you keep keys in a not-checked-in config file and disable code downloads, then only people with the keys can successfully deploy, but nobody can read the keys without sneaking a backdoor into the deployment (potentially not that difficult).
At the end of the day, anyone who can deploy can get at the keys, so the question is whether you think the risk is minimized by storing keys in the datastore (which you might make backups of, for example) or on deployer's machines.
A viable alternative might be to combine the two: Store encrypted API keys in the datastore and put the master key in a config file. This has some potentially nice features:
Attackers need both access to a copy of the datastore and a copy of the config file (and presumably developers don't make backups of the datastore on a laptop and lose it on the train).
By specifying two keys in the config file, you can do key-rollover (so attackers need a datastore/config of similar age).
With asymmetric crypto, you can make it possible for developers to add an API key to the datastore without needing to read the others.
Of course, then you're uploading crypto to Google's servers, which may or may not count as "exporting" crypto with the usual legal issues (e.g. what if Google sets up an Asia-Pacific data centre?).
There's no easy solution here. Checking keys into the repository is bad both because it checks in irrelevant configuration details and because it potentially exposes sensitive data. I generally create a configuration model for this, with exactly one entity, and set the relevant configuration options and keys on it after the first deployment (or whenever they change).
Alternately, you can check in a sample configuration file, then exclude it from version control, and keep the actual keys locally. This requires some way to distribute the keys, though, and makes it impossible for a developer to deploy unless they have the production keys (and all to easy to accidentally deploy the sample configuration file over the live one).
Three ways I can think of:
Store it in DataStore (may be base64 encode to have one more level
of indirection)
Pass it as environment variables through command-line params during deployment.
Keep a configuration file, git-ignore it and read it from server. Here this file itself can be a .py file if you are using a python deployment, so no reading & storing of .json files.
NOTE: If you are taking the conf-file route, dont store this JSON in the static public folders !
If you are using Laravel and want to store your keys in Datastore - this package can make that easy while managing performance using caching. https://github.com/tommerrett/laravel-GAE-secret-manager
Google app engine by default create credential for app engine and inject it in side the environment.
Google Cloud client libraries use a strategy called Application Default Credentials (ADC) to find your application's credentials. When your code uses a client library, the strategy checks for your credentials in the following order:
First, ADC checks to see if the environment variable GOOGLE_APPLICATION_CREDENTIALS is set. If the variable is set, ADC uses the service account file that the variable points to.
If the environment variable isn't set, ADC uses the default service account that Compute Engine, Google Kubernetes Engine, Cloud Run, App Engine, and Cloud Functions provide, for applications that run on those services.
If ADC can't use either of the above credentials, an error occurs.
So point 2 means if you grant the permissions to your service account using IAM Admin you do not have to worry about the passing json keys it will aromatically works.
eg.
Suppose your application running in App Engine Standard and it wants the access to the Google Cloud Storage. To do this you do not have to create new service account just grant the access to the ADC.
REF https://cloud.google.com/docs/authentication/production#finding_credentials_automatically

Resources