Airflow operator to download email to s3 - python-3.x

I'm working on writing a DAG that will download a file from email directly to s3. I figured there would already be operators I can use to help with this. I found the following link but I'm not sure what the parameters 'imap_conn_id' and 's3_conn_id' should be.. any help with this would be appreciated. If you know a better operator with examples, that be great to!

Related

Empty file being created by node-gpg

I'm trying to use node-gpg: https://github.com/drudge/node-gpg for encrypting/decrypting files using GPG.
snippet of code
a file is being created in the fileDecrypted location but doesn't contain anything
Does anyone know what might be wrong/have encountered a similar issue?
Any help would be appreciated
I also looked into this implementation: https://jaygould.co.uk/2019-01-21-decrypting-gpg-file-node-programatically/
and am doing the same steps but still running into the same problem

How to write Postman tests to compare a response body with an external JSON/CSV file using Newman?

It is well documented how to use an external data file with the Collection Runner in Postman to run multiple iterations. However, what I'd like to do is write tests to compare the response body from a GET request to a JSON/CSV file.
From the research I've done it seems this isn't possible using just Postman. I wonder though if it'd be possible with Newman?
I have no experience with Newman but am curious if this might work. Could you create a Postman collection with an env variable called TestJsonFile and write tests comparing values from that to the response body. Then export said collection and using Newman set the TestJsonFile value to that of a JSON/CSV file (--env-var "TestJsonFile=<path_to_file>") and run the collection with that variable?
This seems plausible to me; maybe I could create a Node project to do it all in one place. I'd just like to ask if anyone has done something similar as it seems like it should be doable.
Any critiques or suggestions would be much appreciated, thank you for reading!

Autodesk Forge: Industrial Construction Demo bucket not working

I have cloned the Industrial Construction Demo code https://github.com/petrbroz/forge-industrial-construction but I am having issues with the 'bucket-name'.
In the launch file you need to specify "FORGE_BUCKET": "" instead of the usual URN.
I have written the name of the bucket that contains my file but it does not work. I have tried debugging it but still no luck.
There is no exemplary bucket or files, so I would really like to know:
-What do I need to put in as the bucket name?
-What files need to be in the bucket?
-What file types?
-Any extra information to help me figure out why it isnt working.
Many thanks,
Poppy
My apologies, I haven't explained this in the code sample very well. The demo requires files in Forge to be named in a specific way. I have just updated the README with more details: https://github.com/petrbroz/forge-industrial-construction.

Taurus PassFail Issues when testing Jmeter.jmx script

So I'm learning a little bit about Taurus and was trying to apply some pass fail criteria to my .jmx script. When I try to evaluate a specific sampler, it seems to not run the pass fail criteria at all, but if I were to do a simple avg-rt > 10s, continue as failed , this works, but the issue is I want to evaluate each sampler specifically.
Here is a screenshot of my .yml file
I was using this link as reference to follow but I can't seem to get it to work with my script.
https://dzone.com/articles/running-your-load-tests-with-pass-fail-criteria-a
Any help and advice would be appreciated :)
Thank you!
For me it seems to run pass/fail criteria
As you like screenshots here is the screenshot of Taurus YAML file:
And here is the screenshot of "test.jmx" which basically uses simple single Dummy Sampler:
Just in case here is the link to the official documentation of the subsystem: Pass/Fail Criteria
So the issue was that spaces matter lol.
rewriting it as avg-rt of first_navigate>1s, continue as failed does the trick :)

Configuration file in Azure Functions

I need to read from a JSON configuration file in Azure Function. Is there a way to refer to the file without hard-coding any paths(assuming it's in same directory as the code).
The solution in another answer was something like:
string configuration = string.IsNullOrEmpty(configurationFile) ? "" : File.ReadAllText(Environment.GetEnvironmentVariable("HOME") + #"\site\wwwroot\functionname\filename);
Is there a better way to get this path, or read JSON configuration in general for Azure Function?
Currently, the solution you found is the recommended approach, but with the next release, which is starting to roll out today (5/30/2017), we've introduced a feature to enhance this.
You can learn more about it here:
https://github.com/Azure/azure-webjobs-sdk-script/wiki/Retrieving-information-about-the-currently-running-function
you're able to find out some explanations about this same topic on this another thread, here.
I hope that this information help you.

Resources