Exporting Monitoring Metrics Explorer results to CSV - google-cloud-monitoring

I am looking for a way to export the results from "Monitoring Metrics Explorer" CSV. I am using MQL to fetch required data. But I could not find options to export the data into a CSV file.
I tried visible options available in the Metrics Explorer interface. But no luck.
My expectation is that "Export to CSV" kind of option should have been there in the attached image.
enter image description here
Not sure if I am missing something here.
Regards,
Prasad.

Related

How to read *.txt files in Azure Data Factory?

I'm trying to load data from a file *.txt type to a SQL Data Base by using a Data Flow or Copy Data activity in Azure Data Factory, but I'm not being capable to do it, down below is my try:
File configuration (as you see guys, I'm using the csv option cause' is the unique way that Azure allows me to read it):
Here is the Preview Data shows:
Everything looks fine, but once I use the Data Set in a Data Flow, I get as follow:
It is possible to read a *.txt file with Azure? What I'm doing wrong?
I tried with a sample text file and was able to get the original data in the Source transformation data preview.
Please check if you have selected the correct source dataset in your source transformation. Sometimes, when the source file is changed, it still shows old projections or incorrect projections and data previews. To reset you can change the output stream name or reconnect the source file.
Below is my source dataset connection and source settings.
Source dataset: text file
Dataflow source:

How to use a Tab-Delimited UTF-16le file as source in a Microsoft Azure data Factory dataflow

I am working for a customer in the medical business (so excuse the many redactions in the screenshots). I am pretty new here so excuse any mistakes I might make please.
We are trying to fill a SQL database table with data coming from 2 different sources (CSV files). Both are delivered on a BLOB storage where we have read access.
The first flow I build to do this with azure data factory works perfectly so I just thought to clone that flow and point it to the second source. However the CSV files from the second source are TAB delimited and UTF-16le encoded. Luckily you can set these parameters when you create a dataset:
Dataset Settings
When I verify the dataset by using the "Preview Data" option, I see a nice list with data coming from the CSV file:Output from preview data So it appears to work fine !
Now I create a new dataflow and in the source I use the newly created Data source. All settings I left at default. data flow settings
Now when I open Data Preview and click refresh I get garbage and NULL outputs instead of the nice data I received when testing the data source. output from source block in dataflow In my first dataflow i created this does produce the expected data from the csv file but somehow the data is now scrambled ?
Could someone please help me with what I am missing or doing wrong here ?
Tried to repro and here you could see if you have the Dataset settings,
Encoding as UTF-8 instead of UTF-16 then you will ne able to preview the data.
Data Preview inside the Dataflow:
And if even I try to have the UTF-16LE enabled for the encoding having such issues:
Hence, for now you could change the Encoding and use the pipeline.

using the parameters in Alteryx workflows

I am fairly new to the Alteryx.
I would like to create a process/workflow in Alteryx to import a file from a specified location but that should be controlled by the input parameters.
Kindly help me in this.
Thanks,
RTJ
You can connect an Interface Tool:File Browse via an Interface Tool:Action to change the file selected to an In/Out:Input Data:
You will then want to run the workflow using the Run As Analytic App:
In the developer tool category, you'll find the Dynamic Input tool. This works much like the standard Input Data tool, but can take records in to modify the data it collects.
https://help.alteryx.com/2018.2/DynamicInput.htm
It sounds like you have files in a standard location, but want to be able to dynamically select the ones to load.
Let's say you have a collection of sales files in the format "Sales_20190718.csv" but want to only get sales information for certain dates as specified in your workflow. You can point your Dynamic Input tool to the Sales_20190718.csv, and have it replace the "20190718" part with whatever input you gave to the tool before querying the information.
You could get a similar result by using wildcards in a basic Input Data tool by pulling data from "Sales_*", and ticking the "Output File Name as Field" box. This would load all your sales data (which could take some time) but then you could filter to the relevant files using the new FileName field.

Export data from JIRA into Excel dynamically

I am using JIRA and in each sprint i do some steps to export file from JIRA and copy-paste it in Excel, and then apply my functions on it. This takes me a lot of time so i would like to automate all this steps using Excel.
So, is it possible to export data from JIRA into Excel dynamically ?
Is there a way to do that please?
Thanks
You can export search results directly to Excel: navigate to Issues -> Search for Issues, and run you search. After that you can click on Export -> Excel (All/Current fields).
If this option is not available for you, you can export same search results to CSV file and Excel should be able to import it.
Please, refer to official documentation for more details.
If that is not what you are searching for, please specify your requirements more precisely.
See related answer here. You can export a filter to a csv file with a curl command. You can call this command via Shell in VBA. Then postprocess the output file with vba to integrate the data in your final report file.

How to export Azure Machine Learning output to CSV

I have created an Azure ML experiment which will give the output as predicted probability values and some charts such as bar chart, pie chart, etc. Now I am able to see the outputs in Azure ML's output page.
How can I export my Azure ML experiment results to CSV (or any other similar format)?
You can just configure that by using the modules under Data Format Conversions. Have a look here and here. Documentation is in progress, unluckily.
Once you've trained your model, publish it as a web service. Then from published service, you can Download Excel Workbook. Through this workbook, it will run your web service with the data you input into excel. Then it will show the predicted values.
You can add a module called convert to csv in your experiment.
The Run selected module.
Right click after the module is run and click on 'Download'

Resources