I've created a simple query using Kusto language in Application Insights. This query has been pinned to Azure Dashboard.
I've noticed that there is no any automatic update applied for the table, moreover refresh button on the tile work only first time when you click on it.
Here is the query pinned to Azure Dashboard:
traces
| order by timestamp desc nulls last
| take 10
Based on information specified here I expected that table should be refreshed every 5 minutes. But seems that the table is never refreshed.
Is there are automatic refresh exist for AI queries on Azure Dashboard? Actually I didn't find any mentioning about automatic refresh apart provided above link.
Metrics refresh depends on time range with minimum of five minutes. Logs refresh at one minute.
Related documentation reference
Regarding the refresh button on the tile, thanks for sharing the feedback on that. We have reached out to concerned team to look further into it. Will keep you updated on it.
If you're up for it, you can create a Chrome Extension which hunts the Dashboard DOM and selectively clicks each refresh DIV. Here's the pertinent jquery code:
function doRefresh(){
$( "div.azc-toolbarButton-container" ).each(
function() {
if ($( this ).attr( "title" ).indexOf("Refresh") >= 0)
$( this ).trigger('click');
});
}
I was able to do this. Added a 1 minute setInterval() and it appears to work well.
Those are AnalyticsPart tiles,
they refresh automatically every hour.
Of course, you can initiate a refresh by clicking the refresh button.
Azure dashboard is planning to expose an automatic refresh pile that will enable the user to customize the refresh rate of the dashboard parts. One could then chose to customize the rate to 1 hour, 30 min, etc. AnalyticsPart may onboard to this new pile.
It's still in the designing stages of the azure team.
Related
PROBLEM
We want to track changes in user calendars, but are concerned with how often we'd need to check 2000+ user calendars (Outlook).
Would the process of monitoring over 2000 user calendars present a problem for our network?
WORKFLOW
Trigger (Check for calendar change) -> ACTION (Http: update a DB)
The Trigger below checks a calendar every 2 seconds. Is there a trigger that behaves like a "subscription" object where it simply handles a change notification?
For the question about how often to check the calendar events, it depends on your requirement. In my opinion, if you set check event every 2 seconds(it's a little bit more frequent), you'd better check if your logic app set as run in parallel. You can click the ... button of the trigger and click "Settings". Then check if it is same to below screenshot.
For your question about is there a trigger that behaves like a "subscription". I'm afraid it doesn't exist a trigger which can implement this requirement in logic app. We can also check if any backend api can implement it, we can see the graph api document.
The example in above screenshot is for mailFolders, but it's same with events. We can see it is necessary to specify a user(like me) or a group before the /events. So I don't think we can monitor the subscription events. You can raise a post on Azure feedback page to suggest developer add this feature.
I am creating a custom widget and having some stastics which is coming from jason data. i am able to create widget and fine. Now i want to automate this process so that as soon as jason file arrives in folder it refreshes the graph in widget. Is there any way to do it.
Is there any way to do it.
As far as I know , the Dashboards widget uses the load and reload function to load the widgets.
You can add conditions according to your needs, and then use the load or reload method in the return function to refresh the widget.
Here is an example, you could refer to it.
In addition,the dashboard also has automatic refresh function.
You could select the Automatically refresh the dashboard every 5 minutes option in Dashboards settings. Then the widgets will refresh every 5 minutes.
Hope this helps.
I am trying to extract some data from this website which refreshes every minute. I have tried researching about web scraping and tried chrome extensions but none seem to work for me.
Some background information about the website: it is a website where people go to monitor bid prices for COE (certificate of entitlement for cars in Singapore). Every alternate Wednesday, from 1430 to 1600, I would have to manually copy and paste the data into an Excel spreadsheet before it refreshes every minute.
Details for COE
I have attached screenshots to illustrate further.
This is the website to scrape; https://www.onemotoring.com.sg/1m/coe/coeDetail.html
You can get a very low cost with AWS Lambda with node-js.
Create a Lambda function and trigger it at your cron schedule you want to crawl the website. You can use library like
https://github.com/bda-research/node-crawler
to simplify crawling.
Also,
To get the exact nodes in the page use serverside jquery or any progressive script that can extract elements from the crawled page.
Once you have the details, you can store them under DynamoDB which is a nosql with very low latency.
You can use,
ODM like https://github.com/clarkie/dynogels to access DynamoDB with very less code.
Hope it helps.
I am having the same issues with powerBI ,seems the automatic refresh its failing ,i currently have to click refresh to see new data coming in ,i have configure the the tumblingwindow part e.g tumblingwindow(second,3),done the live to dashboard are the any other settings/factors i have to set for the automatic refresh to work.(its a console app that selects data from database and sends each row to event hubs from event hubs to stream analytics then output is powerBi ).i am assuming the is time restrictions depending on throughput but how do i really calculate the time for tumblingwindow i should set ,i have tried the equation entitycount*60*60/throughput = seconds still no success.
below is a code but still the events take time to reach powerBi even after tumblingwindow(second,3) ,i could stop my application running then delete the dataset from powerbi ,but then the dataset will reappear
EventData data = new EventData(Encoding.UTF8.GetBytes(serializedobjects));
eventHubClient.SendAsync(data);
Overall workflow you describe seems fine and tumbling window size should not cause the data to never show up. You will have to debug this issue, with following steps
Goto Azure portal, inputs page for the stream analytics job and get a "sample". Does it return you any samples?
Goto Monitoring page in azure portal and check if input events and output events are greater than zero. If you see input events but don't see output events, you are likely hitting an error writing to output.
Also check if operation logs has any errors for this job.
Above steps should tell you if something is wrong with event format or the output.
Are you pinning the individual live tile to the dashboard or the entire report? Pinning the entire report does not appear to work.
If you pin the single tile containing the data you want, does that refresh in real time?
I've created a document and I'm sure it is saved in the Vault too since I can fetch it, but the TrueVault dashboard shows me "0 DOCUMENTS Stored In TrueVault". Is this a bug?
The Document count takes some time to load upon refresh of the dashboard. You should see the value updated after waiting some time (less than a minute). A revamp of the dashboard is on our roadmap.