Cognos Report on Bluemix error, Rave not supported? - cognos

I am using Reporting service to display a report I created. I have the following error:
The data and the connection shouldn't be a problem...the report has RAVE visualization engine....might be this the problem? Is it supported?
thanks

The use of reports containing RAVE charts is not currently supported. To have RAVE Viz. available, the visualizations used must be hosted on Bluemix and present within the service for common consumption. In this Beta, we don't yet have a means to host the visualizations from AnalyticZone at this time. It is something we have received feedback on and are looking into for future releases.
Thank you for your interest in using the service - this is valuable feedback!
Anthony Schofield - IBM Platform Product Manager - Business Analytics

Related

How to get Average Visualization Time of an asset in Azure Media Services v3

im currently working in a module of analysis of stadistics of videos from azure media services. I want to ask how can i get some data like average visualization time, number of visualizations and more stuff like that. im pretty sure it has to exist a very easy way to get this data but i cannot find it. I found that application insights could be useful. I have found that i may have to manually track this information. Im working on .net6. An example of code would be awesome. Thanks in advance!
pd: https://github.com/Azure-Samples/media-services-javascript-azure-media-player-application-insights-plugin/blob/master/options.md
I have found that Application Insights could be useful to my problem. Some classes like TelemetryClient (from the package Microsoft.ApplicationInsights) seems to be useful to my problem, but i cant find clear information about them.
No, there is no concept of client side analytics or viewer analytics in Azure Media Services. You have to track and log things on your own on the client side. App Insights is a good solution for this, and there are some older samples out there showing how to do that with a player application.
Take a look at this sample - https://learn.microsoft.com/en-us/samples/azure-samples/media-services-javascript-azure-media-player-application-insights-plugin/media-services-javascript-azure-media-player-application-insights-plugin/
Just WARNING: it is very old and probably very out of date. I would not use much of the code from that sample, as it is using SDK's from 4 year ago. Just use it as guidance at a high level for what the architecture might look like.
Another solution would be to look to a 3rd party service like Mux.com/Data that can plug into any player framework for client analytics.

Automate Data Import to GA4

I am trying to automate refunds report to google analytics 4. I can't find good documentation on importing this data using analytics management API. I came across https://www.npmjs.com/package/#google-analytics/data which seems to be good for pulling reports from GA but couldn't get a way of doing data import.
I am writing a nodejs script and was hoping someone has encountered this scenario before and could share how they accomplished it. Any help or point in the right direction will be really appreciated.
The alternative to the UA Analytics Management api is the Google Analytics Admin API for ga4
To my knowledge it doesn't support data important at this time the API is still under development it may come in the future there is no way to know.
I would suggest looking at the measurement protocol for ga4 you may be able to use that

Questions on Azure Data Explorer normalization

Our Customer currently build out a number of use cases for the client and facilitate the onboarding of logs from 300+ applications. The client is limited on the number of use cases they can support so they have been looking into the option of creating a custom schema with parsers etc.
The focus is insider threat so they are primarily collecting audit/activity logs for these applications.
The challenges they see them are that application audit/activity logs vary greatly and this will make it difficult to bring the data together from multiple applications. The client has a non-standard architecture and ingest their logs through ADX instead of Sentinel and then forward a subset of data for alerting. They also don’t make wide use of native tables yet.
Please do refer the architecture diagram in the attachment.
Question:
Is there a way of normalizing application audit and activity logs so they can build out insider threat use cases over multiple applications?
Any guidance for this requirement would be of great help. Thanks in advance.

Unable to connect to AAS - Resolution of actual cluster endpoint of Azure Analysis Server failed

Can someone help explain the following error message? We get it when connecting from Excel pivot tables to Azure Analysis Services (AAS) .
"Resolution of actual cluster endpoint of Azure Analysis Server: ' ' failed. Response from Server:,
Technical Details :
Root Activity :
Date (UTC) :
Eventually the error went away when I repeatedly tested my pivot table for ten minutes.
I'd like to understand the internal components in Azure that generate such a confusing error. Hopefully there is something that customers can do to investigate, since Microsoft doesn't advertise their service outages very widely .... unless they are "globally impacting" all of their customers (eg. like when AAD went offline around the world last week).
Is this message indicating an outage in AAS itself or is there some other component(s) involved. Note that I was able to connect from SSMS, but not from Excel pivot tables.
If Microsoft were to decide to post a service health alert about this type of message, what service would it be listed under? AAS? Or some type of DNS?
Any clues would be appreciated. I'd like to understand the underlying issue and be able to predict how long it will normally take for these errors to clear up, once we've encountered them.
There are client libraries that Excel uses to connect to AAS. The libraries are based on OLEDB and they are used for connecting to "MSOLAP" which is a term used for both multidimensional and tabular cubes. Depending on whether you are using the x86 or x64 version of Excel, you will need to find and install a different package to deploy the latest libraries to your machine.
I was instructed to install the latest libraries that are available here:
https://learn.microsoft.com/en-us/analysis-services/client-libraries
Updates to the libraries can become available on a regular basis, and it is important to keep your client libraries up to date if you are connecting to azure resources (since they may be changing on a regular basis as well. )
After updating the client libraries, I no longer encounter the obscure message when I run my repro. Moreover, I'm able to simulate connectivity problems and I am now getting more meaningful errors.
As I started working with Microsoft support, they were able to determine that the original message ("Resolution of actual cluster endpoint of Azure Analysis Server failed") was most likely an authentication issue (expired token or similar). Apparently this error happens for a small handful of reasons .
Internally there are a couple steps that happen while Excel pivot tables are authenticating to AAS. The first step retrieves a token from an identity service using an ADAL library, and the next one uses the token to establish a connection to AAS.
I am happy for the help from Azure support to clear this up. Hopefully this information will help others as well. The moral of the story is that Azure services (like AAS) can change over time, and so you must keep updating your local client libraries so that they don't have any integration failures and they don't generate incoherent error messages.

cognos monitoring

Does cognos 8 have an api which can be queried to find out:
the last time a scheduled report has run
was it successful?
if not, what caused it to fail?
it would be preferable for the api to be web based. I have read about a JMX interface but documentation was lacking.
Cognos 8 has an option to enable logging to an Audit Database. The database would then be queryable for details such as when a report ran, what parameters were used, and what if any errors there were.
Link to IBM site for setting up the logging database.
Link to IBM site for setting up the connection to the database.
Basically you create a compatible database with the proper settings, then tell Cognos how to connect to it, then the next time Cognos services start up, it will automatically create necessary tables for logging and begin populating them automatically.
Typically, to accomplish this in the API fashion, you want to use Cognos SDK. SDK allows you to query scheduling and history associated with report runs and see if the request was completed or failed. If failed, you will see History associated with the failure much like you see in the Cognos Administration section when you go to look at the failed runs.
This is a good place to start to look at a sample:
http://www-01.ibm.com/support/docview.wss?uid=swg21343791

Resources