I would like to develop integrations with Acumatica ERP (version 2019 R2) for a BI Application which extracts data from a local Acumatica instance and loads it into an external destination.
In order to test this properly, I need a reasonable amount of data.
The documentation of Acumatica states that a tenant can be created from a snapshot (Refer to: https://help-2019r2.acumatica.com/(W(20))/Help?ScreenId=ShowWiki&pageid=57406416-cf4c-4f91-874e-489d680e8ef0 ).
Are there existing snapshots for tenants which developers can use to test their integrations as entering the data manually is a lot of work?
When creating the site through the configurator you can state the initial data, which will be inserted when the instance is created. I would suggest to select "SalesDemo" and a number of documents and settings would be ready for you to use once the site is created.
Example
Related
We are storing some user info in production that we don't have to keep and thus need to disable custom logging in application insights immediately. I can't change the code in between production rollouts.
Is there any setting or configuration available in application insights?
One solution which will not require code redeployment:
If using Classic Application Insights resource -> migrate to Workspace-based (no code changes required)
Use Workspace transformation DCR
Using them you can either completely drop data or can erase some fields.
Note, though dropped data doesn't appear in Application Insights, since it was ingested, it still will be partially charged (please refer to documentation).
Note, Workspace schema for Application Insights tables is slightly different. Transformation DCR needs to be written in Workspace schema. For instance, "requests" table name is called "AppRequests" in Workspace. You can explore individual fields by opening this table in Workspace directly.
Since ADC is provided by MS as SaaS to customers, is MS taking backups of the dataset and business glossary? If yes, how often and how can a customer get access to the backups for recovery purposes?
Unfortunately, there is no explicit backup/restore feature available for catalogs.
I would suggest you to vote up an idea submitted by another Azure customer.
https://feedback.azure.com/forums/906052-data-catalog/suggestions/33125845-azure-data-catalog-backup-feature
All of the feedback you share in these forums will be monitored and reviewed by the Microsoft engineering teams responsible for building Azure.
The closest way to achieve this with current functionality is to use the Azure Data Catalog REST API to extract all assets and persist them locally (and re import them manually later).
There is a sample application available that demonstrates this technique: Data Catalog Import/Export sample tool.
Over the months of exploration into PBI, started with successfully creating a workspace using PowerBI pro license and ended with hosting a pbi report embedding into my custom MVC site using apps-own-data model.
First experience is maximum allowed embedded tokens running out.
My company decided to create a dedicated A1 core powerbi embedded service in a azure account. Now I have overcame token running out of count issue but seems cringy that my powerbi embedded service besides paused still my embedded site runs and accesses powerbi reports without any interruption.
Previously have created AD using embed tool provided by microsoft. I can see my AD been created in azure portal too.
How this is possible to view a pbi report where my azure powerbi embedded service been paused.
Am i supposed to use those pbi reports without getting billed?
Microsoft has limited information on documentation to clarify my doubts, but the PBI community site is somewhat helpful still having trouble getting clarification for the same.
Help required.
For your question:
How this is possible to view a pbi report where my azure powerbi embedded service been paused. Am i supposed to use those pbi reports without getting billed?
If the A1 Node is paused, then no, you will not be able to see your report or use the service in your front end. It has to be running to deliver the reports in your custom front end. You can still go into the Power BI Service with an assigned Power BI Pro licence and see your report, the workspace that the report has been deployed to, is flagged as 'embedded capacity' that will be shown as a diamond shape next to it.
You allocate the workspace to a capacity by editing the workspace and selecting the 'Advanced' option then 'Dedicated Capacity'
The MS documentation outlines pausing will not deliver content.
Pausing a capacity may prevent content from being available within
Power BI. Make sure to unassign workspaces from your capacity before
pausing to prevent interruption.
Pausing is designed to allow you to stop delivering connect for example, out side business hours, I have a few clients that only run their internal and external report during 7am to 7pm, the other 12 hours the service is paused. The A sku billing costs are reduced to 50%.
Hope that helps
I have some Selenium test code that I need to run in parallel. In order for Selenium to run effectively, certain configurations have to be done on the machine (I.E. zone settings, Chrome and Firefox installs, etc.) and these settings are hard (if not impossible) to apply via an automated approach. I've manually created a VM, done all the setup and created an image following the directions in Microsoft's documentation.
Now I need to setup my code so that I can specify a VM image to use when creating the nodes. I've searched as much as I can and not found any documentation that explains how I can go about doing this. The example in the DotNetTutorial sample doesn't seem to have any way to specify an image.
There is a feedback item here on this same topic and shows the request as started on Jun 1st 2015. I'm hoping this means that it's done now and that it just hasn't been documented well.
Q: How I can specify a custom VM image as the source for my Azure Batch nodes?
https://github.com/Azure/azure-sdk-for-net/blob/AutoRest/src/Batch/Client/changelog.md
• Added support for deploying nodes using custom VHDs, via the OSDisk property of VirtualMachineConfiguration. Note that the Batch account being used must have been created with PoolAllocationMode = UserSubscription to allow this.
Updated Answer on 2017-12-05:
Custom images are now supported through normal Batch accounts (i.e., Batch service pool allocation mode accounts). You will need to specify a valid ARM Image Id and use Azure Active Directory authentication to create custom images (shared key auth does not support custom images).
Updated Answer on 2017-03-17:
Custom images are now supported through "User Subscription" Batch accounts. You can create these types of accounts in Azure Portal or through the newest management SDKs for supported languages.
Previous Answer:
Currently, custom VM images are not supported. As you noted, this is a feature that is being worked on. In addition to uservoice, you can periodically check for product updates at this site.
I have added some tables to my database in in windows azure via entity framework however I am not able to access these tables through the server side scripts (mobile services custom api)and they do not appear through the "MOBILE SERVICES: DATA" section. Do I have to add these tables and set permissions on them manually though the portal to get access to these via the scripts etc? I am sure there is some documentation on this somewhere but have been chasing my tail trying to find it.
The only table that currently appears there is the TodoItem table created by default.
A bit of direct on this would be great
you need to move it to the schema of your Mobile Services App and add the tables, see: http://blogs.msdn.com/b/jpsanders/archive/2013/05/24/using-an-existing-azure-sql-table-with-windows-azure-mobile-services.aspx
You only need to define the table name through the portal interface, its pretty easy to use. This is also when you define if that table requires authentication, and which kind of authentication you will use, also pretty well explained in the interface so I will leave that to you. After you've done this basic layout in azure, the entity framework will take over and define the table details from within your code. (Version and table type Depending) Something like;
private IMobileServiceSyncTable<MyTable> mySyncTable = App.MobileService.GetSyncTable<MyTable>();
Your table names in azure must exactly match the class names that your using in the code to define the tables, this mirroring is how the server maps your data to its intended location in the cloud.
You should now have complete access to your cloud data from the MobileServices API by calling operations on mySyncTable.