How does one access the (g) ID of the obj from the k3d scene objects which I have passed from vtkOBJImporter using the method outlined here.
(g) should be the group name. However I could't find documentation of vtk of where the group name is stored once it is imported by vtkOBJImporter.
.OBJ
From the current source code (https://gitlab.kitware.com/vtk/vtk/-/blob/master/IO/Import/vtkOBJImporter.cxx#L275), it seems the g type is not handle by the vtkOBJImporter.
Feel free to open a feature request (https://gitlab.kitware.com/vtk/vtk/-/issues/new?issue[assignee_id]=&issue[milestone_id]=). Or even a merge request :)
Related
Title says it all, has anyone figured out a way to create and save a file with the equivalent of what would be the enum file.type.MISCBINARY via suitescript suitelet N/file module (this does not exist in any documentation)
I have already found that I can upload any file to whatever ns file type I want regardless of file mime or .ext however I would like to not just upload files to the wrong NS enum just to get the job done.
example, I am currently creating .msg files under the type enum file.Type.MESSAGERFC as this is the closest match however if uploaded through drag and drop the type field is MISCBINARY
I have been through all of the documentation and even dug into several of the object properties, I also took a look at the form object of a MISCBINARY file uploaded through the NS drag and drop feature.
I have tried in the type field:
file.Type.MISCBINARY
'MISCBINARY'
'miscbinary'
'otherbinaryfile'
I also tried to set the file.fileType property by file.fileType = 'MISCBINARY' of course this did not work since this property is readOnly but it was worth a shot.
I'm quite new to Data Factory and Logic Apps (but I am experienced with SSIS since many years),
I succeeded in loading a folder with 100 text-files into SQL-Azure with DATA FACTORY
But the files themselves are untouched
Now, another requirement is that I loop through the folders to get all files with a certain file extension,
In the end I should move (=copy & delete) all the files from the 'To_be_processed' folder to the 'Processed' folder
I can not find where to put 'wildcards' and such:
For example, get all files with file extensions .001, 002, 003, 004, 005, ...until... , 996, 997, 998, 999 (thousand files)
--> also searching in the subfolders.
Is it possible to call a Data Factory from within a Logic App ? (although this seems unnecessary)
Please find some more detailed information in this screenshot:
(click to enlarge)
Thanks in advance helping me out exploring this new technology!
Interesting situation.
I agree that using Logic Apps just for this additional layer of file handling seems unnecessary, but Azure Data Factory may currently be unable to deal with exactly what you need...
In terms of adding wild cards to your Azure Data Factory datasets you have 3 attributes available within the JSON type properties block, as follows.
Folder Path - to specify the directory. Which can work with a partition by clause for a time slice start and end. Required.
File Name - to specify the file. Which again can work with a partition by clause for a time slice start and end. Not required.
File Filter - this is where wildcards can be used for single and multiple characters. (*) for multi and (?) for single. Not required.
More info here: https://learn.microsoft.com/en-us/azure/data-factory/data-factory-onprem-file-system-connector
I have to say that separately none of the above are ideal for what you require and I've already fed back to Microsoft that we need a more flexible attribute that combines the 3 above values into 1, allowing wildcards in various places and a partition by condition that works with more than just date time values.
That said. Try something like the below.
"typeProperties": {
"folderPath": "TO_BE_PROCESSED",
"fileFilter": "17-SKO-??-MD1.*" //looks like 2 middle values in image above
}
On a side note; there is already a Microsoft feedback item thats been raised for a file move activity which is currently under review.
See here: https://feedback.azure.com/forums/270578-data-factory/suggestions/13427742-move-activity
Hope this helps
We have used a C# application which we call through 'app services' -> webjobs.
Much easier to iterate through folders. To call SQL we used sql bulkinsert
After struggling with a template var named {{engine.x.y}}, which was "silently" showing nothing until I changed it for something else such as {{_engine.x.y}}, I saw that other words such as {{action}} seemed to be "reserved", with no "escaping" possibility...
Is there any list of such "reserved words" in order to lessen the "surprise factor" ? I couldn't find any.
EDIT
even more strange, when I use {{engine}} directly (not engine.x.y) it always gets replaced by the string ".hbs"
however in my context (Node v4.6 + email-templates + handlebars) {{action}} works correctly (linked problem above looks specific to Ember.js)
Found it - the issue isn't with Handlebars (which Josh Crozier found to not reserve any names in the comments above), it's with the email-templates library that you're using. The render function it uses sets several properties on the object that gets passed as the template context, meaning that if your object contains properties with the same name, they get clobbered.
The relevant line of code in the email-templates source can be found here, but for the sake of completeness, the properties which get overwritten are:
filename (set to your template's filename)
engine (set to the file extension of the templating engine that's being used)
templatePath (set to the directory that contains your template)
The most obvious solution would be to just change the name, but if that's not an option you may need to file an issue on their repo.
I am getting an import error in a specific environment with a managed CRM 2011 solution. The solution has been imported before into many other environments, but the one in particular where it is failing is throwing the following error:
Dependency Calculation
role With Id = 9e2d2d9b-645f-409f-b31d-3a9c39fcc340 Does Not Exist
I am a bit confused about this. I searched within the solution XML and was not able to find any reference to this particular GUID of 9e2d2d9b-645f-409f-b31d-3a9c39fcc340. I cannot really find it in SQL either, just wandering through the different tables, but perhaps I do not know exactly where to look there.
I have tried importing the solution multiple times. As a desperation effort, I tried renaming all of the security roles in the destination environment prior to importing, but this did not help.
Where is this reference to a security role actually stored? Is this something that is supposed to be within my solution--which my existing CRM deployment is expecting me to import?
How do I fix the problem so that I am able to import this solution?
This is the code we used to fix the issue. We had to run two different scripts. Script A we had to run a total of four times. Run it once, attempt the import, and then consult the log to find the role that is causing the problem--if you receive another error for another role.
To run script A, you must use a valid RoleTemplateId from your database. We just picked a random one. It should not matter precisely which one you use, because you will erase that data element with script B.
After all of the roles are fixed, we got a different error (complaining something about the RoleTemplateId was already related to a role), and had to run script B. That removes the RoleTemplateId from multiple different roles and sets it to NULL.
Script A:
insert into RoleBaseIds(RoleId)
values ('WXYZ74FA-7EA3-452B-ACDD-A491E6821234')
insert into RoleBase(RoleId
,RoleTemplateId
,OrganizationId
,Name
,BusinessUnitId
,CreatedOn
,ModifiedOn
,CreatedBy
)
values ('WXYZ74FA-7EA3-452B-ACDD-A491E6821234'
,'ABCD89FF-7C35-4D69-9900-999C3F605678'
,(select organizationid from Organization)
,'ROLE IMPORT FIX'
,(select BusinessUnitID from BusinessUnit where ParentBusinessUnitId is null)
,GETDATE()
,GETDATE()
,null
)
Script B:
update RoleBase
set RoleTemplateId = NULL
where RoleTemplateID='ABCD89FF-7C35-4D69-9900-999C3F605678'
Perfect solution, worked for me! My only comment would be the error in Script B: it shouldn't clear the template IDs of all roles for the given template, only the template ID of the newly created "fix" role, as follows:
update RoleBase
set RoleTemplateId = NULL
where RoleID='WXYZ74FA-7EA3-452B-ACDD-A491E6821234'
I would've gladly put this in a comment to the answer, but not enough rep as of now.
I am trying to get Subsonic(3.0.0.4) to work(.Net 3.5 - VS2010), simple install that looks at the Northwind database using ActiveRecord. I used the 5 minute demo from http://subsonicproject.com/docs/The_5_Minute_Demo. I added the reference to the dll, un-blocked the T4 templates that I downloaded, changed the Settings.ttinclude to point to the northwind database in 3 places as per video. Dragged the T4 templates (as per video) to my Models folder in VS2010.
If I compile I keep on getting "Type or namespace name 'something' could not be found" There are 13 error (same amount as the amount of tables in the DB..) so I am guessing that there must be some command that turns every table into a class, but where? I right-clicked on ActiveRecord.tt and Context.tt and selected "Run custom tool" with no success.
Any help, I follow this demo to the letter and still get the error. Any help.. please?
Ok, I do not know why but after I added Structs.tt it compiled fine. I still could not get the foreach in the view to work, the error was: foreach statement cannot operate on variables of type 'object' because 'object' does not contain a public definition for 'GetEnumerator'
I could see out of the corner of Rob's video that it did not inherit from System.Web.Mvc.ViewPage and figured out that it should be System.Web.Mvc.ViewPage>
Sometimes clever people(Like Rob) cannot think for proper dummies(like me) Such is life?