How to change the input of a decision operation in the decision center - ibm-odm

I am using the decision center business console, and it allows me to create new action rules or decision tables. Also I can add new variables to the variable sets. But it seems that I cannot modify the INPUT/OUTPUT of the decision operation. I can create a new decision operation, so it seems reasonable that I should be able to add INPUT/OUTPUT variables to it... but I can't figure it out! Please help.

Open rule project map in eclipse and then Add decision operation. Here you find Signature are where you can INPUT/OUTPUT Params.

Related

Is it needed to train Azure Custom translator both ways?

I'm new to azure and am trying to train a translator model. When creating a project, it is asked to choose a source language and a target language. In the list, it can be seen that L1->L2 can be taken but also L2->1. From this raises my question: if I want a model that can translate from one language to another interchangeably L1<->L2, do I need to train 2 models ? One L1->L2 and the other L2->L1. Training is quite expensive and having to do it 2 times seems unpractical.
Each Custom Translator project represents a translation system in one direction. You may choose to build one direction or both directions. You can still translate in both directions if you have built the system only in one direction, Translator will use your custom system in the direction you have built it, and the generic system in the other.
All projects in your workspace access the same data store, so that you can use the same set of documents building a translation system in either direction.

Handle different layout of document using kofax

I am new to KofaxTotalAgility solution, but i am well aware of OCR, OMR and recognition mechanism.
I have two forms in one folder, A and B.
both of them are identical, but due to manual scan there are slight axes change, say 20 pixel right shift, so Layout is slightly differ.
Layout of Image A and Image B are different, position of a form in a page are not fix.
I know, other solution like "abbyy fine reader", provide flexilayout where we can handle this by finding the text and setting up right left top down to automatically identify zones.
As i have started learning KofaxTotalAgility, i am unaware of all option provided by "kofax Transformation Designer".
My question is which Locator should i use, i am currently using/working-on advance zone locator and for one document(Image A) which i set as a reference, extraction is proper. But for other,(Image B) due to layout mismatch text/box field are not getting extracted.
Can anyone point out the right direction from where i can get this case handled properly.
I know, i am asking direct option/solution, any help is highly appreciable.
In general, Kofax Transformations has two groups of locators:
Deterministic. You tell the locator precisely what to do, and how to do it (similar to an imperative approach when programming)
Probabilistic. You just tell your locator what to extract, and it works out the rest (based on AI).
Here's a (non-exhaustive) diagram I created the other day:
When working with forms, you might be tempted to rely on forms-specific locators such as the Advanced Zone Locator. While this locator can account for fields "moving around", for example due to images being jolted, zoomed, or distorted, there are certain limitations. Other locators don't have these limitations - the format locator for example allows you to define a certain pattern (a Regular Expression) that should be matched along with a keyword that has to be found somewhere around that pattern.
For your example, you could create a regex like M|F|X, and then define "Gender" as the keyword that needs to be present on the left.
However, any locator that's ruled by determinism follows Murphy's law - at some point that keyword might change. There could be different languages. And maybe additional letters for certain genders might be added; ultimately breaking your extraction logic.
Enter AI - while Murphy's law still applies when using Group Locators, the difference here is that users can train the system to pick up the new data. Said locator will automatically work out the best way to extract that piece of data. If you used a format locator, the customer would need to get back to you to add additional expressions, or have the keywords changed.
In your particular case, I'd try to use a Trainable Group Locator first. If you already know what you're looking for - for example SSNs that you have somewhere in a database, go for the Database Locator. Use Format Locators as a last resort, as tempting as they may be. Advanced Zone Locators are useful when you deal with forms, but I find myself using them almost exclusively for handprint or checkbox recognition.

What is the role of feature type in AzureML?

I want to know what is the difference between feature numeric and numeric columns in Azure Machine Learning Studio.
The documentation site states:
Because all columns are initially treated as features, for modules
that perform mathematical operations, you might need to use this
option to prevent numeric columns from being treated as variables.
But nothing more. Not what a feature is, in which modules you need features. Nothing.
I specifically would like to understand if the clear feature dropdown option in the fields in the edit metadata-module has any effect. Can somebody give me a szenario where this clear feature-operation changes the ML outcome? Thank you
According to the documentation in ought to have an effect:
Use the Fields option if you want to change the way that Azure Machine
Learning uses the data in a model.
But what can this effect be? Any example might help
As you suspect, setting a column as feature does have an effect, and it's actually quite important - when training a model, the algorithms will only take into account columns with the feature flag, effectively ignoring the others.
For example, if you have a dataset with columns Feature1, Feature2, and Label and you want to try out just Feature1, you would apply clear feature to the Feature2 column (while making sure that Feature1 has the feature label set, of course).

ActivePivot with a rules engine

I have just started on a project which his regulatory in nature and the business area of the IB I work with uses ActivePivot to manage their securities (inventory).
One of the tasks we need to do is that the ActivePivot data set and run some sort of simple rules engine over the data that feeds ActivePivot. There is a little bit of netting involved at the transactional level but it's mostly simply rules using basic operators. I haven't used ActivePivot before but the users are telling me it doesn't really allow them to add fields within the cube which I understand from a technical perspective. I also noted that ActiveViam have a product called ActiveUI which on the surface appears to do this?
Has any one any tips/advice on what worked for them? The business also want a better data visualisation tool (graphs and the likes).. I was looking at tableaux but open to suggestions. Many thanks for any help given.
There is no clear question here so I will answer to your different points one by one:
run some sort of simple rules engine over the data that feeds ActivePivot
Then you can add your rule engine in your project on the data set before feeding ActivePivot as if you were not using ActivePivot afterwards.
users are telling me it doesn't really allow them to add fields within the cube
you cannot add fields once the cube is started but you can update the description of your cube in your project to integrate the new fields brought by your new logic.
I also noted that ActiveViam have a product called ActiveUI which on the surface appears to do this?
ActiveUI is a UI for the ActiveViam products (including ActivePivot), so it provides you (among others) tables, charts to navigate your data.
The business also want a better data visualisation tool (graphs and the likes).. I was looking at tableaux but open to suggestions
ActiveUI can provide you this. ActivePivot follows the standard for OLAP databases (XMLA) so it is also compatible with other XMLA clients like Excel and Tableau. Your BI has probably already chosen which client they would use so you should see with them.

OpenMDAO 1.x: recording desvars, constraints and objective

How can you get information about which variables are design vars, objectives or constraints from the information saved by recorders? It would be useful to print this information to a file to track optimization progress during a run. It looks like the RecordingManager.record_iteration doesn't really allow for this at the moment, since you only pass the root system and a metadata dict meant for optimizer settings.
Would it be possible to add an argument to the RecordingManager.record_iteration called e.g. optproblem, which is a dictionary with dictionaries with desvars, constraints and objective?
A simple OptimizationRecorder could then dump out column formatted files with the quantities for easy plotting during the optimisation.
This is something we have on our list of to-do's for the near future. Our current planned approach is going to be to augment the meta-data (already being saved) of variables with labels identifying them as des-vars, objectives, and constraints. Then you could pull that information out as part of a custom case recorder if you want. We plan on doing it this way because it doesn't require modifying the recorder's api at all. I think we'll have something like this implemented in the next month or so.

Resources