Autodesk Forge - how to use the new Model Derivative IFC pipeline? - node.js

I'm working with autodesk forge, because I have a project for the University, but I have a small problem now.
I read that the IFC Model Derivative processor is currently based on the Navisworks importer, but a better implementation is using Revit instead. As I work mostly with IFC Files this could make a big difference.
I made a small application with the help of the forge-apis npm package, which enables me to easily integrate the Forge APIs.
The Problem is that I don't see an option to use the new Model Derivative IFC pipeline. Is there a way to implement it in my application?

As this point, you'll need to make requests (e.g. using request package). When we completely move to Revit pipeline you can resume with the forge-api package.

Related

Building Python OCR using machine learning

There are a ton of questions dealing with OCR and Machine Learning, I am looking for guidance building my own from scratch.
I have an obscene number of photographs that contain text pertaining to the feature in the photo. The text is the latitude, longitude and id of the feature. I am looking for a way to extract this information into a text file to feed into my GIS.
I am sure Tesseract and Pytesseract would do exactly what I want, I however have a blocker in that I cannot load additional software onto the PC I am working on.
My PC is connected to a strictly controlled and secure network. I cannot install new software. I can however “pip install” any Python libraries needed, using a mobile hotspot. I have installed the Pytesseract library in Python. However if I have understood there is a dependency requiring a windows install file to be downloaded and installed before this works.
So I have decided to try (as a side project) create my own OCR model using Python and whatever libraries I need. The only issue is, there is a ton of information online and trying to find a focused and easy to follow process is not easy.
I am looking for resources detailing step by step what I need to do to create a training dataset, train a model and feed the images into the train model to get an output that makes sense.
I have been using OpenCv to process an image (crop, filter etc) to get bounding boxes of all the identifiable text in the test image. I am not sure where to go from there.
Are there any recommended tutorials online / resources that might make sense to a complete novice? I am using Python 3.5.

How can I generate image with Nodejs

I've made dynamic image generate service like this https://dummyimage.com/ by PHP.
now I'm thinking of make a dynamic image generator with nodejs.
but it's not as easy as thought.
what kind of module can I use for dynamic image generate?
Another possibility is to use the pureimage package which is a pure JavaScript implementation of the HTML Canvas 2D API for NodeJS. It has no native dependencies.
you may use this node module for generating random images. https://www.npmjs.com/package/js-image-generator
I created a thing that generates images from a "function app". Hosting this on Azure costs me only "pennies per month".
You can copy my code and repurpose it. (also function apps support using javascript, which would work similarly)
https://github.com/djangofan/chord-bars

How to efficiently update the API when Swagger spec file is updated? (express, nodejs)

I'm trying to setup a nodejs-express boilerplate for my new project, and this time I want to try doc-driven flow. I've checked couples of packages like swagger-node, swaggerize-express ...etc. They all provide great functionalities.
However, I don't see anything that could support incremental scaffolding when the Swagger file is updated. That means when the spec changes I have to manually check the diff and manually add/modify the new specs. That doesn't sound cool.
Does anyone could share something that is more reasonable? Thanks!!!
Edit:
After trying some frameworks, I decided to use swagger-express-middleware. This framework offers a convenient way to automatically check routes/parameters for your service.
You can use tools like swagger-maven-plugin to incrementally rebuild your server code, which means reading from your api definition and updating/building code as necessary. There are SAAS products like SwaggerHub which enable this as well, by merging code and pushing to git.

securing the source code in a node-webkit desktop application

first things first , i have seen nwsnapshot. and its not helping.
i am building an inventory management system as a desktop app using node-webkit . the project being built is using compoundjs (mvc javascript library). which have a definite folder structure (you know mvc) and multiple javascript files inside them.
the problem is nwsnapshot allows the app to have only a single snapshot file but the logic of application is spread over all the folders in different javascript files.
so how do i secure my source code before shipping it to client? Or any other work-around Or smarter way (yes, i know about obfuscating).
You can use nodewebkit command called nwsnapshot to compile the javascript code into binary which will be loaded into the app without specifying any js file
nwsnapshot --extra-code application.js application.bin
in your package.json add this:
snapshot: 'application.bin'
It really depends on what you mean by "secure".
You can obfuscate your javascript code fairly well (as well as potentially improve performance) by using the Google Closure Compiler.
I'm not aware of any off-the-shelf solutions to encrypt/decrypt your javascript, and honestly I would question the need for that.
Some people think they need to make it impossible to view their source code, because they're used to dealing with compiled languages where you only ship binaries to users. The fact is, reverse-engineering that binary code was never as difficult as some people think it is, so if there's any financial incentive, there is practically no difference between shipping source code and the traditional shipping of binaries.
Some languages have offered genuine encryption of deployed assets, such as Microsoft's SLPS. It seems to me that the market for this was so small that Microsoft gave it to a partner (just my view). The truth is that most customers are not interested in taking your source code; they're far more interested in your ability to service and support that code in an efficient manner, while they get on with their job.
You may consider to merge the JS files into one in the build process and compile it.

XNA 4.0 load external 3D objects on Windows

I'm working on a project where my XNA 4.0 powered 3D engine needs to load external fbx models input by the user, in run time rather than in the default compile time way.
I understand XNA is built to bundle/process complex resources compile time to make the runtime smaller, but as I only need to target Windows I wonder if it is possible to load models with textures externally, and if so, how?
Yes, as #Andrew mentioned, using the built in content pipeline would require a developer install so that the content pipeline is available. Of course, you can parse it yourself and pull out the information at runtime to avoid that dependency. There are people out there doing it ... for example, the guys at sandswept studios have an API to do this, and are willing to discuss commercial agreements (just contact them):
http://thunderfist-podium.blogspot.com/2008/09/fbx-and-xna-part-1-fbx-format-and.html
I found the solution here:
http://create.msdn.com/en-US/education/catalog/sample/winforms_series_2

Resources