I have been trying Meshlab software for smoothing the .stl file and exporting the mesh as .wrl file. I used Laplacian smoothing (Laplacian Smooth with smoothing steps of 2, 1D boundary smooting and contangent weighting). After applying the smoothing filter, it worked nicely. Then I exported the file and saved it. However, when I open the saved file afterwards it goes back to the original un-smoothed version, each time. Same problem persists even if I export the mesh as .stl file. I am not sure if this is a bug or something else.
I don't know if I have missed anything. Do you have any idea what might the problem be?
Thanks in advance.
Related
I would like to apply a specific color curve to some 2000 .tif files.
I am a Windows user and so far I have used GIMP for photo editing.
Using Gimp 2.10 I was able to perform such task working on .JPG files using the batch Image Manipulation plug-in (bimp v 2.6; https://alessandrofrancesconi.it/projects/bimp/).
Work flow so far for Jpegs in GIMP-2.10:
Using I created a color curve working on a jpg file (Colors-> curves).
once happy with corrections I saved the curve in an external file ("myset") which hasbeen saved in '\User\appData\roaming\gimp\2.10\curves'
using bimp plug in I choose Add->color correction
in the new window that pop-up I then select only the checkbox "change the color curv from external file" (or similar, menus are not in english...sorry) and navigate to my "myset" curve file.
finally run the batch
When I tried to do the same BUT WITH the .tif files, I got warnings of the kind "unknown filed tag encountered" at the step of importing in bimp the images to process.
That said, I can open the individual tif files in Gimp (File -> open...).
When I do, I still get the warnings "unknown filed tag encountered", but i can click "OK" on the message window and continue importing the file.
Now the "import TIFF" window show me a "Page 1" icon in the top part, then I can choose if opening the file as "levels" or "image". Both choices seem to give the same result.
At that point I can apply my "myset" curve to the file from the tool Colors-> curves.
One potential solution I've been thinking of is to write a script to do this and call it from the command line. I found something along that line here: https://www.gimpusers.com/forums/gimp-user/11100-curves-spline-batch .
Unfortunately:
I have no experience in writing script-fu scripts and very few on command line.
looking at the example in the above link I cannot figure out how/where to point to the "myset" curve in the script.
looking into the Procedure Brouser I do not know which is the one corresponding to the Color->curve tool. ( possibly someting like gimp-drawable-curves-splines, but again I dont know how to have that refer to "myset")
A copy of my "myset" curve and a some .tif esample files can be found here.
Dows anyone have suggestion on perform batch curve color changes on these tif files similarly to what I describe for the jpg? I am open to other solution then GIMP (but for example I cannot open those tif in rawtherappe - don't know why - so that is less of an option)
IMPORTANT: the I need to preserve the tiff metadata (they're georeferenced)
I am quite new to work working with .stl files and in my research, I found they can be opened in a text editor where faces and vertices were clearly written out and easy to read. However, whenever I download a .stl file and open it with a text editor, I get a series of random characters. The same thing happens when I take a .blend file and export it to a .stl file. In spite of this, the .stl files still render correctly. I have attached a picture of what my problem looks like. If anyone could help me out, I would appreciate it.
Thanks.
The random characters when I open .stl files in a text editor.
I have installed a fresh copy of MeshLab by executing the following commands:
git clone --recursive https://github.com/cnr-isti-vclab/meshlab
bash meshlab/install/linux/linux_setup_env_ubuntu.sh
bash meshlab/install/linux/linux_build.sh
I would like to load some meshes of the ModelNet40 dataset, which contains *.OFF files.
Now, when I load a mesh into the image viewer meshlabjs, then the file seems to look correct:
But when I load the same mesh into MeshLab itself, then it is displayed with strong artifacts and it seems like there happened a strong decimation:
Question: Why do these differences in displaying appear? Is it possible, that there are filters automatically applied, from older installations of MeshLab on my mashine? How can I fix this, to display the "correct" looking meshes inside MeshLab instead?
Note: If I export the "broken" mesh and save it somewhere else and display it inside MeshLabJS, then it looks the same broken way with artifacts as it did in MeshLab before.
It is interesting to mention, that the number of vertices and faces is the same in both cases: Correct geometry and incorrectly loaded geometry.
I have checked MeshLab > Filters > Current Filter Script , but the list there is empty. I am on Ubuntu 19.04 and I use MeshLab version 2020.03.
Here is another example. Incorrectly loaded geometry:
Correctly loaded geometry:
I have .obj file of tree and I want to remove some portion of it. Is it possible to do so in Meshlab?
Yes, both of them support .obj files.
http://meshlab.sourceforge.net/
http://wiki.blender.org/index.php/Extensions:2.6/Py/Scripts/Import-Export/Wavefront_OBJ
No they will not open in Blender! I have a few I'd love to edit, & downloaded blender since it was free, & they will not open in it, no mater what!
I understand that you can manually edit .OBJ meshes by selecting edges or faces in Edit Mode, but is there an easier way to do it?
(ie, the Decimate Modifier does not work. It can take a lot of time reducing polygons on a high poly mesh manually...)
I'm using the LaTeX-Beamer class for making presentations. Every once in a while I need to include screenshots. Those graphics are pixel-based, of course. I use includegraphics like this:
\begin{figure}
\includegraphics[width= \paperwidth]{img/analyzer.png}
\end{figure}
or usually something like this:
\begin{figure}
\includegraphics[width= 0.8\linewidth]{img/analyzer.png}
\end{figure}
This leads to pretty bad readibility of the contained text, so I'm asking for your best practices: How would you include screenshots containing text considering, that I will do the output PDF with pdflatex?
EDIT: I suppose I'm looking for something like an 1:1 presetation of the image within beamer. However, [scale = 1.0] doesn't achieve what I'm looking for.
Your best bet is to scale the image outside of Latex for inclusion, and include it in 1:1 ratio. The scaling done by graphics packages in Latex isn't going to be anywhere near as good as possible from other tools. Latex (Tex) has limited floating-point arithmetic capabilities, whereas an external tool can use sophisticated algorithms to get the scaling better.
Another option is to use only a part of the screenshot, the one you want to concentrate on.
Edit: If you can change the font size before taking the screenshot, that's another option—just increase the font size for the screenshots.
Of course, you can combine the two methods.
I have done exactly what you do and e.g defined
\newcommand{\screenshot}[1]{\centerline{%
\includegraphics[height=7.8cm,transparent]{#1}}} % 7.8in
which worked with whatever style I was using at the time. The files included with this macro were all PNGs created with one the usual Linux screen capture tools.
Edit: You may have to play with the size (height and width) of your input files. It came out rather nice for me (and this was from a presentation in 2006).
How about scaling it as follows:
\includegraphics[scale=0.5]{images/myimage.jpg}
This works for me.
Have you tried to convert the image to .eps or .pdf file and use this file in LaTeX?
Maybe try also latex, dvips and ps2pdf.
Problem might be in used viewer, in Linux I use Document viewer or ePDFViewer and output is much worse than in Adobe Reader or Acrobat, which I use in Windows...