po2html missing html template - python-3.x

I have no coding experience but I need to convert a .po file into an .html file. I am learning how to use this translate toolkit.
I am using translate tool kit here: http://docs.translatehouse.org/projects/translate-toolkit/en/latest/commands/html2po.html
Managed to get far as to install python and things.
I have gotten as far as below, and am unsure how to proceed.
Folder xh has example.po (which has relevant msgids and msgstr) and also example.html (which is blank)
C:\Users\bob>po2html -i C:\Users\bob\Desktop\xh -o C:\Users\bob\Desktop\xh
processing 1 files...
po2html: WARNING: Error processing: input C:\Users\Oh\Desktop\xh\example.po, output C:\Users\Oh\Desktop\xh\example.html, template None: must have template file for HTML files
[###########################################] 100%
How do I create this html template and what does it look like (My only knowledge is that you can create a HTML file in notepad, but I'm not sure what to put in it so example.html is blank atm)

Related

WGET - how to download embedded pdf's that have a download button from a text file URL list? Is it possible?

Happy New Years!
I wanted to see if anybody has ever successfully downloaded embedded pdf file's from multiple url's contained in a .txt file for a website?
For instance;
I tried several combinations of wget -i urlist.txt (which downloads all the html files perfectly); however it doesn't also grab each html file's embedded .pdf?xxxxx <---- slug on the end of the .pdf?*
The exact example of this obstacle is the following:
This dataset I have placed all 2 pages of links into a url.txt:
https://law.justia.com/cases/washington/court-of-appeals-division-i/2014/
1 example URL within this dataset:
https://law.justia.com/cases/washington/court-of-appeals-division-i/2014/70147-9.html
The embedded pdf link is the following:
https://cases.justia.com/washington/court-of-appeals-division-i/2014-70147-9.pdf?ts=1419887549
The .pdf files are actually "2014-70147-9.pdf?ts=1419887549" .pdf?ts=xxxxxxxxxx
each one is different.
The URL list contains 795 links. Does anyone have a successful method to download every .html in my urls.txt while also downloading the .pdfxxxxxxxxxxxxxx file's also to go with the .html's ?
Thank you!
~ Brandon
Try using the following:
wget --level 1 --recursive --span-hosts --accept-regex 'https://law.justia.com/cases/washington/court-of-appeals-division-i/2014/.*html|https://cases.justia.com/washington/court-of-appeals-division-i/.*.pdf.*' --input-file=urllist.txt
Details about the options --level, --recursive, --span-hosts, --accept-regex, and --input-file can be found in wget documentation at https://www.gnu.org/software/wget/manual/html_node/index.html.
You will also need to know how regular expressions work. You can start at https://www.grymoire.com/Unix/Regular.html
You are looking for a web-scraper. Be careful to not break any rules if you ever use one.
You could also process the content you have received through wget using some string manipulation in a bash script.

collada2gltf converter can't produce *.json file

I am reading a book: Programming 3D Applications with HTML5 and WebG , it involve a Vizi framework.
All the examples load the *.json file instead of *.gltf file. Why?
When I load *.gltf, it doesn't load any result, and the collada2gltf converters only produce *.gltf, *.bin, *.glsl files and so on.
What should I do?
.gltf is a JSON file. Try to open it with a text editor and see for youself. .bin and .glsl files are just additional resources, linked from .gltf file. Those are geometry buffers and shaders respectively. So to make it work you should make sure that all the files produced with the converter are also available to a web browser you running your code in.
Also you can try to add -e CLI flag to collada2gltf and it'll embed all the resources into result .gltf file.

SublimeText3 + pandown + pandoc: includes_paths not working

I'm using ST3+pandown+pandoc to convert markdown to PDF. I want to use pandown's includes_paths setting to avoid typing the path to my image directory every time. I haven't been able to get it to work, however. Here's a MWE:
I have a directory structure as follows:
text.markdown
test/img.pdf
In text.markdown, I have:
![](img.pdf)
I've got set includes_paths as follows in Pandown.sublime-settings:
"includes_paths":
[
"test/"
],
But, no dice. I've also tried with an absolute path, ./test, and test. Any ideas?
I think Pandown's includes_paths only applies to Pandoc's --include-in-header, --include-before-body and --include-after-body options, not image locations etc.
From Pandown.sublime-settings about includes_paths:
Pandoc apparently doesn't search for values for its --include
arguments anywhere but the working directory, which makes
working from a standard stylesheet or standard script
sort of tedious.
A workaround, using the graphicx package loaded in the YAML header and \graphicspath:
---
header-includes:
- \usepackage{graphicx}
---
\graphicspath{{test/}}
![](img.pdf)
Pandoc will say that it can't find img.pdf, but the image will be present in the final pdf.

How can I tell what language a file is written in?

I've been trying to find, for a long time, what language this file is written in so that I can decompile it. I have tried to decompile as .luac, .class and also tried to open it as .jar and .rar and .zip.
Although the file extension is .car I have never seen this extension before and there certainly aren't any openers for it on the internet. I have even gone to the point of finding a .car opener, but it wasn't for my .car.
So, I suspect it has just been renamed.
Can anyone tell me what language it is coded in?
I don't know if I'm allowed to post files here, as I have only just joined, but here is a Dropbox link to the file: https://www.dropbox.com/s/y6jd62lfywoskqi/code.car?dl=1
Any help would be appreciated. This is the first line in Notepad++:
rac T D ` constants.lu œÂ tools.stashsaver.lu à scenes.sellitems.lu ˆî scenes.draw.lu ¼ gui.menu.lu 6 scenes.missions.lu ˆP
A quick dump of the file leads to several URLs referencing:
http://www.coronalabs.com/
So based on this I'd say the file was created with the SDK they offer.
https://coronalabs.com/products/corona-sdk/
Corona lets developers use integrated Lua, layered on top of
C++/OpenGL, to build graphic applications. https://en.wikipedia.org/wiki/Corona_%28software%29

Linux: WGET - scheme missing using -i option

I am trying to download multiple files from yahoo finance using wget.
To do that i used a python script to generate a text file with all urls that i need.
When downloading a single file (a csv file) using the following code:
wget ichart.finance.yahoo.com/table.csv?s=BIOM3.SA&a=00&b=5&c=1900&d=04&e=21&f=2013&g=d&ignore=.csv
everything goes OK!
However, when the option -i is added and instead of reading the url directly, but instead reading it from the file, i get the error:
Invalid URL ichart.finance.yahoo.com/table.csv?s=BIOM3.SA&a=00&b=5&c=1900&d=04&e=21&f=2013&g=d&ignore=.csv: Scheme missing
The file that contains the urls is a text file with a single url in each line. The urls are exactly like the one in the first example, but with some different parameters.
Is there a way to correct this?
Thanks a lot for reading!!
To solve the problem I added double-quotes on the links and a web protocol. For example:
"http://ichart.finance.yahoo.com/table.csv?s=BIOM3.SA&a=00&b=5&c=1900&d=04&e=21&f=2013&g=d&ignore=.csv"

Resources