How multiple users can work in the same Keyhole Markup Language (KML) file at the same time? - kml

I have tried to create a solution with SharpSVN to control the version of the file but I have not been successful so far, this issue is not an error that is occurring but rather a question as a way to find a solution. What do I need to know is how multiple users can work in the same Keyhole Markup Language (KML) file at the same time? Thanks.

KML is an XML file so for multiple users to edit and access the latest KML file, there are several possible solutions to consider.
Add the KML file to a shared repository (svn, git, etc.) to allow people to check in changes and update with changes made by others. You could post your KML file to a project on github.com.
Place the KML onto a shared network drive such as Google Drive and give read/write access to users as needed. Users can read or write the KML as needed and the changes will be accessible to other users.
Upload KML to a shared Google Fusion Table then allow users to edit/add the data. The data is then available to download as KML NetworkLink that auto-updates to the latest contents. Here are some tutorials using Google Fusion Tables.

Related

Online file viewer (all formats)

I have some files on S3 and would like to view those files in web. Problem is that the files are not public and I dont want them to be public. Google doc viewer works but condition is, files should be public.
Can I use office web apps to show in browser. Since the files are private, I do not want to store any data on Microsoft servers. It looks like even google doc viewer stores the info while parsing.
What is the cleanest way?
Thanks.
I have looked around for something similiar before and there are some apps you can install locally (CyberDuck, S3 Browser, etc). In the browser has been limited until recently (full disclosure I worked on this project).
S3 LENS - https://www.s3lens.com/
I probably get a minus here, but also Microsoft has an online viewer, which works the same way: the file needs to be publicly accessible.
Here is the link: https://view.officeapps.live.com/op/view.aspx
What I cloud add is that those files need to be publicly accessible only for a short period, i.e. until the page gets opened. So you cloud trick them by uploading the file to be viewed to a public temporary storage in a randomly generated folder and give that url to the online viewer.
Of course this is not that safe, since the file will get as some point to the temp storage and then to Google or Microsoft, but the random path names offer some degree of safety.
I've created recently a small glitch app, which demonstrates what I just explained: https://honeysuckle-eye.glitch.me/
It uploads local files to a temp storage and then opens the viewer from that temp storage; the temp storage only last for one download, so it is pretty safe.

Set limit on number of files user can upload in Liferay Documents and Media?

I've seen in liferay where we can specify the upload size of a file
dl.file.max.size=
But I haven't found a way to specify how to limit the number of files a user (or community) can upload. Obviously, we don't want a user or community upload massive amounts of files and filling up our shared drive. Is there a way to do this?
There's no such quota - that's why you didn't find it. However, you are able to extend Liferay to contain and honor a quota. It has been demonstrated with this app - source code is linked and you might be able to take it as the basis for building your number-of-files-quota. It also doesn't seem to take the number of files into account, but it keeps an eye over the volume stored already and should be easy to extend.
I'm assuming that the spanish user group would happily accept pullrequests if you have a good extension to this plugin.

Document Mng on Redmine: Anyone use DMSF plugin or find an easy way to manage docs in the Files tab with a 'wiki' as a front end?

I'm looking to use Redmine for document management. I know that Redmine is not ideal for this task but there is already a lot of content on the site so I'd like to utilize it if possible.
Redmine currently does not a have great documents module. The files we've uploaded look to be amended on that specific page and it doesn't seem to be able to move to another page (unless you download and re-upload to the proper page).
Idea 1
I see there is a Files section, which could work as a central repository (and you can upload document based on release) however, is there a way to set up a nice-looking 'front-end' page that automatically updates based on new submissions to the Files tab? I envision this front end to be a simple wiki page with the document name, a short description and a links to the file posted in the Files tab.
There are so many documents uploaded to varying pages on the Redmine site. I would only do the whole download and re-upload of files if there was a way to automatically update the 'front end' wiki.
Idea 2
I see there is a DMSF plugin for Redmine. Has anyone used this before and has is solved document management issues? I'd like to hear your feedback. Even if DMSF doesn't totally solve my issue, anything is better than what I have now.
Thanks!
In my opinion DMSF module is a perfect companion for Redmine. We have adopted it in our company. You can easily deal with document versions, webdav access, custom approval workflow, document modifications notification with the extra value of being well integrated with Redmine features (roles, dynamic links in Wiki and issue text and notes).

What are the security/privacy concerns with MHTML (.mht) files?

My customer is currently using MHT files for storing offline representations of browsed web pages. The files are saved and later viewed in Internet Explorer.
When viewing the files we would like to be sure there is absolutely no network activity to the original site or any other site - the content should be browsed 100% offline, and should not have any special "local" privileges as well (i.e. access to file:// protocols etc.). We would like to keep JS running if possible, and we can suffer to consequences of disabled features because of working offline.
We are willing to change the viewer or even the file format (and convert all old mht files as well) if a better solution is suggested.
Thanks for any help on this,
Udi
It is not possible to guarantee that there will be no network activity unless you go to offline mode in Internet explorer. Though the advantage of saving a web page to a mht file is that all the info for displaying the page (including images) are stored in one file instead of several files and folders, making archiving easy, if the web content has links to other pages, clicking on links will initiate network activity.
One option is to post-process the mht file and replace the url links with just the title of the link. For e.g, replacing
<A=20
title=3D"Conduction band"=20
href=3D"http://en.wikipedia.org/wiki/Conduction_band">conduction =
bands</A>
with "Conduction band".

Is it feasible to programmatically change the name of a directory in SharePoint?

To quickly summarise my question:
Is it feasible to programmatically change the name of a directory (with both files and sub-folders) in SharePoint? I am expecting that users will have files checked out on at least some occasions what I am attempting the rename.
The background:
I am currently contracting for a company that produces web based software (ASP.NET) with a configurable document management system. The system can be configured to use different underlying systems, with the most common environment being SharePoint (WSS 3).
I have been assigned a task to extend what has to now been a fairly simple system (simply output files into a fixed directory structure, occasionally read). Having never worked with SharePoint before I am doing some research on best practices, and am attempting to work out what is viable. At this stage I do not have access to a testing environment myself, so am limited to reading up online.
One request is to have the directory structure reflect the name (as one example) of the current client - so all documentation for a client will be in one place, and can be accessed externally via SharePoint or other compatible applications. The specification cites that if the name of the client changes then the directory structure should immediately update. My concern is that this will either directly cause errors (eg. Permission denied) or indirectly cause errors (loss of work for users who have externally checked out files).
As a follow up question if there are concerns with the above, is there a better way to implement the above? I have looked at suggesting the users use views to access the structure in SharePoint, however there is a concern from our BA that users will not be able to directly upload new files into this structure.
Thanks
The issue with Folders in SharePoint is that they are not really folders in the way you would expect of a file system. All files in a SiteCollection are stored in one big-assed table on the Database (checkout the AllDocs table).
I cannot categorically say it is safe to rename the folder without doing a bit of testing, I know that the folders "name" is not the key to accessing the document, despite it appearing to be based on the Url you see in the browser.
The best bet is to do a quick test, but I am pretty sure that your plan will not be a problem.
The potential issue is if any Content Query Web Parts etc rely on specific folders to exist or if any other "code" or "pages" look for that folder and not the folderId.
Save the content of the list before you "attempt" it in production. You don't want to loose data.
Checked out documents will still work the way you expect them to.
You may however have to run a crawl again.

Resources