My customer is currently using MHT files for storing offline representations of browsed web pages. The files are saved and later viewed in Internet Explorer.
When viewing the files we would like to be sure there is absolutely no network activity to the original site or any other site - the content should be browsed 100% offline, and should not have any special "local" privileges as well (i.e. access to file:// protocols etc.). We would like to keep JS running if possible, and we can suffer to consequences of disabled features because of working offline.
We are willing to change the viewer or even the file format (and convert all old mht files as well) if a better solution is suggested.
Thanks for any help on this,
Udi
It is not possible to guarantee that there will be no network activity unless you go to offline mode in Internet explorer. Though the advantage of saving a web page to a mht file is that all the info for displaying the page (including images) are stored in one file instead of several files and folders, making archiving easy, if the web content has links to other pages, clicking on links will initiate network activity.
One option is to post-process the mht file and replace the url links with just the title of the link. For e.g, replacing
<A=20
title=3D"Conduction band"=20
href=3D"http://en.wikipedia.org/wiki/Conduction_band">conduction =
bands</A>
with "Conduction band".
Related
I have some files on S3 and would like to view those files in web. Problem is that the files are not public and I dont want them to be public. Google doc viewer works but condition is, files should be public.
Can I use office web apps to show in browser. Since the files are private, I do not want to store any data on Microsoft servers. It looks like even google doc viewer stores the info while parsing.
What is the cleanest way?
Thanks.
I have looked around for something similiar before and there are some apps you can install locally (CyberDuck, S3 Browser, etc). In the browser has been limited until recently (full disclosure I worked on this project).
S3 LENS - https://www.s3lens.com/
I probably get a minus here, but also Microsoft has an online viewer, which works the same way: the file needs to be publicly accessible.
Here is the link: https://view.officeapps.live.com/op/view.aspx
What I cloud add is that those files need to be publicly accessible only for a short period, i.e. until the page gets opened. So you cloud trick them by uploading the file to be viewed to a public temporary storage in a randomly generated folder and give that url to the online viewer.
Of course this is not that safe, since the file will get as some point to the temp storage and then to Google or Microsoft, but the random path names offer some degree of safety.
I've created recently a small glitch app, which demonstrates what I just explained: https://honeysuckle-eye.glitch.me/
It uploads local files to a temp storage and then opens the viewer from that temp storage; the temp storage only last for one download, so it is pretty safe.
I want to secure static files (images, .txt files) from unauthenticated users. How can I implement the user authentication to the website so that the static files in specific folder also get secured? I have used simple authentication in a login.asp file and started a session for authenticated user and I check the session value for protected .asp files. But I have no idea how to secure static content on Classic ASP website.
The website is hosted on IIS 7 with Integrated pipeline mode.
You already asked this, and I answered it, and I will give you the same answer.
You will need to use BASIC AUTHENTICATION to restrict access on static files in IIS (Classic ASP). Otherwise, you need to save the static content in another format and encrypt it and only make it viewable by people authenticated by your program.
Please don't ask this again, the answers will not be different.
If using Basic Authentification is not your cup of tea, one possibility would be to replace your static files with an ASP file that upon authorization, will output the correct file. If necessary, you can set the ContentType of the Response to the appropriate type. The link http://support2.microsoft.com/kb/173308 show you how to do that with an image stored inside a database but of course, you can take whatever you want as the source of the file. In the case of .TXT files, you can even directly take the file and simply add a small section of ASP code at the beginning for doing the check.
All of this required extra work. There is no way to simply activate some sort of protection with the session state for static files without extra work.
Old question but -- Most MS servers with Classic Asp installed have several default folders which cannot be accessed except via ASP. they are /bin /app_code /app_data and there may be others. It depends on your hosting company. Windows 10 IIS (their cut down dev & test suite) locks these by default. Using ASP code to retrieve and display text and html is very easy but I'm not sure how to do images. If you have very low traffic, one way would be to copy the image file to an unlocked folder and give it a random name, then access it normally in an IMG tag, then delete it after use. (I came here looking for a better method).
Update: The answer to loading images via ASP is here -- displaying images from sql database with classic asp ... see bottom answer by "HeavenCore" and, instead of Response.BinaryWrite rs("ImageBlob"), get the binary of the image into Your variable, eg: BinaryImageData and do Response.BinaryWrite BinaryImageData
I am just starting with Aptana and I don't have the original HTML files for my web site. Is there a way that I can import my whole web site as a project or do I have to open each page from with Aptana and save with the original urls?
Thanks
I use Interachy which is a commercial Mac option. One open source Windows program is HTTrack.
If your site isn't large, it's often feasible to go through page by page and save each one as source. You also need to save all the images and CSS files, and reconstruct the folder directories, though it's goes faster than you might think.
Good luck!
Is there a program that crawls a specified website and will spit out if there is a reference to another website? I have images,video files,pdf's,etc. that I need to give to another developer to finish the port over to their new server.
I just transferred an old site to another person and they are still using my files. I don't know 100% were all the files are and I want to be sure what files I need to give to them. It would be nice to have something like linkchecker that can crawl and if there is a reference to a website root (ex. sub.domain.com) then it will spit out information about it (what page, what is the url).
I don't want to block the site at this point from using the files so that is out.
I'm on a Mac so any terminal program would be just fine.
You could try Sitesucker which can be used to download all the files used on a site (and any it links to depending on the settings). It's OSX (and iPhone) donation-ware so that might be just what you're looking for. I believe it creates a log file of the files it downloads so you could send that if you just want to send the URL's to your colleague instead of the actual files.
You could check out wget. It can recursively (-r option) download a website and save its content to your harddisk. It usually (i.e. if not specified otherwise) downloads everything into directories named like the host.
But be careful not to download the whole internet recursively ;) So be sure to specify correct --domains or --exclude-domains options.
To quickly summarise my question:
Is it feasible to programmatically change the name of a directory (with both files and sub-folders) in SharePoint? I am expecting that users will have files checked out on at least some occasions what I am attempting the rename.
The background:
I am currently contracting for a company that produces web based software (ASP.NET) with a configurable document management system. The system can be configured to use different underlying systems, with the most common environment being SharePoint (WSS 3).
I have been assigned a task to extend what has to now been a fairly simple system (simply output files into a fixed directory structure, occasionally read). Having never worked with SharePoint before I am doing some research on best practices, and am attempting to work out what is viable. At this stage I do not have access to a testing environment myself, so am limited to reading up online.
One request is to have the directory structure reflect the name (as one example) of the current client - so all documentation for a client will be in one place, and can be accessed externally via SharePoint or other compatible applications. The specification cites that if the name of the client changes then the directory structure should immediately update. My concern is that this will either directly cause errors (eg. Permission denied) or indirectly cause errors (loss of work for users who have externally checked out files).
As a follow up question if there are concerns with the above, is there a better way to implement the above? I have looked at suggesting the users use views to access the structure in SharePoint, however there is a concern from our BA that users will not be able to directly upload new files into this structure.
Thanks
The issue with Folders in SharePoint is that they are not really folders in the way you would expect of a file system. All files in a SiteCollection are stored in one big-assed table on the Database (checkout the AllDocs table).
I cannot categorically say it is safe to rename the folder without doing a bit of testing, I know that the folders "name" is not the key to accessing the document, despite it appearing to be based on the Url you see in the browser.
The best bet is to do a quick test, but I am pretty sure that your plan will not be a problem.
The potential issue is if any Content Query Web Parts etc rely on specific folders to exist or if any other "code" or "pages" look for that folder and not the folderId.
Save the content of the list before you "attempt" it in production. You don't want to loose data.
Checked out documents will still work the way you expect them to.
You may however have to run a crawl again.