Editing a webpage with no source - editing

I am a new developer (as in just graduated on the 10th) and was hired by a company to do web development. I was asked to do some minor changes to a site that this company acquired. The problem is that we do not have access to the source code (apparently the people had a bad break up with their previous developers and cannot get the source, I'm not exactly sure). Is there a way I can add links to a site and have it change live? I have Visual Studios, the address, the links, and the videos they will go to, not a hard fix, but I don't know how to edit the site without the source code. Any suggestions? Thanks in advance!

I advise you to talk to a senior or superior and get more information on how to proceed, because getting that code in a less than professional (or legal) way (e.g. using website rippers or something) would be a bad career move ;)
good luck.

interesting situation I should say, the company definetely didnt do its homework before the break-up
I am presuming you answer "yes" for the questions below
Is your company the legal owner of this website?
can you change the name servers or CNames etc
The current website is not Flash or silverlight
if here - you have said "yes" for all the above.
First of all navigate to every page of this website. File save as
each of this page to html(make sure you choose webpage complete -
this will save all the images as well) I realise this will be static, but there is not much you can do here
Get all resources (stylesheets, xsds (if any) , any other images)
Enrich this content based on requirements (i.e. add dynamic content, change logos etc)
Modify the cname or nameserver to point to the location(webserver)
you are in control.
Deploy your enriched and tested code
Educate your company to treat the developers well and when things go wrong, ensure transition is done well
I hope this help and good luck
Krishna

Related

Netsuite Sitebuilder My Reference Checkout 3D secuer

I'm creating a new sitebuilder website, and have successfully set up "My Reference Checkout". However, we haven't been able to get it integrated with 3D secure, meaning that no customers can check out (unless we remove all the security, which we don't want to do).
Looking at netsuite's documentation on SuiteAnswers here, it gives us some very generic pointers about how we have to create some .ss and .js files, and modify some existing front-end code in Custom checkout. However, AFAICT the examples are not useful - they don't tell us where to put these ss & js files, where to modify the front end, and some of the files don't even seen to be related etc.
Does anyone have any better documentation for how to integrate with 3D secure, or please can someone point me in the right direction?
Any help is greatly appreciated!
Thanks!
This is a pretty broad question, and old, so you may have found the answer, but I would recommend reading and searching at this resource: https://devsc.publishpath.com. It uses your netsuite login and password. There is a blog post at this location: http://devsc.publishpath.com/Default.aspx?p=2079436&Add=Show+Post&Key=Show+Post&ContentID=5226425&PostID=1200364&shortcut=developing-your-first-custom-suitecommerce-advanced-module-part-1 that goes through setting up the reference cart if you are using it.
Outside of that here is a brief overview. The SSP and SS pages are setup using a utility in Netsuite under Setup/Site Builder/(Sub-Heading-Web Site Management)SSP applications. Here you set where in the Netsuite file system your SSP and SS pages are. The files themselves should be somewhere in your file structure under your hosting root. They load in order of precedence and have to be deployed. For certain activities (like logging in or checking out) you need to set up touchpoints.

How can I speed up Amazon Affiliate Link creation?

Everyday I need to create 10-15 Amazon Aff links (iframes for a product image+details).
To be able to do this, I need to search for the product on Amazon, then click the product, then click another button, then copy the code, then paste into Wordpress. It quickly gets tedious. Is there a faster way to do this? Any offline apps that could speed the process up?
Thanks
You could check out GeoRiot.
There's a couple of tools to create affiliated links automatically, or you can use the Javascript Snippet in the header of your page, which automatically converts Amazon and iTunes links into affiliated links for you. You'll still have to copy the links and add them to your site, but with the JS snippet, that's all you'll need to do.
Another benefit is the GeoRiot links work for the international Amazon storefronts, so if you have any international users, they'll end up in their local marketplace to purchase the items.
(Full disclosure, I work for GeoRiot so I'm definitely biased here. But I can also answer any other questions that you have about the service.)

Scraping a website where all the data is locked in an XML database?

I am trying to download the full archive files of this website (http://www.afghanislamicpress.com/).
I tried using DeepVacuum (http://www.hexcat.com/deepvacuum/index.html) but the site is dynamic (I think that's the right word).
So you submit a form that gives the article archive, but it only spits out 5 at a time (i.e. per page) and then you have to click through. I want to download all the individual articles for the full data set, but don't want to manually click through.
I know there's some easy way to do this, but not entirely sure how.
Any suggestions for a novice at doing data scraping etc?
The most straightforward solution would be to contact the owner of the website and request their permission to republish their articles, and ask for a digital copy.
You can certainly automate pulling down content that is paged, but it requires some programming effort. The best tool for that imho is HTML Agility Pack.
Please be sure and comply with copyright and licensing terms of the content you are downloading.

SharePoint site space usage

I would like to find out how much space my group's SharePoint site uses (files + version history). However, I only have administrative access to my site, not the entire SharePoint instance, so I have to come up with my own solution. I'm interested in the total, but usage per individual file is also fine.
I've googled everything I could think of but couldn't find much that would help. SharePoint programming seems out of the question since I don't have access to the machine. SharePoint Web Services looked promising but none of the services provided seem to give me what I need. I also found a VBA library that lets me list the versions of a document: Office.DocumentLibraryVersion. However, this type does not include a "size" property - why not?
Anyway, I would be happy with either of the following solutions:
A library or API to be used from VBA, VB, or C# (or any other language, for that matter)
A SharePoint Web Service that provides file size/space usage information
A completely crazy script that uses http to iterate through all the folders/files/versions in the library and does insane pattern matching to figure out the size of each file, then adds them together and returns the grand total (SharePoint du)
I figured SO is the best forum for this question, but a non-programmatic solution is just as welcome. Basically, anything you can come up with would help. At this point, even "this is not possible" would be useful.
Thanks in advance.
There is a hidden page that does this... Cannot find it right now.
Check the 1033 directory and similar to /layouts/usage.aspx.
That page links to /layouts/storman.aspx. Unfortunately that page does not work if your site collection does not have a quota.
Go to Site Settings / Site Usage Report.
If what you are looking for is not there, I don't think you can do it with your level of access.
got to siote actions--> site administration-->site usage reports-->
you can get the site usgae report
if you want to get it in excel chart--> open your site in sharepoint designer-->site-->reports-->usage-->then you can get
usage summary
monthly summary
daily summary
daily page hits
etc

Changing page location after google analytics setup

The current website structure is setup such that all the ASPX pages are in the main folder. It's becoming increasingly difficult to maintain, so I would like to create new folders and move the relevant pages. This would change the URL from say:
http://mydomain.com/DoStuff.aspx
to
http://mydomain.com/DoingFolder/DoStuff.aspx
I fear that this will skew up the google analytics results. Is it recommended I do this change? If so, is there a way to link the page locations of after and before the change?
Also, what would happen when I implement the URL rewrite? Would I run into the same issue again? Anyone?
So in general I think it is a good idea to add the folder for both your users to visually see the section they are in via the URL and to help the search engines figure out the areas and who knows you may even get a (small) SEO benefit out of it.
What I would advise is to setup a second profile in Analytics and then add a filter which removes the folder name from the request and will leave you with the same flat structure in your reports as you have currently. (NB Do this under a new profile with the same tracking code to avoid major mess-ups that you can't undo).
Cheers
Z

Resources