How update a Azure Website via FTP transfer - azure

My requirement is to transfer files like DDL's and config files to the Azure Website from my local system.
I have gone through links and have followed following steps :
Open FTP_Server_name_from_Publish_profile
Supplied User name and password and Able to connect
Cd Site\wwwroot
put "Some_File"
Getting following error:
200 PORT command successful.
150 Opening ASCII mode data connection.
When I am trying to connect via WinSCP client and transferring files via GUI it is working fine.
Also, I tried the above steps for transferring files to the sample FTP servers, and it is working Fine.
Let me know where I am going wrong.

It's not clear what you ask for. I assume you want to automate an update of Azure Website.
As you have succeeded using WinSCP GUI, you can do the same using WinSCP scripting. The WinSCP script would look like:
open ftps://user:mypassword#waws-prod-am1-234.ftp.azurewebsites.windows.net/
put C:\myjob\* /site/wwwroot/App_Data/jobs/type/name/
exit
For details see Automating WebJob Update in WinSCP documentation.

Related

Viewing updates of a dynamically generated log on a remote linux server from a windows machine

I have connected my windows 7 machine via winscp sftp with a cubietruck that runs on debian jessie server.
I need to watch on windows the updated log of an installed app in the linux server.
With winscp I can see this log file. I set the winscp to refresh the remote panel every 10 seconds. I tried to open this log with the following editors:
1) Sublime text with autorefresh plugin
2) Notepad ++ with update silently set
3) Glogg
Unfortunately even though the remote panel was refreshed and watched the log file to grow in size, the file opened in the editors was not updated.
I also tried to set Keep Local Directory up To Date , which creates a replicate file in a local directory in windows. When I had this file opened with the above editors and the remote log file changed then a new file was created on the local disk, without succeeding in watching the updated log.
Does anyone know any solution to this issue or a working alternative ?
You can try klogg. It is a fork of glogg. Glogg opens files in such a way that may prevent other programs from accessing them, that is described here. That has been fixed in klogg(see the issue for details).

Setting up an FTP share folder access from remote desktop

Here is the setup: I have a computer running windows 7 where i connect to a remote desktop. I want to set up a shared folder using FTP so i can copy files from the remote desktop to my local machine. I've created the shared folder using the ISS tool. I set it up so that no credentials are needed (allowed access to all users and anonymous users). I tested it with other computers in the network and it works.
To access the file i want to use NET USE but when i try it from the remote desktop i get asked for a password.
How do I add the user from the remote desktop or what password and username is it expecting.
I am not the admin on the remote desktop, only on the local machine.
I can view the contents from the shared folder if i type ftp://3.xx.xx.x/ . This works on the remote desktop as well.
Please help.... im loosing my mind!!
Thanks
Try using Active mode from the FTP client. Worked for me, I couldn't even get a directory listing before.

Schedule weekly Excel file download to an unique name

We got a database where every Monday a Excel file gets uploaded from a client. The file is always the same name so if we forgot it we lost it. Is there a way how we can make a script that renames the script and gives it the date or a number?
We're using FileZilla to get the files now.
FileZilla does not allow any kind of automation.
You can use this WinSCP script:
open ftp://user:password#host/
get "/path/sheet.xls" "c:\archive\sheet-%TIMESTAMP#yyyy-mm-dd%.xls"
exit
The script connects to the server, downloads the sheet to a local archive under a name like sheet-YYYY-MM-DD.xls.
For details see:
Automate file transfers (or synchronization) to FTP server or SFTP server
Downloading file to timestamped-filename
Then create a task in Windows scheduler to run the winscp.exe every Monday with arguments:
/script="c:\path_to_script\script.txt" /log="c:\path_to_script\script.log"
Logging (/log=...) is optional, but it's recommended.
For details, see Schedule file transfers (or synchronization) to FTP/SFTP server.
(I'm the author of WinSCP)

how to connect and read files of an external Linux server through VB.net programming

I want to connect to a Linux server and read some logs from a folder located in the Linux server through a program written in VB.net.
Is it possible or do I need to have some extra tools or softwares to do the above. (not through Putty or anything similar to it)
Automation Requirements: Go to a folder located in Linux server. Check for a log file. If there are no log files created today then send a mail to me.
I need to write a vb program in VS-2005 to do the above task. Can someone give me a head start or how to go about it.
I am new to Vb programming so any help would be appreciated.
i can neither use SSH protocols nor sockets to do it.For both i would have to install something or the other for which i do not have privileges. Can anyone tell me how can i do it using putty or filezilla tools. I have both available but do not know how to access them through vb programming.
Thanks
You need to find a way to connect linux machine by SSH protocol Please refer this post for your reference executing commands on unix server via visual basic application

linux (CLI) download files via shared dropbox (folder`) link without a account

I was thinking to use dropbox to upload my source code of a web-application. For this folder i would create a shared link. This link i like to use to download all the latest source files on my test server (instead of using s/FTP).
Now i know you can use dropbox with linux by installing their version, but it requires to create account. I don't want to use a account, and for sure don't want to use my own account.
Is there anyway to use a shared (folder) link, and download all the files in that folder command-line, without a account (maybe something like wget) ? There is no need for live-syncing, it would be fine to trigger the download with some bash script.
Thanks.
If you're ok with your links being public (which i think is not a good idea) , then you can just create a file with a list of links to your files and then create a bash script to loop over each line of the file get the link with wget
If you want to use authentication, you'll have to register for a Dropbox API key and then create a script (in python,ruby or java etc) to authenticate and get the files.
If you don't have a specific need for dropbox, i'll recommend you use git (or similar). With git you'll just have to create the repository on your server and clone it on your desktop. Then you can just edit your files and push it to the server.... it's so much easier.
Rogier, github has become the norm for hosting code. There are other options (Sourceforge, Google Code, Beanstalk) or you can set up a private git repository on your own computer.
Somewhere deep in my browser history there's an article about how to do that.
However a little googling turned up http://news.ycombinator.com/item?id=1652414. Let me know if you can't find some satisfactory instructions on your own of how to set up a git repo on your computer.

Resources