I updated to Windows 10 Fall and it destroyed my entire virtual network. It removed my previous virtual switch and created a "Default Switch." It would not allow me to create a new External Virtual Switch.
So I wiped the machine, put a fresh install of windows on there, did all the updates, and I could finally create that External Virtual Switch. The only issue is that when that External Virtual Switch exists, it takes my host's download speed a reduces it to like 1 mb a second (on a 800 mb/s connection).
Anyone know how I can fix this issue?
Related
I have a VM running Windows Server 2012 R2 and it is a Standards D3 VM I need to resize it to DS3_V2 VM but I can't see it in the sizing panel VM list. What I can see is bellow. What is the reason for this?
Probably, the virtual machine’s current host cluster hardware doesn’t support the new series, so it’s not listed as an option.
By deallocating the virtual machine (Azure shutdown, not a guest OS
shutdown), all of the possible options in the region are made
available, and the resize process will migrate the virtual machine to
a new computer cluster.
Get more details from this article.
Deallocating also does not gives the above options. In my case, I did deallocate it and then powered on and again deallocated again then it's showing all sizes available.
Long story short, windows 10 is utterly broken on my laptop after it automatically installed some updates. It's now stuck in a loop which always ends up saying "Undoing changes made to your computer"
I can't get into the BIOS.
I can't get into the Windows Recovery Environment.
Been talking to MS support for far too long so far, so I'm wondering if it's possible to add it as a secondary disk drive to another machine that does work, and manually remove any updates that were installed directly through the filesystem?
The only solution MS were willing to offer was to format the whole drive and re-install windows.
When I moved the HDD into a working windows 7 machine it actually prompted a chkdsk to run over the disk.
It found a whole load of orphaned files, not sure if that was really the cause, but after backing up as many files as I had access to, I put the drive back into the other machine and now it boots.
tl;dr, chkdsk fixed it.
I was stuck in this loop last night
machine configuration : Dell Inspiron
windows 10 (original)
What the one thing you can do is to use an application name Dell usb recovery tool. You will have to format you whole computer be it c:// or any other. You will need an extra hard drive to make a backup.
the process goes like this.
You will have to install the above application on other computer and open it and fill your service tag and make that pen drive bootable with that application.
Now plug in that pen drive to the laptop.
Go for troubleshoot.
Repair.
Install new original os.
It will ask for backup make a backup to other HDD.
Install and recover your backup.
When I use it (its source code is available at https://github.com/Zero3K/ERAM) in a Windows 10 64-bit virtual machine (in driver signature enforcement disabled mode) with it configured to use 100 MB for a RAM Disk and is set as a fixed disk via its cpl, I get a FAT_FILE_SYSTEM blue screen that does not happen in a Windows 7 64-bit virtual machine. Maybe someone could look at its source code to see what is causing it and then offer a fix.
I have trouble getting my VMs into the domain.
Several old machines functioned as "servers" in our environment and when I newly started working here, I wanted to change this cluster into something more appropriate. So I calculated how one big server would cost us a certain amount of money but we would save it in electricity-bills etc.
The server finally arrived (HP Proliant) and I installed a Server 2008 R2 on it (boss wont pay for something newer a.k.a. more expensive). I created .vhd files of the HDDs of the servers I want to virtualize and copied those files into the new server. I activated Hyper-V role and created a new VM and selected one of the .vhd files as the HDD for that VM and started the machine. The machine starts, all goody.
As far as I understand in Hyper-V we got 3 network-settings and for my purpose "external network" is the one to go with. Correct me if I'm wrong. The server is already in the domain. So I thought it should actually work already, but somehow I cannot connect to the domain.
The network settings on the VM is DHCP but it is not getting an IP from the DC. Do I need to spoof the old MAC-Address? Do I have to turn the physical machine that I'm virtualizing off? (I didn't do that because I wanted to check first whether I can get both running at the same time before turning it off. Also I cannot access the environment past business hours, so testing this is when nobody is around is kinda out of question)
The VM has W7 running on it, if it matters. I'm kinda lost what to do as I only find how to join a VM into a virtual network when I try to google for solutions.
DC (physical, x.x.x.1 IP), The new server (physical, x.x.x.82 IP), The VM (virtual, DHCP, current address of its physical copy x.x.x.123 though DHCP is set on it, 123 is just the IP it got).
Regards
Edit: Found the solution. The server has 2 NICs. "Connection" and "Connection 2" or whatever it is called by standard in English. The virtual adapter (3rd NIC so to say) has "Connection 2" in its description too. I got confused and addressed the wrong NIC in the Hyper-V options. I am editing this from the VM that finally has connection to the domain.
Has been resolved. Answer is in the edit of original post.
I used to use MAMP (or just a local Apache/PHP/MySQL stack) to work on web projects. I've since graduated to a live Ubuntu server which is much closer to the production environments for the sites I work on.
Now I'm trying to take this a step further to optimize my workflow. My goal is to have a Linux server running in VirtualBox that automounts a local folder share (from the host) and uses a symlink to gain access to the files (i.e. client:/var/www/dev is a symlink to host:/Users/charlie/dev/).
I don't want to keep my files stored on the virtual server if it can be avoided. I prefer having direct local access to the files and not having to wait for buffering issues between the host and the client. i.e., if I have several files that are located on the client open in my IDE and I close my laptop, as soon as I open it there's a bit of a buffer issue. My IDE has open project(s) that reference folders and files located on a network share that isn't yet available. In the few seconds it takes for the virtual machine to wake up, OSX is already reporting that the share can't be found and was disconnected, the IDE chokes up, etc.
So what am I asking? Well, is this safe / are there obvious pitfalls I'm not seeing / better ways to do this?
Edit: For anyone that stumbles upon this post, the final setup is a Linux virtual machine running in VirtualBox on a Mac with NFS and a symlink from my Apache web root to my mount.
I used NFS Manager (http://www.bresink.com/osx/NFSManager.html) to setup the NFS Server on my host computer with user mapping to my primary account. This ensures that when my VM mounts the NFS share it can do whatever it needs (reading, writing, modifying). Then I added this line to /etc/fstab on my VM to automount the share on boot: "123.456.89.1:/Users/charlie/nfs_share /mnt/nfs_share nfs" (where 123 is my host IP on the virtual NAT).
The result is a killer development environment where I can use Finder, Aptana (or whatever your editor of choice is) Photoshop, etc to work on files locally and simultaneously test them out in my "real" Apache/Lighttpd/MySQL/PHP environment!
I am using the exact same setup for accessing my documents folder between my Ubuntu host and the windows guest. Idem on my iMac. The only issues are when editing on the 2 platforms are the CR/LS, but that will be no issue on your setup.