So I am working on a customer and their path of upgrading to 2016 versions. The first step was to make sure that the VMM 2012 R2 server was updated to latest UR and that I can deploy guest vm´s with 2016.
After the update of VMM to UR11 I checked the list of OS,
So to be able to see the 2016 as a guest OS i have to add a hotfix and that took some time but what ever you do, do not cancel but wait and wait and wait and the never ending progress bar will eventually go away 😉 . And yes you have to add one hotfix for the console and one for the vmm server!
Today I was at Microsoft Sweden and did a webinar on Windows Server 2016 Hyper-V and System Center VMM. This was the first of 5 webinars that Microsoft have this week focusing on the highlights on the new release.
The webinar was in Swedish and I will post a link to it when it will become available!
I am a firm believer that Servers should not be used for the wrong things and thus I have now installed the new System Center VMM 2016 on a Windows Server 2016 Core.
In my home lab I do not have so many hosts so I have used the opportunity to install the SQL 2016 on the same core instance.
As I am installing the SQL on the same machine I had to enable the .net 3.5/2.0 feature on this server and yes I know and can´t agree more, please remove this requirement dear SQL team and move to the future!
Although it is not supported with the wizard for sql install on core it do show some progress through a graphical dialog…
So once that was up and running I installed the ADK for windows 10, and I used the one for Windows 10 1607.
And then I could start the VMM install. And yes there is a command line way of installing the VMM but this time I wanted to see if I could use the wizard in core!
During the installation the wizard complained about my memory that I had assigned to the VM that I was installing on and I could with the superduperfeature in 2016 add more to the running VM without doing any stop and start!
After that I had no more issues and the installation completed successfully!
Well once installed I had to do some patching as at the same time VMM 2016 was released Microsoft also announced the availability of CU1 🙂 and trying to use the short cut from the installation dialog fails on a Server core as those GUI parts are not present! I can though use the Sconfig and the “Download and install updates” option to get the updates I want…
Revised: Based on the SQL req page that have been updated it now is supported to run on SQL standard and from SQL 2012 SP2, the following link on the VMM page though still says 2014 Enterprise but that will be updated. My MVP friend Anders Asp have got info that I share here:
“Official MSFT statement: That is likely a carry over from earlier TP content when we had a bug that installation would fail on Std SQL(TP3?). Standard should work.”
//As you can see the System Center VMM 2016 GA will require a SQL 2014 Enterprise or later, so you will not be able to use a standard SQL to be supported. So if you are upgrading from a VMM 2012 R2 you will also have to upgrade your SQL to the Enterprise level.//
The SQL instance solely used for the System Center is included in the System Center licensing.
During Ignite 2016 in Atlanta, Microsoft announced the technical preview 2 of AzureStack and finally now this friday I got my hardware available (the dang server was not responding on the ILO port and I had to go to the datacenter to give it a kung-fu-devops-kick) so I could deploy the new bits.
First things first! Read the documentation about how to proceed and you will more likely succeed in your deployment!
The download for AzureStack is 20 GB so if you have a slow internet connection it will take some time!
Before getting started i suggest you to run the pre-check script that can tell you if there is some immediate issues,
And then you can unpack and follow the instructions to prepare to vhd-boot into the cloudbuilder disk with the next script:
Once rebooted you want to make sure that you only have one nic enabled and then kick of the deployment which will take about 2-3 hours if you have a decent hardware like me 😛
As you can see the install process uses both desired state and powershell direct (which is a lovely feature in Hyper-V 2016)
And if you are patient and then log in as a azurestack\AzureStackAdmin on the physical machine you will see the status of the deployment. Do not log in as a local user on the server and try to start the deployment again!
Hopefully you will end up with the same result as me:
And then you can log in to the VM MAS-CON01 to connect to the portal,
Maybe I was lucky but I believe that the Stack-Team has done some serious work since TP1 and the deployment process have been thoroughly developed, tested and works really good now.
I have been quite busy lately and not had the time to update the blog so much that I wanted but I will try to add some posts during the summer!
In the project I am right now we are setting up environments in Amazons cloud AWS. I have used their images AMI for a SQL Always On cluster that spans over two nodes.
The AMI is preinstalled with SQL and ready for incorporation in an domain and the Enterprise version can be used with the r3.2xlarge r3.4xlarge and r3.8xlarge.
As you can see each of them have an SSD instance storage that can be used as a temporary volume which suits SQL tempdb perfectly. Just go into the SQL configuration and point its tempdb to the temporary storage! The MSSQL service account is a domain account without administrative rights on the server so that is why I explicit set the rights for the volume…
As the AWS documentations clearly tells us:
An instance store provides temporary block-level storage for your instance. This storage is located on disks that are physically attached to the host computer. Instance store is ideal for temporary storage of information that changes frequently, such as buffers, caches, scratch data, and other temporary content
You can specify instance store volumes for an instance only when you launch it. The data in an instance store persists only during the lifetime of its associated instance. If an instance reboots (intentionally or unintentionally), data in the instance store persists. However, data in the instance store is lost under the following circumstances:
The underlying disk drive fails
The instance stops
The instance terminates
So I created a small powershell script that runs each time the instance boots to set ACL´s on that volume for the SQL to be able to create its tempdb files. No tempdb files = no sql service running….
# Script to check and set ACL on the AWS temp drive Z: for SQL Temp
I created a new Windows 10 client in my Mac that I have in an assignment and that on VirtualBox. I made it to small of course and when trying to add some stuff with onedrive sync I could not succeed …
So to be able to expand the underlying disk in virtualbox I had to go into terminal mode as this is not part of the gui which of course is okay, so find your vdi file and possible if you are afraid, do a copy of the file before and of course the VM have to be stopped during the expansion.
Using command VBoxManage modifymedium disk xxxx.vdi –resizebytes 85899345920
and checking the properties on the virtual disk file I can see that it has been expanded
after that you boot the VM and go into windows disk settings and expand the partition and you are ready to add files 🙂
So the day have come when finally the new TP5 bits have been released! And I of course downloaded and wanted to test to upgrade one of my hyper-v servers in my home lab.
Once installed I tried to migrate a VM from the hyper-v manager at the new TP5 node, I had of course set up kerberos and delegation before but still it gave me an error. To see if it was just in the GUI or also in PowerShell I tried the same move and got the same issue
So powershell remoting to the rescue to test that I could live migrate my VM´s from the TP4 to the TP5 and that worked nicely. I will dig some more into if there is an issue with the AD objects or what causes this and do an update if I find anything…
Finally the Azure Site Recovery service can be reached from the new Azure Portal and the ARM way of doing things! It has been possible to use ASR with PowerShell and the new ARM way for some months but only for a subset of the site recovery services (VMM/Hyper-V).
Not a day to soon! I have a customer that we have engaged in the CSP program and as that is based on the new, the old ASR was not possible to use with that subscription and use another subscription just for ASR sucks..
As you can see on the following screendump I go into the “Getting Started” to select scenario and then follow the guide to complete and in the case with physical and VMware I need to install a process/configuration server on-premise.
Once installed on a Windows 2012 R2 server I connect it to the ASR with the registration file,
One thing to think about using this service is that the process server will if you do not go in and configure the bandwidth settings eat all available internet capacity as my customer so nicely explained…
Configure this to something that works for both you and the company, and with the enhanced ASR where you do not need additional servers in Azure you find this setting in the backup properties.
It is quite easy to start and protecting your workloads and remember that the first 30 days are free 🙂
During last week I was working on some bare metal deployment on some Hyper-V hosts with System Center VMM. We had deployed them before using legacy boot but now we had updated the BIOS to latest version and got into some trouble.. Maybe it was because of the HPE instead of the HP 😉
During the deployment the WinPE got an error and could not connect to the VMM server,
We tried to update NIC drivers and stuff on the WinPE image but that did not help. During the testing we started the server and configured it to boot with EFI instead of legacy boot and olala it worked to connect to the VMM server but thus as the Hyper-V VHD was MBR we got the following error:
The easiest way I could think of right there and then was to create my new GPT based VHD to boot the Hyper-V host with a powershell convert script from the original MBR vhd. The script required to be run on a Hyper-V host so I connected to one of the Hyper-V nodes in the test cluster and ran the script on a patched VM that was sysprepped:
And once that was done I had to update the Physical profile to set the disk to GPT instead of MBR: