Nano Server…Revealed

Nano Server

Nano Server

At Microsoft Ignite, I had a chance to meet and talk with Jeff Snover…a great honor…and his latest project “Windows Nano Server” is very cool.  Windows Nano Server is designed to be as lightweight and compact as possible. ‘Nano Server is a deeply refactoring version of Windows Server with a small footprint and remotely managed installation, optimised for the cloud and a DevOps workflow,‘ as quoted by Jeffrey Snover, Andrew Mason and Alan Back in a joint blog post. ‘It is designed for fewer patch and update events, faster restarts, better resource utilization and tighter security.

The result: as compared to the equivalent Windows Server build, Nano Server offers a 93 percent reduction in storage requirements, 92 percent fewer critical security bulletins, and 80 percent fewer reboots during operation. This is great for Security and Network Admins….I spent a lot of nights during my times as a network admin around patching and worry about what might blow up and this is a welcome change, especially for a Hyper-V environment 😉  Go Microsoft…
Nano Server
Naturally, those benefits come at a cost. ‘To achieve these benefits, we removed the GUI stack, 32 bit support (WOW64), MSI and a number of default Server Core components,‘ the team explained. ‘There is no local logon or Remote Desktop support. All management is performed remotely via WMI and PowerShell. We are also adding Windows Server Roles and Features using Features on Demand and DISM.‘ Despite this, Nano Server remains API-compatible with other Windows Server variants – meaning it should, in theory, be relatively straightforward to port applications across to the platform.

Also, It ships with the baseline version of .NET called CoreCLR, which Microsoft in recent months made open source. The OS does not contain the binaries or metadata that typically increase the footprint and developers are expected to package applications along with dependencies in a single unit of deployment.
Core PowerShell, a minimalistic version of PowerShell refactored to run on CoreCLR, provides Remote management capabilities and Nano Server can be installed on physical hardware or virtualized infrastructure.

When Windows Server starts supporting Docker, Nano Server stands to become the become the preferred OS to run containers.
Nano Server
Windows Nano Server won’t be for everyone. Microsoft has indicated that it is targeting two prime markets for the new OS: cloud applications, which includes the ability to run multiple languages and runtimes in containers, virtual machines or physical servers; and of course its own Cloud Platform infrastructure, with support for Hyper-V compute clusters and Scale-out File Server storage clusters. It’s in virtualization where the biggest benefits will be found: with each virtual machine requiring only seven percent the storage space of previous Windows Server instances and consuming considerably fewer resources while running, the overhead of running a virtualized infrastructure is considerably lessened.

Flexibility is key to delivering a modern data center, and by using the combination of Nano Server and its new container technology Microsoft is making a big shift away from its previous monolithic server model to one that’s more aligned with the way we deliver cloud-scale services. That does mean that Nano Server won’t be for everyone. Customers are going to have to have made the shift to a DevOps model, and to using cloud-scale data center infrastructure practices of which I am a big believer in and frankly why I work for Nutanix.

Microsoft has not yet offered a release date or licensing information for Windows Nano Server. Beta bits are available via MSDN on the Windows Server Technical Preview 2 media. Instructions can be found here to get started, if you want to check it out.
In conclusion, with Nano Server, Microsoft stands a chance to blow them all out the water with keeping Windows relevant in the era of linux, containers and microservices.

Until next time, Rob…

Azure Stack…What is it?

The Ignite 2015 conference in Chicago is where Microsoft made the official announcement of Azure Stack, its private cloud infrastructure for data centers that want to be Azure in their own right. Or in other words, on-premises will be in full parity with Azure Cloud.
AzureStackW1 Azure Stack
Quotes from Brad Anderson from Keynote on Azure Stack
“If you think about Azure, there’s all the infrastructure that you’re aware of, in network, storage and compute. There’s a set of services like IaaS and PaaS that we deliver. And then there’s all your applications, and that, really, is what Azure is,” explained Brad Anderson, Microsoft’s corporate vice president for cloud and enterprise, during a keynote session Monday morning. “Two years ago, we announced we were going to bring portions of this to your data center, and we called it the Azure Pack.”

Portions of this Azure Pack had made their way onto partner vendors’ hardware in the past — in the form of Microsoft Private Cloud Fast Track Program and Dell’s Cloud Platform System. My company, Nutanix was one of the first Private Cloud Fast Track Partners with certified reference architecture.   So we’ve seen private cloud platforms with third-party vendor brands, built around server software made by Microsoft but not called Windows.

What Azure Stack becomes, over and above Azure Pack, is not just a microcosm of Azure, but an extension of Azure itself. As several Microsoft officials confirmed at Ignite, Azure Stack extends the file and object system of Azure into the private space. (And Azure Stack won’t be the only Microsoft technology that does this….Hint, Hint…Hmm…under NDA at moment)

“You want to be able to take those cloud applications, and host them in your environment,” said Anderson. “You’ve told us you want Azure — all of Azure — in your data centers. Azure Stack … is literally us giving you all of Azure to run in your data centers.

I saw early demonstrations of Azure Stack at Ignite, and what I saw was user access policy management system that essentially duplicated the one currently used on the public Azure cloud as shown below.

“The Microsoft Azure Stack gives application owners the ability to ‘write once, deploy anywhere,’ whether it be to your private cloud, a service provider’s cloud, or the public Azure cloud,” reads a post to Microsoft’s server and cloud blog Monday. “Developers will have the broadest access to application development platforms across Windows and Linux to build, deploy and operate cloud applications using consistent tools, processes and artifacts. One Azure ecosystem across public, private and hosted clouds will allow you to participate in a unified, robust partner network for Azure clouds.”

Microsoft’s idea is to make private cloud space and public space addressable and manageable using the same tool set, and by extension, to effectively make data centers into planks, if you will, for Azure. It’s one big reason why the words “Windows Server” are being spoken less and less often by people whom you would think were in charge of it.

Azure Stack Deeper Dive

Now let’s start at the top. When we look at the image below we see the browser experience. In the current version of Azure Pack we have 2 portals, 1 for the tenant and 1 for the admin. In Azure Stack we have 1 browser experience. That experience is also the same across Azure Stack and Azure. So admins as well as the tenants go through the same portal site and leveraging the same portal API’s and extensions.

In the deployment of the portal site there is still an option to scale out to multiple website nodes like we do with a distributed deployment of Windows Azure Pack today. When we go down that rabbit hole, we see the Azure Resource Manager and the Core Management Resource Providers. The Core Management Resource Providers integrate in Azure Resource Manager and all components interact with that. Below in this post,  I will focus on the Azure Resource Manager and the Core Resource Providers. Further down we see the Service Resource Providers. The Service Resource Providers will control and manage the resources it is assigned to. Like the Compute Service Resource Provider will manage the compute resources (nodes), the Storage Resource Provider will manage the storage resources (nodes) and so on…

And that’s really in a nutshell the top to bottom service layout of the Azure Stack.
AzureStackW2 Azure Stack

Let’s look at the portal. The portal is completely redesigned and which allow you to fully personalize. It is highly scalable and have integration across services. When you install new resource providers today in WAP you need to edit the core code for the Azure Pack portal. Then you need to restart the web service process to see the result of that change. With the new design the portal process runs continuously in a separate process and when you extend the portal by adding extensions a workflow will distribute the extensions to all nodes running the portal site. As mentioned before the admin and tenant site are integrated in the same portal.
AzureStackW3 Azure Stack
The portal UI is very nice, but it would be useless if we cannot login to the portal, right? Let me talk about the identity part of the new Azure Stack. In the old portal we had the options to use the SQL .Net membership or we could integrate ADFS to use AD or other federated identity providers (IDP’s). In the new portal they use claims-based authentication and there is native support for the following:

  • Azure Active Directory
  • Windows AD
  • Active Directory Federation Services (ADFS)

From the Azure Resource Manager to the Core Management Resource Providers it will use Windows Authentication or Basic Authentication. The Core Management Resource Providers will use Windows Authentication or an authentication method defined by the Resource Provider.
AzureStackW4 Azure Stack
Now on to the Azure Resource Manager. The Azure Resource Manager is the new Service Management API. It’s as Microsoft calls it “a product” that allows the management of the compute, storage, network. When you, as a tenant, create a resource group it allows you to put all the resources (VM’s, Networks, websites etc…) in a resource group that can be managed as a whole (Create /Add / Update /Delete – aka Life Cycle Management).

With role based access control (RBAC) you, as a tenant, can also provide access to other users that have access based on the permission you assign to the resource group. Also usage is collected for a particular resource group so you can see how much the resources in a resource group will cost.

The Azure Resource Manager will also allow you to put deployments in regions. Regions represents the datacenters of your service provider or your own datacenters. Furthermore the Azure Resource Manager is providing audit logging on your subscriptions and resources. To create resources using the Azure Resource Manager you need to create or use an existing template. A template is a json file what can be edited to define the resources in your deployment.
AzureStackW5 Azure Stack
The Azure Resource Manager will talk to the Core Management services. Let’s look at the components involved in that.

  • The Authorization Service: By using RBAC, it allows us to granular assign permissions to resource groups. Subscriptions are assigned to tenants that have a plan defined.
  • The Subscription Management Service is responsible for managing the Service Plans, Offers and subscriptions. You can even use Azure Resource Manager templates to define new subscriptions based on a template you have defined.
  • The Gallery Service is a core common service that will work across any of the connected services. Admins as well as tenants are allowed to put their own gallery items in it.
  • The Events Service is a collector to collect all events across all the services
  • The Monitoring Service collects metrics from all services.
  • And last but not least we have the Usage Service which will collect the usage per service for each tenant / resource group.

AzureStackW6 Azure Stack

So this what I know so far from MS, but will update this post as I know more. MS is not giving defiant answer, but rumors are beta late fall and Tech Preview in spring. I can’t wait to get the early bird bits to play around with it and when I do I will follow-up on this post to give you more technical information of Azure Stack!

Until next time, Rob…

Microsoft Ignite 2015 – Week in Review…

Gallery

Microsoft Ignite 2015 – Week in Review Well, it was a great show and finally recovering with some sleep under my belt.  Here we go…  Microsoft Ignite 2015 is where I had a chance to meet and talk with the industry leaders … Continue reading

VHD vs. VHDX: Hmm….What’s the Difference

I get asked this all the time….and here is my answer. Microsoft’s virtual hard disk format, VHDX, has some important advantages over the legacy VHD format, as this VHD vs. VHDX comparison shows.

The release of Windows Server 2012 brought many new virtualization improvements, but one that caught the eye of many IT pros was the introduction of the VHDX file format. Windows Server 2012 supports the new format but also lets Hyper-V administrators use the legacy VHD format. With two virtual hard disk formats to choose from, let’s take a minute to talk about VHD vs. VHDX.

Continue reading

Azure ExpressRoute….Explained….

One of the traditional impediments to businesses adopting public cloud computing is the concern over putting all your eggs in one basket. This is where Hybrid Cloud steps in.

The Hybrid Cloud is a description of utilizing a preexisting on-premises datacenter and a cloud solution such as Microsoft Azure to balance the overall solution. Hybrid Cloud requires the need for a constant reliable connection to your datacenter.

Building a Hybrid Cloud and having a customer put this over the shared Public internet connection gives you no SLA or reliability.

Continue reading