I often get asked to do presentations at conferences or user group meetings and although I drive a mean PowerPoint, I feel that showing the actual product and putting it through a few laps adds value and credibility to the production. I’m doing a user group presentation in Montreal next week and I was setting up for it with a colleague of mine. HE asked some interesting questions about why I was setting up my laptop a certain way and I realized that I take for granted that I have been cursed by the demo and presentation gods so many times that I have a few tricks up my sleeve to thwart them. Here are a few of the things I do to minimize the impact of unknown venues with unknown networks:
- Always have a backup of your presentation and and demo VMs. An external drive and/or a cloud drive SkyDrive or Google Drive can be a real saviour when something unexpected happens.
- Always rehearse your demos in the environment you will be presenting in to see how it runs and looks in the venue. Make any changes or restructure the presentation to accommodate for any issues. You don’t want to be surprised on camera.
- Have a backup internet connection avaialbe. A portable hotspot or a mobile phone that has internet connection sharing (like my Nokia Lumia 920)
- Always have a local demo available, even if it’s just a screen recording like Camtasia (full disclosure: TechSmith gives Microsoft MVPs free Camtasia and SnagIt licenses – I also like Faststone Capture since it is inexpensive and also runs as a portable application from a USB key ). Relying on a remote demo is asking for trouble. If you can’t connect to your demo environment for some reason (VPN blocked, network stability, etc.). Also if something goes wrong in the remote location, it is very difficult to troubleshoot.
- If you have multiple systems as part of your demo (virtual or physical) consider using the Sysinternals tool BGINFO or a custom wallpaper with the machine name and/or description to make the different systems readily apparent to the audience (and sometimes to you).
- Make sure your passwords are current and you know what they are. Consider setting demo password properties to “never expire”.
- Set the task bars on your remote demo systems to be in a location other than your primary system so that you don’t get confused as to which task bar you are launching from.
- Explain to the audience the limitations of the demo environment (hardware, data sets, connections to complementary systems, etc.) so that they understand why your demo is designed in a particular way and that it may not be reflective of how a production implementation would work.
I’ve been using Windows 8 and 8.1 since both were in customer preview and I’ve really come to depend on Hyper-V for my demo environment. Before windows 8, I would either boot Server 2008 R2 (or server core) to have a hypervisor available (see my previous blog post about that environment. Before that I would use VMware Workstation or Virtual Box. But they weren’t ideal for every use case as they are type 2 hypervisors not type 1.
I’ve got a few tricks that I use in my demo environment to help build it out and make it present better:
Don’t rely on the Hyper-V Virtual Machine Connection. Enable remote desktop services in your VMs and connect them to an internal network. This allows you to do two things that you cannot do with the Virtual Machine Connection:
- Adjust the screen resolution to meet the needs of the display devices at the venue
- Map local resources like USB drives and printers.
- A cool feature in Windows 8.x and Server 12.x is the ability mount an ISO directly in the OS. Unfortunately, you can’t mount an ISO that is connected through RDP device mapping. You will get the following error:
However, you can mount it in the host OS, it will appear as a DVD drive, and then you access it from the guest VM:
If the demo VM(s) need(s) an internet connection, I like to use ICS to share my wireless connection with my demo VMs. I like this better than the Hyper-V virtual switch bridge because the IP addresses won’t keep changing with the venue. This makes it easier to RDP to them. For step-by-step instruction on how to share a wireless connections try these posts:
I often get asked why I like Hyper-V or why I don’t like VMware. The answer, strangely, isn’t about technology. Anybody that knows me well, knows that I’m not a technology bigot. Meaning I don’t get fanatical about particular companies or pieces of technology. In my house we have six tablets. A Surface RT, a Surface Pro (soon to be replaced by a Pro 2), 3 Android tablets, and an iPad. They all get used on a regular basis. There is no favourite. Just a preference for one device over the other based on the particular use case in question and the strengths of each device at addressing that use case. I’ve used VMware products for years and I like them. They have met many of the requirements I’ve had for a long time.
So how does this relate to Microsoft vs. VMware? Well, I see a lot of fanaticism over VMware. A large percentage IT Pros really love it and many are fanatical about it. They are quick to criticize alternatives (like Hyper-V) without having all of the facts. Another issue is that most people see the results of past consumption and mistake it for current market trends. Let me explain that with an example. Currently Android phones outsell iPhones however, most people see more iPhone sin use that Android phones because iPhones have been around longer have had past sales success. What is being seen is phones that were purchased over the last several years still in use.
Enough digressions. Back to Microsoft and VMware. Historically, VMware has had the edge over Microsoft in the hypervisor market. With Hyper-V 3, most experts would agree that the gap has narrowed enough that for most organizations, the differences are insignificant from a pure technical capabilities perspective. It’s like choosing between a Honda and a Toyota. Both vendors have offerings in every major segment. Most consumers would be equally well served by a Camry or an Accord but preferences still abound. In the virtualization world, there are many other factors to consider such as migration costs, retraining, new licensing, etc. VMware has had very strong technical offerings for a long time and the investments made by many organizations can’t easily be shifted. Of course, historically, there are many examples of a technically superior product being eclipsed (BetaMax vs. VHS, Amiga vs. PC, FLAC vs. MP3). It also isn’t about first or early movers in a market. Consider Blackberry losing 33% market share in 2012 while Android now has nearly 80% market share in the smartphone market. Of course, depending on when you read this the current market share may be very different.
So back to my previous statement “It isn’t about technology”. I’ve shown examples of a superior product losing out as well as examples of an early mover with a dominant market position being eclipsed by a relative newcomer. If not technology, what’s it about then?
Well, I’m an IT Pro. Any IT Pro worth his salt will tell you that the three key elements of a successful IT rollout of any system are People, Process, and Technology. Not necessarily in that order, but all three ingredients are required for success.
As I’ve mentioned previously, VMware has great technology and Microsoft is no slouch either. We can remove people from the equation since both Microsoft and VMware have access to the pretty much the same talent pool and really, the people that matter most aren’t the vendor’s staff but the enterprise customers’ datacenter staff. So a talented VMware administrator could easily be a talented Microsoft administrator. Using the same logic, you might conclude that the processes that are used in enterprise datacenters would also be a wash between VMware and Microsoft implementations and for the most part you’d be right. However I believe Microsoft has an edge. Here’s why:
Microsoft has a long history of supporting cloud/online services that process billions of transactions a year. Consider Hotmail/Outlook.com, XBOX Live, Office 365, Azure, as a few examples with revenue Microsoft has had to develop some fairly robust processes for managing their datacenters. This isn’t new for Microsoft. Consider the ITIL based Microsoft Operations Framework (MOF) currently at version 4.0 has been around since 2000. VMware doesn’t have an online services history to learn the hard lessons of datacenter management or the history of helping customers manage their datacenters from a process perspective. Microsoft has taken the battlefield tested processes they’ve used for over a decade and incorporated many of them into one of the newer and lesser known products in the System Center suite, Service Manager.
Service Manager helps organizations align business processes with technology delivery to create efficiencies in service delivery. The product is tightly integrated with the rest of the system Center suite (especially products like Operations Manager, and Configuration Manager) as well as Active Directory. The rich CMDB provided by Service Manager helps to manage the inevitable VM sprawl that accompanies virtualization. It is also a great platform to bolt on a SAM/ITAM solution like the one from Provance (Full disclosure: Provance is headquartered a few kilometres from my homeand I know many of their staff professionally – We’ve worked on joint projects and I’ve had more than a few drinks with them over the years.).
Until VMware has a similar offering, organizations that want to enable IT Service Management (ITSM) best practices, will find it much easier with a Microsoft private cloud solution than with a VMware solution.
BTW – Market share numbers for last year shows an interesting trend in the hypervisor adoption rates:
Source – Wall street Journal / IDC
Are we in the midst of a Blackberry like decline for VMware?
We keep hearing that it doesn’t matter if your machine is physical or virtual. Your software will still work just fine. That’s true most of the time but there are some exceptions. Monitoring is one of those exceptions. In truth, the monitoring tool will work and will give you accurate information but it will be meaningless.
Remember that in a hypervisor based virtual environment, the guest OS is typically unaware that the hardware has been abstracted and that resource scheduling is taking place to provide shared computing resources based on some preset business rules (some guests may be configured to get more resources than others).
In a scenario like this, a legacy monitoring tool that is targeted at the guest VM, may get false positives concerning resource availability. Typically you might see near 100% CPU consumed. This will be based on the telemetry coming back from the guest VM indicating that it is nearly out of resources, when in reality, there may be more resources available just not committed or allocated to that VM at that point in time.
To get a more realistic and complete view of what is actually happening, the monitoring tool would need to monitor the hypervisor and all of the guests correlating the telemetry form all of them and providing a more holistic view of the availability of resources.
Products like this are starting to emerge. Make sure that when you plan to migrate a server or application to a virtual environment you also plan for the monitoring requirements.