From day to day, admins troubleshoot issues remotely. And, pretty often, they cannot count on another guy who helps them to enable Remote Desktop (RD) on the remote host. There may also be the case when you loose the access to RD on another computer for some reason, and there’s no one in the remote office who can help you. Whatever, I hope you got the point. What do you need to do? Sure, you can just ask a fellow admin to enable RD on the remote host and wait a bit, but what if that’s something really urgent and you are to fix that issue in the middle of the night? Let’s think through what you can do in that case.
Sometimes, guys running home labs do not have licenses for Remote Desktop Services (RDS). Well, that’s not a big deal, you know, because Microsoft provides the 120-day grace period for the platform! However, one day the time runs out and RDS server breaks all the client connections. That day, admins are to choose between reinstalling the server and cheating a bit to reset the 120-day RDS grace period.
In my today’s topic, I discuss why PowerShell behaves like that. Specifically, I shed light on why you cannot run scripts or access a computer on a different domain. Also, I’ll take a closer look at how some cmdlets work.
Today, I’ll talk about a thing that any sysadmin running Hyper-V VMs does (or still dreams about) while managing infrastructure resources: hot modifying assigned to VM memory amount. I’ll discuss not only the feature itself but also how it works on different OS and its impact on the environment stability.
All of us keep an eye on resource consumption within our environments. If a VM needs extra RAM to have the job done, we provide it with some, right? And, we usually run many VMs on our servers each with own purpose and configuration. That’s, actually, why changing the amount of assigned to a VM memory without rebooting it may come in handy. Also, many guys run some parts of their environments on Windows while having other parts run on something from Linux family. Looks pretty hectic in terms of management, doesn’t it?
For quite a long time, System Center Virtual Machine Manager (SCVMM) has a feature called Dynamic Optimization. Its main goal is to automatically rebalance VMs between the participating cluster nodes in case the placement is unequal. Now, this feature has partially became available in Windows Server 2016 in the form of Node Fairness. It balances the workloads among the hosts in a Hyper-V Failover Cluster and automatically live migrates guests from an overloaded node to a less busy one with zero downtime.
Node Fairness goes embedded in Windows Server 2016 and is intended for deployments without SCVMM. SCVMM Dynamic Optimization delivers more versatile functionality than Node Fairness. Regarding this fact, Dynamic Optimization is recommended for balancing workloads among the cluster hosts. However, to use this feature, you need an additional license from the main operating system.
Now that we know what Node Fairness is, let’s take a look at how this service works.
So, we all know about Microsoft’s Storage Spaces Direct (S2D to put it simple) by now. It’s the feature introduced in Microsoft Server 2016 (Datacenter Edition) that pools together server’s storage allowing to build…that’s right: highly available and easily scalable software-defined storage systems. In this article, I’m gonna talk about not as much about its fault-tolerance characteristics themselves, but some hands-on experience, namely: how to replace a failed disk.
Subscribe to my posts
- How to create a bootable USB for Windows Server 2019 installation?
- Where is it smarter to keep databases? Azure SQL vs. Microsoft SQL Server in a VM
- Can SQL Server Failover Cluster Instance run twice as fast as SQL Server Basic Availability Groups in 2-node cluster? Part 3: Comparison time!
- Can SQL Server Failover Cluster Instance run twice as fast as SQL Server Basic Availability Groups in 2-node cluster? Part 2: Studying FCI performance
- Can SQL Server Failover Cluster Instance run twice as fast as SQL Server Basic Availability Groups in 2-node cluster? Part 1: Studying BAG performance