Overview
Direct Answer
A hypervisor is system software that abstracts physical hardware and allocates computing resources to create and manage multiple isolated virtual machines on a single host. It enables multiple operating systems and applications to run concurrently on shared infrastructure whilst maintaining isolation and performance boundaries.
How It Works
The hypervisor intercepts hardware requests from guest operating systems and translates them into instructions executed on the underlying physical processor, memory, and storage. It maintains separate execution contexts and memory spaces for each virtual machine, using techniques such as CPU virtualisation and memory paging to enforce isolation whilst maximising resource utilisation.
Why It Matters
Hypervisors reduce infrastructure costs by consolidating workloads onto fewer physical servers, lower energy consumption, and simplify maintenance operations. They enable organisations to optimise server utilisation rates, improve disaster recovery capabilities, and facilitate workload mobility across physical hosts without application modification.
Common Applications
Data centre operations utilise hypervisors for server consolidation and test environment provisioning. Cloud service providers deploy them as foundational technology for Infrastructure-as-a-Service platforms. Enterprise environments leverage them for legacy application preservation and multi-tenant infrastructure separation.
Key Considerations
Performance overhead from virtualisation can impact latency-sensitive applications, and security boundaries between virtual machines require careful configuration. Licensing costs and the complexity of managing numerous instances must be weighed against consolidation benefits.
More in Cloud Computing
Terraform
Deployment & OperationsAn open-source infrastructure as code tool for building, changing, and versioning infrastructure safely and efficiently.
Content Delivery Network
Architecture PatternsA distributed network of servers that delivers web content to users based on their geographic location.
Cloud Computing
Service ModelsThe delivery of computing services — servers, storage, databases, networking, software — over the internet on demand.
Cloud Orchestration
Service ModelsThe automated arrangement, coordination, and management of complex cloud computing systems and services.
Serverless Computing
Service ModelsA cloud execution model where the provider dynamically allocates resources, charging only for actual compute time used.
gRPC
Architecture PatternsA high-performance remote procedure call framework developed by Google using Protocol Buffers for serialisation.
Cloud Bursting
Strategy & EconomicsA configuration where an application runs in a private cloud and bursts into a public cloud when demand spikes.
AI Infrastructure
Service ModelsThe specialised hardware, software, and networking stack required to train and serve AI models at scale, including GPU clusters, high-bandwidth interconnects, and model serving frameworks.