Concepts
Cloud Temple's IaaS (Infrastructure as a Service) offering is designed to meet the critical needs of business continuity and disaster recovery, with a particular focus on demanding sectors such as industry, banking, and insurance. Built on cutting-edge technologies, this infrastructure ensures maximum availability and optimal performance for your critical workloads.
A Trusted Technology Platform
Cloud Temple's IaaS platform is built on internationally recognized technology partners:
- Compute: CISCO UCS.
- Storage: IBM Spectrum Virtualize, IBM FlashSystem for block storage.
- Networking: JUNIPER.
- Virtualization: VMware, providing a reliable and proven foundation for managing your cloud environments.
- Backup: IBM Spectrum Protect Plus, for backup orchestration and storage.
This architecture is based on the VersaStack model, a collaboration between Cisco and IBM, ensuring broad compatibility with leading software vendors.
A dedicated and automated infrastructure
Although fully automated through APIs and a Terraform provider, Cloud Temple's IaaS offering provides a unique infrastructure:
- Dedicated resources: Compute blades, storage volumes, and software stacks (virtualization, backup, firewalling, etc.) are never shared among clients.
- Maximum predictability: You control virtualization rates, IOPS pressure on storage, and benefit from clear, consumption-based monthly billing.
The platform is certified SecNumCloud by the ANSSI, ensuring a high level of automation and security.
Key Features
- Dedicated and on-demand computing resources (CPU, RAM).
- On-demand storage (multiple classes available).
- Networking resources (Internet, private networks).
- Cross-backups with configurable retention.
- Asynchronous replication for storage or virtual machines.
- Management via the Console or in Infrastructure as Code mode using APIs and the Terraform provider.
Benefits
| Benefit | Description |
|---|---|
| Digital Trust | Data hosting in France and GDPR compliance. |
| Security | Highly secure platform, certified SecNumCloud, HDS (Health Data Hosting), ISO 27001, and ISAE 3402 Type II. |
| High Availability | Platform availability rate of 99.99%, measured monthly, including maintenance windows. |
| Resilience | Business continuity and disaster recovery plans implemented as needed. |
| Automation | Fully automated platform designed to integrate into a digital transformation program. |
| On Demand | Resources available on demand. |
Regions and Availability Zones
The VMware IaaS product is deployed within an availability zone.
An availability zone is part of a region.
This deployment model allows you to choose the location of clusters and enables their distribution across different availability zones (AZ).
This provides better load distribution, maximizes redundancy, and simplifies the implementation of a disaster recovery plan (DRP) in the event of an incident.
Compute
The blades provided by Cloud Temple are of type CISCO UCS B200 or CISCO UCS X210c. They are fully managed by Cloud Temple (firmware, OS version, etc.) through the Cloud Temple console.
Several categories of compute blades are available in the catalog to support your workloads (virtualization, containerization, etc.). These blades feature different characteristics and performance levels to best meet your needs. The compute blade catalog is regularly updated.
When using with a virtualization offering, a hypervisor cluster must consist exclusively of compute blades of the same type (mixing different blade types within a single cluster is not allowed).
| Reference | RAM (1) | Frequency (2) | Number of Cores / Threads | Connectivity (3) | GPU (4) | SKU for VMware Offering |
|---|---|---|---|---|---|---|
| ECO v3 Blade | 384 GB | 2.20/3.0 GHz (Silver 4114 or equivalent) | 20 / 40 threads | 2 × 10 Gbit/s | csp:fr1:iaas:vmware:eco:v3 | |
| STANDARD v3 Blade | 384 GB | 2.40/3.4 GHz (Silver 4314 or equivalent) | 32 / 64 threads | 2 × 25 Gbit/s | csp:fr1:iaas:vmware:standard:v3 | |
| ADVANCE v3 Blade | 768 GB | 2.80/3.5 GHz (Gold 6342 or equivalent) | 48 / 96 threads | 2 × 25 Gbit/s | csp:fr1:iaas:vmware:advance:v3 | |
| PERFORMANCE 1 v3 Blade | 384 GB | 3.20/3.6 GHz (Xeon E-53I5Y or equivalent) | 16 / 32 threads | 2 × 25 Gbit/s | csp:fr1:iaas:vmware:perf1:v3 | |
| PERFORMANCE 2 v3 Blade | 768 GB | 3.00/3.6 GHz (Gold 6354 or equivalent) | 36 / 72 threads | 2 × 25 Gbit/s | csp:fr1:iaas:vmware:perf2:v3 | |
| PERFORMANCE 3 v3 Blade | 1536 GB | 2.60/3.5 GHz (Gold 6348 or equivalent) | 56 / 112 threads | 2 × 25 Gbit/s | csp:fr1:iaas:vmware:perf3:v3 | |
| PERFORMANCE 4 v3 Blade | 512 GB | 2.50/4.1 GHz (Intel 6426Y or equivalent) | 32 / 64 threads | 2 × 25 Gbit/s | 2 × NVIDIA L40S 48GB | csp:fr1:iaas:vmware:perf4:v3 |
Notes:
-
(1) The memory amount delivered corresponds to the physical memory available on the blades. It is not possible to change the physical memory capacity of a blade.
-
(2) Minimum base frequency / turbo frequency, expressed in GHz. By default, processors are configured in the BIOS for maximum performance.
-
(3) Physical connectivity is shared between network and block storage access, as the underlying Cisco platform is converged.
-
(4) The available GPU offering is continuously evolving. As of May 1, 2024, the offering is based on NVIDIA LOVELACE L40S GPUs. By default, the PERF4 blade is delivered with two 48 GB L40S GPUs. Contact support for further details if needed.
The compute offering availability is 99.99%, calculated monthly, including maintenance windows. Eligibility for SLA non-compliance claims requires the creation of an incident ticket. You must also have at least two hosts per cluster and enable the High Availability (HA) feature.
This feature allows your architecture to automatically restart your virtual machines on the second hypervisor in case of failure. If a zone of availability contains only one hypervisor, automatic restart is not possible.
Network
The networking service on Cloud Temple's IaaS platform is built on a VPLS-based network infrastructure, providing flexible and high-performance segmentation to meet client requirements for connectivity and network isolation.
Layer 2 VLANs
The VLANs provided in the IaaS offering are layer 2 types, delivering full network isolation and configurable adaptability according to your needs.
Key Concepts
- Cross-cluster and Availability Zone (AZ) sharing:
- VLANs can be shared across different AZs and clusters belonging to the same tenant.
- Cross-tenant propagation:
- VLANs can be propagated across multiple tenants within the same organization, enabling internal communications.
Network Performance
The network infrastructure ensures low latency for optimal performance:
- Intra-AZ latency: Less than 3 ms.
- Inter-AZ latency: Less than 5 ms.
Key Points
- Flexibility: VLANs can be configured to adapt to multi-cluster and multi-tenant environments.
- High performance: Minimal latency ensures fast and efficient communication between availability zones.
- Isolation and security: Layer 2 VLANs provide strict network segmentation to protect data and traffic.
Block Storage
Cloud Temple offers several classes of block storage based on IBM FlashSystem and IBM SVC technology.
The IBM SVC technology enables delivering the required performance levels for our clients' environments, thanks to the large amount of embedded memory cache in the controllers and the ability to distribute a server's total IOPS across multiple SANs.
It is also used to enable replication of your block storage LUNs across availability zones or to facilitate maintenance operations on storage arrays.
The storage is primarily NVMe flash-based, dedicated to professional workloads. The drives are used by the storage arrays in 'Distributed RAID 6'.
Storage Block Security and Encryption
To ensure the confidentiality of your data at rest, our entire block storage infrastructure integrates a robust hardware-based encryption solution.
- Encryption Type: Data is encrypted directly on the disks (
Data At Rest) using the XTS-AES 256 algorithm. - Compliance: This encryption method complies with the FIPS 140-2 standard, ensuring a high level of validated security.
- Operation: Encryption is applied at the time data is written to the physical storage medium.
It is important to note that this encryption protects data stored on disks. It is not active "on-the-fly," meaning data is not encrypted during storage replication operations between availability zones. Security of transfers is ensured through dedicated, secure communication channels.
The 'Mass Storage' storage class offers mechanical disks for archival needs in an economically efficient context. Several performance tiers are available:
| Reference | Unit | SKU |
|---|---|---|
| FLASH - Essential - 500 IOPS/To | 1 GiB | csp:(region):iaas:storage:bloc:live:v1 |
| FLASH - Standard - 1500 IOPS/To | 1 GiB | csp:(region):iaas:storage:bloc:medium:v1 |
| FLASH - Premium - 3000 IOPS/To | 1 GiB | csp:(region):iaas:storage:bloc:premium:v1 |
| FLASH - Enterprise - 7500 IOPS/To | 1 GiB | csp:(region):iaas:storage:bloc:enterprise:v1 |
| FLASH - Ultra - 15000 IOPS/To | 1 GiB | csp:(region):iaas:storage:bloc:ultra:v1 |
| MASS STORAGE - Archival | 1 TiB | csp:(region):iaas:storage:bloc:mass:v1 |
Note :
- Actual performance for a storage class is tied to the volume actually ordered, based on the "IOPS/To" metric, meaning "maximum IOPS per allocated terabyte",
Thus, a 0.5 TiB volume in the 'Standard' performance class will have an IOPS limit capped at 750 IOPS, Similarly, a 10 TiB volume in the 'Ultra' performance class will have an IOPS limit of 150,000 IOPS,
- The IOPS limit is applied per volume, thus per Datastore in a VMware environment,
- Storage availability is 99.99% measured monthly, including maintenance windows,
- There are no restrictions or quotas on read or write operations,
- There is no billing based on IOPS,
- No performance commitment is provided for the 'Mass Storage' class,
- The minimum size for a storage LUN is 500 GiB,
- When using storage replication, performance must be identical across both availability zones,
- No intelligent optimization mechanisms such as compression or deduplication are implemented: when you reserve 10 TiB of storage, you physically have 10 TiB of usable storage deployed on IBM machines.
- Storage LUNs are dedicated to the client environment.
Usage within the VMware Compute Offering
Within the context of using block storage in the form of a datastore in Cloud Temple’s VMware compute offering, you must take several important considerations into account:
-
Swap file (.VSWP) creation during VM startup: When a virtual machine starts up, it creates a .VSWP file equal in size to its allocated memory on disk. Therefore, to successfully start your virtual machines, you must always have free space in your datastore equivalent to the total memory size of all your virtual machines. For example, if your datastore has 1 TiB of block storage and you want to start 10 virtual machines each with 64 GiB of memory, 640 GiB of disk space will be required. Without sufficient free space, VM startup will be limited by the available capacity to create swap files.
-
Free space required for backup snapshots: The backup service uses instant snapshots. You must therefore always have sufficient free space to allow the creation of a snapshot during a VM backup. The size of the snapshot depends on the amount of write activity from the virtual machine and the time required to perform the backup. Generally, it is recommended to maintain at least 10% free space for this operation.
-
Management of dynamic disks: Exercise caution when using dynamic disks. If you do not properly manage their growth, a lack of physical space can result in the virtual machine freezing (in the best case), or crashing with data corruption (in the worst case). It is crucial to closely monitor available space on your datastores when using this type of disk.
Proactive disk space management is essential to ensure the proper operation of your virtual machines and the reliability of backups. Always ensure you have sufficient space available for swap files, snapshots, and dynamic disk growth.
Backup Storage
The dedicated storage for backing up your virtual machines is automatically provisioned by the platform, up to the quota ordered.
| Reference | Unit | SKU |
|---|---|---|
| MASS STORAGE - Archiving | 1 TiB | csp:(region):iaas:storage:bloc:backup:v1 |
Block Storage Replication
Principles
To enable the implementation of your business continuity plans, when it is not possible to maintain business continuity using application-level mechanisms and virtual machine replication is not suitable, Cloud Temple offers block-level storage replication mechanisms between availability zones within a region.
These replication mechanisms are applied to the storage LUNs of your environments, complementing backup solutions. The decision to use a replication mechanism for a given environment depends on multiple factors, including its criticality, acceptable data loss tolerance, and performance requirements for the application.
Cloud Temple provides two types of replication mechanisms deployed in an active/passive configuration:
-
Asynchronous replication (or 'Global Mirror'): The 'Global Mirror' function provides an asynchronous copy process. When a host writes to the primary volume, the confirmation of the I/O completion is received before the write operation finishes on the secondary volume. If a failover operation is initiated, the application must recover and apply all updates that were not confirmed on the secondary volume. If I/O operations on the primary volume are paused briefly, the secondary volume can become an exact match of the primary volume. This function is comparable to a continuous backup process in which the latest updates are always missing. When using Global Mirror for disaster recovery purposes, you must consider how you intend to handle these missing updates.
-
Synchronous replication (or 'Metro Mirror'): The 'Metro Mirror' function is a type of remote copy that creates a synchronous copy of data from a primary volume to a secondary volume. With synchronous copies, host applications write to the primary volume but do not receive confirmation that the write operation is complete until the data has been written to the secondary volume. This ensures that both volumes contain identical data when the copy operation completes. After the initial copy operation finishes, the Metro Mirror function maintains a fully synchronized copy of the source data at the target site at all times. As of January 1, 2024, the 'Metro Mirror' function is no longer available for sale.
An "active" or "primary" site and a "passive" or "standby" site are then defined. The business continuity plan is activated in the event of a disaster or during a PRA test. The passive site then takes over from the active site.
Asynchronous Replication
When your workloads require short recovery times and application-level replication or virtual machine replication mechanisms are not acceptable or suitable, it is possible to replicate a SAN storage LUN between two Availability Zones within the same region.
This solution provides a RPO of 15 minutes and an RTO under 4 hours. It enables much faster recovery than implementing a backup restoration process.
In an asynchronously replicated storage volume (Global Mirror), the SAN virtualization controllers from both Availability Zones collaborate to perform write operations across both sites. The primary site does not wait for write acknowledgment from the secondary site.
The steps of a write operation are as follows:
- An hypervisor wishes to perform a write operation on a Global-Mirror volume: it sends its request to the SAN controller in its Availability Zone,
- The local SAN controller requests the remote SAN controller to perform the write operation,
- The local SAN controller does not wait for acknowledgment from the remote SAN and proceeds to perform the write locally,
- It then sends the acknowledgment back to the hypervisor that issued the request,
- Hypervisors at the remote site do not directly access the Global Mirror LUN: a service request is required.
| SLA | Description |
|---|---|
| RPO 15 min | In the event of a disaster at the primary datacenter, the maximum amount of data lost corresponds to at most the last 15 minutes of writes. |
| RTO < 4 h | In the event of a disaster at the primary datacenter, business continuity is guaranteed within 4 hours, depending on environment complexity. |
In the event of a PRA activation, the Cloud Temple team performs a presentation operation of the 'Global Mirror' LUN to the remote hypervisors, enabling them to access the data. Therefore, this solution requires reserving compute resources and RAM at the 'standby' site to resume operations in case of disaster.
Using this technology also requires doubling the disk space: the remote site must have exactly the same disk capacity as the local site.
This mechanism may impact application performance by up to 10%. Only storage classes 500 IOPS/To, 1500 IOPS/To, and 3000 IOPS/To are compatible.
| Reference | Unit | SKU |
|---|---|---|
| STORAGE - Global Replication SAN | 1 TiB | csp:(region):iaas:storage:licence:globalmirror:v1 |
Note:
- Since this offering is asynchronous, there is a possibility that some disk operations were not written to the remote site during a disaster. There may therefore be a risk to data consistency, which should be mitigated in the risk assessment of the business continuity plan.
- Block-level storage replication is transparent to virtual machines and applications.
- For this reason, it is important to prioritize application-level replication scenarios, or alternatively virtual machine replication.
- Compute and memory resources at the recovery site may be reduced to optimize costs, provided that a degraded operational state is acceptable for the business during the execution of the business continuity plan.
VMware Virtualization
The VMware Cloud Temple offering qualified SecNumCloud is based on the VMware vSphere technology.
The platform is managed automatically by Cloud Temple (security compliance maintenance, operational readiness maintenance, etc.).
It can be controlled via the Console's graphical interface or through the associated APIs.
Note: For security reasons related to the SecNumCloud qualification,
the customer is not permitted to access the VMware virtualization platform directly (no direct access to vCenter, for example).
Indeed, the SecNumCloud qualification requires complete segregation between the technical assets' management interfaces and the customer's interface (the Console).
- The implemented products are VMware ESXi, VMware vCenter, and VMware Replication.
- The virtualization offering's network does not use VMware NSX technology, but is instead managed physically using Juniper technology and the VPLS protocol.
- The storage does not use VMware vSAN technology, but exclusively IBM SANs over 32G Fiber Channel.
- No hidden optimization techniques are applied (compression, deduplication, etc.).
Definition of a Compute Blade Cluster ('Cpool')
The 'Cpool' is a grouping of VMware ESXi hypervisors, also known as an 'ESX cluster'.
All hosts within a __'Cpool'* belong to the same tenant and the same availability zone (AZ). They must necessarily have the same class:
It is not possible to mix different types of compute blades within the same cluster.
Since all compute blades are delivered with the maximum physical memory, a software-level RAM usage limit is applied at the cluster level to ensure it matches the billed RAM.
By default, each compute blade has 128 GB of memory enabled within the __'Cpool'*.
A __'Cpool'* can contain a maximum of 32 hypervisors. Beyond this limit, a second cluster must be created.
Storage can be shared among __'Cpool'* clusters.
Memory Allocation for a 'Cpool'
RAM reservation is configurable per cluster. You can reduce or increase the amount of RAM to match your cluster-wide requirements.
Be careful not to exceed an average memory utilization of 85% per compute node.
Indeed, VMware's technology uses a compression-based optimization method that can significantly impact the performance of your workloads and complicate diagnostics.
Similarly, excessive memory pressure on your compute nodes will force the hypervisor to offload part of its memory to disk to meet the needs of virtual machines.
This situation, known as 'Ballooning', severely impacts the performance of all virtual machines located on the affected datastore.
Diagnosis becomes complex in this context, as your monitoring will show performance impacts at the CPU level rather than at the memory or storage level.
Also keep in mind that the first action the hypervisor performs when starting a virtual machine is to create a memory swap file (.vswap) on disk, sized exactly to the amount of memory assigned to the virtual machine. You must account for this when sizing your storage.
Each compute node is delivered with 128 GB of memory enabled at the 'Cpool' level, but physically has access to the full amount of allocatable memory.
For example, in a cluster of three hosts of type vmware:standard:v2, the RAM reservation upon activation of the 'Cpool' will be 3 × 128 GB = 384 GB of RAM.
You can extend this up to a maximum of 3 × 384 GB = 1,152 GB of memory.
Minimum memory for a 'Cpool' = number of hosts × 128 GB of memory
Maximum memory for a 'Cpool' = number of hosts × physical memory capacity per compute node
Cloud Temple Virtual Machine Catalogs
Cloud Temple provides you with a catalog of Templates regularly enriched and updated by our teams.
To date, it includes dozens of Templates and images ready to be deployed on your virtual machines.
Hypervisor Updates
Cloud Temple regularly provides updated builds for hypervisors to ensure security patches are applied.
However, updating hypervisors remains your responsibility, as we do not have visibility into your business constraints.
The update process is fully automated. To ensure service continuity, a minimum of two hypervisors is required per cluster during the update. Make sure you have the necessary permissions to perform these actions.
VM Affinity Management
Affinity and anti-affinity rules allow you to control the placement of virtual machines across your hypervisors.
They can be used to manage resource utilization within your 'Cpool'.
For example, they help balance workloads across servers or isolate resource-intensive workloads.
In a VMware 'Cpool', these rules are commonly used to manage virtual machine behavior during vMotion.
vMotion enables moving virtual machines from one host to another without service interruption.
You can configure the following rules through rule management:
-
Affinity Rules: These rules ensure that certain virtual machines run on the same physical host.
They are used to improve performance by keeping virtual machines that frequently communicate together on the same server, thus reducing network latency. Affinity rules are particularly useful in scenarios where performance is critical, such as databases or applications requiring fast inter-server communication. -
Anti-Affinity Rules: Conversely, these rules ensure that certain virtual machines do not run on the same physical host.
They are essential for availability and resilience—for example, to prevent all critical machines from being affected in the event of a single host failure. Anti-affinity rules are crucial for high-availability applications, such as in production environments where fault tolerance is a priority.
For instance, you wouldn’t want both of your Active Directory servers located on the same hypervisor.
When creating a rule, you define the rule type (affinity / anti-affinity), the rule name, its activation status ('Status'), and the virtual machines involved within your hypervisor cluster.
Note: The affinity/anti-affinity rules available in the console are rules between virtual machines (not between hypervisors and virtual machines).
Asynchronous Replication of Your Virtual Machines in VMware Environment
Asynchronous replication of your virtual machines is a mechanism that pushes write operations from the source hypervisor to the standby site at regular intervals.
After an initial hot copy of all active storage to the standby site, only write operations are pushed at regular intervals to the dormant site.
The interval depends on the volume of writes (ranging from every hour to every 24 hours).
VM replication relies on the hypervisor’s snapshot mechanism. As such, this solution shares the same drawbacks, particularly sensitivity to the VM’s write volume, since the snapshot process is recursive and requires closing the snapshot.
A typical example of a machine that does not support VM replication is an FTP server receiving real-time video streams from surveillance cameras. The machine is constantly writing data and will not be able to close a snapshot without pausing the operating system for a significant period (several tens of minutes). If the hypervisor fails to close the snapshot, it will proceed to do so anyway—without any possibility to intervene, except by corrupting the virtual machine.
| SLA | Description |
|---|---|
| RPO of 1H to 24H | In the event of a disaster at the primary datacenter, the maximum amount of data lost corresponds to the last write push to the standby site. |
| RTO < 15 min | Virtual machine startup operation on the remote site after stopping the VM at the primary site. |
In case of need or failure on a machine at the primary site, the mirrored machine at the standby site is activated.
Recovery requires reserving compute and RAM capacity at the standby site. It is also necessary to have the same storage space available at the passive site as at the active site.
| Reference | Unit | SKU |
|---|---|---|
| PRA - VMware Inter-AZ Replication | 1 VM | csp:(region):iaas:vmware:licence:replication:v1 |
Note: The minimum RPO must be defined based on the rate of change on the virtual machine.
Virtual Machine Backup
Cloud Temple offers a native, non-optional cross-architecture backup solution (mandatory for French secnumcloud certification).
Backups are stored in a different availability zone and on a physically distinct datacenter from the one hosting the virtual machine. They are encrypted using the AES 256-bit symmetric key algorithm (cipher mode xts-plain64) to ensure data confidentiality.
This setup protects against major failures in the production datacenter and enables restoration on a secondary datacenter (e.g., in case of fire).
This solution includes:
- Hot off-site backup of all disks,
- Instant presentation and boot of a virtual machine from the mass storage infrastructure, followed by hot reloading onto production SANs,
- Partial file restoration from backups,
- Retention limited solely by allocated mass storage space.
The backup infrastructure is based on IBM Spectrum Protect Plus, a no-agent architecture solution that is easy to use, enables automation of backup processes, and optimizes mass storage usage.
Backup and restore speeds depend on the change rate within the environments.
Backup policies are configurable per virtual machine via the Cloud Temple Console.
Note:
Some virtual machines are incompatible with this backup technology, which relies on the hypervisor’s snapshot mechanisms. These typically include VMs with constant disk write workloads. The hypervisor cannot close the snapshot, which forces the virtual machine to be frozen to complete the closure operation. This freeze can last several hours and cannot be stopped.
In such cases, the recommended approach is to exclude the disk subject to continuous writes and back up the data using an alternative method.
| Reference | Unit | SKU |
|---|---|---|
| BACKUP - Access to IBM Spectrum Protect Plus service | 1 VM | csp:(region):iaas:backup:vm:v1 |
Create a backup policy
To create a new backup policy, you must submit a request to support. Support is accessible via the buoy icon in the top-right corner of the window.
Creating a new backup policy is done through a service request specifying:
Your Organization's name A contact name, email address, and phone number to finalize the configuration The tenant name The backup policy name The retention characteristics (x days, y weeks, z months, ...)
Advanced Data Protection (HSM/KMS)
Cloud Temple offers an advanced virtual machine encryption solution based on Hardware Security Modules (HSM) and a Key Management Service (KMS). This feature enhances the protection of sensitive data through centralized and secure key management, seamlessly integrated into the SecNumCloud environment.
Technical Architecture
The solution is built on a robust security infrastructure composed of:
- HSM (Hardware Security Module) : Thales Luna S790 modules certified FIPS 140-3 Level 3
- KMS (Key Management System) : Thales CipherTrust Manager for centralized key management
- VMware Integration : Communication via the KMIP (Key Management Interoperability Protocol)
High Availability Deployment
The HSM infrastructure is deployed across three availability zones in the FR1 region:
- PAR7S
- TH3S
- AZ07
This distribution ensures maximum high availability and resilience of the encryption service.
Key Hierarchy and Operation
The system uses a cryptographic key hierarchy to ensure data security:
| Level | Key Type | Description | Location |
|---|---|---|---|
| 1 | Root of Trust (RoT) | Master key managed by KMS | HSM Luna |
| 2 | Domain Key (DK) | Domain key per client (multi-tenant isolation) | HSM Luna |
| 3 | Key Encryption Key (KEK) | Key encryption key per VM | CipherTrust Manager |
| 4 | Data Encryption Key (DEK) | Data encryption key per VM | VMware ESXi |
Encryption Process
- Generation: VMware ESXi generates a unique DEK for each virtual machine
- Protection: The DEK is encrypted by the KEK stored in CipherTrust Manager
- Securing: The KEK is itself protected by the HSM key hierarchy
- Storage: The encrypted DEK is stored alongside the VM's configuration files
Security and Compliance
Certifications
- FIPS 140-3 Level 3 : Highest level certification for HSMs
- Common Criteria EAL4+ : Advanced security evaluation
- SecNumCloud : ANSSI qualification integrated into the Cloud Temple environment
Multi-tenant Isolation
- Cryptographic Separation: Each client has an isolated KMS domain
- Dedicated Keys: A specific Domain Key per client
- Audit and Traceability: Full logging of actions per domain
Activation and Usage
VM encryption can be activated with a single click from the Console.
For a detailed step-by-step guide with screenshots, see the VM Encryption Tutorial.
Prerequisites
- Key provider configured: A HSM/KMS provider must be enabled on the vStack
- Virtual machine powered off: The VM must be stopped before encryption
- No active replication: The VM must not be replicated (incompatible with Global Mirror)
- No snapshots: No snapshots must be present
- Service subscription: Advanced protection service must be subscribed
Note: For more details on prerequisites and the complete procedure, refer to the VM Encryption Guide.
Limitations and considerations
Compatibility
- Global Mirror: Encrypted virtual machines are not compatible with Global Mirror replication
- Restore: Backups of encrypted VMs retain their cryptographic protection
- Export: Exporting encrypted VMs requires specific procedures
Performance
- Minimal impact: Hardware encryption ensures optimal performance
- Transparency: No impact on application operation
Recommended Use Cases
This advanced protection solution is particularly well-suited for:
- Sensitive Data: Personal information, financial data, trade secrets
- Regulatory Compliance: GDPR, HIPAA, PCI-DSS, ISO 27001, PDIS requirements
- Critical Sectors: Banking, insurance, healthcare, defense
- Digital Sovereignty: Protection against unauthorized access, even in case of compromise
| Reference | Unit | SKU |
|---|---|---|
| ADVANCED PROTECTION - VM Encryption via HSM/KMS | 1 VM | csp:(region):iaas:vmware:encryption:hsm:v1 |
Note:
- The service requires a specific subscription and is not included in the standard IaaS offering
- Key management remains fully under Cloud Temple’s control within the SecNumCloud environment
- Encryption keys never leave the French, sovereign infrastructure