Code Agency Cloud.

Knowledge is power

Powerful & Scalable cloud hosting.

Code Agency Cloud is the most powerful and scalable cloud hosting build on top of k3s Kubernetes and can host any type of application. 

Our platform is available in several ways and with many options to select from. Therefor, there is no fix price available as every option impact the final price. 
To give you a price indication, you can fill out the form to receive a non-binding quotation and get in touch with our cloud engineers. 

A Kubernetes cluster is a collection of nodes (physical or virtual machines) that are interconnected and used to run containerized applications and manage their lifecycle. Kubernetes is an open-source container orchestration platform that automates the deployment, scaling, and management of applications across a cluster of nodes.

In a Kubernetes cluster, one or more master nodes act as the control plane, managing and coordinating the cluster’s resources and workload scheduling. The worker nodes, also known as minions, are responsible for running the actual containers and executing the tasks assigned by the master nodes.

The cluster operates based on the declarative model, where you define the desired state of your application and Kubernetes ensures that the actual state matches the desired state. This allows you to easily scale your applications, roll out updates, handle failover, and distribute the workload efficiently across the cluster.

Kubernetes provides various features and functionalities to manage containerized applications effectively, including service discovery and load balancing, automatic scaling, self-healing, storage orchestration, and configuration management. It also supports advanced deployment strategies like canary deployments and blue-green deployments.

Overall, a Kubernetes cluster provides a scalable, resilient, and flexible platform for managing containerized applications, allowing organizations to simplify the deployment and management of their applications in a cloud-native environment.

High availability (HA) refers to a system or infrastructure that is designed to remain operational and accessible for an extended period, even in the face of failures or disruptions. The goal of high availability is to minimize downtime and ensure that critical services or applications are continuously available to users.

In a high availability setup, redundancy and fault tolerance are key principles. This involves using redundant components, such as servers, networks, or storage devices, that can seamlessly take over if a failure occurs. By eliminating single points of failure, high availability architectures increase the overall reliability and resilience of the system.

Here are some common techniques and practices used to achieve high availability:

  1. Redundancy: Deploying redundant components, such as multiple servers or network links, so that if one fails, another can take over without interruption.
  2. Load balancing: Distributing the workload across multiple servers to prevent any single server from becoming overloaded and causing service degradation.
  3. Failover: Automatic detection of a failure and quick switching to a redundant component or backup system to ensure uninterrupted service.
  4. Clustering: Grouping multiple servers together in a cluster, allowing them to work together and provide redundancy and failover capabilities.
  5. Data replication: Keeping multiple copies of data synchronized across different systems or locations to ensure data availability and integrity.
  6. Monitoring and automatic recovery: Constantly monitoring the system and automatically taking actions to recover from failures or performance issues.

High availability is particularly important for critical systems and services that require continuous operation, such as e-commerce websites, financial applications, telecommunications networks, or cloud infrastructure. By implementing high availability measures, organizations can minimize downtime, provide better user experience, and ensure business continuity even in the face of hardware failures, software bugs, or other unforeseen events.

Replication is the process of creating and maintaining multiple copies of data or resources. It improves data availability, enhances fault tolerance, and boosts system performance by distributing the workload across multiple instances. Replication can occur at different levels, such as databases, files, or applications, and is used to ensure redundancy, backup, and disaster recovery. It plays a crucial role in achieving high availability and data resilience.

Lorem ipsum dolor sit amet, consectetur adipiscing elit. Ut elit tellus, luctus nec ullamcorper mattis, pulvinar dapibus leo.
Lorem ipsum dolor sit amet, consectetur adipiscing elit. Ut elit tellus, luctus nec ullamcorper mattis, pulvinar dapibus leo.
Lorem ipsum dolor sit amet, consectetur adipiscing elit. Ut elit tellus, luctus nec ullamcorper mattis, pulvinar dapibus leo.
Lorem ipsum dolor sit amet, consectetur adipiscing elit. Ut elit tellus, luctus nec ullamcorper mattis, pulvinar dapibus leo.
Lorem ipsum dolor sit amet, consectetur adipiscing elit. Ut elit tellus, luctus nec ullamcorper mattis, pulvinar dapibus leo. request


This section is about the infrastructure for deploying your application(s)

Clustered hosting and specially with Kubernetes, it is a requirement to have a DNS provider that is capable of API automation to handle the creation and updating of DNS records to automate application deployments and handle failovers, migrations etc...
The load balancer and ingress will handle all the incoming traffic into your cluster and distribute it to your application(s).
This feature is provisioned with * or * (domain name can't be changed). It's typically used to spin up applications when there is no final domain yet or serves as a trial/test purpose to spin up staging or development environments without the need of final FQDN domain.
This option will decide if all hardware needs replication (minimum 3 units) to provide maximum uptime in the eventual case of machine downtime in the datacenter.
This option provides the feature to spin up servers in multiple locations in the world (depending on the hardware vendor locations and compatibility). In the eventual case of a total datacenter disaster, your applications can migrate automatically or failover automatically to machines in a different datacenter without manual intervention required.
This option is to decide if you want to run additional servers with the single purpose as a seperate storage layer for all your applications. If not, it will consume the server hard drive storage but is limited at the machine resource limits.
A seperate storage pool can scale up infinitely and is capable of serving as an internal CDN.
A storage pool typically has lower storage performance speeds than on-device storage types (mostly NVME drives) but has the best flexibility and scalability.


This section is about deploying your application(s). 

This means your applications will always run with minimum 3 replicas
This means your database will always run with minimum 3 replicas
This option means we will deploy an instance of to handle your application events and error monitoring.
This option means we will deploy an instance of ArgoCD to handle your CI&CD processes.