AKS, or Azure Kubernetes Service
Utilise managed Azure Kubernetes service to deploy and scale containers.
Create cloud-native apps, streamline Azure Kubernetes service operations, and experiment with open-source and artificial intelligence.
Overview
Scalable application deployment and management
Drive operational excellence
Utilize unified administration, governance, and monitoring to safely update and enhance Kubernetes environments.
Build resilient global apps
Boost efficiency of containerized workloads from cloud to edge and optimize app performance at scale.
Enhance developer productivity
Utilize generative AI and best-in-class technologies to expedite the development and deployment of applications.
Innovative with AI
By using open-source solutions and experimenting with AI and machine learning, you may investigate novel situations.
Features
Use Kubernetes’ smooth deployment and operation to innovate
A simplified experience with Kubernetes
Use Automatic mode (preview) to automate cluster administration operations like network configuration, node provisioning, upgrades, and scaling.
Curated experience with code-to-cloud
Debugging, automatic node maintenance, continuous integration, and continuous delivery all contribute to the efficiency of developers from start to finish.
Combined logging and monitoring
Detailed information about the health and performance of containerised apps and Azure Kubernetes Service clusters.
Advanced governance and security measures
Strong identification and access control to keep an eye on and preserve container security for large-scale governance.
Deployments from the cloud to the edge
Support for IoT resources, Windows Server, and Linux, along with AKS deployment on your preferred Azure Arc infrastructure.
Safe supply chains for containers
Reliable Azure Kubernetes service solutions with various payment options and click-through deployments are available in the Azure Marketplace.
Security
Integrated security and conformance
- Over the course of five years, Microsoft has pledged to invest USD 20 billion in cybersecurity.
- In 77 countries, it have over 8,500 security and threat intelligence professionals working for us.
- Azure boasts one of the biggest portfolios of compliance certifications in the sector.
Azure Kubernetes Service Pricing
Flexible, tier-based pricing to meet your different workload needs
- Give the Automatic tier (preview) a try and let us know what you think.
- For all production workloads and workloads that need ongoing assistance, use the Standard tier.
How To Getting Started Azure Kubernetes Service
An overview of containers
Understanding containerisation is crucial before beginning to use Azure Kubernetes service.
Software development methods are increasingly using a notion called containerisation, much like the shipping industry uses physical containers to separate distinct cargoes to transport in ships, trains, trucks, and aeroplanes.
The code of an application, together with any necessary libraries, configuration files, and dependencies, are all bundled into a single software package called a container. As a result, developers and IT specialists may produce and implement applications more quickly and safely.
Isolation, portability, agility, scalability, and control over an application’s whole life cycle workflow are all advantages of containerisation. When a container is isolated from the host operating system, it may function independently and is more portable, enabling it to execute reliably and uniformly on any infrastructure and on any platform or cloud.
Azure Kubernetes Service: Components and ideas
The cluster
Azure Kubernetes Service is arranged as a cluster of on-premises or virtual machines at the top level. These devices referred to as nodes share storage, network, and processing power. One master node is linked to one or more worker nodes in each cluster. The master node controls which pods run on which worker nodes, while the worker nodes are in charge of executing pods, which are collections of containerised workloads and applications.
The control plane
Azure Kubernetes Service incorporates a variety of elements that together make up the control plane so that the master node may connect with the worker nodes and a human can communicate with the master node.
Using kubectl, a command-line interface that installs on their local operating system, developers and operators mostly communicate with the cluster through the master node. The kube-apiserver, the Kubernetes API located on the master node, receives commands sent to the cluster via kubectl. The master node’s kube-controller-manager, which is in charge of managing worker node activities, receives requests from the kube-apiserver. The kubelet on the worker nodes receives commands from the master node.
Deploying apps and workloads
App and workload deployment is the next phase in the Kubernetes process. The etc, a key value store database, is constantly updated by the master node with the current configuration and state of the Kubernetes cluster. You will use a YAML file to specify a new intended state to the cluster in order to launch pods with your containerised apps and workloads. Based on preset limitations, the kube-controller-manager uses the YAML file to instruct the kube-scheduler on which worker nodes the workload or application should execute on. The kube-scheduler, which coordinates with the kubelet of each worker node, initiates the pods, monitors the machine status, and oversees resource management.
The intended state that you specify becomes the current state in the etc in a Kubernetes deployment, but the prior state is preserved. Rollbacks, rolling updates, and halting roll outs are all supported by Kubernetes. Replica Sets are also used in the background by deployments to guarantee that the designated number of pods with the same configuration are operating. The Replica Set takes over in the event that one or more pods fail. Kubernetes is referred to as self-healing in this sense.
Structuring and securing Azure Kubernetes Service environments
The final stage in using Azure Kubernetes Service is to organise your workload or application and decide who or what can access it after it has been deployed. You can isolate services, pods, controllers, and volumes from other areas of the cluster while facilitating easy collaboration between them by establishing a namespace, a grouping approach within Kubernetes. Additionally, provide consistent configurations to resources by utilising the namespace notion in Azure Kubernetes Service.
Additionally, every worker node has a kube-proxy that controls how different cluster components can be accessible from the outside. Tokens, certificates, and passwords are examples of sensitive non-public data that should be kept in secrets, which are Azure Kubernetes Service objects that are encoded until execution.
Finally, use role-based access control (RBAC) to define who may see and interact with which areas of the cluster and how they can do so.