- ➡️ A/B Testing - allows deploying multiple versions of a service as independent instances, using load balancers or API gateways to route traffic between versions A and B Centralized observability tools capture real-time performance and user interaction data, which is aggregated for analysis to assess conversion rates and effectiveness. This setup supports safe rollbacks, enabling teams to revert to the stable version instantly if variant B underperforms, facilitating resilient, data-driven iteration.
- 馃敼 Agreed Service Time (AST) - time during which a particular Business Continuity is agreed to be fully available, ideally as defined in the Service Level Agreement.
- 馃搻 Availability (Service) - measures how reliably a system is operational and accessible, typically represented as a percentage of uptime over a set period. High availability is critical, often defined in SLAs, where many organizations aim for “five nines” (99.999%), meaning minimal downtime annually. Fault tolerance and redundancy are essential strategies to achieve high availability, allowing systems to continue functioning even if components fail. Monitoring and scheduled maintenance also help by proactively identifying and addressing issues, reducing unplanned downtime.
- 馃搻 Bandwidth - is the format and size of the messages that are exchanged between the microservices. You can optimize the size by applying compression, filtering, batching, caching techniques or changing the format from verbose JSON to binary Protobuf or Avro.
- 馃洜️ Benchmarking tools
- 馃敼 Bottleneck - refer to limitations within a system that hinder its ability to grow and manage increased workloads effectively. Identifying scalability bottlenecks is crucial for improving system performance, as addressing them can lead to significant enhancements in throughput and response times.
- ⚙️ Cache (in-memory)
- ⚙️ Cache (multilevel)
- L1 (Instruction/Data Level) - lightning-fast caches sit closest to the CPU cores, and their primary role is to store frequently accessed instructions and data. JVM: Ehcache; PHP: APCu/OPCache.
- L2 Application - Distributed cache systems like Redis or Memcached which allows to share data between many application instances.
- L3 DB - Materialized Views.
- L4 HTTP - Varnish.
- 馃敼 Capacity (System)
- ⚙️ CloudFront (AWS) - is a CDN that delivers static and dynamic content reliably, securely, and efficiently.
- ⚙️ Content Delivery Networks (CDN) - geographically distributed group of servers that caches content close to end users.
- 馃敼 Cost Upfront
- 馃搻 Cost Efficiency
- ➡️ Database Horizontal Partitioning
- 馃搻 Database Load
- ➡️ Database Partitioning
- ➡️ Database Replication (master/slave)
- 馃敼 Database Retention
- 馃敼 Database (Retentional)
- ⚙️ Database Shards
- 馃搻 Database Slow Query
- ➡️ Database Vertical Partitioning - in a table entire columns are separated out and put into new, distinct tables.
- 馃敼 Data Consistency (in case of microservices)
- 馃敼 Data Consistency (eventual)
- 馃敼 Data Consistency (strong)
- 馃敼 Data Integrity
- 馃敼 Data Quality
- ➡️ Diagonal Scaling - hybrid of vertical and horizontal scaling.
- ⚙️ Distributed System
- 馃敼 Downtime (DT) [duration/frequency]
- 馃敼 Expenditure
- 馃搻 Fault Tolerance - ability of a system to continue operating without interruption despite encountering failures in one or more of its components. It is a measure of the system's resilience against disruptions (single server failure/data center outage). Strategies: redundancy, sharding, replication, load balancing, failure detection and recovery.
- ➡️ Horizontal Scaling (scale out) - ability for an application to automatically scale by adding/reducing computing nodes as the workload increases/decreases. Usually requires a load balancer. Advantages: easy increasing capacity on the fly, limitless scaling, easier to run fault-tolerance, improve resilience. Disadvantages: complex architectural design, increased costs (resources in data center), complex debugging.
- 馃敼 Idle Resource
- 馃敼 Infrastructure as Code (IaC) - practice where infrastructure provisioning and management are performed through code. This approach automates the setup and scaling of cloud environments, ensuring consistency and eliminating manual errors.
- 馃洜️ k8s - abbreviation of Kubernetes.
- ⚙️ [k8s] Cluster - is a set of nodes that run containerized applications. Containerizing applications packages an app with its dependences and some necessary services. They are more lightweight and flexible than virtual machines. In this way, Kubernetes clusters allow for applications to be more easily developed, moved and managed.
- 馃洜️ [k8s] Cluster Autoscaler (CA)
- 馃洜️ [k8s] Horizontal Pod Autoscaler (HPA)
- ⚙️ [k8s] Node - is a physical or virtual machine that serves as the infrastructure where Pods (containers) run. Each Node provides essential resources, like CPU, memory, and storage, to support applications within the Kubernetes cluster. Nodes run key Kubernetes components such as the kubelet and kube-proxy to manage container operations and network communication. If a Node fails, Kubernetes automatically relocates its workloads to other Nodes, ensuring application stability and high availability across the cluster.
- ⚙️ [k8s] Pod - is the smallest deployable unit and serves as a container for one or more application containers. All containers within a Pod share the same network and storage resources, allowing them to communicate seamlessly and exchange data. Pods are temporary; if a Pod fails, k8s will automatically replace it with a new one to maintain application stability. K8s can also scale an application by creating multiple replicas of a Pod, balancing the load and ensuring availability across different nodes.
- 馃洜️ [k8s] Karpenter - is an open-source, flexible, and high-performance Kubernetes cluster autoscaler, offering advanced scheduling and scaling capabilities. It can improve application availability and cost efficiency by rapidly launching right-sized compute resources in response to changing application load.
- 馃洜️ [k8s] Kubernetes - is used to manage and scale applications running in containers, which are small, isolated environments. It simplifies the reliable management of numerous apps and services, even when they are distributed across multiple servers. Kubernetes automates things like: starting new apps when needed. It's open-sorce.
- 馃洜️ [k8s] Vertical Pod Autoscaler (VPA) - is an autoscaler that enables automatic CPU and memory request and limit adjustments based on historical resource usage measurements.
- 馃敼 Latency
- ⚙️ Load Balancer
- 馃敼 Load Testing
- 馃洜️ OpenShift
- ➡️ Optimization (proactive)
- 馃敼 Parallel Computing
- 馃敼 Performance
- 馃洜️ Prometheus - is an open-source tool used for monitoring and alerting. It collects metrics on system performance, such as CPU usage, memory, and network traffic, and stores them in a time-series database for real-time analysis. It features the PromQL query language, which enables custom metric queries and is often paired with Grafana for advanced data visualization, making it ideal for monitoring Kubernetes and cloud-based environments.
- 馃敼 Quality of Service (QoS)
- ⚙️ Rate Limiter
- ➡️ Read Scaling (1)
- 馃敼 Redundancy (System)
- 馃敼 Region
- 馃敼 Reliability (System)
- 馃敼 Resilience (System)
- ➡️ Resource Allocation Strategies
- 馃搻 Response Time
- ⚙️ Round Robin
- 馃敼 Service Discovery
- ⚙️ Sharding - is a technique of splitting a large dataset into smaller, independent segments called shards, each stored on separate servers. This approach enhances horizontal scalability and performance by distributing the load across multiple servers, allowing each shard to handle a subset of the data. By using a shard key, data is allocated to shards based on a specific criterion, such as range, hashing, or time. Sharding, while powerful, adds complexity in data management and can make cross-shard queries and transactions more challenging to handle.
- 馃敼 Service Level Agreement (SLA)
- 馃敼 Single Point of Failure (SPOF)
- 馃敼 Temporary Unavailability (System)
- 馃洜️ Terraform - Infrastructure as Code (IaC) software tool offered by HashiCorp. It facilitates provisioning and managing your infrastructure on-prem and in the cloud. It can be easily extended with the help of its plugin-based architecture.
- ⚙️ Time Consuming Task
- 馃搻 Traffic
- 馃敼 TTL (Time to Live) - is a mechanism used in computer networking to limit the lifespan or lifetime of data in a network. It was originally designed as a way to prevent network packets from circulating indefinitely in the event of routing loops. TTL is implemented as a counter or timestamp attached to data, which is decremented as it passes through network devices.
- 馃敼 Uptime
- 馃洜️ Varnish - is an HTTP reverse proxy and caching server that accelerates web applications by storing copies of frequently accessed pages or assets in memory. Positioned between users and the backend server, Varnish can serve cached content directly, reducing server load and significantly improving response times. It’s widely used on high-traffic websites for its high performance and flexible configuration options through Varnish Configuration Language (VCL).
- ➡️ Vertical Scaling (scale up) - consists of bolstering the computing power of a single existing server, by increasing the number of CPU cores, memory and disk storage. Virtual server hardware resources are limited to a certain level. Initial costs of Vertical Scaling are lower than Horizontal. Disadvantages: SPOF, hardware limitations, downtime on update (server is offline).
- 馃敼 Workload
- 馃敼 Workload (heavy)
- 馃敼 Workload spike
- 馃敼 Workload distribution
- 馃敼 Workload volumes
- ➡️ Write Scaling (1)
馃搱 Scalability glossary
Subskrybuj:
Posty (Atom)
Brak komentarzy:
Prze艣lij komentarz