Your Web News in One Place

Help Webnuz

Referal links:

Sign up for GreenGeeks web hosting
January 25, 2023 03:50 pm GMT

What are the directions of technical data platforms scalability?

Technical scalability is one of the main drivers of a data platform, as mentioned in the SOFT methodology. But what are the options? In this blog we'll evaluate the direction of scalability and the tradeoffs you might encounter in the decision process.

The main directions of technical scalability

Vertical scalability

Vertical scalability generally means "buy a bigger box": what can't be achieved with a server sized X, could be achieved with a server sized 2X. There's plenty of data platforms that allow this type of scaling and, with all the options available now in the cloud market, the vertical scalability can take you loooooooong way.

Pro:

  • No changes in the architecture
  • Easy estimation of new performance

Cons:

  • Resource constraints on a single machine
  • The cost of a single big machine is greater than the cost of more smaller machines
  • The load is parallelized only within the machine

Horizontal scalability

Horizontal scalability on the other side, allows you to "add more nodes to a cluster". It's usually available in distributed systems where the workload is shared across nodes. Apache Kafka is a perfect example: its distributed nature allows yoi to plug in and out nodes from the cluster and have an automatic rebalancing of the data and responsibilities across nodes.

Pro:

  • Easy/Fast plug in of new resources
  • Cost of more smaller machines is usually less of the cost of a single big machine
  • The load is parallelized across the cluster

Cons:

  • Only available in distributed systems
  • Change in the cluster architecture for every new node added
  • Data is replicated

Diagonal scalability

A mix of the two, is of course possible, in situations where we could both add nodes to a cluster, and at the same time, replace the existing with new nodes of enhanced capacity. Typically the scenario is similar to the Horizontal scalability, the only change is that the new nodes have different (usually increased) specs compared to the previous ones.

Another way of scaling "diagonally" is to create different nodes for different purposes. Taking PostgreSQL as example, one might choose to scale vertically on the primary, but also scale horizontally to create one or more "read only replicas" that could fulfill all the read-only traffic, relieving the primary from part of the workload.

Pro:

  • Freedom of scaling in both directions

Cons:

  • Not always available
  • For some of the technologies, the workload (e.g. write operations) can't really be shared across all nodes

Cross tech scalability

Not all the workloads need to be performed in one single data technology: As the previous example with PostgreSQL read-only replicas, different nodes of a cluster can fulfil different workloads. Taking the example further, one could scale by storing the transactional data in PostgreSQL but then integrating ClickHouse for analytical workloads. This cross tech scalability allows to optimize each component for a single usage and scale independently.

Pro:

  • Workload-optimised data platforms
  • Independent, workload-related, scaling

Cons:

  • Data synch between technologies

Challenges of data platform scaling

If data platform scaling was an easy task, we wouldn't talk a lot about it. There are some challenges that the scaling operation brings with it.

Where does the data reside?

To achieve optimal performance and minimize any network latency, data platforms historically stored co-located data storage and compute. In recent times, a new trend of detaching compute from storage (see Snowflake, neon) allowed further scalability options to increase the disk and the CPU independently at the cost of (a small) additional latency.
The tradeoff is between performance (immediate local access), workload isolation/independent scaling.

Furthermore, if the data is co-located with the compute, it needs to be moved when performing, for example, a vertical scaling.

Time to scale

Depending on the architecture, the time to scale can vary between fractions of a second (in case only a new compute instance needs to be spun up) to hours (in case huge backups need to be restored). Evaluating the time to scale is necessary to understand the scaling thresholds: e.g. if one can scale in 10 seconds, the scaling option could be triggered at 95% capacity, if the scaling takes a few hours, the threshold needs to be lowered.

Downtime

Nobody wants downtime, specifically when needing to scale up to follow a rise in demand. However one must understand and evaluate:

  • how much work can be done without changes in the original system
  • how long is the switchover downtime (if any)
  • how and how fast clients can re-sync connection parameters

Answering the above queries will help understand how to minimize the unavailability of the system.

Scaling-induced additional load

The scaling operation might induce additional load on the system, e.g. when streaming changes from one node to the other. Evaluating how the additional load could impact system performances needs to be done upfront to ensure to provide a sufficient service during the scaling procedure.

Not all workloads are equal

As mentioned before, the type of scaling depends on the architectures and workloads: adding a read-only replica in a specific cloud region is a fairly low cost solution to alleviate some load from a primary PostgreSQL database. That said, the solution doesn't work when a lot of write workloads are in the game.

Backups/Restore

What does it have to do "backup/restore" with scaling? Well, some of the technologies (hi PostgreSQL!) will initially restore a database copy from a backup and then start streaming the delta from the primary. Therefore, in such cases, having a ready to go, and up to date backup allows you to cut the time to scale.

More data jumps more risks

In some of the scenarios presented above, scaling is performed by adding new nodes to a cluster, or adding new technologies in the mix. Even if these options provide great benefit in the parallelism and the separation of workloads they are adding additional steps in the data journey.

Every time there's more than one server storing the data, there's always the chance of not-in-sync replicas. Every time you add a new data hop, there are always additional complexities with network, integration, etc.

Conclusion

This article aims at covering a wide spectrum of scaling possibilities and related issues. When designing a data platform, the scalability problem (not only technical, check out the SOFT methodology to know more) should be an important decision factor to evaluate a solution based on the underline technologies, the expected workloads and the business constraints.


Original Link: https://dev.to/ftisiot/what-are-the-directions-of-technical-data-platforms-scalability-2c9d

Share this article:    Share on Facebook
View Full Article

Dev To

An online community for sharing and discovering great ideas, having debates, and making friends

More About this Source Visit Dev To