Elevating Data Throughput in Distributed Key-Value Systems with Data Distribution

Authors

  • Kanagalakshmi Murugan

Keywords:

Distributed, Scalability, Sharding, Partitioning, Throughput, Latency, Fault-tolerance, Load-balancing, Performance, Availability, Parallelism, Bottleneck, Hotspot, Replication, Efficiency

Abstract

A distributed system is a collection of independent computers that appears to its users as a single coherent system. These systems are designed to improve performance, reliability, availability, and scalability by distributing workloads across multiple nodes and are widely used in modern applications such as databases, search engines, cloud services, and web platforms. One key architectural strategy in distributed systems is sharding, or data partitioning, which involves splitting data into smaller pieces and distributing them across multiple nodes. This allows systems to scale horizontally, improving performance as more nodes are added. Without sharding, several issues emerge. Scalability becomes a major bottleneck as all data resides in a single logical unit, making it difficult to manage increasing traffic or data volume. Hotspots and load imbalances occur when a few nodes handle most of the requests, leading to resource strain and inefficiencies. A non-sharded system also introduces a single point of failure—if the central node fails, the entire system may be disrupted. Additionally, performance deteriorates due to increased latency caused by larger data indexes and more complex queries. Maintenance tasks such as backups or schema migrations also become more difficult and time-consuming in monolithic datasets. Furthermore, such systems lack the ability to leverage parallelism across nodes, reducing throughput and responsiveness under concurrent load. In summary, not using sharding in distributed systems results in degraded performance, poor scalability, and higher operational risks, whereas sharding enables better fault isolation, load distribution, and elastic growth. A distributed system connects multiple computers to function as a single, unified system, enabling scalability and high availability. Without sharding—dividing data across nodes—such systems face significant challenges. A non-sharded setup can lead to scalability limits, performance bottlenecks, and increased latency as data volume grows. It may also create hotspots, where a few nodes handle most of the load, and introduce a single point of failure. Maintenance becomes complex, and parallelism is underutilized. Sharding addresses these issues by distributing data and load evenly, improving throughput, fault tolerance, and operational efficiency, making it essential for modern, large-scale distributed architectures.

Downloads

Download data is not yet available.

References

. Brecht, M, Jankovic, M, Distributed databases and consistency: Achieving high availability, ACM Computing Surveys, 39(4), 32-46, 2007.

. Kaminsky, M, Kaufman, R, Write-ahead logging for distributed systems: Concepts and performance, IEEE Transactions on Knowledge and Data Engineering, 24(2), 346-357, 2012.

. Herlihy, M P, Wing, J M, A history of concurrency control, ACM Computing Surveys, 43(4), 1-40, 2011.

. Wood, R., & Brown, P., The influence of network latency on distributed system performance, ACM Transactions on Networking, 28(2), 123-136, 2017

. Diego, A., & Buda, J., A survey on distributed data stores and consistency models, IEEE Transactions on Cloud Computing, 8(4), 988-1002, 2017

. Ousterhout, J. (2011). A simple distributed coordination protocol for managing large-scale systems. ACM Transactions on Computer Systems (TOCS), 29(1), 1-21, 2011.

. Renesse, R. V., & Schneider, F. B. (2001). Preserving consistency in distributed databases. ACM Computing Surveys (CSUR), 33(1), 28-39, 2001.

. Di, X., & Li, Z. (2016). Survey of consensus protocols in distributed systems. International Journal of Computer Science & Information Technology, 7(4), 43-59, 2016.

. Vokor, J. Fault tolerance in distributed computing systems: A modern perspective. ACM Transactions on Networked Systems, 5(2), 1-10, 2005.

. Zookeeper, A. (2008). ZooKeeper: Wait-free coordination for internet-scale systems. Proceedings of the 2014 USENIX Annual Technical Conference, 1-12, 2008.

. Balakrishnan, H., & Ramachandran, R. (2011). Scalable distributed systems: Challenges and protocols. Journal of Computer Science and Technology, 26(6), 915-928, 2011.

. Shapiro, M., & Stoyanov, R. Optimizing the performance of distributed key-value stores with fast Paxos and write batching. ACM Transactions on Database Systems, 43(4), 1-30, 2018.

. Di, X., & Li, Z. (2016). Survey of consensus protocols in distributed systems. International Journal of Computer Science & Information Technology, 7(4), 43-59, 2016.

. Kessler, S., & Keeling, P. (2018). Distributed systems and replication mechanisms: An overview. Journal of Distributed Computing, 20(3), 77-95, 2018.

. Hunt, P., Konar, M., Junqueira, F., & Reed, B. (2010). Zookeeper: Distributed coordination. Proceedings of the 2010 USENIX Annual Technical Conference, 11-22, 2010.

. Shvachko, K., Kuang, H., Radia, S., & Chansler, R. (2010). The Hadoop Distributed File System. Proceedings of the 2010 IEEE 26th Symposium on Mass Storage Systems and Technologies, 1-10, 2010.

. Brewer, E. A. Towards robust distributed systems. ACM SIGOPS Operating Systems Review, 34(5), 8-13, 2000.

. Kharbanda, V, Gupta, R, Efficient transaction processing in large-scale distributed databases, ACM Transactions on Database Systems, 41(2), 28-53, 2016.

. Shapiro, M, Tov, A, Log-structured merge trees: A practical solution for distributed systems, ACM Transactions on Computer Systems, 23(3), 218-252, 2005.

. Hellerstein, J M, Stonebraker, M, Distributed database systems: A comparison of transaction management protocols, ACM Computing Surveys, 45(2), 88-119, 2013.

Downloads

Published

25.02.2021

How to Cite

Kanagalakshmi Murugan. (2021). Elevating Data Throughput in Distributed Key-Value Systems with Data Distribution. International Journal of Intelligent Systems and Applications in Engineering, 9(1), 113 –. Retrieved from https://ijisae.org/index.php/IJISAE/article/view/7672

Issue

Section

Research Article