IDG Contributor Network: The 8 fallacies of distributed computing are becoming irrelevant
In 1969, the U.S. Department of Defense created ARPANET, the precursor to today’s internet. Around the same time, the SWIFT protocol used for money transfers was also established. These are both early examples of distributed systems: a collection of independent computers that appear to users as a single coherent system.Many come to know they have a distributed system when the crash of a computer they’ve never heard of affects the whole system. This is often the result of assumptions architects and designers of distribution systems are likely to make.In 1994, Peter Deutsch, who worked at Sun Microsystems, wrote about these assumptions to explore what can go wrong in distributed systems. In 1997, James Gosling added to this list to create what is commonly known as the eight fallacies of distributed computing. Traditional approaches, which use time-based replication to architect and build distributed systems, suffer from many of these fallacies and result in systems that are inefficient, insecure and costly to maintain. Modern approaches, using complex mathematics such as the Paxos algorithm, overcome many of these significant hurdles.To read this article in full or to leave a comment, please click here