skip to main content
10.1145/3317550.3321442acmconferencesArticle/Chapter ViewAbstractPublication PageshotosConference Proceedingsconference-collections
research-article
Public Access

The Synchronous Data Center

Published:13 May 2019Publication History

ABSTRACT

Today, distributed systems are typically designed to be largely asynchronous. Designers assume that the network can drop or significantly delay messages at unpredictable times, that there is no way to know how quickly a node might process a message, or how soon it might respond, and that the clocks of different nodes are at most loosely synchronized. These assumptions are certainly safe, but they come at a price: many applications really do need predictable performance, which, on top of an asynchronous system, has to be approximated at great cost and with lots of redundancy, and many distributed protocols for asynchronous systems are much more complex and expensive than their synchronous counterparts.

The goal of this paper is to start a discussion about whether the asynchronous model is (still) the right choice for most distributed systems, especially ones that belong to a single administrative domain. We argue that 1) by using ideas from the CPS domain, it is technically feasible to build datacenter-scale systems that are fully synchronous, and that 2) such systems would have several interesting advantages over current designs.

References

  1. NSF Cloud 3.0 workshop report. http://pages.cs.wisc.edu/~akella/cloud3workshopreport.pdf, Jan. 2018.Google ScholarGoogle Scholar
  2. M. Aguilera and M. Walfish. No time for asynchrony. In Proc. HotOS, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. D. Alistarh, S. Gilbert, R. Guerraoui, and C. Travers. Of choices, failures and asynchrony: The many faces of set agreement. Algorithmica, 62(1):595--629, Feb 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. S. Angel, H. Ballani, T. Karagiannis, G. O'Shea, and E. Thereska. End-to-end performance isolation through virtual datacenters. In Proc. OSDI, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. E. Arjomandi, M. J. Fischer, and N. A. Lynch. Efficiency of synchronous versus asynchronous distributed systems. J. ACM, 30(3):449--456, July 1983. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. M. Attariyan, M. Chow, and J. Flinn. X-ray: Automating root-cause diagnosis of performance anomalies in production software. In Proc. OSDI, Oct. 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. P. Bahl, R. Chandra, A. Greenberg, S. Kandula, D. A. Maltz, and M. Zhang. Towards highly reliable enterprise network services via inference of multi-level dependencies. In Proc. SIGCOMM, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. G. Banga, P. Druschel, and J. C. Mogul. Resource containers: A new facility for resource management in server systems. In Proc. OSDI, 1999. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. P. Barham, A. Donnelly, R. Isaacs, and R. Mortier. Using Magpie for request extraction and workload modelling. In Proc. OSDI, 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. J. Brutlag. Speed matters for google web search. http://services.google.com/fh/files/blogs/google_delayexp.pdf.Google ScholarGoogle Scholar
  11. T. D. Chandra and S. Toueg. Unreliable failure detectors for reliable distributed systems. J. ACM, 43(2):225--267, Mar. 1996. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. A. Chatzieleftheriou, S. Legtchenko, H. Williams, and A. Rowstron. Larry: Practical network reconfigurability in the data center. In Proc. NSDI, 2018. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. A. Chen, W. B. Moore, H. Xiao, A. Haeberlen, L. T. X. Phan, M. Sherr, and W. Zhou. Detecting covert timing channels with time-deterministic replay. In Proc. OSDI, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. M. Y. Chen, E. Kiciman, E. Fratkin, A. Fox, and E. Brewer. Pinpoint: Problem determination in large, dynamic internet services. In Proc. DSN, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. A. Clement, E. Wong, L. Alvisi, M. Dahlin, and M. Marchetti. Making Byzantine fault tolerant systems tolerate Byzantine faults. In Proc. NSDI, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. J. Corbett et al. Spanner: Google's globally-distributed database. In Proc. OSDI, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. A. R. Curtis, W. Kim, and P. Yalagandula. Mahout: Low-overhead datacenter traffic management using end-host-based elephant detection. In Proc. INFOCOM, 2011.Google ScholarGoogle ScholarCross RefCross Ref
  18. T. Dai, J. He, X. Gu, and S. Lu. Understanding real-world timeout problems in cloud server systems. In Proc. IC2E, 2018.Google ScholarGoogle ScholarCross RefCross Ref
  19. C. Dwork, N. Lynch, and L. Stockmeyer. Consensus in the presence of partial synchrony. J. ACM, 35(2), Apr. 1988. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. S. A. Edwards and E. A. Lee. The case for the precision timed (PRET) machine. In Proc. DAC, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. M. J. Fischer, N. A. Lynch, and M. S. Paterson. Impossibility of distributed consensus with one faulty process. J. ACM, 32(2):374--382, Apr. 1985. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. F. C. Freiling, R. Guerraoui, and P. Kuznetsov. The failure detector abstraction. ACM Comput. Surv., 43(2):9:1--9:40, Feb. 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. S. Ghorbani, Z. Yang, P. B. Godfrey, Y. Ganjali, and A. Firoozshahian. DRILL: Micro load balancing for low-latency data center networks. In Proc. SIGCOMM, 2017. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. H. S. Gunawi, M. Hao, T. Leesatapornwongsa, T. Patana-anake, T. Do, J. Adityatama, K. J. Eliazar, A. Laksono, J. F. Lukman, V. Martin, and A. D. Satria. What bugs live in the cloud? A study of 3000+ issues in cloud systems. In Proc. SoCC, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. A. Haeberlen and K. Elphinstone. User-level management of kernel memory. In Proc. ACSAC, 2003.Google ScholarGoogle ScholarCross RefCross Ref
  26. K. Jang, J. Sherry, H. Ballani, and T. Moncaster. Silo: Predictable message latency in the cloud. In Proc. SIGCOMM, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. S. Kandula, R. Mahajan, P. Verkaik, S. Agarwal, J. Padhye, and P. Bahl. Detailed diagnosis in enterprise networks. In Proc. SIGCOMM, Aug. 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. R. Kapoor, G. Porter, M. Tewari, G. M. Voelker, and A. Vahdat. Chronos: Predictable low latency for data center applications. In Proc. SoCC, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. P. Kocher, D. Genkin, D. Gruss, W. Haas, M. Hamburg, M. Lipp, S. Mangard, T. Prescher, M. Schwarz, and Y. Yarom. Spectre attacks: Exploiting speculative execution. ArXiv 1801.01203, 2018.Google ScholarGoogle Scholar
  30. L. Lamport, R. Shostak, and M. Pease. The Byzantine Generals Problem. ACM Trans. Program. Lang. Syst., 4(3):382--401, July 1982. Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. K. S. Lee, H. Wang, V. Shrivastav, and H. Weatherspoon. Globally synchronized time via datacenter networks. In Proc. SIGCOMM, 2016. Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. J. B. Leners, H. Wu, W.-L. Hung, M. K. Aguilera, and M. Walfish. Detecting failures in distributed systems with the Falcon Spy Network. In Proc. SOSP, 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. M. Lipp, M. Schwarz, D. Gruss, T. Prescher, W. Haas, S. Mangard, P. Kocher, D. Genkin, Y. Yarom, and M. Hamburg. Meltdown. ArXiv 1801.01207, 2018.Google ScholarGoogle Scholar
  34. A. P. Markopoulou, F. A. Tobagi, and M. J. Karam. Assessment of VoIP quality over Internet backbones. In Proc. INFOCOM, 2002.Google ScholarGoogle ScholarCross RefCross Ref
  35. K. Ousterhout, P. Wendell, M. Zaharia, and I. Stoica. Sparrow: Distributed, low latency scheduling. In Proc. SOSP, 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. J. Perry, A. Ousterhout, H. Balakrishnan, D. Shah, and H. Fugal. Fast-pass: A centralized "zero-queue" datacenter network. In Proc. SIGCOMM, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. J. Real and A. Crespo. Mode change protocols for real-time systems: A survey and a new proposal. Real-Time Systems, 26:161--197, 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library
  38. W. Reda, L. Suresh, M. Canini, and S. Braithwaite. BRB: BetteR Batch scheduling to reduce tail latencies in cloud data stores. In Proc. SIGCOMM, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  39. T. Ristenpart, E. Tromer, H. Shacham, and S. Savage. Hey, you, get off of my cloud: Exploring information leakage in third-party compute clouds. In Proc. CCS, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  40. N. Shalom. Amazon found every 100ms of latency cost them 1% in sales. GigaSpaces blog, https://blog.gigaspaces.com/amazon-found-every-100ms-of-latency-cost-them-1-in-sales/, 2008.Google ScholarGoogle Scholar
  41. I. Shin and I. Lee. Periodic resource model for compositional real-time guarantees. In Proc. RTSS, 2003. Google ScholarGoogle ScholarDigital LibraryDigital Library
  42. L. Suresh, M. Canini, S. Schmid, and A. Feldmann. C3: Cutting tail latency in cloud data stores via adaptive replica selection. In Proc. NSDI, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  43. V. Vasudevan, A. Phanishayee, H. Shah, E. Krevat, D. G. Andersen, G. R. Ganger, G. A. Gibson, and B. Mueller. Safe and effective fine-grained TCP retransmissions for datacenter communication. In Proc. SIGCOMM, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  44. Y. Wang, H. Wang, A. Mahimkar, R. Alimi, Y. Zhang, L. Qiu, and Y. R. Yang. R3: Resilient routing reconfiguration. In Proc. SIGCOMM, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  45. R. Wilhelm et al. The worst-case execution-time problem. ACM Trans. Embed. Comput. Syst., 7(3):36:1--36:53, May 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  46. S. Xi, C. Li, C. Lu, C. D. Gill, M. Xu, L. T. X. Phan, I. Lee, and O. Sokolsky. RT-OpenStack: CPU resource management for real-time cloud computing. In Proc. CLOUD, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  47. S. Xi, J. Wilson, C. Lu, and C. Gill. RT-Xen: Towards real-time hypervisor scheduling in Xen. In Proc. EMSOFT, 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  48. K. Zarifis, R. Miao, M. Calder, E. Katz-Bassett, M. Yu, and J. Padhye. DIBS: Just-in-time congestion mitigation for data centers. In Proc. EuroSys, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  49. X. Zhang, E. Tune, R. Hagmann, R. Jnagal, V. Gokhale, and J. Wilkes. CPI 2: CPU performance isolation for shared compute clusters. In Proc. EuroSys, Apr. 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  50. D. Zhuo, M. Ghobadi, R. Mahajan, K. Förster, A. Krishnamurthy, and T. E. Anderson. Understanding and mitigating packet corruption in data center networks. In Proc. SIGCOMM, 2017. Google ScholarGoogle ScholarDigital LibraryDigital Library

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Conferences
    HotOS '19: Proceedings of the Workshop on Hot Topics in Operating Systems
    May 2019
    227 pages
    ISBN:9781450367271
    DOI:10.1145/3317550

    Copyright © 2019 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 13 May 2019

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • research-article
    • Research
    • Refereed limited

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader