Download PDF version of this article PDF

Should You Upload or Ship Big Data to the Cloud?

The accepted wisdom does not always hold true.


Sachin Date, e-Emphasys Technologies

It is accepted wisdom that when the data you wish to move into the cloud is at terabyte scale and beyond, you are better off shipping it to the cloud provider, rather than uploading it. This article takes an analytical look at how shipping and uploading strategies compare, the various factors on which they depend, and under what circumstances you are better off shipping rather than uploading data, and vice versa. Such an analytical determination is important to make, given the increasing availability of gigabit-speed Internet connections, along with the explosive growth in data-transfer speeds supported by newer editions of drive interfaces such as SAS and PCI Express. As this article reveals, the aforementioned "accepted wisdom" does not always hold true, and there are well-reasoned, practical recommendations for uploading versus shipping data to the cloud.

Here are a few key insights to consider when deciding whether to upload or ship:

• A direct upload of big data to the cloud can require an unacceptable amount of time, even over Internet connections of 100-Mbps (megabits per second) and faster. A convenient workaround has been to copy the data to storage tapes or hard drives and ship it to the cloud data center.

• With the increasing availability of affordable, optical fiber-based Internet connections, however, shipping the data via drives becomes quickly unattractive from the point of view of both cost and speed of transfer.

• Shipping big data is realistic only if you can copy the data into (and out of) the storage appliance at very high speeds and you have a high-capacity, reusable storage appliance at your disposal. In this case, the shipping strategy can easily beat even optical fiber-based data upload on speed, provided the size of data is above a certain threshold value.

• For a given value of drive-to-drive data-transfer speed, this threshold data size (beyond which shipping the data to the cloud becomes faster than uploading it) grows with every Mbps increase in the available upload speed. This growth continues up to a certain threshold upload speed. If your ISP provides an upload speed of greater or equal to this threshold speed, uploading the data will always be faster than shipping it to the cloud, no matter how big the data is.

Suppose you want to upload your video collection into the public cloud; or let's say your company wishes to migrate its data from a private data center to a public cloud, or move it from one data center to another. In a way it doesn't matter what your profile is. Given the explosion in the amount of digital information that both individuals and enterprises have to deal with, the prospect of moving big data from one place to another over the Internet is closer than you might think.

To illustrate, let's say you have 1 TB of business data to migrate to cloud storage from your self-managed data center. You are signed up with a business plan with your ISP that guarantees you an upload speed of 50 Mbps and a download speed of 10 times as much. All you need to do is announce a short system-downtime window and begin hauling your data up to the cloud. Right?

Not quite.

For starters, you will need a whopping 47 hours to finish uploading 1 TB of data at a speed of 50 Mbps—and that's assuming your connection never drops or slows down.

If you upgrade to a faster—say, 100 Mbps—upload plan, you can finish the job in one day. But what if you have 2 TB of content to upload, or 4 TB, or 10 TB? Even at a 100-Mbps sustained data-transfer rate, you will need a mind-boggling 233 hours to move 10 TB of content!

As you can see, conventional wisdom breaks down at terabyte and petabyte scales. It's necessary to look at alternative, nonobvious ways of dealing with data of this magnitude.

Here are two such alternatives available today for moving big data:

• Copy the data locally to a storage appliance such as LTO (linear tape open) tape, HDD (hard-disk drive), or SSD (solid-state drive), and ship it to the cloud provider. For convenience, let's call this strategy "Ship It!"

• Perform a cloud-to-cloud transfer of content over the Internet using APIs (application programming interfaces) from both the source and destination cloud providers.6 Let's call this strategy "Transfer It!"

This article compares these alternatives, with respect to time and cost, to the baseline technique of uploading the data to the cloud server using an Internet connection. This baseline technique is called "Upload It!" for short.

A REAL-LIFE SCENARIO

Suppose you want to upload your content into, purely for the sake of illustration, the Amazon S3 (Simple Storage Service) cloud, specifically its data center in Oregon.2 This could well be any other cloud-storage service provided by players9 in this space such as (but not limited to) Microsoft, Google, Rackspace, and IBM. Also, let's assume that your private data center is located in Kansas City, Missouri, which happens to be roughly geographically equidistant from Amazon's data centers2 located in the eastern and western United States.

Kansas City is also one of the few places where a gigabit-speed optical-fiber service is available in the United States. In this case, it's offered by Google Fiber.7

As of November 2015, Google Fiber offers one of the highest speeds that an ISP can provide in the United States: 1 Gbps (gigabit per second), for both upload and download.13 Short of having access to a leased Gigabit Ethernet11 line, an optical fiber-based Internet service is a really, really fast way to shove bits up and down Internet pipes anywhere in the world.

Assuming an average sustained upload speed of 800 Mbps on such a fiber-based connection,13 (i.e., 80 percent of its advertised theoretical maximum speed of 1 Gbps), uploading 1 TB of data will require almost three hours to upload from Kansas City to S3 storage in Oregon. This is actually pretty quick (assuming, of course, your connection never slows down). Moreover, as the size of the data increases, the upload time increases in the same ratio: 20 TB requires 2½ days to upload, 50 TB requires almost a week to upload, and 100 TB requires twice that long. At the other end of the scale, a half a petabyte of data requires two months to upload. Uploading one petabyte at 800 Mbps should keep you going for four months.

It's time to consider an alternative.

SHIP IT!

That alternative is copying the data to a storage appliance and shipping the appliance to the data center, at which end the data is copied to cloud storage. This is the Ship It! strategy. Under what circumstances is this a viable alternative to uploading the data directly into the cloud?

The Mathematics of Shipping Data

When data is read out from a drive, it travels from the physical drive hardware (e.g., the HDD platter) to the on-board disk controller (the electronic circuitry on the drive). From there the data travels to the host controller (a.k.a. the host bus adapter, a.k.a. the interface card) and finally to the host system (e.g., the computer with which the drive is interfaced). When data is written to the drive, it follows the reverse route.

When data is copied from a server to a storage appliance (or vice versa), the data has to travel through an additional physical layer, such as an Ethernet or USB connection existing between the server and the storage appliance.

Figure 1 is a simplified view of the data flow when copying data to a storage appliance. The direction of data flow shown in the figure is conceptually reversed when the data is copied out from the storage appliance to the cloud server.

Should You Upload or Ship Big Data to the Cloud?

Note that often the storage appliance may be nothing more than a single hard drive, in which case the data flow from the server to this drive is basically along the dotted line in the figure.

Given this data flow, a simple way to express the time needed to transfer the data to the cloud using the Ship It! strategy is shown in equation 1:

Should You Upload or Ship Big Data to the Cloud?

Where:

Vcontent is the volume of data to be transferred in megabytes (MB).

SpeedcopyIn is the sustained rate in MBps (megabytes per second) at which data is copied from the source drives to the storage appliance. This speed is essentially the minimum of three speeds: (1) the speed at which the controller reads data out of the source drive and transfers it to the host computer with which it interfaces; (2) the speed at which the storage appliance's controller receives data from its interfaced host and writes it into the storage appliance; and (3) the speed of data transfer between the two hosts. For example, if the two hosts are connected over a Gigabit Ethernet or a Fibre Channel connection, and the storage appliance is capable of writing data at 600 MBps, but if the source drive and its controller can emit data at only 20 MBps, then the effective copy-in speed can be at most 20 MBps.

SpeedcopyOut is similarly the sustained rate in MBps at which data is copied out of the storage appliance and written into cloud storage.

Ttransit is the transit time for the shipment via the courier service from source to destination in hours.

Toverhead is the overhead time in hours. This can include the time required to buy the storage devices (e.g., tapes), set them up for data transfer, pack and create the shipment, and drop it off at the shipper's location. At the receiving end, it includes the time needed to process the shipment received from the shipper, store it temporarily, unpack it, and set it up for data transfer.

The Use of Sustained Data-transfer Rates

Storage devices come in a variety of types such as HDD, SSD, and LTO. Each type is available in different configurations such as a RAID (redundant array of independent disks) of HDDs or SSDs, or an HDD-SSD combination where one or more SSDs are used as a fast read-ahead cache for the HDD array. There are also many different data-transfer interfaces such as SCSI (Small Computer System Interface), SATA (Serial AT Attachment), SAS (Serial Attached SCSI), USB (universal serial bus), PCI (Peripheral Component Interconnect) Express, Thunderbolt, etc. Each of these interfaces supports a different theoretical maximum data-transfer speed.

Figure 2 lists the data-transfer speeds supported by a recent edition of some of these controller interfaces.

Should You Upload or Ship Big Data to the Cloud?

The effective copy-in/copy-out speed while copying data to/from a storage appliance depends on a number of factors:

• Type of drive. For example, SSDs are usually faster than HDDs partly because of the absence of any moving parts. Among HDDs, higher-RPM drives can exhibit lower seek times than lower-RPM drives. Similarly, higher areal-density (bits per surface area) drives can lead to higher data-transfer rates.

• Configuration of the drive. Speeds are affected by, for example, single disk versus an array of redundant disks, and the presence or absence of read-ahead caches on the drive.

• Location of the data on the drive. If the drive is fragmented (particularly applicable to HDDs), it can take longer to read data from and write data to it. Similarly, on HDD platters, data located near the periphery of the platter will be read faster than data located near the spindle. This is because the linear speed of the platter near the periphery is much higher than near the spindle.

• Type of data-transfer interface. SAS-3 versus SATA Revision 3, for example, can make a difference in speeds.

• Compression and encryption. Compression and/or encryption at source and decompression and/or de-encryption at the destination reduce the effective data-transfer rate.

Because of these factors, the effective sustained copy-in or copy-out rate is likely to be much different (usually much less) than the burst read/write rate of either the source drive and its interface or the destination drive and its controller interface.

With these considerations in mind, let's run some numbers through equation 1, considering the following scenario. You decide to use LTO-6 tapes for copying data. An LTO-6 cartridge can store 2.5 TB of data in uncompressed form.18 LTO-6 supports an uncompressed read/write data speed of 160 MBps.19 Let's make an important simplifying assumption that both the source drive and the destination cloud storage can match the 160-MBps transfer speed of the LTO-6 tape drive (i.e., SpeedcopyIn = SpeedcopyOut = 160 MBps). You choose the overnight shipping option and the shipper requires 16 hours to deliver the shipment (i.e., Ttransit = 16 hours). Finally, let's factor in 48 hours of overhead time (i.e., Toverhead = 48 hours).

Plugging these values into equation 1 and plotting the data-transfer time versus data size using the Ship It! strategy produces the maroon line in figure 3. For the sake of comparison, the blue line shows the data-transfer time of the Upload It! strategy using a fiber-based Internet connection running at 800-Mbps sustained upload rate. The figure shows comparative growth in data-transfer time between uploading at 800 Mbps versus copying it to LTO-6 tapes and shipping it overnight.

Should You Upload or Ship Big Data to the Cloud?

Equation 1 shows that a significant amount of time in the Ship It! strategy is spent copying data into and out of the storage appliance. The shipping time is comparatively small and constant (even if you are shipping internationally), while the drive-to-drive copy-in/copy-out time increases to a very large value as the size of the content being transferred grows. Given this fact, it's hard to beat a fiber-based connection on raw data-transfer speed, especially when the competing strategy involves copy in/copy out using an LTO-6 tape drive running at 160 MBps.

Often, however, you may not be so lucky as to have access to a 1-Gbps upload link. In most regions of the world, you may get no more than 100 Mbps, if that much, and rarely so on a sustained basis. For example, at 100 Mbps, Ship It! has a definite advantage for large data volumes, as in figure 4, which shows comparative growth in data-transfer time between uploading at 100 Mbps versus copying the data to LTO-6 tapes and shipping it overnight.

Should You Upload or Ship Big Data to the Cloud?

The maroon line in figure 4 represents the transfer time of the Ship It! strategy using LTO-6 tapes, while this time the blue line represents the transfer time of the Upload It! strategy using a 100-Mbps upload link. Shipping the data using LTO-6 tapes is a faster means of getting the data to the cloud than uploading it at 100 Mbps for data volumes as low as four terabytes.

What if you have a much faster means of copying data in and out of the storage appliance? How would that compete with a fiber-based Internet link running at 800 Mbps? With all other parameter values staying the same, and assuming a drive-to-drive copy-in/copy-out speed of 240 MBps (50 percent faster than what LTO-6 can support), the inflection point (i.e., the content size at which the Ship It! strategy becomes faster than the Upload It! strategy at 800 Mbps) is around 132 terabytes. For an even faster drive-to-drive copy-in/copy-out speed of 320 MBps, the inflection point drops sharply to 59 terabytes. That means if the content size is 59 TB or higher, it will be quicker just to ship the data to the cloud provider than to upload it using a fiber-based ISP running at 800 Mbps.

Figure 5 shows the comparative growth in data-transfer time between uploading it at 800 Mbps versus copying it at a 320-MBps transfer rate and shipping it overnight.

Should You Upload or Ship Big Data to the Cloud?

Two Key Questions

This analysis brings up the following two questions:

• If you know how much data you wish to upload, what is the minimum sustained upload speed your ISP must provide, below which you would be better off shipping the data via overnight courier?

• If your ISP has promised you a certain sustained upload speed, beyond what data size will shipping the data be a quicker way of hauling it up to the cloud than uploading it?

Equation 1 can help answer these questions by estimating how long it will take to ship your data to the data center. This quantity is (Transfer Time)hours. Now imagine uploading the same volume of data (Vcontent Megabytes), in parallel, over a network link. The question is, what is the minimum sustained upload speed needed to finish uploading everything to the data center in the same amount of time as shipping it there. Thus, you just have to express equation 1's left-hand side (i.e., (Transfer Time)hours) in terms of (a) the volume of data (Vcontent Megabytes); and (b) the required minimum Internet connection speed (Speedupload Mbps). In other words: (Transfer Time)hours = 8 × Vcontent/Speedupload.

Having made this substitution, let's continue with the scenario: LTO-6-based data transfer running at 160-MBps, overnight shipping of 16 hours, and 48 hours of overhead time. Also assume there is 1 TB of data to transfer to the cloud.

The aforementioned substitution reveals that unless the ISP provides a sustained upload speed (Speedupload) of at least 34.45 Mbps, the data can be transferred faster using a Ship It! strategy that involves an LTO-6 tape-based data transfer running at 160 MBps and a shipping and handling overhead of 64 hours.

Figure 6 shows the relationship between the volume of data to be transferred (in TB) and the minimum sustained ISP upload speed (in Mbps) that is needed to make uploading the data as fast as shipping it to the data center. For very large data sizes, the threshold ISP upload speed becomes less sensitive to the data size and more sensitive to the drive-to-drive copy-in/copy-out speeds with which it is competing. The figure shows that the ISP upload speed at which the data-transfer time using the Upload It! strategy matches that of the Ship It! strategy is a function of data size and drive-to-drive copy-in/copy-out speed.

Should You Upload or Ship Big Data to the Cloud?

Now let's attempt to answer the second question. This time, assume Speedupload (in Mbps) is the maximum sustained upload speed that the ISP can provide. What is the maximum data size beyond which it will be quicker to ship the data to the data center? Once again, recall that equation 1 helps estimate the time required (Transfer Time)hours to ship the data to the data center for a given data size (Vcontent MB) and drive-to-drive copy-in/copy-out speeds. If you were instead to upload Vcontent MB at Speedupload Mbps over a network link, you would need 8 × Vcontent/Speedupload hours. At a certain threshold value of Vcontent, these two transfer times (shipping versus upload) will become equal. Equation 1 can be rearranged to express this threshold data size:

Should You Upload or Ship Big Data to the Cloud?

Figure 7 shows the relationship between this threshold data size and the available sustained upload speed from the ISP for different values of drive-to-drive copy-in/copy-out speeds. The figure shows that the change in the break-even data-transfer size after which the Ship It! strategy becomes faster than the Upload It! strategy is a function of ISP-provided upload speed and drive-to-drive copy-in/copy-out speed.

Should You Upload or Ship Big Data to the Cloud?

Equation 2 also shows that, for a given value of drive-to-drive copy-in/copy-out speed, the upward trend in Vcontent continues up to a point where Speedupload = 8/ΔTdata copy, beyond which Vcontent becomes infinite, meaning that it is no longer possible to ship the data more quickly than simply uploading it to the cloud, no matter how gargantuan the data size. In this case, unless you switch to a faster means of copying data in and out of the storage appliance, you are better off simply uploading it to the destination cloud.

Again, in the scenario of LTO-6 tape-based data transfer running at 160-MBps transfer speed, overnight shipping of 16 hours, and 48 hours of overhead time, the upload speed beyond which it's always faster to upload than to ship your data is 640 Mbps. If you have access to a faster means of drive-to-drive data copying—say, running at 320 MBps—your ISP will need to offer a sustained upload speed of more than 1,280 Mbps to make it speedier for you to upload the data than to copy and ship it.

CLOUD-TO-CLOUD DATA TRANSFER

Another strategy is to transfer data directly from the source cloud to the destination cloud. This is usually done using APIs from the source and destination cloud providers. Data can be transferred at various levels of granularity such as logical objects, buckets, byte blobs, files, or simply a byte stream. You can also schedule large data transfers as batch jobs that can run unattended and alert you on completion or failure. Consider cloud-to-cloud data transfer particularly when:

• Your data is already in one such cloud-storage provider and you wish to move it to another cloud-storage provider.

• Both the source and destination cloud-storage providers offer data egress and ingress APIs.

• You wish to take advantage of the data copying and scheduling infrastructure and services already offered by the cloud providers.

Note that cloud-to-cloud transfer is conceptually the same as uploading data to the cloud in that the data moves over an Internet connection. Hence, the same speed considerations apply to it as explained previously while comparing it with the strategy of shipping data to the data center. Also note is that the Internet connection speed from the source to destination clouds may not be the same as the upload speed provided by the ISP.

COST OF DATA TRANSFER

LTO-6 tapes, at 0.013 cents per GB,18 provide one of the lowest cost-to-storage ratios, compared with other options such as HDD or SSD storage. It's easy to see, however, that the total cost of tape cartridges becomes prohibitive for storing terabyte and beyond content sizes. One option is to store data in a compressed form. LTO-6, for example, can store up to 6.25 TB per tape18 in compressed format, thereby leading to fewer tape cartridges. Compressing the data at the source and uncompressing it at the destination, however, further reduces the effective copy-in/copy-out speed of LTO tapes, or for that matter with any other storage medium. As explained earlier, a low copy-in/copy-out speed can make shipping the data less attractive than uploading it over a fiber-based ISP link.

But what if the cloud-storage provider loaned the storage appliance to you? This way, the provider can potentially afford to use higher-end options such as high-end SSDs or a combination HDD-SSD array in the storage appliance, which would otherwise be prohibitively expensive to purchase just for the purpose of transferring data. In fact, that is exactly the approach that Amazon appears to have taken with its AWS (Amazon Web Services) Snowball.3 Amazon claims that up to 50 TB of data can be copied from your data source into the Snowball storage appliance in less than one day. This performance characteristic translates into a sustained data-transfer rate of at least 600 MBps. This kind of a data-transfer rate is possible only with very high-end SSD/HDD drive arrays with read-ahead caches operating over a fast interface such as SATA Revision 3, SAS-3, or PCI Express, and a Gigabit Ethernet link out of the storage appliance.

In fact, the performance characteristics of AWS Snowball closely resemble those of a high-performance NAS (network-attached storage) device, complete with a CPU, on-board RAM, built-in data encryption services, Gigabit Ethernet network interface, and a built-in control program—not to mention a ruggedized, tamper-proof construction. The utility of services such as Snowball comes from the cloud provider making a very high-performance (and expensive) NAS-like device available to users to "rent" to copy-in/copy-out files to the provider's cloud. Other major cloud providers such as Google and Microsoft aren't far behind in offering such capabilities. Microsoft requires you to ship SATA II/III internal HDDs for importing or exporting data into/from the Azure cloud and provides the software needed to prepare the drives for import or export.16 Google, on the other hand, appears to have outsourced the data-copy service to a third-party provider.8

One final point on the cost: unless your data is in a self-managed data center, usually the source cloud provider will charge you for data egress,4,5,12,15 whether you do a disk-based copying out of data or cloud-to-cloud data transfer. These charges are usually levied on a per-GB, per-TB, or per-request basis. There is usually no data ingress charge levied by the destination cloud provider.

CONCLUSION

If you wish to move big data from one location to another over the Internet, there are a few options available—namely, uploading it directly using an ISP-provided network connection, copying it into a storage appliance and shipping the appliance to the new storage provider, and, finally, cloud-to-cloud data transfer.

Which technique you choose depends on a number of factors: the size of data to be transferred, the sustained Internet connection speed between the source and destination servers, the sustained drive-to-drive copy-in/copy-out speeds supported by the storage appliance and the source and destination drives, the monetary cost of data transfer, and to a smaller extent, the shipment cost and transit time. Some of these factors result in the emergence of threshold upload speeds and threshold data sizes that fundamentally influence which strategy you would choose. Drive-to-drive copy-in/copy-out times have enormous influence on whether it is attractive to copy and ship data, as opposed to uploading it over the Internet, especially when competing with an optical fiber-based Internet link.

References

1. Apple. 2015. Thunderbolt; http://www.apple.com/thunderbolt/.

2. Amazon Web Services. 2015. Global infrastructure; https://aws.amazon.com/about-aws/global-infrastructure/.

3. Amazon. 2015. AWS Import/Export Snowball; https://aws.amazon.com/importexport/.

4. Amazon. Amazon S3 pricing. https://aws.amazon.com/s3/pricing/.

5. Google. Google cloud storage pricing; https://cloud.google.com/storage/pricing#network-pricing.

6. Google. 2015. Cloud storage transfer service; https://cloud.google.com/storage/transfer/.

7. Google. Google fiber expansion plans; https://fiber.google.com/newcities/.

8. Google. 2015. Offline media import/export; https://cloud.google.com/storage/docs/offline-media-import-export.

9. Herskowitz, N. 2015. Microsoft named a leader in Gartner's public cloud storage services for second consecutive year; https://azure.microsoft.com/en-us/blog/microsoft-named-a-leader-in-gartners-public-cloud-storage-services-for-second-consecutive-year/.

10. SCSI Trade Association. Oct 14, 2015. Serial Attached SCSI Technology Roadmap; http://www.scsita.org/library/2015/10/serial-attached-scsi-technology-roadmap.html

11. IEEE. 802.3: Ethernet standards; http://standards.ieee.org/about/get/802/802.3.html.

12. Microsoft. Microsoft Azure data transfers pricing details; https://azure.microsoft.com/en-us/pricing/details/data-transfers/.

13. Ookla. 2015. America's fastest ISPs and mobile networks; http://www.speedtest.net/awards/us/kansas-city-mo.

14. PCI-SIG. 2011. Press release: PCI Express 4.0 evolution to 16GT/s, twice the throughput of PCI Express 3.0 technology; http://kavi.pcisig.com/news_room/Press_Releases/November_29_2011_Press_Release_/.

15. Rackspace. 2015. Rackspace public cloud pay-as-you-go pricing; http://www.rackspace.com/cloud/public-pricing.

16. Shahan, R. 2015. Microsoft Corp. Use the Microsoft Azure import/export service to transfer data to blob storage; https://azure.microsoft.com/en-in/documentation/articles/storage-import-export-service/.

17. The Serial ATA International Organization. 2015. SATA naming guidelines; https://www.sata-io.org/sata-naming-guidelines.

18. Ultrium LTO. 2014. LTO-6 capacity data sheet; http://www.lto.org/wp-content/uploads/2014/06/ValueProp_Capacity.pdf.

19. Ultrium LTO. 2014. LTO-6 performance data sheet; http://www.lto.org/wp-content/uploads/2014/06/ValueProp_Performance.pdf.

20. USB Implementers Forum. 2013. SuperSpeed USB (USB 3.0) performance to double with new capabilities; http://www.businesswire.com/news/home/20130106005027/en/SuperSpeed-USB-USB-3.0-Performance-Double-Capabilities.

Sachin Date (https://in.linkedin.com/in/sachindate) looks after the Microsoft and cloud applications portfolio of e-Emphasys Technologies (www.e-emphasys.com). In his past lives, Date has worked as a practice head for mobile technologies, an enterprise software architect, and a researcher in autonomous software agents. He blogs at https://sachinsdate.wordpress.com. He holds a master's degree in computer science from the University of Massachusetts at Amherst.

Copyright © 2016 held by owner/author. Publication rights licensed to ACM.

acmqueue

Originally published in Queue vol. 14, no. 2
Comment on this article in the ACM Digital Library





More related articles:

Matt Fata, Philippe-Joseph Arida, Patrick Hahn, Betsy Beyer - Corp to Cloud: Google’s Virtual Desktops
Over one-fourth of Googlers use internal, data-center-hosted virtual desktops. This on-premises offering sits in the corporate network and allows users to develop code, access internal resources, and use GUI tools remotely from anywhere in the world. Among its most notable features, a virtual desktop instance can be sized according to the task at hand, has persistent user storage, and can be moved between corporate data centers to follow traveling Googlers. Until recently, our virtual desktops were hosted on commercially available hardware on Google’s corporate network using a homegrown open-source virtual cluster-management system called Ganeti. Today, this substantial and Google-critical workload runs on GCP (Google Compute Platform).


Pat Helland - Life Beyond Distributed Transactions
This article explores and names some of the practical approaches used in the implementation of large-scale mission-critical applications in a world that rejects distributed transactions. Topics include the management of fine-grained pieces of application data that may be repartitioned over time as the application grows. Design patterns support sending messages between these repartitionable pieces of data.


Ivan Beschastnikh, Patty Wang, Yuriy Brun, Michael D, Ernst - Debugging Distributed Systems
Distributed systems pose unique challenges for software developers. Reasoning about concurrent activities of system nodes and even understanding the system’s communication topology can be difficult. A standard approach to gaining insight into system activity is to analyze system logs. Unfortunately, this can be a tedious and complex process. This article looks at several key features and debugging challenges that differentiate distributed systems from other kinds of software. The article presents several promising tools and ongoing research to help resolve these challenges.


George Neville-Neil - Time is an Illusion.
One of the more surprising things about digital systems - and, in particular, modern computers - is how poorly they keep time. When most programs ran on a single system this was not a significant issue for the majority of software developers, but once software moved into the distributed-systems realm this inaccuracy became a significant challenge.





© ACM, Inc. All Rights Reserved.