发明名称 Random next iteration for data update management
摘要 Host machines and other devices performing synchronized operations can be dispersed across multiple racks in a data center to provide additional buffer capacity and to reduce the likelihood of congestion. The level of dispersion can depend on factors such as the level of oversubscription, as it can be undesirable in a highly connected network to push excessive host traffic into the aggregation fabric. As oversubscription levels increase, the amount of dispersion can be reduced and two or more host machines can be clustered on a given rack, or otherwise connected through the same edge switch. By clustering a portion of the machines, some of the host traffic can be redirected by the respective edge switch without entering the aggregation fabric. When provisioning hosts for a customer, application, or synchronized operation, for example, the levels of clustering and dispersion can be balanced to minimize the likelihood for congestion throughout the network.
申请公布号 US8745264(B1) 申请公布日期 2014.06.03
申请号 US201113076932 申请日期 2011.03.31
申请人 Amazon Technologies, Inc. 发明人 Marr Michael David
分类号 G06F15/16 主分类号 G06F15/16
代理机构 代理人
主权项 1. A computer-implemented method of managing data in a networked environment, comprising: under control of one or more computer systems configured with executable instructions, receiving workload data having a specified terminating address for one of a group of host servers selected to process a common set of the workload data for a customer, the group of host servers being connected by a network aggregation fabric including layers of network switches, a path across the network aggregation fabric to each of the host servers in the group of host servers involving a number of connections across the layers of network switches, the group of host servers being selected to process workload data for the customer and being dispersed across a number of network switches for at least a lowest layer of the network aggregation fabric;routing the workload data to a host server of the group of host servers corresponding to the specified terminating address and processing the workload data using a processing host server of the group of host servers; andin response to processing the workload data using the processing host server, selecting a random ordering of the other host servers in the group of host servers and sending updates of the workload data to each of the other host servers in the group of host servers according to the random ordering,wherein updates to be periodically shared across all the host servers in the group of host servers are sent with randomized orderings in order to reduce a statistical likelihood of network congestion due to flow convergence.
地址 Reno NV US