Check out the new USENIX Web site. next up previous
Next: 5 Conclusion Up: FastReplica: Efficient Large File Previous: 3.5 Reliable FastReplica Algorithm


4 Performance Evaluation

We outlined the potential performance benefits of FastReplica in the small in Section 3.3. The goal of this section is to analyze the FastReplica performance in the real Internet environment. Through experiments on a prototype implementation, we will demonstrate the efficiency of FastReplica in the small in a wide-area testbed. Since FastReplica in the small defines the iteration step in the general algorithm, these results will set the basis for performance expectations of FastReplica in the large. Using the generous help of summer interns at HPLabs, we built an experimental testbed with 9 nodes. Table 1 and Figure 10 show the 9 hosts participating in our experiments and their geographic locations.

Table 1: Participating nodes.
$N_0$ hp.com Palo Alto, CA
$N_1$ utexas.edu Austin, TX
$N_2$ umich.edu Ann Arbor, MI
$N_3$ gatech.edu Atlanta, GA
$N_4$ duke.edu Durham, NC
$N_5$ uci.edu Irvine, CA
$N_6$ berkeley.edu Berkeley, CA
$N_7$ mit.edu Cambridge, MA
$N_8$ uiuc.edu Urbana-Champaign, IL


The source node is $N_0$ and is located at the HP site, while the nodes-receivers are at different university sites. In order to perform the sensitivity analysis, we vary the number of participating hosts: in experiments with $k$ participating hosts in replication set, the receivers are $N_1, ... , N_k$ ordered as shown in Table 1.
Figure 10: Geographic locations of hosts.
\begin{figure}
\centering
\def 1 ...
Figure 11: Average replication time for files of different size and a different number of nodes in replication set a) 4 receivers, b) 6 receivers, c) 8 receivers.
\begin{figure*}
\centering\def 1 ...
Using the experimental testbed, we compare the following distribution schemes: The experiments are conducted at an application level. Ideally, the transfer time at each recipient node should be measured from the beginning of the transfer at the source node to the completion of the file download at the recipient node. However, due to clock synchronization problems at different nodes, we measure the file transfer time at each recipient node from the beginning of file download to the end of file download at the corresponding recipient node. Since we are interested in large file transfers, the omission of one-way latency of the first packet from the source to the recipient cannot impact the accuracy of the results. In our study, we consider the two performance metrics introduced in Section 3.3: average replication time and maximum replication time. To analyze the efficiency of FastReplica, we performed its sensitivity analysis for replication of different size files and across different numbers of nodes in the replication set. We experimented with 9 file sizes: 80 Kbytes, 750 Kbytes, 1.5 MBytes, 3 MBytes, 4.5 MBytes, 6 MBytes, 7.5 MBytes, 9 MBytes, and 36 MBytes, and varied the number of nodes in the replication set from 2 to 8. When running experiments with different parameters and strategies, the experiments for the same file size were clustered in time as closely as possible to eliminate biases due to short time scale changes in network and system conditions. In order to eliminate the biases due to longer time scale changes in network and system conditions, we performed the same set of experiments at different times of the day. Each point in the results is averaged over 10 different runs which were performed over a 10 day period. Figure 11 shows the average file replication time for experiments with 4, 6, and 8 recipient nodes in the replication set and files of different sizes. For file sizes larger than 80 Kbytes, FastReplica significantly outperforms Multiple Unicast. The replication time under FastReplica is 2-4 times better than under Multiple Unicast.
Figure 12: Maximum replication time for files of different size and a different number of nodes in replication set a) 4 receivers, b) 6 receivers, c) 8 receivers.
\begin{figure}
\centering
\def 1 ...
Figure 13: FastReplica: average vs maximum replication time for a different number of nodes in replication set a) 4 receivers, b) 6 receivers, c) 8 receivers.
\begin{figure}
\centering
\def 1 ...
Additionally, in most experiments, FastReplica outperforms Sequential Unicast, which approximates the file replication with IP multicast. The explanation is that when Sequential Unicast replicates file $F$ to $n$ nodes, it uses $n$ Internet paths connecting the source nodes to the recipient nodes (while sending only one packet over each common link in those paths). Thus the overall performance is defined by the end-to-end properties of the $n$ paths. Congestion in any of those paths impacts the overall performance of the Sequential Unicast. FastReplica uses the same $n$ paths between the source and recipient nodes to transfer only $1\over
n$-th of file $F$. FastReplica takes advantage of using the additional $(n-1) \times n$ paths between the nodes in the replication set, and each of those paths is used for sending $1\over
n$-th of file $F$. Thus, the congestion in any of those paths impacts FastReplica performance for transfer of only the $1\over
n$-th of file $F$. While the average replication time provides an interesting metric for distribution strategy characterization, the metric representing the maximum replication time is critical, because it reflects the worst case of the replication time among the recipient nodes. Figure 12 shows the maximum replication time for experiments with 4, 6, and 8 recipient nodes in a replication set and files of different sizes. The maximum replication times under Multiple Unicast, as well as Sequential Unicast, are much higher than the corresponding average times for these strategies. For a case of 8 nodes in the replication set, the maximum times under Multiple Unicast and Sequential Unicast are almost 2 times higher than the corresponding average times. The reason is that there is a very limited bandwidth on the path from the source node $N_0$ to the recipient node $N_8$. The performance of this path is practically the same for both Multiple Unicast and Sequential Unicast. This path defines the worst (maximum) replication time among all the recipient nodes in the set. Since FastReplica uses this path to transfer only $1\over
n$-th of file $F$, this ``bad'' path has a very limited impact on maximum replication time and overall performance of FastReplica. Figure 13 shows how close the average and maximum replication times under FastReplica are. These results demonstrate the robustness and predictability of performance results under the new strategy.
Figure 14: Replication time measured by individual receiving nodes for 9 MB file and 8 nodes in replication set.
\begin{figure}
\centering
\def 1 ...
Figure 14 shows the average replication time measured by the different, individual recipient nodes for a 9 MB file and 8 nodes in the replication set (the other graphs for different file sizes and a different number of nodes in the replication set reflect similar trends). There is a high variability of replication time under Multiple Unicast and Sequential Unicast. This is somewhat expected because the file replication times at the individual nodes highly depend on the available bandwidth of the path connecting the source and receiver node. The limited bandwidth of the path between the original node $N_0$ and the receiver node $N_8$ can be observed from these measurements, and it severely impacts the overall performance of both Multiple Unicast and Sequential Unicast. The file replication times under FastReplica across different nodes in the replication set are much more stable and predictable since each node performance is defined by the bandwidth of $n$ paths, each transferring $1\over
n$-th of the original file $F$.
Figure 15: Average file replication time for a different number of nodes in replication set and a) File size of 1.5 MB, b) File size of 9 MB, c) File size of 36 MB.
\begin{figure*}
\centering
\def 1 ...
Figure 16: Speedup in average and maximum file replication time under FastReplica vs Multiple Unicast for a different number of nodes in replication set and a) 1.5 MB file, b) 9 MB file, c) 36 MB file.
\begin{figure*}
\centering
\def 1 ...
Figure 17: Average file replication time for a different number of nodes in replication set and a) File size of 80 KB, b) File size of 750 KB, c) Speedup in file replication time under FastReplica vs Multiple Unicast for 750 KB file.
\begin{figure*}
\centering
\def 1 ...
Figure 15 shows the average replication time for files of 1.5 MB, 9 MB, and 36 MB for a different number of nodes in the replication set. While Multiple Unicast shows a growing replication time for an increasing number of nodes in the replication set, FastReplica and Sequential Unicast demonstrate good scalability for replication sets of different sizes. Additionally, FastReplica consistently outperforms Sequential Unicast for most of the points. Figure 16 shows the average and maximum speedup of file replication time under proposed FastReplica in the small relative to the replication time of Multiple Unicast for files of 1.5 MB, 9 MB, and 36 MB, and a different number of nodes in the replication set. The results consistently show the significant speedup both in average and maximum replication times across considered different file sizes. Finally, Figures 17 a) and b) show the average replication time for 80 KB and 750 KB files and a different number of nodes in the replication set. The files of 80 KB and 750 KB are the smallest ones used in our experiments. For the 80 KB file, FastReplica is not efficient, and the replication time (both average and maximum) is higher than under Sequential Unicast and Multiple Unicast. For the 750 KB file, the replication time under FastReplica is better than under Sequential Unicast and Multiple Unicast strategies, and the average and maximum speedup shown in Figure 17 c) is again significant. These results help to outline the ``border'' parameters for new strategy usage: in our case study, FastReplica works most efficiently for replicating files larger than 0.5 MB. For $n>8$, the ``border'' file size, where FastReplica works most efficiently, may increase correspondingly.
Figure 18: Different configuration with $N_1$ (utexas.edu) being the origin node: speedup in average and maximum replication times under FastReplica vs Multiple Unicast for a different number of nodes in replication set and a) 1.5 MB file, b) 9 MB file, c) 36 MB file.
\begin{figure*}
\centering
\def 1 ...
The results presented in Figure 16 show the significant speedup both in average and maximum replication times under the FastReplica strategy. The additional analysis reveals that the available bandwidth of the paths between the origin node $N_0$ (hp.com) and nodes $N_1, ..., N_7$ (universities' machines) is significantly lower than the cross bandwidth between nodes $N_1, ..., N_7$. only node $N_8$ has a limited incoming bandwidth from all the nodes $N_0, N_1, ..., N_7$, while the outgoing bandwidth from node $N_8$ to $N_1, ..., N_7$ is again significantly higher. In such a configuration, FastReplica utilizes the abundance of additional available bandwidth between the replication nodes in the most efficient way to produce the spectacular results. It is interesting to see how FastReplica would perform when a different node with high bandwidth paths to the rest of the nodes is used as the origin node. We changed the configuration and made node $N_1$ (utexas.edu) to be the origin node, and rerun the experiments again. Figure 18 shows the average and maximum speedup of file replication time under the proposed FastReplica in the small relative to the replication time of Multiple Unicast for files of 1.5 MB, 9 MB, and 36 MB, and a different number of nodes in the replication set in the new configuration. In the new configuration, the average replication times under FastReplica and Multiple Unicast are similar, but the maximum replication time under FastReplica is still significantly better than the maximum replication time under Multiple Unicast. The bandwidth analysis reveals that node utexas.edu is connected to the rest of the nodes via high bandwidth paths with low bandwidth variation across these paths. Our analysis in Section 3.3 with a specially designed example, where the bandwidth matrix $BW$ is defined by equations (6), demonstrates that when the cross bandwidth between some replication nodes is significantly lower than the bandwidth of the original paths from $N_0$ to the recipient nodes $N_1, ..., N_8$ then FastReplica improves the maximum replication time but may have no significant improvement in average replication time.
next up previous
Next: 5 Conclusion Up: FastReplica: Efficient Large File Previous: 3.5 Reliable FastReplica Algorithm