Database Reference
In-Depth Information
[21] Alonso G, Reinwald B, Mohan C. Distributed data management in workflow environ-
ments. In: International workshop on research issues in data engineering high performance
database management for large-scale applications; 1997. p. 82-90.
[22] Rzadca K, Datta A, Buchegger S. Replica placement in P2P storage: complexity and game
theoretic analyses. In: IEEE international conference on distributed computing systems;
2010. p. 599-609.
[23] Yang Y, Liu K, Chen J, Lignier J, Jin H. Peer-to-peer based grid workflow runtime environ-
ment of SwinDeW-G. In: IEEE international conference on e-science and grid computing;
2007. p. 51-58.
[24] Das S, Agrawal D, El Abbadi A. G-Store: a scalable data store for transactional multi-key
access in the cloud. In: ACM symposium on Cloud computing; 2010. p. 163-174.
[25] Juve G, Deelman E, Vahi K, Mehta G, Berriman B, Berman PB, et al. Scientific workflow
applications on Amazon EC2. In: Workshop on Cloud-based services and applications in
conjunction with IEEE international conference on e-Science; 2009.
[26] Ko SY, Hoque I, Cho B, Gupta I. Making cloud intermediate data fault-tolerant. In: ACM
symposium on Cloud computing; 2010. p. 181-192.
[27] Chou CH. Method and apparatus for dynamic bad disk sector recovery. U.S. patent appli-
cation 10/705,809, 2003.
[28] Ewert RA, Schultz SM. Automatic hard disk bad sector remapping. U.S. patent No.
5,166,936, Nov. 24, 1992.
[29] Ruemmler C, Wilkes J. An introduction to disk drive modeling. Computer 1994;27:17-28 .
[30] Huang H, Shin KG. Partial disk failures: using software to analyze physical damage. In:
IEEE conference on mass storage systems and technologies; 2007. p. 185-198.
[31] Huang H, Hung W, Shin KG. Fs2: dynamic data replication in free disk space for improv-
ing disk performance and energy consumption. In: Symposium on operating system prin-
ciples; 2005. p. 263-276.
[32] Patterson D, Gibson G, Katz R. A case for redundant arrays of inexpensive disks
(RAID). In: ACM SIGMOD international conference on the management of data;
1988. p. 109-116.
[33] Li W, Yang Y, Chen J, Yuan D. A cost-effective mechanism for Cloud data reliability man-
agement based on proactive replica checking. In: International symposium on cluster,
Cloud and grid computing; 2012. p. 564-571.
[34] Hughes AK. Making sense of mean time to failure (MTTF), available < http://www.zdnet.
com/blog/hardware/making-sense-of-mean-time-to-failure-mttf/310 > ; 2007.
[35] Wikipedia. Annualized failure rate, available < http://en.wikipedia.org/wiki/Annualized_
failure_rate > ; 2013.
[36] Young JW. A first order approximation to the optimal checkpoint Interval. Commun ACM
1974;17:530-1 .
[37] Gibson GA. Redundant disk arrays: reliable, parallel secondary storage. Vol. 368, Cam-
bridge, MA: MIT press; 1992 .
[38] IDEMA, R2-98: specification of hard disk drive reliability. IDEMA Standards; 1998.
[39] Wikipedia. Magnetic tape data storage, available < http://en.wikipedia.org/wiki/Magnetic_
tape_data_storage > ; 2013.
[40] Amazon. Amazon Glacier, available < http://aws.amazon.com/glacier/ > ; 2013.
[41] Quinlan S, Dorward S. Venti: a new approach to archival storage. In: USENIX conference
on file and storage technologies; 2002. p. 89-102.
[42] Storer MW, Greenan KM, Miller EL, Vorugant K. Pergamum: replacing tape with energy
efficient, reliable, disk-based archival storage. In: USENIX conference on file and storage
technologies (FAST); 2008. p. 1-16.
Search WWH ::




Custom Search