Information Technology Reference
In-Depth Information
connected and linked through SSH protocol, and then information is distributed and
written into these nodes by master node or seed node to achieve remote data center
backup.
3.2
Cross-Platform Data Transfer Using Apache Thrift
Apache Thrift [18] was developed by the Facebook team [19], and it was donated to
the Apache Foundation in 2007 to become one of the open source projects. Thrift was
designed to solve Facebook's problem of large number of data transfers between
various platforms and distinct programming languages, and thus cross-platform RPC
protocols. Thrift supports a number of programming languages [20], such as C++, C#,
Cocoa, Erlang, Haskell, Java, Ocami, Perl, PHP, Python, Ruby, and Smalltalk. With
binary high performance communication properties, Thrift supports multiple forms of
RPC protocol acted as a cross-platform API. Thrift is also a transfer tool suitable for
large amounts of data exchange and storage [21]; when comparing with JSON and
XML, its performance and capability of large-scale data transfer is clearly superior to
both of them. The Input Code is the programming language performed by the Client.
The Service Client is the Client side and Server side code framework defined by
Thrift documents, and read()/write() are codes outlined in Thrift documents to realize
actual data read and write operations. The rest are Thrift's transfer framework,
protocols, and underlying I/O protocols. Using Thrift, we can conveniently define a
multi-language service system, and select different transfer protocol. The Server side
includes the transfer protocol and the basic transfer framework, providing both single
and multi-thread operation modes on the Server, where the Server and browser are
capable of interoperability concurrently.
4
Research Method
4.1
Implementation of HBase and Cassandra Data Centers
The following procedures will explain how to setup HBase and Cassandra data
centers using CentOS 6.4 system, and achieve remote backup. Next, this paper will
test the performance of data centers against reading, writing and remote backup of
large amounts of information.
(1) CentOS's firewall is strictly controlled, to use the transfer ports, one must pre-set
the settings that is a snapshot from the webpage of CentOS's firewall in Chinese
version.
(2) IT manager setup HBase and Cassandra data centers and examine the status of all
nodes.
(3) Forms with identical names must be created in both data centers in HBase
system. The primary data center will execute command (add_peer) [14], and
backup the information onto the secondary data center.
Search WWH ::




Custom Search