Accelerating Computation of DNA Multiple Sequence Alignment in Distributed Environment

®

http://iaesjournal.com/online/index.php/TELKOMNIKA/index

[

USER

i

r---··---·------------;
i

AMセZN@

Accelerating Computation of DNA Multiple Sequence Alignment in
Distributed Environment

BセGᄚN[I@

. セN@



i
.,.I

vャセキ@

• Sub.crlbe / U11>ub>uob"

DOl: 10.11591/telkomnika .v12il2 .6572:

Abstrdct
mオャ|Qーセ@

NセャGqjョ・@
al1=;inmerrt (MS;\) エセ@ '-' t\'_etmiqul' for firdm.;i セQュゥャj[イエL@
"' mJ- "1' lt:e イセゥ[ッウ・、@ー

t(· hセ・ャイGA^エ@


Ti-u5 tecM"l.Ub is Yer'{ •n'•IJ01tant to セオーッイエ@
ヲイ。ァュ・ョエセ@
b1nninq. Th!! セゥイョdQVエ@
dtna1nic pG\ャセュゥョ_@
tH.hn1q•Jl' 1> v"rr

.:.emputation of Star mセエQーャ・@
Sセᄋ|Q・ョ」@
エ|ィN[ュセョ@
•Js•nsi MH5111JO p。ウセQョァ@
lnterfaces HmセiゥN@
The
LNGjャオセエ・、@
by イNャAic|セエ[ョ\j@
セーNゥ・、オ@
E1p•mrnent '"as conducted オセュ@
{..1 セLアオ・ョ」Z@
of セo@
bp
yiold"'d t·v randomly out from Nf"a-rencx. セッエゥオᄋュ\Z・@

of cMャ[ッゥョ・イmク」エュ。hᄋセAb@
エ。セ・ョ@
lrom N".":51 HセjエャッイQゥi@
BNZ・ャ|エセイ@
for bQエ|」セョNLャッG@
!r,forrMt•on). TM 1\'•;••ll's DィoセG、@
lhat エィセ@
ーイッセj、@
tochr.ique could \'lbtaio セャG、オ@
thr Pn
Illustration of data distribution of DNA sequence with MPI can be seen in Figure 6. In
this research, comunication of MPI used point to point communication with a blocking send and
receives operations. Parallelization scheme for MSA is to assign a computer as the data divider
and other computers as data processors. A data divider called rank 0 distributed sequence
pairs using MP/_Send() to other computer as data processors (Figure 7). Data processors
which were symbolized by the rank 1... n received sequence pairs using MPl_Recv() (Figure 8).
Next, each data processors conducted pairwise sequence alignment to compute pairwise
similarity scores in parallel. The calculation results of each data processor were transmitted to
data divider (rank 0) by using MP/_Send(). Data divider with rank 0 received the similarity
scores using MP/_Recv() command and completing the MSA process by selecting a Star

sequence and realigning all sequence to the Star sequence

8282



ISSN: 2302-4046
I
I

------,

I
I

I
r----------IMPl_Send(K1)/ MPl_Recv(K1)--------'f-

,--


セML@

Dividersl :
data
I
: ャPエセᄆqゥjUQZMmp⦅r・」カHsュ。イケ@
\

I ゥヲjャFイョZセ@ェ
I セGBM

MPl_Recv{Similarity Score K1) f MPl_Send(Simi!arity Score K1)

IF""'=wi

I

:
U Qhャエj@


I
I

I
セMmpャ⦅s・ョ、HkRI@

I
1

Rank 1

:

'

'-,--l- - "

I

I


i

i

l. . . . . . . . . . . . . . . . . . . . . . . .

MPl_Seod(Ko) /

L-----------------·----·--·· MPl_Recv(Similarity Score Kn) I MPl_Send(Slmllar!ty Score Kn)···

Rank2

I
I

mpャ⦅r・ッカHkIMᄋセ@
0

·------------:-


ャ[hゥセQL@

D

... - _,RjWl\Jl - ...

Figure 6. Illustration the distribution of sequence data with MPI
llMPI initialization
MPl_lnit(&argc, &argv);
MPl_Comm_rank (MPl_COMM_WOR.LD, &rank);//Computer init_ialization
MPl_Comm_size (MPl_COMM_WORLD, &p_size);//lnitialize the.number of computers
If rank== 0 {
c = ((js*js)'js)/2; II Calculate the many pairwise sequences
p_bts =.p_size-1; II Limit Number of Computers
p=1; II Limit Number of computers for processing data (other than rank 0)
for (i=1; i