Real Time Hand Gesture Movements Tracking and Recognizing System - repository civitas UGM

  • . / 0 1 2 3%)4& $.0 5+,% % +55%65 +%

   !

! ! " #$%#&'(

  ) * +% , ( *

  

2014 Electrical Power, Electronics, Communications, Controls and Informatics

Seminar (EECCIS)

Copyright ©2014 by the Institute of Electrical and Electronics Engineers, Inc. All rights

reserved. Copyright and Reprint Permission

Abstracting is permitted with credit to the source. Libraries are permitted to photocopy

beyond the limit of U.S. copyright law, for private use of patrons, those articles in this

volume that carry a code at the bottom of the first page, provided that the per-copy fee

indicated in the code is paid through the Copyright Clearance Center, 222 Rosewood

Drive, Danvers, MA 01923.

Other copying, reprint, or reproduction requests shouldbe addressed to IEEE Copyrights

Manager, IEEE Service Center, 445 Hoes Lane, P.O. Box 1331, Piscataway, NJ 08855-

1331.

  IEEE Catalog Number Part Number CFP 1432Z-ART

  ISBN 978-1-4799-6947-0 Additional copies of this publication are available from Currant Associates, Inc.

  57 Morehouse Lane Red Hook, NY 12571 USA

  • 1 845 758 0400
  • 1 845 758 2633 (FAX) email: curran@proceedings.com

    For information about producing a conference proceedings and receiving an estimate,

    contact conferencepublishing@ieee.org http://www.ieee.org/conferencepublishing

  2014 Electrical Power, Electronics, Communications, Controls, and Informatics Seminar ( EECCIS)

  Table of Contents

A. Electrical Power

  Characteristics and Suppression of Secondary Arc on 500 kV Transmission Lines for

Single Pole Reclosure Purposes A1_01-07

Kevin Marojahan B.N., Nanang Hariyanto, Muhammad Nurdin Design of Robust-Fuzzy Controller for SMIB Based on Power-Load Cluster Model with Time Series Analysis

A2-08-15

  Ismit Mado, Adi Soeprijanto, Suhartono Impacts ofExhaust Airflow Velocity to Electrical Power Generation of

Small Wind Energy Conversion System (WECS) A3_16-22

Teguh Utomo, Unggul Wibawa Neural Network SVPWM-DTC of Induction Motor for EV Load Model A4_23-28 Sy Yi Sim, Zainal Alam Haron, Nooradzianie Muhammad Zin, Wahyu Mulyo Utomo, Azuwien Aida Bohari, Roslina Mat Ariff Security Constrained Optimal Power Flow Incorporating Preventive and Corrective Control A5_29-34 Rony Seto Wibowo, Tri Prasetya Fathurrodli, Ontoseno Penangsang, Adi Soeprijanto

B. Electronics

  

A Microcontroller-based Yarn Twist Tester B1_35-39

Daniel Santoso, Deddy Susilo, Yahya Y. Prasetyanto Development of Two Axis Solar Tracking Using Five Photodiodes B2_40-44 Munnik Haryanti, Abdul Halim, Arbai Yusuf Performance of LED Lights Installed on DKI Jakarta Streets B3_45-50 Endah Setyaningsih, Joni Fat, Lydwina Wardhani, Ida Zureidar The Development of Computerized Measurement System for Electronic Device Characterization B4_51-55

C. Communications

  Frequency domain Technique For Reducing Chromatic Dispersion C1_56-61 M.F.L Abdullah, Bhagwan Das, Mohd Shah Nor Shahida Investigations on Fractal Square Loop FSS at Oblique Incidence for GSM Applications C2_62-66

F. C. Seman, N. K. Khalid

  The Influence of Hole-Truncated to Characteristic Performance of The Equilateral Triangular C3_67-69 Antenna for Mobile Satellite Communications Muhammad Fauzan Edy Purnomo, Joshapat Tetuko Sri Sumantyo, Vita Kusumasari The Intelligent Energy Harvesting Management Policies in Wireless Sensor Networks with C4_70-77 Directional Water-Filling Algorithm Dion Firmanda, Sigit Wasista, Tri Harsono

  2014 Electrical Power, Electronics, Communications, Controls, and Informatics Seminar ( EECCIS)

D. Controls

  An LMI Approach to Output Feedback Controller of Neutral Systems D1_78-81 Erwin Susanto, Junartho Halomoan, Mitsuaki Ishitobi Design of Geometric-based Inverse Kinematics for a Low Cost Robotic Arm D2_82-86 Muhammad Aziz Muslim, Seif Nura Urfin Forward Kinematics Analysis of a 5-Axis RV-2AJ Robot Manipulator D3_87-92 Mohammad Afif Ayob, Wan Nurshazwani Wan Zakaria, Wan Nurshazwani Wan Zakaria Inverse Kinematics of a Two-Wheeled Differential Drive an Autonomous Mobile Robot D4_93-98 Eka Maulana, Muhammad Aziz Muslim, Akhmad Zainuri

Temperature Control of Liquid Egg Pasteurization System Using PLC D5_99-104

(Programmable Logic Controller) Siemens Simatic S7-200 and HMI (Human Machine Interface) Simatic HMI Panel Tri Oktaviana Putri, Rahmadwati, Bambang Siswojo

E. Informatics

  

Assessing Readiness of IP Networks to Support H.323 Desktop Videoconferencing E1_105-110

Services over Various Scheduling Techniques Using OPNET Eko Fajar Cahyadi

Comparison Binary Search and LinearAlgorithm for German-Indonesian Sign Language E2_111-115

Using Markov Model Fridy Mandita, Toni Anwar, Harnan Malik Abdullah

Content Based Image Retrieval System Based on Watershed Transform for Trademark Images E3_116-120

Cahyo Crysdian

Design of Small Smart Home System Based on Arduino E4_121-125

Andi Adriansyah, Akhmad Wahyu Dani

Enhance Quality Control Management for Sensitive Industrial Products Using 2D/3D Image E5_126-131

Processing Algorithms Murthad Al-Yoonus, Mohammed Saeed Jawad, M.F.L Abdullah, Fares Al-Shargie

Implementation of Modified Vernam Algorithm for Radioactive Transportation E6_132-136

Monitoring System Adi Abimanyu, Dwi Yuliansari N, I Wayan Mustika, Litasari

Real Time Hand Gesture Movements Tracking and Recognizing System E7_137-141

Rudy Hartanto, Adhi Susanto, P. Insap Santosa

Voice Control of Home Appliances using Android E8_142-146

Norhafizah bt Aripin, M. B. Othman

  978-1-4799-6947-0/14/$31.00 ©2014 IEEE 137

  The development of user interfaces has increased rapidly, especially in consumer devices. The use of the touch interface (e.g. touch screen) has commonly in today devices such as computer tablets, Smart Phones, and other control interface on an electronic device. The development of user interface are tends to the use of human body gesture, especially human hand as the main interface, which is more natural and intuitive. However, so far the user interface is still limited to movement on the two dimensions surface, such as the use of the mouse and touch screen.

  The objective of this research is making a system prototype that can track the movement of the hand as well as to recognize the meaning of simple hand gestures command as a replacement of the mouse device. We propose a new approach and algorithms for hand gesture recognition that consist of two-step. The first step is hand gesture pose recognition process. Beginning with the identification process and separation of hand gestures image captured by ordinary webcam from their background images including the body and the head of the user. The segmentation and elimination of background from hand image is obtained using skin color segmentation algorithm combined with reduction of the background image. We use YCrCb color space instead of

  A. Research Objectives

  ETHOD

  M

  ROPOSED

  P

  HE

  II. T

  [5] had developed a framework interface for immersive projection system CAVE in the field of virtual reality by using hand gestures. Hand gesture recognition is performed with the use of blob detection. This recognition model allowing the implementation of an effective interface. Other research on the recognition and tracking of a hand gesture motion done by [6]. Tracking and recognition of hand gestures motion, which is captured by web cameras, are used to control the computer cursor as a replacement of the mouse. [7] conduct research to develop a system using vision-based interface webcam cameras and computer vision technology that allows users to use their hand to interact with the computer. In this research, the user can interact with a computer using their hand movements without the use of a marker. [8] use hand segmentation and color model to track the movement of the hand. Segmentation and tracking process is done by using a color model to extract or identify the gestures hand to implement real-time applications that are better, faster and more accurate.

  The use of pointing device with tactile feedback as a mechanism of interaction with the computer have been examined by [2] in 2008. In the same year, Juan P. Wachs developed a method of vision-based system that capable to interpret the user's hand gesture to manipulate objects on medical data visualization environment [3]. Tao Ni [4] developed the technique called rap Menu (roll-and-pinch menu) interactions to select a specific menu remotely by using hand gesture input.

  Many researches explore various mechanisms of interaction with the computer in a more practical, efficient, and comfortable. Starting from the use of devices like ergonomic keyboard interaction based on gesture hands [1], to the effort to seek mechanisms of interaction without having to get in touch directly with the computer input devices.

  Keywords - hand gesture, human computer interaction, contours, convex hull, Lukas-Kanade Algorithm.

  

Real Time Hand Gesture Movements Tracking and

Recognizing System

Rudy Hartanto

  Abstract— Human computer interaction has a long history to become more intuitive. For human being, gesture of different kind is one of the most intuitive and common communication. However, vision-based hand gesture recognition is a challenging problem, which is involved complex computation, due to high degree of freedom in human hand. In this paper, we use hand gesture captured by web-cam instead of mice, for natural and intuitive human-computer interaction. Skin detections method is used to create a segmented hand image and to differentiate with the background. A contours and convex hull algorithm is used to recognize hand area as well as the number of fingertip of hand gesture image to be mapped with button. Moreover, for detection of hand gesture motion, we use Lucas-Kanade pyramidal algorithm. The result shows that this system can operate well so we can interact with computer using our hand gesture instead using a mouse.

  Jalan Grafika No. 2, Yogyakarta, 55281 Indonesia Email: insap@mti.ugm.ac.id

  Technology Gadjah Mada University

  Department of Electrical Engineering and Information

  P. Insap Santosa

  Jalan Grafika No. 2, Yogyakarta, 55281 Indonesia

  Technology Gadjah Mada University

  Department of Electrical Engineering and Information

  

Adhi Susanto

  Jalan Grafika No. 2, Yogyakarta, 55281 Indonesia Email: rudy@mti.gadjahmada.edu

  Technology Gadjah Mada University

  Department of Electrical Engineering and Information

I. I NTRODUCTION

  RGB color space to eliminate the variation of room illumination. The convexity hull algorithm is applied to recognize hand gesture contour area as well as the number of fingertips to map with the mouse button.

  with Cr1 and Cr2 as well as Cb1 and Cb2 are the range of hand skin color value. As we know that human skin color varies across human races or even between individuals of the same race. Additional variability may be introduced due to changing illumination conditions and/or camera characteristics. Therefore, the range value of hand skin color should be adjust according to the user skin color as well as illumination conditions and/or camera characteristics.

  Convexity defect

  D.

   ! ! ! " % ! ! ! ! " & % ! ) ! " ! ! ! " ( # ! " ! ! " % ! " ' ! ! " % $ Fig. 2. Contour detection.

  Contour is found if there are differences in pixels with its neighbors. The contour can be obtain from binary image based on the results of skin detection between white pixels (the object that has the colors resemble hands) and black pixels (objects other than hands or background). This method however, has the possibility of a high error when an object in an image has a color that is almost the same. Therefore, need to be maintained so that there is only one object in the frame, which is the biggest contour (which in this case is the user's hand). Fig 2 shows the flow chart of contour detection.

  Contour Detection and The contour is a sequence of pixels with same value.

  C.

  (2)

  The second step is tracking and recognizing the moving of user hand that recorded by webcam using Lukas-Kanade pyramidal algorithm. The tracking of hand gesture motion will be used to move the cursor according to the hand movement. Fig. 1 show the system proposed, starting by capture the user hand image, removing the background and segmented process. The detection of hand image is accomplished by detecting the hand contour and determines the hand center point. The fingertips detection is obtained with convexity defect and convex hull algorithm, and will be used as the replacement of mouse buttons click. The motion detection will track the hand motions to move the cursor as a replacement of the mouse function.

  The skin color detection can be obtained by determining the range of Y, Cr, and Cb values respectively, which is the value of the user's skin color to be used as a threshold to determine whether the value of a white pixel (hand image) or black (background). Equation (2) shows the algorithm of skin color detection and threshold process to convert to binary- segmented image.

  Y = 0.299*R + 0.587*G + 0.114*B Cr = (R - Y)*0.713 + 128 (1) Cb = (B - Y)*0.564 + 128

  The segmentation process is used to obtain the hand image to be processed further. We use skin detection algorithm to detect and separate hand skin color with the background due to its simple and fast detection. To improve the reliability from the variation of environment illumination, the RGB color space image will be converted to YCrCb. This conversion will separate the intensity component Y and chromaticity components Cr and Cb. The conversion obtained using (1)

   Segmentation

  B.

  Image Acquisition Skin Detection & Background Removal Noise Filtering & Segmentation Contour Detection & Finding Hand Center Point Convexity Defect Fingertips Detection Motion Detection Mouse Functions Calibration Fig. 1. The Proposed System.

  This method is intended to detect the user's fingers captured by the camera. To extract hand features we use two- steps process. First step is construct a convex hull using Sklansky algorithm. Convex hull is a set of points that constitute a convex polygon covered this points set. Sklansky convex hull from set of points or a simple polygon [9]. It is suitable for polygon with star shape and hand image as well. Sklansky algorithms perform scanning in counter clockwise direction in order to remove a vertex point p from a polygon P. Sklansky defines P as a collection of points that perform a polygon or as a closed region polygon-shaped. Fig. 3 shows the shape of a Sklansky square that illustrating the formation of hull using this algorithm. a. If startpoint.y < box.center.y atau depthpoint.y <

  box.center.y, indicate that startpoint and depthpoint is

  above the center point of contour, or in other word the palm is upward.

  b. If startpoint.y < depthpoint.y, indicate that startpoint is above depthpoint, or in other word the finger is pointed upward.

  c. If the distance between startpoint and depthpoint > box

  Fig. 3 Sklansky Square. [10]

  height/6.5, indicate that the detected hand contour is too long including the user arm, and should be The second step is determining the defect from convext restricted to the palm only. hull that have been formed. Defect can be defined as a

  Fig. 5 shows the flow chart of detecting and counting smallest area in the convex hull surrounding the hand. Search number of fingers. for defect is useful to find hand’s features, so that it can eventually detect the number of fingers. Convexity defects method on this hand can be seen in Fig. 4.

  ) $ ) $ ) ( % ! *

  ! * ! ! ! ! ! * )

  ! ! Fig. 4. Convexity defect of hand image [10].

  The convex hull searching from a set of points Q(CH(Q)) obtained by searching smallest convex set that contain all of

  • "' &

  the points in Q. A convex hull of a set of points Q (CH (Q)) in n dimensions is a whole slice of all convex sets containing Q. Thus, for an N-point {p1, p2, ..., pN} P, the convex hull is a

  () %

  convex combination set that can be expressed as: Fig. 5. Flow chart of detecting and counting number of finger.

  (3)

  F. Motion Detection

E. The Number of Fingers Detection

  The motion detection is accomplished by tracking the user hand motion, and implemented using Lukas-Kanade To detect the number of hand fingers can be implemented pyramidal method, showed in Fig. 6. in five steps:

  1. Determine the center point coordinate (x,y) of minimum area rectangle of hand contour (further called box). defects founded.

  3. Set the number of finger = 0.

  4. For i = 0, to i < defect number do

  If ((startpoint.y < box.center.y) or (depthpoint.y < box.center.y)) And (startpoint.y < depthpoint.y) and (the distence of startpoint and depthpoint > box height/ 6,5) Then number of fingers = 0.

  5. Display detected fingers. The value of variable “number of fingers” will increase Fig. 6. Motion detection flow chart. following the number of defect found, if it satisfies the following three requirements.

G. Mouse Function Implementation

  In this research, the hand gesture motion detection and recognition will be implemented to replace the mouse function. The mouse function can be implemented using mouse event functions from API call to synthesize the mouse motion and button click, which found in user32.dll.

  (a) (b) (c) To implement mouse move, right and left click, double Fig. 8. (a) Original Image, (b) YCrCb image, (c) Segmented image. click and drag function, we use the following mouse event functions.

  B. Contours Detection

MOUSEEVENTF_MOVE

  • Move; is implemented using

  Binary image from segmentation results will then use to

  • Left click; is implemented using contour searching. The contour searching will produce more

  MOUSEEVENTF_LEFTDOWN followed with

  than one contour due to the imperfect segmentation, for

  MOUSEEVENTF_LEFTUP example, there are other similar objects by hand but smaller.

  • Right click, is implemented using

  In order to ensure that only one contour, that is hand contour is

  MOUSEEVENTF_RIGHTDOWN followed with

  processed, and then need to specified areas in which the

MOUSEEVENTF_RIGHTUP

  largest contour. So only, the contour with the greatest area is

MOUSEEVENTF_LEFTDOWN

  • Drag, is implemented using to be formed, while the other will be ignored. This contour continuously while move to the intended directions.

  search results are shown in Figure 9. The applied mouse function according to the number of fingers detected as shown in Fig. 7.

  " " Fig. 9. The resulted contours search .

  " C. Convexity Defect and Number of Fingers Detection

  The process begins with the formation of the convex hull, than followed by searching the defects areas, which can be

  "

  used to detect fingers. Convex hull that is found will be marked by a small circle with blue color, as shown in Figure

  10.

  " " ! Fig. 7. Mouse function implementation.

  XPERIMENT R ESULTS

III. E

  To verify the effectiveness of the proposed approach, we done some testing using I7 1.7 GHz machine and the image Fig.10. Convexity defect of hand fingerprints. sequences are captured at about 10 frames per second.

  The convex hull that has been resulted will be used to

A. Segmentation using Skin Detection determine the number of fingers. In the real world however,

  the illuminations are not perfect and always varies resulted The segmentation process is intended to get hand image that the number of circles that are detected not only in the and separate it with the image background. The RGB image fingertips only. In this case needs to be specified that circle captured from the camera will first converted to YCrCb color use for counting only on the upper part of the image, and the space to separate between the intensity component (Y) and maximum amount is five in accordance with the number of color components (Cr and Cb). This conversion will reduce hand fingers, and then will be calculated and displayed on the the pixels variation due to room illumination variation. Fig. 8 screen. From the observations as shown on Figure 11, this shows the original image in RGB space, image in YCrCb space and segmented image resulted using skin detection. system is capable to detect the number of fingers shown by users from one finger to five fingers.

  Fig. 11 The detection and number of fingers recognition.

  The future work is to improve segmentation process in eliminating the effect of illumination changes. It is possible to use another color space such as HSV to minimize the effect of illumination changes.

  [8] S. Patidar and D. C.S.Satsangi, "Hand Segmentation and Tracking Technique using Color Models," International Journal of Software & Hardware Research in Engineering, vol. 1, no. 2, pp. 18-22, 10 2013.

  [7] R. Bhatt, N. Fernandes and A. Dhage, "Vision Based Hand Gesture Recognition for Human Computer Interaction," International Journal of Engineering Science and Innovative Technology (IJESIT), vol. 2, no. 3, pp. 110-115, 2013.

  [6] D. C. Gope, "Hand Tracking and Hand Gesture Recognition for Human Computer Interaction," American Academic & Scholarly Research Journal, vol. 4, no. 6, pp. 36-42, 11 2012.

  [5] H. Lee, Y. Tateyama and T. Ogi, "Hand Gesture Recognition using Blob Detektion for Immerseive Projection Display System," World Academy of Science, Engineering and Technology, vol. 6, 27 02 2012.

  [4] T. Ni, R. McMahan and D. Bowman, "Tech-note: rapMenu: Remote Menu Selection Using Freehand Gestural Input," in 3D User Interfaces, 2008. 3DUI 2008. IEEE Symposium on , 2008.

  Handler, "A Real-Time Hand Gesture Interface for Medical Visualization Applications," Applications of Soft Computing, Advances in Intelligent and Soft Computing, vol. 36, pp. 153-162, 2006.

  Displays," in Workshop on Multimodal Interaction Through Haptic Feedback, 2008. [3] J. Wachs, H. Stern, Y. Edan, M. Gillam, C. Feied, M. Smith and J.

  EFERENCES [1] V. Adamo, "A New Method of Hand Gesture Configuration and Animation," Information, pp. 367-386, 2004. [2] S. Foehrenbach, W. A. König, J. Gerken and H. Reiterer, "Natural Interaction with Hand Gestures and Tactile Feedback for large, high-res

  R

  Based on the evaluation and discussion above, can be concluded that the system can recognize and track the hand movement and can replace the mouse function to move the mouse cursor and the mouse click function as well. The segmentation algorithm using skin color detection is good enough to separate the hand image with its background. The convexity defect algorithm can work well in detecting the number of user’s hand, to replace the mouse click functions. In general, the system can detect and follow the hand movement so that can be used as user interface in real time.

  D. Motion Detection

  ONCLUSION

  IV. C

  However, there are some limitations of the program, i.e. the change of lighting environment is still influential in the process of segmentation, in particular on the process of the removal of the background.

  The segmentation process is capable in detecting the user's hand gestures and counting the contours. The system also can count the number of user fingers that associated with a command button on a mouse. The hand motion can also been followed properly by the system, and the mouse click function represented by number of fingers detected have work.

  To evaluate the performance of recognition proposed, we examine the system to operate the windows application that is MS Paint using hand gesture to replace mouse as user application interface. Testing is done under normal illumination, system can recognize, and track the movement of user hand. The center of hand gesture represented by a small circle had successfully follow the user's hand movement.

  E. Discussion

  To find a center point of the hand contours is done by searching for a center point of the box enclosing the hand contour, as shown in Fig. 12. Further, by using the box.center.x and box.center.y as the coordinates of the center point contours, mouse cursor can be arranged to follow the contour, represented as a small circle in red.

  Fig. 12. The center point on hand contours .

  The detection of hand movement is required so that the mouse pointer can follow the user's hand gestures. In this process a bound box is formed that will enclose the contours of the hand, the upper left corner of the box is marked with pt1 and bottom right corner marked by pt2 , and will be used as a parameter of the cvRectangle function, to display the box enclosing the contour of the hand, as shown in Figure 12.

  [9] M. M. Youssef, "Hull Convexity Defect Features For Human Action Recognition," The School of Engineering of The University of Dayton, 2011. [10] G. Bradski and A. Kaehler, Learning OpenCV, 1005 Gravenstein Higway North, Sebastopol, CA 95472: O'Reily Media Inc, 2008.

  

Author Index Abdullah, Harnan Malik; E2_111-115 Abdullah, M. F. L.; C1_56-61, E5_126-131 Abimanyu, Adi; E6_132-136 Adriansyah, Andi; E4_121-125 Al-Yoonus, Murthad; E5_126-131 Al-Shargie, Fares; E5_126-131 Anwar, Toni; E2_111-115 Ariff, Roslina Mat; A4_23-28 Aripin, Norhafizah bt; E8_142-146 Ayob, Mohammad Afif; D3_87-92 Bakar, Suheri; B4_51-55 Bohari, Azuwien Aida; A4_23-28 Cahyadi, Eko Fajar; E1_105-110 Crysdian, Cahyo; E3_116-120 Dani, Akhmad Wahyu; E4_121-125 Das, Bhagwan; C1_56-61 Fat, Joni; B3_45-50 Fathurrodli, Tri Prasetya; A5_29-34 Firmanda, Dion; C4_70-77 Halim, Abdul; B2_40-44 Halomoan, Junartho; D1_78-81 Hariyanto, Nanang; A1_01-07 Haron, Zainal Alam; A4_23-28 Harsono, Tri; C4_70-77 Hartanto, Rudy; E7_137-141 Haryanti, Munnik; B2_40-44 Ishitobi, Mitsuaki; D1_78-81 Jalani, Jamaludin; D3_87-92 Jawad, Mohammed Saeed; E5_126-131 Khalid, N. K.; C2_62-66 Kusumasari, Vita; C3_67-69 Litasari; E6_132-136 Mado, Ismit; A2_08-15 Mandita, Fridy Mandita; E2_111-115 Marojahan, B.N., Kevin; A1_01-07 Maulana, Eka; D4_93-98 Muslim, M. Aziz; D2_82-86, D4_93-98 Mustika, I Wayan; E6_132-136 Nurdin, Muhammad; A1_01-07 Othman, M. B.; E8_142-146 Penangsang, Ontoseno; A5_29-34 Prasetyanto, Yahya Y.; B1_35-39 Purnomo, Muhammad Fauzan Edy; C3_67-69 Putri, Tri Oktaviana; D2_82-86 Rahmadwati; D2_82-86 Rudati, P.S.; B4_51-55 Santoso, Daniel; B1_35-39 Santosa, P. Insap; E7_137-141 Seman, F. C.; C2_62-66 Setyaningsih, Endah; B3_45-50 Shahida, Mohd Shah Nor; C1_56-61 Sim, Sy Yi; A4_23-28 Siswojo, Bambang; D2_82-86 Soeprijanto, Adi; A2_08-15, A5_29-34 Sudarsa, Yana; B4_51-55 Suhartono; A2_08-15 Sumantyo, Joshapat Tetuko Sri; C3_67-69 Susanto, Adhi; E7_137-141 Susanto, Erwin; D1_78-81 Susilo, Deddy; B1_35-39 Trisno Y. P.; B4_51-55 Urfin, Seif Nura; D2_82-86

  Utomo, Wahyu Mulyo; A4_23-28 Utomo, Teguh; A3_16-22 Wardhani, Lydwina; B3_45-50 Wasista, Sigit; C4_70-77 Wibawa, Unggul; A3_16-22 Wibowo, Rony Seto; A5_29-34 Yuliansari, Dwi N; E6_132-136 Yusuf, Arbai; B2_40-44 Zainuri, Akhmad; D4_93-98 Zakaria, Wan Nurshazwani Wan; D3_87-92 Zin, Nooradzianie Muhammad; A4_23-28 Zureidar, Ida; B3_45-50