proceedings of engineering and technology innovation , vol. 2, 2016, pp. 04 05 4 a low-cost design of earthquake detector with rescue message deliver using mobile device chuen-ching wang * , gi-hong wei department of information technology, kao yuan university, kaohsiung, taiwan. received 30 january 2016; received in revised form 15 february 2016; accept ed 18 march 2016 abstract the earthquake detector is a complex and expensive device. it is necessary to design a simpler and cheaper earthquake detector for common use. our research employs the ability of android-based mobile device to design a minima l cost detector for sensing the earthquake. therefore, as the earthquake occurring, the mobile device can detect the earthquake and issue a warning alarm to the phone owner. at the same time, a rescue messages including disaster location, earthquake level, and related information to the disaster center. experimental results show that the proposed method can achieve the usage of disaster rescue. keywor ds : earthquake detection, earthquake early warn ing system, ce ll broadcast service 1. introduction because earthquake occurrence is difficult to predict, the earthquake hazard plays a major role in the natural disaster. for seismic hazard mitigation, a practical earthquake forecast method appears to be far from realization, because of the extreme complexity involved in the earthquake processes. there are two types of seismic waves:[1-2]body waves that propagate through the earth’s interior and surface waves that propagate along the earth’s free surface or along other discontinuities in the earth’s interior. body waves traveling faster than the surface waves are classified to two types: compressional (longitudinal) and shear (transverse). compressional waves travel about 1.7 times faster than transverse waves and are often called p waves or primary waves. transverse waves are frequently called s waves or secondary waves. p waves are always the first among seismic waves that reach the recording station. rock particles affected by a propagating p wave oscillate backward and fo rward in the same direction as the wave propagates. in the case of s waves, particles vibrate in planes perpendicular to the direction of travel [3]. in this investigation, an earthquake detector is designed using the accelerator which e xists in the mob ile device to immediate ly ca lculate a c celeration, and therefore the earthquake amp litude can be detected. 2. earthquake detection design due to the rapid development of mobile communicat ions, mob ile devices currently can provide anyone, anywhere, and any when to access the internet through the communication network. for the reason of mobility, we imple ment the proposed method in the mobile device to reduce the cost of earthquake detector. we e xpress the main process shown as follows. first, when earthquake is occurring, the earthquake signal will be detected by g-sensor, which can sense the acceleration and orientation. ne xt , the acceleration is calcu lated by the co mputation module and transferred as earthquake magnitude. the earthquake magnitude can be used for determin ing whether an alarm/ message is issued, or not. that is, if the earthquake magnitude is greater than the specific threshold, then the location information (lat itude and longitude) is obtained by gps wh ich built in mobile device and thus the location information can be passed to the alarm module to issue alarm and rescue messages to notify the specific person or disaster center. in order to obtain the s ma llest possible errors, the phone accelerometer is compared with a standard accelerometer located at the national center for research on ea rthquake engineering (ncree) [4]. fig. 1 shows our tes t phone (marked with red bo x) is set on the accelerometer calibration system. the process is shown as follows : * corresponding aut hor. email: t 90261@cc.kyu.edu.tw proceedings of engineering and technology innovation , vol. 2, 2016, pp. 04 05 5 copyright © taeti (1) set the phone on the fixed standard accelerometer (2) adjust the device parameters (3) repeat step 1 and 2 until the phone output a reasonable range. fig. 1 comparison of the accelerator in mobile device and standard accelerometer at ncree 3. conclusions we have integrated the gravity accelerome ter, gps module and co mmunication modules and other hardware system to imp le ment an earthquake detection app in the mobile device. de monstrations show that the app can detect the earthquake intensity, latitude and longitude informat ion, time of occurrence, phone number and other messages and forward these informat ion to the specified relat ives or disaster center using mobile commun ication. the proposed scheme with advantages of easy installing and no extra hardware cost. acknowledgment the research is supported by the nsc 104-2221-e-244-018-, national science council, taiwan. reference [1] “central weather bureau,” http://www.cwb.gov.tw/ [2] s. l. kra mer, geotechnical ea rthquake engineering, nj: prentice-hall, engelwood cliffs, 2003. [3] a. magro-ca mpe ro, r. l. fle ischer, and r. s. likes, “ changes in subsurface radon concentration associated with earthquakes,” j. geophys. res., vol. 85, no. b6, pp. 3053-3057, 1980. [4] “national center for research on eart hquake engineering,” http://www.ncree.org/  proceedings of engineering and technology innovation , vol. 2, 2016, pp. 26 27 26 oppa-ac: optimal path planning based on ant colony algorithm for temporary isolated node in wsn aripriharta 1, 2 , hendrick 1, 3,* , thi thuy lieu nguyen 1 , gwo-jia jong 1 1 department electrical engineering, national kaohsiung university of applied sciences, kaohsiung, taiwan. 2 state university of malang, indonesia. 3 state polytechnic of padang, indonesia. received 01 april 2016; received in revised form 28 april 2016; accept ed 10 may 2016 abstract in real wireless sensor network application (wsn), several nodes may suffer fro m link fa ilure problem. link failure is a proble m that may exist in the presence of obstacles which blocked the wireless connection between nodes. while the link between nodes is blocked, the node will be temporary isolated from the cluster; thus, their data will not reach the destination node. the temporary node isolation problem becomes more challenging if the data should be arrived on time. the traditional clustering algorithm (leach) is not considered the temporary isolated node, which cause longer waiting time for data delivery in certain rounds. in order to solve this problem, we proposed the ant colony-based optimal path planning algorithm (oppa -ac). the opp-a c was improved the leach algorith m by providing alternative path for temporary isolated nodes and guarantee their data arrived in the destination. based on the experimental result, the opp-a c surpassed the traditional algorithm in term of waiting time. ke ywor ds : tempora ry isolated node, ant colony, opp-ac, wsn 1. introduction in wireless sensor network application (wsn), the nodes may suffer from link failure problem due to the environmental effects, such as: radiation, temperature rising, fog, rain and obstacle. link failure is a problem that may exist in the presence of obstacles which blocked the wireless connection between nodes. therefore, those nodes cannot send data through the original forwarding path for a period of time (or temporal) because their co mmunication range is decreased [2-5] or blocked. sink ch1 temporarily isolated node normal node ch2 ch3 fig. 1 temporary isolated node since lea ch was not considered this problem, those nodes are temporarily isolated fro m their ch. the te mporary node isolation problem beco mes more challenging if the data should be arrived on time. moreover, the temporarily isolated nodes (fig. 1) cause packet drops, increase the average delay, and increase the energy consumption; in turn, the perfo rmance of network become degrades [3]. recently, there has been a few re lated works in te mporarily node isolation p roble m [2-5]. however, those existing works focused on connection reestablishment. there fore, we pre sent the ant colony-based optima l path planning algorith m (oppa -ac) to select the forwarding path for tempora rily isolated nodes . our objec tive is to reduce the waiting time and congestion in wsn. 2. method the structural synthesis of ccpgts will be performed based on the creative design met hodology process [7-8]. the p roposed algorith m *corresponding aut hor. email: hendrickpnp77@gmail.co m proceedings of engineering and technology innovation , vol. 2, 2016, pp. 26 27 27 copyright © taeti consists of two steps: (a) clustering process, (2) data transmission. in the first step, we adopt lea ch fo r the ch selection [1]. the second step involves the proposed path selection mechanis m in order to reduce the end to end delay and packet drops in wsn. in order to determine the forwa rding path for temporarily isolated nodes, we used link quality as the pheromone of ants by favoring received signal strength indication (rssi). the rssi value is collected each time the tempora rily isolated node receive adv message fro m nearby chs. rssi va lues is normalized for fu rther calculation by opp-ac, and it given by: 𝑅𝑆𝑆𝐼𝑛𝑜𝑟𝑚 = 𝑅𝑆𝑆𝐼𝑖 𝑅𝑆𝑆𝐼𝑚𝑎𝑥 (1) the objective function is the lin k quality (lq) wh ich depends on rssi va lue. therefore , the isolated node will selects the path with better lq. moreover, ant colony a lgorith m select a chs node with probability pij k = 𝜏𝑖𝑗 𝛼 𝜂𝑖𝑗 𝛽/ ∑ 𝜏𝑖𝑠 𝛼𝜂𝑖𝑠 𝛽 𝑠𝜖𝑙𝑖𝑠𝑡𝑘 for j∈ 𝑙𝑖𝑠𝑡𝑘 , otherwise pij k = 0. moreover, 𝜂𝑖𝑗 = 1/𝑑𝑖𝑗 and the pheromone (𝜏𝑖𝑗 = 1 𝑅𝑆𝑆𝐼𝑛𝑜𝑟𝑚 ), =0.7 and 𝛽 = 0.3. 3. results and discussion we validated the proposed algorith m through simu lations and comparison with the lea ch in term o f end to end de lay. we simu lated 35 nodes in 50x50m 2 square area, the data length ia 30 byte, elec = 50nj/bit, eda=0.5nj/bit, amp=50nj/bit/m 4 , and fs= 10pj/bit/m 2 . based on the e xperimental result (fig. 2), the opp-ac surpassed the lea ch a lgorith m in term of end to end delay. 4. conclusions in this paper, three new designs of the ccpgt have been generated in a systematic methodology. the feasibility of the new designs is verified by conducting kine mat ic simu lation. the result has shown that the new designs can produce a more wide range of non -uniform output motion than the existing design. therefore, they are better alternatives for driving a variable speed input mechanism. e n d t o e n d d e la y nodes leach opp-ac fig. 2 end to end delay acknowledgement this work was supported by dikti funded by ministry of research and tech nology and higher education, indonesia under contract number: 124.67/e4.4/2014. references [1] m. mehdi afsar and mohammad-h. tayarani-n, “clustering in sensor networks: a literature survey,” journal of network and co mputer applications, vol. 46, pp. 198-226, 2014. [2] a. roy, p. ka r, and s. m isra, “ detection of dumb nodes in a stationary wire less sensor network,” annual ieee india conference (indicon ‘14), 2014, pp. 1-6. [3] p. ka r, a. roy, s. m isra, and m. s. obaidat, “energyeffic ient connectivity reestablish ment in wsn in the presence of du mb nodes,” ieee international conference on co mmunicat ion workshop (iccw ‘15), 2015, pp. 1485-1490. [4] a. roy, a. mondal, and s. misra, “ connectivity re-establishment in the presence of dumb nodes in sensor-cloud infrastructure: a game theoretic approach,” ieee 6th international conference on cloud co mputing technology and science (cloudco m ‘14), 2014, pp. 847-852. [5] s. misra, p. kar, a. roy, and m. s. obaidat, “existence of dumb nodes in stationary wireless sensor networks,” journal of system and software , vol. 91, pp. 135-146, 2014.  proceedings of engineering and technology innovation , vol. 3, 2016, pp. 13 15 13 behavioural assessment of blood-brain barrier opening induced by various ultrasound parameters feng-yi yang * , sheng-fang huang department of biomedical imaging and radiological sciences , national yang-ming university, taipei, taiwan. received 23 march 2016; received in revised form 15 april 2016; accept ed 02 may 2016 abstract focused ultrasound (fus) -induced opening of the blood-brain barrier (bbb) in the presence of microbubbles is prospective for a targeted drug delivery to the bra in lesion. ho wever, the safety of this technology has been addressed recently. the purpose of this study was to assess the behavioural changes follo wing the fus induced-bbb disruption (bbbd). the behavioural assessment of rats were tested using the open field and hole-board fro m day 1 through day 32 a fter undergoing bbbd induced by fus with either a mild or heavy parameter. we found that heavy bbbd induced behavioral changes, including significantly increased locomotor activity and a longer latency to nose poke for baits . additionally, mild bbbd resulted in significant decrease of central activity. therefore, the behavioural changes after fus induced bbbd should be considered before clinica l application. ke ywor ds : behavioural evaluation, focused ultrasound, blood-brain barrier, memory 1. introduction focused ultrasound (fus) with mic robu bbles has offered the potential to produce bbb disruption (bbbd) noninvasively in specific regions of the brain [1, 2]. however, the greatest limitat ion on the use of fus-induced bbbd in clin ical practice consists of safety concerns relating to cavitation in the bra in. mechanica l effects may be responsible for fus-induced bbbd, but inert ial cavitation could usually cause hemorrhaging or apoptosis in the brain tissue from the neighboring vessel [3]. a lthough no significant negative effects resulting fro m histological e xa mination, further evaluation of brain functions follo wing fus-induced bbbd were still needed. such research will increase our knowledge of the behaviora l a lterations that occur following bbbd and allow for a better assessment of the safety of this technique in terms of its effects on bra in functions. therefore , the purpose of this study was to exa mine the impact of bbbd in terms of behavioral altera tions following fus e xposure in the presence of microbubbles. 2. method all procedures involving anima ls were conducted in accordance with the guidelines for the care and use of laboratory animals. th e study protocol was approved by the anima l ca re and use committee of national yang ming university. male sprague-dawley (sd) rats weighing fro m 300 to 350 g we re used in this study. 2.1. ultrasound setup and behavioural assessment the ultrasound system and equipment setup were the same as used in our previous study [4]. the bbb disruption (bbbd) can be quantified based on the extravasation of evans blue (eb), which binds to albumin. activ ity in the open fie ld was tested with the automated fle x-fie ld/open field photobeam activity sy stem on post-sonication days 1, 9, 18, and 32. spatial lea rning ability was studied by means of a hole-board apparatus in which food rewa rds were used as positive motivation. 2.2. number synthesis three rats fro m each group we re prepared for histological observation. the rats were perfused with saline and 10% neutral buffered * corresponding aut hor. email: fyyang@ym.edu.t w proceedings of engineering and technology innovation , vol. 3, 2016, pp. 13 15 14 copyright © taeti forma lin on days 1 and 9 a fter the fus sonication. the brain slices were stained by tunel staining (deadend co lorimetric t unel system, g7130, pro mega, mad ison, wi, usa) in order to detect dna frag mentation and apoptotic bodies within the cells . all values are shown as means ± sem. the behavioral assessment data were analy zed using the mann-whitney u test. other data were analyzed using the unpaired student’s t test. statistical significance was defined as a p value  0.05. 3. results and discussion fig. 1 effects of rats with fus -induced bbbd on the open field activ ity. (a) the nu mber of center entries was significantly increased in the bbbd-heavy group on days 1 and 9 follo wing sonication. (b) but the number of center entries was significantly decreased in the bbbd-mild group on day 18 following sonication. * denotes significant differences compared to the sham group at the same time points. (*, p<0.05; ***, p<0.005, n = 8) bbbd-heavy e xposed rats entered the center of the open field significantly more frequently on days 1 and 9 post-sonication compared with sham group (fig. 1(a)). bbbd-mild e xposure significantly reduced the number of center entries on day 18 post-sonication compared with sham group (fig. 1(b)). the bbbd-heavy group e xhibited a significantly longer latency and latency to first baited hole compared with the sham group (fig. 2(a)). no s ignificant diffe rences in latency and latency to first baited hole were found between the bbbd-mild group and sham group (fig. 2(b)). fig. 2 ho le-board testing was performed on day 12 fo llo wing sonication. the latency and latency to the first baited hole were recorded in (a) the bbbd-heavy group and (b) the bbbd-mild group. * denotes a significant diffe rence relative to sham group. (*, p<0.05; **, p<0.01, n = 8) to our knowledge, the current study is the first to exp lore the e ffect of bbbd induced by fus on behavioral alterations in an anima l model. the findings fro m this study indicate that bbbd-heavy rats produced hyperactivity rela tive to sham animals under open field test. proceedings of engineering and technology innovation , vol. 3, 2016, pp. 13 15 15 copyright © taeti moreover, the bbbd-heavy rats revea led spatial me mo ry impa irment in the hole-board test compared to sham group. in contrast, bbbd-mild rats only e xhibited an xiety-re lated behaviors in the open field test. 4. conclusions in this paper, fus-induced bbbd repre sents a major advance in the targeted drug delivery of the bra in. the current study suggests the possibility that manipulation of fus-induced bbbd might produce a variety of behavioral changes after enhanced drug delivery. further investigations of behavioral alterations in anima ls fo llowing fus -induced bbbd are needed in order to avoid abnorma l bra in function ing in humans following future clinica l uses of fus-induced bbbd. acknowledgement this study was supported by grants from the ministry of science and technology of taiwan (no. most 104-2314-b-010-003-m y3). references [1] f. y. yang, t . t. wong, m. c. teng, r. s. liu, m. lu, and h. f. liang, et a l., “focused ultrasound and interleukin-4 receptor targeted liposomal do xorubic in for enhanced targeted drug delivery and antitu mor effect in glioblastoma multiforme ,” journal of controlled release, vol. 160, no. 3, pp. 652-658, june 2012. [2] f. y. yang, w. y. chang, j. j. li, h. e. wang, j. c. chen, and c. w. chang, “pharmacokinetic analysis and uptake of 18f-fbpa-fr sfter ultrasound-induced blood-brain barrie r disruption for potential enhancement of boron delivery for neutron capture therapy,” journal of nuclear medicine, vol. 55, pp. 616-621, april 2014. [3] f. yang, w. fu, w. chen, w. yeh, and w. lin, “quantitative evaluation of the use of microbubbles with transcranial focused u ltrasound on blood¡vbra in-barrie r disruption,” ultrasonics sonochemistry, vol. 15, no. 4, pp. 636-643, april 2008. [4] f. y. yang, w. y. chang, j. c. chen, l. c. lee, and y. s. hung, “quantitative assessment of ce rebral glucose metabolic rates after blood-brain barrie r disruption induced by focused ultrasound using fdg-m icropet ,” neurolmage, vol. 90, pp. 93-98, april 2014.  proceedings of engineering and technology innovation, vol. 5, 2017, pp. 37 40 applications of thermal images for monitoring surficial temperature changes of naked slope chien-yuan chen * , zhe-hao liu department of civil and water resources engineering, national chiayi university, chiayi, taiwan, roc. received 02 july 2017; received in revised form 06 august 2017; accept ed 10 august 2017 abstract global climate change causes increases in the torrential rainfall brought by typhoons and the monsoon in taiwan. torrential rain in turn causes landslides, debris flows, and the formation of e arth dams. most dams were formed in remote mountainous areas and are difficult to reach for safety evaluation at the beginning of their formation. a long distance and non-destructive testing methodology is necessary for evaluating the safety of landslide d ams. this study used an infrared imager for monitoring naked slopes. the thermography can detect surficial radiation temperature changes in the slope to locate potential unstable areas for further monitoring. this study proposes radiation temperature change (t) per unit of time (δt) as an index (t /δt) for nondestructive monitoring. the index was used for monitoring and analysis of artificial earth dams constructed a t huishun farm in nantou county. the results of the analysis show that the failure zone of the artificial dam exhibited the greatest change in the index and the potential failure mode could be predicted once the dam breached. the proposed model could be used for potential unstable slope monitoring. keywords: thermography, landslide, monitoring 1. introduction typhoon morakot hit taiwan in 2009 and brought torrential rainfall in southern taiwan. the typhoon caused 17 landslide dams [1]. the failure modes of landslide dams include piping, overtopping, and complex modes [2]. stream flow blocks by landslides could cause ponding water upstream and flood disasters downstream once they are breached . natural dam stability evaluation is urgent for dam breach disaster prevention and mitigation. thermal images have been used to examine a wide variety of problems [3-4]. it has been commonly used for monitoring activities of landslides in recent years [5-7]. it has the advantages of non-contact and long distance as camera-like characteristics, and is suitable for both long -term monitoring and short-term inspection for sites with surficial radiation temperature changes . 2. field model test the artificial earth dam field test was conducted at huishun farm in nantou county in central taiwan. the test procedure included construction of an artificial earth dam in the downstream region of landou creek. the dam size was 25 m in length and * corresponding author. e-m ail address: chieny uc@ m ail.ncy u.edu.tw tel.: +886-5-2717686; fax: +886-5-2717693 proceedings of engineering and technology innovation, vol. 5, 2017, pp. 37 40 copyright © taeti 38 2 m in height and it blocked the entire creek [8]. water was stopped upstream first and released during tests. the test was made on 10 april 2014, and modeled the failure processes and modes of natural earth dams. the thermal imager was set up on a bridge for monitoring of the upstream face of the dam and at the side bank for the downstream face (fig. 1). thermal images were tak en every few minutes with the increasing water table during the test until the dam breached . (a) field photography corresponding thermography fig. 1 the artificial earth dam and corresponding thermal image (taken 10 april 2014) 3. methodology the study proposed a new index for monitoring the surficial cha nges of soil slope by water infiltration using thermal images. the index (t /δt) is a measurement of the surficial radiation temperature changes (t) in the time period between the times th e two images were taken (δt). the changes of surficial radiation temperature over time are used for monitoring during the field test for potential failure mode evaluation. the surface of the earth dam downstream was divided into grids for monitoring the temperature changes. there are 27 grids monitored as marked in the image from left to right: a1~a3, b1~b8, c1~c8, and d1~d8 (fig. 1). the temperature is taken as an average within the grid. fifteen successive images were taken for the analysis. a positive value of the index shows the tempe rate is increasing. a negative index shows reduction in the temperate due to seepage and piping with the increasingly blocked water table. 4. results and analysis fig. 2 monitoring of surficial radiation temperature changes on 14:15:56 fig. 3 monitoring of surficial radiation temperature changes on 14:28:39 monitoring began at 14:15:56 and ended at 14:34:26. the surficial temperature variations with time at the monitoring grids for the first two images are shown in fig. 2. all the grids showed a positive index with increasing surficial temperature during the proceedings of engineering and technology innovation, vol. 5, 2017, pp. 37 40 copyright © taeti 39 sunshine, at the beginning of the test. soil erosion and piping occurred at gird no. 26 (d7 in fig. 1) 12 minutes later with a negative value for the index (fig. 3). with increasing time, another piping occurred in grid no. 8 (b5) (fig. 4). it is speculated that the type of failure mode is seepage caused piping if the blocked water table is not rising at the current stage. the dam breached in a complex mode with piping first followed by overtopping failure (fig. 5). the thermal images showed a clear lower index in the grids with potential seepage-induced piping failure, though the dam failed by overtopping with increasing river flow. piping failure occurs in low permeability dam material with high soil shear strength under high discharge flow. it should be note that the piping water in the upper grids could flow down to the lower grids without piping, causing a lower index in all these grids. the res ults show that the index (t /δt) is suitable for monitoring surficial radiation temperat ure changes of naked slopes in a given time period and may be applied for monitoring potential landslide slopes . (a) field photography (b) corresponding thermography (c) the index (t /δt) changes fig. 4 monitoring of surficial radiation temperatu re changes on 14:33:55 (a) field photography (b) corresponding thermography fig. 5 monitoring of surficial radiation temperature changes on 14:34:26 (continued) proceedings of engineering and technology innovation, vol. 5, 2017, pp. 37 40 copyright © taeti 40 (c) the index (t /δt) changes fig. 5 monitoring of surficial radiation temperature changes on 14:34:26 5. conclusions a long distance and non-destructive testing methodology using an infrared imager is adopted for monitoring the safety of artificial dams. the thermography can detect surficial radiation temperature changes in the slope t o locate potential unstable areas for further monitoring. the radiation temperature change per unit of time is proposed as an index (t /δt) for nondestructive monitoring. the index was used for monitoring and analysis of artificial earth dams constructed at huishun farm in nantou county. the results of the analysis show that the piping failure zone of the artificial dam exhibited the greatest change in the grid and the potential failure mode could be predicted. the index reflects the surficial radiation temperature changes and can be used for monitoring the stability of naked slopes . acknowledgement the authors would like to thank the ministry of science and technology, taiwan roc, for financially supporting this research under contract no. most 105-2625-m-415-001. references [1] c. y. chen and j. m. chang, “landslide dam formation susceptibility analys is based on geomorphic features ,” landslide, vol. 13, no. 5, pp. 1019-1033, october 2016. [2] o. korup, “recent research on landslide dams a literature review with s pecial attention to new zealand,” progress in physical geography, vol. 26, no. 2, pp. 206-235, june 2002. [3] i. baroň, d. bečkovský, and l. míčai, “application of infrared thermography for mapping open fractures in deep -seated rockslides and unstable cliffs,” landslides, vol. 11, no. 1, pp. 15-27, february 2014. [4] t. nolesini, w. frodella, s. bianchini, and n. casagli, “detecting slope and urban potential unstable areas by means of multi-platform remote sensing techniques: the volterra (italy) case study,” remote sens ing, vol. 8, no. 9, p. 746, september 2016. [5] j. h. wu, h. m. lin, d. h. lee, and s. c. fang, “integrity assessment of rock mass behind the shot creted slope using thermography,” engineering geology, vol. 80, no. 1-2, pp. 164-173, august 2005. [6] s. martino and p. mazzanti, “integrating geomechanical surveys and remote sensing for sea cliff slope stability analys is: the mt. pucci case study (italy),” natural hazards and earth system sciences, vol. 14, pp. 831-848, april 2014. [7] s. mineo, g. pappalardo, f. rapisarda, a. cubito, and g. di maria, “integrated geostructural, seismic and infrared thermography surveys for the study of an unstable rock slope in the peloritani chain (ne sicily),” engineering geology, vol. 195, pp. 225-235, september 2015. [8] c. y. chen, s. c. chen, and k. h. chen, “earth dam monitoring by using infrared thermography detection,” proceedings of engineering technology and innovation, vol. 4, pp. 40-42, 2016.  proceedings of engineering and technology innovation , vol. 3, 2016, pp. 22 24 22 recognition of concurrent control chart patterns in autocorrelated processes using support vector machine chau-chen torng * , chon-how hwang department of industrial engineering and management, national yunlin university, yunlin, taiwan. received 16 february 2016; received in revised form 18 march 2016; accept ed 09 april 2016 abstract control chart pattern recognition (ccpr) is an important issue in statistical process control because unnatural control chart patterns (ccps) e xhibited on control charts can be associated with specific causes that adversely affect the manufacturing processes. in recent years, many mach ine learning techniques have been su ccessfully applied to ccpr. however, such existing research for ccpr has mostly been d eveloped for identification of basic ccps (sh ift patterns, trend patterns, cyclic pattern and systematic pattern). little attention has been given to the identification of concurrent ccps (two or more basic ccps occurring simultan eously) which are co mmonly encountered in practical manufacturing processes. in addition, these e xisting researches also assume the process data are independently and identically distributed which may not be appropriate for certain manufacturing processes. this study proposes a support vector machine (svm) approach to identify concurrent ccps for a mu ltivariate process with autocorrelated observations which can be characterized by a first order autoregressive (ar(1)) mode l. the nume rical results indicate that the proposed model can effect ively identify two concurrent identical ccps but for those cases involving one trend pattern and one shift pattern, their recognition accuracy deteriorates to around 20% to 50% depending on the autocorrelation coeffic ients used in the data model. ke ywor ds : control chart pattern, support vector machine, autocorrelation process 1. introduction the effectiveness of the use of control charts depends on the ability to recognize patterns. the common types of ccp e xhib ited on statistical process control charts have been formalized in early literature. in most literature, the fo llo wing seven typical types of basic ccp, i.e ., up ward and downwa rd sh ift pattern (usp and dsp); upward and down ward t rend pattern (utpand dtp); cyclic pattern (cp); systemat ic pattern (sp); natural pattern (np), are usually addressed. identificat ion of unnatural patterns canfacilitate early detection of an out -of-control process and the diagnostic search process by narrowing down the set of possiblecauses that must be investigated. for instance, shift patterns may indicate changes in materia l, mach ine or operator, while trend patternsmay indicate tool wear. cyclic patterns may ind icate voltage fluctuation in power supply [1-3]. there is a crucia l need for an auto matic and effective analysis and interpretation methodology for control chart pattern recognition (ccpr), which enables indication of the real state of the manufacturing process. ccpr studies consist of the description, the identification, the e xp lic it classification, and the e xtraction of patterns in data. in recent years, many mach ine learning techniques have been successfully applied to ccpr. ho wever, such e xisting research for ccpr has mostly been developed for identification of basic ccps (sh ift patterns, trend patterns, cyclic pattern and systematic pattern). little attention has been given to the identification of concurrent ccps (two or more basic ccps occurring simu ltaneously) wh ich are commonly encountered in practical manufacturing processes. in addition, these e xisting researches also assume the process data are independently and identically distributed which may not be appropriate for certain manufactu ring processes. for those processes involving continuous manufacturing operations (including the manufacture of food, chemicals, and paper * corresponding aut hor. email: t orngcc@yuntech.edu.tw proceedings of engineering and technology innovation , vol. 3, 2016, pp. 22 24 23 copyright © taeti and wood products), the consecutive observ ations of product characteristics are often highly correlated. the purpose of this paper is to propose a support vector machine (svm ) approach to identify concurrent ccps for a mu ltivariate process with autocorrelated observations. 2. method this study uses simu lations to generate the required sets of ccp e xa mp les for tra ining, testing, and performance evaluation of the proposed model. the mathemat ical e xp ressions for ccp generation are e xp ressed in a general form that consists of process mean, common cause variation and special d isturbance from assign able causes. the equations and parameters used to generate the data points are shown in table 1. table 1 ccp equations and parameters pat t erns equat ions paramet ers nat ural pat t ern 𝑥𝑖 = 𝜇 + 𝑟𝑖𝜎 =0, =1 t rend pat t ern 𝑥𝑖 = 𝜇 + 𝑟𝑖𝜎 + 𝑡 × 𝑖  26.0,10.0t shift pat t ern 𝑥𝑖 = 𝜇 + 𝑟𝑖𝜎 + 𝑢 × 𝑠   10 0.3,0.1 oru s   cyclic pat t ern 𝑥𝑖 = 𝜇 + 𝑟𝑖𝜎 + 𝑎 × sin 2𝜋𝑡 𝛺  0.3,0.1a syst emat ic pat t ern 𝑥𝑖 = 𝜇 + 𝑟𝑖𝜎 + 𝑑(−1) 𝑖  0.3,0.1d the follo wing equations are used to simulate the autocorrelated data. 𝑋1𝑡 = 𝜇 1 + 𝜙11 𝑋1 ,𝑡−1 +𝜙12 𝑋2 ,𝑡−1 + 𝜀1𝑡 (1) 𝑋2𝑡 = 𝜇 2 + 𝜙21 𝑋1 ,𝑡−1 +𝜙22 𝑋2 ,𝑡−1 + 𝜀2𝑡 (2) where 𝜙𝑖𝑗 are autocorrelated coefficients , and 𝜙11 = {1.0, 0.7 0.4, 0.1}, 𝜙22 = {1.0, 0.7, 0.4, 0.1}. for tra ining and testing the concurrent patterns, a svm based on radial basis kernel fun ction is chosen inthis study. related para meters c and γ for this ke rnel were varied in the fixed ranges [2-5, 25]. using the simu lated training and testing exa mp les, the optima l (c, γ) was determined as [1, 0.03]. the observation window sizes of 24 and 32 are selected according to the suggestions by guh and tannock [4] and hachicha and ghorbel [1]. 3. results and discussion the concurrent ccps we re a co mbination of any t wo of the basic unnatural ccps. thus, 13 types of concurrent ccp, na me ly usp mixed with utp (usp+utp), usp mixed with dtp (usp+dt p), usp mixed with cp (usp+cp), usp mixed with sp (usp+sp), dsp mixed with ut p (dsp+utp), dsp mixed with dtp (dsp+dt p), dsp mixed with cp (dsp+cp), dsp mixed with sp (dsp+sp), utp mixed with cp (utp+cp), ut p mixed with sp (ut p+sp), dtp mixed with cp (dt p+cp), dt p mixed with sp (dtp+sp), cp mixed with sp (cp+sp) are addressed in this study. table 2 accuracy of ccpr simu lation results in table 2 show that the rate of accuracy for patterns of utp, dtp, utp+usp, utp+dsp, utp+sp, dtp+usp, dtp+dsp, dt p+sp is worse than the other patterns. it is noticed that if the trend pattern was mixed with other patterns, their recognition performance would decrease. 4. conclusions in this paper, a support vector mach ine approach is proposed to identify concurrent ccps for a mu ltivariate process with autocorrelated observations which can be characterized by proceedings of engineering and technology innovation , vol. 3, 2016, pp. 22 24 24 copyright © taeti afirst order autoregressive (ar (1)) model. the numerical results indicate that the proposed model can effectively identify two concurrent identical ccps but for those cases involving one trend pattern and one shift pattern, their reco gnition accuracy deterio rates to around 20% to 50% depending on the autocorrelation coefficients used in the data model. acknowledgement the support of the minister of education (taiwan), under grant most 104-2221-e224 026 is gratefully acknowledged. references [1] w. hachicha and a. ghorbel, “a survey of control-chart pattern-recognition literature (1991–2010) based on a new conceptual classification scheme,” co mputers & industrial engineering, vol. 63, no. 1, pp.204-222, 2012. [2] l.j. xie, n. gu, d.l. li, and z.q. cao, “concurrent control chart patterns recognition with singular spectrum analysis and support vector machine,” co mputers & industrial engineering, vol. 64, no. 1, pp. 280-289, 2013. [3] w. a. yang, w. zhou, w. liao, and y. guo, “identification and quantification of co ncurrent control chart patterns using extreme -point symmetric mode deco mpos ition and e xtre me learn ing mach ines ,” neurocomputing, vol. 147, pp. 260-270, 2015. [4] r. s. guhand j. tannock, “recognition of control chart concurrent patterns using a neural network approach,”international journal of production research, vol. 37, no.8, pp.1743-1765, 1999.  proceedings of engineering and technology innovation , vol. 2, 2016, pp. 20 22 20 identifying islamic values in the embodiment of makassar traditional houses in barombong coastal mursyid mustafa * , ananto yudono, ria wikantari, afifah harisah department of architectural engineering, hasanuddin university, makassar, indonesia. received 11 march 2016; received in revised form 18 april 2016; accept ed 03 may 2016 abstract indonesia is a country with ma jority muslim population. however, in the arch itectural design, people was not significantly using the isla mic values and tend to be affected by western mo dern architecture on their house embodiment. therefore, this research e xplores a house arch itecture based on islamic values, by taking the case of traditional houses on the co ast. the coastal areas chosen because it still has trad itional houses condensed with local cultural community. using observational techniques, interviews with selected informants and questionnaires, and supported by literature rev iew, this research was analyzed by qualitative method. this paper aims to identify isla mic values in the embodiment of makassar traditional houses in the baro mbong coastal area, and enrich the concept of islamic-va lued on house architecture. the research findings are the e mbodiment of makassar traditional houses in baro mbong coastal area wh ich are generally shaped the house terrace called as bale-ba le and paladang, which are dee med to have isla mic values, like : (1) in isla m, the way to visit a house is coming fro m the front, instead of behind. evidently, this way e xisted in traditional house, (2) ro le of the bale-bale as an e xtension of the main roo m which can be used to interact with neighbors and the community (hablunminannas) and paladang can serve as a transition zone to screen guests before we lco ming the m inside the house (privacy) and guests to greet before the residents of the house is ready to receive, (3) in the tradition of building houses, already recognized the co ncept of space-oriented four wind directions (sulapa’ appaka), which significantly according to the concept of islam which require orientation to the qibla oriented home. the study concluded that with the local wisdom possessed by the makassar traditional house on barombong coastal area discovered the islamic values in its realization. ke ywor ds : ba ro mbong coastal, isla mic va lues , makassar traditional house 1. introduction indonesia is a country with ma jority muslim population. however, in the arch itectural design, people were not significantly using the isla mic values and tend to be affected by western mo dern architecture on their house embodiment. therefore this research e xplores a house arch itecture based on islamic values, by taking the case of traditional houses on the coast. the coastal areas chosen because it still has trad itional houses condensed with local cultural community which is in accordance with statement of rapoport (1977, 2004), gustafson (2001), in wan isma il [1] where the built environment is influenced by culture, belief and past experiences of its inhabitants . makassar ethnicity of south sula wesi in tradition of building their house alway s used a stage-house form, which is in macrocosmos view div ided by three parts, lower part (ra wa balla), the body of the house (kale ba lla ), and upper part. fig. 1 example of makassar traditional house *corresponding aut hor. email: mursyidmust afa58@gmail.com proceedings of engineering and technology innovation , vol. 2, 2016, pp. 20 22 21 copyright © taeti a common makassar traditional house consists of bale-bale -and lego-lego, as shown in the fig. 1. idawa rni [2] e xp la in ba le-bale as a place for socialization, play, rest and work, it’s also cheap and easy to ma ke because it does not require specific space for placing. rad ja et al. [3]also e xa mine the types and characteristics of bale bale and identify ba le bale as a co mmon space for socialization. 2. method this study uses observational method with reference to the literature. observation phase of this research was done by using interviews with people from various social strata in the study area. in addition it a lso conducted observations on aspects of social, cu ltural, and physical morphology of tradit ional houses in the study location. as for the results of the study will be analyzed using qualitative methods. 3. results and discussion after conducting interviews with selected informants and distributing questionnaires and observation of social, cultura l, and physical morphology of the tradit ional house in baro mbong coastal obtained the following results: (1) in islam, the way to visit a house is coming from the front, instead of behind. evidently, this way existed in traditional house. mustafa [4] stated that architectural interpretations of the qur'an regarding honoring or welcoming neighbors can be expla ined that the courtesy visit is required to reach the house from the front and saluted the residents of the house a ma ximu m of three times and if it does not get permission from the owner of the house, guests must go home. this way is also found in traditional malay house as islam-practicing country according to othman [5] whereas a spacious main verandah or serambi located in front of the house as a vital design element for receiving and entertaining guests . (2) ro le o f the ba le -ba le as an e xtens ion of the ma in roo m wh ich can b e used to inte ract with ne ighbo rs and the co mmun ity (hab lun minannas ) and p a ladang can serve as a trans it ion zone to sc reen guests befo re we lco ming the m inside the house (p rivacy ) and guests to greet befo re the residents of the house is re ady to rec e ive. ha kim [6] have selected qur’an ic verses and saying of the p rophet rega rding bu ild ing and isla mic c it ies p lann ing p rinc ip le. a say ings of the prophet re lated to p rivacy found in trad it iona l house is: ‘o ye who be lieve, enter not houses other than you r o wn, unt il ye have asked pe rmiss ion and saluted those in the m: th at is best fo r you, in ord er that ye may h eed.’ th is saying of prophet para lle l to the use o f ba le -ba le and pa ladang as transit ion zon e to scre en guest before we lco ming the m ins ide the house, given the o wner o f the house so me priv acy. (3) in the tradition of building houses , already recognized the concept of space-oriented four wind directions, which is assumed to be associated with building placement orientation, or direction of the shoreline (sulapa’ appaka ph ilosophy) which is significantly according to the concept of islam which require orientation to the qibla-oriented home. akil [7] in his research explain sulapa’ appaka philosophy are visible in the coastal cities whose topography is relatively flat. 4. conclusions in this paper, with the local wisdom possessed by the makassar traditional house on barombong coastal area discovered the islamic values in its realization. acknowledgement this research is grate fully ac knowledged by depart ment of arch itecture enginee rin g, faculty of enginee ring, hasanuddin unive rsity , makassar, south sulawesi, indonesia. references [1] w. h. w. is mail, “ cultural determinants in the design of bugis house,” procedia social and behavioral sciences , vol. 50, pp. 771-780, july 2012. [2] i. asmal, “ba le-bale as a space for social interaction,” procedia-social and behavioral sciences , vol. 179, pp. 176 -182, april 2015. [3] a. m. radja, t. suzuki, y. yoshizumi, m. kousaka, k. ya mauchi, s. matsubara, and t. oku, “study on bale bale in lae lae isle, makassar, indonesia,” journal architecture proceedings of engineering and technology innovation , vol. 2, 2016, pp. 20 22 22 copyright © taeti planning, aij, vol. 77, no. 675, pp. 969-978, may 2012. [4] m. mustafa, r. wikantari, a. harisah, and a. muftiradja, “kajian tekstual nilai-nila i keisla man untuk arsitektur rumah tinggal,” iplbi, universitas sam ratulangi, manado, 2015. (in indonesian) [5] z. othman, r. aird, and l. buys, “privacy, modesty, hospitality, and the design of muslim ho mes: a literature review,” frontiers of architectural research, vol. 4, issue 1, pp. 12-23, march 2015. [6] b. s. ha kim, arabic islamic cities: build ing and planning principles, routledge ed, london, 2013. [7] a. akil, a. yudono, i. latief, and a. m. rad ja , “app licat io n of bug inese makassarese culture in makassar city, indonesia,” international review for spatial planning and sustainable development, vol. 2, no. 4, pp. 68-80, 2014.  proceedings of engineering and technology innovation , vol. 2, 2016, pp. 08 10 8 study on the wear resistance of sinter alloying steel by deep cryogenic treatment chao-hsu cheng * department of mechanical and automatic engineering, chung chou university of science and technology, changhua, taiwan. received 19 february 2016; received in revised form 16 march 2016; accept ed 20 april 2016 abstract sintered alloy steel powder with four d ifferent hardenability of fn0200, fn0208, fd0408, and flnc4808 are applied to form the sintered alloy steel. the sintered alloy steel, being cooled down to room te mperature, is further soaked in -196℃ liquid nitrogen for 4hrs deep cryogenic treatment. with deep cryogenic treatment, the retained austenite in the sintered a lloy steel would change the phase and become martensite, and the apparent hardness of sintered alloy steel would be enhanced. further analy zing the effect of deep cryogenic treatment on the wear resistance of sintered alloy steel, the wear resistance would be improved when the apparent hardness of sintered alloy steel is enhanced. however, the deep cryogenic treatment would change the sintered alloy steel microstructures because of distinct hardenability. such research findings show the significant effect of deep cryogenic treatment on the sintered alloy steel with high hardenability, where the best hardness and wear resistance could be acquired from flnc4808 sintered alloy steel through the deep cryogenic treatment. keywor ds : sintered alloy steel, deep cryogenic treatment, hardenability, wear resistance 1. introduction research on sub-zero treat ment actually has been done in the early 20 th century to change the structure and function of steel by lower than free zing point. in 1937, a. p. gu lyaev [1] first researched sub-zero treat ment fo r h igh-speed steel and proposed the theory of sub-zero down to -80℃. so far, it is known that deep cryogenic treatment (-196℃) could enhance the ratio of ma rtensite and reduce austenite residue to improve the wear resistance of steel. deep cry ogenic treat ment is a lso broadly used for ma chinery industry internationally to enhance the wear resistance of ferrous products, such as die steel [2], powder h igh-speed steel [3], and cast iron [4], and the performance o f non -ferrous products, like magnalium [5]. nevertheless, deep cryogenic treatment has not been applied to the research on sintered alloy steel. this study therefore intends to discuss the effects of deep cryogenic treat ment on the me chanical properties and wear performance of sintered alloy steel. 2. method sintered alloy steel with four d ifferent mu ltiplying factors of fn0200 (fe1.2cu2.5ni 0.3c), fn0208 (fe 1.2cu2.5ni0.9c), fd0408 (fe 1.5cu4.0ni0.5mo0.9c), and flnc4808 (fe 2.0cu2.7ni1.25mo0.2mn 0.9c) is utilized in this study. the alloy contents are shown in table 1. the as -sintered steel are further proceeded deep cryogenic treatment fro m roo m te mpera ture down to cryogenic temperature -196℃ with 0.3℃/ min for 4hr and then slowly increased to room te mperature with 0.3℃/ min. finally, sintering sintered alloy steel and cryogenically treated sintered steel are tempered with 180℃ for 1hr to re move residual stress caused by phase change. furthermore, the mechanical p roperty is e xperimented. the above test pieces are measured the s urface vic kers mic ro hardness (hv 1) with 1kg load. the prepared φ5 cylinder wears test pieces and placed on the pin-on-disc wea r test platform, *corresponding aut hor. email: chaohsu@dragon.ccut.edu.t w proceedings of engineering and technology innovation , vol. 2, 2016, pp. 08 10 9 copyright © taeti fig. 1. the wear test platform rotates an skd11 disc with the speed 60 rp m, and the wear test piece is placed 47.5 mm a way fro m the disc center shaft, while a 2 kg load is loaded on top of the wear test piece. each test piece is proceeded the wear test of the wear distance 15,000m, and the measured wear we ight loss is calculated the wear coeffic ient of the test piece with the following equation. k=       m mm nf m 3  (1) where k: wear coefficient; m: wear amount, g; ρ: density, g/mm 3 ; f: positive load on wear test piece, n; t: wear distance, m. table 1 sintered alloy steel contents fig. 1 pin-on-disc wear test schematic 3. results and discussion 3.1. the hardness of sintered alloy steel the effects of deep cryogenic treatment on the hardness of sintered alloy steel with different hardenability are further analyzed. the as -sintered steel after deep c ryogenic treat ment, where the apparent hardness could be improved by deep cryogenic treatment, regard less of the hardenability. sintered a lloy steel with d istinct hardenability would enhance the apparent hardness with increasing hardenability, revealing that sintered alloy steel would be hardened with deep cryogenic treatment. especially, most of the austenite structure in high ca rbon high hardenability alloy steel (fd0408) is changed into ma rtens ite and the austenite residue is dropped down to the lowest that the apparent hardness is increased from hv1 516 to hv1 733. the improvement of wear resistance could enhance sintered alloy steel parts to a higher level. 3.2. the wear resistance of sintered alloy steel the sintered alloy steel wear test pieces with diffe rent hardenability are placed on the pin-on-disc wear test platform for the 15,000m wear test to measure the wear we ight. the sintering fn0200 a lloy steel with the lowest hardenability is worn the weight 0.0547g with 15,000m, wh ile the sintering flnc4808 a lloy steel with the highest hardenability is mere ly worn 0.0014g with 15,000m; a fter deep cryogenic treat ment, flnc4808 further drops down to 0.0009g, fig. 2. the result e xpla ins that sintered alloy steel could be imp roved the wear resistance with deep cryogenic treatment. fig. 2 wear we ight of sintered alloy steel wea r test substituting the density (ρ), wea r distance (t), test piece wear we ight (m), and wear load (f) of the sintered alloy steel test pieces for eq. (1) to analyze the wea r coefficient, the results are shown in fig. 3. fig. 3 co mparisons of wear coefficient of sintering sintered alloy steel after cryogenic treatment proceedings of engineering and technology innovation , vol. 2, 2016, pp. 08 10 10 copyright © taeti fro m the curve in fig. 3, the sintered a lloy steel with deep cryogenic treatment could effectively reduce the surface wear of sintered alloy steel parts. high carbon high hardenability sintered alloy steel (flnc4808) with deep cryogenic treatment could further enhance the wear coefficient fro m k=0.72(× 10 -6 mm 3 /nm) to k=0.46(× 10 -6 mm 3 /nm), and the metallurgical structure is mostly changed into martensite. among sintered alloy steel with distinct hardenability, the wear coefficient (k=0.46× 10 -6 mm 3 /nm) of flnc4808 sintered alloy steel with deep cryogenic treatment outperforms the wear coeffic ient (k=2.25× 10 -6 mm 3 /nm) of fn0208 sintered alloy steel with deep cryogenic treatment, and even better than fd0408 sintered alloy steel (k=0.93× 10 -6 mm 3 /nm). the result presents that a mong sintered alloy steel with diffe rent hardenability, flnc4808 sintered alloy steel with deep cryogenic treatment could acquire the best wear resistance and effectively reduce the surface wear of sintered alloy steel parts. 4. conclusions (1) as-sintered alloy steel with higher hardenability would enhance the apparent hardness as austenite changes the phase into martensit. (2) austenite of sintered alloy steel with deep cryogenic treatment would change into martensite. such a change enhances the structure apparent hardness. (3) regarding the hardness performance and wear resistance of sintered alloy steel with distinct hardenability, high hardenability alloy steel (flnc4808) with the hardenability above 40 presents the best hardness performance and wear resistance combination. (4) flnc4808 sintered alloy steel with deep cryogenic treatment presents the best wear resistance, which is suitable for machine parts requiring high wear resistance. acknowledgement this research (project no.: nsc 101-262 2-e-235-001-c c3) was cosponsored by the national science council of r.o.c. and jabon international technology co. ltd., which is gratefully acknowledged . references [1] a. p. gu lyaev, “imp roved methods of heat treating high speed steels to improve the cutting properties ,” metallurgy, vol. 12, pp. 65-70, 1937. [2] b. podgornik, et. al., “imp roving tribological p roperties of tool steels thro ugh combination of deep-cryogenic t reat ment and plasma nitriding,” wear, vol. 288, pp. 88-93, 2012. [3] v. leskovšek and b. podgornik, “ vacuu m heat treatment, deep cryogenic treat ment and simultaneous pulse plasma nitrid ing and tempering of p/m s390mc steel,” materials sc ience and engineering: a, vol. 531, pp. 119-129, 2012. [4] t. slatter, r. lewis , and a. h. jones , “the influence of cryogenic processing on wear on the impact wear resistance of low carbon steel and lame lla r graphite cast iron ,” wear, vol. 271, pp. 1481-1489, 2011. [5] k. m . asl, et. al., “ effect of deep cryogenic treatment on microstructure, creep and wear behaviors of az91 magnesium alloy ,” materials sc ience and engineering: a, vo l. 523, pp. 27-31, 2009.  proceedings of engineering and technology innovation , vol. 4, 2016, pp. 01 03 1 spatial analysis of subway ridership: rainfall and ridership kyoung-seon park 1 , sung-hi an 1 , hyunmyung kim 1 , chang-hyeon joh 2 , wonho suh 3,* 1 department of transportation engineering, myongji university, yongin, korea. 2 department of geography, kyung hee university, suwon, korea. 3 department of transportation and logistics engineering, hanyang university, ansan, korea . received 30 january 2016; received in revised form 16 february 2016; accept ed 03 march 2016 abstract in-vehicle congestion of the urban ra ilway system is the most important indicator to reflect the operation state of the urban railway. to provide the good service quality of urban railway, the crowdedness of the urban railway should be managed appropriately. the weather is one of the critical factors for the crowdedness. that is because even though the crowdedness of the urban railway is the same, passengers feel more uncomfortable in ra iny weather condition. indeed if specific sections and stations suddenly are concentrated excessive demand, it will lead far mo re serious problem. there fore, this study analysis the relationship between the nu mber of urban railway passenger and rainfa ll intensity in seoul met ropolitan subway system and then conducts the spatial analysis to deduct passenger demand patterns. this study is expected to be useful base study in order to manage the co ngestion at the urban railway station effectively by considering the different rainfall intensity. ke ywor ds : subway ridership, spatial analysis, ridership analysis 1. introduction the crowdedness of the urban railway is the most important indicator to re flect the operation state of the urban railway. to provide the good service quality of urban railway, the crowdedness of the urban ra ilway should be managed appropriately. to evaluate the crowdedness in the urban railway, quantitative factors and qualitative factors are needed. the weather is one of the critica l factors to influence the crowdedness. this is because even though the crowdedness of the urban railway is the same, passengers feel more uncomfortable in ra iny weather condition. indeed if specific sections and stations suddenly are concentrated excessive demand, it will lead far more serious problem. however, few literatures had been studied to analyze the relationship between the weather and traffic de mand. this is because collecting the traffic de mand and weather data was difficult. however, these data have been opened to the public; it is possible as get detailed information such as passenger data. park and lee analyzed the passenger’s transfer pattern during rainfall in busan [1]. this study showed that the ratio of the passenger’s mode choice is different with the different amount of rain. this research also showed that the ratio of mode choice is more influenced by the rainfall on weekend. lee et a l. conducted the relationship between the number of public transportation passenger and the weather, especially for ra in with the smart ca rd data [2]. this study revealed that both the number of bus and urban railway passengers were reduced in the rain. with these results, they pointed out that the public transportation passengers at seodaemun-gu, dongdaemun-gu, and jung-gu were easily influenced by the rain intensity. yi et a l., ana lyzed the rain intensity and the bus travel time to verify the quality of bus services. the study showed that the quality of the bus services was greatly influenced by rain start in the morning peak hours [3]. however, these researches focused on the specific regions so that it could not show a detail analysis. therefore, this study constructs the database set about the urban railway passenger with the * corresponding aut hor. email: iamwonho@gmail.com proceedings of engineering and technology innovation , vol. 4, 2016, pp. 01 03 2 copyright © taeti rainfa ll intensity in seoul and conducts the sp atial analysis to deduct passenger demand patterns. this paper collects ra infall data which a re collected fro m july to september and urban railway passenger data. with these data, this study analyzed the relationship between the number of urban ra ilway passenger and rainfall intensity. 2. method seoul is one of the biggest cities in the world and eight subway lines are connected so people can reach every single place in seoul. so many people use subway for co mmuting and other travel purposes. according to the kang et al., up to 36 % of people responded that they choose subway as a travel mode [4]. this paper analys es the rainfall intensity to show its impact on de mand of the seoul urban railway with the urban railway passenger data. this study use railway passenger data which are collected fro m ju ly to september 2012 and 2013. daily ra infa ll data which are used are collected fro m the automatic weather stations (aws). to analyse the relationship between the rainfall intensity and the subway passenger ridership, this study selects in that same period rainfa ll data and then conducts a spatial analysis. urban railway pas senger dataset from seoul metro (line 1-4) and seoul metropolitan rapid transit corporation (line 5-8) are obtained and total number of board passengers in july to september 2012, and 2013 are e xtracted. the total number of stations on line 1 to 4 in the 2012-2013 is 119 stations. however, total number of stations on line 5 to 8 is different. in 2012, there are 148 stations on line 5 to 8. however, 9 stations which are gulpocheon station, kkachiul station, bucheon cityhall station, bucheon stadium station, samsan gymnasium station, sangdong station, sinjung-dong station, and chun-ui station data from october 2013 added. 3. results and discussion 3.1. urban railway demand analysis this study set the number of passenger data in level 1 as a standard. based on this standard, this study compares each level’ passenger number and standard and then calculates redu ction ratio. as a result, this study finds out that the number of passengers is decreased during the rainfa ll. in leve l 2, total nu mber of passengers is increased by 0.53% than level 1. at level 3, total number of passengers is decreased by 1.84% and level 4, total number of passengers is decreased by 3.17%. de mand was reduced by 2.65% at level 5. at the level 6, it was reduced by 5.39% . detailed results are given in fig. 1. fig. 1 re lationship between urban railway passengers by rainfall level this study also conducted the passengers’ ridership on each station. the results show that the ridership on yeouinaru station and ttukseom resort station is highly influenced by amount of the rain. this is because people visit these stations for le isure activities. on the other hand, the ridership on samsung station, gasan digital co mplex station, and gangnam station does not have a significant difference by the rain. these stations are highly involved with work trip . contrary to this pattern, hangnyeoul station subway passenger rate is 47% increase fro m relationship. regard less of the rainfall, it can imply that this is because of the hangnyeoul station event. 3.2. spatial analysis on subway network by rainfall level a gis software can be used to store, analyze and allows spatial data layers. first, this study analyzes the high number of the board and alight passenger station in seoul. fig. 2 depicts the high total number of board passenger station analysis data from top 1 to top 10; gang-nam station: 37,181,419persons, jam-sil station: 27,974,601 persons, seoul station: 26,244,942 persons, sillim station: 26,201,372 persons, etc. fig. 2 a lso shows the high total number of alight passenger station: gang -nam station: 38,498,150 persons, jam-sil station: 25,588,647 persons, sillim station: 25,491,673 persons, hongik university station: 25,019,830 persons . proceedings of engineering and technology innovation , vol. 4, 2016, pp. 01 03 3 copyright © taeti fig. 2 board and alight passenger station analysis in seoul subway system 4. conclusions the crowdedness of the urban railway is the most important indicator to re flect the operation state of the urban railway. the weather is one of the critica l factors to influence the crowdedness. that is because even though the crowdedness of the urban railway is the same, passengers feel more uncomfortable in ra iny weather condition. indeed if specific sections and stations suddenly are concentrated excessive demand, it will lead far more serious problem. this study finds out that each station’s passenger is decreased by the rainfall level, e xcept for level 2. the nu mber of passengers in level 6 is decreased about 5.39%. the stations which are highly re lated to leisure activity are sen sitive to the rainfall. also, we ana lyzed the sensitivity area of the rainfa ll level. the station was analyzed by using a high sensitivity for the station leisure activities such as yeouinaru station and ttukseom resort station. on the other hand, sa msung station, gasan dig ital co mple x station, gangnam st ation was analyzed and the sensitivity is low. with the gis progra m, this study conducts the spatial ana lysis to show the relationship between the rainfall intensity and the subway passenger ridership in gis map. this study focuses on the entire date of data and passenger data. to conduct more accurate analysis, the study should consider peak-hour and non-peak hour data. acknowledgement this work is supported by a grant nrf-2014r1a1a2054793 and transportation & logistics research program 15ctap-c097344 of korean government. references [1] k. park and s. lee , “a study on the effect of adverse weather conditions on public transportation mode choice,” journal of civil engineering, ko rean society of civil engineers , vol. 32, no. 1, pp. 23-31, 2012. [2] k. lee, j. eu m, s. you, j. min, and k. yang, “the impact of ra in on public t ransit ridership in seoul,” the korean society for railway, 2014. [3] c. yi, j. ko, y. kang, and t. lee, “the impact of ra infall on public transport service in seoul focusing on the changes in punctuality and speed of bus service”, journal of korea planners association, vol. 46, no. 7, pp. 73-87, 2011. [4] y. kang, c. yi, and s. lee , “a study on the factors affecting crowd ing degree in the subway train considering the seoul metropolitan subway network and the land use of its catchment area,” journal of korea planners association, pp. 203-218, 2014.  proceedings of engineering and technology innovation , vol. 3, 2016, pp. 25 27 25 a wireless sensor network-speech recognition scheme using deployments of multiple kinect microphone array-sensors ing-jr ding * and shih-kai lin department of electrical engineering, national formosa university, yunlin, taiwan. received 22 february 2016; received in revised form 28 march 2016; accept ed 13 april 2016 abstract speech recognition has successfully been utilized in lots of applications recently. with the development of the kinect sensor device from microsoft, speech recognition could be further promoted to be used in an ubiquitous environment where a wireless sensor network using kinect sensors is deployed. this study develops a wire less sensor network (wsn)-speech recognition scheme using deployments of mu ltiple kinect microphone-array sensors. presented speech recognition by kinect-wsn could effectively capture the acoustic data made from the talking speaker and then perform the corresponding voice command control on certain target. in this study, different strategies to deploy multiple kinect microphone-array sensors for constructing an ubiquitous kinect-wsn speech recognition environment are investigated. several different acoustic sensing data fusion methods are also explored for achieving superior performance on kinect-wsn speech recognition. the presented method in this paper is evaluated the efficiency and effectiveness in an 5m×5m laboratory env ironment in which any of four test speakers is to ma ke the voice command anywhere. developed kinect microphone array sensor-deployed wsn speech recognition in this work is finely utilized in various different applications in control. keywor ds : speech recognition, kinect microphone arraysensor, wire less sensor network 1. introduction speech recognition has been a matured technique for human machine-interaction (hci) in the recent years. with the development of internet of things (iots) technology nowadays, the smart home scenario that most of equipment and devices in a fa mily are connected and commun icated with each other via wireless networks will be a practical integrated application. conventional speech recognition viewed as the category of voice control interactions is that the voice command data provided by the specific user is acquired by the mic rophone in a very short distance from the user [1, 2]. such voice-control speech recognition application can be widely seen in speech recognition on the smart phone platform and speech recognition on the central mu ltimedia control panel platform in car. in the new technology of iots nowadays, different to conventional voice command-control speech recognition, in order to control all things connected to the internet in a home, office or other indoor environments, a new strategy for speech recogn ition developments, wireless sensor network (wsn)-speech recognition will be attracted much attention and a new and challengeable technique issue. this study explores the utilization of the kinect sensor [3, 4] for sensing and then acquiring the acoustic voice command of the user in an office environment. as many persons know, in addition to gesture recognition by kinect [5-7], the kinect device can also be used to perform speech recognition due to the embedded microphone array design composed of four microphones [8]. speech recognition in this work will be performed in an acoustic sensing area that is properly deployed by multiple kinect microphone-arrays. in the presented scheme of wsn-speech recognition by multip le kinect microphone array-sensors in this paper, technical issues such as the (1) deployment method of the kinect mic rophone array, (2) the establishment of client server-based wireless sensor network by kinect sensors, (3) investigations of acoustic sensing data fusion methods, and (4) the possible application with practice in a real life using presented wsn-speech recognition by kinect microphone array sensors will be considered, which will be detailed in the following section . * corresponding aut hor. email: ingjr@nfu.edu.t w proceedings of engineering and technology innovation , vol. 3, 2016, pp. 25 27 26 copyright © taeti 2. speech recognition via wireless sensor network by using kinect microphone array-sensors the frame work of wsn-speech recognition by kinect microphone array-sensors explored in this study mainly contains kinect sensor deployments by two kinect microphone-arrays, client-server wsn establishments using tcp/ip protocol, acoustic sensing data fusion using a simple and computationally fast strategy. the developed framework is further performed in an application of voice sensing and remote control to the multimedia player component on a smart phone, which is depicted in fig. 1. fig. 1 wsn-speech recognition by deployed kinect microphone array-sensors and its control application fig. 2 sensed data fro m a kinect microphone array-sensor for data fusion calculations (four-channel voice data contained simultaneously in the unique kinect sensor) as depicted in fig. 1, a voice command fro m a speaking user is sensed by two kinect microphone array-sensors that are properly deployed in an office space. each sensed data from each of these two kinect sensors is sent to the server via the tcp/tp protocol, and the server end performs data fusion calculations for determining the recognition result of the sensed voice command. after the data fusion estimate, the recognized voice command is then sent to the smart phone device via the bluetooth protocol to carry out a series of functional control on the multimedia player application program. in this work, two kinect microphone array-sensors are designed to be appropriately localized inside to acquire almost most of all possible voice data. fig. 2 shows sensed data in a form of four channels. all these data come fro m the unique kinect microphone array-sensor. two kinect microphone array-sensors deployed in this work are composed of 2 sets of the four-channel sensed voice data. these data are then considered the content of the voice command using data fusion calculations. the data fusion strategy employed in this study is a voice energy-based method. as shown in fig. 2, each voice data fro m the kinect sensor have the different values of energy. the value of the voice energy is dependent on the amplitude value of the data in certain time duration. the microphone in the kinect microphone array has the large value of voice energy in case the voice data source (i.e. the speaking user) is located extreme ly near the microphone. co nversely, when the microphone in the kinect microphone array is far away from th e voice data source, the estimated voice energy to this microphone will be significantly small. based on the above design thought-line, the primary principle of the data fusion method in this study is that the microphone with the sensed data of large-sized values will have more effects on the recognition decision of voice commands. the simplest method based on such the designed fusion principle is that the data fusion result is the recogn ition outcome of the microphone receiver where the sensed data has the largest values of energies. in the part of control applications, the blu etooth (bt) protocol is employed in this work to handle transmissions of the fused recognition command. the bt connection tunnel is firstly established in an initialization process to form a peer-to-peer connection pair between the server (the command provider) and the end -device of the smart phone (the command receiver). for speeding up command transmissions via bt, a co mmand table containing a series of labels, each label with a text form representing a corresponding voice command, is properly devised. the multimedia player application platform in the smart phone will be finely operated by “remotely sensed voice commands made by the speaking user” under the regulation of the presented method. proceedings of engineering and technology innovation, vol. 3, 2016, pp. 25 27 27 copyright © taeti 3. conclusions in this paper, a wireless sensor network-speech recognition approach is presented by deploying the kinect microphone array-sensors. co mpared to conventional speech recognition, the presented frame work considering sensor deployments, sensing data fusion, wireless communication scheme establishments, and possible e xtension application with p ractice provides an acoustic sensing way for command control in the application of internet of things. in addition, the presented approach with the use of kinect sensors will a lso avoid the property of ‘surveillance’ and therefore can be much mo re acceptable by the users. acknowledgement this research is partially supported by the ministry of sc ience and technology (most) in taiwan under grant m ost 104-2815-c-150-01 7-e. references [1] i. j. ding, c. t . yen and d. c. ou, “a method to integrate gmm, svm and dtw for speaker recognition,” international journal of engineering and technology innovation, vol. 4, no. 1, pp. 38-47, 2014. [2] i. j. ding and y. m. hsu, “an hmm -like dynamic t ime wa rping scheme for automat ic s peech recognition,” mathematica l proble ms in engineering, vo l. 2014, artic le id 898729, 8 pages , 2014. [3] i. tashev, “kinect development kit : a toolkit for gestureand speech based human-machine interaction,” ieee signal processing magazine, vol. 30, no. 5, pp. 129–131, 2013. [4] z. zhang, “m icrosoft kinect sensor and its effect,” ieee mult imedia , vol. 19, no. 2, pp. 4-10, 2012. [5] i. j. ding and c. w. chang, “an eigenspace-based method with a user adap tation scheme for hu man gesture recognition by using kinect 3d data,” applied mathemat ical modelling, vol. 39, no. 19, pp. 5769-5777, 2015. [6] i. j. ding and c. w. chang, “feature design scheme for kinect-based dtw human gesture recognition,” multimedia tools and applications , pp. 1-16, july, 2015 [7] k. qian, j. niu and h. yang, “developing a gesture based re mote human-robot interaction system using kinect,” international journal of smart ho me, vol. 7, no. 4, pp. 203-208, 2013. [8] k. ku matani, t. ara kawa , k. ya ma moto, j. mcdonough, b. ra j, r. singh and i. tashev, “microphone array processing for distant speech recognition: towards rea l-world deployment,” proc. asia-pacific signal & information processing association annual summit and conference (apsipa asc), 2012.  proceedings of engineering and technology innovation , vol. 3, 2016, pp. 01 03 1 hybrid ofdm/ook modulations in ocdma scheme for free space optics chih-ta yen * , chia-yu liu department of electrical engineering, national formosa university, yunlin, taiwan . received 01 february 2016; received in revised form 15 march 2016; accept ed 02 april 2016 abstract the study proposes hybrid analog/digital transmission format scheme wh ich integrated optical code-division mult iple -access (ocdma) and polarization mu lt iple xing technique in free space optics (fso) trans mission. orthogonal frequency division mu ltiple xing (ofdm ) transmits as the analog format and (on -off keying) ook transmits as digita l format in the study, respectively. in the proposed hybrid ocdma system, it has high-speed transmission, signal security and low cost…etc. the multip le access interference (mai) can be e fficiently e liminated by using the balanced detection scheme at the receive end. keywor ds : optical code-division multiple-access (ocdma), free space optics (fso), orthogonal frequency division multiplexing (ofdm), multiple access interference (mai) 1. introduction in recent years, peoples become closer and closer relationship between the networks. be cause of hybrid coaxia l cable transmission network change into a fiber optical trans mission network, making people have a better internet e xperience. in the current study, we use the method of different states of polarizations to transmit analog/digital signals and th en into optical code-division mult iple -access (ocdma) networks. due to polarization modulator without using electronic circuit or impressed voltage to control, and has the same intensity of light signals in the outputs, which can mit igate fiber nonlinear e ffects and ma king the overa ll structure greatly improve the reliability of the tran smission side. at present, wired transmission interface is using fiber or coa xial cable to transmit signal. but in some place because of terrain effect (e.g., cliffs desert, both ends of the bridge). wireless transmission system is one of the solutions to solve topographical constraints , but it is always disturbed by channel interfe rence and shelter, severely reducing its effectiveness . we take advantage of the free space optics (fso) approach as the way for last mile trans missions [1]. in this paper, we present optical spectral amp litude coding ocdma (sac-ocdma) combin ing with polarization mu lt iple xing to transmit analog signal of orthogonal frequency division mu ltiple xing (ofdm) and dig ital signal of on-off keying (ook) signals at the sa me time . in the past year, sac-ocdma system is attracted more attention because of the multip le access interference (mai) elimination and pre serve the orthogonality among the users in the system [2]. ofdm technique has been widely adopted in broadband wired and wireless co mmunicat ion systems , because it is robustness to the inter-symbol interfe rence (isi) and multi-path fading by a dispersive channel [3]. 2. system structure description fig. 1 and 2 illustrate the structure of the proposed base stations (transmitters) and control station (receivers). this ocdma mult iple xing structure can transmit analog signal and digita l signal at the same time and still preserve the ability of mai e limination. we transmit ofdm signal as analog signal and ook signal as digita l signal. in this hybrid system, each base station needs a spectral encoder and a decoder in the receiver. the dfb laser array source is co nnected a 1 ( 1)n  splitter which p roviding light source to each base station, shown in fig. 1. the dbf laser source performs as optical carri* corresponding aut hor. email: chihtayen@gmail.com proceedings of engineering and technology innovation , vol. 3, 2016, pp. 01 03 2 copyright © taeti ers pass through a polarizer a polarization co ntrolle r (pc), and a polarization bea m splitter (pbs). the pbs allocates the orthogonal horizontal and vertical states of polarizat ions (sops) upper and lower arms, respectively. between base station and control station, fso is used to be this structure channel. the control station received a composite signal comprising the transmitted chips of all the active base station. as shown fig. 2, the received signal in the control station is divided in to (n−1) -th parts by using the 1×(n−1) splitter, and passed into each receiver. the signal was produced vertical and horizontal po larization states by using pbs. two sops of vertica l and horizontal output from the pbs which are split into an upper branch and the lower branch. the two diffe rent signals of vertical and horizontal passed balance detector, wh ich co mputed the correlation d iffe rence between the two spectra and eliminate the interfe rence of base station with un-matching codeword ind ividually. the upper branch uses ofdm de modulator to revert analog origina l signal and lo wer branch uses digital ook de modulator to revert dig ital ook signal. encoder #j ofdm transmitter #l pc pbs#1 vertical sop horizontal sop mzm mzm ( ) l c i base station #l 1 x k c o m b in e rdfb laser array source 1xk splitter pbc base station #1 base station #(n-1) digital signal transmitter #l fso channel fig. 1 the transmitter of proposing system 1x2 splitter 1x2 splitter ( ) l c i ( )lc i ( ) l c i ( )lc i + + ofdm demodulator #l desired data #l receiver #l decoder #j horizontal sop 1 x (n -1 ) s p litte r pbs#2 vertical sop receiver #1 receiver #(n-1) control station digital signal demodulator #l desired data #l pd #l, v1 pd #l, v2 pd #l, h1 pd #l, h2 fig. 2 the receiver of proposing system before analysis of the proposed system [4], the ofdm signal of the k-th base station is )(, to rfk : )2cos()( ,,, tfato rfkrfkrfk  (1) where ,k rfa is the a mplitude of ofdm signal, and ,k rff is the carrier frequency. and the analog ofdm signal is modulated by a mult i-wavelength dfb laser as optical carriers and then pass through awg encoder is:     cn i krfksk irectictopto 1 , )()()(1)(  (2) where ( )kc i denote the i-th element of the k -th whc codeword,  is the modulation inde x, sp is the transmitted optical power, and ( )rect i function is given by:     0 0 ( ) 2 2 2 2 2 c c c c rect i u v v n i n u v v n i n                        (3) at the receiver end, the instantaneous power spectral density (psd) of the vertical and horizontal sop’s received optical signals can be expressed by the following: , 1 1 ( , ) 1 ( ) 2 ( ) ( ) cnk sr v s o rf k i k p r v t p q t c i rect i             (4)       k k nc i kk sr h irecticb p vr 1 1 )()( 2 )(  (5) where psr is the received effective power at the input port of each receiver,  is the light source bandwidth, k is the number of base station. the relat ionship between bit-error-rate (ber) can employ as:   1 / 8 2 ber erfc snr (6) 3. performance analysis and discussion fso performance by several para meters for reference including geo metric loss, link marg in, the received power and bit error rate. in this paper, it considers the received power and bit error rate [5]. fig. 3(a) and 3(b ) illustrates the bit error rate (ber) re lative to active users in d ifproceedings of engineering and technology innovation , vol. 3, 2016, pp. 01 03 3 copyright © taeti ferent bit rate at the analog/digital signals, re spectively. it had more active user when it had low bit rate. (a) analog signal (b) digital signal fig. 3 illustrates the bit error rate (ber) rela tive to active users in different bit rate 4. conclusions in this study, we present hybrid an alog/digital spectral ocdma system. th is system transmits two different kinds of format signal wh ich contain analog and dig ital signals by using polarizat ion multip le xing technique. the proposed hybrid system can transmit analog and digital signals at the same time such as digital t v and mob ile phone system simu ltaneously. fso is the main t ransmission channel in the study, which has the advantages of low cost, easy construction and terrain without restrictions … etc. the simu lation results show that both ofdm signal as analog signal and ook modulation as digital signal in the proposed hybrid ocdma system perfo rm good performance in fso transmission. acknowledgement this study was supported in part by the ministry of sc ience and technology most 104-2221-e-150-044. references [1] p. t jondronegoro, “free-space optics for fixed wire less broadband,” pp. 2-20, oct. 2004. [2] z. wei, h. m. h. shalaby, and h. ghafouri-shira z, “modified quadratic congruence codes for fiber bragg-grat ing-based spectral-a mp litude-coding optical cdma systems,” j. lightwave technol., vol. 19, pp. 1274-1281, sep. 2001. [3] m. se lvi and k. murugesan, “performance of ofdm based fso commun ication systems using m-ary psk modulation,” international journal of co mputer applications , vol. 49, pp. 41-45, july 2012. [4] j. f. huang and c. t. yen, “phase noise suppression in multilevel optical code-division multiple-access network coding system with embedded orthogonal polarizations ,” optical engineering, vol. 45, pp. 065005, june 2006. [5] v. sharma and g. kaur, “modelling of ofdm-odsb-fso trans mission system under different weather conditions ,” advanced computing and co mmunication technologies (acct ), pp. 154-157, april 2013.  proceedings of engineering and technology innovation , vol. 2, 2016, pp. 14 16 14 synthesis and annealing study of siliconand fluorine-containing low bandgap conjugated polymers for solar cell applications cheng-hsien wu, cian-hao huang, pei-ni song, tzong-liu wang * department of chemical and materials engineering, national university of kaohsiung, kaohsiung, taiwan. received 07 march 2016; received in revised form 12 april 2016; accept ed 09 may 2016 abstract two donor-acceptor alternating conjugated copolymers were synthesized as the donor materials of the active layer in polymer solar cells via stille coupling reaction. both copolymers consisted of dithienosilole as the donor unit and nonfluorinated 2,1,3-benzooxadiazole or fluorinated 2,1,3-benzooxadiazole as the acceptor unit. the nonfluorinated polymer and fluorinated polymer were designated as p1 and p2, respectively. the structures of copolymers were confirmed by ft-ir, 1 h nmr, and 13 c nmr. optoelectronic properties of the polymer were investigated and observed by uv-vis spectrum, photoluminescence spectrum, and cyclic voltammetry. both polymers exhibited a panchromatic absorption ranging from 300 nm to 1100 nm and displayed low band gaps of 1.46 ev and 1.42 e v, respectively. both of the synthesized materials were used as the donor material in the bulk heterojunction (bhj) solar cells and then power conversion efficiency (pce) measurements were conducted in different weight ratios of the polymer:pc61bm blends . compared to the nonfluorinated one, the fluorinated polymer exhibited a higher pce of 2.67% at room temperature under the illumin ation of am 1.5 (100 mw/cm 2 ). study of the effect of annealing on the performance of the p1/pc61bm devices using a 1:2 blend ratio of p1:pc61bm exhibited the highest short-circuit current density of 5.88 ma/cm 2 and a power conversion efficiency of 2.76% at annealing temperature of 125 o c. the bulk heterojunction polymer solar cell employing p2 and pc61bm at a blend ratio of 1:2 exhibited the highest short-circuit current density of 6.44 ma/cm 2 and a power conversion efficiency of 3.54% at annealing temperature of 150 o c. keywor ds : donor-acceptor, solar cell, dithienosilole, 2, 1, 3-benzooxadiazole, low band gap, power conversion efficiency 1. introduction except the “weak donor–strong acceptor” concept for the design of d-a alternating copolymers with low band gaps and high vocs for solar cell applications [1], adding elec tron-withdrawing groups to the polyme r backbone is also considered as an effic ient approach. recently, fluorine has attracted much attention as an electron-withdrawing group in the synth esis of conjugated polyme rs for h igh -efficiency solar cells [2-3]. according to a theoretical study by brédas et a l., the introduction of fluorine atom as a substituent of the conjugated polymer would lo wer both of the hom o and lum o energy levels of the synthesized polymer [4]. on the othe r hand , s ilo le-cont ain ing a romat ic mo iet ies hav e attracted mu ch attent ion as ne w bu ild ing units o f con jugated po ly me rs because of the ir un ique optoe lect ron ic p ro p ert ies. as rega rds the ir ca rbon ana logues, silo le-conta in ing po ly me rs have quit e lo w-ly ing lum o leve ls, wh ich are att ributed to the fact that the -orb ita l of the si-c bond can e ffect ive ly interacts with the -orb ita l o f the butad iene frag ment [ 5]. by trans forming pcpdt bt to the s ilo le-conta in ing po ly me r, poly{[4,4'-b is(2-ethy lhe xy l)d ith ieno (3,2-b; 2', 3'-d )silo le ]-2,6-d iy l-a lt-(2,1,3-ben zoth id ia zo l e)-4,7-d iy l} h as been re cent ly rea lized and a pce of as high as 5.6% has been ach ieved [6]. there fo re , the d ithieno[3,2-b :2',3'-d]-s ilole (dt s) un it has frequent ly used as th e s i*corresponding aut hor. email: t lwang@nuk.edu.t w proceedings of engineering and technology innovation , vol. 2, 2016, pp. 14 16 15 copyright © taeti lo le -conta in ing donor un it in the deve lop ment of high-efficiency pscs . here in, we have investigated the effect of silole-containing and fluorination on the photovoltaic performances of the d -a conjugated copolymers. t wo d-a copoly mers co mprising dts unit and nonfluorinated and fluorinated 2,1,3-benzoo xad iazo le (bo) unit has been synthesized to study the effect of silicon and fluorine substitution. the effect of annealing on the device performance of the poly mer solar ce lls fabricated fro m the nonfluorinated and fluorinated polymers has been investigated in detail. 2. method 2.1. synthesis of polymers in this study, two d-a type copolymers were synthesized via stille coupling reaction. for convenience, they are designated as p1 and p2. the synthetic routes are shown in fig. 1. fig. 1 polymer synthesis 2.2. annealing study the dev ic e st ruct ure o f pscs fo r cu rrent density -vo ltage (j -v) me asu re ments is it o/ pedot : pss/po ly me r: pc61bm/ lif/a l. the films of act ive layers we re ann ea led d irect ly on top o f a hot p late in the glove bo x, and the te mpe ratu re was mon ito red by using a thermo couple touch ing the top o f the su bstrates. a fte r re mova l fro m the hotp late , the substrates were immed iat e ly put onto a meta l plate at the roo m te mp eratu re . ult rav io let -v isib le (uv-vis ) spectroscopic ana lysis was conduct ed on a perkin– elme r la mbda 35 uv-vis spectrophoto mete r. the j-v cu rves we re measured using a ke ith ley 2400 sourc e mete r, unde r illu min at ion fro m a sola r simu lato r. the intensity o f so la r s imu lato r was set with a prima ry re fe rence c e ll and a spectra l correct ion facto r to g ive the pe rformanc e u n der the am 1.5 (100 mw/c m2) g loba l re fe rence spectrum (iec 60904-9). 3. results and discussion fig. 2 j-v characteristics of devices for p1 a fter annealing at different te mperatures for 30 min fig. 3 j-v characteristics of devices for p2 a fter annealing at different te mperatures for 30 min figs. 2 3 show the j-v characteristics under white light illu mination (100 mw/c m 2 ) for photovoltaic devices subjected to thermal annealing. for the p1 poly mer, the device annealed at 125 °c shows the best efficiency of 2.76% , while the p2 poly mer shows the optimu m efficiency of 3.54% at 150 °c. thus, it is c lear that proceedings of engineering and technology innovation , vol. 2, 2016, pp. 14 16 16 copyright © taeti annealing treatment affects the device perfo rmance significantly and that there e xists opt imu m range of the annealing te mperature. it is noteworthy that the voc values of p2 a re higher than those of the p1 polymer because the introduction of f into the conjugated backbone has lowered the hom o and lumo energy levels of the copolymer. due to the values of voc a re re markable for the p2 poly me r, the power conversion efficiencies of the solar ce ll for p 2 a re much improved. 4. conclusions two d–a type copolymers based on dts as the donor unit and nonfluorinated and fluorinated bo as the acceptor unit have been synthesized and e mp loyed as the donor materia ls in the active layer of bhj -type polyme r solar cells. as a result of annealing treatment at an optimu m condition (125 °c/30 min ), the pv cell performance of p1 was dra matica lly enhanced and the power conversion efficiency of device reached to 2.76%. through the modification by fluorination on the bo units, the photovoltaic pe rformance of p 2 at 150 °c was much imp roved due to the increased jsc and enhanced voc and the power conversion efficiency of the device reached 3.54% under wh ite light illu mination (100 mw/cm 2 ). acknowledgement we gratefully acknowledge the support of the ministry of science and technology in taiwan through grant nsc 99-2221-e-390-001-m y3. references [1] h. zhou, l. yang, s. stoneking, and w. you, “a weak donor-strong acceptor strategy to design ideal poly mers for organic solar cells ,” acs appl. mater. & interfaces vol. 2, pp. 1377-1383, 2010. [2] h. y. chen, j. h. hou, s. q. zhang, y. y. liang, g. w. yang, y. yang, l. p. yu, y. wu, and g. li, “poly me r solar ce lls with enhanced open-circuit voltage and efficiency,” nat. photonics vol. 3, pp. 649-653, 2009. [3] h. x. zhou, l. q. yang, a. c. stuart, s. c. price, s. b. liu, and w. you, “development of fluorinated benzothiadiazo le as a structural unit for a poly mer solar ce ll of 7 % efficiency,” angew. chem., int. ed. vol. 50, pp. 2995-2998, 2011. [4] j. l. brédas and a. j. heeger, “influence of donor and acceptor substituents on the electronic characteristics of poly (para phenylene vinylene) and poly (paraphenylene),” che m. phys. lett., vol. 217, pp. 507-512, 1994. [5] s. ya maguchi, y. ita mi, and k. ta mao, “ group 14 metalloles with thienyl groups on 2,5-positions:  effects of group 14 ele ments on their π-e lectronic structures ,” organometa llics , vol. 17, pp. 4910-4916, 1998. [6] j. hou, h. y. chen, s. zhang, g. li, and y. yang, “synthesis, characterizat ion, and photovoltaic properties of a low band gap polymer based on silole-containing polythiophenes and 2,1,3-benzothiadia zole,” j. am. che m. soc., vol. 130, pp. 16144-16145, 2008.  proceedings of engineering and technology innovation , vol. 3, 2016, pp. 07 09 7 design of slim led coupler for collimated light source chun-hao sun, tun-chien teng * department of mechatronic engineering, national taiwan normal university, taipei, taiwan. received 25 february 2016; received in revised form 28 march 2016; accept ed 16 april 2016 abstract this study proposed a slim coupler to collimate the light e mitted fro m an led, which can be used as a collimated light source. the coupler is substantially a 2-d co mpound parabolic collector (cpc); its bottom surface has longitud inally e xtending v-groove microstructures thereon. the parabolic side wa lls re flect the propagating light to converge only in the transverse direction, but the v-groove mic rostructures reflect the light and make its propagating direction gradually rotates around a central a xis of the cpc. therefo re, the angular d istribution of the light finally converges after several times of reflection on both side walls and v-groove microstructures. moreover, the illu minance distribution on the outlet of the v-groove cpc becomes more uniform than a cpc without v-groove mic rostructures. the effects of v-groove microstructures on both angular and illu minance distributions of the light emerging fro m the outlet of the cpc is analyzed, and the feasibility of providing a uniform co llimated light source is verified by conducting simula tion. keywor ds : collimation, coupler, led, compound parabolic collector, illumination 1. introduction the liquid crystal display (lcd) needs an e xtra p lanar light source such as a backlight. as energy-saving is paid more attention an u ltra-collimated planar light source (ucpls) becomes an attractive solution to ra ise energy efficiency of the lcd. the ucpls concentrates the light toward the observer to avoid waste of energy. moreover, it is essential for achieving some advanced functions to greatly raise energy efficiency. severa l typica l designs of ucpls were proposed. the first is to fabricate a light guide plate (lgp) with special microstructures thereon that directly reflect the light propagating within the lgp into the norma l direct ion [1]. the second is to form grat ing dots on the lgp surface to diffract the light propagating within the lgp into the normal d irect ion. [2]. the third is to utilize an lgp co mposed of a stack of mu ltip le layers of diffe rent refractive inde x to ma ke the inner light emerge with narrow vertica l angular distribution [3]. the fourth is to pre-collimate the light emitted fro m an led by an ext ra optical co mponent before it enters an lgp and then to reflect the collimated light into the norma l d irection through microstructures on the lgp [4]. a mong those designs, using an e xtra pre-collimating co mponent is a simple r way but needs thicker or wider volu me to accommodate the component, which is adverse to a slim backlight. thus, we proposed a ucpls design, in which a slim led coupler was used to pre-collimate the light e mitted fro m an led and the overall thickness was 5 mm [5]. in this paper, we further analy ze the effects of para meters of the coupler on both spatial and angular d istrib ution of the output light to attain the optima l condition. 2. principle the led coupler used in the backlight must account for both light mixing and light-collimation. the coupler has a profile of the cpc to convert divergent light entering a small area (inlet) into collimated light emerging from a large area (outlet). the relationship between the angular distribution range (i.e., the “half-angle”) of the light entering the inlet and that of the outlet of cpc can be expressed as follows : 1 2 2 2 1 1 sin sin t n t n    (1) 2 11 2 22 2 1 )sin( )sin(   n n a a  (2) * corresponding aut hor. email: walt er.teng@ntnu.edu.tw proceedings of engineering and technology innovation , vol. 3, 2016, pp. 07 09 8 copyright © taeti eq. (1) is applicable for the 2-d case, where t1 and t2 is the thickness of the inlet and outlet of the cpc, respectively; eq. (2) is applicab le for the 3-d case, where a1 and a2 is the area of the inlet and outlet of the cpc, respectively; and n1 and n2 respectively represent the refractive inde x of the mediu m surrounding the inlet and outlet. the two equations indicate that the degree of collimation of the light e merging fro m the outlet depends on the ratio of the outlet area to the inlet area. as the ratio increases, both width and thickness of the cpc outlet become larger, thus adverse to backlight application. moreover, output a uniform collimated light bea m is another focus if we intend to transversely connect mu ltip le of the cpcs side by side to form sufficiently wide and uniform co llimated light source for the lgp. for a cpc that can be accommodated in a slim backlight whose space is relatively wide but very thin, the horizontal angular distribution of the output light of the cpc would be re latively narrow, but the vertical angular distribution would be re latively wide. to solve this issue, an approach combining a cpc with v-groove microstructures was proposed. the coupler was designed as a thin and long cpc with a rectangular cross -section, and an array of v-groove microstructures longitudinally e xtended along the z-a xis was form on its bottom surface as shown in fig. 1. the light propagating along the z-a xis within a thin but wide cpc e xhib its a narrow horizontal angular distribution (projected on the x-z plane) and a re latively wider vertica l angular distribution (projected on the y -z p lane). when a light beam with a narrow horizontal angle but comparatively wider vertical angle is incident to the slope facet of the v-groove microstructures, it is reflected by total internal re flection (tir), and then continues propagating conversely with a narro w vertica l but co mparatively wider horizontal angle. such a light beam with a wider horizontal angle is more likely to hit the curved surface on the left or right side of the cpc. consequently, it is reflected by tir, and then continues propagating with a narrower horizontal angle; in other words, the beam b ecomes more collimated. after many similar cycles, the original light propagating within the cpc converges both vertically and horizontally. in the process, the length of the cpc and the angle of the apex of the v-groove microstructures have effects on both spatial and angular distributions of the output light. fig. 1 cpc with v-groove microstructures 3. results and discussion in this study, we established an optical model whose related para meters are detailed as follows. a cpc was made of poly methyl methacrylate (pmma), whose dimensions of its inlet and outlet were 3.6 1.2 (mm) and 13.82 2.4 (mm), respectively; its length was 50– 70 mm. a series of simu lations we re performed for different conditions such as: ape x angles of v-grooves and types of the profile of the cpc. ape x angles included 90 o , 110 o , 130 o , and 150 o ; types of the profile had three: width and thickness simultaneously increasing (case a); width increasing first and then thickness (case b); thickness increasing first and then width (case c). table 1 uniformity of illu minance vs. ape x angles case a case b case c 0 o 0.80/0.87 0.88/0.90 0.83/0.85 90 o 0.69/0.76 0.79/0.90 0.69/0.85 110 o 0.92/0.94 0.97/0.98 0.95/0.95 130 o 0.84/0.99 0.98/0.99 0.80/0.97 150 o 0.89/0.98 0.98/0.99 0.78/0.93 table 1 lists uniformity of illu minance in the three cases under the conditions of different apex angles. uniformity is defined as the ratio of minimal illu minance to ma xima l. angle of zero means the bottom of the cpc is bare flat; the uniformity data behind the slash are those of the cpc that is e xtended by an extra length of 20 mm with a constant cross -section. the results indicate that the apex angle of 90 o is adverse to uniformity for the three cases; apex angles of 110 o , 130 o , and 150 o imp rove uniformity for case a and b; only the apex angle of 110 o improves uniformity for case c. case b performs the best uniformity; case c performs the worst. however, a ll the uniformity is improved for the three cases when the cpc is extended. proceedings of engineering and technology innovation , vol. 3, 2016, pp. 07 09 9 copyright © taeti angular distributions of output light of the cpc for case a, b, and c are shown in fig. 2. the left insets are those in the vertical direct ion; the right insets are those in the horizontal direc tion. the effects of ape x angles of v-groove microstructures on angular distribution are demonstrated. the v-groove microstructure narrows the angular distribution in the vertical, and up to 47% at most as co mpared with the cpc without microstructures. however, it also slightly broadens the angular distribution in the horizontal. a mong the three cases, case b has narrower and smooth vertical angular d istrib ution, but its horizontal angular distribution is wider. considering uniformity of illu minance, case b is an optimal solution. fig. 2 angular distributions of output light of the cpc: (a) case a; (b) case b; case c 4. conclusions in this paper, effects of v-groove mic rostructures on both angular and illu minance distributions have been demonstrated for the three types of cpcs. in the slim volu me , the vertica l angular distribution and illu minance uniformity of the output light is narrowed and improved, respectively. the cpc of the profile with width increasing first and then thickness is an optima l solution for a slim coupler p roviding a uniform collimated light source because of its smooth angular distribution and e xcellent illu minance uniformity. references [1] j. h. lee , h. s. lee, b. k. lee , w. s. choi, h. y. cho i, and j. b. yoon, “simp le liquid crystal display backlight unit co mprising only a single-sheet mic ropatterned polydimethylsilo xane (pdm s) light-guide plate,” opt. lett., vol. 32, no. 18, pp. 2665-2667, 2007 [2] s. r. park, o. j. kwon, d. shin, and s. h. song, “ grating mic ro-dot patterned light guide plates for led backlights,” opt. express , vol. 15, no. 6, pp. 2888-2899, 2007. [3] j. w. pan, c. w. fan, “high lu minance hybrid light guide plate for backlight mo dule applicat ion,” opt. expres s, vol. 19, no. 21, pp. 20079-20087, 2011. [4] w. h. yang, h. h. lin, c. j. hsu, and y. n. pao, “ led coupler lens array for one dimensional co llimat ing backlight,” idw., vol. 11, pp. 1399-1400, 2011. [5] t. c. teng and l. w. tseng, “slim p lanar apparatus for converting led light into collimated polarized light uniformly e mit ted fro m its top surface,” opt. express , vol. 22, no. s6, pp. a1477-a1490, 2014.  proceedings of engineering and technology innovation , vol. 3, 2016, pp. 31 33 31 experimental study of the crack control of concrete by self-healing of synthetic fiber reinforced cementitious composites synthetic fiber heesup choi 1,* , masumi inoue 1 , takechi ai 1 , tomoya nishiwaki 2 , myungkwan lim 3 , hyeonggil choi 4 1 department of civil engineering, kitami institute of technology, hokkaido, japan . 2 department of architecture and building science, tohoku university, sendai, japan . 3 graduated school of engineering, hankyong national university, kyonggido, korea . 4 faculty of environmental technology, muroran institute of technology, hokkaido, japan . received 25 january 2016; received in revised form 10 february 2016; accept ed 06 march 2016 abstract in this study, it is possible to disperse effec tively cracked using synthetic fiber, an e xa mination of the most suitable self-healing conditions was performed on the above crack width 0.1mm. as a result, e ffective c rack dispersion using polyvinyl alcohol (pva ) fibers with pola r oh groups, as well as improved self-healing for cracks that are la rger than 0.1 mm in width, posing concerns of co2 gas and cl penetration, were observed. also, co3 2 reacts with ca 2+ in the concrete crack, resulting in the precipitation of a carbonate compound, caco3. based on this, it is deemed possible for the recovery of effec tive watertightness to be made fro m c racks that are la rger than 0.1 mm in width. in addition, it was determined that, as for the most suitable self-healing conditions in the inside and surface of the cracks, calciu m hydro xide (ca(oh)2) solution with co2 mic ro-bubble was more effective in pro moting the self-healing capability than water with co2 micro-bubble. keywor ds : micro crack, pva, co2 micro-bubble, s elf-healing,ca(oh)2, caco3 1. introduction although concrete is one of the most widely used construction materials, it is characterized by substantially low tensile strength in comparison to its compression strength, and occurrence of cracks is unavoidable. in addition, cracks progress due to environmen tal conditions including da mage by freezing, neutralization and salt damage, etc. moreover, detrimental da mages can occur in concrete structures due to the permeat ion of deteriorat ing ele ments such as cl and co2, etc. meanwh ile, under an environment in wh ich moisture is being supplied, if the width of the crac k is sma ll, a phenomenon of self-healing in which the portion of the crack is filled in due to the rehydration of the ce ment particles and precip itation of caco3 has been confirmed. as for the self-healing mechanis m of concrete, the reaction between ca 2+ in concrete and co3 2 dissolved in water produces caco3, a carbonate compound that does not easily dissolve in water. this phenomenon leads to the filling and c losing of crac ks [1]. the calc ite crystal reactions are shown in equations [1-3] h2o + co2⇔h2co3 ⇔h + + hco3-⇔2h + + co3 2 (1) ca 2+ + co3 2 ⇔ caco3 {phwater> 8} (2) ca 2+ + hco3 ⇔ caco3 + h + {7.5 <phwater<8} (3) therefore, this study effectively disperses the cracks in the cracks in the ce mentitious composite materia ls by using synthetic fiber, and, for the cracks with a width of more than 0.1mm, e xecuted the review of the optima l * corresponding aut hor. email: hs-choi@mail.kitami-it.ac.jp proceedings of engineering and technology innovation , vol. 3, 2016, pp. 31 33 32 copyright © taeti self-healing conditions along with the revie w of a diverse range of self-healing performance. 2. experimental overview the mixture proportions of the mo rtar a re summarized in table 1. portland ce ment (c, density: 3.16 g/c m 3 , mean d ia meter: 10 μ m), quartz sand as the fine aggregate (s, surface-dry density: 2.61 g/cm 3 , mean dia meter: 180 μ m), and a high-performance water reducing agent as an admixture (sp, density: 1.05 g/cm 3 , ma in constituent: polycarboxylate-based superplasticizer) we re used. as for the synthetic fibers, pva (fiber dia meter: 40 μ m, fiber length: 12 mm, density: 1.3 g/c m 3 ) and polyethylene (pe) (fiber dia mete r: 12 μ m, fiber length: 12 mm, density: 0.97 g/c m 3 ) and polypropylene (pp) (fiber dia mete r: 65 μ m, fiber length: 12 mm, density: 0.91 g/cm 3 ) were used. table 1mixture proportions of the mortar type s/c (wt.%) w/c (wt.%) sp/c (wt.%) fiber (vol.%) pva 0.4 0.3 0.4 1.2 pe 0.45 pp 0.3 table 2 experimental factors and conditions experimental factors conditions fiber pva, pe, pp self healing water + micro-bubble ph 6.0 ca(oh)2 + micro-bubble ph 8.5 temperature 20°c crack (target of crack width: 0.3mm) tensile load period of self-healing 7 days as the experimental procedure, in st ep a, an analysis of the permeability coeffic ient i mmed iately following the introduction of the cracks by the tensile loading test (prior to self-healing), and an analysis of the types of the hydrates prior to self-healing by using the raman spectroscopy were executed. in st ep b, a comparison and evaluation were made by using the method applied in st ep a in order to eva luate the changes in the permeability of each of the specimens due to self-healing, changes in the structure within the cracks and a quantitative evaluation the self-healing precip itated substances. the experimental factors and conditions are summarized in table 2. 3. results and discussion 3.1. permeability coefficient as the results of the experiment, w+m b in st ep b, when compared with that in st ep a, displayed the trend of an increase in the re sistance to permeability by about 40 fo lds for pva, 3.5 fo lds for pe and 1.5 folds for pp(fig 1). meanwhile, ca +mb in st ep b, when co mpared with that in st ep a, displayed the trend of an increase in the resistance to permeability by about 460 folds for pva , 60 fo lds for pe and 6.0 folds for pp (fig 2). in addition, in the comparison of the permeability coeffic ient follo wing self-healing, ca+m b, in co mparison to w+mb, displayed the trend of improve ment in the re sistance to permeability by appro ximately 15 folds for pva, 17 fo lds for pe and 4.0 fo lds for pp (fig 3). fro m the afore mentioned results, it can be discerned that the resistance to permea bility is improved in the order of pva>pe>pp, regardless of the conditions of self-healing. in particular, pva with oh-radica l d isplays a more e ffective self-healing performance [2], and, it was confirmed that the conditions of ca+m b are more advantageous than the conditions of w+mb for the pro motion of self-hea ling performance. therefore , for the micro crac ks with a width of mo re than 0.1mm for wh ich substantial permeat ion of deteriorating ele ments fro m the e xternal into the internal aspects of the concrete is anticipated, it is deemed that the generation and precipitation of the self -healing substances were pro moted due to the mixing of the pva fiber with the oh radica l a long with the introduction of the conditions of self -healing of the saturated ca(oh)2 solution (ca 2+ ) that contains co2 micro bubbles (co3 2). 3.2. review of the crack section due to self-healing fig 4 shows the e xperimental results of ra man spectroscopy. here, the locations of the occurrence of the peak of a wave generated by a laser at the crack section to which the white colored precipitated substance of pva specimen is attached and that at the sectio ns without cracks were co mpared. as the results of the e xperiment, with the peak of the wave of ca co3powder as the subject of comparison, there was almost no occurrence of the peak in the proceedings of engineering and technology innovation , vol. 3, 2016, pp. 31 33 33 copyright © taeti wave of the sections without cracks that coincides with that of the caco3. ho wever, the peak of the wave in the c rack section accurately coincides with the peak of the wave of ca co3 powder. accordingly, it is concluded that the ma jority of the white colored precipitated substance was generated following self -healing due to caco3. fig. 1 permeab ility coeffic ient of water + micro bubble (w + mb). fig. 2 permeability coeffic ient of ca (oh)2 + micro-bubble (ca + mb). fig. 3 co mparison of permeability coefficient ratio fig. 4 precipitated substances of self-healing 4. conclusions in this paper, it was confirmed that the effective restoration of water t ightness and the ma jority of the self-hea ling products was due to ca co3 by using synthetic fiber with polarity, along with the effect of inducing a multip le number of hairline cracks. in addition, it was confirmed that the self-healing conditions of the saturated ca(oh)2 solution, wh ich supplied the co2 micro-bubble, displayed the most effective self-healing performance in the surface and internal sections of the cracks. acknowledgement this work was supported by the young researcher program of lixil 2015. references [1] c. edvardsen, “water permeability and autogenous healing of c racks in concrete,” aci materials journal, vol. 96, no. 4, pp. 448-454, 1999. [2] d. ho mma , h. m ihashi, and t. nishiwa ki, “self-healing capability of fiber re inforced cementit ious composites ,” journal of advanced concrete technology, vol. 7, no. 2, pp. 217-228, 2009. 0.0000001 0.000001 0.00001 0.0001 0.001 0.01 0.1 1 pva pe pp c o e ff ic ie n t o f w a te r p e rm e a b il it y (m /s e c ) step. a step. b 0.0000001 0.000001 0.00001 0.0001 0.001 0.01 0.1 1 pva pe pp c o e ff ic ie n t o f w a te r p e rm e a b il it y (m /s e c ) step. a step. b 0.0001 0.001 0.01 0.1 1 pva pe pp r a ti o o f c o e ff ic ie n t o f w a te r p e rm e a b il it y w+mb ca+mb 1 = standard 0 5 10 15 20 25 0 500 1000 1500 2000 in te n s it y ( c o u n ts ) raman shift/cm-1 0 5 10 15 20 25 0 500 1000 1500 2000 in te n si ty ( c o u n ts ) raman shift/cm-1 0 5 10 15 20 25 0 500 1000 1500 2000 in te n si ty ( c o u n ts ) raman shift/cm-1 non crack caco3 crack caco3 p owder  proceedings of engineering and technology innovation, vol. 7, 2017, pp. 20 24 preliminary study on a system for visualization of big data in smes yasuo uchida 1,* , miao xinyun 1 , seigo matsuno 1 , yasushi iha 2 , makoto sakamoto 3 1 department of business administration, national institute of technology, ube college, ube, japan 2 department of media information engineering, national institute of technology, okinawa college, nago, japan 3 department of computer science and systems engineering, university of miyazaki, miyazaki, japan received 19 july 2017; received in revised form 30 july 2017; accept ed 03 august 2017 abstract the 2012 white paper on information and co mmun ications in japan issued by the ministry of internal affa irs and co mmunications of japan advocates use of big data under its “special the me: ict -induced and accelerated disaster recovery and japan’s re-birth.” however, the analysis in the japan users association of informat ion systems’ white paper on its 2014 it trend survey for co mpanies reports that less than 10% of co mpanies utilize big data, and it would appear that progress in its use is centered on large firms. under such conditions, use of big data is becoming a challenge for the purpose of ensuring the survival and success of sm es as we ll. as a result, r&d and technological support for smes are becoming pressing issues. however, at present there has been almost no academic research concerning policies and future direct ions for use of big data at smes. accordingly, th is study conducted the modelizat ion of the procedure fo r v isualization of big data fo r sm es. specifically, we o rganized the procedure as a tutorial, fro m obtaining data of japanese hot -spring areas using web scraping, to visualizing them using the visualization software cytoscape keywor ds: big data, visualization, smes, cytoscape 1. introduction this study is intended to research and develop a system for visualizat ion of big data suited to smes, as a tactica l informat ion tool to support sm es’ strategies for success under conditions of increasing ly intense global co mpetition. that is, it aims to probe a fra me work that is easy to adopt and superior in terms of operability for the collection, storage, analysis, a nd use of big data. at the same time, it also a ims to eluc idate e mpirica lly the ideal fo rm of a strategic information infrastructure for smes and challenges in its operation and administration. in this study, we carried out a preparatory consideration of visualization of big data by smes. specifically, we organized the procedure as a tutorial, fro m obtaining data of japanese hot-spring areas using web scraping, to visualizing the m using the visualization software cytoscape. 2. trends in use of big data at smes in japan at present, there are very few e xa mp les of successful use of big data by japanese smes. in addition, how big data is used at smes depends on individual planning by each company. accordingly, this paper will begin by summarizing measures taken and research trends related to the use of big data at japanese smes. it also will e xa m ine a nu mber of e xa mp les of early adopters. * corresponding author. e-mail address: uchida@ube-k.ac.jp tel.: +81-836-35-7567; fax: +81-836-35-7567 proceedings of engineering and technology innovation, vol. 7, 2017, pp. 20 24 copyright © taeti 21 for e xa mple , the report “ enriched living and economy fro m connected it : the va lue and re liab ility of big data”[1] fro m the research group on it infrastructure for living and the economy of the informat ion -technology promotion agency, japan (ipa) (it infrastructure for living and the economy of the informat ion -technology promotion agency, japan) both e xpla ins in simp le terms what big data means for managers of co mpanies aiming to provide new services using big dat a and identifies results such as expansion of business opportunities through summa rizing e xa mp les of early adopters of big data, advantages and issues in service realizatio n, and efforts to resolve these. in addition, the 2014 white paper on small and mediu m enterprises in japan [2] fro m the sma ll and mediu m enterprise agency mentions use of data on corporate transactions (big data) as a “key” to revitalization of regional economies. looking at the activit ies of sm es in the fie ld, in nove mber 2014 the osaka cha mber of co mmerce and industry published the results of a survey intended to ascertain matters such as expectations, needs, and issues involved in use of bi g data by second-tier co mpanies and smes [3]. while the results of this survey show that approximate ly 81% of co mpanies are interested in “informat ion (data)” as “useful for manage ment purposes,” respondents also identify the following as the top three “issues in use” of data:  “difficulty of understanding the cost-effectiveness of use of information (data)” (64.9%)  “lack of human resources to analyze information (data)” (56.9%)  “lack of understanding of methods of using information (data)” (34.0%) accordingly, we decided to proceed with research and development focusing on these three points. first of all, we identified as a necessary condition the ability to use personal computers having specificat ions like those used in ord inary administrative-leve l operations instead of high -priced co mputers, to keep costs down as much as possible. we also decided t o use, in principle, software such as open -source software that can be used free of charge as tools needed for analysis and visualizat ion. another prerequisite we identified was that the data analysis must be of a degree capable of being conducted b y emp loyees who have the skill levels needed to analyze data using spreadsheet software (such as microsoft exce l), since it is difficult for sm es to secure staff that have sp ecialized data analysis skills. furthermo re, we decided to provide hints on use of data by describing specific examples of methods of their use. 3. visualization of big data 3.1. steps from data collection through visualization the data subject to visualizat ion can be broken down into two main categories. the first consists of data in the possession of the co mpany itself. in this case, the co mpany has ascertained the content of the data sufficiently and it is easy for it t o process the data on its own. the other category consists of data that is present on the internet. in this case, it is difficult to understand the structure of the data and they are not easy to obtain. however, somet imes sm es will want to obtain and utilize these data . accordingly, this study will consider the steps used when obtaining and processing data present on the in ternet. since the main objective of this study is to illustrate a data processing model, we limited the purposes of visualization itself to the foll owing content:  subject data to be collected: data on hot-springs resorts in japan, published on the internet  purpose of visualization: to visualize the locations and water qualities of hot -springs resorts  steps in visualization: obtaining data through web scraping [4], conducting a number o f preprocessing steps, and then using cytoscape [5] to import the data as network information and visualize it in the form of graphs. proceedings of engineering and technology innovation, vol. 7, 2017, pp. 20 24 copyright © taeti 22 3.2. data acquisition and processin g when obtaining data through web scraping, the permission of the data provider must be obtained in advance. there is a need to consider how to avoid burdenin g the servers and network when actually obtaining the data. besides , the end-user license agree ment must be co mplied with fo r the data obtained. a lthough we used the python language [6] as a software environment for obtain ing and processing data, we arrang ed the model as one consisting of steps that could be used even by non-specialists, with consideration for ease of use. (1) analysis of web pages there was a need to analy ze the data structure of web pages and identify the data obtained. th is can be done using the view source” feature of a web browser (fig. 1). fig. 1 example of displaying a web page’s source (2) we used a python program to obtain the desired data from within web pages through web scraping. in this study, we obtained only data on the names and water quality of hot-springs resorts . (3) we used the python program to look up the latitude and longitude of the hot -springs resorts in google maps [7]. (4) we processed the above data using exce l and other tools and saved it as network data. an e xa mp le of the fo rmat of the data is provided below. in th is case, we used latitude as the y-a xis va lue on the graph a fter inverting positive and negative signs, since display coordinates and axial directions on the monitor are reversed . sample network data format: prefecture name, hot-springs resort name, water quality, x coordinate (longitude), y coordinate (latitude) 3.3. visualization using cytoscape cytoscape is a tool for visualizat ion of networks (through a graph structure). for this reason, the subject of processing needs to have a network structure. accordingly, we decided to analy ze the locations of hot -springs resorts and their local prefecture capitols, as an exa mp le of a network. fig. 2 shows an exa mple of v isualization of information on hot -springs resorts in ya maguchi prefecture resulting fro m loading network data to cytoscape and color-coding the informat ion by water quality. proceedings of engineering and technology innovation, vol. 7, 2017, pp. 20 24 copyright © taeti 23 in the center of the graph is the ya maguchi prefecture capitol. fro m this graph, the reader can identify the mutual positioni ng from the latitudes and longitudes on the map and the water quality from the color coding of the hot -springs resort names. fig. 2 visualization of hot-springs resorts in yamaguchi prefecture 4. discussion in this study, ultimately we visualized structured data. first of all, the origina l data source of web page source text (htm l) is semistructured data [8]. we followed the method of web scraping to obtain data fro m the content of web pages. ne xt, when the data obtained as in this study are composed of multip le files, there is a need for steps such as data abstraction and combination. this p rocess requires use of data-processing tools and programming languages. in addition, occasionally it is impossible to apply general-purpose tools to conversion of semistructured to structured data, and in such cases one must rely on programming languages. also, in the case of a locale such as japan that employs mult ibyte characters, sometimes text code conversion is required [9]. 5. conclusion this study employed a preparatory consideration of a system for visualizat ion of big data at smes, eluc idating a number of require ments. that is, it showed that in processes such as data collection and data processing there are many cases in which it is difficult to process the data using general-purpose tools alone. topics for future study include development of independent tools to supplement general-purpose tools as well as development of general-purpose models for the steps involved in visualization and preparation of tutorials suitable for use by smes. acknowledgement this work was supported by jsps kakenhi grant number 15k03639 . references [1] it infrastructure for living and the economy of the information -technology promotion agency, japan, “enriched living and economy from connected it: the value and reliability of big data,” http://www.ipa.go.jp/files/000001884.pdf. [2] small and medium enterprise agency, “2014 white paper on small and medium enterprises in japan,” http://www.chusho.meti.go.jp/pamflet/hakusyo/h26/pdf/h26_pdf_moku ji.ht ml. [3] the osaka chamber of commerce and industry, “results of survey on use of big data,” press release, 2014. [4] l. richardson, “beautiful soup,” available via https://www.crummy.com/software/beautifulso up/. cited 26 january 2016 https://www.crummy.com/software/beautifulsoup/ proceedings of engineering and technology innovation, vol. 7, 2017, pp. 20 24 copyright © taeti 24 [5] cytoscape consortium, “cytoscape,” http://www.cytoscape.org/. [6] python software foundation, “python,” https://www.python.org/. [7] python software foundation, “pygeocoder,” https://pypi.python.org/pypi/pygeocoder. [8] d. quass, a. rajaraman, y. sagiv, j. ullman , and j. widom, “querying semi structured heterogeneous information,” journal of systems integration, vol. 7, no. 3, pp. 381-407, 1997. [9] “the unicode consortium,” http://unicode.org/.  proceedings of engineering and technology innovation, vol. 7, 2017, pp. 31 36 using text mining to extract issues for school: an empirical study of the social platform-dcard hsin-yi wang, shu-fen chiou * , jung-wen lo department of information management, national taichung university of science & technology, taichung, taiwan, roc. received 18 july 2017; received in revised form 15 august 2017; accept ed 30 august 2017 abstract nowadays, social network within sentiment analysis has become the main trend in text mining domain. there are many platforms have been analyzed, such as facebook, twitter, instagram, and so on. in our manuscript, we attempt to extract the information about the sentiment polarity of messages (positive, neutral or negative) in a social platform “dcard”. the users of dcard are taiwanese college students, and anonymou s post is being used this in social platform, therefore, the user can express their opinion more freedom. we use dcard to the sentiment polarity of messages in extract the information about the school; moreover, the school could get the feedback from this finding to improve their policy. in this paper, we used python to scrap the web page, and the sentiment lexicon would be built. keywords: text mining, big data, social platform, sentiment 1. introduction nowadays, internet is used for communication widely. people used the internet to be browsing the web and collecting data (79.9%), using community websites (24.1%), playing online games (19.4%), listening to music or watching movies (17.8%), (15%) shopping online (10.3%) [1]. for community websites, people prefer communicating via the internet services over talking face-to-face or writing letters. they are more often writing blogs or posting messages on social networks and the personality will be presented by habitual vocabularies they used. in this research, we t ry to s analyzed chinese vocabularies on a social platform named dcard [2]. with the accumulation of large amounts of educational data, the use of advanced statistical techniques, such as data mining, exploring the potentially useful information or realizing some knowledge from bunch of data [3]. another important material for structured data is unstructured data composed of free text ; for example, scientific research papers, patented technical documents, qualitative interview data, and open questionnaires the content of the analysis, etc. with the analysis or research value. analysis of these unstructured data exploration methods, relying on the free text of the advanced processing and statistical operation s, that is, text mining technology[4]. dcard anonymous posting mechanism has successfully attracted tens of thousands of domestic and foreign college students to become platform members, this research is using dcard community behavior of the chinese corpus , the part of the analysis from the dcard text behavior can be analyzed everyone comment on the community on the web. * corresponding author. e-m ail address: fen057@ nutc.edu.tw tel.: +886-4-22196391 proceedings of engineering and technology innovation, vol. 7, 2017, pp. 31 36 copyright © taeti 32 in this paper, we use the word surveying technology to analyze the administrative quality of the campus in order to avoid the loss of the students, but also to attract the resources and improve the quality of education. the results data can be used as reference for the follow-up research reference and campus administrators. 2. related works in this section, we first introduce the technique which we use to extract chinese vocabularies. and we describe the social platform named dcard 2.1. python python is a widely used high-level programming language for general-purpose programming, created by guido van rossum and first released in 1991. an interpreted language, python has a design philosoph y which emphasizes code readability (notably using whitespace indentation to delimit code blocks rather than curly braces or keywords), and a syntax which allows programmers to express concepts in fewer lines of code than possible in languages such as c++ or java. the language provides constructs intended to enable writing clear programs on both a small and large scale [5]. 2.2. dcard dcard is the first exclusive college student dating site platform in taiwan. the official website is show in fig. 1 [6]. the dcard platform attracts many college students to join to expand their circle of friends, but to join must provide school e -mail, with the real name and photos through the administrator can be certified . dcard using anonymous system to register, many s tudents published articles to take anonymous system to express their own emotions, and these acts through the text presented in the article. fig. 1 dcard website 2.3. text mining in a large number of data, there are digital structure of the structured data and text, sound, image of the unstructured data. there are already hundreds of ways to deal with structured data. [7] text mining from biological literature is emerging as one of the main issues in bioinformatics research, and nlp methods are regarded as being useful to raise the potential of text mining from this literature. while the techniques are separated relatively into domain-portable, reference materials, except for corpora. [8]. proceedings of engineering and technology innovation, vol. 7, 2017, pp. 31 36 copyright © taeti 33 text mining is a new field, the text of the survey can be collec ted from the text of meaningful messages, for the file can also be analysed out of his specific purpose of the target, the text is undiscovered, invisible, and difficult in the algorithm. h owever, in modern culture, the document is the most common tool for the formal exchange of messages. the field of textual exploration usually involves a document whose function is to convey a factual message or opinion. [9] the steps of knowledge exploration : [10]  data collection  data cleansing  data conversion  application of exploration technology  the results are presented and interpreted the methods of knowledge exploration:  association analysis  classification  clustering  summarization  prediction  sequence analysis 3. our proposed method the work presented in this article covers, on the one hand, the extraction of information about user’s positive/neutral/negative sentiments from the text title they write. we investigate the detection of sentiment changes with respect to the “usual” sentiment of each us er [11]. in fig. 2, we present the methods ’ flowchart with each purpose below: fig. 2 method table 1 marking positive 其實我們的學校真的挺不錯的(actually our school is really good) negative 畢業證書之學校爛行政(graduation certificate of the school rotten administration) neutral 問學校語言學習軟體(ask for school language learning software) proceedings of engineering and technology innovation, vol. 7, 2017, pp. 31 36 copyright © taeti 34 table 2 dcard to word web text table 3 preprocessing 2 before the process after the process 【沒朋友 徵室友,我又來了】 (【no friends levy roommate, i have come】) 沒朋友徵室友我又來了 (no friends levy roommate, i have come) table 4 keywords main classification keyword course 課程 課名 老師 administration 足球隊(football team) 行政(administrative) 校園(camp us) step 1: we need data collection and marking. this research using the article title to classification to positive , neutral and negative includes school course and administration tow part. we give an example in table 1. step 2: we use python to extract dcard website to the text content. we extract dcard website by using python api url. we extract dcard website convert to text content, combine and export to document. table 2 are our examples. step 3: sorting out the data we export, and remove the extra number and english letter (the web content which is not created by the dcard user), and convert the full shape character to half shape character. step 4: remove the punctuation in the data (e.g. #, semicolon, comma and space) like table 3. step 5: create keywords. classification the pre-process data we created for the school course and administration. for an example in table 4, we set the course and its name for the school course partition, and the “soccer team” for administration due to the students usually use 4. data analysis and results 4.1. data we collect the dcard information from 2014 to 2017 in table 5. the results of the original data have number of 24939. the data have been 4941 after pre-process finishing. we the separate these into "administration", "courses" and "other" three classifications : table 5 data main classification data administration 1098 courses 729 other 3114 table 6 keywords dictionary main classification keywords administration 行政 (administrative) 足球隊 (football team) 停車場 (parking lot) 學校 (school) 校園 (camp us) 系學會 (dep artment of science) 智慧大師 (m aster of wisdom) courses 課 (class) 課程名稱 (class name) 老師名稱 (teacher) rs (rs) 多益 (toeic) 科系 (dep artment) proceedings of engineering and technology innovation, vol. 7, 2017, pp. 31 36 copyright © taeti 35 table 7 polarity administration courses positive 251 211 negative 321 215 neutral 526 303 4.2. keyword dictionary in the classification keywords shown in table 6, we found that students had certain classification of keywords . for instance, students often use the "足球隊" to describe the attitude of the campus administration. the following keywords which are usually being used: 4.3. data polarity we will separate relatively into positive, negative and neutral, the following is a review of a variet y of data classification number: 5. discussion in fig. 3, studies have investigated the title dcard reviews, we found that 22% of students expressed the article ideas for the administration, 15% of people have questions or ideas for courses in dcard, and the others are expressed other deman ds. fig. 3 data fig. 4 administrative in fig. 4, we also found that students in the comments can really achieve the administration and course, there are 1098 administrative from administration. 251 is the positive title, 321 negative title and 526 neutral title. 52 percent of the students published their perspective, point of view and opinions on the dcard, 29% of whom expressed negative title, which showed that students were less likely to be administratively in school (including school systems). fig. 5 course there are 729 titles in the course, with 211 positive titles, 215 negative titles and 303 neutral titles, which shows that st udents have 58% of the ideas and suggestions in the course or teacher, and the 42% of the people have other problem in course. we can see the results in fig. 5. proceedings of engineering and technology innovation, vol. 7, 2017, pp. 31 36 copyright © taeti 36 6. conclusions this research explores students' perceptions of campus quality, extracts web pages, use keywords and the terminology of classified sentences to achieve the desired results. use the positive and negative sentences to explore the ideas, perspectives, point of views of students on the administration and the curriculum the negative title, it can be extracted separately review and achieve to improvement the purpose campus administration and curriculum. furthermore, future research can be published on dcard and use the text to explore the students for the campus administration ideas and recommendations make research more in-depth can improve the improvement. there are two limitations to this paper, which can be discussed in subsequent studies : 1. research methods can only see the school administration is missing, but cannot see what kind of administrative issues are 2. title cannot tell what kind of problem, is it can only know the advantages and disadvantages references [1] s. w. wu, “applying text mining technique to analyze the software quality characteristics of mobile games, ” in department of industrial management, national pingtung university of science and technology , pp. 1-49, 2013. [2] x. z. chang and y. m. li, “using text mining to predict personality based on social behavior,” 2012. [3] r. s. baker and k. yacef, “the state of educational data mining in 2009: a review and future visions ,” journal of educational data mining, vol. 1, no. 1, pp. 3-17, 2009. [4] y. h. tseng and y. i. lin, “the application of content mining techniques to the analysis of educational evaluation research trends ,” journal of research in education sciences, vol. 56, no. 1, pp. 129-166, 2011. [5] g. van rossum, python, https://zh.wikipedia.org/wiki/python#.e5.8f.82.e8.80.83.e6.96.87.e7.8c.ae, 1991. [6] c. y. jian, “dcard,” https://www.dcard.tw/, 2011. [7] c. m. young, “applies by text mining to the support systems for coding icd-9-cm —a study of admission note and discharge summary,” taipei medical university information management, pp. 1-68, 2004. [8] j. d. kim, t. ohta, y. tateisi, and j. tsujii, “genia corpus —a semantically annotated corpus for bio -textmining,” bioinformatics, vol. 19, pp. i180-i182, 2003. [9] m. hall, e. frank, g. holmes, b. pfahringer, p. reutemann, and i. h. witten, “the weka data mining software: an update,” acm sigkdd explorations newsletter, vol. 11, no. 1, pp. 10-18, 2009. [10] y. h. tseng, “research and development on automatic information organization and subject analysis in recent decades ,” journal of educational media and library sciences , vol. 51, pp. 3-26, 2014. [11] a. ortigosa, j. m. martin, and r. m. carro, “sentiment analysis in facebook and its application to e-learning,” computers in human behavior, vol. 31, pp. 527-541, 2014. microsoft word 6-#934-typesetting finish pdf列印.doc proceedings of engineering and technology innovation, vol. 5, 2017, pp. 31 36 causal relationships among factors affecting the buildability in executing construction projects in vietnam hai pham1,*, soo-yong kim2, truong-van luu3, phu-son lam4 1interdisciplinary program of construction engineering and management, pukyong national university, busan, korea. 2department of civil engineering, pukyong national university, busan, korea. 3department of civil engineering, international university, ho chi minh city, vietnam. 4department of civil engineering, ho chi minh city university of technology, vietnam. received 07 july 2017; received in revised form 10 august 2017; accepted 13 august 2017 abstract this paper aims to identify causal relationships among factors affecting the buildability in executing construction projects in vietnam. first, through a comprehensive literature review, the study identified a full set of attributes that affect the buildability of construction projects. subsequently, through discussions and interviews with experts, the study determined main attributes affecting the buildability in the context of the construction industry in vietnam. after that, a survey questionnaire was developed which was based on identified attributes. the data after collected, was analysed by structural equation model (sem) and the expected result from the sem model is a structural model. this structural model includes four factors, namely: design applies advanced machinery, materials and construction methods; design uses available resources at local for saving costs; design applies safe construction methods; and design uses prefabricated and typical components. the key finding of the study is to provide insights into causal relationships among factors affecting the buildability in executing construction projects in vietnam. keywords: buildability, structural equation model, construction projects, vietnam 1. introduction the construction industry is one of the most developed industries in vietnam in recent years. there are many projects to be completed corresponding to the expectation of investors as well as meeting aesthetic and quality requirements. however, there are also many projects completed with differences to initial designs. particularly, many projects after completed design cannot be constructed or have delays in construction schedule, cost overruns due to revisions of drawings to suit practical requirements. currently, there are many methods of project implementation in vietnam but design-bid-build method is still the most common method. for this traditional method, design and construction companies usually work separately. thus a completed design which cannot match the construction is unavoidable. therefore, to finish projects with the high efficiency in terms of aesthetics, ease for construction as designed, limitation of many revisions, reduction of incurred costs and schedule; design companies need to consider more carefully about the * corresponding author. e-mail address: haiphamcpm@gmail.com tel.: +825-1-6297718 proceedings of engineering and technology innovation, vol. 5, 2017, pp. 31 36 copyright © taeti 32 buildability for each design. to obtain this purpose, this study needs to finish the following two objectives: identify attributes that affect the buildability for the process of project implementation in the context of the construction industry in vietnam; and discover causal relationships among constructs affecting the buildability for projects deployment. 2. research methodology to achieve the study objectives above, firstly a comprehensive literature review is conducted to determine attributes that affect the buildability for projects deployment in the context of the construction industry in vietnam. then, a survey questionnaire is developed to collect data for the analysis of the study. the questionnaire consists of two parts: the first part presents the attributes affecting the buildability, the respondents are asked to evaluate the items based on a five-point likert scale according to convention: 1 (strongly disagree ) to 5 (strongly agree). the second section is designed to collect personal information of the respondents such as years of work experience as well as positions in projects. the next step, a structural equation modelling (sem) is deployed to discover causal relationships that affect the buildability for implementing projects. the sem model in this study includes two models: the measurement model and structural model. the purpose of the measurement model is to evaluate the extent to which observed variables load up their underlying constructs. the purpose of the structural model is to discover causal relationships among constructs affecting the buildability. the fit of both the measurement model and structural model is assessed by several goodness of fit (gof) indices. if the two models achieve the minimum requirements of these indices, they are considered to be fit into the data of the study and accepted. finally, discussion of the result of the sem model is also shown. the next section summarizes the factors influencing the buildability to be identified through a literature review and the implementation of the sem model. 3. structural equation model (sem) 3.1. confirmatory factor analysis (cfa) in testing measurement scale, the cfa method in structural equation model has many advantages rather than traditional methods. the reason is that cfa allows the testing of theory structure of measurement scale such as relationship between a research concept and other concepts without bias due to measurement error. moreover, cfa can also check convergence value as well as discriminant value of measurement scale without taking many research steps as traditional methods [10-13]. the model fit is assessed by many goodness of fit (gof) indices as follows:  chi-square (cmin).  chi-square/df (cmin/df).  comparative fit index (cfi).  tucker & lewis index (tli).  root mean square error approximation (rmsea). the model is considered as suitable to data when p-value (chi-square test) higher than 0.05. however, chi-square is dependent on sample size. if a model gets the values of gfi, tli, and cfi higher than 0.9 [10-13]; cmin/df lower than 2; rmsea lower than 0.08, then the model is considered as suitable to data or compatibility to data. proceedings of engineering and technology innovation, vol. 5, 2017, pp. 31 36 copyright © taeti 33 seven main factors covering attributes for the sem model are denoted as in table 1. these attributes are found out based on previous studies of the authors, such as: [1-9] table 1 factors affecting the buildability attributes factors affecting the buildability denoted design applies advanced materials design applies advanced machinery, materials and construction methods b1 design applies advanced machinery and construction methods design considers many options for projects foundation with complex geological conditions design considers supply conditions to imported construction equipment and materials design considers relocation of temporary prop structures to create more space for construction design uses available resources at local for saving costs design uses available resources at local for saving costs b2 design considers minimization of materials waste design considers convenient traffic conditions for workers in sites design ensures foundation stability for projects being executed and surrounding buildings design applies safe construction methods b3 design considers safe construction methods when working underground ground and shape of floors are not too complex design considers sufficient workspace for workers and machinery during the construction design minimizes complex components minimization of complex components and maximization of standardized components for fast construction b4 design maximizes standardized components for fast construction design minimizes complex components design considers safe construction conditions b5 design maximizes standardized components for fast construction design considers the operation possibility of machinery and equipment on sites work coordination and construction process consideration b6 design considers safe conditions for projects when heavy equipment dropped design of construction process considers creating easy conditions for components installation design uses prefabricated and typical components b7 designers, contractors and project management boards cooperate closely and regularly the fig. 1 presents attributes as well as factors affecting the buildability for projects deployment. fig. 1 the original cfa model proceedings of engineering and technology innovation, vol. 5, 2017, pp. 31 36 copyright © taeti 34 fig. 2 the final cfa model (standardized) the final cfa model (fig. 2) is revised several times by adding errors to observed variables according to modification indices (mi). after that, variables which have standardized weights lower than 0.5 are eliminated. data of the study is evaluated based on goodness of fit (gof) indices: chi-square/df = 1.53 < 2; cfi = 0.95 > 0.9; gfi = 0.93 > 0.9; tli = 0.91 > 0.9; rmsea = 0.064 < 0.08. therefore, the final cfa model fits into the data of the study and accepted. the standardized weights are higher than 0.5 and the unstandardized weights are also statistically significant (p-value = 0.00), so the factors achieve the convergent validity. the correlation coefficients among the factors in the model are lower than 0.9 thereby the scale achieves discriminant validity. 3.2. the structural model from the results obtained in the previous analysis of cfa, combining with the consideration of relationships of the main factors, a structural model (fig. 3) is proposed to describe the mutual influence of the main factors as follows: fig. 3 the structural model among the main factors the result of the structural model among the main factors: chi-square/df = 1.64 < 2; cfi = 0.94 > 0.9; gfi = 0.93 > 0.9; tli = 0.91 > 0.9; rmsea = 0.07 < 0.08. therefore, the structural model among the main factors fits into the data of the study and accepted. from the structural model above, it can be concluded that design applies advanced machinery, materials and construction methods has the positive influence on design uses prefabricated and typical components. design consultants generally use proceedings of engineering and technology innovation, vol. 5, 2017, pp. 31 36 copyright © taeti 35 traditional structural systems (such as reinforced concrete systems poured at sites) without much attention to advanced machinery, materials and construction methods. prefabricated components often have advantages in terms of quality as produced in factories and prestressed components will meet requirements on the aesthetic. the increase of using advanced machinery and construction methods also helps to shorten time for construction of prefabricated components. design enhancing the application of new materials will also help reduce costs for projects such as using light brick to make partition will help to reduce loads thereby reducing the volume of foundation and frame systems. the increase of using lightweight and good quality materials as well as advanced machinery and construction techniques will improve safe construction methods as designing. design using locally available materials to reduce transportation and supply costs will negatively impact on design which considers use of prefabricated components. projects which currently use reinforced concrete structures poured in sites under traditional methods are very popular, thereby local workers are mostly familiar with construction process. in addition, contractors are also easy to find the supply of locally available materials. this contributes to limitation of using prefabricated components and advanced construction methods. 4. bootstrap test to assess the reliability of estimates, in quantitative research methods using sampling method, one sample is usually divided into two subsamples. the first subsample is used to estimate model parameters, and the second is for reassessment. another method is to repeat the study by other samples. two methods above are generally impractical because structured methods often require large samples, so this takes more time and cost [14]. in such cases, bootstrap is a suitable method to replace [15]. bootstrap is a repeated sampling method with replacement. the collected sample consists of 132 observations. the sample for bootstrap test consists of 500 observations. the test result is given in the following table: table 2 error of standardized regression weights parameter se se-se mean bias se-bias cr b2  b1 0.111 0.004 0.663 -0.009 0.005 -1.80 b3  b1 0.132 0.004 0.475 0.008 0.006 1.33 b7  b1 0.472 0.015 1.074 0.016 0.021 0.76 b7  b2 0.462 0.015 -0.597 -0.032 0.021 -1.52 the result in table 2 shows that there is no statistically significant difference in parameter values between the collected sample and the sample for bootstrap test. all cr values are smaller than z = +/1.96 of the normal distribution (no statistical significance at the confidence level 95%). thus, the estimated parameters of the sem model are reliable. 5. conclusions in this study, a theoretical model is developed and tested rigorously using data collected through survey questionnaires covering the factors affecting the buildability for projects deployment in the context of the construction industry in vietnam. the sem method is used to empirically validate this theoretical model to discover causal relationships between constructs. the result of the sem model indicates that design applies advanced machinery, materials and construction methods has the positive influence on design uses prefabricated and typical components. design enhancing the application of new proceedings of engineering and technology innovation, vol. 5, 2017, pp. 31 36 copyright © taeti 36 materials will also help reduce costs for projects such as using light brick to make partition will help to reduce loads thereby reducing the volume of foundation and frame systems. the increase of using lightweight and good quality materials as well as advanced machinery and construction techniques will improve safe construction methods as designing. design using locally available materials to reduce transportation and supply costs will negatively impact on design which considers use of prefabricated components. however, there are also some limitations to this study: the sem model is validated by the data collected from the respondents working in the construction industry in vietnam. for this reason, a recommendation made is that there should be further studies to be carried out to discover causal relationships between constructs affecting the buildability in other regions. consequently, similarities and differences will be summarized and compared to this study. references [1] p. t. i. lam, f. w. h. wong, and a. p. c. chan, “contributions of designers to improving buildability and constructability,” journal of design studies, vol. 27, no. 4, pp. 457-479, july 2006. [2] r. m. zin, m. z. a. majid, and c. fadhil, “neural network model for design constructability assessment,” jurnal teknologi, vol. 40, pp. 27-40, june 2004. [3] code of practice on buildable design, 2013 ed. singapore, building and construction authority, august 2013. [4] a. m. jarkas, “an investigation into the influence of buildability factors on productivity of in situ reinforced concrete construction,” ph.d. thesis, university of dundee, 2005. [5] e. saghatforoush, b. trigunarsyah, and e. g. too, “assessment of operability and maintainability success factors in provision of extended constructability principles,” 9th international congress on civil engineering, may 2012. [6] a. m. jarkas, “the effects of buildability factors on formwork labor productivity of grade beams,” journal of construction engineering, vol. 25, no. 2, pp. 240-265, august 2010. [7] a. m. jarkas, “buildability factors that influence micro-level formwork labour productivity of beams in building floors,” journal of construction in developing countries, vol. 16, no. 1, pp. 1-18, 2011. [8] a. m. jarkas, “the effects of buildability factors on rebar fixing labour productivity of beamless slabs,”australasian journal of construction economics and building, vol. 10, no. 1-2, pp. 16-35, january 2010. [9] a. griffith, “a critical investigation of factors influencing buildability and constructability,” ph.d. thesis, heriot-watt university, uk, 1984. [10] j. f. h. jr, w. c. black, b. j. babin, and r. e. anderson, multivariate data analysis, 7th ed. person prentice hall, nj, 2010. [11] j. b. ullman, “structural equation modeling: reviewing the basics and moving forward,” journal of personality assessment, vol. 87, no. 1, pp. 35-50, august 2006 [12] r. e. schumacker and r. g. lomax, a beginner’s guide to structural equation modeling, 3rd ed., new york, ny, taylor and francis group, 2010. [13] k. schermelleh-engell and h. moosbrugger, “evaluating the fit of structural equation models: tests of significance anddescriptive goodness-of-fit measures,” methods of psychological researchonline, vol. 8, no. 2, pp. 23-74, may 2003. [14] j. c. anderson and d. w. gerbing, “structural equation modeling in practice: a review and recommended two-step approach,” psychological bulletin, vol. 103, no. 3, pp. 411-423, 1988. [15] r. e. schumacker and r. g. lomax, a beginner’s guide to structural equation modeling, 2nd ed., new jersey: lawrence erlbaum associates, 2004. 4___peti#7790__38-44 proceedings of engineering and technology innovation, vol. 19, 2021, pp. 38-44 low profile wideband dual-ring slot antenna for biomedical applications shilpee patil 1,* ,vinod kapse 1 , shruti sharma 1 , anil kumar pandey 2 1 department of electronics and communication engineering, noida institute of engineering and technology, greater noida, utter pradesh, india 2 department of electronics and communication engineering, gl bajaj institute of technology and management, greater noida, utter pradesh, india received 26 may 2021; received in revised form 22 june 2021; accepted 23 june 2021 doi: https://doi.org/10.46604/peti.2021.7790 abstract in this study, a low-profile, co-planar waveguide (cpw) fed, wideband, and dual-ring slot antenna design for biomedical applications is proposed. the proposed antenna has a total area of 10 mm × 10 mm and a height of 0.4 mm, and is designed by using a thin and biocompatible fr4 epoxy (εr = 4.4) substrate to accomplish human body isolation and great flexibility obtained by implantation. this wideband antenna covers a large bandwidth of industrial scientific and medical (ism) frequency band, including 902.8 mhz to 928 mhz, 1.395 ghz to 1.4 ghz, 1.427 ghz to 1.432 ghz, 2.4 ghz to 2.485 ghz, and above. the simulation results of return loss, voltage standing wave ratio (vswr), impedance matching, gain, and radiation pattern of the proposed antenna are obtained through high frequency structure simulator (hfss) 14 software. keywords: biomedical application, dual-ring slot, ism band, implantable antenna 1. introduction the application of antenna is as much as a vast antenna topic itself. some of the applications include automobiles, flush-mounted applications, satellites, cars, aircrafts for biomedical purposes, etc. for biomedical purposes, the antenna is not only used as a device to take and pass body information, but is also used for monitoring, tracking, and for many other purposes. however, for the implantation of an antenna, the size of the antenna is one of the main issues introduced for antenna design. as for introducing an antenna inside the human body, the overall dimensions of the antenna must be tiny and flexible, showing a large bandwidth and a low specific absorption rate (sar) value. many of the research articles depict various issues in the field of antenna design technology to implement a small antenna for biomedical application. in previous years, several multiband slot antennas have been proposed [1-5]. das and mitra [6] present a small, wideband, flexible, and implantable slot antenna design that provides high gain. to improve the antenna gain, their proposed metamaterial (mtm) array has a very high value of epsilon on the superstrate material of an antenna. the gain of that antenna has been enhanced by about 3db using mtm. bashir et al. [7] present a miniaturized wideband implantable antenna for biomedical application. this reported antenna contains the radiating element with rectangular and circular slots and a ground plane with a rectangular slot on it. the maximum gain achieved by this antenna is -12 dbi at 2.45 ghz. the study shows how the parameters such as bandwidth and gain of antenna are varied by positioning the designed antenna in different human body phantoms, e.g., skin, muscles, stomach, small intestine, large intestine, etc. xu et al. [8] present an antenna with an open-end slot on the ground plane, which is dual band in nature and has resonance at 2.45 ghz in the industrial scientific and medical * corresponding author. e-mail address: shilpeepatil21@gmail.com tel.: +919811143811 proceedings of engineering and technology innovation, vol. 19, 2021, pp. 38-44 (ism) band and at 402 mhz in the medical implant communication system (mics) band. the result shows a measured bandwidth of 52.6% at the lower band. for the upper band, the observed resonant frequency shifts to 2.47 ghz and a measured bandwidth of 4.4% is obtained for s11 < -10 db. nesasudha and fairy [9] present a low-profile antenna design for biomedical applications in the frequency range of 2.65 to 3 ghz. this antenna is evident with three distinct layers of substrate materials, i.e., fr4 substrate, bakelite substrate, and roger substrate (ro3010). among the three substrates, the fr4 possesses a return loss of -32.6 db which is the best result for the radiation of the antenna. guo and liu [10] present a single-fed miniaturized circularly polarized microstrip patch antenna. the influence of different body phantoms is presented to calculate the sensitivity of the designed antenna. the simulated value of impedance bandwidth of the reported antenna covering from 2.36 to 2.55 ghz is 7.74%. xiao and li [11] propose an antenna design for an implantable medical device. a miniature semi-circular implantable planar inverted-f antenna (pifa) is designed and proposed in this work. there are three arc-shaped slots that are embedded in a semi-circular patch due to which the proposed antenna achieves reduction in size. chauhan et al. [12] present radio frequency (rf)-linked implantable medical devices. the antenna presented in [13] is implemented to work in the ism band (2.4-2.4835 ghz). the antenna is inserted within a silicon layer and fed by a whole strip line. the result shows the thin dimension of the antenna, which allows the antenna to flex along the curvature of the body without high performance loss, having a higher bandwidth and a good return loss. yang and xiao [14] present a wideband single-fed implantable antenna for the application of ism biomedical at 2.4 ghz. the simulated result of the s11 bandwidth is below -10 db ranging from 2.24 to 2.59 ghz (35%). for the low mode, the maximum of the realized gain -20.8 dbi at boresight is obtained. yang et al. [15] present a design to develop an implantable circularly polarized microstrip patch antenna embedded in a lossy material by addressing its total quality factor. liu et al. [16] design and verify a single-fed miniaturized wide-beamwidth circularly polarized implantable antenna operating in the ism band, for subcutaneous real-time glucose monitoring applications. xia et al. [17] and gani et al. [18] present two antenna systems that operate on multi bands, including the mics band and the ism band. in this study, two triangular rings with equal spacing are designed with the co-planar waveguide (cpw) feeding technique. at the operating frequency of 2.45 ghz, s11 is less than -10 db that shows its proper nature for implantation. the proposed design is a low-profile antenna, which shows a wideband from 0.78 ghz to 5 ghz including 902.8 mhz to 928 mhz, 1.395 ghz to 1.4 ghz, 1.427 ghz to 1.432 ghz, and 2.4 ghz to 2.485 ghz. it covers the ism band and demonstrates a wide range of applications. 2. design configuration the proposed antenna design consists of two concentric triangular ring slots of the equal width d with cpw fed. the design is fabricated on 10 mm × 10 mm fr4 epoxy substrate with a dielectric constant 4.4 and a height of 0.4 mm. therefore, the overall dimension of the antenna is 10 × 10 × 0.4 mm 3 . although a different type of feeding technique is available, in the proposed antenna, the cpw fed with 50 ω matching is used by using appropriate ground plane spacing (s) and trace width (w). fig. 1(a) shows the top view of the designed antenna, where d is the spacing between the triangular rings, s is ground plane spacing, and w is tracing width. their values are presented in table 1. for the design and simulation of the proposed antenna, high frequency structure simulator (hfss) software [19] is used in this study. hfss is a three-dimensional electromagnetic software for designing and simulating high frequency antennas, and is commonly used for antenna design and for the design of complex rf electronic circuits. table 1 design parameter of implantable slot antenna parameter s l d h w a b value (mm) 0.5 10 0.75 0.4 1.05 60 40 39 proceedings of engineering and technology innovation, vol. 19, 2021, pp. 38-44 (a) top view of antenna (b) antenna design in hfss software (c) antenna design in hfss software (continued) fig. 1 antenna configuration 3. antenna performance in biomedical applications, the designed implantable antenna should be situated in a body-like model. for this purpose, skin, fat, muscle, etc., have been used. the proposed antenna is simulated inside the muscle box having a relative permittivity of 52.79 and bulk conductivity of 1.705 as shown in fig. 2. to achieve the radiation characteristics of the proposed antenna, the total height of the muscle box is taken 10 mm whereas the radiation box is taken 20 mm. the value of a and b is given in table 1. fig. 2 antenna placement in muscle box and radiation box 3.1. return loss and input impedance the proposed dual concentric triangular slot antenna is simulated inside the muscle model. the simulated result for return loss characteristics is illustrated in fig. 3. fig. 3 s11 parameter of antenna 40 proceedings of engineering and technology innovation, vol. 19, 2021, pp. 38-44 the s11 characteristics of the proposed antenna represent the amount of power reflected from the antenna, so the s11 parameter is known as the reflection coefficient or return loss of the antenna. therefore, according to power radiation, if s11 is 0 db, it means 100% reflection. in other words, if no power is into the antenna, all power gets reflected. on the other hand, if s11 is less than -10 db, it means 90% of available power is already delivered to the antenna and 10% gets reflected back which is the best result for an antenna. fig. 3 shows the values from 0.78 ghz to 5 ghz under -10 db, representing that at least 90% input power is delivered to the device and the reflected power is less than 10%, which is sufficient for many applications. the simulated s11 results show that the proposed antenna covers many ism band frequencies which include 902.8 mhz to 928 mhz, 1.395 ghz to 1.4 ghz, 1.427 ghz to 1.432 ghz, 2.4 ghz to 2.485 ghz, and above. it is useful for many wireless applications. due to the radiating triangular slot patch at the top of the antenna, the ground is assumed like an effective magnetic current. the image current with respect to the ground assumed as a perfect electric conductor has the same magnitude and phase. fig. 4 illustrates the input impedance characteristics for the proposed antenna at the desired band. fig. 4 input impedance characteristics of antenna 3.2. voltage standing wave ratio the voltage standing wave ratio (vswr) describes the amount of mismatch between the proposed antenna and its feed line connected to it. the value of vswr is in the range of 1 to infinity. vswr is 1 when return loss is infinity. fig. 5 illustrates the value of vswr between 2.5 db to 5 db of the proposed antenna. fig. 5 voltage standing wave ratio 3.3. radiation pattern the three-dimensional radiation pattern describes the strength of radiation emitted in a particular direction as well as the gain of an antenna in terms of electric and magnetic fields. the distance of radiations from the origin represents the strength of radiation emitted in that direction. fig. 6 shows that the power transmitted along the z axis is maximum. 41 proceedings of engineering and technology innovation, vol. 19, 2021, pp. 38-44 fig. 6 3d radiation pattern of the proposed antenna e-plane and h-plane are basically the principal planes for the propagation of waves or radiation fields, where the e-plane shows the plane containing the electric field and h-plane shows the plane containing a magnetic field. co-polarization means that the polarization of the transmitting and receiving antenna is the same, and cross polarization means that the polarization of the transmitting and receiving antenna is different. the proposed antenna has been simulated at 0.915 ghz and 2.45 ghz. fig. 7(a) represents the x-z plane graph and fig. 7(b) represents the y-z plane graph at 0.915 ghz. similarly, fig. 7(c) represents the x-z plane graph and fig. 7(d) represents the y-z plane graph at 2.45 ghz. (a) the x-z plane at 0.915 ghz (b) the y-z plane at 0.915 ghz (c) the x-z plane at 2.45ghz (d) the y-z plane at 2.45ghz fig. 7 radiation pattern of the antenna 3.4. the simulated surface current distribution to accomplish an effective operation of the proposed antenna, it is required to cover most of the impedance bandwidth by axial ratio bandwidth of the antenna. to introduce the current distribution perturbation on the surface of the proposed antenna, dual triangular shaped slots are cut into the ground with cpw feed line. fig. 8 illustrates the simulated surface current distribution at 2.45 ghz for time phase 0˚ and 90˚. as a result, an improved current distribution is obtained within the operating band. 42 proceedings of engineering and technology innovation, vol. 19, 2021, pp. 38-44 (a) phase angle 0˚ (b) phase angle 90˚ fig. 8 the simulated surface current distribution at 2.45 ghz 3.5. antenna gain the simulated value of gain for the proposed antenna with respect to its desired frequency is plotted in fig. 9. the frequency is in the range of 0.78 ghz to 5ghz. the maximum simulated gain is observed -17.0411 db at 0.92 ghz and -14.572 db at 4.02 ghz. the noticeable radiation characteristics with acceptable gains for desired bands and a wideband including ism band are achieved with a small size compared to other implantable antennas. this proposed antenna is compared with different designs of the implantable antennas in table 2. fig. 9 gain vs. frequency table 2 performance comparison with different implantable antennas ref. frequency (ghz) bandwidth (mhz) gain (dbi) dielectric material antenna shape volume (mm 3 ) [7] 2.45 483 -12 rogers ultralam circular 9.8 [15] 2.45 73 -20.47 rogers rt/duroid 6010 zig zag 21 [16] 2.45 300 -17 rogers 3210 circular 91.75 [17] 2.45 246 -21.2 rogers rt/duroid 6010 spiral 31.5 [18] 2.45 105 -22.3 rogers rt/duroid 6010 zig zag 52.5 the proposed work 2.45 4220 -14.6 fr4 epoxy triangular 40 4. conclusions the proposed work presents a cpw fed slot antenna and its design procedure, which is suitable for biomedical applications at a very low cost. this antenna is characterized by its small size, wideband, and single feeding property. the proposed antenna has appreciable gain and good radiation pattern characteristics over a large impedance bandwidth which 43 proceedings of engineering and technology innovation, vol. 19, 2021, pp. 38-44 covers a wide range from 0.8 ghz to 5 ghz and above. this antenna is for biomedical purpose as it covers the ism band range, including 902.8 mhz to 928 mhz, 1.395 ghz to 1.4 ghz, 1.427 ghz to 1.432 ghz, and 2.4 ghz to 2.485 ghz. the maximum gain of the antenna is -17.0411db at 0.92 ghz and -14.572db at 4.02 ghz. conflicts of interest the authors declare no conflict of interest. references [1] j. pourahmadazar, c. h. ghobadi, j. nourinia, n. felegari, and h. shirzad, “broadband cpw-fed circularly polarized square slot antenna with inverted l-strip for uwb applications,” ieee antennas wireless propagation letters, vol. 10, pp. 369-372, april 2011. [2] b. k. kanaujia, s. k. gupta, j. kishor, and d. gangwar, printed antennas: theory and design, boca raton: crc press, 2020. [3] s. patil, a. k. singh, b. k. kanaujia, and r. l. yadava, “a compact, dual wide-band circularly polarized, modified square ring slot antenna for c and ku band applications,” international journal of microwave and wireless technologies, vol. 11, no. 2, pp. 182-189, october 2019. [4] s. patil, a. k. pandey, and v. k. pandey, “a compact, wideband, dual polarized cpw-fed asymmetric slot antenna for wireless systems,” journal of microwaves, optoelectronics, and electromagnetic applications, vol. 19, no. 3, pp. 343-355, september 2020. [5] s. patil, a. k. singh, b. k. kanaujia, and r. l. yadava, “a low-profile triple-band circularly polarized wide slot antenna for wireless systems,” international journal of microwave and wireless technologies, vol. 11, no. 1, pp. 67-75, august 2019. [6] s. das and d. mitra, “a compact wideband flexible implantable slot antenna design with enhanced gain,” ieee transactions on antennas and propagation, vol. 66, no. 8, pp. 4309-4314, may 2018. [7] z. bashir, m. zahid, n. abbas, m. yousaf, s. shoaib, m. a. asghar, et al., “a miniaturized wide band implantable antenna for biomedical application,” uk/china emerging technologies, october 2019, pp. 1-4. [8] l. j. xu, y. x. guo, and w. wu, “dual-band implantable antenna with open-end slots on ground,” ieee antennas and wireless propagation letters, vol. 11, pp. 1564-1567, 2012. [9] m. nesasudha and j. j. fairy, “low profile antenna design for biomedical applications,” international conference on signal processing and communication, july 2017, pp. 139-142. [10] y. guo and c. liu, “miniaturized implantable circularly polarized patch antenna,” ieee international workshop on electromagnetics, august 2014, pp. 271-272. [11] s. q. xiao and r. q. li, “antennas design for implantable medical devices,” ieee international conference on computational electromagnetics, february 2015, pp. 61-63. [12] a. chauhan, g. chauhan, and g. kaur, “implantable antennas in biomedical applications,” international conference on computational intelligence and communication networks, december 2015, pp. 1-5. [13] t. a. aleef and a. biswas, “design and measurement of a flexible implantable stripline-fed slot antenna for biomedical applications,” international conference on electrical engineering and information communication technology, september 2016, pp. 1-5. [14] z. j. yang and s. xiao, “a wideband implantable antenna for 2.4 ghz ism band biomedical application,” international workshop on antenna technology, march 2018, pp. 1-3. [15] z. j. yang, l. zhu, and s. xiao, “an implantable circularly polarized patch antenna design for pacemaker monitoring based on quality factor analysis,” ieee transactions on antennas and propagation, vol. 66, no. 10, pp. 5180-5192, august 2018. [16] x. y. liu, z. t. wu, y. fan, and e. m. tentzeris, “a miniaturized csrr loaded wide-beamwidth circularly polarized implantable antenna for subcutaneous real-time glucose monitoring,” ieee antennas and wireless propagation letters, vol. 16, pp. 577-580, july 2016. [17] z. xia and h. li, “miniaturized dual-band implantable antenna,” ieee international conference on microwave and millimeter wave technology, may 2018, pp. 1-3. [18] i. gani and h. yoo, “multi-band antenna system for skin implant,” ieee microwave and wireless components letters, vol. 26, no. 4, pp. 294-296, march 2016. [19] “ansys hfss,” https://www.ansys.com/en-in/products/electronics/ansys-hfss, 2019. copyright© by the authors. licensee taeti, taiwan. this article is an open access article distributed under the terms and conditions of the creative commons attribution (cc by-nc) license (https://creativecommons.org/licenses/by-nc/4.0/). 44 microsoft word 3-peti#5291 23-29.docx proceedings of engineering and technology innovation, vol. 16, 2020, pp. 23 29 development and characterization of thermal insulation geopolymer foams based on fly ash michał łach 1,* , janusz mikuła 1 , wei-ting lin 2 , patrycja bazan 1 , beata figiela 1 , kinga korniejenko 1 1 institute of materials engineering, faculty of materials engineering and physics, cracow university of technology, cracow, poland 2 department of civil engineering, national ilan university, yilan, taiwan received 18 february 2020; received in revised form 30 april 2020; accepted 04 june 2020 doi: https://doi.org/10.46604/peti.2020.5291 abstract the main purpose of the article is to present the differences in the parameters of geopolymer foams obtained in the same way, which is associated with difficulties in controlling the foaming process. difficulties in controlling the foaming process of geopolymers are the direct reason for the lack of implementation of such materials nowadays. the article shows the results for experimental research, especially research into insulation, physical, and mechanical properties for the designated foamed materials. microspheres (75%), sand (5%) and fly ash (20%) were used to produce foamed geopolymers. hydrogen peroxide was the foaming agent. heat conduction coefficients of 0.08-0.07 w/mk were obtained. the material density was obtained at the level of 363-375 [kg/m 3 ] and the compressive strength was 520-683 [kpa]. the results showed that geopolymers can be a good alternative to conventional insulation materials, but the foaming technology should be developed so that it is stable and allows for reproducible material parameters. keywords: geopolymer, thermal insulation, mechanical properties, inorganic foams 1. introduction for many years, scientists have been developing insulation materials that have the best durability, non-flammability, and a relatively low price. one of such materials is foamed geopolymer. however, in order to make a foamed geopolymer be a real alternative to replace the currently used insulation materials, it must be met the relevant strength standards and passed the non-flammability test. different values of thermal conductivity and strength of foamed geopolymers are given in scientific paper. nevertheless, they usually differ from the thermal conductivity of popular materials such as polystyrene [1-2]. the best values in terms of thermal conductivity that scientific studies have shown so far are those in studies by liefke [3]. this is a material called trolit, which has a thermal conductivity of 0.037 w/mk. the selected parameters of this material are presented in table 1. nonetheless, since the material was first described [3-4], such a good parameter for geopolymer foams has not been noted again. scientists have tested a lot of compositions by using different materials and are trying to create the optimal material. a good quality of geopolymer insulating material which has density of 313 kg/m 3 , a compressive strength of 0.63 mpa, and a thermal conductivity 0.07 w/mk was obtained, for example, from a mixture of metakaolin and biomass ashes [5]. usually various types of additives are introduced such as diatomite [6]. the replacement of the diatomite in the formulation affects the morphological structure of the foams and then influences their mechanical properties due to a different type of * corresponding author. e-mail address: michal.lach@pk.edu.pl proceedings of engineering and technology innovation, vol. 16, 2020, pp. 23 29 24 chemical bond between sodium silicate and the solid phase. in particular, the mechanism of consolidation of hybrid foams based on the diatomite has changed from geopolymerisation to silicate polycondensation. therefore, the strength parameters improved with the increase of diatomite content [6]. metakaolin-based geopolymers foamed with h2o2, which the amounts from 0.25 wt.% up to 1.25 wt.%, were characterised by a compressive strength from 6 mpa (for 0.25% h2o2) to 0.36 mpa (for 1.25% h2o2). thermal conductivity was from 0.298 w/mk (for 0.25 wt.% h2o2) up to 0.172 w/mk (for 1.25 wt.%h2o2) [7]. to obtain the optimal parameters of the geopolymer foams, the metakaolin and fly ash are often used together. samson and cyr [8] proved that the most optimal composition for their type of raw materials was 62.5% metakaolin with 12.5% fly ash and 25% activator. the composition was expanded using h2o2 in amounts of 1, 1.5, and 2 wt.% and densities from 225 to 506 kg/m 3 and thermal conductivities from 0.07 to 0.12 w/mk were obtained. these studies have also shown that the use of surfactants has a huge impact on the porous structure [8]. it is also possible to use rice husks as a foaming agent [9]. additives to geopolymers are also very often introduced to improve strength properties. for instance, strength parameters of geopolymers can be improved by adding the lime and the gypsum [10]. fly ash and metakaolin beyond their use for not only use in the production of geopolymers, but also used in other technologies, for example, they can be used as a filling material in mineral composites [11]. table 1 physical characteristics of trolit™ foams [1, 2] bulk density [kg/m 3 ] 200 800 max. temperature of application [ o c] 1000 max. thermostability [ o c] 1200 thermal conductivity [w/mk] ≥0.037 (depending on density) pore parameter [mm] 0.5-3.0 fire protection [din 4102] inflammable, classification a1 compressive strength [n/mm 2 ] 0.5-2.0 tensile strength [n/mm 2 ] ≈250 this study presents the results obtained from foamed geopolymer materials based on fly ash. the results of compressive and bending strength tests depending on the density of the obtained materials are to be presented. in this study, the authors want to draw the problems when foaming geopolymers occur and show the fact that despite of attempting to control foaming conditions, the process will be different each time. the material described in this research differs significantly from that described in the literature because it contains up to 75% of microspheres. only 20% by the weight of fly ash was subjected to the foaming process. this material which is the combination of the open porosity obtained from foaming geopolymer and closed porosity obtained by microspheres is produced. 2. materials and methods fly ash from the skawina chp production plant was used as a precursor. the oxide composition of the precursor, determined by the xrf (x-ray fluorescence), is shown in table 2. table 2 oxide content of the fly ash from the skawina chp plant oxide sio2 al2o3 fe2o3 cao mgo na2o k2o so3 tio2 p2o5 bao content [%] 55.89 23.49 5.92 2.72 2.61 0.59 3.55 0.16 1.09 0.82 0.20 the geopolymers were produced by using technical sodium hydroxide flakes and an aqueous sodium silicate solution of r-145 module 2.5, with a molar density of about 1.45 g/cm 3 . the mixture did not use distilled water. the alkaline solution was prepared, for instance, the solid sodium hydroxide was poured over an aqueous solution of sodium silicate and water. the solution was thoroughly mixed and allowed to equilibrate until a constant concentration and temperature were achieved. the geopolymer foams were prepared in such a way that solid constituents, such as fly ash, microspheres, and sand were mixed until reaching a homogeneous state, and afterwards an alkaline solution -10m sodium hydroxide solution with the addition of sodium silicate in the ratio 1:2.5 was introduced. the material was mixed for 10 minutes. an appropriate amount of proceedings of engineering and technology innovation, vol. 16, 2020, pp. 23 29 25 hydrogen peroxide was then introduced and continuously mixed for approx. 1 minute. after the mixing process was complete, the material was poured into a mould and exposed at a temperature of 75°c for a period of 24 hours. subsequently, it was the heating process; the element was dismantled and stored under laboratory conditions for a period of 28 days in order to obtain full strength. examples of the foamed geopolymer materials are presented in figs. 1 and 2. fig. 1 an example of foamed geopolymer panels fig. 2 an example of foamed geopolymer panels the thermal conductivity coefficient was calculated in accordance with pn-en 12667:2002. the measurements were carried out under the steady-state conditions of heat transfer at an average sample temperature amounting to 10°c, for the samples which dimensions were 300×300×100 mm. the applied temperature difference was 20°c for the thickness of the sample (t1=0°c, t2=20°c, tmean=10°c). the tests were conducted in a closed test chamber (having thermally-insulated walls) with the one-sample plate apparatus hfm 436/3/0, netzsch, provided with a heat-flow density sensor. the results were compared with the certified reference values irmm-440 no 30. the bulk density of the product was calculated as a mass ratio per unit volume of the sample, without reference to the certified specimen. the flexural strength test was conducted in accordance with en 12089 on samples of dimensions 300×50×150 mm. compressive strength tests were carried out according to en 826 on samples of dimensions 50×50×50 mm. 3. results the tests were implemented for one and the same composition, whose composition and heat transfer coefficients are given in table 3. however, three different variants of geopolymer panels were obtained due to the different courses of the foaming process. five different samples and measurements were made for each variant. for the obtained results, the standard deviation was calculated and the values of which were quite huge and exceeded several percent because of the heterogeneous porosity of the material. macroscopic and visual observations showed that the distribution of pores in the samples is uneven. the authors of the study often face the problem of uneven distribution of pores and their different sizes. the process of foaming geopolymers is difficult to control and attention. moreover, it should be paid to stabilise the produced foams. these problems cause a large dispersion in the obtained test results, despite the same process and composition, different results are obtained. table 3 composition of foamed geopolymers composition amount h2o2 thermal conductivity [w/mk] geofoam1 75% microspheres +5% sand +20% flyash 3% wt. 0.08094 geofoam2 75% microspheres +5% sand +20% flyash 3% wt. 0.07430 geofoam3 75% microspheres +5% sand +20% flyash 3% wt. 0.07492 table 4 test results of mechanical properties of obtained geopolymer foams density [kg/m 3 ] flexural strength [kpa] compressive strength [kpa] geofoam1 375.8±5.9 447.61±85.6 683.3±28.87 geofoam2 363.8±7.3 362.91±72.2 520.0±34.64 geofoam3 364.0±18.0 369.38±74.8 526.7±80.8 table 4 and fig. 3 show the results of testing the strength properties of the foamed geopolymer panels. the consequences of tests on the flexural strength and compressive strength are presented, depending on the density of the geopolymer material. the highest value of flexural strength is 447 kpa and the compressive strength is 683 kpa. proceedings of engineering and technology innovation, vol. 16, 2020, pp. 23 29 26 fig. 3 graphical representation of the density and strength of the materials it is the results of the sem / eds scanning microscope below showing the morphology for one of the samples to illustrate the heterogeneity of the material (figs. 4 and 5). point chemical composition analyses and visual observations show that not all the fly ash particles have been expanded. their composition indicates a smaller amount of na and an increased amount of ca (point 4). the observed structure is heterogeneous. fig. 4 microstructure of foamed geopolymer (a) point 1 (b) point 2 fig. 5 chemical point analysis 0 100 200 300 400 500 600 700 800 geofoam1 geofoam2 geofoam3 [k g /m 3 ]; [k p a ] flexural strength[kpa] compressive strength [kpa] density [kg/m3] proceedings of engineering and technology innovation, vol. 16, 2020, pp. 23 29 27 (c) point 3 (d) point 4 (e) point 5 fig. 5 chemical point analysis (continued) there are also variables at content points; for example, the fe and si/al ratios are variable. obviously, all these parameters affect the quality of the geopolymer foams obtained. in the drawing, points 3-5 show spherical particles, probably microspheres used to make geopolymers. they are characterized by a much higher content of al than the foamed structure revealed in points 1 and 2. thanks to the introduction of hollow spherical particles in the form of microspheres, it was possible to achieve both open and closed pore structure. the researchers would like to pay special attention to the heterogeneity of the obtained materials. the heterogeneity comes from the difficulties in controlling the foaming process and problems with the stabilization of the produced foams. 4. summary figs. 1 and 2 show the geopolymer materials characterized by uneven pore size distribution. it’s more distinctly visible in fig. 2 due to the low stability of the pores produced and their tendency to join. foamed structures disintegrate very quickly and proceedings of engineering and technology innovation, vol. 16, 2020, pp. 23 29 28 create larger pores coming to the surface. this is one of the reasons for the great difficulty in foaming geopolymer materials. scientists are attempting to incorporate various types of stabilizers into geopolymer foams. the uneven distribution of pores is also found in the photos from the scanning electron microscope in fig. 4. the geopolymer foams have similar density values in the range of 363-375 kg/m 3 . compressive and bending strengths vary for individual samples but to a small extent. similarly, the thermal conductivity coefficient differs for individual tests. despite the fact that the materials were made in the same way and their composition is identical, different parameters of the finished products were obtained. it is proved that the process of foaming geopolymers is very difficult to control. the lack of the stability of the obtained foams can lead to different parameters such as the heat transfer coefficient. even small differences may have higher thermal conductivity coefficients in this case. it should also be emphasized that in the presented research only 20% of fly ash was used to compose the geopolymer mixture and only these 20% of the components were subjected to the foaming process. if more fly ash and fewer microspheres were used, which additionally stabilize the produced foams, the differences in the results obtained would be greater. difficulties in controlling the foaming process, foam stability, and uneven pore size distribution are one of the main reasons for the lack of commercial implementations of this type of material. all research achievements in this area, new experiences are the most beneficial and one should strive to overcome these problems. the parameters of currently obtained foamed geopolymers are different, unfortunately, from the developed materials based on resins such as irmm-440. these materials are characterized by a very low density of 66-74 kg/m 3 and a heat transfer coefficient of 0.029 w/mk. when designing the geopolymers, the researchers should seek to obtain such the attractive parameters. the obtained thermal conductivity values of foamed geopolymers are better than when using the microwave method of foam production. when using microwaves to produce foamed geopolymers, it was possible to obtain a minimum thermal conductivity 0.2387w· m-1· k-1 and higher density values [12]. 5. conclusions foam geopolymer materials based on fly ash are interesting alternatives to currently popular materials such as polystyrene, mineral wool, and glass. they are a non-flammable material characterized by relatively good insulation parameters. the tests showed that the geopolymer foams, at a density of 375 kg/m3, have a compressive strength of about 0.68 mpa and the bending strength of 0.45 mpa. as the density decreases, the strength properties decrease. it is expected that this material will soon be implemented on an industrial scale as an alternative for the currently used insulation materials in construction. preparation for implementation, however, requires further research related to the stability and repeatability of the obtained geopolymer foams. it should also be striven to make it possible to use various waste materials for the production of foamed inorganic polymers. acknowledgements this work was supported by the polish national centre for research and development under the lider grant no: lider/31/0168/l-10/18/ncbr/2019. conflicts of interest the authors declare no conflict of interest. proceedings of engineering and technology innovation, vol. 16, 2020, pp. 23 29 29 references [1] m. łach, k. korniejenko, and j. mikuła, “thermal insulation and thermally resistant materials made of geopolymer foams,” procedia engineering, vol. 151, pp. 410-416, december 2016. [2] m. łach, d. mierzwiński, k. korniejenko, and j. mikuła, “geopolymer foam as a passive fire protection,” matec web of conferences, vol. 247, pp. 1-6, december 2018. [3] e. liefke, “industrial applications of foamed inorganic polymers,” geopolymere’99 proceedings, : institut géopolymère, december 1999, pp. 189-200. [4] j. davidovits, geopolymer chemistry and application, 4th ed. saint-quentin: institut géopolymère, 2015. [5] a. n. murri, v. medri, e. papa, l. laghi, c. mingazzini, and e. landi, “porous geopolymer insulating core from a metakaolin/biomass ash composite,” environments, vol. 4, no. 4, pp. 1-13, december 2017. [6] b. liguori, i. capasso, v. romeo, m. d auria, m. lavorgna, d. caputo, s. iannace, et al., “hybrid geopolymeric foams with diatomite addition: effect on chemico-physical properties,” journal of cellular plastics, vol. 53, no. 5, pp. 525-536, february 2017. [7] n. a. jaya, y. m. liew, m. m. a. b. abdullah, and c. y. heah, “porous metakaolin geopolymers with tailored thermal conductivity,” iop conference series: materials science and engineering, vol. 551, no.1, pp. 1-5, august 2019. [8] g. samson and m. cyr, “porous structure optimisation of flash-calcined metakaolin/fly ash geopolymer foam concrete,” european journal of environmental and civil engineering, vol. 22, no. 12, pp. 1482-1498, april 2017. [9] h. i. riyap, c. n. bewa, c. banenzoué, h. k. tchakouté, c. h. rüscher, e. kamseu, et al., “microstructure and mechanical, physical and structural properties of sustainable lightweight metakaolin-based geopolymer cements and mortars employing rice husk,” journal of asian ceramic societies, vol. 7, no. 2, pp. 199-212, april 2019. [10] v. sahu, a. srivastava, and v. gayathri, “effect of lime and gypsum on engineering properties of badarpur fly ash,” international journal of engineering and technology innovation, vol. 6, no 4, pp. 294-304, september 2016. [11] r. v. upadhyaya and t. g. suntharavadivel, “optimization of fly ash and metakaolin content in mineral based cfrp retrofit for improved sustainability,” international journal of engineering and technology innovation, vol. 9, no. 3, pp. 171-181, may 2019. [12] m. nadeem, e. ulhaq, f. ahmed, m. asif rafiq, g. hameed awan, and m. zain-ul-abdein, “effect of microwave curing on the construction properties of natural soil based geopolymer foam,” construction and building materials, vol. 230, pp. 117074, january 2020. copyright© by the authors. licensee taeti, taiwan. this article is an open access article distributed under the terms and conditions of the creative commons attribution (cc by-nc) license (https://creativecommons.org/licenses/by-nc/4.0/).  proceedings of engineering and technology innovat ion, vol. 6, 2017, pp. 17 22 the development of a new paper pallet machine chi-wei chen * department of marketing and distribution management, tajen university, pingtung, taiwan, roc. received 13 july 2017; received in revised form 15 july 2017, accept ed 19 july 2017 abstract in modern industry world, logistics plays an import role of enhancing the effic iency of shipping goods by use of pallets. the shipping pallets used in modern industry sites are generally made o f meta l, plastic, wood, or paper. among them, the wood pallet is used most often by many firms bec ause of low manufacturing cost. however, a wood pallet is not the best shipping selection by use in the area of air cargo transportation because of its weight and fro m the perspective of environ mental preservation. thus, for a ir cargo transportation, paper pallet is a better shipping selection because of its light weight advantage and also a good shipping tool for upgrading the competitiveness of taiwan’s electronic industry by reducing its air cargo shipping cost. in current paper pallet manufacturing mach ine requires many workers to work with and also generates many wasted papers, which is not a good way fro m the perspective of reducing cost and environmental preservation. a new manufacturing approach was thus proposed and a new machine was also designed to manufacture the newly designed paper pallet. the new paper pallet design was successfully imp le mented and the new prototype machine was also successfully developed. at least 20% of paper was saved comparing to the same size of paper pa llet made by the cu rrent best existing paper pallet machine. and the workers used in new manufacturing process of making paper pallet were a lso reduced at least 10 workers comparing to the process of current existing paper pallet manufacturing machine. keywor ds: paper pallet design, paper pallet machine, air cargo transportation , supply chain management, logistics management 1. introduction during the product manufacturing process and shipping process of finished goods, logistic s area was not being paid enough attention by many industries. thus, in 1962, peter drucke r ca lled the area of logistic s is “the economy’s dark continent” [1-2]. the ma jor reason is that many manufacturing firms are always seeking or ma king new machines, or develop new manufacturing processes to increase their productivities. in recent 20 years, there are more manufacturing firms realize to increase the efficiency of supply chain operation is also an import mean to enhance overall productivities. thus, this paper is intended to propose a new paper pallet design and a new paper pallet machine to enhance the efficiency of logistics operation . traditional ma rine shipping cargo is in the size of 20 or 40-foot long(x8-foot width x 8-foot height) container and usually use pallets with standard specification of 80x120c m or 100x120c m(euro spec.), 110x110c m(jis spec.), 110x110co m or 100x120c m(cns -ta iwan spec.) . in 2007, singapore, shanghai, hong kong ,shen zhen, yingkou(liaonian, china), kaoshiung are in the top 10 lis t of ma rine cargo import and e xport amount (20-foot equivalent unit, tue) , which was shown in table 1 and the international airport air cargo shipping amount was shown in table 2[3]. *corresponding author. e-mail address: cthomas@tajen.edu.tw tel: +886-8-7624002 ext.3213 proceedings of engineering and technology innovation, vol. 6, 2017, pp. 17 22 copyright © taeti 18 table 1 2007 world harbor ranking by import/export cargo cargo import/export amount (tues) ranking harbor name country tues 1 singapore singapore 27,935,500 2 shanghai china 26,152,400 3 hong kong china 23,998,440 4 shenzhen china 21,103,800 5 yingkou (liaonian, china) 13,713,000 6 busan south korea 13,254,703 7 rotterdam netherlands 10,790,604 8 dubai uae 10,653,026 9 kaoshiung taiwan 10,256,829 10 hamburg germany 9,917,180 in the case of marine shipping cargo, each tue needs at least about 10 standard pallets to support unit loads of products. therefore, huge a mount of pallets are needed in this globalizat ion modern world. also huge amount of pallets are made of wood and some special goods requested plastic pallets. and some heavy products need metal pa llets to carry goods. wood pallets are most often used by manufacturing firms because of its low cost. however, if goods are manufactured in ta iwan or china and are needed to be shipped to developed countries, such as usa, uk, or most european countries, wood pallets are needed to add up smoking treat ment and discard cost(usually burning cost), which is not a cost effect method to handle overse a shipping. the most difficult tas k for ta iwan or ch ina manufacturing firms to reduce the shipping cost is that an effic ient mach ine is not existed in modern wo rld wide. paper pa llets are always only one -third of weight a mount of the same size wood pallets. in the area of air ca rgo shipping, paper pallets have large advantages of reducing shipping cost. and about 75% of wood pallets could be replaced by paper pallets. regarding a ir cargo transportation, as in table 2, ta iwan and ch ina occupied top list of international a ir cargo ran king and indeed both taiwan and ch ina have many e lectronic products are needed to be shipped to countries worldwide. thus paper pallets are in big he lp of reducing shipping cost by us ing its light weight advantage and also in the perspective of environmental preservations. table 2 2008 international airport ranking by import/export cargo cargo import/export amount (tons) ranking international airport country tons 1 memphis international airport usa 3,695,438 2 hong kong international airport china 3,660,901 3 shanghai pudong international airport china 2,602,916 4 incheon international airport south korea 2,423,717 10 singapore changi airport singapore 1,883,894 15 taiwan taoyuan international airport taiwan 1,493,120 the proposed newly designed paper pallet mach ine is going to generate a new environ mental protection business with a manufacturing process without waste water and waste air produced. its manufacturing material could be fro m recycled papers and its scraps could be recycled and shipped to paper manufacturing factories to be used as raw materials. 2. existing paper pallet machine and paper pallet designs as mentioned above, both taiwan and china have la rge a mount of electronic products need to be shipped by a ir cargo to many fore ign countries worldwide. thus, the demand of paper pallets is numerous. co mparing to the end price between wood and paper pallets, wood pallets will not have much more advantages than paper pallets. ho wever, in the case of air cargo s hipping, paper pallets will be a better selection than wood pallets. in 2009, eu ropean union was intended to force other countries sellers to use paper pallets to ship goods to enter european un ion countries. this dec ision was hold back, because there is no efficient paper pallet machine available in recent industry markets. proceedings of engineering and technology innovation, vol. 6, 2017, pp. 17 22 copyright © taeti 19 in taiwan, there are about 30 sma ll co mpanies in the business as paper pallets suppliers. the manufacturing method used is all in a very old fashion way, pure ly made by hu man hands and simple tools, and without any mechanica l design s and calculations. the situation in china is that some manufacturing firms are using a paper pallet machine designed by netherland , which required many (at lease 20) worke rs to finish a complete paper pallet. the paper pallet machine developed by netherland was not designed in a highly automat ion way because during its manufacturing process that it still needs many human workers to perform tasks for work -in-p rocess. the manufacturing process es, including stage 1 and stage 2 shown in fig. 1 and fig. 2 respectively, are as follows: stage 1. manufacturing the core material of a bee-hives type paper pallet: fig. 1 the flow chart of making core materials of a bee-hives type paper pallet stage 2. forming a paper pallet fig. 2 the flow chart of making a complete bee-hives type paper pallet this type of paper pa llet is strong enough to support a unit load o f e lectronic products whose average we ight is usually less than 500kgs. usually for the bee-hives type paper pallets, static carrying strength could go up to more than 1000kgs, while dynamic carrying we ight usually less than 1000kgs because of the strength limits of the be -hives type core materia l. fro m cost management point of view, the bee-hives type paper pallet consumes too much human powers. 3. a proposed new paper pallet design and a new designed paper pallet m achine 3.1. a new paper pallet design the proposed paper pallet design has about 20 years of innovation patent protection. the main design of paper pallet ma in board differs fro m that of bee-hives type paper pallet ma in board is the core material inside the main board. the core material of the proposed paper pallet has wave type structure and during the manufacturing process, the core materia l does not need human powers to perform any cutting procedure and thus will not generate any scraps. the core mate ria l of e xisting bee-h ives type paper pallets need human powers (at least 2 persons) to perform stretching and cutting procedure before putting core materia l inside industry papers for succeeding wrapping procedure. the major drawback of the previous process is the time needed for bee-hives type core materia l after pasting glue, heating and drying glues used for adhere outside industry papers and core materia l. the density of glue and the temperature of heating box are key factors affecting the manufacturing efficiency. proceedings of engineering and technology innovation, vol. 6, 2017, pp. 17 22 copyright © taeti 20 the proposed wave type structure also faced the challenge of density of glue used and the heating temperature adjustment, which we re overcome eventually. the majo r challenges of wave type structure core materia l could be separated into two parts. first part is during shaping the wave type structure procedure. it will need very strong forces to transform a flat carton paper to wave type structure, while the wave height is the key factor of affecting final strength of paper pallets. second part is to place a wave type structure core material inside a pre -mold ing carton flat paper and to perform a wrapping procedure. as mention above, in this stage, the glue density and heating temperature setting are very important factors to affec t the strength of final paper pallets. this paper ma inly concentrates efforts on the main board design. the manufacturing of legs used for the proposed paper pallet ma in board is a minor proble m co mparing to the consideration for designing the paper palle t machine. the legs could be manufactured by special designed small molds. the samples of the proposed new paper pallet design are shown in fig 3 as follows: fig. 3 the samples of proposed paper pallets 3.2. the proposed new paper pallet machine in order to produce the paper pallets mentioned above, new manufacturing procedures are developed and are shown in fig. 4 as follows . a prototype machine was also developed to perform the following procedures : fig. 4 the proposed new paper pallet machine working procedures as shown in fig. 4, whole manufacturing procedures without any cutting process, therefore, it will not generate any scraps. the ma jor variab le cost for the ma in boards of paper pallet is fro m the weight of carton papers and indu stry papers used, and also fro m the wo rking hours of human powe rs used. just simply co mparing the manufacturing procedures between bee-hives type paper pallet and the proposed paper pallet, the latter shows a more efficient manufacturing procedure. 3.2.1. wave type structure shaping in this step, a design engineer needs to calculate the strength of carton flat paper to be used [4] and the feeding speed of ma in machine to perform wave type structure shaping [5-8]. if the selection of paper strength is not correct, it will increase the risk of tearing apart the carton flat papers during the shaping wave type structure process , which is an important procedure that proceedings of engineering and technology innovation, vol. 6, 2017, pp. 17 22 copyright © taeti 21 any engineer needs to be very cautious to avoid making mistakes. if a carton flat paper was tearing apart in the paper pallet mach ine, the whole machine has to be shut down and the process of removing the scraps inside the paper pallet machine is needed. the situations mentioned above must be tested during test runs. as the green light starts for making the main boards of proposed paper pallet, the machine must be sure that it will perform smoothly. the feed ing speed of ma king each ma in board could be adjusted up to 60 ma in boards per minute, however, the co ming cha llenge is the pasting and heatin g speed must be adjusted to match up the feeding speed of main machine. 3.2.2. gluing, pasting and heating procedures the key factors of adhering carton flat paper with wave type structure are not the a mount of glue used, nor the brand na me of glue used, but the density of glue used and the time needed to mix the surface fibers tangled together for both carton flat papers and wave type structure core materia ls. when an engineer find the right density of glue and the time needed to put both carton flat papers and wave type structure together, he has solved one of the most fetal problems of making paper pallets. for the p roposed paper pallet mach ine, before starting the machine, a stack of carton flat papers will put in front of mach ine in order to feed into the machine and performing wave type structure shaping procedure. in the same time, the bottom position of the paper pallet mach ine which is designed and placed under surface ground 2 meters depth with ma jor functions of feeding in carton flat papers, g luing, pasting on the bottom of the wave type structure. as soon as accomplishing the procedures mentioned above, the following step will be performing heating procedure to adhere both parts. the following step is wrapping another pre-mold ing carton flat paper on the work-in-process item which is a wave type structure was attached on a carton flat paper. the purpose of this step is to avoid the work-in-process item decreasing its peak heights. as its peak height decreased, the strength of the paper pallet decreased. the holding force could be applied by e xtending the machine structure of a strip of conveyors and apply ing jig and fixtures to withhold the work-in-process on the conveyors while continue its glue drying process, and which can be added up to 10, maybe 100 meters or more , if needed, depending on the space of factories one has or what budget one has. 3.3. purchasing cost, weight, and other factors among wood pallets, existing paper pallets in taiwan, and the proposed paper pallet table 3 comparisons of wood pallets and paper pallets (according to a pallet with 100cmx100c m, dynamic carrying weight 800kgs) proposed paper pallets wood pallets existing paper pallets (in taiwan) weight/each 7kgs 18~40kgs 8~12kgs manufacturing process shaping, pasting cutting, sawing, nailing gluing, pasting, cutting market price(nt$) 250 280~300 260~450 waste treatment recycled burning or burying recycled contaminant none wood scraps paper scraps automation feasible half automation hand made as shown in table 3, market price and light weight advantages of proposed paper pallets are better than that of wood pallets. co mpare to the producing speed between existing paper pallets manufacturers in taiwan and the proposed paper pallet mach ine, the latter shows much better automation advantage. the proposed paper pallet machine a lso could be applied in china, since many taiwan’s e lectronic manufacturers have invested in china and their products are needed to be shipped to countries worldwide. proceedings of engineering and technology innovation, vol. 6, 2017, pp. 17 22 copyright © taeti 22 3.4. other advantages of using paper pallets final step of discarding wood pallets is burning which will generate 100kgs of carbon dio xide per 100c mx100c m wood pallet. it was indicated in 2006[9] that one ton of carbon right trading price in european union ma rket is 20~25euro. thus, in the case of shanghai port (china), the ma rine shipping a mount is 26,152,400 t ues (in 2007). if 75% of marine shipping amount could be transferred to paper pellets shipping, the shipping amount will be 19,614,300 tues. each tue ne eds about 10 pallets to accomp lish unit load tasks, therefore, yearly shanghai port will consume 196,143,000 paper pallets. each paper pallet will reduce 100kgs of carbon dioxide generation which yearly could reduce 196,143,000 x 100 kgs = 19,614,300 tons of carbon dio xide. the previous a mount could be transferred to money wh ich is 19,614,300 x 20 euro = 392,286,000 euro per year. another fact is using recycled papers to ma ke pa llet s could reduce cutting trees. thus, using the previous data fro m shanghai port, yearly it could use 196,143,000 paper pa llets. each paper pallet we ights 7.5 kgs, yearly shanghai could generate 196,143,000 x 7.5kg / 1000kg = 1,471,070 tons of recycled papers. one ton of recycled paper will save cutting 17 trees, wh ich means that shanghai port yearly could save 1,471,070 x 17trees = 25,008,190 trees. 4. conclusions and future studies in this paper, a ne w design of paper pallet was proposed and a new paper pallet mach ine was also introduced. the ne w proposed machine could resolve the issues related to the paper pallet machine used in ch ina that was designed by netherland. the labor cost is increasing dramat ically late ly in ch ina, especially in southern-east coast area. low labor wage is no longer a huge advantage of investing manufacturing firms in ch ina. electronic manufacturing firms in ta iwan or china need to pay more attention to enhance the efficiency of global logistic management issues. also, the advantages of using paper pallets to acquire e xtra economic returns by engaging carbon right trading market and save cutting trees to protect our environment were a lso introduced. econo mic development a lso could ma int ain the sustainable environment preservations. the issue regarding shearing force of ma king wave type structure for the proposed paper pallet needs to be re-e xa mined carefully. expe riments could also be adopted for testing the strength ranges of carton flat papers used without tearing apart during the shaping procedure. an alternative is to develop a different method of feeding carton papers, not only increasing feeding speed but also reducing the risks of occurring carton paper jams. references [1] j. d. wisner, g. k. leong, and k. tan, principles of supply chain management-a balanced approach, south western, a division of thomson learning, inc., 2005. [2] peter f. drucker, “the economy’s dark continent,” fortune, p.14, april 1962. [3] douglas c. long and e. s. su, global logistics management, 2nd ed. taipei:hwatai publishing co., pp.248-249, june 2010.(in chinese) [4] a. c. ugural, mechanics of materials , taipei;new york:mcgra w-hill, 1991. [5] m. f. spotts, t. e. shoup, design of machine elements, 7th ed., new jersey: simom & schuster, 2002. [6] j. l. moum, z. lin, s. s. shei, etc., applied force-kinematics, taipei:gou-li publisher, january 1995.(in chinese) [7] mikell p. groover, automation, production system, and computer integrated manufacturing, new jersey:prentice-hall, inc., 1987. [8] mikell p. groover, fundamentals of modern manufacturing-materials, processes, and systems, 2nd ed., new jersey: prentice-hall, inc., 2002. [9] liberty times, taipei, 10 june, 2006.(in chinese) 3___peti#9025___20-26 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 20-26 evaluation of the shapley additive explanation technique for ensemble learning methods tsehay admassu assegie * department of computer science, injibara university, injibara, ethiopia received 05 december 2021; received in revised form 02 january 2022; accepted 03 january 2022 doi: https://doi.org/10.46604/peti.2022.9025 abstract this study aims to explore the effectiveness of the shapley additive explanation (shap) technique in developing a transparent, interpretable, and explainable ensemble method for heart disease diagnosis using random forest algorithms. firstly, the features with high impact on the heart disease prediction are selected by shap using 1025 heart disease datasets, obtained from a publicly available kaggle data repository. after that, the features which have the greatest influence on the heart disease prediction are used to develop an interpretable ensemble learning model to automate the heart disease diagnosis by employing the shap technique. finally, the performance of the developed model is evaluated. the shap values are used to obtain better performance of heart disease diagnosis. the experimental result shows that 100% prediction accuracy is achieved with the developed model. in addition, the experiment shows that age, chest pain, and maximum heart rate have positive impact on the prediction outcome. keywords: model explanation, random forest, ensemble method, interpretability 1. introduction feature visualization and model explanation techniques play an important role to develop interpretable heart disease diagnostic models and improve the diagnostic accuracy of ensemble learning methods. to implement automated ensemble methods for heart disease diagnosis, model explanation techniques are required [1]. recent research focuses on analyzing the prediction outcome of ensemble learning methods and the interpretation of automated models for creating trust, reliability, and adoption of the ensemble learning methods. the explanation of ensemble learning methods is significant in getting insights into the prediction process of such methods and improving the accuracy of heart disease diagnosis [2]. in addition to improving the prediction accuracy, improving the interpretability of the predicted outcome is important for the adaptability and reliability of such automated methods because the explainability of the automated models can increase the trust and reliability of the prediction outcome. random forest is one of the most widely implemented methods for heart disease prediction due to high precision [3]. designing and implementing interpretable ensemble learning methods for heart disease diagnosis has become one of the recent research topics in machine learning [4]. the reason is that, in recent years, the size of heart disease datasets is growing and the complexity of heart disease diagnosis increases the need for ensemble learning methods to obtain better prediction outcome. however, the application of ensemble learning methods to heart disease diagnosis requires additional research effort to develop the automated models that are automatically from the dataset features. moreover, the prediction results of heart disease diagnosis provided by ensemble methods are more promising than those provided by linear methods such as support * corresponding author. e-mail address: tsehayadmassu2006@gmail.com tel.: +251-9-21114923 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 20-26 vector machines [5]. thus, to develop reliable and adaptable ensemble learning methods for heart disease diagnosis, model interpretation is highly required. the reasoning for the prediction outcome made by the ensemble learning methods is important for domain experts to make decisions on whether the outcome is accepted. this study aims to improve the accuracy of an existing ensemble learning method for heart disease prediction by employing the feature visualization technique and the shapley additive explanation (shap) values with the real-world heart disease datasets obtained from kaggle data repository. experiments are conducted to analyze the features with significant influence on the prediction outcome by the shap values. overall, this study aims at achieving the following specific objectives. 1) the study provides a comprehensive and thorough analysis of the literature on the shap model explainability technique. 2) the study explores heart disease features that have significant influence on the prediction outcome of an ensemble method using random forest. 3) the study designs and implements a trusted and reasonable ensemble-based heart disease diagnosis method. 4) the study evaluates the performance of the developed ensemble method for heart disease diagnosis and analyzes the effectiveness of the shap technique in explaining the ensemble method. 2. literature review recent research has shown significant improvement in the performance of heart disease diagnosis using ensemble learning methods such as random forest on complex training samples [6]. the evaluation and interpretation of ensemble learning models for heart disease diagnosis have become an active field of research [7]. model interpretation is gaining much attention and research effort due to the fact that data scientists have noticed the importance of explaining why machine learning models have predicted an instance as a positive or negative class in the medical domain. interpretability is aimed to increase the trustworthiness and reliability of the predictive outcome made by machine learning models on risky decisions, particularly in the medical domain such as the prediction of diabetes disease [8]. interpretability has become a major concern in developing robust and trusted machine learning models for heart disease diagnosis. numerous previous works have investigated the predictive models of heart disease using machine-learning ensemble methods and deep learning algorithms. literature also shows that high accuracy is attained with the ensemble and deep learning methods [9]. however, the models are not widely adopted due to the lack of interpretability and transparency. model transparency, auditability, and explainability of ensemble learner models are critical to provide a maximally accurate heart disease prediction [10]. additionally, domain experts, regulators, or policy makers demand medical diagnostic models to be transparent and interpretable. thus, in medical diagnosis, the very simple predictive models such as logistic regression or decision trees are still widely employed for medical diagnosis. however, the superior predictive power of modern machine learning algorithms in solving complex problems with the growing size of datasets is not leveraged with simple models. thus, much research effort is needed to develop understandable medical diagnostic models and frameworks for making the “black box” machine learning models transparent, auditable, and explainable. the interpretability of models is gaining much research attention due to three reasons. first, predictive models are widely employed in different areas. second, simple models are not effective to provide good performance on complex problems. lastly, there is growing failure of adoption of complex models due to the interpretability and transparency problems. interpretable and explainable machine learning is regarded as a process, encompassing three high-level stages [11]. first, model interpretability is critical to understand ensemble learning models and datasets. second, model interpretability is important to improve the performance of models using explainable artificial intelligence (ai) methods. third, model interpretability and explanation are applied to refine and optimize machine-learning models. for instance, this study presents an application where interpretability is important. for heart disease prediction, the goal is to identify the patients to be classified into heart disease positive or heart disease negative classes. specifically, the problem is framed as a binary classification problem where the goal is to predict whether a patient is suffering from heart disease or not by considering the 21 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 20-26 related factors: heart rate, maximum heart rate achieved (thalach), presence or absence of chest pain (cp), fasting blood sugar level (fbs), cholesterol level (chol), presence or absence of the exercise-induced angina, total resting blood pressure (trestbps), sex, age, slope, resting electrocardiogram (restecg), old peak, and other patient conditions. the contribution of this study is the use of the shap values to explain and interpret the prediction outcome of random forest with the shap technique. the literature shows that only few works are published regarding the explainability of ensemble learning methods on heart disease diagnosis [12]. however, the existing models do not explain the prediction outcome. to the best of the author’s knowledge, there are no prior published works on the use of interpretable random forest-based heart disease prediction models. this study explores the application of shap in the machine-learning context for medical diagnosis using heart disease datasets as a case study. overall, the study is designed to answer the following research questions: (1) how to increase the reliability of ensemble learning methods such as a random forest model for heart disease diagnosis? (2) how does the random forest model make predictions with heart disease datasets? (3) how to improve the prediction outcome of an ensemble learning method for heart disease diagnosis? (4) what is the influence of heart disease features on the prediction outcome of the random forest model? overall, the design and implementation process of the proposed automated heart disease diagnosis system is demonstrated in fig. 1. the purpose of this study is to develop a more interpretable and accurate automated heart disease diagnosis model using the random forest model and the shap technique. the reliability of the proposed model will be illustrated with the model prototype, and verified with real-world heart disease datasets. fig. 1 the process of model explanation with shap 3. design methodology and implementation the complexity of ensemble learning methods and the growing size of real-world heart disease datasets limit the ability to understand what a model has learned or why a given prediction is made, acting as a barrier to the adoption of machine learning [13]. additionally, from a legal or regulatory perspective, it is required to explain the prediction outcome of a machine-learning model. machine learning explainability (mlx) is a process of explaining and interpreting an ensemble learning model. to conduct the study, the author collects 1025 real-world heart disease datasets from university of california irvine (uci), obtained from a publicly available kaggle data repository, to the scientific community for experimentation. in this study, shap is employed to generate different explainers: objects used to compute the shap values for each sample in the datasets. these values represent the impact of certain features on the prediction outcome. positive values contribute to increasing the final probability, and negative values contribute to decreasing the probability. the experiment is performed on the jupyter notebook in python 3.7. the datasets are in comma-separated value (csv) file format. to be used for analysis, the data is imported to the jupyter notebook python environment using panda’s data analysis library in the form of the data frame. the libraries used for data transformation include pandas, matplotlib, scientific learning kit, and shap. to develop a model for heart disease diagnosis, the author employs random forest algorithms. the datasets are divided into a training set and a testing set. the training set includes 70% of the datasets or 717 observations, and the testing set includes 30% of the datasets or 308 observations. heart disease dataset random forest model shap random forest model random forest model 22 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 20-26 4. existing design fig. 2 shows an existing ensemble learning-based automated heart disease diagnosis system. as can be seen, a model is developed using a feature selection method to improve the prediction outcome. while the application of feature selection improves the prediction accuracy of the system, the reliability and interpretability of the existing model limit the adaptability of such a model in the real-world scenario. thus, to address the existing gap (i.e., the lack of reasoning) for the prediction outcome of the ensemble learning model on heart disease diagnosis, the study proposes a new model with the help of the shap technique to increase the trust of domain experts on using automated models in decision-making. overall, this study aims to introduce the shap technique for model explanation as demonstrated in fig. 1. fig. 2 the existing design 5. experimental results and discussion this section presents the application of shap values for explaining and interpreting the prediction outcome of a random forest model on heart disease diagnosis. the experiment is conducted with shap for heart disease prediction with random forest algorithms. to evaluate the proposed random forest model, the author employs shap for improving the model performance and interpretability. the experiment is designed to explain the prediction output of the random forest model (i.e., the prediction of heart disease positive or heart disease negative). the author employs the heart disease prediction outcome as a case study to evaluate the interpretability of a random forest model using shap. the empirical analysis is conducted to evaluate shap-scores using heart disease datasets. all experiments are implemented in python 3.7 on jupyter notebook environment. the experiment is performed on intel(r) core(tm) i7-8565u cpu, 1.80 ghz, 1992 mhz, 4 core(s), 8 logical processor(s) using microsoft windows 10 pro 64-bit version. for data management and optimization, panda’s library is used. the shap values of each feature in heart disease datasets are demonstrated in fig. 3. as observed from fig. 3, chest pain (cp), the number of major vessels colored by fluoroscopy (ca), and thalassemia (thal) are the factors that contribute to the model’s prediction. thus, the number of major vessels colored by fluoroscopy (ca), chest pain (cp), and thalassemia (thal) are the features that have the greatest influence on the prediction outcome of the ensemble model. as demonstrated from the experimental result in fig. 3, a patient with chest pain is most likely predicted as a heart disease patient by the developed random forest model. the number of major vessels colored by fluoroscopy (ca) has the second most impact on the model output. as observed from the plot given in fig. 3, the features with lower shap values include resting electrocardiogram (restecg), fasting blood sugar level (fbs), and slope. feature selection with chi-square and correlation feature scaling using min-max scaler cleveland heart disease dataset data pre-processing classification using ensemble-learning algorithms performance analysis result and analysis 23 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 20-26 the features with lower shap values have less impact on the model output as compared to the features with higher shap values. the features having high shap values impact the model output on the positive class (i.e., heart disease patients). in contrast, the features with lower shap values have positive impact on the predictive outcome of heart disease negative (i.e., patients without heart disease). the feature “resting electrocardiogram (restecg)” shows that the lowest shap value (approximately 0.4 shap value as illustrated in fig. 3) has positive impact on the predictive outcome of the negative class (class o). another important plot for analyzing the impact of features on the model output is the shap waterfall plot, demonstrated in fig. 4. fig. 4 demonstrates the impact of heart disease features on the random forest model output on heart disease prediction. as observed from fig. 4, the number of major vessels colored through fluoroscopy (ca) has the highest impact for the class of heart disease positive, but is the second most impactful feature for the positive prediction outcome using the shap summary plot demonstrated in fig. 3. ranked below the number of major vessels colored by fluoroscopy (ca), chest pain (cp), and thalassemia (thal) are the features which influence the model’s prediction towards the class of heart disease positive. in contrast, slope, age, resting electrocardiogram (restecg), exercise-induced angina, and total resting blood pressure (trestbps) influence the model’s prediction towards the class of heart disease negative. features such as maximum heart rate (thalach), fasting blood sugar level (fbs), and cholesterol level (chol) have lower effects on the model’s prediction as compared to other heart disease features as demonstrated in fig. 4. fig. 5 demonstrates the interaction plot between maximum heart rate (thalach) and cholesterol level (chol). as illustrated in fig. 5, maximum heart rate (thalach) and cholesterol level (chol) have a linear relationship with each other, showing that they are highly correlated. the strong correlation between them indicates their dependency on each other. fig. 3 shap values of each feature fig. 4 shap waterfall plot fig. 5 shap interaction plot of maximum heart rate (thalach) vs cholesterol level (chol) fig. 6 roc curve of the developed model 24 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 20-26 fig. 7 interaction plot of total resting blood pressure (trestbps) vs age fig. 8 interaction plot of maximum heart rate (thalach) vs age fig. 6 demonstrates the receiver operating characteristic (roc) curve of the random forest model for heart disease diagnosis. as can be seen, the area under the roc curve is 1.00, which shows that the model is perfect for the detection of heart disease. moreover, the area score or the area under curve (auc) value of 1.00 evidently proves that the model performs well and does not have the prediction outcome of false positive and false negative classes on the test set conducted in the experiment. fig. 7 demonstrates the shap interaction plot between total resting blood pressure (trestbps) and age. as demonstrated in fig. 7, the shap interaction values for total resting blood pressure (trestbps) are higher for the patients above the age of 50 years. thus, the plot evidently proves that there is high positive interaction between age and total resting blood pressure (trestbps). fig. 8 illustrates the shap interaction value for maximum heart rate (thalach) and age. as observed from fig. 8, the shap interaction value for maximum heart rate (thalach) is higher for the patients above the age of 45 years. thus, the plot appears to prove that there is high positive interaction between age and maximum heart rate (thalach). 6. conclusions this study presents an explainable and automated heart disease diagnosis model using the shap model explanation technique and random forest-based ensemble learning method. with the shap technique, an interpretable random forest model with 100% prediction accuracy is successfully developed. the shap technique is effective for providing the intuition and the reasoning behind the predictive outcome of ensemble learning methods such as random forest for heart disease diagnosis. for future work, it is recommended to use various model explanation techniques and machine learning algorithms, such as extreme boosting and deep learning, to identify whether a patient is suffering from heart disease or not. the results of future work will facilitate the trusted and reliable automated intelligent heart disease diagnosis models that would become alternatives to the healthcare experts. conflicts of interest the author has no conflicts of interest. references [1] s. liu, x. wang, m. liu, and j. zhu, “towards better analysis of machine learning models: a visual analytics perspective,” visual informatics, vol. 1, no. 1, pp. 48-56, march 2017. [2] k. aas, m. jullum, and a. løland, “explaining individual predictions when features are dependent: more accurate approximations to shapley values,” artificial intelligence, vol. 298, article no. 103502, september 2021. [3] a. chatzimparmpas, r. m. martins, i. jusufi, and a. kerren, “a survey of surveys on the use of visualization for interpreting machine-learning models,” information visualization, vol. 19, no. 3, pp. 207-233, july 2020. 25 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 20-26 [4] j. zhou, a. h. gandomi, f. chen, and a. holzinger, “evaluating the quality of machine learning explanations: a survey on methods and metrics,” electronics, vol. 10, no. 5, article no. 593, march 2021. [5] p. linardatos, v. papastefanopoulos, and s. kotsiantis, “explainable ai: a review of machine learning interpretability methods,” entropy, vol. 23, no. 1, article no. 18, january 2021. [6] h. s. yan, m. c. tsai, and m. h. hsu, “an experimental study of the effects of cam speeds on cam-follower systems,” mechanism and machine theory, vol. 31, no. 4, pp. 397-412, may 1996. [7] d. farrugia, c. zerafa, t. cini, b. kuasney, and k. livori, “a real-time prescriptive solution for explainable cyber-fraud detection within the igaming industry,” sn computer science, vol. 2, no. 3, article no. 215, may 2021. [8] k. futagami, y. fukazawa, n. kapoor, and t. kito, “pairwise acquisition prediction with shap value interpretation,” the journal of finance and data science, vol. 7, pp. 2244, november 2021. [9] m. chaibi, e. m. benghoulam, l. tarik, m. berrada, and a. e. hmaidi, “an interpretable machine learning model for daily global solar radiation prediction,” energies, vol. 14, no. 21, article no. 7367, november 2021. [10] p csókaa, f. illés, and t. solymosi, “on the shapley value of liability games,” european journal of operational research, in press. [11] c. m. viana, m. santos, d. freire, p. abrantes, and j. rocha, “evaluation of the factors explaining the use of agricultural land: a machine learning and model-agnostic approach,” ecological indicators, vol. 131, article no. 108200, november 2021. [12] s. n. payrovnaziri, z. chen, p. rengifo-moreno, t. miller, j. bian, j. h. chen, et al., “explainable artificial intelligence models using real-world electronic health record data: a systematic scoping review,” journal of the american medical informatics association, vol. 27, no. 7, pp. 1173-1185, 2020. [13] k. dissanayake and m. g. m. johar, “comparative study on heart disease prediction using feature selection techniques on classification algorithms,” applied computational intelligence and soft computing, vol. 2021, article no. 5581806, 2021. copyright© by the authors. licensee taeti, taiwan. this article is an open access article distributed under the terms and conditions of the creative commons attribution (cc by-nc) license (https://creativecommons.org/licenses/by-nc/4.0/). 26  proceedings of engineering and technology innovation, vol. 7, 2017, pp. 01 07 a designed enclosure unit for loop-mediated isothermal amplification arayan wongniyom 1 , jyh-jian chen 2,* 1 department of tropical agriculture and international cooperation, national pingtung university of science and technology, pingtung, taiwan, roc. 2 department of biomechatronics engineering, national pingtung university of science and technology, pingtung, taiwan, roc. received 17 july 2017; received in revised form 05 august 2017; accepted 25 august 2017 abstract the gene amplification technique is used in medical science and biotechnology popularly. the loop-mediated isothermal amplification (lamp) is a nucleic acid amplification method that can be utilized in low-resource settings and does not require expensive or complex instruments. the annealing process of lamp is performed at a constant temperature (around 65 °c) for an hour. from previous works, phase change materials (pcms) have been applied for the latent heat energy storages due to their isothermal operation during phase transitions. the utilization of the melting temperature of the paraffin wax to maintain the suitable temperature for lamp process is introduced in our study. to enhance the heat transfer inside the energy storage unit, several pin fins are constructed on the inner surface of the unit which is 50 mm × 50 mm × 50 mm in inner size. the influence of various numbers of pin fins on the temperature distribution is examined by the numerical simulation. the heat transfer inside the paraffin enclosure with fins is better than that without fin. it can be found that the temperature uniformity in the pcm inside the enclosure with fins shows much better. though the numbers of the fin are increased, some transfer energy from the heat source remains at the enclosure wall and cannot penetrate into the pcm. the duration for all pcm melted from solid phase to liquid phase is over 10 minutes. in the future, the geometric parameters on the duration for phase transition of pcm inside the lamp chamber will be investigated. keywords: phase change materials, lamp, paraffin wax 1. introduction the detection of nuclei-acid sequences is popularly used in medical research, biotechnology and food science. the loop-mediated isothermal amplification (lamp) is a biotechnique that can amplify deoxyribonucleic acid (dna) within 60 minutes under isothermal conditions (50-65 °c). this technique, a highly sensitive technique that can be used in low-resource settings and does not require expensive or complex instruments [1], is suitable for small-scale hospitals, primary care facilities and clinical laboratories in developing countries [2-3]. in previous studies, the lamp technique was able to yield the amplification of 109-bp of dna [4], and applied to detect the virus as follows: mycobacterium [1], tuberculosis [5], m. avium and m. intracellulare in sputum [3], porcine kobuvirus [6], etc. * the lamp process is to anneal dna and primers with an appropriate constant temperature. to build a reaction device for lamp process, one of the most important issues is to design a reaction chamber that can be kept at a constant temperature for a period of time. when heat energy is supplied to or extracted from a phase change material (pcm), there is only a little * corresponding author. e-mail address: chaucer@mail.npust.edu.tw tel.: +886-8-7703202 ext. 7029, 7839(l) mailto:chaucer@mail.npust.edu.tw proceedings of engineering and technology innovation, vol. 7, 2017, pp. 01 07 copyright © taeti 2 change of temperature inside the pcm. therefore, pcm can be used to stabilize the temperature in an application for lamp process. in general, pcm refers to solid-liquid phase changes only [7]. paraffin wax is the most commonly used commercial organic pcm because of its high heat of fusion, varied phase change temperatures, negligible supercooling, lower vapor pressure in a melt and chemically inert and stable behavior [8]. however, due to low thermal conductivity of paraffin wax, it requires large surface area for heat exchange [9]. kandasamy et al. [10] studied about using the paraffin wax as the pcm-based heat sink. the results from simulations and experiments have shown that pcm-based heat sink can be used in intermittent-use devices. nayak et al. [11] improved the cooling of the electronics devices by using pcm with fins made of thermal conductivity enhancer material. pakrouh et al. [12] presented a numerical investigation to optimize the geometry of pcm-based fin heat sinks. the result showed the fin height affects the base temperature. hosseinizadeh et al. [13] carried out a comparison between heat sink with and without pcm. their results revealed that fins thickness larger than 4 mm did not improve the performance of the heat sink. kamkari et al. [14] demonstrated the experimental results of pcm melting in an enclosure with and without horizontal partial fins. results indicated that when the surface-averaged nusselt number was reduced, increasing the number of fins decreased the melting time and increased the total heat transfer rate. lamp has been utilized for the detection of dna and/or rna from a wide range of bacterial and viral pathogens [15-17]. curtis et al. [15] presented a non-instrumented nucleic acid amplification (nina) heating device. the heat from the exothermic chemical reaction was enough to anneal dna sample for 60 minutes. however, the changing of the ambient temperature had an effect on the temperature stability during the amplification process and it was difficult to control the heat from the chemical reaction. in our lamp device, a cartridge heater is utilized to generate heat energy for melting pcm and then the melting pcm is used for stabilizing the sample temperature during the dna amplification. this process does not need any thermal controller and the integration of the lamp device was easy. 2. numerical methodology a schematic diagram of three-dimensional the enclosure physical model consist of lid, fin array and cavity, the dna sample put on the top surface of the lid is shown in fig. 1(a). considering the symmetry of the geometry, only one quarter of the physical model is included for computation. the pin fin array and the cavity are made of aluminum which contains the pcm. a uniform heat power is applied at the bottom surface of the cavity. initially, the solid pcm and enclosure are in thermal equilibrium with the ambient temperature, 20 °c. the solid and liquid phases of pcm are homogeneous and isotropic. thermal resistance across the wall of the container is neglected and lateral sides of the rectangular enclosure are well insulated. the melting point of pcm is 65 °c. the effect of natural convection within the melt pcm is neglect because the space into enclosure is moderately narrow and fin array barrier velocity flow when phase transition to liquid. the pcm undergoes the transition from the solid state to the liquid state. there are governing equations as follows. heat conduction in liquid region: 𝜕𝑇𝑙 𝜕𝑡 = 𝛼𝑙 𝜕2𝑇𝑙 𝜕𝑥2 + 𝛼𝑙 𝜕2𝑇𝑙 𝜕𝑦2 + 𝛼𝑙 𝜕2𝑇𝑙 𝜕𝑧2 , 𝑡 > 0 (1) heat conduction in solid region: 𝜕𝑇𝑠 𝜕𝑡 = 𝛼𝑠 𝜕2𝑇𝑠 𝜕𝑥2 + 𝛼𝑠 𝜕2𝑇𝑠 𝜕𝑦2 + 𝛼𝑠 𝜕2𝑇𝑠 𝜕𝑧2 , 𝑡 > 0 (2) interface temperature: 𝑇(𝑥, 𝑦, 𝑧, 𝑡) = 𝑇𝑚 (3) proceedings of engineering and technology innovation, vol. 7, 2017, pp. 01 07 copyright © taeti 3 stefan condition: 𝑘𝑠 𝜕𝑇𝑠 𝜕𝑥 − 𝑘𝑙 𝜕𝑇𝑙 𝜕𝑥 = 𝐻𝑓 𝜌 𝑑𝑋 𝑑𝑡 , 𝑡 > 0 (4) where 𝛼 is thermal diffusivity. k is the thermal conductivity. 𝛾 is the melt fraction during the phase change process and is defined by the following relations: 𝛾 = 0 , if 𝑇 < 𝑇𝑠; 𝛾 = 0 , if 𝑇 > 𝑇𝑙 ; 𝛾 = 𝑇−𝑇𝑠 𝑇𝑙−𝑇𝑠 , if 𝑇𝑠 < 𝑇 < 𝑇𝑙 . tm is melting temperature, the subscript l and s is liquid and solid phase respectively. a grid system is composed of 32768 to 61952 hexahedral cells, which depends on the numbers of fin. the numbers of fin are varied from 4 to 49 pin fins, h is fin height, which is varied from 20 mm to 47 mm and the fin maintains 2 mm thickness of whole models shown in fig. 1(b). a constant heat power of 14 w is applied at the bottom surface of the enclosure. a finite volume approach is adopted by using computational fluid dynamics (cfd) software esi-cfd (v2006, cfd research corporation, huntsville, al, usa) to model the pcm melting process. a structured grid system with good orthogonality is recommended for accurate prediction and reliability. in this study the geometric model for the three dimensional rectangular structure is considered. in the computation of the enthalpy of the simulation model, the spatial discretizations are performed using a second-order upwind scheme. the conditions of convergence can be divided into two kinds. the first one is the maximum number of the solver iterations. the second one determines the convergence criteria to be used. in this study, we perform 100 sweep times and use 5×10-5 as the convergence criterion. grid sensitivity is evaluated by using different element sizes to ensure that the simulated results are independent of element size. the element size is repeated until the similarity in results is within the acceptable solution tolerance (0.1%). this study use the pcm, which has the melt point near the amplification process of lamp technique is the rubitherm® rt64hc paraffin wax has a solid density of 880 kg/m 3 , a liquid density of 780 kg/m 3 , a specific heat of 2000 j/kg-k, a latent heat of 2.3×10 5 j/kg and a thermal conductivity of 0.2 w/m-k. aluminum has a density of 2700 kg/m 3 , a specific heat of 880 j/kg-k and a thermal conductivity of 190 w/m-k. (a) the quarter part of the enclosure and pcm (b) bottom view of the heat sinks with 4, 9, 20, 28, 36, and 49 pin fins fig. 1 the numerical models of 2 mm thickness fin 3. results and discussion in this section, we present the thermal characteristics obtained from the three dimensional simulations and the results of the parametric studies are shown. the discussion concerns the calculated temperatures and melt fractions of the pcm as functions of time. in our study, heat conduction is the main heat transfer mechanism inside the enclosure. the heater generates heat flux is attached to the bottom of the aluminum enclosure with a high thermal conductivity. with an increase in time, the proceedings of engineering and technology innovation, vol. 7, 2017, pp. 01 07 copyright © taeti 4 heat energy from pin fin array and cavity through to the solid pcm, melting of the pcm and then heating of the molten pcm to temperature above its melting point. the material of the fins is identical to the enclosure. therefore the temperature of the enclosure and the fins increase more than that of pcm. the temperature of the lid might exceed the temperature suitable for amplification process in lamp technique. the purpose of the fins is enhancing thermal conductivity of pcm that contain inside an enclosure, the heat flux flow with heat conduction from the heat source at the bottom cavity aluminum to pcm. the large disparity thermal conductivity between aluminum and pcm that make effect to the intensive heat flux flow in material. the fins array is proficiently constrained the temperature for annealing the sample on the top surface enclosure shown in fig. 2(a). in order to assess the influence of fin, melt fraction with time variation is evaluated for the pcm with fins and without fin (a) temperature contour of 36 pin fins and fin height of 47 mm at top view at 110 s (b) compared melting fraction varied (c) compared melting fraction varied fin height. fig. 2 the melt contour inside enclosure of pcm and melt fraction effected of number of pin fin and fin height after 1800s, there are three cases of pcm no complete melting to liquid as follows: without fin, 20pins fin and 47mm fin height, and 49 pins fin and 20mm fin height fig. 2(b) and 2(c) show the temporal variations of melt fractions during melting processes for finned enclosure and enclosure without fun. a rapid increase in temperature of pcm is observed with fins than that of pcm without fins. this is because the longitudinal finned material is able to transfer heat from the heat source quickly. the higher heat transfer rate was induced by the presence of fins near the heat source causes faster melting of pcm. thus, the total melting time is reduced compared to the system without fins. the fins accelerate the heat transfer rate from the surface of heat source to and through the pcm. the melt fraction of pcm with 49 pin fins and fin height of 47 mm is higher than that of other cases. regarding to the enclosure with 36 pin fins, the melt fraction is lower than that with 49 pin fins, but is almost the same. therefore, the enclosure with 36 pin fins and fin height of 47 mm is capably performed as well. proceedings of engineering and technology innovation, vol. 7, 2017, pp. 01 07 copyright © taeti 5 (a) compare without fin with 49 pin fins, 47 mm fin height (b) at base cavity (c)pcm fig. 3 average temperature profiles the gradients of the temperature contours are large at the bottom of the enclosure during the observed times because the applied heat source is at the bottom of the enclosure. the pcm has a low thermal conductivity. therefore, the large heat flow along aluminum wall causes high temperature, the pcm temperature is improving almost lid temperature when applied fin array the 49 pins fin and 40mm fin height shown in fig. 3(a). although concentrated temperature at the base cavity in 49 pins and 40 mm fin height is most (neglect without fin case) fig. 3(b), the temperature profiles of pcm almost 49 pins and 47 mm fin height case after 900s. the solid pcm was heated by the adjacent liquid pcm. hence fin array enhanced and increased the melting rate. the extended surfaces of the fin array are inserted into pcm and lead to more heat transfer than the enclosure without fin, as shown in fig. 3(c). the fin array and cavity structures cause the heat energy penetrated into the pcm. the largely different thermal conductivity between aluminum and pcm has the effect to intensive gradient temperature occur on surface heat transfer of aluminum, demonstrated in fig. 4(a,b), and the space without fin remain solid state indicate that the heat transfer between liquid to solid not well. the melting rate is enhanced by pin fin array. the number of pin fin and fin height directly influence the surface heat transfer. when the surface heat transfer is increased, the heat energy for phase transition is reduced in fig. 4(c). it is because the large heat flux through surface fin to pcm change to complete the melting, the liquid pcm cannot absorb energy. the influence of various fin heights and numbers of pin fin on melt fraction when each fin surface heat transfer is almost equal is shown in fig. 4(c). the surface heat transfer are 22840 mm 2 with 49 pin fins and 20 mm fin height, 22528 mm 2 proceedings of engineering and technology innovation, vol. 7, 2017, pp. 01 07 copyright © taeti 6 with 28 pin fins and 47 mm fin height and 22520 mm 2 with 20 pin fins and 47 mm fin height, respectively. the 49 pin fins and 40 mm fin height has surface heat transfer 30680 mm 2 less surface heat transfer than 49 pin fins and 47 mm fin height (33424 mm 2 ) and use latent heat energy for phase transition. the difference of the melt fractions among these cases are within 10% after the beginning of the duration of 300 s. therefore, the latent heat energy storage depends on the surface of the heat transfer. the saved energy for each model can be found by integrating the area under the relative latent heat energy and time. the slope of each graph indicates the energy to be used within 1800s. the group of 49 pin fins saves more energy than others with large space between adjacent fins. (a)top view cross-section at half of fin length (b) front view cross-section at middle of first row of fin (c) latent heat energy and surface heat transfer fig. 4 temperature contours of 36pin 47mm fin height model at 300 (a) the area under each the graph (b) the slope value each graph of latent heat energy fig. 5 the latent energy in each case used to melt pcm in 1800s proceedings of engineering and technology innovation, vol. 7, 2017, pp. 01 07 copyright © taeti 7 4. conclusion in this paper, the numerical study of the designed pcm enclosure for lamp was presented. the transient simulation of heat conduct problem of an enclosure with pcm is shown with several pin fins inside the enclosure. the main parameters about enhancing the heat transfer were the number of pin fin and the fin height. the simulation results show an enhancement of melting rate and a decreasing temperature gradient near the heating wall of enclosure with fin array extended from lid. the thorough pin fin position distribution and the number of pin fin can be used to save the energy to melt pcm. the simulation study shows the pin fin array technique can improve the potential of pcm for applying in the lamp machine. acknowledgment the authors would like to thank the ministry of science and technology of the republic of china for financially supporting this research under contract no. most 105-2313-b-020-007-. references [1] e. aryan, m. makvandi, a. farajzadeh, k. huygen, p. bifani, and s. l. mousavi et al, “a novel and more sensitive loop-mediated isothermal amplification assay targeting is6110 for detection of mycobacterium tuberculosis complex,” microbiological research, vol. 165, no. 3, pp. 211-220, 2010. [2] y. mori and t. notomi, “loop-mediated isothermal amplification (lamp): a rapid, accurate, and cost-effective diagnostic method for infectious diseases,” journal of infection and chemotherapy, vol. 15, no. 2, pp. 62-69, 2009 [3] t. iwamoto, t. sonobe, and k. hayashi, “loop-mediated isothermal amplification for direct detection of mycobacterium tuberculosis complex, m. avium, and m. intracellulare in sputum samples,” journal of clinical microbiology, vol. 41, no. 6, pp. 2616-2622, 2003. [4] t. notomi, h. okayama, h. masubuchi, t. yonekawa, k. watanabe, and n. amino et al, “loop-mediated isothermal amplification of dna,” nucleic acids research, vol. 28, no. 12, pp. e63-e63, 2000. [5] c. c. boehme, p. nabeta, g. henostroza, r. raqib, z. rahim, and m. gerhardt et al., “operational feasibility of using loop-mediated isothermal amplification for diagnosis of pulmonary tuberculosis in microscopy centers of developing countries,” journal of clinical microbiology, vol. 45 no. 6, pp. 1936-1940, 2007. [6] c. li, j. chen, h. shi, x. zhang, d. shi, and x. han et al., “rapid detection of porcine kobuvirus in feces by reverse transcription loop-mediated isothermal amplification,” virology journal, vol. 11, no. 1, pp. 73, 2014. [7] h. mehling and l. f. cabeza, heat and cold storage with pcm, berlin, springer, 2008. [8] m. k. rathod and j. banerjee, “thermal stability of phase change materials used in latent heat energy storage systems: a review,” renewable and sustainable energy reviews, vol. 18, pp. 246-258, 2013. [9] m. akgün, o. aydın, and k. kaygusuz, “experimental study on melting/solidification characteristics of a paraffin as pcm,” energy conversion and management, vol. 48, no. 2, pp. 669-678, 2007. [10] r. kandasamy, x. q. wang and a. s. mujumdar, “transient cooling of electronics using phase change material (pcm)-based heat sinks,” applied thermal engineering, vol. 28, no. 8, pp. 1047-1057, 2008. [11] k. c. nayak, s. k. saha, k. srinivasan, and p. dutta, “a numerical model for heat sinks with phase change materials and thermal conductivity enhancers,” international journal of heat and mass transfer, vol. 49, no. 11, pp. 1833-1844, 2006. [12] r. pakrouh, m. j. hosseini, a. a. ranjbar, and r. bahrampoury, “a numerical method for pcm-based pin fin heat sinks optimization,” energy conversion and management, vol. 103, pp. 542-552, 2015. [13] s. f. hosseinizadeh, f. l. tan, and s. m. moosania, “experimental and numerical studies on performance of pcm-based heat sink with different configurations of internal fins,” applied thermal engineering, vol. 31, no. 17, pp. 3827-3838, 2011. [14] b. kamkari and h. shokouhmand, “experimental investigation of phase change material melting in rectangular enclosures with horizontal partial fins,” international journal of heat and mass transfer, vol. 78, pp. 839-851, 2014. [15] k. a. curtis, d. l. rudolph, i. nejad, j. singleton, a. beddoe, and b. weigl et al., “isothermal amplification using a chemical heating device for point-of-care detection of hiv-1,” plos one, vol. 7, no. 2, p. e31432, 2012. [16] j. singleton, j. l. osborn, l. lillis, k. hawkins, d. guelig, and w. price et al., “electricity-free amplification and detection for molecular point-of-care diagnosis of hiv-1,” plos one, vol. 9, no. 11, p. e113693, 2014. [17] j. singleton, c. zentner, j. buser, p. yager, p. labarre, and b. h. weigl, “instrument-free exothermic heating with phase change temperature control for paper microfluidic devices,” spie proceedings, vol. 8615, p. 86150r, 2013.  proceedings of engineering and technology innovation, vol. 12, 2019, pp. 26-34 job mix formulation for bituminous concrete grade ii using c# programming v. venkat ramayya 1,* , v. vinayaka ram 2 , s. krishnaiah 3 1 department of civil engineering, mahatma gandhi institute of technology, hyderabad, india 2 department of civil engineering, bits, hyderabad, india 3 department of civil engineering, jntua, anantapuramu, india received 16 january 2018; received in revised form 04 march 2018; accepted 17 april 2018 abstract bituminous concrete mix design is a complex process and different methods are proposed by a number of researchers. marshall’s method of mix design is one such process, which became very popular among the practitioners in india and elsewhere. superpave method of mix design which takes into account climatic and traffic factors for aggregate and binder selection. it has gained its prominence and is observed to be superior to any other well-known mix design processes. aggregate gradation adopted for the mix design plays a crucial role in the performance of the mix. many researchers have developed mechanisms to work out job mix formulations (jmf) so as to reach the appropriate aggregate mix in the field. the analytical phase of jmf deals with the method of blending of aggregates from the identified stockpiles in order to fulfill the requirements of gradation as specified by agencies for a given layer of flexible pavements. methods proposed for aggregate blending includes the graphical methods, trial and error method and optimization techniques. the aim of these techniques is two-fold; the first is to optimize the cost involved in blending from the known unit cost of individual stockpiles; and the other is to fulfill the gradation recommendations of the local agencies and parameters suggested in other methods. in this research, an attempt has been made to develop c# (c-sharp) application in .net framework using the sequential search technique to work out the best blend of aggregate particles by ensuring the requirements mentioned in morth’s (ministry of road transport & highways, india) recommended practice. bailey’s method and the maximum density line have been recommended in 0.45 power chart and incorporating optimization criteria. grade ii bituminous concrete has been chosen for the research. it was clearly observed that the fine aggregate fraction in the stockpiles influences the final job mix formulation. the devleloped tool can accommodate up to 5 stockpiles. it was also observed that the bailey’s requirements are more stringent when compared with morth requirements as far as jmf is considered. the developed tool is capable of generating the blends with recycled asphalt pavement (rap) aggregate blends. keywords: bituminous concrete, maximum density line, bailey’s parameters, aggregate blending, c# programming, job mix formula. 1. introduction the flexible pavement is a layered system where the load transfer takes place through the grain to grain contact [1]. in each of these layers, an aggregate blend with different sizes will be used along with the bitumen binder, which forms the binder matrix and holds the aggregates in place. the bituminous concrete mix design is a complex process. marshall’s method of mix *corresponding author. e-mail address: ramayya@mgit.ac.in tel.: +91-99669 20375 proceedings of engineering and technology innovation, vol. 12, 2019, pp. 26-34 27 design is the most popular in india and elsewhere [2].the marshall specimen prepared using the optimum binder content (obc) shall satisfy the parameters like stability, flow, volume filled with bitumen (vfb), volume in mineral aggregates (vma) and volume of voids (vv) [9]. agencies from different countries have standardized the requirements of these parameters and it is mandatory to verify these parameters before adopting a given mix in the field. a number of internal and external factors are responsible for distress in flexible pavements. aggregate gradation is one of the internal factors. the gradation adopted is going to influence the performance of dense graded, open graded and stone matrix asphalt mixes. research has shown that aggregate gradation will influence the performance of porous asphalt mixes. these pavements are now in use for effective stormwater management.the performance of porous asphalt mixes is studied through permeability and air voids [3]. the aspects of strength and durability of these mixes are related to the gradation [3 ]. the aggregate gradation adopted is known to have its influence on mobility, ability to be compacted, stiffness, durability and permeability of the asphalt mixes [4]. mixes with ensured continuity of gradation were observed to have better rutting resistance [6]. apart from the gradation chosen for the mix, the nominal maximum particle size (nmps) is also observed to have an influence on pavement performance. with the increase of nmps, the rutting resistance is observed to improve while mixes with smaller nmps, they found to exhibit better raveling resistance [7]. therefore it can be stated that the gradation of the mix should be given utmost importance and the method of grade selection should be rational to ensure satisfactory pavement performance. several methods were proposed by researchers for assessment of gradation requirements of layers in flexible pavements [8]. since it is not practically possible to sieve the aggregates in the field, the target gradation is attained by blending aggregate stockpiles from one or more quarry sites in a specific proportion called job mix formula. the methods proposed for blending includes graphical method, triangular chart method & asphalt institute graphical method and optimization methods [8]. graphical methods cannot be used when number of stockpiles is more than 3. as of today, there are no tools available for integrating different philosophies of gradation selection and hence there is a potential for research in this area. 2. objective of the research in view of the research gap identified, the objective of the present research is to develop an integrated software tool using c# programming for job mix formulation (jmf) of grade 2 of bituminous concrete mix wearing course. morth specifications, bailey’s parameters, 0.45 power gradation chart and cost per unit of the stock pile have been simultaneously considered to workout different combinations of aggregate blends. 3. problem formulation the blends are to be filtered as per three philosophies mentioned previously. the objective function is . 1 min c p cp m i j j j    (1) pp jmfupper i (2) pplowerjmfi  (3) 1 1 p m j j   (4) the jmf for a specified blend is worked out as proceedings of engineering and technology innovation, vol. 12, 2019, pp. 26-34 28 1 1 jmf p pp i to n m i j ij j     (5) ci unit material cost of the ith blend m number of designated stock piles in the mix pj proportion of aggregates of the jthstockpile cpj cost per unit of jth stockpile jmfi job mix formula for ith particle n number of designated particles in the mix ppij percent passing of the ith particle for jth stock pile obtained from sieve analysis pplower recommended lower percent passing ppupper recommended higher percent passing all particles in the mix should satisfy the condition specified in the eqs. (2) and (3) above for a successful trial blend. apart from the requirements of gradation, bailey’s method addresses the issues related to the ability of the mix to get compacted, segregation and continuity of gradation. the method is not very popular, due to the lack of extensive field verification of the pavement stretches designed using this method [13]. in this method, guidelines have been laid down through three parameters viz. coarse aggregate ratio (ca ratio), coarser portions of the fine aggregate ratio (fac) and finer portions of fine aggregate ratio (faf). these ratios will be calculated based on nmps of the mix, which is one sieve greater than the sieve on which at least 10% of the particles are retained as per trb circular e-c044 [13]. the three ratios specified are linked to half, primary, secondary and tertiary control sieves so as to distinguish the coarser and finer portions of the mix. 4. inputs for determination of jmf in order to find the proportion of aggregate blends for a specified type of layer and hence jmf for a stated nmps, it is required to collect samples from the stockpiles of quarry sites and perform the sieve analysis. the percent passing of each of the particle sizes is to be obtained in sieve analysis and will become input for determination of the jmf. the maximum density line is plotted using eq. (6). where size i size of ith particle in the mix sizemsa maximum size aggregate 5. need for a computational tool the spreadsheets are now being used to find a blend which satisfies the grading requirements of the specified layer fulfilling eqs. (2) and (3) for all the particles in the mix. this is a trial and error method with several proportions satisfying the grading requirements. in order to generate all those proportions which fulfilled the requirements of gradation and to address the issues of compaction, segregation and cost per unit of the blend, excel spreadsheet has its own limitations. to accomplish this, an interactive software interface is developed by using the c# (c-sharp) language. c# is a very useful tool which supports the use of forms in windows environment with graphic user interface (gui) as front end, the user can input data and it is possible to manipulate the data with suitable computations, manage and export data *100 0.45 pp size i i size msa          (6) proceedings of engineering and technology innovation, vol. 12, 2019, pp. 26-34 29 files in the back end using object oriented programming (oop) concept. thus c# programming has been chosen to develop the interactive gui software for the assessment of jmf in the current research. the software requires microsoft net 4.5 framework or higher version for execution. 6. need for a computational tool in order to select a suitable blend, the following criteria have been adopted and embedded in the software. (i) the proportions generated by the software will be examined for the morth requirement. (ii) bailey parameters will be evaluated for the percent passing of each blend obtained. the parameters will be worked out as per the following formulae. where ca ratio coarse aggregate ratio pp half sieve percent passing half sieve pcs primary control sieve = 0.22 x nmps scs secondary control sieve = 0.22 x pcs tcs tertiary control sieve = 0.22 x scs pp pcs percent passing primary control sieve pp scs percent passing secondary control sieve pp tcs percent passing tertiary control sieve the factor 0.22 had arrived after considering the analysis of 2-d and 3-d packing of different shaped particles. 2-d analysis has shown the particle diameter ratio equal to 0.155 (all round) and 0.289 (all flat) with an average of 0.22 for angular and subangular particles [12]. these parameters will be computed from percent passing of each blend generated in the process. (iii) for each of these proportions generated by the software, the level of gradation closer to the maximum density line is evaluated by the statistical parameter root mean square error (rmse) [5]. where ppmax density percent passing corresponding to the i st rticle for the maximum density line in the mix pp jmf percent passing corresponding to the i st rticle in the trial blend i index for the particle in the mix n total number of particle sizes in the mix the blend of aggregates which fulfill the requirements of morth and bailey with the least of rmse and with the least cost per unit of the blend will be recommended for final implementation. in the present research work, the nominal maximum particle size was chosen as 12.5 mm. the ca ratio between 0.50 and 0.65 and other two ratios in the range 0.35 to 0.50 have been chosen, as recommended in bailey’s specifications. 100 pp pp pp ca ratio pcshalf sieve half sieve                (7) pp fa ratio pp scs c pcs  (8) pp fa ratio pp tcs f scs  (9) ( ) 2max. 1 pp pp rmse n n ijmfdensity i     (10) proceedings of engineering and technology innovation, vol. 12, 2019, pp. 26-34 30 (1) ca ratio: the three ratios discussed above are chosen based on nmps. it was observed that the mixes with lower values than recommended ca ratio, and are subjected to segregation and are generally gap-graded mixes, while the mixes with higher values of ca ratio will have issues related to continuity of gradation and difficulty in compaction. (2) fac ratio: while the mixes with lower values of fac ratios will have non-uniform gradation and will have problems of compaction, the higher values will indicate the presence of excessive amounts of fine aggregate leading to tender mixes. the mixes with higher values of fac ratio will show a hump in the 4.75 mm region and below when plotted on a 0.45 power gradation chart. (3) faf ratio: this ratio explains how fine aggregates pack together and influence the voids in the mix. typically, dense graded mixes will have faf less than 0.50. also, vma increases with the decrease in this ratio. 7. need for a computational tool the flow chart for the process incorporating the above said criteria is presented in fig. 1. fig. 1 the process flow chart 7.1. sample source code the sample source code is presented below for the control ‘find jmf’ button. private void button3_click_1(object sender, eventargs e) { //find jmf button string str1, str2, str3, str4, str5; for (j = 0; j < 10; j++) { proceedings of engineering and technology innovation, vol. 12, 2019, pp. 26-34 31 str1 = datagridview2.rows[j].cells[1].value.tostring(); ppsize1[j] = double.parse(str1); } for (j = 0; j < 10; j++) { str2 = datagridview2.rows[j].cells[2].value.tostring(); ppsize2[j] = double.parse(str2); } for (j = 0; j < 10; j++) { str3 = datagridview2.rows[j].cells[3].value.tostring(); ppsize3[j] = double.parse(str3); if (noofcoarsepiles == 2) ppsize3[j] = 0; } for (j = 0; j < 10; j++) { str3 = datagridview2.rows[j].cells[4].value.tostring(); ppsize4[j] = double.parse(str3); } for (j = 0; j < 10; j++) { str5 = datagridview2.rows[j].cells[5].value.tostring(); mf[j] = double.parse(str5); } //location of maximum density line for (i = 0; i < 10; i++) { strtemp = datagridview2.rows[i].cells[0].value.tostring(); size[i] = double.parse(strtemp); ssr[i] = math.pow(size[i], 0.45); if (i > 0) maxden[i] = 100 * math.pow(size[i] / size[0], 0.45); else maxden[0] = 100; } int looper = 0, morth = 0; // three stockpiles in coarse aggregates if (noofcoarsepiles == 3) {//outermost loop for (s1 = 1; s1 <= 95; s1++) {//1 tester = 0; for (s2 = 1; s2 <= 95; s2++) {//2 if (98 s1 s2 > 0) {//3 s34 = 98 s1 s2; proceedings of engineering and technology innovation, vol. 12, 2019, pp. 26-34 32 {//4 for (s3 = 1; s3 <= 95; s3++) {//5 bandmchecker = 0; s4 = s34 s3; if (s4 > 0) {//6 if ((s1 + s2 + s3 + s4 + s5) == 100) {//7 tester = 0; newrmssum = 0; jmfsum = 0; bandmchecker = 0; looper = looper + 1; for (i = 0; i < 10; i++) { jmf[i] = 0.01 * (s1 * ppsize1[i] + s2 * ppsize2[i] + s3 * ppsize3[i] + s4 * ppsize4[i] + 2 * mf[i]); jmfsum = jmfsum + jmf[i]; if (jmf[i]<=upper[i]&&jmf[i]>= lower[i]) tester = tester + 1; newrmssum = newrmssum + (jmf[i] maxden[i]) * (jmf[i] maxden[i]); } 8. results of software for local quarry stockpiles sample sieve analysis was performed for all the four aggregate stockpiles, procured from a local quarry. the results of sieve analysis are considered for generating the alternative aggregate blends through the software developed during the present study. the user can edit the data of percent passing and generate the blends. fig. 2 results of the software showing generated blends the results of c# program are presented in fig. 2 for two aggregate stockpiles. in the initial trial, the gradation of dust could generate blends fulfilling only morth gradation. subsequently, change of gradation of dust could give blends to fulfill both morth and bailey criteria; therefore it can be observed that gradation of dust will influence jmf. out of 4656 combinations of possible blends generated by the software, 36 blends had fulfilled morth requirements (as displayed in yellow colour) and 26 blends fulfilled bailey’s requirements. the minimum and maximum rmse values were observed to be proceedings of engineering and technology innovation, vol. 12, 2019, pp. 26-34 33 7.52 and 9.53 respectively. the dependency of bailey’s parameters on aggregate proportions is presented in fig. 3. no significant observations could be made with regard to the variation of other stockpiles with bailey parameters and hence not discussed. the successful jmf of the 3 blends with minimum dust and maximum ca # 1 & ca # 2, maximum dust with minimum ca #1 and maximum dust with minimum ca # 2 fulfilled both morth & bailey parameters with rmse values, 8.68, 10.54 and 10.45 respectively have been chosen from the results and are plotted and presented in fig.4. also, superpave control points were also marked on the plot for verification. even though the restricted zone has lost its significance [4] in superpave gradation, out of research interest, it was also plotted in fig. 3. it can be clearly observed that all the 3 blends with designated % feed have fulfilled superpave control points, while the blend with minimum dust, maximum ca #1 and ca #2 is observed to be the closest to the maximum density line, with least rmse of 8.68 and also with the lowest cost per unit and hence recommended for final implementation. fig. 3 variation of ca # 1 with ca ratio for dust 6 fig. 4 aggregate gradation for selected blends with % feed along with superpave control points fulfilling morth & bailey criteria 9. conclusions this research has presented the development of software for the assessment of jmf for bc grade ii. excel spreadsheets are been widely used to find the desired blend limiting with few successful blends. the software is developed by the c# programming for arriving at the jmf had to fulfill morth & bailey requirements. it is capable of generating blends fulfilling morth and bailey parameters; and will be able to assess the nearness to maximum theoretical density line along with the cost per unit of the blend. a sample source code was also presented. the results of a case study were also discussed wherein the gradation of dust obtained from a local quarry and was observed to fulfill morth requirements with 36 combinations of aggregate blends and 26 combinations to fulfill bailey requirements. the gradation of dust plays a vital role in designing the jmf. from the study, it can also be concluded that the bailey requirements are more stringent than morth and the superpave gradation. the tool developed with c# programming in the .net framework will facilitate the designers to arrive at most optimal blend and will serve in the initial phase of the mix design process fulfilling recommendations of local agencies, bailey parameters and superpave. conflicts of interest the authors declare no conflict of interest. proceedings of engineering and technology innovation, vol. 12, 2019, pp. 26-34 34 references [1] y. h. haung, pavement analysis and design, 2nd ed, pearson education, usa. [2] active standard, “standard test method for marshall stability and flow of asphalt mixtures,” west conshohocken, pa, 2015. [3] t. n. mansour, bradley j. putman, “influence of aggregate gradation on the performance properties of porous asphalt mixtures,” journal of materials civil engineering, vol. 25, no.2, pp. 281-288, 2013. [4] cooley jr, l a. zhang, jianping, kandhal, p. s. hand, a. j. epps martin a, “significance of restricted zone in superpave aggregate gradation specification,” transportation research circular no. e-c043, transportation research board, national research council, washington, d.c., 2001. [5] r. j. hyndman, a. b. koehler, “another look at measures of forecast accuracy,”. international journal of forecasting, vol. 22, no. 4, pp. 679-688, 2006. [6] b. e. ruth et al, “aggregate gradation characterization factors and their relationships to fracture energy and failure strain of asphalt mixtures,” journal of association of asphalt paving technologies. vol. 71, pp. 310-344, 2002. [7] h. liu, p. hao and j. xu, “effects of nominal maximum aggregate size on the performance of stone matrix asphalt,” applied sciences, vol. 7, no. 126, 2017. [8] mario t. tabiucanon, p. adulbhan, stephen and s. y. chen, “a probabilistic programming model for blending aggregates,” applied mathematical modelling, vol. 4, pp. 257-260, 1979. [9] the asphalt institute, “mix design methods for asphalt concrete and other hot-mix types,” ms-2 (4th edn), maryland, 1974. [10] w. r. vavrik, “bailey method for gradation selection in hot-mix asphalt design transportation research circular,” trb circular number e-c044, 2002. [11] specifications for base and surface course, section 500, materials for structures, morth, government of india, 2015. [12] w. r. vavrik, w. j. pine, g. huber, s. h. carpenter and r. bailey, “the bailey method of gradation evaluation: the influence of aggregate gradation and packing characteristics on voids in the mineral aggregate,” journal of the association of asphalt paving technologists, vol. 70, pp. 132-175, 2001. [13] w. r. vavrik, w. j. pine and s. h. carpenter, “aggregate blending for asphalt mix design: the bailey method,” presented at the 81st annual meeting of the transportation research board, washington, d.c., vol. 1789, pp. 146-153, 2002. [14] w. r. vavrik, “quality control of hot-mix asphalt: controlling aggregate interlock through the bailey method,” in proceedings of the 1st international conference on quality control and quality assurance of construction materials, dubai municipality, united arab emirates, 2001. copyright© by the authors. licensee taeti, taiwan. this article is an open access article distributed under the terms and conditions of the creative commons attribution (cc by-nc) license (https://creativecommons.org/licenses/by-nc/4.0/). https://www.sciencedirect.com/science/article/pii/0307904x80901924#! https://www.sciencedirect.com/science/article/pii/0307904x80901924#! 4___peti#8870___27-33 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 27-33 development of non-contact real-time monitoring system for animal body temperature kuo-hsiung tseng 1,* , meng-yun chung 1 , wei-hong jhou 1 , wei-jer chang 2 , chen-han xie 2 1 department of electrical engineering, national taipei university of technology, taiwan 2 morepets company limited by shares, taiwan received 09 november 2021; received in revised form 13 february 2022; accepted 14 february 2022 doi: https://doi.org/10.46604/peti.2022.8870 abstract body temperature is an important indicator of health monitoring. however, since animals are covered with fur, it is difficult to obtain their accurate body temperature with the traditional infrared measurement technology. to deal with this problem, this research proposes a non-contact real-time monitoring system using an infrared method combined with object detection. the system is developed based on general infrared thermal imaging technology and an infrared thermal imaging module with an image tracking algorithm. yolo is used to detect animals, and a thermal imaging camera is used to measure the body surface temperature of animals. the result shows that the proposed system can accurately measure the body temperature of animals without being influenced by animals’ fur. in the future, it can be applied to monitor the body temperature of sick animals in veterinary hospitals. keywords: animal body temperature sensing, internet of things, image detection, infrared thermography 1. introduction infrared thermometers were developed in the mid-1980s [1-2]. with pyroelectric sensors, infrared thermometers can effectively detect the heat radiation emitted from a target surface and measure the temperature of the surface in a non-contact way [3-4]. infrared thermometers have been extensively used for body temperature measurement, but their limitation is that they can only conduct close-distance measurement with a small detection range. infrared thermography is a technique combined with image processing to generate visual images with different colors indicating different temperatures on different parts of the target surface [5-7]. this technique is characterized by remote observation, timeliness, and heat distribution observation. notably, it has been used in animal medical diagnosis for decades, e.g., the diagnosis for mammals. mammals generate heat energy by metabolism to maintain constant body temperature [8]. as their skin surface temperature can reflect the condition of tissue metabolism and blood circulation, the lesion of surface inflammation or circulation abnormality can be found when the body temperature is observed to be abnormal. thus, body temperature is an important index for diagnosing diseases. many studies attempted to use infrared thermometers to measure the body temperature of animals, including cats, pigs, and cattle [9]. however, due to animals’ fur and physiological structures, the probability of error is quite high when using infrared thermometers, making it difficult to replace conventional thermometers with infrared thermometers. furthermore, when using conventional thermometers in veterinary clinics, veterinarians have to insert the thermometers into animals’ anuses to measure the core body temperature, making the animals suffer from severe stress and discomfort and even resulting in injuries of the animals or the veterinarians. moreover, it is impractical to use conventional thermometers for repeated measurement, especially for hospitalized animals that require constant monitoring to check whether there is fever or hypothermia. hence, there is an urgency to develop a non-contact continuous animal temperature monitoring system to meet the existing clinical need [10]. * corresponding author. e-mail address: f10473@mail.ntut.edu.tw tel.: +886-2-27712171#2173; fax: +886-2-27317187 advances in technology innovation, vol. 21, 2022, pp. 27-33 in addition to the study need and clinical need, there is also a commercial need that motivates the researchers to develop a non-contact monitoring system for measuring the body temperature of animals. livestock’s commercial value is highly dependent on the health conditions of livestock. however, there is still no effective approach to monitor the health conditions of livestock because of the limitations in fields, manpower, and equipment. infrared thermography is a possible solution that can monitor animals without direct contact and simultaneously measure the temperature distribution of multiple targets in the field of view [11]. infrared thermography is free from the effect of ambient light source conditions and thus is highly suitable for animal temperature monitoring, aiding the early detection of infectious diseases. this research uses the infrared method combined with object detection to develop a non-contact real-time monitoring system to accurately measure the body temperature of animals, without being influenced by animals’ fur. in the rest of this article, the research method and measurement equipment used are explained in section 2, the results and discussion are presented in section 3, and the concluding remarks are provided in section 4. 2. materials and research method 2.1. experimental construction the experimental process of this study is shown in fig. 1. first, a field (e.g., veterinary hospitals, ranches, or farms) is selected as the site where the experiment is conducted. then, the type of animal subjects is selected, and the non-contact and contact methods are used to measure the body temperature of animal subjects. after that, the obtained temperature data are compared and sent to a veterinarian and the owners of animal subjects. in this study, a veterinary hospital is chosen as the experimental field, and an adult pet dog is selected as the animal subject. this study performs target detection for image recognition in the early stage [12]. a visible-light/infrared twin-lens module is used for detection, and is combined with an image tracking algorithm for animal temperature monitoring with target tracking and positioning functions. two groups are set up in this study: a control group (the traditional body temperature measurement conducted by a veterinarian) and an experimental group (the non-contact infrared measurement using the proposed system). an adult pet dog is used for both groups. in the early setup, the compensation for fur interference cannot be regulated accurately, and the module’s precision is adjusted by comparing the data of contact measurement. it must be noted that, unlike human beings, animals cannot tell their conditions. thus, in this study, the dog subject’ body temperature is initially measured to examine and determine fever or other symptoms. subsequently, the measured body temperature is transferred via the internet of things (iot) to the computer of the verification domain as the criterion of diagnosis and treatment for the veterinarian. in addition, as the owner of the dog subject is concerned about the status of the dog, instant information is also sent via iot to the owner’s mobile phone for first-hand updates on the dog’s condition at the hospital [13]. a temperature and humidity sensor is used to detect the dog subject’s internal temperature and humidity, determining whether the subject’s condition is influenced by external factors. fig. 1 experimental process field selection pet selection non-contact temperature measurement contact temperature measurement data comparison send message veterinarian owner pet´s state 28 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 27-33 an flir c5 thermal imaging camera is used in this study to measure the body temperature of the dog, as shown in fig. 2. the specification of the camera is shown in table 1. the camera has a size of 135 × 85 × 25 mm and is only 0.19 kg, which is suitable for conducting the measurement. it can store more than 5000 images by its internal memory and onboard flir ignite cloud connectivity with wi-fi. the infrared sensor provides 160 × 120 (19,200 pixels) images, and can be used under the ambient temperature of -10 to 50°c to measure the subjects having the temperature of -20 to 400°c. the flir camera is a long wave infrared (lwir) camera and receives infrared radiation with a wavelength of 8-14 µ m. under the atmospheric window, water vapor and carbon dioxide have weak absorption of the radiation. this facilitates the infrared body temperature measurement. with the ambient temperature of 15 to 35°c, the accuracy of measuring the subjects with 0 to 100°c is ±3°c, and the accuracy of measuring the subjects with 100 to 400°c is ±3%. parameters such as ambient temperature, ambient humidity, and the emissivity of the subjects can be adjusted. the lightness and portability of the camera make it easy to set up experiments. before each experiment, the environmental parameters (temperature and humidity) of the camera are adjusted according to the indoor temperature and humidity count value. the emissivity of the animal fur can also be set in the range of 0.98-1. fig. 2 flir c5 infrared thermal imaging camera table 1 specification of an flir c5 thermal imaging camera model c5 camera size (l × w × h) 138 × 84 × 24 mm infrared ray (ir) sensor 160 × 120 (19,200 pixels) operating temperature range -10 to 50°c (14 to 122°f) object temperature range -20 to 400°c (-4 to 752°f) spectral range 8 to 14 µ m accuracy at ambient temp. 15 to 35°c (59 to 95°f): 0 to 100°c (32 to 212°f): ±3°c (±5.5°f) 100 to 400°c (212 to 752°f): ±3% company flir country estonia 2.2. infrared radiation in spectra, the wavelength exceeding 0.75 µm is called infrared ray (ir). from the wavelength 0.75 µm to the longest wave band, ir can extend to 1000 µm. the wavelength range covered in this section can be divided into three according to the application domains: 0.75-2.5 µm (near-ir), 2.5-25 µm (mid-ir), and above 25 µm (far-ir). 2.2.1. planck’s law planck’s law, also known as planck’s black body radiation law, is a mathematical relationship proposed by a german physicist max planck in 1990. it refers to the relationship between the radiance and frequency of electromagnetic radiation emitted from a black body at an arbitrary temperature t [14]. 2.2.2. wien’s displacement law wien’s law, also known as wien’s displacement law, was proposed by wilhelm wien. it is a law describing the inverse relationship between the peak wavelength of spectral radiosity of a black body’s electromagnetic radiation and its temperature in physics [15]. the relationship is expressed as eq. (1). wien’s displacement law also describes that the hotter an object is, the shorter the radiation wavelength or the higher the radiation frequency is. 29 advances in technology innovation, vol. 21, 2022, pp. 27-33 max b λ t = (1) where λmax is the peak wavelength of radiation (unit: m), and t is the absolute temperature of the black body (unit: k). b is the proportional constant, called wien’s displacement constant; the value is 2.897 7729 × 10 –3 m· k. 2.2.3. stefan-boltzmann law the stefan-boltzmann law, also known as stefan’s law, means that the total energy radiated from the surface of a black body per unit area in unit time (radiosity or energy flux density of object) j* is proportional to the fourth power of thermodynamic temperature t (absolute temperature) of the black body [16]. the relationship is expressed in eq. (2). * 4 j σt= ε (2) where the radiosity j* has the dimension of power density given in eq. (3). according to the international system of units, the standard unit of the radiosity j* is j/(s × m 2 ), i.e., w/m 2 . 2 energy time distance× (3) the standard unit of absolute temperature t is k. ϵ is the radiation coefficient of the black body; if the black body is an absolute black body, then ϵ = 1. the scale factor σ is called stefan-boltzmann’s constant or stefan’s constant. given that it can be worked out of the known basic physical constants of the nature, it is not considered a basic physical constant. the value of the constant is expressed as eq. (4). 5 4 8 1 2 4 2 3 2π k σ 5.670400 10 js m k 15c h − − − −= = × (4) where k is the value of boltzmann’s constant 1.38064852 × 10 -23 j/k, h is the planck’s constant 6.62607015 × 10 −34 (j· s), and c is the light velocity in a vacuum with the value of 3.00 ×10 8 m/s. 2.3. object detection the traditional thermal imaging equipment mostly performs data acquisition for a single object. thus, the information analysis and processing are local and have low volume computation. to the extent of the information of the whole space, the obtained thermal image information will expand greatly, and the demand for computing resources will grow sharply, becoming the system load [17]. without an intelligent zed improvement method, the system will fail. therefore, this project uses the visible-light/infrared twin-lens module and visible light features for recognizing image objects. meanwhile, target locking is established automatically to reduce the range of image information analysis to improve the excessive operation [18]. this study uses yolo object detection technology, where the image input to the prediction result of yolo only depends on one convolutional neural network (cnn). multiple bounding boxes are predicted simultaneously by using cnn, and the probability of an object is calculated for each box. the object is recognized in the local region compared to the sliding window and region proposals [19]. yolo aims at the full image in the training and operational phases, paving the way for a better detection effect on the background. the background error detection rate of yolo is only half of that of a fast region-based convolutional neural network (fast r-cnn). further, yolov4 has better recognition speed and accuracy and less hardware requirement than yolov3 [20]. this study employs the jetson nano developer edition for development. this edition is characterized by powerful computation capability, lightness, easy disposition, and applicability to mobile equipment. 30 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 27-33 3. results and discussion this study employs one adult dog as the animal subject for non-contact and contact temperature measurements (the contact one is the control group and the non-contact one is the experimental group). for the control group, the body temperature of the adult dog is measured using the rectal thermometer. as can be seen in table 2, the dog’s body temperature got from the rectal thermometer is higher than the normal human body temperature which is in the range of 36-37°c. for the experimental group, the flir c5 thermal imaging camera is used. the eye and abdomen temperatures of the dog are measured at a fixed time in the morning from january 26 2021 to february 02 2021. as shown in figs. 3-5, the square frame is the measuring area of a thermal imager, and the white point is the center point of the measuring area. the temperatures at the red and blue points are the maximum and minimum temperatures in the measuring area, respectively. in fig. 3, it is observed that the density of fur distribution can influence the temperatures measured. the red point is the area with sparse fur and the blue point is the area with fur. the maximum temperature measured in the abdomen area occurs in the region covered with sparse fur. meanwhile, in fig. 4, the region having the maximum temperature measured on the face is the eye. the eye temperature is different from the abdomen temperature. it can be seen that the difference in fur density affects the value of temperature, as shown in fig. 5. this is the reason why this study applies the infrared temperature measurement to animals. in addition to the material emissivity, the fur density also causes measurement difficulties and errors. as shown in fig. 6, it is observed that the eye temperature is approximate to the actual body temperature, and the side temperature is a little lower than the actual body temperature. table 2 result of the contact temperature measurement date temperature (°c) 2021/1/26 39.4 2021/1/27 39.0 2021/1/28 38.6 2021/1/29 38.7 2021/1/30 38.7 2021/1/31 38.6 2021/2/1 38.7 2021/2/2 39.0 fig. 3 flank temperature fig. 4 eye temperature fig. 5 effect of fur density 24 26 28 30 32 34 36 38 2021/1/26 2021/1/27 2021/1/28 2021/1/29 2021/1/30 2021/1/31 2021/2/1 2021/2/2 t e m p e ra tu re (˚ c ) date eye temperature(°c) abdomen temperature(°c) fig. 6 result of the non-contact temperature measurement 31 advances in technology innovation, vol. 21, 2022, pp. 27-33 4. conclusions in asian countries (e.g., taiwan), the number of dogs and cats could reach 3 million in 2020, exceeding the number of children below the age of 15. with the increasing number of dogs and cats, the number of animal hospitals has also been increased to more than 1,700. to ensure that the aging pets receive adequate medical care, the pet owners pay close attention to their pets’ medical service quality. since the intelligentized monitoring systems can save the labor cost for animal hospistals and thus reduce the medical payments for pet owners, admitting such intelligentized monitoring systems will be the future trend. in clinical scenarios, the body temperature of hopitalized animals need to be continuously monitored by veterinarians. however, when veterinarians employ traditional thermometers to measure the temperature through animals’ anuses, the animals are severely discomforted and the veterinarians may be injured if the animals defend themselves. to deal with this problem, a non-contact continuous animal temperature monitoring system is desired. this study combines non-contact infrared measurement with object detection, and uses an optical lens to conduct the measurement. specific regions can be extracted by the yolo algorithm to reduce excessive operation information. the temperatures of the subjects with specific positions can be captured by infrared thermal imaging technology and instantly conveyed to veterinarians. with the proposed system, the risk and inconvenience induced by contact measurement can be avoided. acknowledgements the authors would like to thank the ministry of science and technology (most 109-2622-e-027-033-) for the financial support of this research. also, the authors would like to thank mr. jue-hao chen (department of electrical engineering, national taipei university of technology, taiwan) and yuan lin animal hospital for their kind assistance. conflicts of interest the authors declare that there are no conflicts of interest regarding the publication of this study. statement of ethical approval all procedures performed in studies involving animals were in accordance with the ethical standards of the institution or practice at which the studies were conducted. references [1] j. l. hatfield, “measuring plant stress with an infrared thermometer,” hortscience, vol. 25, no. 12, pp. 1535-1538, december 1990. [2] e. j. sadler, c. r. camp, d. e. evans, and j. a. millen, “corn canopy temperatures measured with a moving infrared thermometer array,” transactions of the american society of agricultural engineers, vol. 45, no. 3, pp. 581-591, 2002. [3] t. matsukawa, m. ozaki, t. nishiyama, m. imamura, and t. kumazawa, “comparison of infrared thermometer with thermocouple for monitoring skin temperature,” critical care medicine, vol. 28, no. 2, pp. 532-536, february 2000. [4] s. abbasbandy, “homotopy analysis method for heat radiation equations,” international communications in heat and mass transfer, vol. 34, no. 3, pp. 380-387, march 2007. [5] c. a. balaras and a. a. argiriou, “infrared thermography for building diagnostics,” energy and buildings, vol. 34, no. 2, pp. 171-183, february 2002. [6] b. b. lahiri, s. bagavathiappan, t. jayakumar, and j. philip, “medical applications of infrared thermography: a review,” infrared physics and technology, vol. 55, no. 4, pp. 221-235, july 2012. [7] s. bagavathiappan, b. b. lahiri, t. saravanan, j. philip, and t. jayakumar, “infrared thermography for condition monitoring—a review,” infrared physics and technology, vol. 60, pp. 35-55, september 2013. [8] l. s. peck, m. s. clark, s. a. morley, a. massey, and h. rossetti, “animal temperature limits and ecological relevance: effects of size, activity, and rates of change,” functional ecology, vol. 23, no. 2, pp. 248-256, march 2009. 32 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 27-33 [9] d. m. newsom, g. l. bolgos, l. colby, and j. a. nemzek, “comparison of body surface temperature measurement and conventional methods for measuring temperature in the mouse,” journal of the american association for laboratory animal science, vol. 43, no. 5, pp. 13-18, september 2004. [10] g. j. tattersall and v. cadena, “insights into animal temperature adaptations revealed through thermal imaging,” the imaging science journal, vol. 58, no. 5, pp. 261-268, july 2010. [11] p. a. warn, m. w. brampton, a. sharp, g. morrissey, n. steel, d. w. denning, et al., “infrared body temperature measurement of mice as an early predictor of death in experimental fungal infections,” laboratory animals, vol. 37, no. 2, pp. 126-131, april 2003. [12] n. a. othman and i. aydin, “a new deep learning application based on movidius ncs for embedded object detection and recognition,” 2nd international symposium on multidisciplinary studies and innovative technologies, pp. 1-5, october 2018. [13] t. h. chung, w. s. jung, e. h. nam, j. h. kim, s. h. park, and c. y. hwang, “comparison of rectal and infrared thermometry for obtaining body temperature of gnotobiotic piglets in conventional portable germ free facility,” asian-australasian journal of animal sciences, vol. 23, no. 10, pp. 1364-1368, october 2010. [14] j. c. cuevas and f. j. garcía-vidal, “radiative heat transfer,” acs photonics, vol. 5, no. 10, pp. 3896-3915, september 2018. [15] j. k. baird and t. r. king, “a wien displacement law for impact radiation,” international journal of impact engineering, vol. 23, no. 1, pp. 39-49, december 1999. [16] w. shi, j. cao, q. zhang, y. li, and l. xu, “edge computing: vision and challenges,” ieee internet of things journal, vol. 3, no. 5, pp. 637-646, october 2016. [17] g. montambaux, “generalized stefan-boltzmann law,” foundations of physics, vol. 48, no. 4, pp. 395-410, march 2018. [18] q. zhu, h. zheng, y. wang, y. cao, and s. guo, “study on the evaluation method of sound phase cloud maps based on an improved yolov4 algorithm,” sensors, vol. 20, no. 15, article no. 4314, august 2020. [19] j. m. zurada, introduction to artificial neural systems, saint paul: west publishing, 1992. [20] a. bochkovskiy, c. y. wang, and h. y. m. liao, “yolov4: optimal speed and accuracy of object detection,” https://arxiv.org/pdf/2004.10934v1.pdf, april 23, 2020. copyright© by the authors. licensee taeti, taiwan. this article is an open access article distributed under the terms and conditions of the creative commons attribution (cc by-nc) license (https://creativecommons.org/licenses/by-nc/4.0/). 33 6___peti#8695___50-59 proceedings of engineering and technology innovation, vol. 22, 2022, pp. 50-59 effect of gear materials on the surface contact strength of spur gears eniyew eskezia * , misganaw abebaw department of mechanical engineering, institute of technology, university of gondar, gondar, ethiopia received 14 october 2021; received in revised form 21 january 2022; accepted 22 january 2022 doi: https://doi.org/10.46604/peti.2022.8695 abstract in gear applications, the initiation of cracks at or near the contact surfaces of gear mates occurs mostly due to the surface contact mode of gear failure. this study aims to investigate the influence of gear materials on the contact strength of spur gears. in this study, four different gear materials are selected and the contact stress on spur gear mates is analyzed. hertz’s contact stress equation and ansys 16 are used for the theoretical analysis and finite element (fe) method, respectively. the results of hertz’s equation are compared with the results of ansys 16. the results show that the contact stress on spur gear mates varies when different gear materials are used in both methods. this indicates that the surface contact strength of spur gears is greatly influenced by the type of gear materials. keywords: spur gear, contact stress, analytical method, hertz’s equation, finite element method 1. introduction gears are the basic machine elements in a mechanical power transmission system and most industrial rotating machines. humans have used them for different applications for several years. however, some problems still exist. for instance, gear tooth surface damage, which is caused by wear and surface fatigue, can result in the change of material properties in different working conditions [1-3]. scholars have investigated the solution to the problems by using various methods. among them, theoretical and numerical methods have been used frequently due to low research expenses. contact stress is the main cause of gear failure. it initiates the crack propagation around the contact surfaces of gear mates, and stress concentration may develop based on it [2, 4-5]. to increase gears’ load-carrying capacity and component reliability, it is important to improve the surface strength of gears. surface strength is the factor that determines the gear life rather than bending strength due to the large enough teeth profile of gears. different techniques were adopted by researchers to increase the working ability, efficiency, and working life of gears. furthermore, the surface life of gears can be improved by subjecting the gear materials to different surface treatment processes. gear parameters also influence the gear performance greatly [3-6]. mechanical properties of a material are the properties that encompass a response to an applied load. they reveal the elastic and plastic behavior of the material upon applied forces. these properties are used to classify and identify metals. strength, ductility, hardness, toughness, impact strength, etc. are some mechanical properties of metals. strength is the ability of metals to withstand deformation due to external loads, whereas ductility is the tendency of metals to be drawn into wire before fracturing under tensile loads [4-5]. the main purpose of this study is to investigate the influence of gear materials on the contact strength of spur gears by using the theoretical method (hertz’s contact stress equation) and finite element (fe) method (ansys workbench). both gear designers and manufacturers could take advantages of the findings in this study. * corresponding author. e-mail address: enyew.eskezia@uog.edu.et tel.: +251-9-12802371; fax: +251-058-1141240 proceedings of engineering and technology innovation, vol. 22, 2022, pp. 50-59 2. materials, methods, and conditions the contact stress on spur gears is the main concern in the research field. the analytical, experimental, and fe methods are the basic tools to compute and predict the contact stress on gears. any method out of the three can support the other one for validating the results [3]. 2.1. analytical methods and conditions gear failure is categorized into two general groups: bending and contact failure. bending failure occurs in the root of gear teeth, whereas contact failure occurs on the gear surface when the gear strength is inadequate. the bending stress can be calculated by the lewis formula, whereas the hertz’s equation is used for the contact stress. the design procedures developed by american gear manufacturing association (agma) and international organization for standard (iso) were also used by the researchers to design spur gears for different applications [5-8]. in machine elements, which work under pressure loading and have rolling-sliding motion, the value of contact stress depends on the size of contact areas. the contact stress can be calculated based on elastic theory. the damage of gear teeth due to the contact stress commonly occurs either on the pitch line (only rolling movement) or away from the pitch line (both sliding and rolling movements) [9]. 2.1.1. contact problem modeling using parallel cylinders the contact problem modeling is created by using parallel cylinders in order to apply the hertz’s equation to spur gear mates. the mating gears are considered equivalent to the contact cylinders with the same radius of curvature at contact point [9]. fig. 1 below illustrates the contact modeling. as shown in fig. 2 below, the shape of the contact area of the two cylinders is rectangular (2b × l). however, the stress is distributed along the width (2b) with an elliptical shape. the maximum pressure on this contact area occurs along the midway of the width (b). the value of this maximum pressure is calculated by eq. (1). (a) cylindrical contact in compression [10] (b) contacted spur gear teeth [11] fig. 1 contact problem modeling using parallel cylinders fig. 2 ellipsoidal-prism pressure distribution [9] 51 proceedings of engineering and technology innovation, vol. 22, 2022, pp. 50-59 from the above figure, the maximum pressure, pmax, can be formulated as: 2 max p f bl= π (1) where f is the total contact force. half of the contact width of the pressure distribution can be obtained by: 4 e e fr b le = π (2) where ee is the equivalent modulus of gear material and re is the equivalent radius of curvature. 1 2 1 1 1 e e e e = + (3) 1 2 1 1 1 e r r r = + (4) at pitch point, r1 = rp and r2 = rg. this is given by: sin 2 p p r d= α (5) sin 2 g g r d= α (6) where d refers to pitch diameter, p refers to pinion, g refers to gear, and α refers to pressure angle. the equivalent radius will be: sin 1 1 1 1 2 1 1 ( ) e p g p g r r r d d = = + + α (7) the tangential and normal force components (ft and fn) and the torque (tp) can be related to the pinion and gear. cos cos pt n p tf f f r = = = α α (8) substituting eqs. (2), (7), and (8) into eq. (1), the maximum contact pressure or hertz stress can be obtained. cos p e max e p t e p lr r = απ (9) in the gear contact kinematics, there are three steps as identified in fig. 3. in the beginning, the contact occurs through a combination of rolling and sliding movements between the teeth. then, the sliding movement has the frictional effect, and there is simple rolling contact in the pitch diameter region. after this point, both sliding and rolling movements will occur again [12]. in this study, the authors have considered some assumptions. the materials in contact are considered homogeneous. the load which is normal to the contact tangent plane causes the contact stress. in other words, there are no tangential forces between the surfaces as the contact area is considered very small. thus, the frictional effect is totally neglected. 52 proceedings of engineering and technology innovation, vol. 22, 2022, pp. 50-59 fig. 3 mechanics of gear tooth contact [12] 2.1.2. material requirements the crucial aspect to consider in gear engineering and design is the selection of gear materials. strength, durability, and cost (both material and manufacturing cost) are the most important factors when selecting the gear materials. the purpose of material selection is to find the right combination of physical properties that satisfy the requirements of certain applications at the lowest cost. depending on the shape, nature, and final application of gears, designers and manufacturers choose one from an unlimited number of gear materials to create the finished product. the most commonly used materials include many types of steel, brass, bronze, cast iron, ductile iron, aluminum alloys, powder metals, and plastics [13-14]. table 1 below displays the mechanical properties of four engineering gear materials selected for this study. table 1 properties of gear materials selected for this study selected gear materials modulus of elasticity (gpa) poisson’s ratio density (kg/m 3 ) case-hardened wrought steel 206 0.3 7800 astm class 35 cast iron 114 0.29 7150 gray cast iron (gg-30) 91 0.25 7050 beryllium-aluminum alloy (albemet 162) 179 0.165 2070 2.1.3. geometrical design of spur gear mesh fig. 4 below shows the standard meshing spur gears. the standard meshing spur gears are the reference circles of two gears that contact and roll each other. based on some set values, all the geometrical specifications of the gears are calculated [15]. in this study, the gear and pinion are considered to have identical geometrical parameters which are presented in table 2. the input parameters of the mating gear pair (gear train) are specified in table 3. fig. 4 geometry of standard meshing spur gears [15] table 2 specification of the meshing spur gears no. description of gear geometry symbol formula value 1 module m set value 2 mm 2 number of teeth z 30 3 pressure angle α 20° 4 face width b 12 mm 5 pitch circle diameter d � � � � � 60 mm 6 base circle diameter db �� � � � �� 56.38 mm (a) point of first contact (b) pitch point contact (c) point of last contact a direction of roll direction of frictional forces 53 proceedings of engineering and technology innovation, vol. 22, 2022, pp. 50-59 table 2 specification of the meshing spur gears (continued) no. description of gear geometry symbol formula value 7 outside circle diameter do �� � � � (� + 2) 64 mm 8 addendum ha ℎ� � � 2 mm 9 center distance a � � 1 2 (� + �) � � 60 mm 10 pitch co-efficient re �� � 1 2 �� ( 1 �� + 1 �� ) 5.13 table 3 input parameters for the gear train 2.1.4. model calculation of contact stress analysis by hertz’s equation now, let’s take grey cast iron (gg-30) material for both gears. by using eq. (10), the material co-efficient (equivalent modulus) can be obtained as shown in eq. (11). 2 2 1 2 1 1 1 ee v v e e = − − + (10) 2 2 4 4 1 48513.3 mpa 1 0.25 1 0.25 9.1 10 9.1 10 ee = =− − + × × (11) the hertz contact stress can be calculated using eq. (9). 955000 48513.3 2258.03 mpa 20 30 5.13 cos 20 h max p × = = = × × × × σ π (12) 0.56 2258.03 1264.41 mpa vonmiss = × =σ (13) by using the same approach, the maximum hertzian contact stresses of the other gear materials selected for this study are calculated. the maximum contact stresses calculated based on the theoretical analysis (hertz’s equation) are presented in table 4. table 4 maximum theoretical contact stress for the selected gear materials by hertz’s equation selected gear materials maximum contact stress by hertz’s equation (mpa) case-hardened wrought steel 1194.56 astm class 35 cast iron 1177.75 grey cast iron (gg-30) 1264.41 beryllium-aluminum alloy (albemet 162) 1325.82 2.2. finite element method the fe method is a numerical technique used for solving the physical problems encountered in engineering and mathematical modeling. finite element analysis (fea) is the simulation of any given physical phenomenon by applying the fe method. fea is used to mathematically model the physical phenomenon and solve the complex structural, thermal transfer, fluid flow, and no. input parameters value 1 nominal input power (p) 10 kw 2 pinion speed (n) 100 rpm 3 nominal torque on the pinion shaft �� � 9550 � � � 955 nm 54 proceedings of engineering and technology innovation, vol. 22, 2022, pp. 50-59 multi-physics problems. engineers use fea to reduce the number of experiments and/or physical prototypes and optimize the machine elements in their design phase for faster and better product developments while saving expenses. fea uses mesh generation techniques to divide a complex physical problem into finite small elements coupled with the fe algorithm [16-19]. many contact problems of rotational machine components can be solved by using the fe method. ansys workbench, one of the frequently used fe software, provides a computerized approach for predicting the response of the object to the applied loads [6, 19]. in the present study, ansys 16.0 workbench package is used to determine the maximum allowable contact stress on spur gears. the steps in ansys 16 workbench solution procedures are explained in the following sub-sections. the fe results will be compared and discussed with the theoretical results in section 3. 2.2.1. constructing the geometric model of the matting spur gears the geometric model of the gear teeth profile is constructed by using the solidworks modeling software. the geometrical data of the gears are taken from table 2. the generated model is saved in the ansys supported geometry file formats such as iges and acis. fig. 5 shows a three-dimensional (3d) geometrical model of a single spur gear done by the solidworks software. fig. 5 three-dimensional model of a spur gear by using solidworks 2.2.2. importing the model to ansys workbench the generated gear model is imported to the ansys workbench interface for the analysis purpose. all the dimensions and geometries of the gear mesh are calculated and presented in table 2 above. the imported model of assembled gears is presented in fig. 6 below. fig. 6 assembled spur gear set 2.2.3. specifying the material properties engineering materials may be specified after the geometry is imported. the specific properties of gear materials under analysis, such as young’s modulus, poisson’s ratio, and density, are used for fea. in the ansys workbench interface, the engineering data can be selected from the analysis tab and the corresponding values for each material can be inserted. 2.2.4. conducting the mesh refinement meshing is the process of dividing the entire model into small cells so that at every cell the equations are solved. it gives an accurate solution and improves the quality of the solution. mesh refinement is conducted in order to generate fine mesh elements. fig. 7 shows the mesh model of the gear mates. 55 proceedings of engineering and technology innovation, vol. 22, 2022, pp. 50-59 (a) refined mesh of gear assembly (b) close-up view of tooth flanks fig. 7 meshed assembly in ansys workbench simulation 2.2.5. setting the boundary condition boundary condition refers to the external load on the border of the model structure. this is normally done by considering that the tooth contact is at the pitch point as illustrated in fig. 3(b). in the lower gear, the fixed support boundary condition is introduced. besides, the other gear is subjected to frictionless support and torque/moment in a clockwise direction as illustrated in fig. 8 below. fig. 8 setting the boundary condition 3. results and discussion 3.1. results the von-mises stress at the contact region for the contact model of mating spur gears is displayed on the solution processor of ansys 16 workbench. the images below (figs. 9-12) show the values for four different materials considered in this study. (a) equivalent stress for case-hardened wrought steel (b) close-up view of tooth fig. 9 von-mises stress for case-hardened wrought steel 56 proceedings of engineering and technology innovation, vol. 22, 2022, pp. 50-59 (a) equivalent stress for grey cast iron gg-30 (b) close-up view of tooth fig. 10 von-mises stress for grey cast iron gg-30 (a) equivalent stress for astm class 35 cast iron (b) close-up view of tooth fig. 11 von-mises stress for astm class 35 cast iron (a) equivalent stress for albemet 162 (b) close-up view at tooth fig. 12 von-mises stress for beryllium-aluminum alloy (albemet 162) 3.2. discussion the main aim of this study is to analyze the change of the contact stress on spur gears with different gear materials. the analysis is done by varying the gear materials under constant loading conditions and geometric parameters by using the hertz theory and fe methods (ansys 16). the results of the analysis by these two methods are compared and presented in table 5. the selection of appropriate gear materials for different applications is the key task in gear design and/or manufacturing processes. the use of these different gear materials provides a range of contact stress which is useful in material selection [9]. as illustrated in table 4, the results of contact stress in spur gear mates using the theoretical (hertz’s equation) and fe method (ansys 16) agree with each other with a maximum error of 2.9%. this consequently indicates that the fe simulation (ansys workbench) is compatible and almost identical with the theoretical analysis (hertz’s equation) for the materials under analysis. this fact is demonstrated in fig. 13. 57 proceedings of engineering and technology innovation, vol. 22, 2022, pp. 50-59 table 5 comparison of maximum contact stress for the selected gear materials selected gear materials maximum contact stress by ansys 16 (mpa) maximum contact stress by hertz’s equation (mpa) error (%) case-hardened wrought steel 1195.2 1194.56 -0.05% astm class 35 cast iron 1202.6 1177.75 -2.1% grey cast iron (gg-30) 1231.3 1264.41 2.6% beryllium-aluminum alloy (albemet 162) 1286.8 1325.82 2.9% fig. 13 shows that the contact stress from each material is different, though the analysis is done by keeping all conditions constant except the type of materials. this indicates that the type of gear materials has a great influence on gear failure due to the contact stress. 4. conclusions the surface strength of the gear tooth is a crucial parameter to prevent the failure of mating gears. in this study, the theoretical approach is used to predict the surface contact stress on the spur gear mates, and is done on different gear materials to study their influence on the surface contact strength of spur gears. moreover, the study shows that the effective way of predicting the surface contact stress is to use the 3d geometrical contact model of spur gears under the fe method (ansys 16). the accuracy of the fe results for the selected materials is verified by comparing the fe results with the analytical results using hertz’s standard formula. the meshed in-volute spur gear is properly modeled by rotating cylindrical contacts for the theoretical investigation of surface contact stresses under the applied torque/moment. the 3d spur gear model is done in the modeling software (solidworks) and imported to ansys 16 workbench for fea for the selected gear materials. the correlation between the theoretical formulation and fe results is compared with tolerable accuracy. the comparison shows that the results are well-matched with a maximum of 2.9% accuracy error. both the results of fe and theoretical methods show that the value of contact stress varies as the gear materials are changed. this indicates that the type of gear materials has a direct influence on the surface contact strength of spur gear mates. thus, gear designers and/or gear manufacturers have to consider gear materials as a factor for the surface contact failure of gears. conflicts of interest the authors declare no conflicts of interest. 1100 1150 1200 1250 1300 1350 case-hardened wrought steels astm class 35 cast iron grey cast iron (gg-30) beryllium-aluminum alloy (albemet 162) c o n ta c t s tr e s s i n m p a hertizian contact stress contact stress from ansys 16 fig. 13 comparison of the contact stresses obtained by theoretical analysis and ansys simulation 58 proceedings of engineering and technology innovation, vol. 22, 2022, pp. 50-59 references [1] n. d. narayankar and k. s. mangrulkar, “contact stress and bending stress analysis of spur gear by analytical method,” international journal on theoretical and applied research in mechanical engineering, vol. 6, no. 1-2, pp. 1-3, 2017. [2] p. s. rao and c. vamsi, “contact stress and shear stress analysis of spur gear using ansys and theoretical,” international journal of modern studies in mechanical engineering, vol. 2, no. 2, pp. 9-14, 2016. [3] a. d. kalani, j. a. vadher, and r. k. jani, “fatigue failure of spur gears–a review,” international journal of modern trends in engineering and research, vol. 4, no. 11, pp. 182-186, november 2017. [4] p. singh, g. verma, and d. l. singh, “design and analysis of stress induced in spur gear tool profile using catia and ansys,” international journal of engineering and techniques, vol. 3, no. 5, pp. 118-124, september-october 2017. [5] x. deng, “analysis and prediction of gear fatigue life,” iop conference series: earth and environmental science, pp. 1-5, april 2019. [6] p. s. rao, n. sriraj, and m. farookh, “contact stress analysis of spur gear for different materials using ansys and hertz equation,” international journal of modern studies in mechanical engineering, vol. 1, no. 1, pp. 45-52, june 2015. [7] p. d. devan and v. r. muruganantham, “contact and bending stress analysis of gears–a review,” international journal for research in applied science and engineering technology, vol. 5, no. 12, pp. 2561-2565, december 2017. [8] k. sivakumar, t. m. saravana, s. balaguru, r. sabarish, and r. hariharan, “contact stress analysis of a spur gear tooth pair for two different materials,” international journal of applied engineering research, vol. 10, no. 68, pp. 344-349, january 2015. [9] r. l. norton, cam design and manufacturing handbook, new york: industrial press inc., 2002. [10] b. gupta, a. choubey, g. v. varde, “contact stress analysis of spur gear,” international journal of engineering research and technology, vol. 1, no. 4, june 2012. [11] a. r. hassen, “contact stress analysis of spur gear teeth pair, ” world academy of science, engineering, and technology, vol. 58, no. 1, pp. 597-602, 2009. [12] d. walton and a. j. goodwin, “the wear of unlubricated metallic spur gears,” wear, vol. 222, no. 2, pp. 103-113, november 1998. [13] “selection of gear material,” https://gearmotions.com/selection-of-gear-material/, december 05, 2017. [14] j. r. davis, gear materials, properties, and manufacture, ohio: asm international, 2005. [15] khk stock gears, “calculation of gear dimensions,” https://khkgears.net/new/gear_knowledge/gear_technical_reference/calculation_gear_dimensions.html, december 20, 2017. [16] m. j. khan, a. mangla, and s. h. din, “contact stress analysis of stainless steel spur gears using finite element analysis and comparison with theoretical results using hertz theory,” international journal of engineering research and applications, vol. 5, no. 4, pp. 10-18, april 2015. [17] j. liu, c. wang, and w. wu, “research on meshing stiffness and vibration response of pitting fault gears with different degrees,” international journal of rotating machinery, vol. 2020, article no. 4176430, 2020. [18] r. dharshini and m. vasundara, “analysis of load factors and modes of failure on spur gear,” international research journal of engineering and technology, vol. 5, no. 6, pp. 1618-1625, june 2018. [19] m. bozca, “investigation of the effective parameters of scuffing failure in gears,” ukrainian journal of mechanical engineering and materials science, vol. 6, no. 1, pp. 1-7, july 2020. copyright© by the authors. licensee taeti, taiwan. this article is an open access article distributed under the terms and conditions of the creative commons attribution (cc by-nc) license (https://creativecommons.org/licenses/by-nc/4.0/). 59  proceedings of engineering and technology innovation, vol. 22, 2022, pp. 01-09 manufacture and design an apparatus for measuring the thermal resistance of building materials ammar saliby * , béla kovács department of mathematics analysis, miskolc university, miskolc, hungary received 17 february 2022; received in revised form 22 april 2022; accepted 23 april 2022 doi: https://doi.org/10.46604/peti.2022.9500 abstract a good thermal design of buildings plays a key role in reducing the thermal loads of heating and air conditioning systems and thus reducing energy consumption. this study develops a steady-state apparatus for measuring the thermal resistance of building materials by using an absolute technique. the apparatus is designed to accommodate the full and actual size of the tested samples of building materials used in the syrian market to ease testing. therefore, suitable samples do not have to be built for the apparatus; the thermal resistance is measured for many samples, and the results are compared with the thermal insulation code for buildings in the syrian arab republic. the results show that the difference between the resistance values obtained from the experiments and the syrian thermal insulation code is due to the materials used in the local markets and their non-compliance with the code. these materials need testing and quality control during the manufacturing process. keywords: thermal resistance, thermal conductivity coefficient, steady-state, absolute technique 1. introduction in this era, wasted energy is the biggest problem in industrial, commercial, and residential fields. all these fields have a common factor, i.e., buildings. a significant amount of energy is used to make the buildings comfortable to live or work in. however, a considerable amount of energy is wasted on the building envelope. this problem is unavoidable, but the amount of energy wasted could be minimized. additionally, building materials could be improved to make better designs for energy-sustainable buildings. furthermore, updating standards and regulations can help keep up with the latest metrology technology. generally, there are two topics that receive much attention in material engineering, i.e., structure and thermal. this study will focus on the thermal side. there are multiple thermal parameters used for energy calculations in heat transfer applications. these values are available in books and references. however, with the emergence of new materials, these sources need new-found methods with these parameters [1]. the heat transfer discipline depends on explaining thermal energy transfer and its rate under specific conditions. the key elements in determining heat transfer are thermal properties, temperature difference, and dimensions. these factors form the definition of thermal resistance [2]. thermal resistance (r) in this study is limited to conduction only. it is defined as the resistance of energy flow in a structural element through its thickness [2]. therefore, increasing this resistance means the ability of the structural elements to resist the energy flow and reduce the heat transfer. these materials are considered poor conduction materials, called thermal insulators. the thermal resistance is calculated by dividing the thickness of the material (δ) by its thermal conductivity (λ), and its unit is m 2 .°c/w, as shown in eq. (1). * corresponding author. e-mail address: a.saliby@hotmail.com proceedings of engineering and technology innovation, vol. 22, 2022, pp. 01-09 2 r    (1) the thermal resistance increases when the thermal conductivity decreases or the heat sink area becomes larger while the spreading thermal resistance value stays the same. also, the effective thermal conductivity rises when the heat source area is smaller, and the heat sink area is larger [3-5]. additionally, the thermal resistance increases when the thickness of the material increases. the cost of building materials increases exponentially with wall thicknesses. therefore, finding the optimal thickness can minimize the energy waste and the overall cost of the building. many methods have been developed and used to measure the thermal resistance of different materials [6]. however, most of these methods have accuracy limits based on the temperature and thermal conductivity of the samples. proceeding from the importance of thermal resistance in heat transfer calculations, it is necessary to determine and measure the resistance value experimentally. this study identifies the methods of measuring the thermal resistance, develops an apparatus to measure the thermal resistance, and tests some samples of building materials by using the developed apparatus. the thermal resistance of these samples is also calculated theoretically based on the thermal insulation code for buildings in the syrian arab republic. finally, a comparison is made between the thermal resistance values obtained based on the code and the apparatus. 2. methodology and techniques 2.1. steady-state method in the steady-state method, temperatures are measured at a heat source and a heat sink. the temperature difference is calculated at a known distance and a constant heat flow through a sample. fig. 1 shows a schematic diagram of steady-state methods. (1) absolute technique: this method is used for the sample in the form of a cuboid or cylinder. the sample is placed between a heat source and a heat sink, and the temperature difference is measured over a known distance to the sample by thermocouples, which are the most widely used devices because of their cheapness and measurement accuracy of less than 1% [7]. (2) comparative technique: in this method, the tested sample is placed with another sample made of a material with a known thermal conductivity coefficient. the thermal conductivity is derived from the measured temperature gradients in the tested sample and the thermal conductivity of the reference materials [8]. (3) radial heat flow technique: this method allows experiments to be carried out at various temperatures, especially when the temperatures are more than 700°c, where the heat loss by radiation cannot be neglected. normally, this method uses the sample with a circular or duct shape. heat is provided to the sample by using the heaters at the axis, and the heat flows out diagonally. thermocouples are used to measure the temperature with an accuracy of ±0.1°c [9]. (4) parallel thermal conductance technique: this method is suitable for the sample with small diameters, such as carbon fibers, with a diameter of 10-100 mm. with this method, it is difficult to measure the temperature by thermocouples and determine the heat flow. a parallel heat conduction method is proposed for the needlelike sample that cannot support heaters and thermocouples [10]. before measuring the thermal resistance of the sample, the sample is measured concerning the sample support first to determine the heat loss rates related to the sample support. then, the test sample is connected to the sample support, and the thermal resistance is measured again. proceedings of engineering and technology innovation, vol. 22, 2022, pp. 01-09 3 (a) absolute technique (b) comparative technique (c) radial heat flow technique (d) parallel thermal conductance technique fig. 1 steady-state methods 2.2. transient-state method the transient-state method determines the energy dissipation as a function of time. several transient methods have been developed to eliminate the negatives of the steady-state such as the heat loss and long waiting time to reach a constant temperature difference. fig. 2 presents a schematic diagram of transient-state methods. (1) pulsed power technique: this method is similar to the steady-state method, but the heating current is a square pulse wave. one of the advantages of this method is that it allows quick measurements, especially when there are many samples [11]. (2) hot wire technique: this method measures the temperature rise at a known distance from the linear heat source. the method assumes a one-dimensional heat flow inside the sample, which means that the heat source is linear with an infinite length and infinitesimal diameter. when there is an electric current of constant intensity passing through the hot wire, the heat conduction coefficient of the sample can be derived. the available time is tested by using the resulting temperature change at a known distance from the hot wire during a period. the hot wire method is used to measure the materials with a low coefficient of thermal conductivity (such as plastics and fibers) [12]. (3) transient plane source (tps) technique: this method uses a thin metal strip or a metal disc as a flat heat source and a temperature sensor. the metal disc is electrically isolated and placed between two test samples; all other surfaces are thermally insulated. a small constant electric current is applied to the metal disc during the experiment to heat it. the thermal properties of the test samples can be determined by observing the temperature increase for a short period, and the measurement accuracy of the temperature sensor is ±0.01°c [13]. (4) laser flash technique for thermal diffusivity: thermal contact resistance is a significant source of error when measuring the temperature. the thermal diffusion method is used by laser flash to sense the temperature without contact to achieve high accuracy. this method uses rapid optical heating, and the temperature is measured using thermocouples, an infrared detector, or an optical pyrometer. the sample is prepared by coating or spraying a layer of graphite or any high emissivity material on both sides, to make an absorbent from the front side and an emitter on the backside to sense the temperature [14]. proceedings of engineering and technology innovation, vol. 22, 2022, pp. 01-09 4 (a) pulsed power technique (b) hot wire technique (c) transient plane source technique (d) laser flash technique for thermal diffusivity fig. 2 transient-state methods there are two general types of apparatus for determining the thermal resistance of materials. the first type (the hot plate apparatus), which is the most common, is used to determine the thermal resistance from one material surface to the other surface. the apparatus consists of a heating section lined with double-sided copper plates. the copper plates are divided into two parts. these two parts are completely separated by an air gap that prevents any metallic contact between the inner and outer parts of the heating plate, as the inner part is the test part. the outer part is a protective and insulating ring to prevent heat loss during the experiment [15-17]. the second type (hotbox apparatus) is used to determine the total thermal resistance of the air on one side to the air on the other side of the material. the hot plate apparatus is generally accepted to determine the heat resistance. the main goal of any device is to obtain accurate results. this result can be achieved by taking accurate temperature measurements and achieving a correct heat flow direction through the material. special precautions must be taken to prevent heat loss from the test values and ensure that the heat flow lines are perpendicular to the test surfaces. 3. equipment and procedures used in apparatus design 3.1. purpose of apparatus the design of the apparatus, which is based on the absolute technique in the steady-state method, aims to measure the thermal resistance by applying a constant heat flow during the experiment and measuring the temperature difference between the heat source and the heat sink. 3.2. apparatus design and manufacture the thermal resistance measurement apparatus is designed based on the absolute technique in the steady state, where a constant heat source heats the sample, and the temperature difference is measured over a distance known to the sample by temperature sensors. a thermal insulator surrounds the sample and the heating part to ensure the one-dimensional heat transfer and reduce the heat loss by convection and radiation. fig. 3 shows the design of the apparatus in solidworks, where the dimensions are set based on the building materials. proceedings of engineering and technology innovation, vol. 22, 2022, pp. 01-09 5 fig. 3 apparatus design in solidworks 3.3. basic parts the thermal resistance measurement apparatus for building materials mainly consists of a metal structure made of iron bars and a number of thermal insulation types depending on the temperatures to prevent the heat loss. table 1 shows the basic parts of the apparatus. the insulation materials, such as polystyrene, mineral wool, and glass, can be replaced with foam geopolymer materials based on fly ash. they are a non-flammable material characterized by relatively good insulation parameters [18]. fig. 4 illustrates the horizontal projection of the apparatus with its constituent parts. table 1 basic parts of the proposed apparatus no. part name dimensions length (cm) width (cm) height (cm) 1 metal frame 60 60 40 2 aluminum plate 40 1 20 3 mineral wool 40 5 40 4 polystyrene board 40 10 40 (a) dimensions of the apparatus (b) the apparatus after assembly fig. 4 horizontal projection of the apparatus after assembly 3.4. other parts the apparatus contains other systems to complete its work. these systems are the heat source, control system, and sensors. an electrical heater is used as the heat source to provide heat to the system. the control system is used to maintain the temperature at around 200°c, display the temperature from the sensors, and log the data. the instruments used in these systems are described below. proceedings of engineering and technology innovation, vol. 22, 2022, pp. 01-09 6 fig. 5 thermocouples (type k) fig. 6 arduino uno r3 board fig. 7 max6675 assembled printed circuit board (pcb) fig. 8 lcd (type 2004a) fig. 9 power display fig. 10 electric heater on the aluminum plate fig. 11 power supply (1) thermocouples: six k-type thermocouples, which have a length of 100 mm and a diameter of 5 mm, are shown in fig. 5. (2) data logger: the code is designed and written to be compatible with the heat resistance apparatus based on the arduino uno r3 microcontroller. it is an open-source electronic board to easily develop many control projects using the open-source programming language arduino c, as shown in fig. 6 [19]. (3) analog-digital converter (adc) and cold junction compensation: as shown in fig. 7, this study selects an max6675 integrated circuit compatible with k-type thermocouples and a 12-bit adc with a heat-sensing diode [20]. (4) temperature display and data recording: as shown in fig. 8, a liquid-crystal display (lcd) screen (4 × 20) is installed to read the temperature values measured by thermocouples. (5) display power provided: a power screen, as shown in fig. 9, is installed to read the power values provided by the electric heater. (6) electric heater: an electric heater works as a heat source, wraps the heater, and is fixed by brackets on the aluminum plate. the electric heater with a length of 1 m and a diameter of 5 mm is installed as shown in fig. 10. (7) electrical power supply: as shown in fig. 11, an electrical supply source (12 v/5 a) is installed to supply the equipment with constant voltage. 3.5. apparatus calibration apparatus calibration is done to determine the amount of heat flowing through the sample and to determine the error that results from the heat loss by convection, radiation, contact resistance, and resistance of aluminum plates. the calibration is done using a material with a known thermal conductivity coefficient, such as gypsum with dimensions of 40 × 20 cm, a thickness of 1.6 cm, and a thermal conductivity coefficient of 0.16 w/m.°c. the heat is presented on the front side of the sample at a temperature of tai = 200°c by an electric heater. after reaching the steady state, the mean temperature at the front face of the sample is tai = 200°c, and the mean temperature at the opposite side is tao = 85°c. fig. 12 shows the mean temperatures tai and tao. through this calibration, the value of the heat provided is known without losses by convection, radiation, contact resistance, and resistance of aluminum plates. the error value from the losses is included in the calculation code for arduino uno. proceedings of engineering and technology innovation, vol. 22, 2022, pp. 01-09 7 4. tests and experiments experiments are conducted on the materials used in the syrian local markets, where two types of the most famous samples are selected to verify their thermal resistance and compare the results with the syrian thermal insulation code. (1) the sample is a hollow cement block: the sample has a thickness of 10 cm and is placed within the apparatus. the experiment is conducted by adding heat to the front side of the sample using the electric heater. after the system reaches the steady state, the mean temperature at the front side (the heat source) tai is 200°c, and the mean temperature at the opposite side tao is 85°c, the temperature difference δt is 115°c. the heat of 125 w is introduced to the system, resulting in the thermal conductivity coefficient λ of 1.28 w/m.°c. fig. 13 illustrates the tested sample’s dimensions, and fig. 14 shows the mean temperature tai measured on the front side of the sample (heat source) over time and the mean temperature tao on the backside of the sample (heat sink). (2) the sample is a wall section: the wall section contains a hollow cement block with a thickness of 10 cm and two layers of cement (plaster) with a thickness of 2 cm. the experiment is conducted by adding heat to the front side of the sample using the electric heater. after the system reaches the steady state, the mean temperature at the front side (the heat source) tai is 200°c, the mean temperature at the opposite side tao is 65°c, and the temperature difference δt is 135°c. the heat of 125 w is introduced to the system, resulting in the thermal conductivity coefficient λ of 1.62 w/m.°c. fig. 15 illustrates the tested sample’s dimensions, and fig. 16 shows the mean temperature tai measured on the front side of the sample (heat source) over time and the mean temperature tao on the backside of the sample (heat sink). fig. 13 dimensions of the hollow cement block 0 50 100 150 200 250 0 0.5 1 1.5 2 t e m p e ra tu re [ °c ] time [hrs.] tai tao 0 50 100 150 200 250 0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 4.0 4.5 t e m p e ra tu re [ °c ] time [hrs.] tai tao fig. 14 mean temperatures (tai and tao) of the hollow cement block fig. 12 mean temperatures (tai and tao) of the gypsum board proceedings of engineering and technology innovation, vol. 22, 2022, pp. 01-09 8 fig. 15 dimensions of the wall section fig. 16 mean temperatures (tai and tao) of the wall section as shown in table 2, the physical and thermal characterization of the tested materials includes: length (a), width (b), weight (w), density (ρ), thermal conductivity coefficient (λ), thermal resistance (r), thermal resistance according to the code [21] (r*), absolute error, and relative error. table 2 physical and thermal characterization of materials materials characteristics hollow cement block wall section a (cm) 40 40 b (cm) 20 20 δ (cm) 10 14 w (kg) 12.1 16.4 ρ (kg/m 3 ) 1513 λ (w/m.°c) 1.282 1.624 r (m 2 .°c/w) 0.078 0.086 r * (m 2 .°c/w) 0.100 0.139 absolute error 0.022 0.053 relative error 22 % 38 % 5. conclusions this study designs an apparatus for measuring the thermal resistance of the building materials used in the syrian arab republic. the results indicate that the thermal resistance values obtained by the experiment differ from the values obtained by the thermal insulation code for buildings in the syrian arab republic. this difference is due to the materials used in the local markets and their non-compliance with the code. these materials need testing and quality control during the manufacturing process. knowing the heat resistance value is necessary to calculate the lost or gained thermal energy through the structural element and to judge the efficiency of thermal insulation with various structural elements. the lower the value of the thermal conductivity coefficient, the greater the thermal insulation capacity and the higher the percentage of savings in thermal energy used for heating and air conditioning purposes. to measure the thermal resistance, it is essential to develop an apparatus which can work in different positions to achieve the horizontal and vertical heat transfer, determine the amount of heat, and reduce the measurement errors and heat loss. 0 50 100 150 200 250 0.0 1.0 2.0 3.0 4.0 t e m p e ra tu re [ °c ] time [hrs.] tai tao proceedings of engineering and technology innovation, vol. 22, 2022, pp. 01-09 9 nomenclature r thermal resistance (m 2 .°c/w) tai mean temperature (heat source) (°c) tao mean temperature (heat sink) (°c) w weight (kg) δ thickness (cm) λ thermal conductivity coefficient (w/m.°c) ρ density (kg/m 3 ) conflicts of interest the authors declare no conflicts of interest. references [1] y. a. çengel, heat transfer: a practical approach, 2nd ed., new york: macgraw-hill, 2003. [2] f. p. incropera, et al., fundamentals of heat and mass transfer, 6th ed., new york: john wiley & sons, 2007. [3] m. kim, et al., “numerical case study and modeling for spreading thermal resistance and effective thermal conductivity for flat heat pipe,” case studies in thermal engineering, vol. 31, article no. 101803, march 2022. [4] m. kim, et al., “numerical investigation and modeling of thermal resistance and effective thermal conductivity for two-phase thermosyphon,” case studies in thermal engineering, vol. 27, article no. 101358, october 2021. [5] j. h. moon, et al., “boiling-driven, wickless, and orientation-independent thermal ground plane,” international journal of heat and mass transfer, vol. 167, article no. 120817, march 2021. [6] d. zhao, et al., “measurement techniques for thermal conductivity and interfacial thermal conductance of bulk and thin film materials,” journal of electronic packaging, vol. 138, no. 4, article no. 040802, december 2016. [7] standard test method for steady-state heat flux measurements and thermal transmission properties by means of the guarded-hot-plate apparatus, astm c177, 2019. [8] standard test method for thermal conductivity of solids by means of the guarded-comparative-longitudinal heat flow technique, astm e1225, 2004. [9] standard test method for steady-state heat transfer properties of pipe insulation, astm c335/c335m, 2017. [10] b. m. zawilski, et al., “description of the parallel thermal conductance technique for the measurement of the thermal conductivity of small diameter samples,” review of scientific instruments, vol. 72, no. 3, pp. 1770-1774, march 2001. [11] o. maldonado, “pulse method for simultaneous measurement of electric thermopower and heat conductivity at low temperatures,” cryogenics, vol. 32, no. 10, pp. 908-912, april 1992. [12] standard test method for thermal conductivity of refractories by hot wire (platinum resistance thermometer technique), astm c1113/c1113m, 2009. [13] standard test method for measurement of thermal effusivity of fabrics using a modified transient plane source (mtps) instrument, astm d7984, 2016. [14] standard test method for thermal diffusivity by the flash method, astm e1461, 2013. [15] m. m. terzić, et al., “development of a single-sided guarded hot plate apparatus for thermal conductivity measurements,” thermal science, vol. 20, no. 1, pp. 321-329, 2016. [16] w. hemminger, et al., “a guarded hot-plate apparatus for thermal conductivity measurements over the temperature range -75 to 200 [°c],” international journal of thermophysics, vol. 6, no. 5, pp. 489-498, 1985. [17] u. hammerschmidt, “guarded hot-plate (ghp) method: uncertainty assessment,” international journal of thermophysics, vol. 23, no. 6, pp. 1551-1570, 2002. [18] m. łach, et al., “development and characterization of thermal insulation geopolymer foams based on fly ash,” proceedings of engineering and technology innovation, vol. 16, pp. 23-29, august 2020. [19] “arduino uno r3,” http://www.arduino.cc, june 2022. [20] “cold-junction-compensated k-thermocoupleto-digital converter (0°c to +1024°c),” http://www.maximintegrated.com, june 2021. [21] thermal insulation code for buildings in the syrian arab republic, syrian engineers association, 2006. copyright© by the authors. licensee taeti, taiwan. this article is an open access article distributed under the terms and conditions of the creative commons attribution (cc by-nc) license (https://creativecommons.org/licenses/by-nc/4.0/). http://www.arduino.cc/ http://www.maximintegrated.com/  proceedings of engineering and technology innovation, vol. 13, 2019, pp. 10-19 impact of dfig in wind energy conversion system for grid disturbances sasmita behera * , sudeep kumar behera, bibhuti bhusan pati department of electrical & electronics engineering, veer surendra sai university of technology, burla, odisha, india received 09 april 2018; received in revised form 11 december 2018; accepted 25 january 2019 abstract in this work, a grid-connected doublyfed induction generator (dfig) is studied for the transient and steady response. the vector control technique controls the pulse width modulation (pwm) of both the back-to-back converters interfacing rotor to the grid. reactive power supply and dc bus voltage are managed by the grid-side inverter. active power and rotor angular speed are adjusted by the machine side inverter facilitating power generation for varying wind. the effect of voltage and frequency deviation from the grid on the control is observed. the controllers are found to work satisfactorily except for large frequency variation. the current harmonics are also within the allowed limit. the proposed controllers are expected to satisfy the revised grid code for wind energy. keywords: vector control, pwm, voltage sag, voltage swells 1. introduction increasing share of wind power in grid connection draws attention to the study of its effect on power system stability. in the wind energy conversion system (wecs), different generators are in use. the majority of variable speed wecs have doubly fed induction generator (dfig) [1]. this is due to the efficient energy conversion, low cost, less mechanical stress etc. [2]. the increased proportion of wind power has forced the grid codes to be modified from its isolation by crowbar [1] to ride through the control during the grid disturbance [3]. active and reactive power control for changing wind speed is accomplished by a stator flux oriented vector control with its real axis aligned to the stator flux vector rotating at the synchronous speed in [4]. the performance of such a system for various balanced and unbalanced fault conditions is explored in [5 -6]. there are many possible ways to control the outputs of the generator with different orientation of the real axis of the reference frame [7]. minor but frequent disturbances in the grid such as voltage dip, sag and swell in the grid lead to poor of power quality, reliability,and equipment failure [8]. such disturbances associated with voltage can be supplemented by reactive power balance. different external reactive power support devices have also been in research, but with the additional cost [9-11]. in [12], upholding of dfig by internal control for low voltage in stator voltage oriented reference frame is studied. in another study, the swell case of voltage is observed in [13]. xiao et al. [14] have tentatively defined a maximum limit beyond which the crowbar protection should be activated. selectively harmonics have been eliminated by proper control of dfig in [15]. to minimize rotor current that activates crowbar during low voltage, additional feed-forward control has been proposed in [16]. instead of voltage oriented vector control, a direct current control [17] has been proposed for faster control. mohseni et al. [18-20] have proposed switching to hysteresis control during transient rather than utilizing proportional-integral (pi) current control for steady-state in order to satisfy grid code for the ride-through requirement. but, harmonics with hysteresis control is beyond the standard. pi with resonance control has been proposed in [21-22], to reduce torque pulsation and unbalanced voltage [23]. in addition, to reduce the transient time, torque oscillation for voltage swells virtual impedance control strategy has been proposed in [24]. *corresponding author. e-mail address:sasmitabehera2000m@gmail.com proceedings of engineering and technology innovation, vol. 13, 2019, pp. 10-19 11 the model predictive control (mpc) for switching converter instead of pwm has been reviewed in [25]. though it was faster than pwm, the thd and steady-state performance were inferior. in a comparative study, tremblay et al. [26] have shown robustness to noise and better control of the dfig by pi control in vector control in comparison to direct power control (dpc) and direct torque control (dtc) approach. in recent literature, a demagnetizing current control based on vector control has been proposed for faster control for low voltage [27]. the studies are also limited to sag, swell or harmonic reduction. having reviewed the literature for improvement of control to meet the latest grid codes for wecs, the vector control with pi controller is still unsurpassed though many other control techniques, which have been developed [28-32]. herein, the vector control technique controls the pwm of both the back to back converters interfacing rotor to the grid to maintain reliable power. the sections 2-5 embody the equations governing the different parts of the model. response to grid disturbances is presented in section 6 concluding the impact of dfig in this condition. 2. model of wt-dfig system the power of the wind drives the wind turbine (wt) coupled to dfig. with the gearbox in between, the rotational speed is stepped up suitably and the dfig generates electricity from the mechanical input. fig. 1 shows a dfig vector control strategy. for the accurate design of the dynamic model in the d-q frame, the transient reactance of the generator is considered. the equivalent circuits of dfig and induction machine are analogous [7, 33]. fig. 1 overall structure of the dfig model definingthe different electrical circuit parameters in the d-q frame as in eqs. (1-5), it yields   qr rr sss d l l e )( '  (1)   dr rr ms q l l e )( '  (2) lx ssss  (3)                  l l lx rr m ssss 2 '  (4) r l t r rr '0 (5) proceedings of engineering and technology innovation, vol. 13, 2019, pp. 10-19 12 the dfig is modelled as in eqs. (6) to (10) :     ixe t v l l esixx t rv dt dix qssq s dr rr m ddsss s sds s dss '' ' 0 '' ' 0 ' 1 1 1         (6)     ixe t v l l esixx t rv dt dix dssq s qr rr m qqsss s sqs s qss        '' ' 0 '' ' 0 ' 1 1 1  (7)   ixxe t v l l e dt de qssdqr rr m sqs d s '' ' 0 ' ' 1  (8)   ixxe t v l l e dt de dssqdr rr m sds q s '' ' 0 ' ' 1  (9) dr rr mdr dsi il l  (10) i il lqr rr qr m qs  (11) where , l , r , x , v , e , i are flux linkage, inductance, reactance, voltage, emf, and current respectively. the first subscript d, q stands for the directors and the quadrature component respectively, and the second subscript r and s represent as rotor and stator respectively. the repetition of subscript s and r for l stand for the corresponding self-inductance, whereas single m stands for mutual inductance. t is the synchronous speed, s is the slip, 0 ' t is the electrical time constant of the rotor,the superscript ′ represents a transient condition. equating the lhs of eqs. (3) and (4), and neglecting rotor resistance dsv .then substituting for dr and qr , it reduces to ( ) ( ) dr s r rr qr s r m qs v l i l i        (12) ( ) ( ) qr s r rr dr s r m ds v l i l i       (13) 3. the drive train to model drive train, the two-mass model is simple and it amply represents the dynamics of wt suitable for transient state analysis [34]. the two masses are the generator-gearbox and the hub with blades. mathematically, it is expressed by tt dt d h sw s t   2 (14)   )1( s dt d strt t  (15) tt dt ds h seg 2 (16) dt d dkt t stss    (17) where th and gh represent the inertia of the hub and the generator, respectively, t is the turbine angular speed, r is the generator rotor angular speed, t is the shaft torsion angle, sd and sk are the is the stiffness and damping coefficients of shaft respectively, st is the shaft torque, wt is the effective torque input to the wt, et is the electromagnetic torque. wt and et are given by proceedings of engineering and technology innovation, vol. 13, 2019, pp. 10-19 13 v cr t w t pa w 3 25.0    (18) / e s s t p  (19)    0068.0)54.0 116 (5176.0 21   ec i i p (20)  3 1 035.0 08.0 11     i (21) t w r v    (22) where ap is the air density, r is the radius swept by wt blade, wv is the velocity of wind,  is the pitch angle,  is the blade tip speed ratio, sp is the stator active power, pc is the power coefficient. by controlling t , maximum pc and maximum power from wind can be achieved. 4. converter modeling from fig. 2(b), the active power equation is: dcr gp p p  (23) where r p is the rotor power that charges the dc bus capacitor with dc p , and delivers g p to the grid. these are given by r dr dr qr qrv i v ip   (24) g dg dg qg qgv i v ip   (25) dc dc dc dc dc c dt dv v i vp    (26) where v stands for the voltage and i the current with the second subscript r and g represent for rotor and grid respectively. dc v , dc i and c are the voltage, current,and capacitance of the capacitor in dc link. now, rewriting eq. (18), the power balance in terms of voltages and currents gives: ( )dc vdc dg dg qg qg dr dr qr qr dv c v i v i v i v i dt     (27) 5. controller modeling fixing d-axis of the d q reference frame with dsv , qsv nullifies and dsv equals the grid voltage. gp and reactive power g q can be regulated separately by dg i and qg i respectively in a decoupled manner by vector control resolving three phases to d q components [35]. the stator active power and voltage can also be adjusted separately by qr v and dr v as described below. 5.1.machine-side control the controller on the rotor-side retains the terminal voltage to set point and controls the dfig output active power varying with the wt torque [27]. the control of active power and voltage are achieved by qr v and drv , respectively. the control structure is given in fig. 2(a), while the control is expressed as: proceedings of engineering and technology innovation, vol. 13, 2019, pp. 10-19 14  2 1 1 21 2p p i i m rrqr qr r s ds r s qrpv x i x s i s ik k k k l l        (28)  2 3 3 23 4p p i i rrdr dr r s drvv x i x s ik k k k l      (29) where the gain of the pi controller is xy k . the first subscript x with notation p and i represent proportional and integral gains respectively. the second subscript 1, 2, 3 are for the power controller, machine-side current controller, and grid voltage controller, respectively. dg ef i r and dg ef i r are the references for the respectivecomponents of the current on the rotor, s ref v  is the reference for stator terminal, ref p is the reference of active powerfound from r p b t ef tb p    (30) where tb is the wt angular speed base, bp is the maximum output active power at tb . 5.2.grid-side control the grid-side converter controller (fig. 2(b)), controls dcv by dgi and the reactive power by dgi . the change in intermediary states 1 x , 2x and 3x , shown in fig. 2(b) are expressed as: 1 dc ref dc dc dx dt v v v    (31) 2 1pdg idgdg ref dg dc dg dx dt k ki i v x i      (32) 3 qg ref qg dx dt i i   (33) 2 2pg igdg dt dx k kv x  (34) 3 3igqg pg dt dx k kv x  (35) (a) machine-side converter control (b) grid-side converter control fig. 2 the control of converters [12] where pdg k and idg k are the proportional and integral gains of the dc bus voltage controller, respectively. its output dg ef i r is the set point for dg i ; pg k and ig k are the proportional and integral gains of the grid-side converter current regulator, respectively, dc ef v r is the dc bus voltage set point and dq efi r is the reference for dqi [7]. dc link voltage drop due to increased active power demand due to or wind speed variation in sub-synchronous speed operation is controlled by a change in dg v (cf. eq. (22)). reactive power is controlled by controlling dgv . pwm gating pulses are generated in a similar manner from above two grid voltages as shown in fig. 2(a). proceedings of engineering and technology innovation, vol. 13, 2019, pp. 10-19 15 6. result discussion the wecs considered is integrated to a 120 kv, 3 phase 2500 mva grid with transformers in between. the model is built in matlab/simulink with system specifications: w v :15 m/s,dfig: 9kw, 575 v, 6 pole, 1440 rpm, 60 hz; dc bus voltage 1150 v with a capacitor of 60 mf; frequency of the pwm carriers 1.62 khz and 2.7 khz for machine-side and grid-side converters respectively.to reduce the injection of harmonics from dfig via the grid side converter to the grid, the switching frequency is kept high. 6.1. balanced grid the stator currents and voltages of dfig under steady-state condition (the wind speed constant at 15m/s) are depicted in fig. 3(a). the active power (blue) produced by dfig is 0.9 pu which is excluding electromechanical losses from the mechanical input, which is similar to the findings in [4]. the reactive power (green) reference is 0 kvar. the dc bus voltage remains stable at 1150 v. the rotor angular speed is 1.2 pu, the rated value. (a) no disturbance (b) with voltage sag fig. 3 simulation results of performance parameter variation with times 6.2. voltage sag on the grid grid voltage sag occurs due to generation outage, line outage and switching of a heavy fluctuating load. when voltage sag of 0.15 pu [16, 21] is uniformly applied in all phases at 0.05 s for 9 cycles, the effects are depicted in fig. 3(b). normall y the deviation is limited to ±10% as per ieee 1159. abnormal violations of grid voltage, which sag beyond ±10%, are taken to verify the effectiveness of the control system. on abrupt deviation of the grid voltage from its rated, the stator current rises and the active power p fluctuates and the active power control competently brings it to its rating within the next 4 cycles. for voltage sag reactive power is supported by dfig with an increased injection of vqg and adjusting vdr . it supplies reactive power to the grid at that time and settles to 0 kvar by appropriate control action. for a variation of up to 25%, such rise of current due to sag has a negligible effect on dfig and the controller works well. the dc bus voltage oscillates at the switching instants and comes down to its set value of 1150 v by the grid side converter. the rotor speed is constant (1.2 pu) as v w is fixed. proceedings of engineering and technology innovation, vol. 13, 2019, pp. 10-19 16 6.3. voltage swells on the grid under symmetrical rise of voltage by 0.2 pu at 0.05 s continuing for 9 cycles, the variations are depicted in fig. 4(a).the grid voltage may swell under outage of a large load. when the grid voltage swells, the stator current falls to maintain powerconstant. the reactive power q falls suddenly due to a decrease in power demand and then comes back to 0 kvar within 4 cycles which is better than that in [8] and [13]. it sustains as per australian grid code [20]. finally, the system settles to its set value. 6.4. fall of grid frequency when the grid frequency drops from 60 hz rated to 58 hz at 0.05 s, the proposed control strategy does not act upon it as seen from fig. 4(b). such condition of frequency fall may arise due to loss of large generation. though the case is quite unusual, as the iec standard limits grid frequency deviation to ±0.5 hz, it is shown to check the prominent effect in the simulation. it is indicative of the modification of control required to counter to such large change in frequency. however, for the deviation within standard limits the controller performs to stabilize. since the machine rated frequency is 60 hz, it is not responding to 58 hz, and the machine collapses with a gradual fall of rotor angular speed and dc bus voltage. compared to the grid, the capacity of the generator is small, hence this dfig can’t support frequency stabilization of the grid. (a) voltage swell (b) with voltage sag fig. 4 simulation results of performance parameters variation with times 6.5. current harmonics fig. 5 thd of the stator current the harmonics in the stator current of the dfig are also analyzed and given in fig. 5. the dominant odd harmonics have suppressed considerably. the thd is 1.89 %, which is satisfied as per the standards. proceedings of engineering and technology innovation, vol. 13, 2019, pp. 10-19 17 7. conclusions a dfig-wt system with a back-to-back vsc is studied with its pwm controlled by pi controllers without a crowbar. the impact of dfig is investigated under steady state, for a large deviation of ±20% in voltage and 3% in the frequency of the grid. the stator voltage, current, active and reactive power exchange between the grid and the dfig are observed. in conclusion, the dfig supports grid within stability limits for the cases being discussed except for variation in frequency with the implementation of the vector control strategy. in the grid-connected dfig, the harmonic suppression measure has already been presented. since the harmonicfilter at the input to the controller, it is expected that the control will work in the presence of harmonics. the stator current has 1.89% thd, well within the limit. further study can be focused on the improvement of control performance in the larger rating of dfig and unbalanced faults. acknowledgement the facilities created by the project aicte/8023/rid/rps/19/11-12 dt.15.5.12 of the department entitled “transient stability analysis and control of power system with excitation control”, are gratefully acknowledged. conflicts of interest the authors declare no conflict of interest. references [1] j. lópez, e. gubía, e. olea, j. ruiz, and l. marroyo, “ride through of wind turbines with doubly fed induction generator under symmetrical voltage dips,”ieee transactions on industrial electronics,vol. 56, no. 10, pp. 4246-4254, 2009. [2] r. mittal, k.s. sandhu, and d.k. jain, “an overview of some important issues related to wind energy conversion system (wecs),” international journal of environmental science and development, vol. 1, no. 4, pp. 351-363, 2010. [3] a.k. pathak, m.p. sharma, and m. bundele, “a critical review of voltage and reactive power management of wind farms,”renewable and sustainable energy reviews,vol. 51, pp. 460-471, 2015. [4] b. c. babu and k.b. mohanty, “doubly-fed induction generator for variable speed wind energy conversion systems modeling &simulation,” international journal of computer and electrical engineering, vol. 2, no. 1, pp. 1793-1806, february, 2010. [5] k. s. rao and m. v. kumar, “analysis of doubly fed induction generator under varies fault conditions,” international journal of engineering research and applications, vol. 3, no. 6, pp. 2102-2106, november-december 2013. [6] x. yan, g. venkataramanan, p.s. flannery, y. wang, and b. zhang, “evaluation of the effect of voltage sags due to grid balanced and unbalanced faults on dfig wind turbines,”european power electronics and drives journal,vol. 20, no. 4, pp. 51-61, 2010. [7] f. wu, x.p. zhang, k. godfrey, and p. ju, “small signal stability analysis and optimal control of a wind turbine with doubly fed induction generator,”iet generation, transmission and distribution,vol. 1, no. 5, pp. 751-760, 2007. [8] n. edomah, “effects of voltage sags, swell and other disturbances on electrical equipment and their economic implications,” in iet electricity distribution-part 1, 20th international conference and exhibition oncired, 2009, pp. 1-4. [9] m.e. hossain, “low voltage ride-through capability improvement methods for dfig based wind farm,”journal of electrical systems and information technology, vol. 5, no. 3, 2018. [10] s. agalar and y.a. kaplan, “power quality improvement using sts and dvr in wind energy system,”renewable energy,vol. 118, pp. 1031-1040, 2018. [11] s. priyavarthini, c. nagamani, g.s. ilango, and m.a. a. rani, “an improved control for simultaneous sag/swell mitigation and reactive power support in a grid-connected wind farm with dvr,”international journal of electrical power and energy systems,vol. 101, pp. 38-49, 2018. [12] r. d. shukla and r. k. tripathi, “dynamic performance of dfig based wecs under different voltage sag,”international journal of chemtech research,vol. 5, no. 2, pp. 980-992, 2013. proceedings of engineering and technology innovation, vol. 13, 2019, pp. 10-19 18 [13] s. choudhury, k.b. mohanty, and b.k. debta, “investigation on performance of doubly-fed induction generator driven by wind turbine under grid voltage fluctuation,” inieee 10thinternational conference onenvironment and electrical engineering, 2011, pp. 1-4. [14] s. xiao, h. geng, h. zhou, and g. yang, “analysis of the control limit for rotor-side converter of doubly fed induction generator-based wind energy conversion system under various voltage dips,”iet renewable power generation,vol. 7 no. 1, pp. 71-81, 2013. [15] m. kesraoui, a. chaib, a. meziane, and a. boulezaz, “using a dfig based wind turbine for grid current harmonics filtering,”energy conversion and management,vol. 78, pp. 968-975, 2014. [16] j. liang, w. qiao, and r.g. harley, “feed-forward transient current control for low-voltage ride-through enhancement of dfig wind turbines,”ieee transactions on energy conversion,vol. 25, no. 3, pp. 836-843, 2010. [17] s. li, t.a. haskew, k.a. williams, and r.p. swatloski, “control of dfig wind turbine with direct-current vector control configuration,”ieee transactions on sustainable energy,vol. 3, no. 1, pp. 1-11, 2012. [18] m. mohseni, m.a. s. masoum, and s.m. islam, “low and high voltage ride-through of dfig wind turbines using hybrid current controlled converters,”electric power systems research,vol. 81, no. 7, pp. 1456-1465, 2011. [19] m. mohseni, s.m. islam, and m. a. s. masoum, “enhanced hysteresis-based current regulators in vector control of dfig wind turbines,”ieee transactions on power electronics,vol. 26, no. 1, pp. 223-234, 2011. [20] m. mohseni and s.m. islam, “transient control of dfig-based wind power plants in compliance with the australian grid code,”ieee transactions on power electronics,vol. 27, no. 6, pp. 2813-2824, 2012. [21] h. xu, w. q. zhang, h. nian, and j. li, “improved vector control of dfig based wind turbine during grid dips and swells,” inieeeinternational conference onelectrical machines and systems (icems), 2010, pp. 511-515. [22] j. hu, h. nian, h. xu, and y. he, “dynamic modeling and improved control of dfig under distorted grid voltage conditions,”ieee transactions on energy conversion,vol. 26, no. 1, pp. 163-175, 2011. [23] h. nian, y. song, p. zhou, and y. he, “improved direct power control of a wind turbine driven doubly fed induction generator during transient grid voltage unbalance,”ieee transactions on energy conversion,vol. 26, no.3, pp. 976-986, 2011. [24] z. xie, x. zhang, x. zhang, s. yang, and l. wang, “improved ride-through control of dfig during grid voltage swell,”ieee transactions on industrial electronics,vol. 62, no. 6, pp. 3584-3594, 2015. [25] j. rodriguez, m.p. kazmierkowski, j.r. espinoza, p. zanchetta, h. a. rub, h.a. young, and c.a. rojas, “state of the art of finite control set model predictive control in power electronics,”ieee transactions on industrial informatics,vol. 9, no. 2, pp. 1003-1016, 2013. [26] e. tremblay, s. atayde, and a. chandra, “comparative study of control strategies for the doubly fed induction generator in wind energy conversion systems: a dsp-based implementation approach,”ieee transactions on sustainable energy,vol. 2, no. 3, pp. 288-299, 2011. [27] m.k. döşoğlu, “a new approach for low voltage ride through capability in dfig based wind farm,” international journal of electrical power and energy systems,vol. 83, pp. 251-258, 2016. [28] manaullah, a.k. sharma, h. ahuja, and a. singh, “performance comparison of dfig and scig based wind energy conversion systems,” in innovative applications ofcomputational intelligence on power, energy and controls with their impact on humanity, pp. 285-290, november 2014. [29] s.k.tiwari, b. singh, and p.k. goel, “design and control of microgrid fed by renewable energy generating sources,”ieee transactions on industry applications,vol. 54, no. 3, pp. 2041-2050, 2018. [30] d.s. l. simonetti, a.e. a. amorim, and f.d. c. oliveira, “doubly fed induction generator in wind energy conversion systems,” inadvances in renewable energies and power technologies, pp. 461-490, 2018. [31] y. deng, z. xing, and q. zhang, “analysis of electromagnetic transient characteristics of doubly-fed induction generator under grid voltage swell,”cpss transactions on power electronics and applications,vol. 3, no. 2, pp. 111-118, 2018. [32] s. e. d. silveira, s. m. silva, and b. j. c. filho, “fault ride-through enhancement in dfig with control of stator flux using minimized series voltage compensator,”iet renewable power generation,pp.1234-1240, april 2018. [33] a. panda and m. tripathy, “optimal power flow solution of wind integrated power system using modified bacteria foraging algorithm,”international journal of electrical power and energy systems,vol. 54, pp. 306-314, 2014. [34] e.b. muhando, t. senjyu, a. uehara, and t. funabashi, “gain-scheduled h ∞ control for wecs via lmi techniques and parametrically dependent feedback part ii: controller design and implementation,”ieee transactions on industrial electronics,vol. 58, no. 1, pp. 57-65, 2011. proceedings of engineering and technology innovation, vol. 13, 2019, pp. 10-19 19 [35] m. yamamoto and o. motoyoshi, “active and reactive power control for doubly-fed wound rotor induction generator,” ieee transaction on power electronics, vol. 6, no. 4, pp. 624-629, 1991. copyright© by the authors. licensee taeti, taiwan. this article is an open access article distributed under the terms and conditions of the creative commons attribution (cc by-nc) license (https://creativecommons.org/licenses/by-nc/4.0/).  proceedings of engineering and technology innovation, vol. 15, 2020, pp. 42 51 a learning-based em clustering for circular data with unknown number of clusters shou-jen chang-chien, wajid ali, miin-shen yang * department of applied mathematics, chung yuan christian university, taoyuan, taiwan received 04 february 2020; received in revised form 04 march 2020; accepted 21 april 2020 doi: https://doi.org/10.46604/peti.2020.5241 abstract clustering is a method for analyzing grouped data. circular data were well used in various applications, such as wind directions, departure directions of migrating birds or animals, etc. the expectation & maximization (em) algorithm on mixtures of von mises distributions is popularly used for clustering circular data. in general, the em algorithm is sensitive to initials and not robust to outliers in which it is also necessary to give a number of clusters a priori. in this paper, we consider a learning-based schema for em, and then propose a learning-based em algorithm on mixtures of von mises distributions for clustering grouped circular data. the proposed clustering method is without any initial and robust to outliers with automatically finding the number of clusters. some numerical and real data sets are used to compare the proposed algorithm with existing methods. experimental results and comparisons actually demonstrate these good aspects of effectiveness and superiority of the proposed learning-based em algorithm. keywords: clustering, circular data, mixtures of von mises distributions, em algorithm, learning schema 1. introduction since von mises [1] proposed a distribution on circular data, watson and williams [2] considered statistical inferences for von mises distributions. afterward, circular data were widely applied in biology, geology, medicine, oceanography, and meteorology [3-5]. clustering is a useful tool for data mining. according to the statistical point of view, clustering methods can be generally divided into two categories. one is a model-based approach. another is a nonparametric approach. in the model-based approach, the expectation and maximization (em) algorithm [6-8] is the most used method. for a nonparametric approach, an objective function of dissimilarity measures is generally considered in which partitional clustering is the most popular, such as k-means [9-11], fuzzy c-means (fcm) [12-13], and possibilistic c-means (pcm) [14-15]. generally, the em algorithm was popularly used in the analysis of circular data based on mixtures of von mises distributions [16-19]. however, these em clustering algorithms are sensitive to initials and not robust to outliers with a given number of clusters a priori. in this paper, we construct a learning-based schema for em, and then propose a learning-based em algorithm on mixtures of von mises distributions for clustering circular data that is free of initials and robust to outliers. the proposed clustering method can also automatically find an optimal number of clusters. we apply the proposed algorithm to real circular data and several comparisons are given to demonstrate its effectiveness and superiority. 2. learning-based em clustering for circular data in yang et al. [20], they proposed a robust em clustering for gaussian mixture models. the algorithm can automatically find an optimal cluster number of data sets and solve the drawbacks of em that is sensitive to initial values. in this paper, we * corresponding author. e-mail address: msyang@math.cycu.edu.tw proceedings of engineering and technology innovation, vol. 15, 2020, pp. 42 51 43 will modify the approach proposed by yang et al. [20] such that it can handle circular data. we mention that yang et al. [21] used the idea of the robust em clustering on the mixtures of von mises-fisher distributions. we know that von mises-fisher distributions are well used to handle data higher than three dimensions, i. e. data on the unit hypersphere. however, the mixtures of von mises distributions are well for handling circular data. in this paper, we especially focus on circular data based on the mixtures of von mises distributions with angles measured by sine and cosine that is different from yang et al. [21]. suppose that the circular data set 𝑋 = {𝜃1, 𝜃2, ⋯ , 𝜃𝑛} is a random sample of size n from a mixture model 1 ( ; , ) ( ; )       c i i i i f a a f (1) where ai denotes the proportion of the ith class with 1 1   c ii a (2) and 𝑓𝑖 is the pdf of ith class with parameter ∅𝑖 . let class variable 𝑧 = {𝑧1, 𝑧2, ⋯ , 𝜃𝑐 } with 𝑧𝑖 = {𝑧𝑖1, 𝑧𝑖2, ⋯ , 𝑧𝑖𝑛 } 𝑇 where 𝑧𝑖𝑗 = 𝑧𝑖 (𝜃𝑗 ) = 1 if 𝜃𝑗 arises from the ith class and 𝑧𝑖𝑗 = 0 if 𝑥𝑗 arises from other class, for 𝑖 = {1, 2, ⋯ , 𝑐} and 𝑗 = {1, 2, ⋯ , 𝑛}. thus, the joint probability density of the complete data { 𝜃1, 𝜃2, ⋯ , 𝜃𝑛 , 𝑧1, 𝑧2, ⋯ , 𝑧𝑐 } shall be 1 2 1 2 1 1 ( , , , , , , ; , ) [ ( ; )]           ijn c n c j i i i j i z f z z z a a f (3) the log-likelihood for the complete data is given by 1 2 1 2 1 1 ( , ; , , , , , , ) (ln ( ; ))           n c em n c j i ij i i j i l a z z z z a f (4) the em algorithm consists of two steps with “expectation”, shorten by e, and “maximum”, shorten by m. in the e step, the expectation 1 ( ) ( ) / ( )        c ij j i i j i s s s j s e z a f a f (5) is used to substitute for missing data 𝑧𝑖𝑗 . in the m step, estimates of parameters are used by maximizing 1 1 ( ) ( )(ln ( ; ))        c n em i j ij j i i j i e l e z a f (6) with the restriction of eq. (2). the von mises distribution is the most common used distribution on circular data. if we consider that the density 𝑓𝑖 = (𝜃; ∅𝑖 ) is the von mises distribution 𝑉𝑀 = (𝑣𝑖 ; 𝑘𝑖 ) with mean direction 𝑣𝑖 and the concentration 𝑘𝑖 with -1 0 ( ; ) ( ; , )=(2 ( )) exp ( cos( ))      i i i i i i i i f f v k i k k v , 0 2 , 0       i k (7) 2 -1 0 0 ( ) (2 ) exp ( cos( )     i i ii k k v d (8) is the modified bessel function of order zero, then update equations for 𝑎𝑖 , 𝑣𝑖 , and 𝑘𝑖 can be obtained as follows: 1 /    n j i ia z j n (9) 11 1 sin( ) tan ( ) cos( )        n j ij n j ij j i j z v z (10) proceedings of engineering and technology innovation, vol. 15, 2020, pp. 42 51 44 11 1 cos( ) ( ) )        n j ij n j ij j i i z v k a z (11) where 𝐴−1(𝑥) a function that can be computed from batschelet’s table (see fisher [3]). in order to solve drawbacks for em, yang et al. [20] give an entropy term of the proportion to the em log-likelihood function. according to their idea, we propose the learning-based em objective function for mixtures of von mises distributions as follows: 1 n 1 11 ( , , , , , , ) (ln ( ; , )) ln          c c c ij i i j i i i j ii j z a v k z a f v k n a (12) thus, in the e step, the expectation �̂�𝑖𝑗 = 𝐸(𝑧𝑖𝑗 |𝜃𝑗 ) is used to substitute for missing data 𝑧𝑖𝑗 with 1 ˆ ( ) ( ; , ) / ( ; , )       c ij ij j i i j i i s s j s s s z e z a f v k a f v k (13) in the m-step, under the constraint eq. (2), we need to maximize 1 n 1 11 ( ( , , , ; , , )) (ln ( ; , )) ln          c c c ij i i j i i i j ii e j z a v k z a f v k n a (14) taking partial derivatives with respect to 𝑎𝑖 , and then the following update equation is obtained ( ) ( ) ( ) ( ) ( ) 1 (ln ln       c new em old old old old i i i i s s s a a a a a a (15) where 1 ˆ /   nem i ijj a z n (16) eq. (15) provides us a method to seek the number of clusters. because the number of clusters is not larger than n, ( ) 1/ new i a n (17) or ( ) 0 new i a (18) is not reasonable. in this situation, we must discard the ith class. thus, the new number of clusters can be obtained with ( ) ( ) ( ) ( ) ( ) | { | 1/ , 1, 2, , }|    new old new new old i i c c a a n i c (19) in order to satisfy the constraints ( ) 1 1 new c kk a   (20) and ( ) 1 ˆ 1 new c k jk z   (21) we adjust 𝑎𝑘′ and �̂�𝑘′𝑗 by proceedings of engineering and technology innovation, vol. 15, 2020, pp. 42 51 45 ( ) 1 = /    new c k k s s a a a (22) ( 1) 1 ˆ ˆ ˆ= /      t c k j k j sj s z z z (23) the purpose of the parameter β is to control competition of proportions. following yang et al. [20], we set β as follows: ( ) ( ) ( ) ( )1 ( ) ( ) ( ) exp (| ) 1 min{ , }      new c new old emi i ci old old c n a a a c a e (24) where ( ) ( ) ( ) max old old c i i a a (25) ( ) max em em c i i a a (26) ( ) ( ) ( ) 1 ln   old c old old i ii e a a (27) when the cluster number c is stable, let 𝛽 = 0. we also use the m-step to estimate 𝑣𝑖 and 𝑘𝑖 . we take partial derivatives with respect to 𝑣𝑖 and 𝑘𝑖 , respectively. then, update equations are obtained as follows: 1 1 1 tan ( sin( ) / cos( ))        n n i ij j ij j j j v z z (28) 1 1 1 ( cos( ) / )       n n i ij j j ij j j k a z v z (29) to solve the initialization problem, we let the initial number of clusters be n. it means that (0) (0) (0) 1 2 1 2 ( , , , ) ( , , , )   n n v v v (30) (0) (0) (0) 1 2 ( , , , ) (1/ , 1/ , , 1/ ) n a a a n n n (31) we then consider the maximum likelihood estimate of the concentration of von mises distribution as initial value 𝑘 𝑖 (0) with (0) 1 ( ), 1, ,    i k a r i c (32) where 2 2 1/ 2 1 1 1 1 [( cos )] ( sin )       n n j j r j j n n (33) is the sample mean resultant length. thus, the proposed learning-based em for circular data can be summarized as follows: learning-based em algorithm for circular data step 1: fix 𝜀 > 0. give initials 𝛽(0) = 1, 𝑐 (0) = 𝑛, 𝑎𝑖 (0) = 1 𝑛 , and assign (𝑣1 (0) , 𝑣2 (0) , ⋯, 𝑣𝑛 (0) )= (𝜃1, 𝜃2, ⋯, 𝜃𝑛). step 2: compute 𝑘𝑖 (0) using eq. (32) and set 𝑡 = 1. step 3: compute �̂�𝑖𝑗 (0) with (0) i a , (0) i  , and (0) i  using eq. (13). proceedings of engineering and technology innovation, vol. 15, 2020, pp. 42 51 46 step 4: compute 𝑣𝑖 (𝑡) with �̂�𝑖1 (𝑡−1) , �̂�𝑖2 (𝑡−1) , ⋯, �̂�𝑖𝑛 (𝑡−1) using eq. (28). step 5: update to 𝑎𝑖 (𝑡) with �̂�𝑖1 (𝑡−1) , �̂�𝑖2 (𝑡−1) , ⋯, �̂�𝑖𝑛 (𝑡−1) and 𝑎𝑖 (𝑡−1) using eq. (15). step 6: compute 𝛽(𝑡) with 𝑎𝑖 (𝑡) and 𝑎𝑖 (𝑡−1) using eq. (24). step 7: update 𝑐 (𝑡−1) to 𝑐 (𝑡) by discard those clusters with 𝑎𝑖 (𝑡) ≤ 1/𝑛 and adjust 𝑎𝑖 (𝑡) and �̂�𝑖𝑗 (𝑡−1) by eqs. (22) and (23). if 𝑡 ≥ 60 and 𝑐(𝑡−60) − 𝑐 (𝑡) = 0, then let 𝛽(𝑡) = 0. step 8: compute �̂�𝑖 (𝑡) with �̂�𝑖1 (𝑡−1) , �̂�𝑖2 (𝑡−1) , ⋯, �̂�𝑖𝑛 (𝑡−1) using eq. (29). step 9: compute �̂�𝑖𝑗 (𝑡) with 𝑎𝑖 (𝑡) , 𝑣𝑖 (𝑡) , and �̂�𝑖 (𝑡) using eq. (13). step 10: compute 𝑣𝑖 (𝑡+1) with �̂�𝑖1 (𝑡) , �̂�𝑖2 (𝑡) , ⋯, �̂�𝑖𝑛 (𝑡) using eq. (28). step 11: compute 𝑑(𝑣 (𝑡+1), 𝑣 (𝑡)) in a convenient norm 𝑑. if 𝑑(𝑣 (𝑡+1), 𝑣 (𝑡)) < 𝜀, stop else 𝑡 = 𝑡 + 1 and return to step 5. 3. examples and comparisons in this section, we make comparisons between the proposed learning-based em and em algorithms on numerical and real data sets. example 1: in fig. 1(a), there is a 2-cluster data set which is generated from a mixture of two von mises distributions 0.4vm(60°, 6.5)+0.6vm(180°, 7.5) with 200 data points. if the angle is an observation, then x-coordinate represents cos and y-coordinate represents sin. we first implement em with c = 2 for this data set. the clustering results of em are shown in fig. 1(b). the data set is well separated, and so we obtain the same clustering results of em with most random initial values. the estimates parameters are as: �̂�1 = 0.4274, �̂�2 = 0.5726, �̂�1 = 61.0254 °, �̂�2 = 178.2695 °, �̂�1 = 6.54, �̂�2 = 7.426. we can see that em has good clustering results and estimates close to the given parameters. we use the learning-based em without a priori cluster number. figs. 1(c)-1(f) are the states of the cluster centers (denoted by “*” symbol) at iteration = 0, 5, 10, 20, respectively. the cluster centers also imply the mean directions. we can see that the cluster number decreases from 200 to 2. therefore, the learning-based em can find an optimal cluster number with c * =2. the clustering results of the learning-based em are the same with em (with a priori cluster number c=2), as shown in fig. 1(b). the parameter estimates of the learning-based em are as: �̂�1 = 0.4257, �̂�2 = 0.5743, �̂�1 = 60.8349 °, �̂�2 = 178.1072 °, �̂�1 = 6.54, �̂�2 = 7.426. the results are closed to those of em. to compare the performance of the proposed learning-based em with em, we use the criterion of mean squared error (mse) to evaluate the accuracy of the two algorithms. we generate 100 data sets from 0.4vm (60°, 6.5)+0.6vm(180°, 7.5) and computer mses from em and the learning-based em. the mse is defined as sum of se of each data set for mse 100   where the squared error (se) for v is defined as 2 2 1 2 ˆ ˆ( 60) ( 180) se 2      from the 100 data sets, we obtain that the mse value from em is 2.3561, and the mse value from the learning-based em is 2.2193. the proposed learning-based em performs better than em. proceedings of engineering and technology innovation, vol. 15, 2020, pp. 42 51 47 (a) 2-clusters data set (b) clustering results of em and the learning-ased em (c) the states of cluster centers from the learning-based em at iteration = 0 (d) the states of cluster centers from the learning-based em at iteration = 5 (e) the states of cluster centers from the learning-based em at iteration = 10 (f) the states of cluster centers from the learning-based em at iteration = 20 fig. 1 2-cluster circular data set with their clustering results furthermore, we consider the influence of noisy points on the proposed learning-based em and em algorithms. we add 50 randomly noisy points which are generated from the interval [0°, 240°] into 2-cluster data set and call noisy data set, as shown in fig. 2(a). the noisy points are denoted by the "+" symbol. we first implement em with c = 2 for the noisy data set. fig. 2(b) shows the clustering results of em. compare fig. 2(b) with fig. 1(b), we find that one of the original 200 data points is classified into another cluster. the estimates parameters are as follows: �̂�1 = 0.4223 , �̂�2 = 0.5477 , �̂�1 = 61.9240 ° , �̂�2 = 178.4338 °, �̂�1 = 4.8590, �̂�2 = 6.5400. we also implement the learning-based em with unknown number c of clusters for the noisy data set. the final states of the cluster centers are shown in fig. 2(c). the learning-based em find the optimal cluster number c * =2 after 51 iterations. the clustering results of the learning-based em are shown in fig. 2(d). compare fig. 2(d) with fig. 1(b), we find that two clusters in the original 200 data points are the same. the estimates parameters are as -1.5 -1 -0.5 0 0.5 1 1.5 -1.5 -1 -0.5 0 0.5 1 1.5 x y -1.5 -1 -0.5 0 0.5 1 1.5 -1.5 -1 -0.5 0 0.5 1 1.5 x y -1.5 -1 -0.5 0 0.5 1 1.5 -1.5 -1 -0.5 0 0.5 1 1.5 x y iteration = 0 -1.5 -1 -0.5 0 0.5 1 1.5 -1.5 -1 -0.5 0 0.5 1 1.5 x y iteration = 5 -1.5 -1 -0.5 0 0.5 1 1.5 -1.5 -1 -0.5 0 0.5 1 1.5 x y iteration = 10 -1.5 -1 -0.5 0 0.5 1 1.5 -1.5 -1 -0.5 0 0.5 1 1.5 x y iteration = 20 (convergence) proceedings of engineering and technology innovation, vol. 15, 2020, pp. 42 51 48 follows: �̂�1 = 0.4490, �̂�2 = 0.5510, �̂�1 = 61.1672 ° , �̂�2 = 178.1286 ° , �̂�1 = 4.8590, �̂�2 = 5.8520. to compare the influences of noisy points on the two algorithms, we compute the increasing rate of mse. we generate 50 randomly noisy points from the interval [0°, 240°] for the data set. then we compute mses for em and the learning-based em from 100 noisy data sets. the increasing rate is defined as (the noisy data set) (the original data set) (the original data set) mse mse increasing rate 100% mse    (34) table 1 shows these mses with its increasing rate. from table 1, we find the increasing rate of the learning-based em is smaller than em, which means that the noisy points have a smaller influence on the learning-based em. (a) noisy data set (b) clustering results of em (c) the final states of cluster centers of the learning-based em (d) clustering results of the learning-based em fig. 2 2-cluster data set with 50 noisy points and their clustering results table 1 mses for em and the learning-based em with its increasing rates algorithms em learning-based em mse(the original data set) 2.3561 2.2193 mse (the noisy data set) 2.8719 2.4621 increasing rate 21.89% 10.94% example 2: in this example, we apply the proposed learning-based em and em algorithms to a real data set, sudden infant death syndrome (sids) data of mooney et al. [17]. the sids data set consists of the sids cases in the uk by a month of death for the years 1983-1998, exhibiting a seasonal pattern with a winter peak. mooney et al. [17] modified all of 12 months to 31 days, i.e. the sids data were month-corrected to 31 days by multiplying february by 1.097 and the 30 day months by 1.033. then they found that the sids data for the year 1998 could be fitted by a mixture of two von misses distributions. table 2 shows the corrected number of sids cases for the year 1998. we transform table 2 to circular data. we map the 12 months from 0 to 360 in degree. every month has equal length (31 days), and so [0, 360] should be divided into 12 equal length intervals. january corresponds to the interval [0°, 30°], february corresponds to the interval [30°, 60°] and so on. based on -1.5 -1 -0.5 0 0.5 1 1.5 -1.5 -1 -0.5 0 0.5 1 1.5 x y -1.5 -1 -0.5 0 0.5 1 1.5 -1.5 -1 -0.5 0 0.5 1 1.5 x y -1.5 -1 -0.5 0 0.5 1 1.5 -1.5 -1 -0.5 0 0.5 1 1.5 x y iteration = 51 (convergence) -1.5 -1 -0.5 0 0.5 1 1.5 -1.5 -1 -0.5 0 0.5 1 1.5 x y proceedings of engineering and technology innovation, vol. 15, 2020, pp. 42 51 49 frequency in each month, we generate random numbers from the uniform distribution on the corresponding interval. 402 observations and its rose diagram (circular histogram) are shown in fig. 3(a) and fig. 3(b), respectively. clearly, we can se e two modes on the intervals [150°, 180°] and [330°, 360°] from fig. 3(b). we implement em with cluster number c=2 for the data set in fig. 3(a). the clustering results of em are affected by different initials. figs. 3(c)-3(e) are results with different initials. if we implement em for this data set with 100 random initials, then we obtain 23 of 100 results as shown in fig. 3(c), 72 of 100 results as shown in fig. 3(d), and 5 of 100 results as shown in fig. 3(e). we also implement the learning-based em without giving c for the data set in fig. 3(a). fig. 3(f) shows the cluster number decreases from 402 to 2. therefore, the learning-based em finds an optimal cluster number c * =2. the final states of cluster centers are shown in fig. 3(g). we can see that the locations of two mean directions are consistent with two modes in fig. 3(b). the clustering results of the learning-based em are shown in fig. 3(h). obviously, the results are different from em. note that the learning-based em sets all data points as initials. thus, an initial condition is not necessary for the learning-based em. if we perform the learning-based em for 100 times, then 100 clustering results are the same. the estimated parameters of the learning-based em, em (with fig. 3(d)) and mooney et al. [17] are shown in table 3. we find the estimates of the learning-based em are closed to mooney et al. [17]. according to mean direction estimates, we know that there are two peaks in sids cases for the year 1998. one peak is in early june and the other is in early december. table 2 corrected number of sids case for the year 1998 jan feb mar apr may jun jul aug sep oct nov dec total 40 31 25 26 29 33 25 20 27 40 43 63 402 table 3 estimates of parameters for sids data algorithms em learning-based em mooney et al. [17] �̂�1 153.32° 151.84° 154.48° �̂�2 337.66° 340.33° 340.52° (a) the sids data set (b) rose diagram to sids data for the year 1998 (c) clustering results of em with different initials (d) clustering results of em with different initials fig. 3 clustering results of learning-based em and em algorithms for sids data set -1.5 -1 -0.5 0 0.5 1 1.5 -1.5 -1 -0.5 0 0.5 1 1.5 x y 20 40 60 80 30 210 60 240 90 270 120 300 150 330 180 0 -1.5 -1 -0.5 0 0.5 1 1.5 -1.5 -1 -0.5 0 0.5 1 1.5 x y -1.5 -1 -0.5 0 0.5 1 1.5 -1.5 -1 -0.5 0 0.5 1 1.5 x y proceedings of engineering and technology innovation, vol. 15, 2020, pp. 42 51 50 (e) clustering results of em with different initials (f) cluster number obtained by the learning-based em under different iterations (g) final states of cluster centers of the learning-based em (h) clustering results of learning-based em fig. 3 clustering results of learning-based em and em algorithms for sids data set (continued) 4. conclusions in general, the em algorithm for clustering circular data quite depends on initializations with a given number of clusters a priori. to solve these drawbacks of em, a learning-based em algorithm is proposed for circular data in this paper. the proposed learning-based em can automatically find an optimal number of clusters for different circular data sets without any initialization and also robust to outliers. some numerical data from mixtures of von mises distributions and real circular data are used for comparisons of the proposed learning-based em with em algorithms. furthermore, we also consider noisy points and outliers in data sets. these comparisons and experimental results actually demonstrate the effectiveness and usefulness of the proposed learning-based em clustering algorithm for circular data. although the proposed learning-based em works well for circular data, it is only fitted for 2-dimensional angular data sets. in our future work, we will further construct a learning-based em algorithm for spherical data sets, i.e. 3-dimensional angular data, and then apply it to extrasolar planet taxonomy. conflicts of interest the authors declare no conflict of interest references [1] r. von mises, “uber die ‘ganzzahligkeit’ der atomgewicht und verwandte fragen,” physikal z., vol. 19, pp. 490-500, 1918. [2] g. s. watson and e. j. williams, “on the construction of significance tests on the circle and the sphere,” biometrika, vol. 43, no. 3/4, pp. 344-352, december 1956. [3] n. i. fisher, statistical analysis of circular data, cambridge:cambridge university, in press, october 1995. -1.5 -1 -0.5 0 0.5 1 1.5 -1.5 -1 -0.5 0 0.5 1 1.5 x y -1.5 -1 -0.5 0 0.5 1 1.5 -1.5 -1 -0.5 0 0.5 1 1.5 x y iteration = 39 (convergence) -1.5 -1 -0.5 0 0.5 1 1.5 -1.5 -1 -0.5 0 0.5 1 1.5 x y 0 5 10 15 20 25 30 35 40 2 50 100 150 200 250 300 350 400 450 iteration c lu st e r n u m b e r proceedings of engineering and technology innovation, vol. 15, 2020, pp. 42 51 51 [4] n. masseran, a. m. razali, k. ibrahim, and m. t. latif, “fitting a mixture of von mises distributions in order to model data on wind direction in peninsular malaysia,” energy conversion and management, vol. 72, pp. 94-102, april 2013. [5] l. p. rivest and s. kato, “a random‐effects model for clustered circular data,” canadian journal of statistics, vol. 47, no. 4, pp. 712-728, august 2019. [6] a. p. dempster, n. m. laird, and d. b. rubin, “maximum likelihood from incomplete data via the em algorithm,” journal of the royal statistical society, series b, vol. 39, no. 1, pp. 1-22, september 1977. [7] g. j. mclachlan and k. e. basford, mixture models: inference and applications to clustering, new york: marcel dekker, 1988. [8] j. yu, c. chaomurilige, and m. s. yang, “on convergence and parameter selection of the em and da-em algorithms for gaussian mixtures,” pattern recognition, vol. 77, pp. 188-203, december 2017. [9] j. macqueen, “some methods for classification and analysis of multivariate observations,” proc. 5th berkeley symposium on mathematical statistics and probability, university of california, in press, vol. 1, no. 14, pp. 281-297, june 1967. [10] d. pollard, “quantization and the method of k-means,” ieee transaction on information theory, vol. 28, no. 2, pp. 199-205, march 1982. [11] m. s. yang and k. p. sinaga, “a feature-reduction multi-view k-means clustering algorithm,” ieee access, vol. 7, pp. 114472-114486, august 2019. [12] j. c. bezdek, pattern recognition with fuzzy objective function algorithms, new york: plenum, in press, july 1981. [13] m. s. yang and y. nataliani, “a feature-reduction fuzzy clustering algorithm with feature-weighted entropy,” ieee transactions on fuzzy systems, vol. 26, no. 2, pp 817-835, april 2018. [14] r. krishnapuram and j. m. keller, “a possibilistic approach to clustering,” ieee transaction on fuzzy systems, vol. 1, no. 2, pp. 98-110, may 1993. [15] m. s. yang, s. j. chang-chien, and y. nataliani, “a fully-unsupervised possibilistic c-means clustering method,” ieee access, vol. 6, pp. 78308-78320, december 2018. [16] r. bartels, “estimation in a bidirectional mixture of von mises distributions,” biometrics, vol. 40, no. 3, pp. 777-784, september 1984. [17] j. a. mooney, p. j. helms, and i. t. jolliffe, “fitting mixtures of von mises distributions: a case study involving sudden infant death syndrome,” computational statistics and data analysis, vol. 41, no. 3/4, pp. 505-513, october 2002. [18] n. sanusi, a. zaharim, s. mat, and k. sopian, “a weibull and finite mixture of the von mises distribution for wind analysis in mersing, malaysia,” international journal of green energy, vol. 14, no. 12, pp. 1057-1062, september 2017. [19] y. ban, x. alameda-pineda, c. evers, and r. horaud, “tracking multiple audio sources with the von mises distribution and variational em,” ieee signal processing letters, vol. 26, no. 6, pp. 798-802, march 2019. [20] m. s. yang, c. y. lai, and c. y. lin, “a robust em clustering algorithm for gaussian mixture models,” pattern recognition, vol. 45, no. 11, pp. 3950-3961, may 2012. [21] m. s. yang, s. j. chang-chien, and w. l. hung, “learning-based em clustering for data on the unit hypersphere with application to exoplanet data,” applied soft computing, vol. 60, pp. 101-114, june 2017. copyright© by the authors. licensee taeti, taiwan. this article is an open access article distributed under the terms and conditions of the creative commons attribution (cc by-nc) license (https://creativecommons.org/licenses/by-nc/4.0/). 1___peti#8420__01-11 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 01-11 a method based on only currents for determining fault direction in radial distribution networks integrated with distributed generations ngo minh khoa * , tran xuan khoa faculty of engineering and technology, quy nhon university, binh dinh, vietnam received 05 september 2021; received in revised form 12 december 2021; accepted 13 december 2021 doi: https://doi.org/10.46604/peti.2021.8420 abstract nowadays, more distributed generations (dgs) are connected to a radial distribution network, so conventional overcurrent relays cannot operate correctly when a fault occurs in the network. this study proposes a method to determine the fault direction in a three-phase distribution network integrated with dgs. the obtained pre-fault and fault currents are utilized to extract their phasors by the fast fourier transform, and the phase angle difference between the positive-sequence components of the pre-fault and fault currents is used. moreover, the method only uses the local current measurement to calculate and identify the phase angle change of the fault current without using the voltage measurement. matlab/simulink software is used to simulate the three-phase distribution network integrated with dgs. the faults with different resistances are assumed to occur at backward and forward fault locations. the simulation results show that the proposed method correctly determines the fault direction. keywords: directional protection relay, distributed generation, fault direction, positive-sequence current, radial distribution network 1. introduction integrating distributed generations (dgs) into a radial distribution network has many significant benefits, such as reducing the total power loss, improving the voltage quality, etc. however, there are some existing problems on a radial distribution network integrated with dgs, i.e., the power flow, control, operation, reliability, security, and protection problem of the network [1-2]. an adaptive quantum-inspired evolutionary algorithm was then proposed to improve the power flow and voltage profile in the distribution network integrated with dgs [3-6]. in addition, the protection problem of the distribution network integrated with dgs is studied in many works [7-10]. conventional overcurrent protection relays are usually used to protect a radial distribution network. many researchers only use the current measurement at the relays via current transformers (cts) to operate when the current measurement exceeds the pickup values set in the relays. because the power flow of the distribution network integrated with dgs is changed depending on the penetration level of the network sources, the conventional overcurrent protection relays will not correctly determine the fault direction when a fault occurs in the network [7]. in horak’s work [8], a scheme of directional overcurrent relays was designed based on the phase relationship of voltages and currents at the relay location to determine the fault direction. in the work of jang et al. [9], an adaptive approach for relay protection was applied to a distribution network integrated with a wind farm. the method proposed in eissa’s work [10] * corresponding author. e-mail address: ngominhkhoa@qnu.edu.vn tel.: +84 988 371 737 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 01-11 2 utilized a novel current polarized directional element technique to determine the fault direction on a transmission line. to completely overcome the protection problem, the directional overcurrent protection relays are applied to ensure the selection factor of the relays in the situation. for the directional overcurrent protection relays, the direction of the fault current flowing through the relay location is determined by using the phase angle of voltage and current. the voltage polarity is used as a reference value for determining the fault direction [11]. therefore, voltage transformers are utilized to transfer the high voltage on the primary side to the low voltage on the secondary side and to feed the protection relays. similarly, cts are simultaneously utilized to feed the small-scale currents to the relays. in general, when a short circuit occurs in the network, the fault current phasors at the relays are usually located in two distinct areas: the forward location and the backward location, as shown in fig. 1. this study will exploit the information from these two areas to develop an innovative method for determining the fault direction in a radial distribution network integrated with dgs. i iforward u ibackward fig. 1 two fault identification areas for directional overcurrent protection relays motivated by the above-mentioned works, an innovative method is developed in this study for determining the fault direction in a radial distribution network integrated with dgs. in summary, the main contributions of this study cover three aspects: (i) the proposed method only uses the current measurement at the relay location to determine the phase angle difference between the positive components of the pre-fault and fault currents; (ii) the proposed method is embedded in a relay to detect and protect all types of faults that occur in a radial distribution network integrated with dgs, including phase-to-ground (lg) faults, phase-to-phase (ll) faults, two-phase-to-ground (llg) faults, and three-phase-to-ground (lllg) faults; and (iii) the proposed method is more cost-effective for investing in voltage transformers, as compared to the conventional methods. the rest of this study is organized as follows. section 2 presents the literature review. section 3 describes the background methodology of the proposed method for determining the fault direction in a radial distribution network integrated with dgs. the simulation results and discussion are contained in section 4, and finally section 5 concludes this study. 2. literature review several published works are related to the methods applied to directional overcurrent protection relays in power systems [12-18]. voima et al. [12] presented an adaptive protection scheme to protect the medium voltage networks integrated with dgs, particularly in island operation modes. ukil et al. [13] proposed a novel approach to detect the possibility of fault direction using only the currents at the relays. yousfi et al. [14] developed a method based on the adaline neural network and the instantaneous power theory for extracting the online symmetrical components and phase angle from the fault current. eissa [15] proposed a new technique for directional overcurrent protection based on the post-fault current signals and directional reference current signals. the voltage measurement at the relays did not require determining the fault direction in the technique. proceedings of engineering and technology innovation, vol. 20, 2022, pp. 01-11 3 furthermore, samet et al. [16] developed a high-speed algorithm for determining the fault current, which only used the current measurement at the relays. in the algorithm, the sign of the summation of multiplied faults by the samples of pre-fault current and the direction of power flow in a normal power system was a criterion to determine the fault direction. the effectiveness of the method was shown by the speed for determining the fault direction in less than one-eighth of the cycle of power frequency. samet et al. [16] also proposed a novel directional overcurrent protection scheme for the distribution networks integrated with dgs. this protection scheme calculated the fault direction using a micro-genetic algorithm through numerical relays which were located on the network to protect and detect any changes in the configuration as well as recalculate the setting of directional overcurrent protection relays. on the other hand, to find the optimum relay setting for the minimum time to interrupt the power supply, nascimento et al. [17] and khond et al. [18] developed a new technique using the linear programming problem approach to optimize the relay setting in distribution networks. the power flow in the distribution networks with the penetration of dgs is changed depending on the level of penetration. the adaptive and flexible algorithms for directional overcurrent protection relays were mentioned in many publications [19-25]. brahma et al. [19] developed an adaptive protection scheme applied to directional overcurrent relays to protect the distribution networks with dgs. balyith et al. [20] proposed another novel protection scheme without the need of communication assistance to determine the relay setting to minimize the relays’ overall operating time. zhan et al. [21] proposed a genetic algorithm for the location and sizing optimization of dgs in distribution networks by investigating their relay protection coordination. another adaptive overcurrent coordination scheme based on the evolution algorithm was developed in the work of shih et al. [22] to enhance the relay sensitivity and overcome the drawbacks of dgs. jia et al. [23] presented an improved scheme based on high-frequency impedance to manage the adaptability problem when determining the fault direction in the network with inverter-interfaced renewable energy generations. to overcome the challenges of overcurrent protection in the distribution networks integrated with dgs, the local measurement information was used to detect the operating status and the faulted section in the network [24]. in the work of jones et al. [25], directional overcurrent relays were used to solve difficult problems for distribution feeder protection with high penetration of dgs. concerning the protection issue, ieee standard 1547-2003 [26] presents the criteria and requirements for the interconnection of dgs with power systems, and ieee standard p1547.4 [27] provides alternative approaches and good practices for the design, operation, and integration of dg island systems with power networks. therefore, these two ieee standards are considered in this work. 3. proposed method for determining fault direction 3.1. single-phase network the fault direction in a radial distribution network integrated with dgs can be determined by the proposed method, which uses the phase angle difference between the pre-fault and fault currents at the protection relay location. the method is developed based on the phase angle change of the fault current compared with the pre-fault current, as shown in fig. 2. the dg at the busbar a is connected to the grid source via two segments: line 1 (from the busbar a to the busbar b) and line 2 (from the busbar b to the busbar c). it is assumed that dg is generating the power to supply a load located at the busbar b and transferring it to the grid source via line 2. therefore, the power flow direction in the distribution network is normally from the busbar a to the busbar c. at the end b of line 2 (from the busbar b to the busbar c), a protection relay is established to get the current from ct. the proposed method is applied to the relay to determine the fault direction when a fault occurs in the lines. two fault locations, including the backward fault location (f1) on line 1 and the forward fault location (f2) on line 2, are investigated in the network as shown in fig. 2. proceedings of engineering and technology innovation, vol. 20, 2022, pp. 01-11 4 r relay zdg udg f1 f2 zgrid ugrid z11 z12 z21 z22 a b c ct i dg source grid source line 1 z1=z11+z12 line 2 z2=z21+z22 qdg pdg load cb2cb1 fig. 2 power network integrated with dgs when a fault occurs at f1 and f2, the pre-fault current and the fault current are calculated as follows. the pre-fault current at the relay location is: a c u u i z − = (1) where ua and uc are the voltages at the busbar a and the busbar c, respectively; z = z1 + z2 is the impedance from the busbar a to the busbar c. when an lllg fault occurs at f1, the fault current flow from the grid side to f1 is given as follows: c f1 f1 u i z = (2) where zf1 = z2 + z12 is the impedance from the busbar c to f1. similarly, when an lllg fault occurs at f2, the fault current flow from dg to f2 is given as follows: a f2 f2 u i z = (3) where zf2 = z1 + z21 is the impedance from the busbar a to f2. when a fault occurs at f1 and f2, the fault current flow through ct is determined as follows: 1 f1 i i i= − (4) substituting eqs. (1) and (2) into eq. (4), the following equations are established: a c c 1 f1 u u u i z z − = − (5) 2 f2 i i i= + (6) substituting eqs. (1) and (3) into eq. (6), the following equation is established: a c a 2 f2 u u u i z z − = + (7) it is assumed that the voltage at the busbar a (ua) and the voltage at the busbar c (uc) have the same voltage magnitude and phase angle. f1 and f2 are assumed at the busbar a and the busbar c, respectively. the fault impedances zf1 and zf2 are equal to the impedance of the network z. from eqs. (5) and (7), it is obvious that the currents i1 and i2 are opposite phase angles. in general, the relationship between these currents is depicted in fig. 3. proceedings of engineering and technology innovation, vol. 20, 2022, pp. 01-11 5 i1 ua uc i2if2if1 i ∆ϕ1 ∆ϕ2 note: ∆ϕ1 = ϕ1 – ϕ > 0 ∆ϕ2 = ϕ2 – ϕ < 0 -if1 fig. 3 vector diagram representing the relationship among the currents i1, i2, and i as can be seen in fig. 3, the phase angle difference between the pre-fault current and the fault current is positive (∆φ1 > 0) when a fault occurs at f1. reversely, the phase angle difference between the pre-fault current and the fault current is negative (∆φ2 < 0) when a fault occurs at f2 (∆φ2 < 0). 1 1 0ϕ ϕ ϕ∆ = − > (8) 2 2 0ϕ ϕ ϕ∆ = − < (9) 3.2. three-phase network for a three-phase power network, four types of faults (i.e., lg, ll, llg, and lllg faults) can appear in the network [28-30]. therefore, to perform fault analysis in the network as shown in fig. 2, a positive equivalent rule is used to calculate the fault current flow at f1 and f2. the positive-sequence equivalent diagrams for the fault locations at f1 and f2 are illustrated in figs. 4(a) and (b), respectively, where the additional impedance z∆ depends on the fault types shown in table 1. the proposed method for determining the fault direction in the distribution network integrated with dgs can be described as follows: step 1: faults are detected by comparing the fault current with the pickup current at the relay. step 2: the pre-fault and fault currents at the protection relay location are acquired. step 3: the phasors of the pre-fault and fault currents are estimated by using the fast fourier transform. step 4: the positive-sequence current components are computed from the phasors. step 5: the phase angle difference is calculated. step 6: the fault direction is determined according to the positive equivalent rule. zdg udg f1 zgr id ugr id z11 z12 z21 z22 a b ci pos z∆ ct zdg udg f2 zgr id ugr id z11 z12 z21 z22 a b c z∆ ct i pos (a) for the fault at f1 (b) for the fault at f2 fig. 4 positive-sequence equivalent circuit to calculate fault types proceedings of engineering and technology innovation, vol. 20, 2022, pp. 01-11 6 table 1 additional impedance of four fault types fault location fault type additional impedance z∆ f1 lg ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) zero zero zero zero zero neg neg neg neg neg dg 11 12 2 grid dg 11 12 2 grid zero zero zero zero zero neg neg neg neg neg dg 11 12 2 grid dg 11 12 2 grid z z z z z z z z z z z z z z z z z z z z z ∆ + + + + + + = + + + + + + + + + ll ( ) ( ) ( ) ( ) neg neg neg neg neg dg 11 12 2 grid neg neg neg neg neg dg 11 12 2 grid z z z z z z z z z z z ∆ + + + = + + + + llg ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) zero zero zero zero zero neg neg neg neg neg dg 11 12 2 grid dg 11 12 2 grid zero zero zero zero zero neg neg neg neg neg dg 11 12 2 grid dg 11 12 2 grid zero zero zero zero zero dg 11 12 2 grid zero dg 1 z z z z z z z z z z z z z z z z z z z z z z z z z z z z ∆ + + + + + + + + + + + + + + = + + + +( ) ( ) ( ) ( ) ( ) ( ) neg neg neg neg neg dg 11 12 2 grid zero zero zero zero neg neg neg neg neg 1 12 2 grid dg 11 12 2 grid z z z z z z z z z z z z z + + + + + + + + + + + lllg 0z ∆ = f2 lg ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) zero zero zero zero zero neg neg neg neg neg dg 1 21 22 grid dg 1 21 22 grid zero zero zero zero zero neg neg neg neg neg dg 1 21 22 grid dg 1 21 22 grid z z z z z z z z z z z z z z z z z z z z z ∆ + + + + + + = + + + + + + + + + ll ( ) ( ) ( ) ( ) neg neg neg neg neg dg 1 21 22 grid neg neg neg neg neg dg 1 21 22 grid z z z z z z z z z z z ∆ + + + = + + + + llg ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) zero zero zero zero zero neg neg neg neg neg dg 1 21 22 grid dg 1 21 22 grid zero zero zero zero zero neg neg neg neg neg dg 1 21 22 grid dg 1 21 22 grid zero zero zero zero zero dg 1 21 22 grid zero dg 1 z z z z z z z z z z z z z z z z z z z z z z z z z z z z ∆ + + + + + + + + + + + + + + = + + + +( ) ( ) ( ) ( ) ( ) ( ) neg neg neg neg neg dg 1 21 22 grid zero zero zero zero neg neg neg neg neg 21 22 grid dg 1 21 22 grid z z z z z z z z z z z z z + + + + + + + + + + + lllg 0z ∆ = 4. simulation results to evaluate the effectiveness of the proposed method, the three-phase power network presented in fig. 2 is simulated using the matlab/simulink software. the parameters of the power network elements are given in table 2. the nominal frequency of the network is 50 hz, and the nominal voltage is 22 kv. four types of faults, including lg, ll, llg, and lllg faults, are emulated at f1 and f2, respectively. it is assumed that each fault is started at the time of 0.1 seconds, and the total simulation time is set at 0.3 seconds. besides, the fault resistance at the fault locations is established at different values to analyze the performance of the proposed method. the simulation results of the fault current at the relay location are performed in the time domain. table 2 parameters of the three-phase power network element parameters dg source the frequency f = 50 hz the voltage u = 1.05 × 22 kv the short-circuit power sn = 300 mva the positive-sequence impedance z1 = 0.1613 + j0.0051 ω the zero-sequence impedance z0 = 0.4840 + j0.0154 ω grid source the frequency f = 50 hz the voltage u = 1.05 × 22 kv the short-circuit power sn = 500 mva the positive-sequence impedance z1 = 0.0968 + j0.0031 ω the zero-sequence impedance z0 = 0.2904 + j0.0092 ω line ab the line length l = 25 km the positive-sequence impedance z1 = 0.1153 + 0.3299 ω/km the zero-sequence impedance z0 = 0.413 + 1.0430 ω/km line bc the line length l = 25 km the positive-sequence impedance z1 = 0.1153 + 0.3299 ω/km the zero-sequence impedance z0 = 0.413 + 1.0430 ω/km load the nominal voltage u = 22 kv the power rating p = 10 mw the power factor pf = 1.0 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 01-11 7 the fault current in a circuit is changed from the original values in a steady-state operating mode to the faulty ones. thus, to carry out the fault simulations in this section, the steady-state voltages and currents of the network are calculated by the power flow of the network. it is assumed that dg is generating the power to the grid source via line 1 and line 2. the steady-state results of the distribution network are shown in table 3. it is obvious that the voltages at the three busbars are 1.042 pu, 0.9436 pu, and 1.043 pu, respectively. the current measurement at the overcurrent relay at the busbar b is the magnitude of 0.3599 pu and the phase angle of 42.39°. after establishing the initial steady-state values as shown in table 3, fault simulations are carried out to verify the proposed method above. the simulation results for the two fault locations (f1 on line 1 and f2 on line 2) are simulated and analyzed as follows. table 3 steady-state results of the distribution network bus name voltage current ua (pu) ub (pu) uc (pu) ia (pu) ib (pu) ic (pu) busbar a 1.042∠-2.028° 1.042∠-122° 1.042∠118° 1.132∠-7.822° 1.132∠-127.8° 1.132∠-112.2° busbar b 0.9436∠-25.1° 0.9436∠-145.1° 0.9436∠-94.9° 0.3599∠42.39° 0.3599∠-77.61° 0.3599∠162.4° busbar c 1.043∠-29.84° 1.043∠-149.8° 1.043∠90.16° 0.3563∠42.19° 0.3563∠-77.81° 0.3563∠162.2° the simulation results of the phase angle difference for f1 on line 1 are illustrated in fig. 5. figs. 5(a), (b), (c), and (d) show the results of lg faults, ll faults, llg faults, and lllg faults, respectively. for each case, five fault resistances (0, 5, 10, 15, and 20 ω) are set at the fault locations for evaluating the phase angle difference between the positive-sequence of the pre-fault current and the fault current at the relay location. it is obvious that the phase angle difference is dramatically increased from the value of 0 degrees at the time t = 0.1 seconds. after that, it reaches a positive value. this information is used to determine the backward fault direction. for f2 on line 2, the phase angle difference between the positive-sequence components of the pre-fault current and the fault current is shown in fig. 6. in this case study, five fault resistances (0, 5, 10, 15, and 20 ω) are also set at the fault locations. all the case studies are simulated in matlab/simulink and the results are shown in figs. 6(a), (b), (c), and (d) for the lg faults, ll faults, llg faults, and lllg faults, respectively. as can be seen in fig. 6, the phase angle difference is dramatically decreased at the time t = 0.1 seconds. it then reaches a negative value. therefore, the fault direction of f2 is the forward fault direction. in addition, the fault resistance at the two fault locations is varied from 0 to 20 ω with a step of 1 ω in each case study to evaluate the effectiveness of the proposed method. for each fault resistance, the apparent impedance from the grid source and the dg source to the fault location is also changed in both the magnitude and phase angle; these simulation results are shown in fig. 7. as can be seen in fig. 7, the x-axis and y-axis show the real and imaginary components of the fault current at the relay location, respectively. these simulation results show clearly that the red and blue nodes in fig. 7 represent the forward and backward faults, respectively. (a) lg faults (b) ll faults fig. 5 phase angle difference when faults occur at f1 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 01-11 8 (c) llg faults (d) lllg faults fig. 5 phase angle difference when faults occur at f1 (continued) (a) lg faults (b) ll faults (c) llg faults (d) lllg faults fig. 6 phase angle difference when faults occur at f2 fig. 7 two distinct areas for the backward faults and the forward faults 5. discussion in this study, a method is developed based only on the currents at the relay location for determining the fault direction in a radial distribution network integrated with dgs. a typical radial distribution network is modeled and simulated in the matlab/simulink software. the power flow of steady state in the network is the necessary initial condition to determine the proceedings of engineering and technology innovation, vol. 20, 2022, pp. 01-11 9 fault direction when faults occur in the network. two fault locations, including the backward and forward locations, are considered to confirm the capability of the proposed method. for each case study, the positive-sequence components of the pre-fault and fault currents are extracted using the fast fourier transform, and then the phase angle difference is calculated to determine the fault direction. this work also establishes four types of faults, including lg faults, ll faults, llg faults, and lllg faults, as well as the different fault resistances ranging from 0 to 20 ω. 6. conclusions this study proposes an innovative method for determining the fault direction in a radial distribution network integrated with dgs. the fast fourier transform is applied to extract the phasors of the pre-fault and fault currents at the relay location. the proposed method determines the fault direction based on the phase angle difference between the positive-sequence components of the pre-fault and fault currents at the relay location. the analysis results confirm that the phase angle difference is positive for the faults in the backward direction and negative for the faults in the forward direction. the effectiveness of this method is verified by performing the simulation of a three-phase radial distribution network integrated with dgs. four types of faults with different fault resistances and locations are simulated to evaluate the method. the simulation results confirm that the protection relay applied by this method determines the fault direction correctly. furthermore, because the proposed method only requires the local current measurement without the voltage measurement, it can be easily implemented in conventional non-directional overcurrent relays. the directional overcurrent relays applied by the proposed method can be utilized for future smart grids, displacing the traditional directional overcurrent relays that utilize the reference voltage phasors for estimating the fault direction. conflicts of interest the authors declare no conflict of interest. nomenclature ����� voltage of the grid source ��� �� zero-sequence impedance of the segment from f1 to b ����� impedance of the grid source ��� impedance of the segment from b to f2 � ���� � � negative-sequence impedance of the grid source � �� � � negative-sequence impedance of the segment from b to f2 ����� �� zero-sequence impedance of the grid source ��� �� zero-sequence impedance of the segment from b to f2 ��� voltage of the dg source ��� impedance of the segment from f2 to c �� active power of the dg source ��� � � negative-sequence impedance of the segment from f2 to c ��� reactive power of the dg source ��� �� zero-sequence impedance of the segment from f2 to c ��� impedance of the dg source �△ additional impedance � �� � � negative-sequence impedance of the dg source �� phase-a voltage ��� �� zero-sequence impedance of the dg source �� phase-b voltage ��� impedance of the segment from a to f1 �� phase-c voltage ��� � � negative-sequence impedance of the segment from a to f1 �� phase-a current ��� �� zero-sequence impedance of the segment from a to f1 �� phase-b current ��� impedance of the segment from f1 to b �� phase-c current � �� � � negative-sequence impedance of the segment from f1 to b references [1] w. a. elmore, protective relaying: theory and applications, 2nd ed., new york: marcel dekker, 2003. [2] k. kauhaniemi and l. kumpulainen, “impact of distributed generation on the protection of distribution networks,” 8th iee international conference on developments in power system protection, pp. 315-318, april 2004. proceedings of engineering and technology innovation, vol. 20, 2022, pp. 01-11 10 [3] g. manikanta, a. mani, h. p. singh, and d. k. chaturvedi, “simultaneous application of distributed generator and network reconfiguration for power loss reduction using an adaptive quantum inspired evolutionary algorithm,” international journal of energy technology and policy, vol. 17, no. 2, pp. 140-179, april 2021. [4] g. manikanta, a. mani, h. p. singh, and d. k. chaturvedi, “effect of voltage dependent load model on placement and sizing of distributed generator in large scale distribution system,” majlesi journal of electrical engineering, vol. 14, no. 4, pp. 97-121, december 2020. [5] g. manikanta, a. mani, h. p. singh, and d. k. chaturvedi, “adaptive quantum-inspired evolutionary algorithm for optimizing power losses by dynamic load allocation on distributed generators,” serbian journal of electrical engineering, vol. 16, no. 3, pp. 325-357, 2019. [6] g. manikanta, a. mani, h. p. singh, and d. k. chaturvedi, “placing distributed generators in distribution system using adaptive quantum inspired evolutionary algorithm,” 2nd international conference on research in computational intelligence and communication networks, pp. 157-162, september 2016. [7] p. b. eriksen, t. ackermann, h. abildgaard, p. smith, w. winter, and j. r. garcia, “system operation with high wind generation,” ieee power and energy magazine, vol. 3, no. 6, pp. 65-74, october 2005. [8] j. horak, “directional overcurrent relaying (67) concepts,” 59th annual conference for protective relay engineers, pp. 1-8, april 2006. [9] s. i. jang, j. h. choi, j. w. kim, and d. m. choi, “an adaptive relaying for the protection of a wind farm interconnected with distribution networks,” ieee pes transmission and distribution conference and exposition, pp. 296-302, september 2003. [10] m. m. eissa, “evaluation of a new current directional protection technique using field data,” ieee transactions on power delivery, vol. 20, no. 2, pp. 566-572, april 2005. [11] r. v. r. de carvalho, f. h. t. vieira, s. g. de araújo, and c. r. lima, “a protection coordination scheme for smart grid based distribution systems using wavelet based fault location and communication support,” ieee pes conference on innovative smart grid technologies, pp. 1-8, april 2013. [12] s. voima, h. laaksonen, and k. kauhaniemi, “adaptive protection scheme for smart grids,” 12th iet international conference on developments in power system protection, pp. 1-6, april 2014. [13] a. ukil, b. deck, and v. h. shah, “smart distribution protection using current-only directional overcurrent relay,” ieee pes innovative smart grid technologies conference europe, pp. 1-7, october 2010. [14] f. l. yousfi, d. o. abdeslam, t. bouthiba, n. k. nguyen, and j. merckle, “adaline for online symmetrical components and phase-angles identification in transmission lines,” ieee transactions on power delivery, vol. 27, no. 3, pp. 1134-1143, june 2012. [15] m. m. eissa, “current directional protection technique based on polarizing current,” international journal of electrical power and energy systems, vol. 44, no. 1, pp. 488-494, january 2013. [16] h. samet, t. ghanbari, m. a. jarrahi, and h. j. ashtiani, “efficient current-based directional relay algorithm,” ieee systems journal, vol. 13, no. 2, pp. 1262-1272, september 2018. [17] j. p. nascimento, n. s. d. brito, and b. a. souza, “an adaptive overcurrent protection system applied to distribution systems,” computers and electrical engineering, vol. 81, 106545, january 2020. [18] s. v. khond and g. a. dhomane, “optimum coordination of directional overcurrent relays for combined overhead/cable distribution system with linear programming technique,” protection and control of modern power systems, vol. 4, no. 1, pp. 1-7, april 2019. [19] s. m. brahma and a. a. girgis, “development of adaptive protection scheme for distribution systems with high penetration of distributed generation,” ieee transactions on power delivery, vol. 19, no. 1, pp. 56-63, january 2004. [20] a. a. balyith, h. m. sharaf, m. shaaban, e. f. el-saadany, and h. h. zeineldin, “non-communication based time-current-voltage dual setting directional overcurrent protection for radial distribution systems with dg,” ieee access, vol. 8, pp. 190572-190581, october 2020. [21] h. zhan, c. wang, y. wang, x. yang, x. zhang, c. wu, et al., “relay protection coordination integrated optimal placement and sizing of distributed generation sources in distribution networks,” ieee transactions on smart grid, vol. 7, no. 1, pp. 55-65, january 2016. [22] m. y. shih, a. conde, z. leonowicz, and l. martirano, “an adaptive overcurrent coordination scheme to improve relay sensitivity and overcome drawbacks due to distributed generation in smart grids,” ieee transactions on industry applications, vol. 53, no. 6, pp. 5217-5228, december 2017. [23] k. jia, z. yang, y. fang, t. bi, and m. sumner, “influence of inverter-interfaced renewable energy generators on directional relay and an improved scheme,” ieee transactions on power electronics, vol. 34, no. 12, pp. 11843-11855, december 2019. proceedings of engineering and technology innovation, vol. 20, 2022, pp. 01-11 11 [24] p. mahat, z. chen, b. bak-jensen, and c. l. bak, “a simple adaptive overcurrent protection of distribution systems with distributed generation,” ieee transactions on smart grid, vol. 2, no. 3, pp. 428-437, september 2011. [25] d. jones and j. j. kumm, “future distribution feeder protection using directional overcurrent elements,” ieee transactions on industry applications, vol. 50, no. 2, pp. 1385-1390, april 2014. [26] interconnecting distributed resources with electric power systems, ieee standard 1547-2003, 2011. [27] draft guide for design, operation, and integration of distributed resource island systems with electric power systems ieee standard p1547.4, 2011. [28] ieee recommended practice for conducting short-circuit studies and analysis of industrial and commercial power systems, ieee standard 3002.3-2018, 2019. [29] n. m. khoa, n. a. toan, and d. d. tung, “experimental study on fault ride-through capability of vsc-based hvdc transmission system,” archives of electrical engineering, vol. 70, no. 1, pp. 173-188, march 2021. [30] n. m. khoa and d. d. tung, “locating fault on transmission line with static var compensator based on phasor measurement unit,” energies, vol. 11, no. 9, 2380, september 2018. copyright© by the authors. licensee taeti, taiwan. this article is an open access article distributed under the terms and conditions of the creative commons attribution (cc by-nc) license (https://creativecommons.org/licenses/by-nc/4.0/). 5___peti#8556_in press proceedings of engineering and technology innovation, vol. 21, 2022, pp. 34-43 modified digital correlation technique for accurate phase measurement in multi-frequency bio-impedance analysis sruthi s * , rasika dhavse, jignesh sarvaiya department of electronics engineering, sardar vallabhbhai patel national institute of technology, gujarat, india received 25 september 2021; received in revised form 05 december 2021; accepted 06 december 2021 doi: https://doi.org/10.46604/peti.2021.8556 abstract in bio-impedance analysis (bia), high-frequency low-amplitude alternating current (ac) signals can incur time delays due to the capacitive nature of human cell membranes, and the characteristics of human tissues can be assessed from these delays in terms of phase changes. to accurately measure the phase changes, this work proposes a modified digital correlation-based phase measurement method. the accuracy of the general correlation technique is improved through digital direct synthesis (dds) and digital correlation of unipolar square input signals. the proposed method is established through memory management and frequency adjustment. the result shows that, compared to the existing methods, the proposed method needs fewer hardware components, has better accuracy of 0.2° and higher frequency compatibility from 5 khz to 1 mhz, and requires lower cost (140 usd). the method can be applied for the bia of all types of tissues (recently used in covid detection and care) and for the applications where efficient phase measurement is required. keywords: phase measurement, bio-impedance analysis, direct digital synthesis, unipolar square signal, digital correlation 1. introduction time delays can occur when signals pass through different media. these delays are depicted as the phase differences of the propagated signals with respect to the original signals generated from sources. the measurement of phase differences can efficiently help assess the characteristics of the media. thus, the delays between test signals and reference signals need to be determined precisely [1]. interferometry, phase doppler anemometry (pda) and ultrasound imaging, signal demodulation of bio-signals, surface acoustic wave (saw) sensing, coriolis mass flowmetry (cmf), space systems, etc. are some of the applications which demand a precise measurement of phase differences. a phase measurement system should be analyzed based on its accuracy, sensitivity, simplicity, portability, response time, operating frequency, and noise immunity. in the case where synchronous detection/frequency synchronization is to be carried out to test the reliability of phase measurement techniques, the test signals and reference signals are generated from the same source. in bio-impedance analysis (bia), high-frequency low-intensity sinusoidal alternating current (ac) signals are applied to the tissue under test (tut) through electrodes by a drive module, and the voltage response of tut is measured by a sense module through electrodes, as shown in fig. 1. the signal used for excitation suffers phase alteration while traveling through tut. this is caused due to the capacitive nature of cell membranes of the cells constituting tissues. the signal used for excitation is an ultra-weak signal which also gets affected by additive noise sources like motion artifacts, other bio-signals, etc. [2]. it is essential to employ a precise ac constant source to excite tissues [3]. for a very weak signal, the detection system should be highly sensitive and accurate. three dispersion mechanisms (α, β, and γ) are used to characterize the anomalous electric properties of * corresponding author. e-mail address:sruthisindu@gmail.com tel.: +917069778270 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 34-43 bio-materials when subjected to the external ac signals of low frequency to very high frequency. due to the capacitive nature of cell membranes, the tissue characteristics including intracellular, extracellular, and membranous properties are obtained only in the β dispersion range (5 khz to 1 mhz) [2, 4]. recently, the bia during covid care is a hot topic of research [5-6]. the purpose of the study is to devise a digital correlation-based phase measurement method for bia. the method is intended to be simple, portable, cost effective, accurate, and high frequency compatible. for the first time, the phase measurement for bio-impedance signals by using digital correlation method is demonstrated. tut sense module drive module drive electrodes sense electrodes bio-impedance device fig. 1 simple block diagram depicting bia 2. related work conventional phase measurement methods in bia, such as lissajous pattern analysis [7], virtual vector voltmeter [7], xor-filter [8], sine fitting [9], and zero crossing detectors [10], are easy to implement but are outdated. they are not feasible in portable devices and have frequency limitations. yang et al. [11] used an off-the-shelf chip ad8302 for bio-impedance phase measurement since it directly gives the magnitude and phase of the signal under study. this technique is expensive and is sensitive to direct current (dc) components, so it is not effective. olivarez et al. [12] utilized an lcr meter to measure bio-impedance; the meter is expensive and bulky, and is not application specific. analog phase measurement systems are characterized by bulky outdated designs, long-term voltage drifts, non-linearity, and cross talks at high frequencies. digital electronics and arm-based computational methods could minimize the effect of these characteristics, leading to an ultraprecision measurement. an accurate phase measurement is traditionally realized by using digital signal processing (dsp). the typical digital-type phase measurement schemes include digital zero crossing detector, fast fourier transform (fft), quadrature detection, phase-locked loop (pll), and correlation. hu et al. [13] used a zero-crossing algorithm for the sake of simplicity and accuracy, but it is done in an off-line manner due to hardware constraints. also, the digital zero crossing detection is influenced by harmonic distortion and high computation time. kamat et al. [14] used a discrete fourier transform (dft) based off-the-shelf ad5933, and controlled the system using field programmable gate array (fpga) which is computationally efficient. the dft based measurement is affected by additive noises as it involves inverse tan function. also, the frequency-domain method suffers from spectral leakage. the quadrature detection adds an additional phase shift of 90° followed by the use of sophisticated multipliers leading to complexity [1]. pll is also employed to lock into the phase of low amplitude signals, but pll is available as an off-the-shelf (analog and digital) and built-in module (in fpga, software, and neuronal) and requires high quality external circuits [15]. bertotti et al. [16] suggested a peak detection-based phase measurement, but it requires two reference sources increasing hardware complexity. the state-of-art devices such as sine-phase [17] and eliko [18] have focused on the accuracy compromising the cost. therefore, it is necessary to devise a phase measurement method which is simple, compatible with portable devices, cost effective, accurate, and high frequency compatible. these features lead to effective phase measurement. knapp et al. [19] introduced a generalized cross correlator (gcc) based method to estimate phase differences by locating the cross-correlation peak of filtered versions of two received signals. the optimum performance is attained only when the signals and noises are gaussialy distributed [19]. on the basis of accuracy, the correlation-based phase measurement tends to yield better results if modified in an application specific way. drawing inspiration from their work [19], liu et al. [20], luo et 35 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 34-43 al. [21], shen et al. [22], liang et al. [23], etc. utilized the correlation method for time delay/phase difference measurement. till now, the correlation method has been applied to the estimation of time delay in the field of pda, interferometry, and remote sensing. however, the digital correlation-based phase measurement is not employed in many viable applications due to certain limitations, such as frequency drift errors and phase errors. this study proposes a digital correlation-based phase measurement method, which is simple, portable, cost effective, accurate, and high frequency compatible. for the first time, the phase measurement for bio-impedance signals by using digital correlation method is demonstrated. the experimental functionality and reliability of this method is supported by simulation and analytical treatment. the exhaustive error analysis is carried out, and the remedial measures are suggested and implemented. finally, the proposed setup is compared with the existing ones. 3. hardware realization for phase measurement using digital correlation in a bio-impedance system, a drive module offers a high-frequency low-amplitude excitation current to tut through drive electrodes, and a sense module measures the impedance of tut from the voltage response through sense electrodes. the tissue impedance is to be obtained as magnitude and phase. the methodology employed here for bio-impedance phase measurement is a part of a bio-impedance measurement device whose development is under progress. the excitation signal can incur phase changes while passing through a biological tissue. the phase difference between the excitation signal and the response signal is measured by the sense module to give bio-impedance phase. the accurate measurement of phase has always been critical. for testing the reliability of digital correlation, synchronous measurement approach is used. it requires the introduction of the phase difference between the reference signal (like excitation signal) and test signal (dummy bio-impedance signal). these signals can then be processed by digital correlation method, and the phase difference can be obtained. 3.1. experimental setup for phase measurement the controllers used in dsp boards do not accept negative signals. hence, a dc offset is often added before processing them. a typical method of introducing a delay in signals is to propagate the signal through resistor capacitor (rc) networks. rc networks are unfavorable in the present application as they remove dc components and also entail the use of variable passive components with changing frequencies. another option is to use a two-channel programmable function generator, which can produce two sine waves with a dc offset. the phases from 1° to 360° and the frequencies from 5 khz to 1 mhz can be set using it. however, such generator is bulky and cannot be a part of portable devices. as such, direct digital synthesis (dds) chips are employed to produce the reference signal and the test signal in this work. the experimental setup for performing the phase measurement is as shown in fig. 2. dds 1 (ad9837) vref vdd= 5v dgnd vout agnd fsync sclk sdata dds 2 (ad9837) vdd= 5v dgnd vout agnd fsync sclk sdata vdd=5 v lt1017 4 2 3 8 1 c1 lt1017 4 2 3 8 1 10µ f 10µ f tm4c123gxl gnd scl sda vdd= 5 v pf_2 pf_3 d1 d2 controllercomparators sine generators x(p) y(p) vtest cs1 cs2 sdata sclk 1 6 * 2 l c d c2 vdd=5 v display phase =0 phase =0 phase =phase = fig. 2 schematic diagram of hardware realisation for phase measurement 36 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 34-43 the reference (vref) and test (vtest) sinusoidal signals of 0.624 mvpp and the frequencies from 5 khz to 1 mhz are generated using dds chips. the phase measurement using digital correlation method is performed using arm cortex m4 based texas board tm4c123gxl [24]. the board can accept signals through analog and digital pins. the analog pins of the board receive the signals and convert them to digital signals before processing them. however, sampling the signals with a 10-bit analog-to-digital converter (adc) can affect the phase information and can add a/d quantization noise. in order to avoid this, the conversion of sine waves to unipolar square waves is done and then given to the digital pins of the board. dds (ad9837 from analog devices, inc.) are employed to produce the reference signal and the test signal. dds ensures faster switching between output frequencies. additionally, it has fine frequency resolution of 0.06 hz, wide spectrum of operation till 8 mhz, and signal-to-noise ratio (snr) of -63 db. the calculation of frequency word (fw) is shown in eq. (1), and the required frequency is set by sending 16-bit control word through serial peripheral interface (spi) bus. 1 . . 2 . . ndes freq fw ref freq = × (1) where n1 is 28 for an in-built frequency register, the desired frequency (des. freq.) is the frequency of the input sinusoidal signals, and the reference frequency (ref. freq.) is 16 mhz. the phase register content is set after calculating its content as shown in eq. (2) and sending its content as 16-bit control word. 22 2 n phase register = × φ π (2) where n2 is 12, phase register is 12 bits, and � is true phase. for example, the data line value of c200 is the equivalent 16-bit control word for setting the phase of dummy bio-impedance signal as 45°. comparators are used to convert sine waves to square waves without altering the phase information. it is essential to select a comparator integrated circuit (ic) which can sense low input voltages with favorable bandwidth and match input impedance with respect to previous stages. lt1067 from analog devices, inc. is experimentally chosen from a set of comparators, as it can detect a minimum of 100 mv input sine wave and convert it into unipolar square wave. in real time applications, weak signals are magnified by subsequent amplifier stages, and hence the sensitivity of 100 mv is ample. it also functions on single supply, and this feature is essential to avoid negative signals entering the subsequent digital board. comparator circuits generate unipolar square waves. noise interferences are kept away by using constant current source of 215 kω output impedance and 218 µ v root mean square noise for excitation [3]. symmetrical instrumentation amplifiers with common mode rejection ratio (cmrr) above 80 db for rejection of electrode noise and motion artifacts are also used. the simulative noise analysis of comparator circuit is carried out in multisim, and a considerable low value of output noise of 41 pv/hz power is obtained. this work only focuses on the method of phase measurement, and other real time measurement challenges are considered during the development of bio-impedance devices. the texas board with 64 pin and 80 mhz crystal has two spi buses, one of which is used for the communication and control of dds chips. the reference square signal (c1) and test square signal (c2) obtained at the outputs of comparators are given to the digital pins of the board (pf2 and pf3) and then processed digitally to deduce the phase information. this simple method uses only two comparators and a digital board. the comparator works on 5 v dc single supply, and the board supplies this power. the board is universal serial bus (usb) powered. 37 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 34-43 3.2. phase measurement using digital correlation the phase difference of � is induced between two unipolar square signals using dds. these signals are processed by tm4c123gxl board after acquiring their samples through digital pins. the floating point unit (fpu) of the board performs digital correlation by utilizing the samples of input signals. even though the input signal is continuous, the large sample size conserves the phase information within the board. both auto correlation and cross correlation of waves are performed to obtain the correlation coefficient as presented in eq. (3). 0 2 2 0 0 [ ( ) x][ ( ) y] [ ( ) x] [ ( ) y ] n p n n p p x p y p x p y p = = = − − = − − ∑ ∑ ∑ γ (3) where γ is the correlation coefficient, n is the sample number of input unipolar square signals, p represents the sample number, x(p) is the reference unipolar square signal, x� is the mean of x(p), y(p) is the test unipolar square signal, and y� is the mean of y(p). c1 and c2 signals have an amplitude of 5 v with an offset of 2.5 v. the digital pin senses and stores 1 for 5 v and 0 for 0 v. x� and y� are 0.5 considering the stored inputs. the maximum value of γ occurs when � is 0°, and the minimum value of γ occurs when � is 180°. these values are 1 and -1 respectively, and γ is 0 at 90°. this establishes a relationship between phase difference and γ, as shown in eq. (4). ˆ 90 90 ° °= − ×φ γ (4) where �� is the measured phase in degrees, and γ is the correlation coefficient. once the correlation is performed and the correlation coefficient is obtained, the phase difference is displayed on a liquid crystal display (lcd) monitor in degrees. this value is compared with the known value set by the dds chip. the observation is recorded for a set of phase values from 0° to 360° in a range of frequencies from 5 khz to 1 mhz. it is necessary to analyze the errors occurred during the experimentation and resolve them to apply the method in real time applications to improve the accuracy. the digitalread() command accepts the samples of unipolar signals. 4. modelling for phase measurement using digital correlation once the relationship between phase difference and correlation coefficient is established, it is necessary to further analyse the functionality and challenges of the digital correlation method before its real time utilisation. 4.1. modelling by simulation unipolar square waves are generated in matlab environment using square function. the square waves are defined by amplitude (0.5 v), offset (0.5 v), phase, number of samples (200-5000), and frequency (5 khz to 1 mhz). the phase adjustment in the degrees from 0° to 360° (�) is converted to radians and is introduced to one of the signals treated as the test signal. the correlation coefficient is calculated as per eq. (3), and the measured phase (��) is obtained using eq. (4). the true phase and measured phase for 2000 samples at 5 khz are recorded, and the error is tabulated in table 1. a maximum error of 0.2° (the best possible accuracy) is observed at 5 khz for true phase from 0 to 180. for the true phase greater than 180°, the value of γ repeats and hence the �� value is also repeated. this is not an issue in phase measurement as the phase differences in all major applications are confined within 180°. also, it is observed that as frequency increases with the fixed sample size, the error increases (addressed in section 5.1.). 38 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 34-43 table 1 simulative phase measurement using digital correlation true phase, �(°) correlation coefficient, (γ) measured phase, �� (°) error, �� � (°) 0.000 1.000 0 0 1.000 0.992 0.895 -0.105 10.000 0.948 9.895 -0.105 45.000 0.501 44.978 -0.022 50.000 0.449 49.835 -0.165 60.000 0.337 59.910 -0.090 90.000 0.000 89.955 -0.045 120.000 -0.107 119.820 -0.180 150.000 -0.331 149.865 -0.135 180.000 -1.000 180.000 0.000 270.000 0.000 90.000 -180.000 360.000 1.000 0.000 -360.000 4.2. experimental modelling the phase measurement using digital correlation method is performed by the fpu of arm cortex m4 based texas board tm4c123gxl, and the experimental setup is illustrated in section 3.1. the digital pins are configured as input pins, and the unipolar signals are applied to these pins. 2000 samples of these signals are stored as arrays on which the correlation operation (repeated multiplication and cumulative addition of array elements/input samples) is performed to extract γ (eq. (3)). the true phase set by dds is measured using γ (eq. (4)) as the measured phase. the result obtained at 5 khz is tabulated in table 2. during the experimental modelling of the proposed method, it is seen that a maximum permissible error of 0.870°is observed at 5 khz. the error is more in experimental modelling than in simulation due to analog elements like comparators. the measurement process is repeated for different frequencies. it is observed that as the frequency increases with the fixed sample size, the error also increases just like in the simulation of digital correlation. the reasons for this error increase are discussed in section 5.1. table 2 experimental phase measurement using digital correlation true phase, �(°) correlation coefficient, (γ) measured phase, �� (°) error, �� � (°) 0.000 1.000 0 0 1.000 0.992 1.800 0.800 10.000 0.948 10.680 0.680 45.000 0.501 45.860 0.860 50.000 0.449 50.640 0.640 60.000 0.337 60.750 0.750 90.000 0.000 90.760 0.760 120.000 -0.107 121.020 0.020 150.000 -0.331 150.87 0.870 180.000 -1.000 180.000 0.000 270.000 0.000 90.660 -180.000 360.000 1.000 0.000 -360.000 5. results and discussion the characteristics of biological tissue are assessed using bio-impedance. bio-impedance can be represented as magnitude and phase varying with respect to time. sinusoidal carrier signals with bio-impedance information are processed to extract magnitude and phase. the proposed method uses digital correlation to measure the bio-impedance phase. sinusoidal signals are conditioned into unipolar square signals before phase measurement. the correlation method is used to compare the 39 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 34-43 unipolar square signals and yield a correlation coefficient. the correlation coefficient represents the phase information. the phase measurement using correlation method is made viable through a series of simulative and experimental analyses. initially, the simulation and experimentation of phase measurement is carried out at a certain frequency. then, the frequency is varied. other parameters like amplitude, offset, sample size, etc. are kept the same for all measurements. it is seen that the occurring error increases with respect to frequency for a fixed sample size. 5.1. error analysis for the simulation of phase measurement using digital correlation the frequency and sample size of signals are varied in the simulation of phase measurement using digital correlation. the error is plotted versus true phase for various frequencies and sample sizes as shown in fig. 3. at 5 khz, as the sample size increases from 200 to 5000, the error (2.5° to 0.1°) decreases. this is because the sample size denotes the sampling frequency. according to nyquist criterion, the sampling frequency should be at least twice the input frequency. in the simulation, a single cycle is divided into samples. for an input frequency of 5 khz, the number of samples should be at least 2000. any value below that (e.g., 200) does not extract full phase information due to under-sampling whereas a large sample size (e.g., 5000) yields more information from input signals. as the frequency increases to 10 khz, the sample size of 5000 is not enough and need to be increased to fulfill nyquist criteria. a small error occurs even if the sample size is optimum with respect to frequency. this error is due to the harmonics of fundamental frequency of square wave and is negligible at higher frequencies. (a) at 5 khz for 200 samples (b) at 5 khz for 2000 samples (c) at 5 khz for 5000 samples (d) at 10 khz for 2000 samples fig. 3 simulation of phase measurement using digital correlation 5.2. experimental error analysis for phase measurement using digital correlation as explained in section 4.2., the phase measurement of signals with 5 khz is carried out smoothly with a maximum error of 0.870° for 2000 samples. the input signals at the digital pins are sampled by digitalread() and stored as global variables in random access memory (ram). the digital read time for 2000 samples is found out through programming as 9.8 ms, and the 40 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 34-43 sampling frequency is hence 200 khz. the input frequency is changed by programming dds and is set to a higher frequency of 100 khz. it is seen that if the sample size is kept the same, there is a drastic increase in phase error just like in simulation. the reason for the phase error is explained in section 5.1. the error is nullified by increasing the sample size, but arm cortex m4 board has a 32 kb ram which restricts the sample size to 3800 (2000 is a safe value to avoid instability issues). in order to process the signals with higher frequencies, more memory in ram is needed. the limitation of ram affecting sample size is confirmed by trying to use an atmega 328p arduino board instead of texas board. the sample size of input signals is limited in atmega328p to 300 due to 2 kb ram. external rams (like 23lc1024, 47lc16, etc.) can be interfaced to the board through i2c or spi to accommodate large sized global variable arrays to solve this shortcoming. another alternative is to use boards with microcontrollers having ram with compatible large memories (like arduino due, tms320f28335, etc.). the bia experimental setup using modified phase measurement technique for a biological tissue (egg) is as shown in fig. 4. tetrapolar bio-impedance (bi) measurement is done by inserting needle electrodes up to 1 cm into the tissue of a hard-boiled country egg. bi is measured using lcr meter, and bi phase measurement is conducted at 5 khz, 50 khz, and 100 khz. lcr meter has frequency limitation up to 200 khz and uses voltage excitation with bipolar measurement [2].the results are tabulated in table 3. fig. 4 bio-impedance measurement of biological tissue (egg) table 3 comparison of bi measurement with lcr meter and proposed method frequency (khz) bi using lcr meter bi using digital correlation phase (°) phase (°) 5 -89.65 -88.12 50 -64.56 -65.04 100 -30.19 -30.33 5.3. comparison between the proposed method and the existing methods the existing methods of bio-impedance measurement are compared based on signal processing type, accuracy, frequency compatibility, cost, and applications. the proposed phase measurement method is a part of the bio-impedance device which is being developed. the technique used here is the correlation method [20-23]. to analyze where the proposed method stands, a comparison between the proposed method and other traditional and contemporary techniques is made, as tabulated in table 4. a careful study of these methods reveals that the traditional techniques are bulky, slow, and limited by operating frequency. other advanced techniques overcome the frequency limitation up to a certain extent and offer a better accuracy. however, they are complex. the commercially available phase detectors are expensive, sensitive to dc components, frequency limited, and offer moderate accuracy. the proposed digital correlation method is a part of bio-impedance device which is superior to all above techniques in terms of accuracy, frequency compatibility, portability (stand-alone module), speed, and cost effectiveness. the developed bio-impedance device with the present method of phase measurement costs around 140 usd. a/d quantization errors are nil in this method because there is no adcs. 41 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 34-43 table 4 comparative analysis of phase measurement methods ref. method (°) type of signal processing accuracy (°) max. operating frequency (khz) approx. cost (usd) applications [1] quadrature detection analog 0.1 100 200 impedance measurement [7] lissajious loop pattern analog 1 100 280 impedance measurement [7] virtual vector voltmeter analog 1 100 3485 bio-impedance measurement for body mass [8] xor-filter analog 1 10 150 bio-impedance measurement for urolithiasis [9] sine wave fitting simulative 0.1 1 278 impedance measurement [10] auto-balance bridge analog 1 10000 200 bio-impedance of cell [11] phase detector (ad8302) analog 0.3 100 2787 bio-impedance measurement [12] lcr meter analog 0.1 20 2000 bio-impedance measurement for diabetes neuropathy [14] fpga method of samples digital 0.003 100 500 impedance cardiography [15] phase-locked loop digital 0.1 10000 250 impedance measurement [16] peak detection-based phase measurement analog 0.3 100 200 impedance measurement [17] sine phase software based 1 16600 6272 impedance measurement [18] eliko software based 0.1 349 6690 impedance measurement [20] correlation digital 0.15 1000 range finder [23] sine-tan method analog 0.01 1000 5000 interferometry [25] fft-inverse discrete fourier transform digital 0.1 10000 impedance measurement this study modified digital correlation digital 0.002 1000 140 bio-impedance analysis 6. conclusions on the basis of the studies conducted, it can be concluded that the digital correlation-based phase measurement method is a reliable means to characterize the media through which the signals pass. this characterization is done by measuring the delay/phase changes incurred by the signals. in the hardware realization, a reference wave from sources is compared with the wave at the receiving end and looked upon for phase differences by the experimental setup. the phase measurement using correlation method is not widely used in many viable applications including bio-impedance phase measurement due to the demerits such as frequency drift and phase errors. these issues are addressed in this work in a systematic manner. the best possible accuracy of 0.2° is obtained by realizing the proposed method for measuring the induced phase differences between two sine waves in a frequency range of 5 khz to 1 mhz. this modified phase measurement technique is a part of sense module of bio-impedance device under development. this modified method can be incorporated in portable devices for any application which needs accurate phase measurement, as it uses fewer numbers of hardware components and can be operated on usb power. conflicts of interest the authors declare no conflict of interest. references [1] m. sedlacek and m. krumpholc, “digital measurement of phase difference—a comparative study of dsp algorithms,” metrology and measurement systems, vol. 12, no. 4, pp. 427-448, january 2005. [2] s. grimnes and o. martinsen, bioimpedance and bioelectricity basics, london: academic press, 2008. 42 proceedings of engineering and technology innovation, vol. 21, 2022, pp. 34-43 [3] s. sruthi, r. dhavse, and j. n. sarvaiya, “precise, portable and wide band alternating current source for extremely low current values,” analog integrated circuits and signal processing, vol. 102, no. 2, pp. 273-282, february 2020. [4] f. simini and p. bertemes-filho, bioimpedance in biomedical applications and research, cham: springer international publishing, 2018. [5] h. p. f. x. moonen, f. j. l. zanten, l. driessen, v. de smet, r. slingerland-boot, m. mensink, et al., “association of bioelectric impedance analysis body composition and disease severity in covid-19 hospital ward and icu patients: the biac-19 study,” clinical nutrition, vol. 40, no. 4, pp. 2328-2336, october 2021. [6] m. n. r. jafari, z. esmaeilbeig, and m. nemati, “overcoming the bottleneck in covid-19 detection: a machine-learning approach to improve accuracy of electrochemical impedance spectroscopy (eis) detection sensitivity,” open access journal of biomedical science, vol. 3, no. 1, pp. 676-679, october 2020. [7] j. g. webster and h. eren, measurement, instrumentation, and sensors handbook: electromagnetic, optical, radiation, chemical, and biomedical measurement, boca raton: crc press, 2017. [8] a. silverio and j. d. tatel, “development of a variable frequency impedance measuring system,” 14th international symp. on integrated circuits, pp. 564-567, december 2014. [9] p. m. ramos, f. d. silva, and a. c. serra, “improving sine-fitting algorithms for amplitude and phase measurements,” 17th international measurement confederation world congress, pp. 614-619, june 2003. [10] n. li, h. xu, w. wang, z. zhou, g. qiao, and d. d. li, “a high-speed bioelectrical impedance spectroscopy system based on the digital auto-balancing bridge method,” measurement science and technology, vol. 24, no. 6, 065701, may 2013. [11] y. yang, j. wang, g. yu, f. niu, and p. he, “design and preliminary evaluation of a portable device for the measurement of bioimpedance spectroscopy,” physiological measurement, vol. 27, no. 12, 1293, 2006. [12] j. p. olivarez, f. a. olivares, a. p. medina, j. d. carmona, a. r. agundis, a. e. calderon, et al., “bioimpedance phase angle analysis of foot skin in diabetic patients: an experimental case study,” innovation and research in biomedical engineering, vol. 36, no. 4, pp. 233-239, 2015. [13] g. hu, j. xu, g. w. auner, j. smolinski, and h. ying, “digital phase detection approach and its application for aln dual-mode differential surface acoustic wave sensing,” sensors and actuators b: chemical, vol. 132, no. 1, pp. 272-279, january 2008. [14] d. k. kamat, s. nimkar, and p. m. patil, “bio-impedance computation based on hardware implementation using fpga,” international journal of electrical engineering, vol. 6, no. 1, pp. 12-19, january-june 2014. [15] “what is direct digital synthesiser dds,” https://www.electronics-notes.com/articles/radio/frequency-synthesizer/dds-direct-digital-synthesis-synthesizer-what-isbasics.php, 2012. [16] f. l. bertotti, m. s. hara, and p. j. abatti, “a simple method to measure phase difference between sinusoidal signals,” review of scientific instruments, vol. 81, no. 11, 115106, 2010. [17] “sinephase impedance analyzer,” https://sinephase.com/impedance-analyzer-model-16777k-lcr-meter/, august 2016. [18] “impedance spectroscopy,” https://eliko.ee/quadra-impedance-spectroscopy/, december 2017. [19] c. knapp and g. carter, “the generalized correlation method for estimation of time delay,” ieee transactions on acoustics, speech, and signal processing, vol. 24, no. 4, pp. 320-327, august 1976. [20] y. z. liu and b. zhao, “phase-shift correlation method for accurate phase difference estimation in range finder,” applied optics, vol. 54, no. 11, pp. 3470-3477, april 2015. [21] j. luo and e. e. konofagou, “a fast normalized cross-correlation calculation method for motion estimation,” ieee transactions on ultrasonics, ferroelectrics, and frequency control, vol. 57, no. 6, pp. 1347-1357, june 2010. [22] y. l. shen and y. q. tu, “correlation theory-based signal processing method for cmf signals,” measurement science and technology, vol. 27, no. 6, 065006, may 2016. [23] y. r. liang, h. z. duan, h. c. yeh, and j. luo, “fundamental limits on the digital phase measurement method based on cross-correlation analysis,” review of scientific instruments, vol. 83, no. 9, 095110, september 2012. [24] texas instruments, “tiva™ tm4c123gh6pm microcontroller data sheet,” https://www.ti.com/lit/ds/symlink/tm4c123gh6pm.pdf, 2014. [25] w. wang, y. tang, x. zhang, and k. wang, “design of a recursive single-bin dft algorithm for sparse spectrum analysis,” cluster computing, vol. 20, no. 2, pp. 1483-1492, april 2017. copyright© by the authors. licensee taeti, taiwan. this article is an open access article distributed under the terms and conditions of the creative commons attribution (cc by-nc) license (https://creativecommons.org/licenses/by-nc/4.0/). 43  proceedings of engineering and technology innovation, vol. 1, 2015, pp. 19 26 19 cop y right © taeti credit card risk assessment using artificial neural networks hsi-chieh lee 1, 2, chen-chi yu 1 , tzu-miao lin 1 1 department of information management, yuan ze university, chungli, taoyuan, taiwan r.o.c. 2 department of computer science and information engineering, national quemoy university, kinmen, taiwan r.o.c. received 16 april 2015; received in revised form 08 may 2015; accept ed 18 june 2015 abstract in recent years, the supply and demand of the plastic currency market has been rapidly increasing, especially, for the growth rate in the credit card market. it has increased about 16 times in the last 10 years. many banks devoted to make a large investment in credit card marketing for the sake of getting the maximum profits in the worldwide market. however, most banks are trying to reduce the requirements for credit card application in order to increase the motivation of the customers for applying their credit cards. as a result, many banks somehow ignore the risk management of credit card approval which leads to the increases of bad debt in the credit card market. when this scenario happens year by year, those banks will not get profits from the credit card market but a great loss. in this study, a total of 113,048 entries were used which included fundamental customer data, credit card data, and customer history data from joint credit information center (jcic) of taiwan. we used the characteristics of artificial neural networks and grey theory to find out the potential factors of the bad credit and finally used the correlation method to find out the higher (important) relative variables (parameters) of bad credit. 80,000 entries were randomly selected as training data and the remaining 33,084 entries were used as testing data. the experimental results shown that the accuracy of forecasting rate for the proposed early warning system was an overall of 92.7%. these results suggested that the once the collections of the new customer data were available, the proposed approach could be used as an early warning system which can be used to decrease the risk of credit card approval. ke ywords : credit card risk assessment, artific ia l neural networks, self-organizing map (som), grey theory, electronic commerce 1. introduction daily items for human beings have been developed completely and in wide array to satisfy the diverse needs of individuals, given the ceaseless advancement of technologies. the money market, an indispensable part of human livelihood, has also transformed from the “barter system” of early times, to transactions with currency, and contemporary transactions with plastic money. in taiwan, the use of plastic money for transactions is common and everywhere. this implies that plastic money has been gradually and surely replacing cash in transactions, and has emerged as the primary form of currency. however, the excessive inflation of credit limit granted in favor of the consumers and the improper approval of credit limits by banks resulted in high debts beyond the means of the consumers for retirements and high non-performing loans for the banks. this is the problem of the so-called “credit card slaves” in taiwan, which hit the society hard wave after wave. in light of the said issued, this paper proposes to apply the neural network with an attempt to map out an effective early warning system for detecting the characteristics of potential credit risks of card holders to avoid the loss of banks due to improper review and approval of credit cards. 1.1. background and motivation according to joint credit information center [1], there were 3,676,000 credit cards circulated in 1995. this number increased to 45,494,000 in 2005. similarly, credit card loans rose from nt$272,387 million in 1995 to nt$1,420,984 in 2005. the data of loans from credit card charges and cash advances provided a clue to the reasons why banks scrambled for a share of the credit card market. the statistics of loans from credit card charges increased by 7.5 times in the decade of 1995 to 2005. the cash advance business also mushroomed by almost 15 times in the same period (fig. 1, 2). fig. 1 volume of credit cards in circulation over the years *corresponding aut hor, email: imhlee@gmail.com proceedings of engineering and technology innovation, vol. 1, 2015, pp. 19 26 20 cop y right © taeti cop y right © taeti cop y right © taeti cop y right © taeti cop y right © taeti fig. 2 growth trend of loans from credit card charges and cash advances yet, the number of new cards issued, the amount of credit card loans, and cash advances significantly reduced in 2006. this phenomenon was the result of one card holder possessing many credit cards because banks tended to lower the entrance barrier for granting the applications for credit cards and cash cards in order for a larger quantity of issuance. after the “credit card slaves” storm, banks have learned their lesson. they changed from focusing on the quantity of card issuance to screening for clients with good credit standing and therefore attempted to construct an efficient early warning system under objective conditions to reduce non-performing loans. this remains the motive of this study. 1.2. the objective of study this study adopts the method of data mining whereby the potential credit risks of credit card clients are explored based on the characteristics of neural network clustering. in practice, the personal information of clients, the borrowing records of the clients at joint credit information center (jcic) and the patterns of spending by credit card holders and the payment records are included for analysis for finding out the factors related to bad credit of card holders and delinquency in payment. these factors will be used as the variables in the analysis for constructing an early warning system for credit evaluation. 1.3. literature review literature on risk management in the financial sector will be introduced in this section. theories of the neural network model also will be discussed here. 1.3.1. risk management risk literally means the possibility of suffering loss. risk can be managed through identification, analysis, and responding measures. risk can be analyzed from the perspectives of general causes, the probability of occurrence, the effect of risk, and the damage or loss that is caused (converted into monetary form). rosenbloom (1972) defined risk as “the uncertainty of the occurrence of loss”0, webster’ s dictionary defined risk as the opportunity of damage or loss or being in danger. these definitions focused on the possibility of a negative result. further. williams & henis (1981) pointed out that: risk management is the management method through identification, assessment and control of risk at minimal cost in order to minimize the loss that is caused 0. focusing on risk management of banks, lin (1997) proposed a few principles [4]: (1) properly evaluate the influences of different risks on the profitability of banks and these influences shall be subject to good financial management; (2) properly evaluate the capacity of the bank management in early warning of the banking sector and assessment of the strength and weakness of the bank; (3) understand the capacity of the bank in assuming risks on the basis of the competitive advantage of the bank in the industry and take the changes in the macro economic environment under control in pursuing its goal with relevant adjustment of risk level; (4) viable risk management and monitoring systems will be necessary: (5) good management and the training of good professionals. risk management is important to enterprises of any industry. better risk control will help to lower the intensity of loss. 1.3.2. neural network model a neural network is one form of artificial intelligence and is a data processing system imitating the biological neurotic system. with large number of linked artificial receptive neurons, this system imitates the learning capacity of a biological neurotic system and simulates learning and memory through a parallel computation framework. it is applied to categorization, reflection, identification and optimization functions. the neural network system is a simulation of the biological neurotic system in operation through ceaseless learning and correction of data for correct output. the neural network system features large volume of parallel processing, non-linear output, and multiple layering in forecasting. the processing element (or pe) is the fundamental component unit of the neural network system and the processing unit can be subcategorized as input and output values. the sum of the output values from each layer is the input value of the upper layer. generally, the function of the product of the input value and weighed links is expressed in the following equation: ( ) ij i j i a f net f w x           a=the output signal of neural network pe xi=external input of data net=cluster function θj=threshold value of neural network pe proceedings of engineering and technology innovation, vol. 1, 2015, pp. 19 26 21 wij=the linked weights of all neura l network pe in between data paths of the pe. the symbol denotes the strength of the effect of the ith pe of the former layer on the jth pe of the latter layer. f is the transfer function of neural network pe. it is the conversion of the sum of another pe input value and the linked weights to output value through the pe conversion function. the conversion function usually sorts out the output value like the sigmoid function and is mathematically expressed as: 1 ( ) 1 net f net exp    the output of s function fa ll between 0 to 1, and the s function input/output curve is usually skewed. therefore, the net signal is very small and approaches zero. at this point, the output value approaches infinity. if net is a positive value and approaches infinity, the output va lue approaches 1. besides, another commonly used conversion function is hyperbolic tangent function and its output is also in an s shape, and is mathematically expressed as follows: ( ) net net net net exp exp f net exp exp      the hyperbolic tangent function is symmetrical with its origina l point and is a function w ith output function between -1 and 1. this function works faster than other neural functions in computation for results [5]. 1.3.3. neural network model this section gives an introduction on the application software, techniques, and relational analys is. the clustering techniques and neural network algorithm used clementine of spss proposed by kohonen. grey relational analysis is adopted for the sampling process of card users in poor credit standing. 1.3.4. clustering technique clustering is the method through which data with the same or similar characteristics are grouped. clustering is a form of non-monitoring learning skill and in the process of clustering the distance between each data and the center of the cluster will be mapped out. clusters of the shortest distance to the center will be formed into the same group. the main feature of clustering is that no attribute has been assigned to the data and clustering is d irectly conducted. after the clustering process, the clusters will be based on the results of clustering to trace the characteristics of the cluster elements and the criteria for clustering. taking this study as an example, the symptoms of c lients w ith potential risk are unclear before the process. therefore, clustering is used for digging out the characteristics of different clusters. with the different intens ifies of the clusters, different weights are assigned to construct a system for forecasting clients with bad credit standing. the self-organizing map neural network, or som, was proposed by kohonen (1988) [6]. in brief, this system is a network framework with 2 layers, the input layer and output layer (or competitive layer). the characteristic of som is to search the characteristics of the targets and reflect the vector of any dimension to a one or two dimensiona l map (fig. 3.2). the output layer neurons will base on the characteristics of the input and display in the output space in topological structure. this map reflects the characteristics of the input vector. therefore, this is called som. the strength of the neural network proposed by kohonen is that it is a parallel neural network structure and provides the characteristics of analysis and the relation of the organizations. therefore, it has high confident level as a technique of categorization. assuming there are k neurons in this network. if wkt (n) x (n) = max kt wi (n) x (n) i=1, 2,…a this means that the kth neuron is the ultimate winner. adjustment of the chain values            1 ? j j k j w n x n w n if j k w n w n        where, x=(x1,x2,...,xp )t means the input vector of the network; wj=[wj1,wj 2,...,wjp]t, represents the chain value vector of the jth neuron. p represents the dimension, and η is the parameter of learning rate, n represents the dispersion time. usually, each chain value of the neuron will be normalized with growing rate at 1. therefore, the rule for sorting out the winner could be corrected as: fig. 3 dimension kohonen model proceedings of engineering and technology innovation, vol. 1, 2015, pp. 19 26 22 cop y right © taeti cop y right © taeti cop y right © taeti cop y right © taeti cop y right © taeti if ∥x-wk∥<∥x-wj∥, then neuron k will be the winner until the formation of the characteristic map (fig. 3). 1.3.5. clustering technique teng (1982) proposed the grey system theory for studying the cognition of human be ings on different research and things from simple to sophistication. it is just like the waves in the sea coming one after another. in other words, many things are neither absolute ly black (blank) nor white (fully transparent) to human be ings. instead, it is in a state of grey (half understood and half unknown), which implies that information is not complete or cannot be described in full detail [7]. grey relation ana lys is refers to certain degree of association between 2 elements of the systems or events. if the elements become congruent due to changes, the association between the two is high, or vice versa. degree of association is the variance of the geometric shapes of the curves. assuming a reference value x0, and n comparative values, then x1, x2,…, xn, the relationa l coefficient ξ(xi) of each comparative value matrix and the reference value can be expressed by the following equation:                   0 0 0 0 | | | | | | | | i i i k i k i i i i k min min x k x k max max x k x k k x k x k max max x k x k           ζ(zeta) is the differentiation coefficient and 0 < ζ < 1 while , implies that it is the smallest absolute value between two values (comparative and reference values). after the comparison of a ll values is completed, the smallest value of the small values is sorted and marked as δ min. similarly, represent the greatest absolute value between two values (comparative and reference values). after the comparison of all values is completed, the greatest value of the great values is sorted and marked as δ max. relationa l coeffic ient ξ(xi) can be simplified and expressed as maxoi(k) maxmin )(       k i furthermore, the relationa l coeffic ient may vary at different points of the reference value and the comparative value along the curve, and there is more than one value of the degree of association. therefore, if a ll the points a long the curve are linked and make it a relational coeffic ient value, this will be the mean of such value and is expressed by ri. the equation of degree of association ri will be:  k n r i n k i  1 1   finally, after adding up a ll relationa l coeffic ient values and then dividing by the total number of values, we get the degree of association. due to the demands on credit card risk assessment and prediction capability of artific ia l neural networks, there are many literatures regarding the deployment of nove l techniques on the problem of credit card risk assessment in recent years [8-10]. 2. experiment 2.1. study sample processed data 113,048 grouping of data testing data 33,048 training data 80,000 use neural network clustering to find out the characteristics of factors most influential on bad credit status like characters of the clusters for constructing the credit early warning system) are the variables meet the requirements screen out the variables credit early warning system n y is the forecast ideal adjust the weights of the parameters end y n fig. 4 flow figure of constructing the early warning system samples of this study inc luded the master file of card holders, credit card payment record information, and the information of jcic, which made up a total of 113,048 entries. thirteen variables are sorted and the samples are divided into training group and testing group. 80,000 samples were selected at random for the training group. the characteristics of bad credit standing of this group are traced through neural network analys is. weights were assigned to these characteristics for constructing an objective client credit evaluation system. the remaining 33,084 samples were subject to testing under the testing group (fig. 4). the result from forecasting is compared with the actual bad credit status clients of proceedings of engineering and technology innovation, vol. 1, 2015, pp. 19 26 23 the banks exemplified by failure to renew credit cards or confirmed by the processing banking officers. if the variance between the forecast and the actual result is wide , it means that more influentia l factors exist and additiona l weight shall be assigned in the scoring. the same princ iple is applied for repeated adjustment of weights. if the forecast approximates the ratio of clients in poor credit standing, it means the characteristics of the variables used in this study are sufficient to forecast clients of potentia lly poor credit status. 2.2. the benchmark for banks to make judgment on poor credit status different ba nks have different criteria for judging poor credit status of c lients. ba nks who provided sa mple for this study offered both interna l and externa l information for judging poor credit status of c lie nts. they are : i. interna l information: this is the information ava ilable from the database of the ba nks and inc ludes : 1) loan information of the ba nk – overdue cash card accounts of more tha n 15 da ys of de linquency for the last payment. 2) loan information of the bank – overdue loan accounts of more than 30 da ys of de linquenc y for the last payment. 3) marked by the bank as gd (1~9) une mployed/ins uffic ient inc omes. 4) marke d by the bank as gf (1~9) dec lare d le ga lly inc ompetent. 5) marke d by the bank as gl (1~9) to be ass iste d by guara ntor for settleme nt.6) marked by the bank as gm (1~9) bearing informa tion on ass ignment of right to debts for repayment. 7) marked by the bank as gn (1~9) retirement of loa ns by a re lative. 8) tota l loans from the bank exceed approved limits. ii. externa l information: this inc ludes information of the c lie nts from other banks , jcic transaction record, loa n record and payment record. externa l information us ua lly c onta ins the follow ings : 1) loan information –unusua l credit approva l. 2) loan information – record of dec lined account. 3) there were more than 3 “insuffic ient pa yme nt for the minimum” or “overdue account for payment in full” of cre dit cards over the las t s ix months. 4) delinque ncy of cash card payment for more tha n 31 days (inc lus ive) over the las t s ix months. 5) delinque ncy of cons umer loa ns for tha n 31 da ys (inc lus ive) over the las t s ix months. 6) supple mentary/s pec ia l me ntion information indicated as une mployed/ins uffic ient inc omes. 7) supple mentary/s pec ia l mention informa tion from jcic indicated dec lared le ga lly incompetent. 8) supple mentary/s pec ia l me ntion information from jcic indicated ass ignme nt of right to debts for retirement of loans. 9) supple mentary/s pec ia l me ntion information from jcic indicate d retirement of loans by a relative. 10) tota l loa ns from other banks exceed approved limits. further to the aforementioned interna l and external information, other special requirements are availed to the banks for reference: 1) no renewal of credit cards for student card holder who holds more than 3 cards. 2) supplementary proof of financia l capacity is required if the tota l unsecured loans exceeds 22 times of monthly salary. 3) if any of the said conditions applies to particular clients of the bank in credit judgment, such client shall be classified as “no card renewal” client. 2.3. analysis of the clustering results the clustering characteristics of the kohonen mode l are adopted and cross computed with a ll input variables. the result showed that each cluster shows their resemblance. under this method, there are 22 clusters grouped. data for the 22 clusters are permuted in sequential order. clusters ranked at the higher echelon of the hierarchy are the variables denoting clients in better credit standing, and vice versa. the table of variables ranking showed that the top 6 poor credit accounts, clusters 6, 9, 14, 18, are congruent with the 6 clusters confirmed by the banks as in poor credit status through ranking and manual confirmation (therefore renewal of cards is declined). therefore, this study further explores the factors and characteristics of the 4 clusters for poor credit standing. they will be listed as the reference in subsequent forecasting of clients of poor credit standing. the relations between the variables and the status of bad credit are expressed in terms of grey relation (fig. 5). fig. 5 forecasting on clusters of bad credit status we can see the comparison of relations between the clusters and the factor values directly from fig. 5. in this study, cluster 18 is used as a control for comparison. grey relation analys is is introduced as follows: i. standardization: the characteristics are sorted through analys is under the kohonen mode l. the data of the 4 clusters are listed (the primary data indicated the ranking among the clusters. the greater value of the data, the poorer the credit status). ii. find proceedings of engineering and technology innovation, vol. 1, 2015, pp. 19 26 24 cop y right © taeti cop y right © taeti cop y right © taeti cop y right © taeti cop y right © taeti out the maximum variance of the clusters and the minimum variance, for finding out the maximum value and the minimum value , we must first find out the “mapping of variance” table for the maximum value of variance as =1.59, which means δ max=1.59. the minimum variance is =0 and is expressed by δmin=0. iii. calculation of relationa l coefficient: it is ξi (k). assuming ζ=0.5, then the relational coefficient of x1 and x0, ξi (k), can be expressed as follows: (1) (2) 57.0 59.15.057.0 59.15.00 max)3( maxmin )3( 01 1           (3) 5.0 59.15.040.0 59.15.00 max)4( maxmin )4( 01 1           (4) the same principle applies to the computation of all other variables.  k n r i n k i  1 1   (5) iv. calculate for the value of: the degree of association between comparative value x1 and corresponding reference value x0.   5723.0 4 5.057.064.01 4 1 1 4 1 1     kr k  the same principle applies to the computation of all other variables. v. the above results are ranked in sequentia l orders and the following result is yielded. r4>r5 >r6> r7> r8> r13> r2> r14> r9 >r11> r16> r12> r1> r15> r1> r17> r3 2.4. construct the early warning system mapping of variables for judgment (fig. 6) can be given the benchmark for mark deduction of each variable. for example, the factor of “cash advance”, if the data field is “y”, it means 8% of the scores for this factor shall be deducted. the sample princ iple applies to the computation of all other variables. the user interface of the early warning system (fig.7) for the forecasting of potentia l risks of credit card holders. enter the id card number of the card holder and press “activate for scoring”, the system will retrieve the data of this holder from the database and process each count of data for determining if they meet the default standards. the total scores for this id will be shown in the fie ld of “total score”. and recommendation will be shown in the fie ld of “advice result”, including the recommendation for “renewal of card” or “no renewal”, and so forth. outstanding amount 7% pass_due_amt stauts of payment 8% pcd payment record in the last 6 months(1~6)48% pay_code_12 payment range 7% pmt cash advance 8% cash approved limit/payement range 8% limit / pmt type of loan 7% account_code is it greater than 0 is it>=4 is it>=2 is it greater than 800,000 y is it smaller than 0.00015 is it a, b, or c full score=100 if the conditions set forth in each field are met, deduct marks from the field. total scores oustanding amount/payment range 7% pass_due_cdoe / pmt is it greater than 0 fig. 6 mapping of variables for judgment fig. 7 the user interface of the early warning system 1 59.15.00 59.15.00 max)1( maxmin )1( 01 1           64.0 59.15.074.0 59.15.00 max)2( maxmin )2( 01 1           proceedings of engineering and technology innovation, vol. 1, 2015, pp. 19 26 25 3. conclusions fig. 8 distribution of “renewal” and “no renewal” accounts judged by the banks currently, each bank has its own method of evaluation in credit risk management. for transparency of information, each bank can access to the database of jcic for the information of its clients on loans w ith other banks, credit information and related payment record. banks can be more accurate in forecasting credit risk if they could have more complete information. in this study, the common fie lds and data of jcic are adopted, and the clients of the banks are subject to analys is with the data provided. neural network system and grey relation analysis are applied to find out the variables possibly related to poor credit status of clients. this section gives a description on the empirical findings of this study. fig. 8 is the population distribution of the clients for deciding the renewals of cards in the testing group. the distribution shows that 93% of all samples constituted c lients being granted the renewal of cards. only 4% of the clients are determined as “no renewal” of cards due to poor credit standing. there are 3% of the samples judged by the banks as “no renewal” after supplementary information on financ ial pos ition has been provided, and shown in the credit record as between “renewal” and “no renewal” of cards (fig. 8). table 1 shown that t1 stands for forecasted “renewal of card” in this study while t2 stands for “renewal of card” determined by banks. f1 stands for forecasted “no renewal of card” in this study while f2 stands for “no renewal of card” determined by banks. (t1,t2)( ※ noyr 2) the forecasting rate is 91.21%, and (f1,f2) the forecasting rate is 1.46%. the total of the two is 92.67%, which means that the forecasted results under the early warning system are congruent with the actual judgment of the banks. in table 2, 49% of the forecasted clusters (t1, f2) were randomly selected by banking professiona l for verifying the credit status of the card holders. the findings indicated that of the 160 samples being selected (note 3), 33.33% are indeed in good credit standing but were erroneously judged by the banks as “no renewal of card”. this would be significant loss of the banks for losing such good potential clients. further, 1.82% of the forecasted clusters (f1, t2) in table 3 are subject to analysis. the findings ind icated that of the 60 randomly selected samples (note 4), 33.39% are indeed clients in potentia l poor credit status but erroneously judged by the banks as “renewal of card” clients. in other words, the proportion of bad debts from 1/3 of these clients is much higher than other clients. the banks shall pay attention and take precautions. this study has led us to know that banks have shortcomings in making judgment on renewal or no renewal of cards. this study also helps to explore a group of margina l c lients falling in between “renewal” and “no renewal” and associated strongly w ith bad credit status, which is critical for decis ion of renewal or no renewal. this could be served as reference for the banks in the evaluation of credit card clients. table 1 comparing the forecasted rate of the early warning system banks this study t2 f2 t1 91.21% 5.49% f1 1.84% 1.46% table 2 t1f2 distribution of sample clustering t1f2( 160 samples) renewal % no renewal % judged by banking officers 53 33.33 106 66.67 table 3 f1t2 distribution of sample clustering f1t2(60 samples) renewal % no renewal % judged by banking officers 20 33.90 39 66.10 ※note 1: “no renewal for student cards” refers to the same student who holds more than 3 credit cards or applies for credit cards with more than 3 banks at the same time. if these conditions are met, the banks shall treat this group of clients as potentially in poor credit status and determined as “no renewal of card”. ※ note 2: (t1, t2) in this study stands for: (t1) forecasted as “renewal of card” and (t2) the judgment of the banks is also “renewal of cards”, and so forth. ※note 3: at the proportion of 5.49% of all samples, 10% or 160 samples are selected at random. ※note 4: at the proportion of 1.82% of all samples, 10% or 60 samples are selected at random. proceedings of engineering and technology innovation, vol. 1, 2015, pp. 19 26 26 cop y right © taeti cop y right © taeti cop y right © taeti cop y right © taeti cop y right © taeti re fe re nces [1] joint credit information center, taiwan, http://www.jc ic.org.tw/index. htm [2] j. s. rosenbloom, “a case study in t. khabaza and c. shearer (1995), data mining with clementine,” integral solutions limited, 1972. [3] c. a. willia ms and r. m. henis, risk management and insurance, 4th ed., new york: mcgraw-hill, 1981. [4] c. r. lin, finance innovation and risk management, 1st ed., 1997. [5] r. a. neibet, “h ow to choose a data mining sui te,” http://www. dmre view.c om/artic le_s ub.cfm? articleid=1000465, 2004. [6] t. kohonen, self-organizing maps, 3rd ed. springer, berlin, heidelberg, 2001. [7] c. l. teng, h. kuo, grey theory and its application, 1996. [8] j. n. crook, d. b. edelman, l. c. thomas, “recent developments in consumer credit risk assessment,” european journal of operational research, vol. 183, no. 3, 16, pp. 1447-1465, december 2007 [9] l. yu, s. wang, k. k. lai,“2007 credit risk assessment with a multistage neural network ensemble learning approach,” expert systems with a pplications , vol. 34, no. 2, pp. 1434-1444, february 2008 [10] b. twala, “multiple classifier application to credit card risk assessment,” expert systems with a pplications , vol. 37, no. 4, pp. 3326-3336, april 2010. 3___peti#7217__24-34 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 24-34 supervised learning based classification of cardiovascular diseases arif hussain 1 , hassaan malik 1, 2, * , muhammad umar chaudhry 1, 3, 4 1 department of computer science, national college of business administration and economics, multan, pakistan 2 department of computer science, university of management and technology, lahore, pakistan 3 department of electrical and computer engineering, sungkyunkwan university, suwon, korea 4 aihawks, multan, pakistan received 26 february 2021; received in revised form 21 august 2021; accepted 22 august 2021 doi: https://doi.org/10.46604/peti.2021.7217 abstract detecting cardiovascular disease (cvd) in the early stage is a difficult and crucial process. the objective of this study is to test the capability of machine learning (ml) methods for accurately diagnosing the cvd outcomes. for this study, the efficiency and effectiveness of four well renowned ml classifiers, i.e., support vector machine (svm), logistics regression (lr), naive bayes (nb), and decision tree (j48), are measured in terms of precision, sensitivity, specificity, accuracy, matthews correlation coefficient (mcc), correctly and incorrectly classified instances, and model building time. these ml classifiers are applied on publically available cvd dataset. in accordance with the measured result, j48 performs better than its competitor classifiers, providing significant assistance to the cardiologists. keywords: cardiovascular disease, machine learning, artificial intelligence 1. introduction cardiovascular disease (cvd), which is considered to be one of the most fatal diseases, includes a cluster of diseases such as aortic aneurysm, angina, rheumatic heart disease, coronary heart disease, congenital heart disease, stroke, peripheral arterial disease, etc. [1]. cvd affects heart muscles (myocardial infarction) and vasculatures (hypertension). the only two methods to reduce the occurrence of fatal events due to cvd are reducing cholesterol intake and increasing proper exercise. the detection of cvd in the early stage is a difficult and crucial step. cvd are complex and heterogeneous, as it is caused by multiple factors including environmental and behavioral factors. artificial intelligence (ai) has an ability to exploit the big data largely used in patient care. the big data refers to the large volume and variety of the data that are complex in nature. conventional data processing methods and techniques are insufficient to effectively examine a large amount of data [2]. the big data generally displays high capacity, reliability, and various types of features. as a result, scalability and accuracy is improved by using the big data techniques as compared to the traditional methods such as regression and correlation-based models. the big data is based on four vs [3]: volume, velocity, variety, and veracity. the techniques of ai and machine learning (ml) in the big data have achieved tremendous success that helps commerce and fabrications, in order to highlight the hidden patterns and to predict future prospects. researchers use ai methods to track down the disease related contexts, e.g., controlling tuberculosis [4] and predicting extended-spectrum β-lactamases (esbl) [5] and heart diseases [6]. * corresponding author. e-mail address: f2019288004@umt.edu.pk tel.: +92-333-3873355 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 24-34 ml methods are not only used for disease classification, but also widely used in industries. ml based framework are effectively used for evaluating the data collected from smart meters to analyze the original and fake data [7]. several types of ml classifiers are also used for the identification of chatter vibration [8]. in addition, ml techniques are also utilized to distinguish cyber-attacks to perform the paradigm and authentication [9]. ml algorithm is not a panacea for prediction related tasks. even a perfect ml model depends upon the quality and magnitude of the dataset from which it is trained. there are variant kinds of algorithms for classification and prediction of cvd outcomes. this study provides a comparative analysis among the performance matrices of ml classifiers, i.e., support vector machine (svm), logistics regression (lr), naive bayes (nb), and decision tree (j48), using the cvd dataset in diagnosing and making significant decisions to assist the clinical and health experts. the objective of this study is to evaluate and achieve the best accuracy with the lowest error rate in diagnosing the disease. consequently, the efficiency and effectiveness of those approaches in terms of precision, sensitivity, specificity, accuracy, correctly and incorrectly classified instances, and model building time are evaluated. in summary, the main contribution of this study is discussed below: (1) the four ml classifiers applied for cvd diagnosis and the detailed state-of-the-art ml classifiers are discussed. (2) the comparative analysis of ml classifiers is observed in terms of the performance evaluation matrices for cvd classification. (3) j48 model achieves the highest classification accuracy. 2. related work ai is a comprehension of mathematical designing that has the capability to improve health services by using new health care strategies and procedures. in ml, classification is considered to be the most important and essential part. a lot of research has been conducted by using ml to classify cvd from health care related datasets, proving that ml has good performance. haq et al. [6] measured the efficiency of seven algorithms for ml, i.e., k-nearest neighbors (knn), lr, decision tree (dt), nb, artificial neural network (ann), svm, random forest (rf), and measured the efficiency of three algorithms for selecting features, i.e., relief feature selection, minimal-redundancy-maximal-relevance (mrmr), and least absolute shrinkage and selection operator (lasso) to analyze the best classifier for heart diseases. lr is proven to be the most accurate classifier with the accuracy of 84% before applying feature selection. svm achieves the accuracy of 80% after applying feature selection algorithms. li et al. [10] applied different ml approaches to classify heart diseases. they proposed a diagnostic system that was accurate and efficient in producing the significant results. the heart disease dataset was pre-processed by removing the null values. they applied feature selection techniques to find the appropriate outputs. the result of experiment clarifies that the proposed feature selection algorithm (fcm/m) is suitable to diagnose the heart disease with svm. as compared to previous methods, the proposed system (srm-fcm/m) achieves better precision. latha and jeeva [11] presented a method of ensemble classification on cleveland heart dataset to increase the accuracy of weak algorithms by combining multiple classifiers [12]. the study found that ensemble techniques such as bagging and boosting effectively improved the predicted accuracy of weak classifiers in identifying the risk of heart disease. a maximum increase of 7% accuracy was achieved with the help of ensemble classification. raza et al. [13] evaluated different ml techniques and combined them with different feature selection methods to find the best technique for classification [14]. they applied four heart disease datasets and different methods like principal component analysis, chi-squared testing, relief, and symmetrical uncertainty to create distinctive feature sets [15]. by doing this, they achieve the significant accuracy of 85% on all datasets. alaa et al. [16] performed a comparison analysis of the ml techniques 25 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 24-34 based on autoprognosis framework, traditional techniques, and non-traditional techniques on the improvement of cvd and risk predictions. the research found that the ml based autoprognosis framework has the highest accuracy rate for cvd risk prediction as compared to traditional and non-traditional techniques. soni [17] compared different methods for heart disease prediction. the author combined ml methods with data mining techniques to be used in clinical decision support system. nb, sequential minimal optimization (smo), rf, decision table algorithms were used for the detection of heart diseases. few parameters including electrocardiogram (ecg) results were also employed. moreover, the author focused on multiparameter classification using weka tool. different parameters are analyzed, including ecg and heart rate variability (hrv). by using these data, nb and dt achieved good diagnostics accuracy of both 85% in predicting heart diseases. fitriyani et al. [18] proposed a method for heart disease diagnosis at the initial stage. they designed a module named clinical decision support systems (cdss) for heart disease prediction. moreover, they also conducted a comparative analysis with other medical modules and achieved the classification accuracy of 95.90% and 98.40% for statlog and cleveland datasets, respectively. 3. materials and methods 3.1. dataset for this study, cvd dataset [19] is used to develop a cvd diagnosis system based on ml. the sample size of the dataset is 70 thousand patients containing 11 features and one target value. the input features have three types: objective (realistic information), examination (result of medical checkup), and subjective (patient’s information). there are two classes on the target output mark, which reflect whether a cardio patient or a regular subject is the individual entity being examined. therefore, the extracted dataset consists of a matrix of 70k/12 attributes. the complete information and definition of 70k instances with 12 dataset characteristics is given in table 1. table 1 feature information of cvd dataset id parameters short code description 1 age ag age in days 2 height hgt height in cm 3 weight wgt weight in kg 4 gender gen male = 2; female = 1 5 systolic blood pressure sb_hi upper value (mmhg) 6 diastolic blood pressure db_lo lower value (mmhg) 7 cholesterol chl 1: normal; 2: above normal; 3: well above normal 8 glucose glu 1: normal; 2: above normal; 3: well above normal 9 smoking smk 1 = yes; 0 = no 10 alcohol alc 1 = yes; 0 = no 11 physical activity act 1 = yes; 0 = no 12 presence or absence of cvd cardio 1 = yes; 0 = no 3.2. the proposed methodology the proposed framework is designed (fig. 1) to classify cardio patients. for this study, the performance of variant kinds of ml predictive models for cvd diagnosis on full features of datasets is tested. the well renowned ml classifiers, i.e., svm, lr, nb, and dt (j48), are applied in the system. the model’s validation and performance evaluation matrices are computed. the methodology of the proposed system is divided into four phases. in the first phase, the pre-processing of cvd dataset is done by removing the duplicates and null values. secondly, the dataset is divided with the ratio of 85% for training and 15% for testing process. predictive ml classifiers are applied in the third phase. in the fourth and the final phase, the classifier’s performance is calculated in terms of sensitivity, specificity, matthews correlation coefficient (mcc), and f1-score. 26 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 24-34 fig. 1 framework for cvd prediction 3.3. dataset pre-processing dataset pre-processing is a crucial step for achieving the significant outputs of ml classifiers, which should be trained and tested in a powerful manner. pre-processing techniques such as standard scalar and removal of missing values are enforced to the dataset for adequate use in the classifiers. all these data pre-processing techniques are used in this study. 3.4. machine learning classifiers classification is the process of predicting the class of given data points. ml classification algorithms are used to classify the data into their respective domains. in this study, four well-known predictive ml classifier algorithms with their hypothetical background are presented. 3.4.1. logistic regression lr algorithm is used to observe the discrete set of classes [20]. in lr, binary and multiclass classification methods are performed. in binary classification, the prediction is completed by the predictive variable y when y ∈ �0,1�, where 0 is known as negative class and 1 is positive class. in multiclass classification, the prediction is applied to the predictive value y when y ∈ �0,1,2,3�. in order to classify binary classes [0, 1], the hypothesis �� ��� is devised, and the threshold output value is �� at 0.5. if the hypothesis �� � 0.5, it will predict y = 1; if the value is less than 0.5, it will predict y = 0. the hypothesis tends to limit the cost function between 0 and 1, e.g., 0 � �� � 1. the sigmoid function of lr can be written as: 1 1 ( ) xh x fθ −+= ℓ (1) lr is applied on the dataset to measure the probability of the occurrence of binary outcomes such as healthy (0) and cardiac (1) patient. therefore, eq. (1) of the sigmoid function is used to convert the outcomes into the categorical values. similarly, the cost function of lr can be written as: 1 1 1 ( ) cost ( , ) i i i j h x y m θ θ = =   ∑ (2) eq. (2) shows that the cost function takes two parameters in the input. ������ is hypothesis function, ������ is the output, and m represents the number of samples. the cost function is used to measure how badly the lr model is performing for predicting the relationship between y and x. before fitting the parameters to training the data with eq. (2), a constraint is added to prevent the model from overfitting. however, gradient descent (gd) as the optimization algorithm is applied by minimizing j(θ) as a function of θ to find optimal parameters. it takes partial derivative of j with respect to θ, and updates θ via each iteration with a selected learning rate of 0.5 until gd is converged. 27 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 24-34 3.4.2. support vector machine svm is considered to be the most powerful ml technique. svm is a supervised learning algorithm which is mostly used for classification problems as support vector classification (svc) [21]. this approach seeks to classify hyperplanes that are capable of unravelling datasets into elevated-dimensional spaces of features. the margin is called the distinction between datasets, and the svm maximizes the margin [22]. when dealing with binary classification problems, the instances are alienated with the hyperplane �� × � +� 0, where w is known as dimensional coefficient vector, b is the offset value, and x is the dataset value. some of the popular kernel choices are linear ��� × ���, polynomial ���� × �� +�� �, � > 0, ! � 2�, and radial basis function exp%−� ∥ �� −�( ∥ )* [23]. the purpose of using svm is to identify a hyperplane in binary dimensional space that distinctly classifies the data points. therefore, to separate the binary classes of data points, hyperplanes use the maximum distance between data points of both classes. the linear separable hyperplane is used for decision boundaries that assist in classifying the data points. the data points falling on either side of the hyperplane can be attributed to different classes. the size of the hyperplanes is dependent on the numbers of input features. when there is a misclassification done by the model on the prediction of the class of the data point, the loss is included along with the regularization parameter to perform gradient update. however, the required features are extracted from the dataset and split into the training and testing data. the learning rate of the model is set to 0.0001 and the regularization parameter λ is set to 1/epochs. therefore, the regularizing value reduces the number of epochs increased. 3.4.3. naive bayes nb is an algorithm for supervised learning based on the conditional probability theorem to evaluate the class function vector [24]. nb algorithm assumes that all the variables in the dataset are not correlated to each other. it can be written as: ( | ) ( ) ( | ) ( ) p x c p c p c x p x × = (3) where p(c|x) is the posterior probability of class c; p(c) is the probability of class c; p(x|c) is the likelihood probability of the predictor given class c; p(x) is the prior probability of predictor. when working with the dataset, the values of the dataset associated with each class are distributed according to a normal distribution. the developing process of nb model is that the data is defined by a normal distribution with no co-variance between dimensions. this nb model is fit by finding the mean and standard deviation of the points within each label. the cvd database of various patients is based on 11 columns including gender, age, height, and weight (table 1). the target is whether the patient is normal or having cardiac problems. if the targeted column p(c) has value 1, then it means that the person has heart problems, and if the value is 0, then it means that the person is healthy. 3.4.4. decision tree classifier the dt (j48) supervised algorithm is based on tree shape structures where every node is a decision node [25]. decisions are taken on the basis of an algorithmic method that identifies the ways to divide a dataset based on diverse conditions [26]. in dt, internal and external nodes are linked together, which make the decision done easily and efficiently. 4. results and discussion this section contains the outcomes produced by the ml classifiers. the outcomes of ml classifiers are evaluated in terms of different performance evaluation matrices. the cross-validation (cv) approach is applied, and the performance matrices of each ml classifiers are also calculated. the details are given in the following subsections. 28 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 24-34 4.1. k-cross-validation cv is probably considered to be the most important techniques used to validate the stability of the ml model on new data [27]. in k-fold cv, the dataset is separated into k equal size in which k-1 groups are used for training and the rest is for testing purposes. the performance of the classifiers is based on the k results. in this study, k =15 cv is applied, which means that 85% data is applied for training process and the rest is for testing the classifiers. 4.2. performance evaluation matrices various performance evaluation matrices are used in this study. the binary confusion matrix 2 × 2 is used because two categories are included in the proposed output variables. in addition, it provides two types of correct and incorrect prediction. a confusion matrix is shown in fig. 2. fig. 2 confusion matrix classification accuracy shows the performance, and the classification error shows the incorrect classification of the model. they can be written as: fp fn classification error tp tn fp fn + = + + + (4) tp tn accuracy tp tn fp fn + = + + + (5) sensitivity is also known as “true positive (tp) rate”. it confirms that if a diagnostic test is positive, then the subject is a cardio patient. it can be written as: 100% tp sensitivity tp fn = × + (6) specificity is the “true negative (tn) rate” classifier for detecting negative cases. when the diagnostic test is negative, the person is healthy. it can be mathematically written as: 100% tn specificity tn fp = × + (7) the mathematical expression of precision is written as: 100% tp precision tp fp = × + (8) mcc represents the prediction ability of a classifier with the values which vary between [−1, +1]. if mcc classifier has the value +1, this indicates that the predictions of the classifier are ideal; if mcc classifier has the value -1, it indicates that the classifier produces incorrect predictions; the value near 0 implies that random predictions are created by the classifier. mcc’s mathematical expression is written as: 29 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 24-34 ( ) ( ) ( ) ( ) tp tn fp fn mcc tp fp tp fn tn fp tn fn × − × = + + + + (9) receiver operating characteristic curve (roc) analysis compares the tp rate and false positive (fp) rate in the classification results while the area under the curve (auc) characterizes the roc of classifier. the larger the value of auc is, the more effective the performance of the classifier is. 4.3. results and discussion this section contains the outcomes generated by the classification models in terms of variant kinds of performance evaluation matrices. four different ml algorithms are tested for accuracy, i.e., nb, lr, j48, and svm on full features of cvd dataset. all computations are performed in weka software on an intel® core™ i7-6700 cpu @ 3.40 ghz personal computer with installed memory (ram) of 16 gb. the 15-fold cv test is applied, and the data is analyzed visually to figure out the distribution of values in terms of effectiveness and efficiency. the effectiveness of all four classifiers in terms of model building time, accuracy, correctly and incorrectly classified instances is calculated (table 2). the results of the 15-fold performance assessment of four classifiers nb, lr, j48, and svm on cvd dataset are shown in table 2. the processing time of each classifier with nb is 0.16 seconds in fig. 3, which is very fast computationally compared to the lr, j48, and svm. furthermore, the processing time of svm is 148.93 seconds which indicates that svm is a lazy learner. from comparative graph (fig. 4), it is easily noted that j48 has the highest value of correctly classified instances and the lower value of incorrectly classified instances while nb has the lowest value of correctly instances and the higher value of incorrectly classified instances. in table 3, nb shows poor performance that has 58.87% classification accuracy, 27.77% specificity, 58.9% sensitivity, and 22.6% mcc. the nb regression specificity value is 27.77%, indicating that the likelihood of diagnostic tests is negative, and the individual does not have cvd. in addition, 58.9% sensitivity shows the likelihood that the diagnostic test is positive for the person. for lr, the experiment is performed by using 15-fold cv. however, the performance of lr is good as shown in table 3. table 2 evaluation of classifiers in term of time, correctly instances, and incorrectly instances evaluation criteria classifiers nb lr j48 svm time taken to build a model 0.1 1.28 3.33 148.93 correctly classified instances 41209 50502 51159 45405 incorrectly classified instances 28791 19498 18841 24595 0.16 1.28 3.33 148.93 0 20 40 60 80 100 120 140 160 nb lr j48 svm time in seconds fig. 3 classifier processing time in seconds 30 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 24-34 fig. 4 comparative graph of correctly and incorrectly classified instances table 3 classifier 15-fold cv classification performance evaluation evaluation criteria classifiers nb lr j48 svm accuracy (%) 58.87 72.1457 73.0843 64.8643 specificity (%) 27.77 67.79 70.19 59.87 sensitivity (%) 58.9 72.1 73.1 64.9 mcc 0.226 0.445 0.462 0.299 precision (%) 64.4 73.3 73.2 65 the lr classifier achieves 72.1457% accuracy, 67.79% specificity, 72.1% sensitivity, and 44.5% mcc. similarly, svm has the 64.86% classification accuracy, 59.87% specificity, 64.9% sensitivity, and 29.9% mcc. the performance of j48 is excellent which achieves the highest 73.0843% classification accuracy, the best 70.19% specificity, 73.1% sensitivity and 46.2% mcc value. the output of j48 outperforms the other three classifiers in terms of precision, sensitivity, specificity, and mcc, as shown in fig. 5. j48 has 73.0843% predictive accuracy, 73.1%sensitivity, and 70.19% specificity. lb, which has 72.1547% classification precision, is the second significant classifier. the worst performance is observed in terms of precision, sensitivity, and specificity for nb classifiers. in fig. 6, the accuracy 73.0843% obtained by j48 is greater than the accuracy obtained by nb, lr, and svm (which varies between 58.87% and 72.1457%). it is observed that j48 and lr have high mcc values, while nb and svm have the lowest mcc values with 15-fold cv on the full feature’s dataset, as depicted in fig. 7. therefore, it means that j48 and lr are the highest prediction ability classifiers as compared to nb and svm. simulation errors are found to test the better performance of classifiers. to do so, the measured efficacy of all four classifiers is in provisos of kappa statistic (ks), mean absolute error (mae), root mean squared error (rmse), relative absolute error (rae), and root relative squared error (rrse). ks, mae, and rmse show numeric values while rae and rrse are in percentage. the detailed results are shown in table 4. the graphical representation for simulation of errors is presented in fig. 8. 0 10000 20000 30000 40000 50000 60000 nb lr j48 svm correctly classified instances incorrectly classsified instances 0 10 20 30 40 50 60 70 80 nb lr j48 svm accuracy (%) specificity (%) sensitivity (%) fig. 5 performance in terms of precision, specificity, and sensitivity of classifiers 31 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 24-34 table 4 summary of simulation errors evaluation criteria classifiers nb lr j48 svm ks 0.17 0.44 0.4617 0.29 mae 0.41 0.39 0.3611 0.35 rmse 0.53 0.43 0.43 0.59 rae (%) 83.85 79.09 72.21 70.27 rrse (%) 106.10 87.60 86.91 118.55 from table 4, the outcomes represent the chance of having a best classification (0.4617) with the least warning error rate (0.3611) formed by j48. furthermore, the second-best classification (0.4429) with error rate (0.3955) is produced by lr. nb and svm have the highest value of error rate, which explain the large number of incorrectly classified instances for each algorithm. the value of mae is observed to evaluate the quality of the classifiers. from table 4, it is clearly identified that j48 and svm have minimum mae values as compared to other classifiers nb and lr (fig. 9). 58.87 72.1457 73.0843 64.8643 0 10 20 30 40 50 60 70 80 nb lr j48 svm accuracy 0.226 0.445 0.462 0.299 0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5 nb lr j48 svm mcc 0 0.2 0.4 0.6 0.8 1 1.2 ks mae rmse rae rrsse nb lr j48 svm fig. 6 classifier accuracy test with 15-fold cv on complete dataset features fig. 7 mcc of all four classifiers with 15-fold cv fig. 8 compared simulation error with respect to ks, mae, rmse, rae, and rrse 32 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 24-34 j48 and svm have the nearest predictive values to the target values as compared to lr and nb. the confusion matrix is used for evaluating the performance of a classification model. the confusion matrix of all four classifiers is presented in fig. 10 below. (a) j48 (b) lr (c) svm (d) nb fig. 10 confusion matrix 5. conclusions and future work in this modern era, the amount of data collected from health surveillance has rapidly increased with the advancement of the medical technology. it is foreseeable that ai platform will enable the analysis of multiple infectious diseases including cvd, which is considered to be one of the main causes of death worldwide. in this study, variable generalization performance of supervised learning algorithms for cvd diagnosis was evaluated on cvd dataset by using four well-known supervised ml classifiers. in accordance with the obtained output, the higher probability of achieving an improved result can be inferred in terms of the accuracy of the cvd classification when j48 is applied. in the future work, the ensemble of pre-trained models would be applied for classifying cvd. conflicts of interest the authors declare no conflict of interest. statement of ethical approval for this type of study, statement of human rights is not required. statement of informed consent for this type of study, informed consent is not required. references [1] j. mackay, g. a. mensah, s. mendis, and k. greenlund, the atlas of heart disease and stroke, geneva: world health organization, 2004. [2] z. s. wong, j. zhou, and q. zhang, “artificial intelligence for infectious disease big data analytics,” infection, disease, and health, vol. 24, no. 1, pp. 44-48, february 2019. 0.3 0.32 0.34 0.36 0.38 0.4 0.42 0.44 nb lr j48 svm mean absolute error fig. 9 mae for nb, lr, j48, and svm 33 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 24-34 [3] s. schneeweiss, “learning from big health care data,” the new england journal of medicine, vol. 370, no. 23, pp. 2161-2163, june 2014. [4] a. blumenthal, “artificial intelligence to fight the spread of infectious diseases,” https://phys.org/news/2018-02-artificial-intelligence-infectious-diseases.html, february 20, 2018. [5] k. e. goodman, j. lessler, s. e. cosgrove, a. d. harris, e. lautenbach, j. h. han, et al., “a clinical decision tree to predict whether a bacteremic patient is infected with an extended-spectrum β-lactamase-producing organism,” clinical infectious diseases, vol. 63, no. 7, pp. 896-903, october 2016. [6] a. u. haq, j. p. li, m. h. memon, s. nazir, and r. sun, “a hybrid intelligent system framework for the prediction of heart disease using machine learning algorithms,” mobile information systems, vol. 2018, 3860146, december 2018. [7] m. elsisi, k. mahmoud, m. lehtonen, and m. m. darwish, “reliable industry 4.0 based on machine learning and iot for analyzing, monitoring, and securing smart meters,” sensors, vol. 21, no. 2, 487, january 2021. [8] m. q. tran, m. k. liu, and m. elsisi, “effective multi-sensor data fusion for chatter detection in milling process,” the international society of automation transactions, in press. [9] m. elsisi, m. q. tran, k. mahmoud, d. e. a. mansour, m. lehtonen, and m. m. darwish, “towards secured online monitoring for digitalized gis against cyber-attacks based on iot and machine learning,” institute of electrical and electronics engineers access, vol. 9, pp. 78415-78427, june 2021. [10] j. p. li, a. u. haq, s. u. din, j. khan, a. khan, and a. saboor, “heart disease identification method using machine learning classification in e-healthcare,” institute of electrical and electronics engineers access, vol. 8, pp. 107562-107582, june 2020. [11] c. b. c. latha and s. c. jeeva, “improving the accuracy of prediction of heart disease risk based on ensemble classification techniques,” informatics in medicine unlocked, vol. 16, 100203, july 2019. [12] k. w. johnson, j. torres soto, b. s. glicksberg, k. shameer, r. miotto, m. ali, et al., “artificial intelligence in cardiology,” journal of the american college of cardiology, vol. 71, no. 23, pp. 2668-2679, june 2018. [13] a. raza, a. mehmood, s. ullah, m. ahmad, g. s. choi, and b. w. on, “heartbeat sound signal classification using deep learning,” sensors, vol. 19, no. 21, 4819, january 2019. [14] r. spencer, f. thabtah, n. abdelhamid, and m. thompson, “exploring feature selection and classification methods for predicting heart disease,” digital health, vol. 6, pp. 1-10, march 2020. [15] k. shameer, k. w. johnson, b. s. glicksberg, j. t. dudley, and p. p. sengupta, “machine learning in cardiovascular medicine: are we there yet?” heart, vol. 104, no. 14, pp. 1156-1164, july 2018. [16] a. m. alaa, t. bolton, e. di angelantonio, j. h. rudd, and m. van der schaar, “cardiovascular disease risk prediction using automated machine learning: a prospective study of 423,604 uk biobank participants,” public library of science one, vol. 14, no. 5, e0213653, may 2019. [17] v. d. soni, “detection of heart disease using machine learning techniques,” international journal of scientific and technology research, vol. 9, no. 8, pp. 285-288, august 2020. [18] n. l. fitriyani, m. syafrudin, g. alfian, and j. rhee, “hdpm: an effective heart disease prediction model for a clinical decision support system,” institute of electrical and electronics engineers access, vol. 8, pp. 133034-133050, july 2020. [19] h. malik, m. s. farooq, a. khelifi, a. abid, j. n. qureshi, and m. hussain, “a comparison of transfer learning performance versus health experts in disease diagnosis from medical imaging,” institute of electrical and electronics engineers access, vol. 8, pp. 139367-139386, june 2020. [20] b. alić, l. gurbeta, and a. badnjević, “machine learning techniques for classification of diabetes and cardiovascular diseases,” 6th mediterranean conference on embedded computing, pp. 1-4, june 2017. [21] x. zhai and c. tin, “automated ecg classification using dual heartbeat coupling based on convolutional neural network,” institute of electrical and electronics engineers access, vol. 6, pp. 27465-27472, may 2018. [22] s. r. saufi, z. a. b. ahmad, m. s. leong, and m. h. lim, “challenges and opportunities of deep learning models for machinery fault detection and diagnosis: a review,” institute of electrical and electronics engineers access, vol. 7, pp. 122644-122662, august 2019. [23] m. mohammed, m. b. khan, and e. b. m. bashier, machine learning: algorithms and applications, boca raton: chemical rubber company press, 2017. [24] r. y. goh and l. s. lee, “credit scoring: a review on support vector machines and metaheuristic approaches,” advances in operations research, vol. 2019, 1974794, march 2019. [25] n. friedman, d. geiger, and m. goldszmidt, “bayesian network classifiers,” machine learning, vol. 29, pp. 131-163, november 1997. [26] m. mohri, a. rostamizadeh, and a. talwalkar, foundations of machine learning, cambridge: massachusetts institute of technology press, 2018. [27] r. r. isnanto, a. f. rochim, d. eridani, and g. d. cahyono, “multi-object face recognition using local binary pattern histogram and haar cascade classifier on low-resolution images,” international journal of engineering and technology innovation, vol. 11, no. 1, pp. 45-58, january 2021. copyright© by the authors. licensee taeti, taiwan. this article is an open access article distributed under the terms and conditions of the creative commons attribution (cc by-nc) license (https://creativecommons.org/licenses/by-nc/4.0/). 34 4___peti#7653__35-56 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 artifact removal methods in eeg recordings: a review mariyadasu mathe 1, 3, * , padmaja mididoddi 2 , battula tirumala krishna 1 1 department of electronics and communication engineering, jawaharlal nehru technological university kakinada, andhra pradesh, india 2 department of electronics and communication engineering, velagapudi ramakrishna siddhartha engineering college, andhra pradesh, india 3 gandhi institute of technology and management, andhra pradesh, india received 10 may 2021; received in revised form 21 july 2021; accepted 22 july 2021 doi: https://doi.org/10.46604/peti.2021.7653 abstract to obtain the correct analysis of electroencephalogram (eeg) signals, non-physiological and physiological artifacts should be removed from eeg signals. this study aims to give an overview on the existing methodology for removing physiological artifacts, e.g., ocular, cardiac, and muscle artifacts. the datasets, simulation platforms, and performance measures of artifact removal methods in previous related research are summarized. the advantages and disadvantages of each technique are discussed, including regression method, filtering method, blind source separation (bss), wavelet transform (wt), empirical mode decomposition (emd), singular spectrum analysis (ssa), and independent vector analysis (iva). also, the applications of hybrid approaches are presented, including discrete wavelet transform adaptive filtering method (dwt-afm), dwt-bss, emd-bss, singular spectrum analysis adaptive noise canceler (ssa-anc), ssa-bss, and emd-iva. finally, a comparative analysis for these existing methods is provided based on their performance and merits. the result shows that hybrid methods can remove the artifacts more effectively than individual methods. keywords: eeg signal, canonical correlation analysis (cca), ssa-bss, emd-iva 1. introduction electroencephalogram (eeg) is a voltage test recording the electrical activity of the neurons in the brain. eeg is obtained in various frequencies including delta (1-4 hz), theta (4-8 hz), alpha (8-13hz), beta (13-30 hz), and gamma (greater than 30 hz) [1]. eeg brain rhythms with different frequency ranges are shown in table 1 [2]. eeg is suitable for brain signal acquisition because of its non-intrusiveness, portability, high temporal resolution, and cost-effectiveness. brain signal acquisition and processing are mainly used for clinical research purposes to diagnose multiple brain disorders, e.g., sleep disorders, depression, schizophrenia, and epileptic disorder [3]. however, eeg processing and analysis are affected by a couple of conditions, such as technical and physiological conditions. during the signal acquisition process, the most prevalent disorders (artifacts) are ocular, muscle, and cardiac disorders, and electrooculogram (eog) is cumbersome [4]. table 1 basic brain rhythms and their frequency range [2] serial number brain rhythm frequency range state of the brain 1 delta 1-4 hz sleeping/unconscious 2 theta 4-8 hz imagination 3 alpha 8-13 hz calm consciousness 4 beta 13-30 hz focused consciousness 5 gamma >30 hz peak performance * corresponding author. e-mail address: mmathe@gitam.edu tel.: 9866420248 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 these artifacts, power line interference, impedance fluctuation, and wire movement collide and intrude the energy with eeg signals. due to these artifacts, it is difficult to obtain the correct diagnosis and analysis of eeg. therefore, artifact removal methods are required to reduce the artifacts and preserve the originality of the actual eeg signals. the placement of electrodes in the 10-20 system is shown in fig. 1. fig. 1 depicture of 10-20 electrode placement system [5] researchers developed automated approaches and techniques wherein the eeg signals are recorded using singleand multi-channels. different methods are proposed for the mitigation of artifacts from singleand multi-channels. the accuracy of the methods using single-channels is less than the one using multi-channels, but the complexity over time is less when using single-channels. the basic process includes recording the eeg data with precaution and removing the epochs in eeg data, which can cause the removal of some valuable data [6]. generally, artifact removal techniques are categorized into two types: regression-based method and blind source separation (bss) based method [7]. so far, no specific method can remove all the artifacts. researchers are trying to suppress the artifacts by improving the existing techniques and combing two and more methods to get the artifact-minimized eeg signals. the main objective of this study is to give an overview on the existing methodology to remove the artifacts in eeg recordings and to choose the better method based on artifact removal types. the remaining paper is organized as follows. the artifacts affecting the eeg recordings are discussed in section 2. literature review is clearly explained in section 3. the primary artifact removal techniques’ advantages and disadvantages proposed by various researchers are discussed in section 4. the comparative analysis for multiple artifact removal methods is presented in section 5. 2. about the artifacts human brain is a combination of billions of neurons, and the activity of this brain region is measured by recording the electrical signals generated by the neurons. eeg is such a signal recording acquired by placing the electrodes on the scalp. the recorded eeg signals include the signals from the brain neurons and the neurons of other body parts. the signals generated by other than brain neurons are called artifact signals. these artifacts are due to ocular artifacts (oas), muscle artifacts, cardiac artifacts, electrode placement, impedance mismatching, etc. [8]. non-physiological artifacts can be removed by simple methods such as high-pass filtering and low-pass filtering because these artifacts’ frequency does not overlap with the desired eeg signals [9]. however, the removal of physiological artifacts requires different methods because the frequency spectrum of physiological artifacts overlaps with the desired eeg signals [10]. oas are generated from eye movements (ems) and eye blinks, and are propagated to the scalp region and overlap with the eeg signals. the oa frequency range is petite and is recorded by eog by placing the electrodes (one above the eye and the other on one side of the eye) as shown in fig. 2. eeg signals are deteriorated by eog signal components and vice versa, so bi-directional interference is introduced [11]. the amplitude of eog is higher than the amplitude of eeg. muscle artifacts are originated from different muscle groups such as contraction, stretch, talking, sniffing, swallowing, etc. the frequency range of 36 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 muscle artifacts is vast and is overlapped entirely with the actual eeg signals. the amplitude of muscle artifacts depends on the activity of the muscle and is measured by electromyography (emg). because of the broad spectrum, muscle artifacts cannot be obtained by single-channels. the autocorrelation of muscle artifacts is very small and independent of the eeg signals both temporally and spatially [8]. cardiac artifacts are originated from the heart and are measured by electrocardiography (ecg). cardiac artifacts follow some regular patterns, so removing the cardiac artifacts is simple by using reference signals. fig. 2 eog and emg elements of a sleep montage setup [5] for the better analysis of human brain and the diagnosis of patients, artifacts must be removed. generally, the recorded eeg signals are between the ranges of 0.01 hz to 100 hz. different types of the brain rhythms and the artifacts generated by the brain and body are shown in fig. 3 and fig. 4. (a) delta wave (b) theta wave (c) alpha wave (d) beta wave (e) gamma wave fig. 3 graphical representation of different brain rhythms fig. 4 different physiological artifacts [8] 37 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 3. literature survey 3.1. related works in 2013, zou et al. [12] have addressed an automatic algorithm for general artifact identification. it was composed of electrode-scalp impedance information and an event-associated feature-oriented clustering algorithm. the former identified the non-biological artifacts, and the latter identified the artifacts with physiological origins. it could efficiently remove, separate, and detect the non-biological and physiological artifacts. it efficiently improved the signal quality. the classification accuracy was also improved. mahajan and morshed [13] have proposed a uniquely robust, unsupervised, and computationally speed statistical algorithm that employed modified multiscale sample entropy (mmse) and kurtosis to recognize the independent eye blink artifactual components automatically. these components were denoised by the biorthogonal wavelet decomposition. the persistent neural activity was preserved in the independent components (ics), and the artifactual activity was removed. their proposed method attained average specificity, sensitivity, and execution time, which are better than the ones in other previously existing methods. in 2014, acharjee et al. [14] have described the conflict of gradient artifact removal from the eeg signals recorded in a concurrent manner with functional magnetic resonance imaging (mri) acquisition. the artifacts were estimated by the quasi periodicity and its similarity over distinct channels with the help of independent vector analysis (iva). their proposed method utilized spatio-temporal information to evaluate the artifacts for a specific channel with the support of spatial dependences. hence, it offered robustness in terms of uncontrollable variations such as fluctuations and head movement during acquisition. in 2015, khatun et al. [15] have evaluated the unsupervised wavelet transform (wt) approach for removing oas in the case of a single-channel eeg system. a group of 7 raw eeg dataset was investigated. two familiar wt techniques, i.e., stationary wavelet transform (swt) and discrete wavelet transform (dwt), were used. four wt basis functions were utilized for removing oas, i.e., bior4.4, sym3, coif3, and haar, having statistical threshold (st) and universal threshold (ut). their proposed method was an efficient tool for real-time applications. in 2016, tibdewal et al. [16] have performed detection of em artifacts and classification of non-artifactual/artifactual eeg time series together with multiple em artifactual zones. artificial neural network (ann) was employed for the identification. the researchers also proposed a computationally fast, robust, simple, and novel time-amplitude algorithm. their proposed method provided the ann model accuracy for the classification of the artifacts, and detected and marked the multiple em artifactual zones automatically and appropriately. liu et al. [17] have introduced a practical and efficient filtering algorithm based on multiscale entropy and multivariate empirical mode decomposition (memd) for computing the depth of anesthesia (doa). the mean entropy found the artifact-free intrinsic mode functions (imfs). the impact of distinct levels was investigated by means of the simulated data. the complexity was also computed for monitoring the doa. the ann was used for the range mapping for correlating the measurements with bispectral index (bis). the bis and entropy values contained a more robust correlation, and it also efficiently filtered the artifacts. in 2017, barua et al. [18] have proposed an algorithm which is known as automated artifact handling in eeg (arte). it was employed as a pre-processing step in the driver monitoring application, and functioned based on the wavelets, hierarchical clustering, and independent component analysis (ica). it was analyzed on a driver sleepiness study, and was evaluated with the help of a clinical neurophysiologist using quantitative measures. the outcomes demonstrated that it minimized the effect of artifacts in the recorded eeg signals. it was a data-driven one, and it did not depend on manually defined thresholds or extra reference signals. sai et al. [19] have introduced a new technique for artifactual component identification with the help of a pre-trained support vector machine (svm). the researchers also proposed an extendable and robust system that enabled removing the artifacts without arbitrary thresholding. the target artifacts were removed in a successful manner. the training 38 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 used a group of features such as amplitude range, shannon’s entropy, variance, and kurtosis. the multiple forms of artifacts in the multi-channel eeg were also accommodated. li et al. [20] have introduced a unified framework based on canonical correlation analysis (cca) to remove artifacts. the raw signals constructed a pair of matrices based on the sources via the autocorrelation maximization. these sources were eradicated by placing them as zeros, and the remaining sources reconstructed the artifact-free eeg. the actual and simulated recorded data assessed this framework. it lessened the associated motion artifacts and contained benefits for eliminating the gradient artifacts with respect to the computational complexity and combination of characteristics. roy and shukla [21] have compared the effectiveness of enhanced empirical mode decomposition (eemd) with distinct interpolations based on an artifact removal technique. the input single-channel eeg signals were transformed to multi-channel signals, and the motion artifact randomness was eradicated by dwt filtering and cca. their proposed method acted as a supplement for the traditional algorithms and provided enhancements in difference signal to noise ratio (dsnr) and several performance parameters. in 2019, chang et al. [22] have evaluated artifact subspace reconstruction (asr) during the simulated driving experiments. the ic classifier and ica separated the artifacts for assessing the efficiency of asr. the temporal activity power was minimized. it enhanced the quality of a subsequent ica decomposition. the brain activities were retained, and non-brain signals were removed. their proposed method was an automatic and powerful artifact removal technique for online real-time eeg applications or offline data analysis like brain-computer interfaces and clinical monitoring. cheng et al. [23] have addressed an efficient technique for removing different artifacts for the single-channel eeg. it was a combination of second-order blind identification and singular spectrum analysis (ssa) method. the possible cases contaminated by ecg, eog, and emg artifacts were analyzed. their proposed method removed different artifacts and acted as a promising tool for biomedical signal processing applications. saini et al. [24] have introduced a robust framework for detecting and eliminating oas based on turning point count and variational mode decomposition (vmd). it exploited the efficiency in two stages, i.e., vmd-i and vmd-ii. it was composed of four components: eeg signal decomposition into two modes using vmd-i, rejection of low-frequency baseline components, processed eeg signal decomposition into three modes using vmd-ii, rejection of mode containing oas based on turning point count-based threshold criteria. the oas were eliminated with minimal loss in the entire local rhythms and reconstructed eeg signals. liu et al. [25] have labeled a robust and effective muscle artifact removal technique with the help of fast multivariate empirical mode decomposition (fmemd) and cca in the case of few-channel eeg. the input eeg recordings were decomposed into various multivariate imfs. the cca processed the imfs for measuring the underlying sources. in the final step, the sources with less autocorrelations were eradicated. the average correlation coefficient (cc) and the accuracy were largely consistent. their proposed method worked much robust in the low sampling rate based on the benchmark and real data. dora and biswal [26] have suggested an ecg artifact correction algorithm for the automatic diagnosis and analysis of single-channel eeg signals. it employed a modified and enhanced signal decomposition version for attaining band limited intrinsic mode functions (blimfs). it was helpful when the signal was composed of correlated properties. it suppressed the qrs complexes and estimated the ecg reference. it offered less distortion and was computationally more intensive than other methods. it overcomes the shortcomings such as the requirement of a reference ecg channel and the requirement of r-r interval or amplitude thresholding for qrs complex identification. in 2020, dora et al. [27] have introduced an adaptive ssa algorithm for muscle artifact removal. the mobility threshold was decided in an adaptive manner by neural network regressor (nnr). their proposed method was trained using several contamination levels. it discriminated among several contamination levels and performed superior to the traditional single-channel algorithms. yedurkar and metkar [28] have proposed a technique for removing physiological artifacts and 39 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 positioning the epileptic region. a hybrid method was recommended based on multi-resolutional analysis and adaptive filtering (mraf) . in the initial step, the eeg signals were decomposed by dwt that localized the epileptic region. the multi-resolutional soft thresholding removed the abrupt changes. moreover, the adaptive filtering (af) eradicated the low-frequency components that described the physiological artifacts. it retained most of the seizure signals that were available in the validated datasets. islam et al. [29] have introduced a technique for removing motion-related artifacts in the field of epilepsy. initially, it involved the eeg signal recording with the help of a wearable eeg headset. it was colored using few motion artifacts produced in a lab-controlled experiment. it was followed by the spectral and temporal spectral signal characterization and artifact removal by ica. the actual clinical eeg data tested it, and the outcomes described an average improvement in accuracy for detecting the seizures and prediction. bajaj et al. [30] have addressed an algorithm based on wavelet packet decomposition (wpd) that permitted the presumed artifact removal by optimizing intuitive parameters. it was composed of 2 tuning parameters and 3 operating modes. it was compared with comparative wavelet-oriented technique and ica-oriented techniques. the performance could be enhanced by proper optimization of the parameters for an individual predictive method. it was accomplished better for artifactual noise removal and neural information retention computed using cc and mutual information (mi) than the comparative wavelet-oriented technique. in 2021, noorbasha and sudha [31] have labeled the conflict of eog artifact removal and separation of different cerebral activities in a single-channel contaminated eeg. a new method based on ica and ssa together with swt was proposed. the single-channel contaminated eeg signals were transformed into multivariate information. then, ica separated the source signals as distinct ics. the swt accomplished the thresholding for preserving the eeg signals and dividing the actual artifacts. their proposed method provided better artifact separation characteristics than the traditional approaches. 3.2. utilized datasets several kinds of datasets are used in the considered previous research, including the manual dataset, mendeley, mit-bih polysomnographic database, eeg during mental arithmetic task database, synthetic dataset, simulated dataset, temple university hospital (tuh) eeg database, gold-standard dataset, epileptic eeg signal dataset from the local hospital, freiburg eeg dataset, semi-simulated dataset, real dataset, open-source dataset, etc. table 2 lists different kinds of datasets used in the considered works in a detailed manner. most of the considered works utilize the manual dataset, followed by a real dataset, simulated dataset, mit-bih polysomnographic database, etc. the manual dataset is used in 50% of the works, real dataset was used in 20% of the works, simulated dataset was used in 15% of the works, mit-bih polysomnographic database was used in 10% of the works, and the remaining datasets was used only in 5% of the works. table 2 different datasets used in the considered works dataset previous work manual dataset [12-15, 18-19, 22-23, 25, 31] mendeley [24] mit-bih polysmnographic database [24, 26] eeg during mental arithmetic task database [24] synthetic dataset [31] simulated dataset [17, 20, 27] tuh eeg database [27] gold standard dataset [28] epileptic eeg signal dataset from local hospital [28] freiburg eeg dataset [29] semi-simulated dataset [26] real dataset [16-17, 20, 30] open-source dataset [21] 40 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 3.3. simulation platforms the platforms used in the considered works are given in fig. 5. here, matlab and labview softwares are used for the implementation. 92% of the works use the matlab software, and 8% use the labview software. 3.4. performance measures the performance measures used in the analyzed research are depicted in table 3, including accuracy, sensitivity, specificity, mean square error (mse), root mean square error (rmse), relative root mean square error (rrmse), etc. among the considered works, time measure is used in 25% of the works, cc is used in 45% of the works, mi is used in 25% of the works, accuracy is used in 20% of the works, sensitivity is used in 25% of the works, specificity is used in 15% of the works, mse is used in 10% of the works, mean absolute error (mae) is used in 20% of the works, rrmse is used in 25% of the works, rmse is used in 15% of the works, signal to noise ratio (snr) is used in 30% of the works, power spectral density (psd) is used in 10% of the works, signal to artifact ratio (sar) is used in 10% of the works, and the remaining measures such as euclidean distance, normalized mean square error (nmse), spectral coherence, agreement rate, signal quality index (sqi), relative error (re), normalized root mean square error (nrmse), variance, power, maximum absolute error (max), normalized maximum absolute error (nmax), normalized root mean square difference (nrd), percentage root mean square difference (prd), autocorrelation, peak signal to noise ratio (psnr), precision, false alarm rate (far), coherence, mean entropy value (mev), dsnr, and spectral distortion improvement are used only in 5% of the contributions. table 3 performance measures used in the previous research ref. time cc mi accuracy sensitivity specificity mse mae rrmse rmse snr psd sar miscellaneous [12] √ euclidean distance [13] √ √ √ √ √ √ spectral coherence and agreement rate [14] √ [15] √ √ √ nmse [16] √ [17] √ √ mev [18] √ √ √ sqi, re, and nrmse [19] √ √ √ [20] √ √ √ [21] √ dsnr and spectral distortion improvement [22] variance and power [23] √ √ √ √ [24] √ √ √ √ √ √ max, nmax, nrd, and prd [25] √ √ √ √ autocorrelation [26] √ √ √ √ [27] √ √ √ psnr [28] √ √ √ precision [29] √ √ far [30] √ √ coherence [31] √ √ √ 92% 8% contribution matlab labview fig. 5 simulation platforms used in the literature 41 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 4. survey on the existing artifact removal techniques 4.1. artifact avoidance and rejection the first basic technique to remove artifacts is to instruct the patient to shun unnecessary actions such as ems and stay calm until the completion of the process. however, since it is not possible to keep quiet, artifacts are recorded in the eeg signals [3]. the second way is to remove the epochs in eeg data by observation. this method is also not available because the data is vast, and the expertise must be analyzed appropriately. therefore, a separate automatic procedure is needed to remove the artifacts as much as possible [32]. 4.2. regression methods the regression method is a straightforward and commonly used method for removing artifacts in eeg signals. in this method, a reference signal is required to remove artifacts. eeg signals are generally most affected by eog and emg signals. the regression method depends on the subtraction of artifact signals from the eeg signals. it assumes that oas and eeg signals are linearly uncorrelated [33]. identifying the reference artifact signals is very difficult because eeg and eog are both neural activities [2]. while subtracting eog signals, some useful information related to the original eeg signals was also removed; loss of data is there. the simple mathematical analysis is shown here in eq. (1). ( ) ( ) ( ) ; with 1, 2, ,oseeg l eeg l eog l l o= + = ⋯γ (1) here, the recorded eeg signals are denoted as eegs the fraction of the eog signals is denoted as �, and the original eeg signals are denoted as eego respectively. 4.3. filtering methods 4.3.1. adaptive filtering the performance of wt depends upon the mother wavelet effectively. the wt efficiency is improved if the artifact component’s frequency does not overlap with the usage frequency of the eeg signals. unfortunately, ecg, emg, and eog artifact frequencies are coinciding with the desired eeg signals. in ica, the improvement performance of artifact removal depends on the reference signal and a large amount of data. however, af overcomes the problem of extensive data and frequency band overlap [34]. fig. 6 presents a block diagram to implement af. fig. 6 basic block diagram of adaptive filter [34] more than one channel can be selected as a reference channel, and af is employed mainly to minimize power line interference artifacts. however, this has two particular limitations: (i) the noise content in the original signals must be corrected with some reference noise signals; (ii) the original signal helpful information may be uncorrected with the reference noise. 42 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 4.3.2. kalman filtering kalman filter (kf) is one of the appropriate estimators for the analysis of linear systems and signals. it is a recursive filtering algorithm to remove the artifacts in eeg signals. the inputs used for this algorithm are the state of the system, recorded signal, and past observations. in this algorithm, a reference signal is not required; kf estimates the state of the system and the reference signal from the contaminated part of the eeg signals. the recorded eeg signals are denoted as eegr, true eeg signals are denoted as eegt, and oas are denoted as eog. kf uses a state-space model to represent the complete analysis. the state-space model of the kf is as follows [35]: ( ) ( 1) ( 1) ( 1)x n ax n bu n cw n= − + − + − (2) ( ) ( ) ( )z n hx n v n= + (3) where z(n) is the measured signal; x(n) is the system’s state; a, b, c, and h are the matrices obtained from different regions of the acquired eeg signals; w(n) and v(n) are the process noise and measurement noise respectively; u(n) is the control input for the system. the measured signal is the combination of eegt and artifactual signals like eog [36]. ( ) ( ) ( ) t z n eeg eog hx n v n= + = + (4) after estimating eeg and eog components, kf is applied to remove the artifacts. in this algorithm, two steps are there. (1) estimate the state variables. (2) compare the previous estimation with the measured signal to obtain an improved version of the state variables. the disadvantage of kf is that it assumes the errors are gaussian, and the system is a linear model. the measurement of covariance matrices for process noise and measured noise is tough to calculate because the practical recorded eeg signals are not a linear combination of the underlying source signals. the extension of kf is extension kalman filter (ekf) and unscented kalman filter (ukf), which work on nonlinear systems. 4.4. blind source separation (bss) bss is a separation approach of underlying sources from the mixed signals without knowing prior information about the origin of signals and mixing parameters. the main plan of bss is to recover the sources from mixed signals [37]. let us consider x as an observed eeg signal from the scalp electrodes, which is a linear mixture of underlying sources (s) with unknown mixing matrix a, eq. (5) is obtained. x as= (5) bss algorithm is reverted to recover the underlying sources. y ws= (6) where y is the estimation of sources, and w is an un-mixing matrix. these algorithms are used to remove artifact components and to reconstruct artifact-free eeg signals. in the following section, some of the bss algorithms are discussed. 4.4.1. principle component analysis (pca) principle component analysis (pca) is one of the simple data reductions and bss techniques to remove the artifacts using the orthogonality principle [8]. berg and scherg proposed pca to minimize oas. pca is a multiple data analysis procedure that involves the transformation of many correlated variables into a smaller number of uncorrelated variables. pca is defined 43 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 as “linear projection that transforms multivariate data into a set of components called principle components (pcs)” [38]. pcs are calculated using a simple algorithm called singular value decomposition (svd). in svd, eigenvalues are computed from the covariance matrix (c) [36]. most significant eigenvalues are related to the highest variance, and the lowest eigenvalues are related to the lower variance. assume the recorded eeg signals (eegr) combine the original eeg signals (eego) and the artifacts from eog and emg, eq. (7) is obtained. r oeeg eeg eog emgα β= + × + × (7) where α and β are real constants. the drawback of pca is the orthogonality principle only because it cannot separate the artifacts if they have the same amplitude as original eeg data. the orthogonality principle is not always held due to the nonlinearity between the electrical signals generated by brain neurons from eye blinks and ems compared with original eeg signals [39]. the advancement in pca is done by robust pca and kernel pca [40]. the performance of pca is better compared to the regression method and filtering methods. pca can remove oas, but other artifacts cannot be removed. therefore, researchers preferred alternative methods such as ica, cca, empirical mode decomposition (emd), iva, ssa, etc. 4.4.2. independent component analysis (ica) ica is a powerful bss method to remove the artifacts in eeg signals. it converts the multi-dimensional data into different ics. in this method, the observed eeg signals’ primary assumption is a linear mixture of underlying independent sources [39]. pca is unable to differentiate the artifacts if the neural activity and artifacts have the same amplitude. assume that the recorded eeg signals x(k) use n channels and that the number of underlying sources s(k) are n [2], eq. (8) is obtained. ( ) ( )x k as k= (8) where x(k) = [x1(k), x2(k), …, xn(k)] t and s(k) = [s1(k), s2(k) , …, sn(k)] t . t and k are transpose operator and time index operator, respectively. a ϵ r nxn is mixing the unknown matrix. the goal of ica is to recover s from x without knowing a and s. for this, a separate un-mixing matrix w should be found such that eq. (9) is obtained. ( ) ( )y k w x k= (9) where y is the expected value of s. y is precisely equal to s if and only if w = a -1 . this model is by assuming no noise model. for experimental conditions, some noises are also included. a simple model of how the electrodes are placed to record the eeg signals from the brain is shown in fig. 7. fig. 7 a simple model of recording eeg signals by electrodes [2] in the preprocessing of eeg signals using ica, two main steps are there; one is centering, and the second is whitening [37]. in the centering step, the eeg data is centered by subtracting the mean from the data. in the whiting step, the eeg data is de-correlated. ica uses higher-order statistics (hos) to identify independent artifactual components. unlike pca, 44 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 orthogonality is not required; only statistical independence is needed. the main drawback of ica is that the number of channels used to collect the data should be huge compared to underlying artifactual sources. the second drawback is that ica removes the artifacts whose topologies are stereotyped like eog, but muscle artifacts are not stereotyped because the muscle artifacts are from different sources. 4.4.3. canonical correlation analysis (cca) hotelling in 1936 has proposed cca. the entire concept of this approach is to find the vectors for two dataset variables so that the correlation between the projections of the variables on this basis vector is mutually maximized. this analysis relies on the coordinate system where the variables are described; however, if there is a strong linear relationship between the sets of multi-dimensional variables, they may depend on the coordinate system. this correlation relation may not be visible, but it may seek a pair of linear transformations. each set of variables are transformed, and the corresponding coordinates are correlated to the maximum [41]. cca is also treated as one of the powerful tools used in bss to remove artifacts in eeg signals. this analysis mainly uses second-order statistics (sos) with a less computational cost than ica (which uses hos) [42]. cca removes muscle artifacts easily because muscle artifacts do not have stereotyped topography [43]. muscle artifacts are distributing a total range of frequency spectrum. the autocorrelation of muscle artifacts is very low, while the eeg signal autocorrelation is high. cca uses this property to remove the emg artifacts which are mainly affected in the frequency range of 20hz-50hz. consider that there are a multi-channel eeg signal x(t) and its delayed version y(t) such that y(t) = x(t-1). the basis vectors are x = w t x and y = w t y. in this scenario, cca maximizes the correlation between x and y by removing the mean of x and y [3]. , , [ ][ ] ( , ) max max max ( )( )[ ] [ ] [ ] [ ] t t tt x y x xy t tx y x yt t t t t t x xx y yy yx x x y w w w w e w w w c we xx x y w c w w c we xx e yy e w w e w w ρ = = = (10) where cxx and cyy are auto-covariance matrices of x and y, respectively. cxy is the cross-covariance matrix of x and y. by calculating the derivative of eq. (10) with respect to wx and wy and equating to zero, the following equations are derived. ^ ^ 1 1 2 x x xx xy yy yx c c c c w wρ− − = (11) ^ ^ 1 1 2 y y yy yx xx xy c c c c w wρ− − = (12) where ρ is the canonical cc. ρ is maximized such that artifact components can be removed. if ρ is minimized, then these components approach to artifactual features. 4.5. wavelet transform (wt) the observed eeg signals are not entire because physiological and non-physiological signals contaminate the observed signals. for the better analysis of eeg signals, artifact signals are removed. in pca and ica, the drawback is maintaining orthogonality and large data requirements. these drawbacks are overcome by wt. wt decomposes the time-domain eeg signals by convolving a scaled and translated wavelet function [44]. , 1 ( ) ( ) a b n b n aa ψ ψ − = (13) where a and b are real numbers. n = 1, 2, 3, …., n. dwt removes oas because the wavelets have resembled oas [45]. here, some of the basic wavelets are shown in fig. 8. 45 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 (a) haar (b) shannon or sinc (c) daubechies 4 (d) daubechies 20 (e) gaussian or spline (f) biorthogonal (g) mexican hat (h) coiflet fig. 8 graphical representation of different basic wavelets dwt is a non-redundant and outstanding transform technique that aims to remove the artifacts from the time domain eeg signals. in this analysis, the signal is passed through a series of low-pass and high-pass filters to attain approximate and detailed coefficients [44]. the process is iteratively repeated until the desired frequency is reached. when this is achieved, the output is down-sampled by a factor of 2 and up-sampled to get the reconstructed signal. the basic block diagram of dwt is shown in fig. 9. fig. 9 dwt graphical decomposition representation [44] (hpf = high-pass filter; lpf = low-pass filter) dwt’s advantage is that it can provide good temporal resolution for high frequency components and better frequency resolution for lower frequency components [19]. the major drawback of dwt is its time invariance. however, eeg data is not time-invariant and is randomly varying. this problem can be overcome by swt, but the drawback is redundant information and slowness. in matlab, it is implemented by swt function [46]. 4.6. empirical mode decomposition (emd) the practical recorded eeg signals are random stochastic signals, but not static signals. in the standard methods of artifact removal, the eeg signals are projected onto the basis vectors. in wt, the basis vectors are predefined. the experimentally recorded eeg data is not following the same pattern for different patients and, at other conditions, follows the random nature. using predefined basis vectors to remove artifacts is not perfect because some proper eeg signals are also removed. a new method is proposed in 1998 to process nonlinear stochastic random data to overcome this problem [34]. emd decomposes the eeg signals into some oscillatory functions called imfs [47]. the derived imfs must follow two criteria. (1) in the whole dataset, the number of extremes and the number of zero-crossings must be equal or differ by at most one. (2) at any point, the mean value of the envelope is given by the local maxima, and the minima must be zero. 46 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 it is a data-driven method; unlike wt, the decomposition is not done by using predefined bass functions. here are the steps for performing emd [48]: (1) for the given eeg signal x(t), identify all the local maxima and minima of the given signal. (2) interpolate between the maxima and minima to calculate the upper envelope u(t) and lower envelope l(t), respectively. (3) calculate the average m1 from the lower and upper envelopes and subtract it from the signal x(t) to get the first imf. 1 ( ) ( ) 2 u t l t m + = (14) 1 1 ( )h x t m= − (15) (4) check if h1 satisfies the two criteria for imf. if not, repeat the steps (1)-(3). after i th iteration, the conditions are satisfied. 1( 1) 1 1i i i h m h− − = (16) then, c1 = h1i becomes the first imf. the stopping criteria are explained by huang et al. [49-50], which are given by cauchy convergence test. the first imf is expected to have a high frequency. 1 1 ( )x t c r− = (17) where r1 is the first remainder, which contains remaining frequencies. now, r1 is the input signal for the next sifting process. the second imf is c2. 1 2 2 r c r− = (18) 1n n n r c r − − = (19) this procedure is stopped when the residue is not satisfied with the imf criteria and is a monotonic function. the given eeg signal is represented as: 1 ( ) [ ( ) ( )] i i i i x t c t r t = = +∑ (20) emd decomposes the signals based on amplitude and frequency information. the emg artifacts cannot be separated from the eeg signals due to the overlapping at higher frequencies, but these kinds of emg artifacts can be removed using some standard forms of recent artifact removal techniques. 4.7. singular spectrum analysis (ssa) ssa is a powerful decomposition method for the removal of artifacts in single-channel and multi-channel eeg signals. the bss method is mainly applied to multi-channel data. in ssa, the procedure is primarily divided into two steps, and each step is further divided into two phases [51]. (1) decomposition: (i) time delay embedding; (ii) svd. (2) reconstruction: (i) grouping; (ii) diagonal averaging. 47 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 consider that there are n sampled signals x(t) = x1, x2, …, xn, then in the first embedding step, the x(t) is mapped as x whose size is k × p [52]. 1 2 2 3 1 1 2 1 , , , p p n l l n x x x x x x x x x x x x x + +         = =           ⋯ ⋯ ⋯ ⋮ ⋮ ⋱ ⋮ ⋯ (21) where l is the embedding size. the general value l = n/2. xi is a lagged vector. the second step is svd. by using svd, the trajectory matrix x is defined as: i i i i x vuλ= (22) where λi is eigenvalue; ui and vi are left and right singular eigenvectors. now the trajectory matrix x is written as: 1 l i i x x = = ∑ (23) as the indices in are divided into r groups. the matrix x after grouping is: 1 j r i j x x = = ∑ (24) in the last step of diagonal averaging, the sub-matrix ��� is translated to new series by the averaging of anti-diagonal elements. the final trajectory matrix is represented as: 1 ( ) ( ) l rc j j x t x t = = ∑ (25) the ssa method can remove muscle artifacts better compared to bss methods. by combing the ssa with other bss methods, artifacts can be removed further. 4.8. independent vector analysis (iva) the eeg data is contaminated by noises/artifacts while recording from the scalp. there are many methods for removing eog artifacts, e.g., ica, cca, emd, ssa, etc. the removal of muscle artifacts is difficult; only a few methods are there because muscle artifacts have high amplitude and variable topographical distribution, and overlap with the entire eeg spectrum [53]. ica separates the muscle artifacts effectively using hos by linearly un-mixing the eeg data into ics if and only if the artifacts are stereotyped. however, muscle artifacts are not stereotyped because these are generated from a group of muscles. cca extracts the sources maximally autocorrelated and mutually uncorrelated using sos. comparing with eeg signals, muscle artifacts have a low correlation. cca uses this property to remove the artifacts. cca and ica are applicable mainly for multi-channel data. kim et al. [54] have proposed an iva approach similar to ica, and have addressed the permutation issue when dealing with acoustic source separation. in recent times, this approach has been applied to eeg signals and could remove muscle artifacts successfully. iva operation which relies on “source extraction from different datasets and maximally dependent sources” within the dataset is maximally independent [53]. in this analysis, the source independence within the dataset and dependence between the datasets exist in parallel. in iva, the source component vector (scv) was defined from different datasets [54]. 48 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 the p th scv is �� = ��� , �� �,…,�� � (p = 1, 2, …, p), which is a random vector independent of all other scvs. �� � (m =1, 2, …, m) represents the p th source component in the m th dataset. the main intention of iva approach is to identify the independent scv from the multi-dimensional dataset, which is achieved by minimizing the mi among the scv estimates �̂�. 1 2 ˆ ˆ ˆ[ ; ; ...... ] iva p i i s s s≜ (26) where ���̂�� is the mi within the p th scv. initially, iva assumes that the data follow laplacian distribution (iva-l), which uses hos without considering linear dependence using sos. anderson et al. [53] proposed iva-g with the gaussian distribution model, which uses sos and does not consider linear dependence. recently, chen et al. [55] in 2017 implemented a novel iva algorithm with multivariate power exponential distribution (mpe), which uses both sos and hos to get both advantages like linear independence and correlation. the drawback of iva is computational complexity and simulation time. it is mainly applied in laboratory applications, not in online applications. 4.9. hybrid methods artifact removal in eeg signals is vital for analyzing brain conditions and diagnosing the patient if any abnormalities are there. researchers are trying some new techniques by combing more than one existing previously discussed method for better removal of artifacts. by combining one or more methods, the advantages of each technique can be used to develop a new modality for completely removing the artifacts. a hybrid method utilizes the combination of several filters to eliminate the artifacts. it can eliminate the significant artifacts from the eeg signals without adjusting the interpretation of the neural state in the real and simulated eeg data. it improves the previous versions of total variation denoising and emd variants. 4.9.1. discrete wavelet transform adaptive filtering method (dwt-afm) for removing the artifacts in eeg signals using ica, a reference signal is required. this problem can be resolved by using dwt. the reference eog signal can be generated from the given data by selecting low-frequency wavelet coefficients to remove oas [56]. unlike ica, a large amount of information is also not required for better removal of artifacts. however, the disadvantage of dwt is that it cannot remove the artifacts effectively if the spectrum of artifacts overlap with the original eeg signals [44]. the afm is also called adaptive noise canceler (anc). anc can solve this problem. the drawback of anc is that it requires a reference signal for the removal of artifacts. by combing dwt and anc, both spectral overlap and reference signal problems can be solved [33]. for oa removal, dwt is used for generating reference eog signal, and this reference signal is used as an input reference for anc. with this hybrid method, the removal of artifacts is better compared to using dwt and anc independently. 4.9.2. discrete wavelet transform blind source separation (dwt-bss) dwt-bss is one of the hybrid methods to remove the artifacts more effectively. the bss technique is generally applicable to multi-channel data. the disadvantages of both methods are discussed in previous sections. to overcome these disadvantages by combing these methods, the advantages in both methods can be combined. based on single-channel and multi-channel, one of these methods can be used as the first method to apply. first, dwt is applied to decompose the time series signal into wavelet coefficients for the single-channel technique, and these coefficients are applied to bss (ica or cca) [46]. for multi-channel data, dwt or bss can be used first, and the resultant signal is applied to the second method for the removal of artifacts. if ica and cca are applied first, the time domain signal is decomposed into ics and ccs, and by using dwt, these coefficients are further decomposed to better remove the artifacts from the eeg data [19]. 49 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 4.9.3. empirical mode decomposition blind source separation (emd-bss) bss techniques are used for decomposing multi-channel data into separate ics and ccs using ica and cca, respectively. these techniques are mainly helpful for laboratory experiments, where multi-channel data is available. however, for ambulatory or single-channel data analysis, different methods should be used, e.g., dwt, emd, ssa, etc. the performance can be increased in terms of artifact removal using a hybrid model combing multiple methods. dwt decomposes the data using predefined mother wavelets [57]. however, for nonlinear and non-stationary stochastic data like eeg signals, dwt is not sufficient. emd is a data-driven technique that effectively decomposes randomly varying data like single-channel eeg signals into imfs [10]. then, by using any bss like ica or cca, these imfs can be further decomposed to remove artifacts. for multi-channel data, emd can also apply channel-wise to decompose into imfs [43]. 4.9.4. singular spectrum analysis adaptive noise canceler (ssa-anc) ssa-anc is one of the hybrid methods for the removal of artifacts from single-channel and multi-channel eeg signals. it can be applied to portable devices, where multi-channel data collection is not possible because ica and cca are multi-channel data methods. in the anc method for eog artifact removal, a reference eog artifactual signal is required. in a hybrid method like ssa-anc, the reference signal for removing eog artifacts for anc is given by ssa, like dwt, based on eigenvalues. in dwt, last few low-frequency components are used for generating the eog reference signal [56]. in ssa, the reference signal is also generated from the lower eigenvalues, which are developed while applying svd for the trajectory matrix [58]. the performance is improved in terms of artifact removal compared with ssa and anc because the input reference eog signal is given for anc, which is generated from the output of ssa. 4.9.5. singular spectrum analysis blind source separation (ssa-bss) ica and cca are powerful bss methods that use sos and hos to remove oas and muscle artifacts from the recorded eeg signals [57]. bss methods are multi-channel techniques, and each one has its own advantages and disadvantages. bss methods are combined with ssa as a hybrid method to remove artifacts from the recorded eeg signals. ssa decomposes the single-channel or multi-channel eeg data, and the decomposed data is further processed by bss methods to remove artifacts. 4.9.6. empirical mode decomposition independent vector analysis (emd-iva) iva uses sos and hos for the removal of artifacts [55]. by combing with emd, iva removes the artifacts effectively compared with iva independently. emd method decomposes the single-channel signals into multiple imfs, and these imfs are applied to iva to remove artifacts effectively. to overcome the disadvantages in emd, noise-assisted emd is proposed, which is eemd. eemd is applied to single-channel, but memd is applied to multi-channel data also. by combining memd and iva, zou et al. [12] proposed a method memd-iva for removing the artifacts from multi-channel signals. 5. comparative analysis of the existing methods a tabulated survey of literature and methods are presented in table 4. no independent method can remove all the artifacts. the categorization of procedures is done based on the reference channel requirement. the automatic process can be applied online, whether single-channel or multi-channel, but prior information is required to improve the performance to remove artifacts. some methods are used for physiological artifact removal, and some are for non-physiological artifact removal. regression methods are elementary for the removal of artifacts but require reference channels and multi-channel data. filtering methods remove the artifacts by first estimating the artifactual components and subtracting them from the original eeg contaminated signals, but part of the eeg signals is also removed while subtracting. in kf, the first artifactual region is 50 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 identified, and by estimating noise, it will be subtracted from only the contaminated area [35]. bss methods are very effective and helpful for the removal of artifacts. most researchers use ica to remove artifacts but are restricted to stereotyped artifacts only [61]. muscle artifacts are not stereotyped and thus cannot be removed by ica effectively. cca can overcome this problem, but it increases computational cost and simulation time, making cca not being applied in online applications. ica and cca are multi-channel data processing methods that cannot be used for ambulatory analysis where few channels or single-channel data is available. table 4 comparative analysis of artifactual methods discussed in the survey method reference electrode required prior knowledge automatic online single-/ multi-channel ref. regression yes no yes no multi [33] af yes yes yes yes both [34] kf no yes yes yes both [35] pca no no no no multi [8] ica no no yes yes multi [2] cca no no yes yes multi [3] dwt no no no no both [44] emd no no no yes both [48] ssa no no yes no both [52] iva no no no no both [54] dwt-afm no no yes yes both [56] dwt-bss no no no no both [46] emd-bss no no no no both [10, 59] ssa-anc no no yes yes both [57] ssa-bss no no no no both [57] emd-iva no no no no both [55, 60] emd and ssa methods can be applied for both single-channel and multi-channel, but the disadvantage is mode mixing and aliasing in emd. ssa method removes only one type of artifacts. an extension of ica from 1-d data is iva technology, which can be used to remove both ocular and muscle artifacts as it takes advantage of both cca and ica methods. automatic methods can remove artifacts without manual intervention. regression methods can be used as intuitive if the reference channel is there [8]. bss methods are used as automatic if a subsequent procedure like svm is applied to verify artifactual components. individually each method has its advantages and disadvantages. researchers are combing more than one method like a hybrid method to get better results to use the benefits of both approaches. in hybrids methods, multi-channel strategies can be applied for single-channels by applying single-channel techniques first. the output is applied as the input for the multi-channel approach. eemd-ica method removes the eog artifacts, and eemd-cca removes emg artifacts from a single-channel, even though ica and cca are multi-channel methods [62]. eemd decomposes the single-channel data into multiple imfs. these imfs can be applied to ica or cca for further processing to remove artifacts. compared to wt-ica, eemd-ica removes artifacts effectively because eemd is a data-driven method. fmemd-cca is used to remove muscle artifacts from a few channel methods with less computational time [25]. wt-ica combined with svm was proposed to remove oas automatically. the recursive least square (rls) notch filter method is proposed to remove ecg artifacts because regression and filtering methods cannot delete the ecg artifacts [63]. the memd-iva method is proposed to remove the muscle artifacts from the acquisition channels, in which the inter-channel interference is also considered one of the noises [60]. some recommendations are suggested to remove muscle artifacts from a few channel data [64]. all the above methods remove the artifacts effectively by hybrid application. apart from these methods, researchers will propose one optimal method to remove all the artifacts in the future. 51 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 6. conclusions eeg signals are generated from the brain and recorded from the scalp using electrodes. different physiological and non-physiological artifacts contaminate these time-domain eeg signals. in this study, many methods are summarized from various author’s publications, but no specific method can eradicate the artifacts with high accuracy and efficiency. the advantages and disadvantages of each method are discussed in this study. some methods are used primarily to remove artifacts such as eog, emg, and ecg, but no specific method can remove all the artifacts. regression methods and filtering methods require artifactual reference signals for the removal of artifacts. bss methods remove the artifacts effectively, but they require multi-channel extensive data and are computationally complex. dwt, emd, and ssa methods are used for single-channel and multi-channel, but the efficiency of removing artifacts is not up to the level. by combining more than one method, i.e., the hybrid method, the efficiency is improved compared to individual methods. still, cost and complexity is increased, and this may not be suitable for online applications. in this aspect, researchers are finding an optimal solution for minimizing the artifacts in the eeg signals. in the future, with the existing artifact removal techniques, one specific method will be implemented by combing present trends, e.g., machine learning, artificial intelligence (ai), and genetic algorithm (ga), with increased accuracy and efficiency for the removal of artifacts. nomenclature eeg electroencephalogram max maximum absolute error bss blind source separation nrd normalized root mean square difference oa ocular artifacts prd percentage root mean square difference emg electromyography nmse normalized mean square error ecg electrocardiography psnr peak signal to noise ratio mmse modified multiscale sample entropy far false alarm rate iva independent vector analysis mev mean entropy value wt wavelet transform dsnr difference signal to noise ratio swt stationary wavelet transform af adaptive filtering dwt discrete wavelet transform eog electrooculogram st statistical threshold ekf extension kalman filter ut universal threshold ukf unscented kalman filter em eye movement pca principle component analysis ann artificial neural network pc principle component doa depth of anaesthesia svd singular value decomposition bis bispectral index cca canonical correlation analysis arte automated artifact handling in eeg anc adaptive noise canceler svm support vector machine cc correlation coefficient asr artifact subspace reconstruction mi mutual information eemd enhanced empirical mode decomposition sos second-order statistics sqi signal quality index hos higher-order statistics imf intrinsic mode function ai artificial intelligence memd multivariate empirical mode decomposition ga genetic algorithm fmemd fast multivariate empirical mode decomposition snr signal to noise ratio blimf band limited intrinsic mode function mse mean square error ssa singular spectrum analysis mae mean absolute error nnr neural network regressor re relative error mraf multi-resolutional analysis and adaptive filtering rmse root mean square error wpd wavelet packet decomposition rrmse relative root mean square error ica independent component analysis nrmse normalized root mean square error ic independent component psd power spectral density tuh temple university hospital sar signal to artifact ratio 52 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 conflicts of interest the authors declare no conflict of interest. statement of ethical approval (a) statement of human rights for this type of study, statement of human rights is not required. (b) statement on the welfare of animals for this type of study, statement on the welfare of animals is not required. statement of informed consent for this type of study, informed consent is not required. references [1] h. hu, s. guo, r. liu, and p. wang, “an adaptive singular spectrum analysis method for extracting brain rhythms of electroencephalography,” peerj, vol. 5, e3474, june 2017. [2] a. k. abdullah, c. z. zhang, a. a. a. abdullah, and s. lian, “automatic extraction system for common artifacts in eeg signals based on evolutionary stone’s bss algorithm,” mathematical problems in engineering, vol. 2014, 324750, august 2014. [3] m. m. n. mannan, m. a. kamran, and m. y. jeong, “identification and removal of physiological artifacts from electroencephalogram signals: a review,” ieee access, vol. 6, pp. 30630-30652, 2018. [4] r. zafar, a. qayyum, and w. mumtaz, “automatic eye blink artifact removal for eeg based on a sparse coding technique for assessing major mental disorders,” journal of integrative neuroscience, vol. 18, no. 3, pp. 217-229, 2019. [5] a. morley, l. hill, and a. g. kaditis, “10-20 system eeg placement,” https://www.ers-education.org/lrmedia/2016/pdf/298830.pdf, march 15, 2016. [6] s. k. goh, h. a. abbass, k. c. tan, a. al-mamun, c. wang, and c. guan, “automatic eeg artifact removal techniques by detecting influential independent components,” ieee transactions on emerging topics in computational intelligence, vol. 1, no. 4, pp. 270-279, august 2017. [7] s. r. sreeja, r. r. sahay, d. samanta, and p. mitra, “removal of eye blink artifacts from eeg signals using sparsity,” ieee journal of biomedical and health informatics, vol. 22, no. 5, pp. 1362-1372, september 2018. [8] x. jiang, g. b. bian, and z. tian, “removal of artifacts from eeg signals: a review,” sensors, vol. 19, no. 5, 987, march 2019. [9] s. s. menon and k. krishnamurthy, “a study of brain neuronal and functional complexities estimated using multiscale entropy in healthy young adults,” entropy, vol. 21, no. 10, 995, october 2019. [10] k. t. sweeney, s. f. mcloone, and t. e. ward, “the use of ensemble empirical mode decomposition with canonical correlation analysis as a novel artifact removal technique,” ieee transactions on biomedical engineering, vol. 60, no. 1, pp. 97-105, january 2013. [11] m. chavez, f. grosselin, a. bussalb, f. d. v. fallani, and x. navarro-sune, “surrogate-based artifact removal from single-channel eeg,” ieee transactions on neural systems and rehabilitation engineering, vol. 26, no. 3, pp. 540-550, march 2018. [12] y. zou, v. nathan, and r. jafari, “automatic identification of artifact-related independent components for artifact removal in eeg recordings,” ieee journal of biomedical and health informatics, vol. 20. no. 1, pp. 73-81, january 2016. [13] r. mahajan and b. i. morshed, “unsupervised eye blink artifact denoising of eeg data with modified multiscale sample entropy, kurtosis, and wavelet-ica,” ieee journal of biomedical and health informatics, vol. 19, no. 1, pp. 158-165, january 2015. 53 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 [14] p. p. acharjee, r. phlypo, l. wu, v. d. calhoun, and t. adalı, “independent vector analysis for gradient artifact removal in concurrent eeg-fmri data,” ieee transactions on biomedical engineering, vol. 62, no. 7, pp. 1750-1758, july 2015. [15] s. khatun, r. mahajan, and b. i. morshed, “comparative study of wavelet-based unsupervised ocular artifact removal techniques for single-channel eeg data,” ieee journal of translational engineering in health and medicine, vol. 4, pp. 1-8, march 2016. [16] m. n. tibdewal, r. r. fate, m. mahadevappa, a. k. ray, and m. malokar, “classification of artifactual eeg signal and detection of multiple eye movement artifact zones using novel time-amplitude algorithm,” signal, image, and video processing, vol. 11, no. 2, pp. 333-340, february 2017. [17] q. liu, y. f. chen, s. z. fan, m. f. abbod, and j. s. shieh, “eeg artifacts reduction by multivariate empirical mode decomposition and multiscale entropy for monitoring depth of anaesthesia during surgery,” medical and biological engineering and computing, vol. 55, no. 8, pp. 1435-1450, august 2017. [18] s. barua, m. u. ahmed, c. ahlstrom, s. begum, and p. funk, “automated eeg artifact handling with application in driver monitoring,” ieee journal of biomedical and health informatics, vol. 22, no. 5, pp. 1350-1361, september 2018. [19] c. y. sai, n. mokhtar, h. arof, p. cumming, and m. iwahashi, “automated classification and removal of eeg artifacts with svm and wavelet-ica,” ieee journal of biomedical and health informatics, vol. 22, no. 3, pp. 664-670, may 2018. [20] j. li, y. chen, f. taya, j. lim, k. wong, y. sun, et al., “a unified canonical correlation analysis-based framework for removing gradient artifact in concurrent eeg/fmri recording and motion artifact in walking recording from eeg signal,” medical and biological engineering and computing, vol. 55, no. 9, pp. 1669-1681, september 2017. [21] v. roy and s. shukla, “effective eeg motion artifacts elimination based on comparative interpolation analysis,” wireless personal communications, vol. 97, no. 4, pp. 6441-6451, december 2017. [22] c. y. chang, s. h. hsu, l. pion-tonachini, and t. p. jung, “evaluation of artifact subspace reconstruction for automatic artifact components removal in multi-channel eeg recordings,” ieee transactions on biomedical engineering, vol. 67, no. 4, pp. 1114-1121, april 2020. [23] j. cheng, l. li, c. li, y. liu, a. liu, r. qian, et al., “remove diverse artifacts simultaneously from a single-channel eeg based on ssa and ica: a semi-simulated study,” ieee access, vol. 7, pp. 60276-60289, 2019. [24] m. saini and u. satija, “an effective and robust framework for ocular artifact removal from single-channel eeg signal based on variational mode decomposition,” ieee sensors journal, vol. 20, no. 1, pp. 369-376, january 2020. [25] y. liu, y. zhou, x. lang, y. liu, q. zheng, y. zhang, et al., “an efficient and robust muscle artifact removal method for few-channel eeg,” ieee access, vol. 7, pp. 176036-176050, 2019. [26] c. dora and p. k. biswal, “correlation-based ecg artifact correction from single channel eeg using modified variational mode decomposition,” computer methods and programs in biomedicine, vol. 183, 105092, january 2020. [27] c. dora, r. n. patro, s. k. rout, p. k. biswal, and b. biswal, “adaptive ssa based muscle artifact removal from single channel eeg using neural network regressor,” innovation and research in biomedical engineering, in press. [28] d. p. yedurkar and s. p. metkar, “multiresolution approach for artifacts removal and localization of seizure onset zone in epileptic eeg signal,” biomedical signal processing and control, vol. 57, 101794, march 2020. [29] m. s. islam, a. m. el-hajj, h. alawieh, z. dawy, n. abbas, and j. el-imad, “eeg mobility artifact removal for ambulatory epileptic seizure prediction applications,” biomedical signal processing and control, vol. 55, 101638, january 2020. [30] n. bajaj, j. r. carrión, f. bellotti, r. berta, and a. de gloria, “automatic and tunable algorithm for eeg artifact removal using wavelet decomposition with applications in predictive modeling during auditory tasks,” biomedical signal processing and control, vol. 55, 101624, january 2020. [31] s. k. noorbasha and g. f. sudha, “removal of eog artifacts and separation of different cerebral activity components from single channel eeg—an efficient approach combining ssa-ica with wavelet thresholding for bci applications,” biomedical signal processing and control, vol. 63, 102168, january 2021. [32] m. a. klados, c. papadelis, c. d. lithari, and p. d. bamidis, “the removal of ocular artifacts from eeg signals: a comparison of performances for different methods,” 4th european conference of the international federation for medical and biological engineering, pp. 1259-1263, november 2008. [33] y. chen, q. zhao, b. hu, j. li, h. jiang, w. lin, et al., “a method of removing ocular artifacts from eeg using discrete wavelet transform and kalman filtering,” ieee international conference on bioinformatics and biomedicine, pp. 15-18, december 2016. [34] a. zhang and w. li, “adaptive noise cancellation for removing cardiac and respiratory artifacts from eeg recordings,” 5th world congress on intelligent control and automation, pp. 15-19, june 2004. 54 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 [35] h. shahabi, s. moghimi, and h. zamiri-jafarian, “eeg eye blink artifact removal by eog modeling and kalman filter,” 5th international conference on biomedical engineering and informatics, pp. 356-367, october 2012. [36] a. mur, r. dormido, and n. duro, “an unsupervised method for artefact removal in eeg signals,” sensors, vol. 19, no. 10, 2302, may 2019. [37] b. mijović, m. de vos, i. gligorijević, j. taelman, and s. van huffel, “source separation from single-channel recordings by combining empirical-mode decomposition and independent component analysis,” ieee transactions on biomedical engineering, vol. 57, no. 9, pp. 2188-2196, september 2010. [38] v. roy and s. shukla, “a survey on artifacts detection techniques for electro-encephalography (eeg) signals,” international journal of multimedia and ubiquitous engineering, vol. 10, no. 3, pp. 425-442, 2015. [39] g. wang, c. teng, k. li, z. zhang, and x. yan, “the removal of eog artifacts from eeg signals using independent component analysis and multivariate empirical mode decomposition,” ieee journal of biomedical and health informatics, vol. 20, no. 5, pp. 1301-1308, september 2016. [40] x. chen, a. liu, h. peng, and r. k. ward, “a preliminary study of muscular artifact cancellation in single-channel eeg,” sensors, vol. 14, no. 10, pp. 18370-18389, october 2014. [41] d. r. hardoon, s. szedmak, and j. shawe-taylor, “canonical correlation analysis: an overview with application to learning methods,” neural computation, vol. 16, no. 12, pp. 2639-2664, december 2004. [42] w. de clercq, a. vergult, b. vanrumste, w. van paesschen, and s. van huffel, “canonical correlation analysis applied to remove muscle artifacts from the electroencephalogram,” ieee transactions on biomedical engineering, vol. 53, no. 12, pp. 2583-2587, 2006. [43] x. chen, x. xu, a. liu, m. j. mckeown, and z. j. wang, “the use of multivariate emd and cca for denoising muscle artifacts from few-channel eeg recordings,” ieee transactions on instrumentation and measurement, vol. 67, no. 2, pp. 359-370, february 2018. [44] s. khatun, r. mahajan, and b. i. morshed, “comparative study of wavelet-based unsupervised ocular artifact removal techniques for single channel eeg data,” ieee journal of translational engineering in health and medicine, vol. 4, 2000108, 2016. [45] p. gajbhiye, r. k. tripathy, a. bhattacharyya, and r. b. pachori, “novel approaches for the removal of motion artifact from eeg recordings,” ieee sensors journal, vol. 19, no. 22, pp. 10600-10608, november 2019. [46] b. s. raghavendra and d. n. dutt, “wavelet enhanced cca for minimization of ocular and muscle artifacts in eeg,” world academy of science, engineering, and technology, vol. 5, no. 9, pp. 419-424, january 2011. [47] e. karatoprak and s. seker, “an improved empirical mode decomposition method using variable window median filter for early fault detection in electric motors,” mathematical problems in engineering, vol. 2019, 8015295, 2019. [48] b. yang, t. zhang, y. zhang, w. liu, j. wang, and k. duan, “removal of electrooculogram artifacts from electroencephalogram using canonical correlation analysis with ensemble empirical mode decomposition,” cognitive computation, vol. 9, no. 5, pp. 626-633, october 2017. [49] n. e. huang and z. wu, “a review on hilbert-huang transform: method and its applications to geophysical studies,” reviews of geophysics, vol. 46, no. 2, rg2006, june 2008. [50] n. e. huang, m. l. c. wu, s. r. long, s. s. shen, w. qu, p. gloersen, et al., “a confidence limit for the empirical mode decomposition and hilbert spectral analysis,” proceedings of the royal society a: mathematical, physical, and engineering sciences, vol. 459, no. 2037. pp. 2317-2345, september 2003. [51] s. xu, h. hu, l. ji, and p. wang, “embedding dimension selection for adaptive singular spectrum analysis of eeg signal,” sensors, vol. 18, no. 3, 697, march 2018. [52] q. liu, a. liu, x. zhang, x. chen, r. qian, and x. chen, “removal of emg artifacts from multichannel eeg signals using combined singular spectrum analysis and canonical correlation analysis,” journal of healthcare engineering, vol. 2019, 4159676, 2019. [53] m. anderson, t. adali, and x. l. li, “joint blind source separation with multivariate gaussian model: algorithms and performance analysis,” ieee transaction on signal process, vol. 60, no. 4, pp. 1672-1683, april 2012. [54] t. kim, t. eltoft, and t. w. lee, “independent vector analysis: an extension of ica to multivariate components,” international conference on independent component analysis and signal separation, march 2006, pp. 165-172. [55] x. chen, a. liu, q. chen, y. liu, l. zuo, and m. j. mckeown, “simultaneous ocular and muscle artifact removal from eeg data by exploiting diverse statistics,” computers in biology and medicine, vol. 88, pp. 1-10, september 2017. [56] h. peng, b. hu, q. shi, m. ratcliffe, q. zhao, y. qi, et al., “removal of ocular artifacts in eeg—an improved approach combining dwt and anc for portable applications,” ieee journal of biomedical and health informatics, vol. 17, no. 3, pp. 600-607, may 2013. 55 proceedings of engineering and technology innovation, vol. 20, 2022, pp. 35-56 [57] l. zou, x. chen, g. dang, y. guo, and z. j. wang, “removing muscle artifacts from eeg data via underdetermined joint blind source separation: a simulation study,” ieee transactions on circuits and systems ii: express briefs, vol. 67, no. 1, pp. 187-191, january 2020. [58] a. k. maddirala and r. a. shaik, “removal of eog artifacts from single channel eeg signals using combined singular spectrum analysis and adaptive noise canceler,” ieee sensors journal, vol. 16, no. 23, pp. 8279-8287, december 2019. [59] x. chen, a. liu, j. chiang, z. j. wang, m. j. mckeown, and r. k. ward, “removing muscle artifacts from eeg data: multi-channel or single-channel techniques,” ieee sensors journal, vol. 16, no. 7, pp. 1986-1997, april 2016. [60] x. xu, x. chen, and y. zhang, “removal of muscle artefacts from few-channel eeg recordings based on multivariate empirical mode decomposition and independent vector analysis,” electronics letters, vol. 54, no. 14, pp. 866-868, july 2018. [61] x. chen, q. liu, w. tao, l. li, s. lee, a. liu, et al., “remae: user-friendly toolbox for removing muscle artifacts from eeg,” ieee transactions on instrumentation and measurement, vol. 69, no. 5, pp. 2105-2119, may 2020. [62] x. chen, q. chen, y. zhang, and z. j. wang, “a novel eemd-cca approach to removing muscle artifacts for pervasive eeg,” ieee sensors journal, vol. 19, no. 19. pp. 8420-8431, october 2019. [63] c. dai, j. wang, j. xie, w. li, y. gong, and y. li, “removal of ecg artifacts from eeg using an effective recursive least square notch filter,” ieee access, vol. 7, pp. 158872-158880, october 2019. [64] x. chen, x. xu, a. liu, s. lee, x. chen, x. zhang, et al., “removal of muscle artifacts from the eeg: a review and recommendations,” ieee sensors journal, vol. 19, no. 14, pp. 5353-5368, july 2019. copyright© by the authors. licensee taeti, taiwan. this article is an open access article distributed under the terms and conditions of the creative commons attribution (cc by-nc) license (https://creativecommons.org/licenses/by-nc/4.0/). 56  proceedings of engineering and technology innovation , vol. 2, 2016, pp. 06 07 6 the sensor collaboration structure of industrial 4.0 hao-ting lin * department of mechanical and computer-aided engineering, feng chia university, taichung, taiwan. received 02 february 2016; received in revised form 15 march 2016; accept ed 05 april 2016 abstract as informat ion and technology grow up globally, the business should look for the opt ima l productivity effic iency and add smart manage ment to ma intain the co mpany’s inte rests and value. in order to achieve the smart productivity, vast businesses pay lots of attention to invest in industrial 4.0 and change their business model. in this research, the sensor collaboration structure will be proposed and established to connect the information base and thing base. human can ma ke decisions and communicate between the physical domain and the cyber domain. in the physical doma in, the sensors can gather the data from the machine, the production line, the factory and the product. by sensor collaboration, the signals and info rmat ion can transfer quickly and simu ltaneously in the cyber domain. through the networking between information base and thing base, dec ision make rs can easily predict, judge and handle the tasks. as a result, in formation base, thing base and decision makers can beco me a triangle system to achieve the vertical integration and level contact. ke ywor ds : industrial 4.0, sensor collaboration, physical domain, cyber domain 1. introduction nowadays, the pursuing goal for the manufacturing industry is to establish the smart fac tories. the first industrial revolution is to exe cute machines instead of human to conduct large-scale p roduction. the second industrial revolution is to add the concept of assembly line for improving production effic iency. in the third industrial revolution, with ele ments of electronics and informat ion technology, the production costs are declined. with the development of globalization, the factory must be supplemented with intelligent manage ment to facilitate more rapid changes in the industry pulsation. co mbining with the advantages of three previous industrial revolution, the industry 4.0 is coming to search for the highest effic iency of production resources and to achieve the purpose of the intelligent production. in this paper, the sensor collaboration structure will be proposed for the smart factory. the concept of industrial 4.0 has been popular in eu rope, especially in ge rman manufacturer industries. smart mach inery can contin uously share inventories, problems info rmation and demands through the internet, so the production process and lead time can be fle xib ly adjusted. cyber-physical production systems (cppss) can not only connect machines with each other, but also create the whole values of the stages in the product life cycle. sensors and control ele ments make the machines, factories, transportation, internet and humans together [1]. fig. 1 shows the history of the industry revolution. fig. 1 history of the industry revolution 2. the architecture of industrial 4.0 the architecture of industrial 4.0 which has the vertical integration of the s mart production system, the horizontal integration of the ne tworks, through engineering interaction, and exponential technologies. 2.1. the vertical integration of the smart production system cppss ma ke the factories vary their inventories and demands fle xibly. smart factories *corresponding aut hor. email: haotlin@fcu.edu.t w proceedings of engineering and technology innovation , vol. 2, 2016, pp. 06 07 7 copyright © taeti establish customized and personal production through data integration and smart sensing techniques. cppss a lso pro mote the automation manage ment, ma intenance manage ment and resource connection. 2.2. the horizontal integration of the network s the horizontal integration which is an o ptimized real-time networks can create the higher value and mo re fle xib le for proble m solving. the internet based production system can provide the integrated platform for logistics, inventories, production, marketing and sales b etween in-house and outsourcing. 2.3. through engineering interaction industrial 4.0 has a cross -disciplinary through engineering to connect overall supply chain from products to customers. data and information can be obtained in each stage of the product life cycle to establish the model fro m the prototype to the end product. 2.4. exponential technologies artificia l intelligence (ai), advanced robot and sensing technology can make the production personalized and fle xible. ai can not only an alyze big data, design the more fle xible path for the automated guided vehicle (a gv) and save costs of the supply chain management (scm), but also enhance the cooperation between h uman and machinery. in addition, the 3d p rinting provide the new solution for the production mode to reduce inventories and lead time for the firm. 3. conclusions in this paper, the proposed architecture of industrial 4.0 has the vertical integration, the horizontal integration, through engineering and e xponential technologies. the sensor collaboration for industrial 4.0 can collect and interact with hu man and machinery. managers can co rrectly ma ke a decision by the cyber-physical system. acknowledgement this research was sponsored by ministry of science and technology, taiwan under the grant most 104-2218-e-035 -017. references [1] x. yue, h. cai, h. yan, c. zou, and k. zhou, “cloud-assisted industrial cyber-physical systems: an insight,” microprocessors and microsystems , vol. 39, no. 8, pp. 1262-1270, 2015.  proceedings of engineering and technology innovation, vol. 2, 2016, pp. 01 03 1 optical design of direct projection bicycle light hsuan-jung tsai * , tsang-yen hsieh, jyh-liang wang department of electronic engineering, ming chi university of technology, new taipei city, taiwan. received 01 february 2016; received in revised form 15 march 2016; accepted 03 april 2016 abstract optical lighting systems usually demand specific illuminance distributions to comply with regulatory requirements for various applications. this study demonstrates an optical design of bicycle light integrated with a special fresnel lens and a reflector. to enhance the light intensity, the proposed bicycle light adopts direct projection rather than reflection. the fresnel lens is used to concentrate light and achieve the specific distribution of illuminance. the reflector provides a parabola reflective surface and reduces the divergence of light to greatly cast out parallel light, which improves the illuminance. for optimum case, the simulation is based on the total led power of 180 lm, and the optical power is analyzed as 85 lm with a distance of 10 m away from the bicycle light. the illuminance distribution of bicycle light and cut-off line can meet the requirements of german stvzo regulations. keywords: fresnel lens, bicycle light, reflector, direct projection, stvzo 1. introduction bicycle light provide people with safe riding at night. it requests enough brightness on the road and should avoid the light radiation to the people’s eye. therefore, the optical design with cut-off line is important and indeed necessary. german road traffic licensing regulations (germany stvzo regulations) clearly define the use of bicycle safe and bicycle lighting requirements as following [1] (1) at least 20 lux in hv, the maximum illuminance emax shall be found at this point. if emax is found lateral to hv, its value shall not exceed 1.2 times the value of hv. (2) within the range delimited by an angle of 4° to both sides starting from hv up to and including point l1 respectively r1 as well as 1.5° below hv up to and including point 2: at least half of the maximum illuminance emax. (3) up to 5° below hv and between point 2 and point 3 including the latter: at least 2.5 lux and from there 4° to the left and to the right up to and including the points l5 respectively r5: at least 2.0 lux. (4) in the horizontal plane 3.4° above hv and beyond (zone 1): not more than 2.0 lux. optical design of bicycle light can usually classified as reflective mode and direct mode. reflective mode requires a reflective apparatus. direct mode uses the convex lens and forms optical imaging or non-imaging depending on the application. in this study, a direct projection bicycle light integrated with a reflector is considered and investigated due to the better light efficiency. fig. 1 the illustration of germany stvzo regulations [1] 2. method the requirements and constraints of optical design are based on the germany stvzo regulations. the regulations require power consumption of 6.2w at 12v. the luminous flux of light source embedded in headlights should exceed a value of 42 lm at test voltage. the required illuminance must correspond to the guideline shown in fig. 1. *corresponding author. email: pot.martian@gmail.com proceedings of engineering and technology innovation, vol. 2, 2016, pp. 01 03 2 copyright © taeti fig. 2 the fresnel lens designed with inward grooves facing [3] a typical imaging fresnel lens with inward grooves facing has an analytical solution indicated by tver’yanovich [2]. the structure of fresnel lens will be performed and based on the creative design methodology process. in accordance with fig. 2, three equations can be set up to describe the lens. the prism angle α is the goal of a simulation written as [2-3], where n is a material medium, ω is divergence angle of light, α is light emission angle, β is the sum of α and ω, r is lens radius, and f is lens focal length. 𝑛 𝑠𝑖𝑛 𝛼 = 𝑠𝑖𝑛 𝛽 (1) 𝑡𝑎𝑛 𝜔 = 𝑅 𝑓 (2) 𝛽 = 𝛼 + 𝜔 (3) 3. simulation and result the optical characteristics of bicycle light are designed to fit german stvzo regulations. 3d modeling and components of the proposed bicycle light, which is equipped a commercial led lamp (cree xlamp xp-g2) as the light source. a specific fresnel lens is adopted in the optical design of bicycle light. the fresnel lens combines two focal points together to concentrate light and presents a rectangle distribution of illuminance, shown in fig. 3. fig. 3 (a) 3d cad of fresnel lens, and (b) the simulated illuminance distribution of the led lamp with the specific fresnel lens the simulated illuminance distribution of the led lamp with the specific fresnel lens only provide central illuminance with rectangle shape, and does not extend the illuminance distribution to reach the requirements of l1 and r1. thus, the optical design of bicycle light combined with fresnel lens cannot fit german stvzo regulations. moreover, a reflector is proposed to improve the optical design. the reflector provides a parabola reflective surface and reduces the divergence of light to greatly cast out parallel light, revealed in fig. 4. it evidently indicates that the simulated illuminance distribution can be coherent with the real illuminance of projection. fig. 4 (a) simulated illuminance distribution of the led lamp with a reflector, (b) the real illuminance of projection with a distance of 1 meter, and (c) the reflector entity fig. 5 schematic mechanisms of direct projection bicycle light fig. 5 gives the final design of direct projection bicycle light. the design combines the specific fresnel lens and reflector. the combination demands some modulations of the physical dimensions to achieve the optimum simulation. by the trial of optical simulation to fine tune the design of bicycle light, an optimum illuminance proceedings of engineering and technology innovation, vol. 2, 2016, pp. 01 03 3 copyright © taeti distribution can be made. fig. 6 demonstrates the simulation result of the optimum design. the simulated illuminance distribution is similar as an extended rectangle and in compliance with the requirements of germany stvzo regulations. fig. 6 simulated illuminance distribution of the proposed direct projection bicycle light 4. conclusions a direct projection bicycle light is proposed in this research. the optical design of bicycle light is integrated with a special fresnel lens and a reflector. the fresnel lens combines two focal points to concentrate light and achieve the specific distribution of illuminance. the reflector provides a parabola reflective surface and reduces the divergence of light. the simulated illuminance distribution of bicycle light and cut-off line can meet the requirements of german stvzo regulations. references [1] road traffic licensing regulations, germany: government printing office, 2012. [2] e. v. tver’yanovich, “profiles of solar engineering fresnel lenses,” gelioteknika, vol. 19, no. 6, pp. 31-34, 1984. translated into english in applied solar energy, vol. 19, no. 6, pp. 36-39, 1984. [3] r. leutz and n. suzuki, nonimaging fresnel lenses, paris: springer, may 2001. cut-off line liline microsoft word 7-#990-typesetting finish 列印pdf.doc proceedings of engineering and technology innovation, vol. 7, 2017, pp. 37 40 traffic safety evaluation based on vision and signal timing data won-ho suh*, jin-woo park, jung-ryul kim department of transportation and logistics engineering, hanyang university, ansan, korea. received 18 july 2017; received in revised form 11 august 2017; accepted 30 august 2017 abstract with the advancement of recent image processing technologies numerous imaging applications have been developed to collect traffic information. in this study, image processing technology is applied to roadway safety to detect and measure the conflict between pedestrians and vehicular traffic and investigate the correlations between the conflict and signal timing data. by using an image processing technique, pedestrian and vehicular movements are detected and the distance between these two are calculated to detect conflicts between them. in this study, machine learning techniques are utilized to facilitate more accurate and efficient detection. it is assumed that the intersection is accident-prone if conflicts between pedestrian and vehicular traffic appear more frequently than other intersections and correlations exist depending on the status of signal timing, for example at the end of green signal for a particular movement. this approach is expected to identify accident-prone intersections without actually experiencing crashes; therefore, this would potentially reduce social costs associated with traffic accidents. keywords: image processing, vehicle detecting, traffic safety, pedestrian safety 1. introduction pedestrians are vulnerable road users. when they are involved in collisions with vehicles, there are considerably higher chances of being severely or fatally injured, while a large number of collisions between vehicles are property damage only. surrogate safety measures has been developed as a complementary method to improve pedestrian safety and offer more in depth analysis than relying on historic accident data alone. there is a growing effort to use recent image processing technologies for a conflict analysis between vehicles and pedestrians. in this study, vehicle tracking and pedestrian tracking technologies currently developed and used are reviewed. 2. vehicle tracking model-based tracking utilizes a priori knowledge of typical objects. for a vehicle tracking, vehicles in an image captured from a video are tracked. these methods recognize vehicles. however, this method tends to fail when there is an occlusion for the target object. also, this tracking method is limited in recognizing all the vehicle types on the roadway. region-based tracking method is to identify connected regions of the image associated with each target vehicle. regions are usually acquired from background subtraction. also, this method utilizes kalman filters to process available information from the image including motion, size, color, shape, texture, and centroid. it is known that this tracking method is * corresponding author. e-mail address: wonhosuh@hanyang.ac.kr tel.: +82-31-400-5154; fax: +82-31-436-8147 proceedings of engineering and technology innovation, vol. 7, 2017, pp. 37 40 copyright © taeti 38 computationally efficient and the detection accuracy is generally good in free-flowing traffic conditions. however, if traffic is congested – when background images do not change over a short period of time and vehicles partially occlude one another, this method tends to have problems, since separating individual vehicles are becoming challenging. sometimes, such vehicles are identified as one vehicle, resulting in low detection rates. contour-based tracking method uses the boundary curves of the moving object. for example, this method tracks pedestrians by identifying the contour of a pedestrian’s head. this method is known to have more efficient description of objects than region-based methods. however, this method still shares the weakness of region-based tracking when traffic condition is congested. feature-based tracking does not track object as a whole in image processing. instead, this method tracks unique features. there features can be points or lines of the object which are distinct from other objects. when the target object is partially occluded, some of the distinctive features of the moving object remain visible and detectable, resulting in successful tracking. the advantage of this method over other tracking methods is that the same algorithm can be applied for tracking in different lighting or traffic conditions. for example, this method utilizes window corners or bumper edges during the day when the entire object is visible. during the night, this method uses tail lights, since other features are not detectable in night time lighting conditions. currently, this method is advanced and computational efficient allowing this method to be used in real time processing. an example of vehicle tracking is provided in fig. 1. fig. 1 vehicle tracking [1] 3. pedestrian tracking to identify conflicts between vehicles and pedestrians, all movements of vehicular and pedestrian traffic need to be tracked from video frames. especially, this is challenging since pedestrians sometimes move in group and change directions, while vehicular movements are somewhat predictable. to overcome the issues and increase the accuracy of tracking, there are three tracking methods which are currently popular; tracking by detection, tracking using flow, and tracking with probability. tracking by detection method detects objects using background modeling and subtraction with the current image or deformable templates. tracking using flow is similar to feature based tracking. this method is popular in traffic monitoring and pedestrian counting. tracking with probability uses a bayesian tracking framework. this method also utilizes other filers and methods, such as particle filters, markov chain, and monte carlo methods. an example of pedestrian tracking is provided in fig. 2. also, fig. 3 presents an example of vehicle and pedestrian tracking in an intersection. proceedings of engineering and technology innovation, vol. 7, 2017, pp. 37 40 copyright © taeti 39 fig. 2 pedestrian tracking [1] fig. 3 vehicle and pedestrian tracking [2] 4. calculating conflicts to calculate conflicts between vehicular and pedestrian traffic, potential accident-prone conditions need to be identified. these conditions are illustrated in fig. 4 and fig. 5. based on the calculated time gap between the vehicle and pedestrian, conflict data can be obtained. fig. 4 left turning vehicle conflict example [3] fig. 5 thru moving vehicle conflict example [3] 5. preliminary results and future works the research team is in a process of developing vehicle and pedestrian tracking method. the team was successfully tracked vehicles in a video collected from an intersection in seoul, korea (fig. 6). the team is planning to continue to develop the tracking algorithm and calculate the conflicts between vehicular and pedestrian traffic. the team is utilizing machine learning techniques to reduce the detection error and increase the efficiency of the program. the output of this research is proceedings of engineering and technology innovation, vol. 7, 2017, pp. 37 40 copyright © taeti 40 expected to improve traffic analysis [4] by detecting pedestrian and vehicular movements and facilitate more accurate and efficient detection. it is assumed that the intersection is accident-prone if conflicts between pedestrian and vehicular traffic appear more frequently than other intersections and correlations exist depending on the status of signal timing, for example at the end of green signal for a particular movement. this approach is expected to identify accident-prone intersections without actually experiencing crashes; therefore, this would potentially reduce social costs associated with traffic accidents. fig. 6 vehicle tracking program snapshot acknowledgment this work was sponsored by nrf-2017r1d1a1a09000606 and transportation & logistics research program id-97344 of the korean government. references [1] k. ismail, t. sayed, and n. saunier, “automated analysis of pedestrian-vehicle: conflicts context for before-and-after studies,” transportation research record: journal of the transportation research board, vol. 2198, pp. 52-64, 2010. [2] a. laureshyn, “superpixel based road user tracker,” workshop on the comparison of surrogate measures of safety extracted from video data, transportation research board 93rd annual meeting, january 2014. [3] t. madsen, “automatic detection of conflicts at signalized intersections,” workshop on the comparison of surrogate measures of safety extracted from video data, transportation research board 93rd annual meeting, january 2014. [4] w. suh, “mitigating initialization bias in transportation modeling applications,” proceedings of engineering and technology innovation, vol. 3 pp. 34-36, 2016.  proceedings of engineering and technology innovation , vol. 2, 2016, pp. 11 13 11 fire resistant study for long span beams made of fr 490 steels in-kyu kwon * department of fire protection engineering, kangwon national university, samcheok, korea. received 22 february 2016; received in revised form 02 april 2016; accept ed 10 may 2016 abstract building could be severely influenced by fire and it can be demo lished so that the ma jor structural ele ments such as columns and beams required to be coated by fire protection materials according to each condition. especially, a steel fra med building regarded as a weaken structures when the build ing is engulfed with a severe fire condition. therefore, the fire resistant steel (fr 490) has been developed to improve the stru ctural properties of an ordinary structural steels at high temperature. generally, the fire resistance of structural beam built with a structural steel can be measured by horizontal furnace using a standard testing method by each nation’s standard. ho wever, the real length of the bea m showed the tendency longer than that of the horizontal furnace. therefore, it is considered more important to know the rea l fire resistant performance of structural bea ms. in this paper, to evaluate the fire resistant performance of long span beams made of the fr 490, the fire engineering method is done us ing not only materials properties at high temperatures but a heat transfer, a stress theory. the results showed that as the length of beams are longer, the ma ximu m load bearing capacities at high temperatures are lower and the deflections are increased. ke ywor ds : fr 490, long span beam, fire re sistant performance, structural stability at high temperature 1. introduction when steel fra med buildings are covered with a severe fire, the structural stabilit ies can be damaged and this can cause a progressive co llapse. therefore, a ll most of countries require the prima ry structural elements such as columns, beams, trusses the fire protection according to building situation. in ko rea, the fire protection depends on the buildings height and use. the way to satisfy the fire protection is divided into two parts. one is to obey the building regulation and standard, the other is to design with an e ngineering technique. the last one is known to an advanced engineering des ign or a performance based fire design. the na me can be used differently by nation or by person who uses it. the basic concept of the advanced engineering d esign consists of a fire scenario and a structural analysis using mechanical properties and heat transfer theory. until now, the advanced fire engineering method is known to practice to most of countries but not my country. at now, my country is trying to accept the perfo rmance-based engineering method in building industry and this paper is one of proces s to do it. in general, the structural steel is known as a weaken materia l when it e xposed to a fire. be cause the atomic structure is inclined to recede away as the temperature of it is getting high. therefore, to protect the normal strength of the structural steel at high temperature situation, a proved thickness of passive fire protective ma terials should be applied. a fr 490 had developed to lessen the tendency to reduce mechanical properties such as yield strength and an elastic modulus at high temperature rather than those of sm 490. the fr 490 has the same mechanica l property as the sm 490 has. the purpose of this study is to evaluate the structural stability of long span beams made of the fr 490 at high temperature using mechanica l propert ies and heat transfer and heat stress analysis. 2. method the structural stability of long span beam can be measured by vertical furnace in terms of deflection with loading. but the vert ical furnace is fixed with a ma ximu m d imension, 4 meters in length and 3 meters in width. but the size of the *corresponding aut hor. email: kwonik@kangwon.ac.kr proceedings of engineering and technology innovation , vol. 2, 2016, pp. 11 13 12 copyright © taeti vertical furnace is dependent on each fire facility. therefore, when the longer bea m is required the fire resistance, the furnace is no longer fun ctioned. fro m middle 1900’s, a ca lculation method started fro m european nations based on an engineering development and as a reasonable alternative to furnace test. until now, korea didn’t utilize the fire engineering method and has used a prescriptive method fro m early 1970’s. therefore , even though the fr 490 showed a better performance than that of sm 490, the steel construction fie ld not allow the fr 490 as a structural steel. to utilize the better structural steel than sm 490 into the construction market, an analytica l evaluation is conducted using mechanical pro perties with long span beams. applied mechan ical properties are shown in table 1. table 1 mechanical properties of fr 490 properties temperature (℃) regression equation properties t≤500 cold value (315 mpa) 500