Microsoft Word - numero_42_art_25.docx A. De Santis et alii, Frattura ed Integrità Strutturale, 42 (2017) 231-238; DOI: 10.3221/IGF-ESIS.42.25 231 Classification of ductile cast iron specimens: a machine learning approach Alberto De Santis, Daniela Iacoviello Department of Computer, Control and Management Engineering Antonio Ruberti, Sapienza University of Rome, via Ariosto 25, 00185, Rome, Italy desantis@dis.uniroma1.it, orcid.org/0000-0001-5175-4951 iacoviello@dis.uniroma1.it, orcid.org/0000-0003-3506-1455 Vittorio Di Cocco, Francesco Iacoviello Department of Civil and Mechanical Engineering, Università di Cassino e del Lazio Meridionale, via G. Di Blasio 43, 03043 Cassino (Fr), Italy v.dicocco@unicas.it iacoviello@unicas.it, orcid.org/0000-0002-9382-6092 ABSTRACT. In this paper an automatic procedure based on a machine learning approach is proposed to classify ductile cast iron specimens according to the American Society for Testing and Materials guidelines. The mechanical properties of a specimen are strongly influenced by the peculiar morphology of their graphite elements and useful characteristics, the features, are extracted from the specimens’ images; these characteristics examine the shape, the distribution and the size of the graphite particle in the specimen, the nodularity and the nodule count. The principal components analysis are used to provide a more efficient representation of these data. Support vector machines are trained to obtain a classification of the data by yielding sequential binary classification steps. Numerical analysis is performed on a significant number of images providing robust results, also in presence of dust, scratches and measurement noise. KEYWORDS. Ductile cast irons; Graphite nodules; Machine learning approach. Citation: De Santis, A., Iacoviello, D., Di Cocco, V. Iacoviello, F., Classification of ductile cast iron specimens: a machine learning approach, Frattura ed Integrità Strutturale, 42 (2017) 231-238. Received: 25.06.2017 Accepted: 15.08.2017 Published: 01.10.2017 Copyright: © 2017 This is an open access article under the terms of the CC-BY 4.0, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. INTRODUCTION iscovered in the years 1943-48, ductile cast irons (DCIs) offer a really interesting combination of cast irons peculiarities (first of all, castability) and of carbon steels mechanical properties (e.g., toughness), [1]. Small additions of elements like Mg or Ce allow to modify the graphite elements shapes, from lamellae (extremely D A. De Santis et alii, Frattura ed Integrità Strutturale, 42 (2017) 231-238; DOI: 10.3221/IGF-ESIS.42.25 232 dangerous) to spheroids (with a decrease of the stress intensification near the graphite elements): these grades are widely used to produce pressure pipes and fittings, in the automotive industry (e.g., crankshafts), in road and construction application. Graphite elements morphology peculiarities (e.g. shape, dimension, distribution) are crucial to define the DCI mechanical properties. Image analysis has been using extensively in the last two decades in order to automatically characterize specimens in material science, [2-4]. The aim is to provide quantitative characterization of the materials in order to determine mechanical properties and establish relationship with damaging mechanisms, [5-6]. Nevertheless up to now the official guide of the International Standard [7] is applied almost manually for visual inspection. Only few attempts have been made to automatize the classification of microstructural image data [8]. In this paper the aim is to provide an automatic procedure to classify specimens according to the American Society for Testing and Materials (ASTM) standard with respect to the graphite elements shape, the “Type” parameter. As will be recalled in Section II, the shape is the first characteristic to be evaluated in order to determine whether a graphite has a desirable shape or not. In case the shape is not nodular, different levels are possible and it could be determined if the graphite has vermicular aspect or if it contains exploded nodules and so on. Given the images classified by two experts, useful features are extracted and re-arranged by principal components analysis (PCA) [9] in order to enhance the informative and useful content of the data. The classification is performed by support vector machine (SVM) suitably trained, [10]; it is a versatile tool useful to classify signals of different nature [11-12]. The classes identified with respect to the Type in the ASTM 2016 are seven; nevertheless binary classifiers are trained in order to simplify the classification step and guarantee the modularity of the procedure. The paper is organized as follows: in Section II, after the description of the data, the image analysis and features extraction is described. Then the training and classification procedure by the SVM is outlined. In Section III numerical results are proposed and discussed, whereas in Section IV conclusions and future work are presented. MATERIALS AND METHODS n this section the procedure for the image acquisition and classification is outlined. Different DCIs have been considered, focusing the attention only on the graphite elements morphological peculiarities and not on the metal matrix microstructure. Specimens have been obtained by means of a metallographic preparation according to the following procedure: - specimen sectioning operation by abrasive cutting; - specimen mounting; - specimen grinding (decreasing grit sizes for abrasive papers up to P1200) and polishing (6 micron diamond followed by 1 micron diamond on low napped polishing cloths); - observation of the metallographically prepared specimen by means of a Light Optical microscope (LOM); Graphite elements characterization is usually performed by means of a visual inspection and a qualitatively evaluation according to the standards [7, 13]. The standardized procedure is based on the visual comparison between the observed images and the charts that are available in the standards. To classify automatically images of ductile cast iron specimens the idea is to extract features useful to describe the specimens and, once the classes of interest are defined, train a classifier able to assign each image to the specific class. This implies the identification of a sort of signature of the images, so that once a new unknown image is proposed, it could be classified by evaluating its signature. On the basis of the International Standard ASTM [7] the information to be retrieved from the images are: - the shape, in particular a measure of its nodularity in shape; the classes with respect to the shape are indicated by: Type I-II-III-IV-V-VI-VII; - the distribution of the graphite in the specimen: it is particularly important in rating the flake graphite and the distribution is described by the letters A-B-C-D-E; - the size of the graphite particles, and the classes are indicated by 1-2-3-4-5-6-7-8 depending on the actual dimension ; - the nodularity, measured as the percentage of the nodular particles present in the microstructure; - the nodule count evaluated as the number of nodules per mm2 at a magnification of 100x. In Fig. 1 examples of specimens belonging to the Type I, IV-and VII (whose differences between them are more evident) are proposed. I A. De Santis et alii, Frattura ed Integrità Strutturale, 42 (2017) 231-238; DOI: 10.3221/IGF-ESIS.42.25 233 It is worth noting that, starting from the classification with respect to the shape (the Type), all the characterizations (distribution, size, nodularity, nodule count) could be further particularized with respect to the other properties, suggesting a sequential procedure for the classification. Therefore, first it will be determined the type-class to which the specimen belongs and then the other characterizations will be established. (a) (b) (c) Figure 1: Examples graphite morphologies: a) Type I; b) Type VI; c) Type VII; [7]. An efficient procedure to classify the specimens with respect to the type is to use binary classifiers in a sequential way: Step 1: with a binary classifier C1 first establish if a specimen could be assigned to Type I class or not - If it belongs to Type I class one can refine the classification with respect to the other characteristics. - If the specimen does not belong to Type I one proceeds to Step 2; Step 2: with a binary classifier C2 establish if the specimen (that is not of TypeI) may be classified of Type II or of Type III-IV-V-VI-VII. If it belongs to Type II, then again one refines the classification with respect to the other characteristics, otherwise one goes to step 3 using another binary classifier C3 and so on. As it can be noted the core of the global procedure is the binary classification step. From now on we will refer to the first step in which one wants to classify a specimen as belonging to the Class 1 (Type I specimen) or the Class 2 (Type II-III-IV-V-VI-VII specimens), thus determining the classifier C1. Therefore it will be possible to distinguish the specimens with normal and well-formed nodules with respect to all the other situations. In Fig. 2 a scheme of the overall classification procedure, simplified when considering only three types, is presented. The classifiers distinguishes the specimens on the basis of suitable features that are evaluated from a simplified representation of the image obtained by using a segmentation procedure; then the features are efficiently modified by the principal components analysis that provides the most efficient data representation. Finally a classifier is obtained by using the support vector machine. The block diagram of the binary classification step is outlined in Fig.3. It consists of two steps; a first one is off-line, aiming at determining the classifier after a proper data processing (image segmentation, features computation and extraction) and training. The second step is on-line, and represents the application of the classifier over images of specimens not used for training. A. De Santis et alii, Frattura ed Integrità Strutturale, 42 (2017) 231-238; DOI: 10.3221/IGF-ESIS.42.25 234 Figure 3: Block diagram of the classification procedure Image analysis and features extraction Given an image, it could be noted that, though it is of good quality, it requires a segmentation process in order to evaluate the properties of each nodule and their spatial distribution. The segmentation with respect to the gray level allows to represent the data with a reduced number of gray levels, thus allowing to retrieve useful information on the nodules, such as the area, or the eccentricity, or their spatial distribution, for example. Different segmentation methods could be applied, [14,15] and in this case, with the nodules well defined over the background, the results obtained with different methods are quite equivalent. Moreover, since the images are of good quality, a binarization is sufficient to enhance the nodules with respect to the background and to determine the properties of interest. The features to be extracted from the images should be chosen in order to determine the best characterization of the data. The indications in the International Standard ASTM 2016 suggest that useful information to be retrieved to determine the classifier C1 concern the roundness of the nodules and their area. Therefore the following features are identified: - features if , 1, 2, 3i  that are the number of nodules with area (in pixels) in the intervals  1 25, 125I  ,  2 126 500I  ,  3 501, 900I  , respectively. Nodules with area less than 25 pixels are discarded since could be associated to dust or measurement noise; nodules with are greater than 900 pixels are in general not present; - feature 4f defined as the number of elements with area greater than the minimum one (25 pixels) normalized with respect to the area of the background: it is a measure of the presence of the nodules; - features jf , 5, 6, 7j  that are the solidities of the nodules in the three intervals iI , 1, 2, 3i  respectively; the solidity is defined as the area of the nodule over the convex area, that is the area of the smallest convex polygon that can contain the nodule; - features kf , 8, 9,10k  that are the eccentricities of the nodules in the three intervals iI , 1, 2, 3i  respectively, and are a measure of the roundness of the nodules. Therefore, given a set of images of specimens jS , 1, 2,...,j n , a vector of 10fn  features is calculated  1 2 3 4 5 6 7 8 9 10F f f f f f f f f f f for each image; these information are collected in a dataset matrix D of dimension fn n , where on the k -th row the fn features of the specimen kS are collected. The fn features have been chosen in order to determine the best characteristics useful to distinguish specimens of Class 1 with respect to specimens of Class 2; nevertheless if one uses directly these features to train a classifier, maybe they don’t A. De Santis et alii, Frattura ed Integrità Strutturale, 42 (2017) 231-238; DOI: 10.3221/IGF-ESIS.42.25 235 represent at best the data, or maybe some of them yields the same information. To this aim the Principal Component Analysis, that will be herein briefly recalled, yields the best data representation, [16]. The PCA is a linear data transformation aiming at reducing the redundancy of the data covariance matrix and maximizing the information retrieved; in the new reference coordinate the new variables are independent one another. One can consider the features selections, when a subset of the original features is considered, or the features extraction, when a new set of features is built suitably weighting the information of interest. Of course, when the dimensionality of the data is reduced it is mandatory to quantify the loss of information. In this paper the PCA are used aiming at the features extraction. More precisely, the covariance matrix DC of size f fn n of the data matrix D is evaluated and its eigenvalues  1,..., n f  are sorted according to decreasing order. The corresponding unit eigenvectors iv , 1, 2, ..., fi n are the directions of maximum variance of the data; the transformation yielding the new data representation in the principal components Z is: Z D V  (1) where 1 n fV v v      is the matrix constituted by the ordered eigenvectors. Therefore, for example, the first principal component is:    1 11 1 1 1 1p n fZ Z Z D v D v   being 1D the first row of matrix D . Generally the number of principal components pn is chosen in order to retrieve the p-percentage of the information content, that is: 1 1 100 % n p i i n f i i p         (2) It means that from now on, instead of trying to classify the data collected in the matrix D of dimension fn n , the data to be considered are the first pn principal components. Training and classification The PCA allows to reduce the dimensionality of the data preserving adequately the information; therefore now each image X is described by a new set of feature. The aim is to determine a classifier able to assign each set of feature (and therefore each image) to Class 1 or to Class 2. To train a classifier able to separate the available data into two classes, the set of n images is split into two groups, the training set, trN , and the test set testN . To the data corresponding to images belonging to the Class 1 it is assigned label 1, whereas label 0 is assigned to the data belonging to the Class 2. The training set trN is divided into two groups, 1trN and 2trN ; the first one is used to train the classifier; the second one 2trN is used to determine the classification accuracy. The support vector machine determines the optimal hyperplane that splits the data into two groups, [17]; it is a tradeoff between the requirement of minimizing the error on misclassified points and maximizing the Euclidean distance between the closest points, see Fig.4. The optimal hyperplane is obtained as the solution of the quadratic programming problem: , , 1 1 min 2 n T i w b i w w H      A. De Santis et alii, Frattura ed Integrità Strutturale, 42 (2017) 231-238; DOI: 10.3221/IGF-ESIS.42.25 236 with the constraint:   1 , 0Ti i i iy w x b      , where w is the vector of the points perpendicular to the separating hyperplane and H>0 is a penalty parameter on the error term. Figure 4: Representation of the classification problem. To make the elements ix of the two classes linearly separable, the data are mapped into a richer space, and the separating hyperplane is determined in that space. A possible choice for the mapping function  is the radial basis function and, denoting with 2 the 2L -norm, for the kernel function it is assumed: 2 2 2 ( , ) ( ) ( ) exp 2 i jT i j i j x x K x x x x             The two parameters to be evaluated, H and  , may be determined during the training phase, by using the 10-fold cross validation, [18]. The classification is performed by the SVM algorithm LIBSVM 3.18, [19- 20]. Once the optimal parameters  ,H   have been determined, the classifier is trained; the classification accuracy, evaluated on the 2trN , is defined as the percentage of correctly classified data with the optimal choice  ,H   and it is a property of the classifier. With this calculation the off-line phase of the classification procedure is over. The obtained classifier is tested over the test set testN , not used for the training, simulating the situation of unlabeled data. The percentage of misclassified images is the error of the classifier. The same procedure is applied to train the classifier C2 able to assign a specimen (not belonging to Type I class) to Type II class or to Type III-IV-V-V-VII class and so on, according to the scheme of Fig. 2. NUMERICAL RESULTS AND DISCUSSION n this section the results of the classification procedure are described. As could be noted in the International Standard [7], the specimens of Type I, II and III, though they could present a similarity between each other, they differ significantly from the other types. Therefore out attention will be focused in Type I, II and III, even if the overall analysis may be extended to all the types’ classification. The first step is the classification of a specimen as of Type I or of Type II-III. If the specimen is of Type II-III a further classification procedure starts in order to decide whether the specimen is of Type II or III. I A. De Santis et alii, Frattura ed Integrità Strutturale, 42 (2017) 231-238; DOI: 10.3221/IGF-ESIS.42.25 237 A set of 192n  images of specimens is considered, 64 are of specimens of Type I, 64 of specimens of Type II and 64 of specimens of Type III. The images have been previously classified by an expert, manually. To obtain the features a binarization procedure is applied; it has been chosen the binarization by the discrete level set approach [15] and the ten features described in Section 2 have been evaluated, thus obtaining three matrices of size 64 10 , collected together in the data matrix D , 192 10 . To deal with data with comparable magnitude, a normalization is applied. The covariance matrix DC of size 10 10 of the data matrix D is evaluated; after evaluating its eigenvalues, by using formula (2) 6pn  principal components are considered, thus preserving the percentage of more than 94% of the original information. The training set trN contains 45 images: 27 of Type I, randomly chosen among the set of 64 Type I data, and 27 of Type II and III randomly chosen among the set of 128 images of specimens of these type. The test set is constituted by 20 images, equally distributed between Type I and Type II-III. The 1trN contains 40 elements and the remaining 14 are used for the set 2trN . The number of images of specimens of Class 1 (i.e. Type I specimens) and of Class 2 (i.e. Type II- and III, equally distributed) is the same in the groups involved in training and testing steps to avoid polarization in the result. As said, the parameters  ,H   are determined by the 10-fold cross validation that provides also the optimized value for b . The used SVM algorithm LIBSVM 3.18 is a simple and efficient open source software. The classification accuracy is calculated as the average value of the accuracy evaluated for 20 different random choices of the training and the test sets, to be sure that the results do not depends on lucky choices, obtaining a percentage of success over 99%. With this calculation the off-line step is over. The results over the test set (containing images not used in the training phase) yield a percentage of success of 97.3% 2.7 . The results of the classifier C1 appears satisfactory; moreover it has been also investigated if the classifier C1 makes a mistake more often with images of Class 1 (Type I data) or with images of Class 2 (Type II and III data), and among the Class 2 if more errors are made when testing with images of Type II or III. This unbundled test on 10 images of each type, repeated 20 times, shows that images of Type III are always correctly classified (percentage of success of 100%), whereas the results on Type I and Type II yield percentage of success of 97.5% 5.5 and 94.5% 10.5 , respectively. A possible explanation could be that images of Type III are a little bit more different with respect to the Type I, than the images of Type II. For the images classified by the classifier C1 as belonging to Class 2 the second classifier C2 must be applied in order to discriminate the images of Type II and those of Type III. Also in this case all the results have been repeated for 20 different random choices of the training and test sets. The classifier C2, trained using only images of Type II and Type III, has a classification accuracy of 98.9%. The test accuracy provides a percentage of success of almost 100% on a test set of 10 images belonging to Type II class and of 98.9% 3.15 on a test set of 10 images of Type III. The results of the classifier C2 are even more satisfactory with respect to those of classifier C1, since the training has been more specific. The classifier C2 has the aim of determining the class membership of images of Type II and III; when applied to an image of Type I, for example if the classifier C1 has provided an erroneous classification, in more than 91% the C2 classifier assigns the specimen of Type I to the class of Type II images. This is the correct choice, being the images of Type II the more similar to the ones of Type I. CONCLUSIONS AND FUTURE WORK n this paper an automatic procedure to support the classification of microstructure of graphite in iron castings is proposed. By training binary support vector machine classifiers it is possible, in an efficient way, to determine the type of the specimen according to the American Society for Testing and Materials guidelines and therefore to proceed in the classification specifying the size, the nodularity and the nodule count. Three classes (Type I, Type II and Type III) may be identified by the proposed procedure, but it could be extended to as many classes as needed. The choice of using binary classifiers operating sequentially is determined aiming at yielding a simple, efficient and modular procedure. I A. De Santis et alii, Frattura ed Integrità Strutturale, 42 (2017) 231-238; DOI: 10.3221/IGF-ESIS.42.25 238 The classifier uses features evaluated on the original specimens’ images and successively suitably transformed by principal components analysis that reduces the complexity and yields a more efficient representation of the information. The results appear satisfactory, and future work will be devoted in: - classify the images of the specimen with respect to all the properties (size, nodule count,…); - determine the most suitable features in order to better characterize each nodule present in the specimen; - consider different classification schemes, for example by using polling systems, evaluating their robustness. REFERENCES [1] Labrecque, C., Gagne, M., Ductile Iron: Fifty Years of Continuous Development, Canadian Metallurgical Quarterly, 37( 5) (1998) 343–78. [2] Bonnet, N., Multivariate statistical methods for the analysis of microscope image series: applications in materials science, Journal of Microscopy, 190 (1998) 2-18. [3] Filho, P.P.R., Moreira, F.D.L., de Lima Xavier, Gomez, F.G, S.L., dos Santos, J.C., Freitas, F.N.C, Freitas, R.G., New analysis method application in metallographic images through the construction of mosaics via speeded up robust features and scale invariant feature transform, Materials, 8 (2015) 3864-3882. [4] Papa, J.P., Pereiram, C.R., de Albuquerque, Silva, V.H.C C.C., Falcao, A.X, Tavares, J.M.R.S., Precipitates segmentation from scanning electron microscope images through machine learning techniques, Lecture Notes in Computer Science Series, 6636 (2011) 456-468. [5] Di Cocco, V., Iacoviello, F., Rossi, A., Iacoviello, D., Macro and microscopical approach to the damaging micromechanisms analysis in a ferritic ductile cast iron, Theoretical and applied fracture mechanics, 69 (2014) 26-33. [6] De Santis, A., Di Bartolomeo, O., Iacoviello, D., Iacoviello, F., Quantitative shape evaluation of graphite elements in ductile iron, Journal of Materials Processing and Technology, 196 (1-3) (2008) 292-302. [7] ASTM standard A247 – 16a, Standard test method for evaluating the microstructure of graphite in iron castings, (2016) 1-13. [8] Decost, B.L., Holm, E.A., A computer vision approach for automated analysis and classification of microstructural image data, Computational materials science, 110 (2015) 126-133 [9] Jolliffe, I.T., Principal Component Analysis, 2nd edition, Springer, (2002). [10] Cristianini, N., Shawe-Taylor, J., An Introduction to Support Vector Machines and other kernel-based learning methods, Cambridge University Press, New York, NY, USA, (2000). [11] Nguyen, B.P., Tay, W.L.,Chui, C.K. :,Robust biometric recognition from palm depth images for gloved hands, IEEE Transactions on human-machine-systems, 45 (6) (2015) 799-805. [12] Subasi, A., Gursoy, M.I., EEG signal classification using PCA, ICA, LDA and support vector machine, Expert systems with applications, 37 (2010) 8659-8666. [13] ASTM standard E2567 – 13a, Standard test method for determining nodularity and nodule count in ductile iron, (2013) 1-4. [14] Otsu, N.,Threshold selection method for gray-level histograms, IEEE Transactions od Systems, Man, and Cybernetics, 1 (1979) 62-66. [15] De Santis, A., Iacoviello, D., Discrete level set approach to image segmentation, Signal, Image and Video Processing, Springer-Verlag London, 1(4) (2007) 303-320. [16] Song, F., Guo, Z., Mei, D., Feature selection using principal component analysis, 2010 International Conference on System Science, Engineering Design and Manufacturing Informatization, (2010) 27–30. [17] Hsu, C.W., Chang, C.C., Lin, C.J., A practical guide to support vector classification, Bioinformatics, 1 (1) (2003) 1–16. [18] Efron, B., Estimating the error rate of a prediction rule: improvement on cross-validation, Journals- American Statistical Association, 78 (1983) 316–331. [19] Chang, C-C, Lin, C-J., LIBSVM: A library for support vector machines,, ACM Transactions on Intelligent Systems and Technology, 2 (27) (2011) 1-27. Software available at http://www.csie.ntu.edu.tw/~cjlin/libsvm. [20] Cortes, C., Vapnik, V., Support vector networks,, Mach. Learn., 20 (1995) 273–297. << /ASCII85EncodePages false /AllowTransparency false /AutoPositionEPSFiles true /AutoRotatePages /None /Binding /Left /CalGrayProfile (Dot Gain 20%) /CalRGBProfile (sRGB IEC61966-2.1) /CalCMYKProfile (U.S. Web Coated \050SWOP\051 v2) /sRGBProfile (sRGB IEC61966-2.1) /CannotEmbedFontPolicy /Error /CompatibilityLevel 1.4 /CompressObjects /Tags /CompressPages true /ConvertImagesToIndexed true /PassThroughJPEGImages true /CreateJobTicket false /DefaultRenderingIntent /Default /DetectBlends true /DetectCurves 0.0000 /ColorConversionStrategy /CMYK /DoThumbnails false /EmbedAllFonts true /EmbedOpenType false /ParseICCProfilesInComments true /EmbedJobOptions true /DSCReportingLevel 0 /EmitDSCWarnings false /EndPage -1 /ImageMemory 1048576 /LockDistillerParams false /MaxSubsetPct 100 /Optimize true /OPM 1 /ParseDSCComments true /ParseDSCCommentsForDocInfo true /PreserveCopyPage true /PreserveDICMYKValues true /PreserveEPSInfo true /PreserveFlatness true /PreserveHalftoneInfo false /PreserveOPIComments true /PreserveOverprintSettings true /StartPage 1 /SubsetFonts true /TransferFunctionInfo /Apply /UCRandBGInfo /Preserve /UsePrologue false /ColorSettingsFile () /AlwaysEmbed [ true ] /NeverEmbed [ true ] /AntiAliasColorImages false /CropColorImages true /ColorImageMinResolution 300 /ColorImageMinResolutionPolicy /OK /DownsampleColorImages true /ColorImageDownsampleType /Bicubic /ColorImageResolution 300 /ColorImageDepth -1 /ColorImageMinDownsampleDepth 1 /ColorImageDownsampleThreshold 1.50000 /EncodeColorImages true /ColorImageFilter /DCTEncode /AutoFilterColorImages true /ColorImageAutoFilterStrategy /JPEG /ColorACSImageDict << /QFactor 0.15 /HSamples [1 1 1 1] /VSamples [1 1 1 1] >> /ColorImageDict << /QFactor 0.15 /HSamples [1 1 1 1] /VSamples [1 1 1 1] >> /JPEG2000ColorACSImageDict << /TileWidth 256 /TileHeight 256 /Quality 30 >> /JPEG2000ColorImageDict << /TileWidth 256 /TileHeight 256 /Quality 30 >> /AntiAliasGrayImages false /CropGrayImages true /GrayImageMinResolution 300 /GrayImageMinResolutionPolicy /OK /DownsampleGrayImages true /GrayImageDownsampleType /Bicubic /GrayImageResolution 300 /GrayImageDepth -1 /GrayImageMinDownsampleDepth 2 /GrayImageDownsampleThreshold 1.50000 /EncodeGrayImages true /GrayImageFilter /DCTEncode /AutoFilterGrayImages true /GrayImageAutoFilterStrategy /JPEG /GrayACSImageDict << /QFactor 0.15 /HSamples [1 1 1 1] /VSamples [1 1 1 1] >> /GrayImageDict << /QFactor 0.15 /HSamples [1 1 1 1] /VSamples [1 1 1 1] >> /JPEG2000GrayACSImageDict << /TileWidth 256 /TileHeight 256 /Quality 30 >> /JPEG2000GrayImageDict << /TileWidth 256 /TileHeight 256 /Quality 30 >> /AntiAliasMonoImages false /CropMonoImages true /MonoImageMinResolution 1200 /MonoImageMinResolutionPolicy /OK /DownsampleMonoImages true /MonoImageDownsampleType /Bicubic /MonoImageResolution 1200 /MonoImageDepth -1 /MonoImageDownsampleThreshold 1.50000 /EncodeMonoImages true /MonoImageFilter /CCITTFaxEncode /MonoImageDict << /K -1 >> /AllowPSXObjects false /CheckCompliance [ /None ] /PDFX1aCheck false /PDFX3Check false /PDFXCompliantPDFOnly false /PDFXNoTrimBoxError true /PDFXTrimBoxToMediaBoxOffset [ 0.00000 0.00000 0.00000 0.00000 ] /PDFXSetBleedBoxToMediaBox true /PDFXBleedBoxToTrimBoxOffset [ 0.00000 0.00000 0.00000 0.00000 ] /PDFXOutputIntentProfile () /PDFXOutputConditionIdentifier () /PDFXOutputCondition () /PDFXRegistryName () /PDFXTrapped /False /CreateJDFFile false /Description << /ARA /BGR /CHS /CHT /CZE /DAN /DEU /ESP /ETI /FRA /GRE /HEB /HRV (Za stvaranje Adobe PDF dokumenata najpogodnijih za visokokvalitetni ispis prije tiskanja koristite ove postavke. Stvoreni PDF dokumenti mogu se otvoriti Acrobat i Adobe Reader 5.0 i kasnijim verzijama.) /HUN /ITA /JPN /KOR /LTH /LVI /NLD (Gebruik deze instellingen om Adobe PDF-documenten te maken die zijn geoptimaliseerd voor prepress-afdrukken van hoge kwaliteit. De gemaakte PDF-documenten kunnen worden geopend met Acrobat en Adobe Reader 5.0 en hoger.) /NOR /POL /PTB /RUM /RUS /SKY /SLV /SUO /SVE /TUR /UKR /ENU (Use these settings to create Adobe PDF documents best suited for high-quality prepress printing. Created PDF documents can be opened with Acrobat and Adobe Reader 5.0 and later.) >> /Namespace [ (Adobe) (Common) (1.0) ] /OtherNamespaces [ << /AsReaderSpreads false /CropImagesToFrames true /ErrorControl /WarnAndContinue /FlattenerIgnoreSpreadOverrides false /IncludeGuidesGrids false /IncludeNonPrinting false /IncludeSlug false /Namespace [ (Adobe) (InDesign) (4.0) ] /OmitPlacedBitmaps false /OmitPlacedEPS false /OmitPlacedPDF false /SimulateOverprint /Legacy >> << /AddBleedMarks false /AddColorBars false /AddCropMarks false /AddPageInfo false /AddRegMarks false /ConvertColors /ConvertToCMYK /DestinationProfileName () /DestinationProfileSelector /DocumentCMYK /Downsample16BitImages true /FlattenerPreset << /PresetSelector /MediumResolution >> /FormElements false /GenerateStructure false /IncludeBookmarks false /IncludeHyperlinks false /IncludeInteractive false /IncludeLayers false /IncludeProfiles false /MultimediaHandling /UseObjectSettings /Namespace [ (Adobe) (CreativeSuite) (2.0) ] /PDFXOutputIntentProfileSelector /DocumentCMYK /PreserveEditing true /UntaggedCMYKHandling /LeaveUntagged /UntaggedRGBHandling /UseDocumentProfile /UseDocumentBleed false >> ] >> setdistillerparams << /HWResolution [2400 2400] /PageSize [612.000 792.000] >> setpagedevice