Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Sonification of time dependent data

2002, VDM@ ECML/ …

VDM@ECML/PKDD2002 Proceedings International Workshop on Visual Data Mining 19th August, 2002, Helsinki, Finland Edited by Simeon J. Simoff, Monique Noirhomme-Fraiture and Michael H. Böhlen in conjunction with 13th European Conference on Machine Learning (ECML'02) and 6th European Conference on Principles and Practice of Knowledge Discovery in Databases (PKDD'02), 19-23 August 2002, Helsinki, Finland © Copyright 2002. The copyright of these papers belongs to the paper's authors. Permission to copy without fee all or part of this material is granted provided that the copies are not made or distributed for direct commercial advantage. Proceedings of the 2nd International Workshop on Visual Data Mining VDM@ECML/PKDD'2002, in conjunction with 13th European Conference on Machine Learning (ECML'02) and 6th European Conference on Principles and Practice of Knowledge Discovery in Databases (PKDD'02), 19-23 August 2002, Helsinki, Finland S. J. Simoff, M. Noirhomme-Fraiture and M. H. Böhlen (eds) Workshop Web Site: http://www-staff.it.uts.edu.au/~simeon/vdm_pkdd2002/ Foreword Visual data mining is a collection of interactive methods for knowledge discovery from data, which integrates human perceptual capabilities to spot patterns, trends, relationships and exceptions with the capabilities of the modern digital computing to characterise data structures and display data. The underlying technology builds on visual and analytical processes developed in various disciplines including data mining, information visualisation, and statistical learning from data with algorithmic extensions that handle very large, multidimensional, multivariate data sets. The growing research and development in visual data mining offers machine learning and data mining communities complementary means of analysis that can assist in uncovering patterns and trends that are likely to be missed with other non-visual methods. Consequently, the machine learning and data mining communities have recognised the significance of this area. The first edition of the workshop took place at the ECML/PKDD conference in Freiburg, following a similar workshop at the ACM KDD conference in San Francisco. The first edition of the workshop offered to the ECML/PKDD2001 participants a mixture of presentations on state-of-art methods and techniques, with controversial research issues and applications. A report about this workshop has been published in SIGKDD Explorations 3 (2), pp. 78-81. The presentations were grouped in three streams: Methodologies for Visual Data Mining; Applications of Visual Data Mining; and Support for Visual Data Mining. The workshop included also two invited presentations – one from Erik Granum, the head of the Laboratory of Computer Vision and Media Technology, Aalborg University (on the behalf of the 3DVDM group), who presented an overview of the unique interdisciplinary 3DVDM group, its current projects and research opportunities there; and one from Monique Noirhomme-Fraiture, who demonstrated 2D and 3D visualisation support for visual data mining of symbolic data. The workshop brought together a number of cross-disciplinary researchers, who were pleased with the event and there was consensus about the necessity of turning it into an annual meeting, where researchers, both from the academia and industry can exchange and compare both relatively mature and green house theories, methodologies, algorithms and frameworks in the emerging field of visual data mining. Meantime a discussion list on visual data mining (vdm@it.uts.edu.au) was setup to facilitate this exchange. This workshop has been initiated and organised in response to this interest. Being a second edition, the workshop this year is aiming to create a stimulating atmosphere for open discussions of the crossdisciplinary theoretical foundations, frameworks, interactive methods, algorithms and environments for visual data processing and analysis; novel applications and utilisation of other perceptual channels. Consequently, the papers selected for presentation at the workshop are grouped in the following sessions: Visual Data Pre-processing; Visual Environments for Data Mining and Analysis; Interactive Visual Data Mining Algorithms and Applications; and ‘Perceptual’ Data Mining. The works selected for presentation at this workshop form more cohesive body of work, which indicates that the field has made a step forward towards achieving some level of maturity. We would like to thank all those who submitted their work to the workshop. As part of the ECML/PKDD joint conference series, the workshop follows a rigid blind peer-review process. All papers were extensively reviewed by at least three referees drawn from the program committee. Special thanks go to them. Once again, we would like to thank all those, who supported this year’s efforts on all stages – from the development and submission of the workshop proposal to the preparation of the final program and proceedings. Simeon J. Simoff Monique Noirhomme-Fraiture Michael H. Böhlen Workshop co-Chairs July 2002 ii Workshop Chairs Simeon J. Simoff Monique Noirhomme-Fraiture Michael H. Böhlen Program Committee Michael Ankerst Boeing, USA James L. Alty Loughborough University, UK Katy Börner Indiana University, USA Alberto Del Bimbo Universitá degli Studi di Firenze, Italy Maria F. Costabile Universita' di Bari, Italy Alex Duffy University of Strathclyde, UK Erik Granum Aalborg University, Denmark Markus Hegland Australian National University, Australia Maolin Huang University of Technology Sydney, Australia Alfred Inselberg Multidimensional Graph Ltd, Israel Carlo Lauro University of Naples, Italy Donato Malerba Università degli Studi, Italy Carl H. Smith University of Maryland, USA Michael Schroeder City University, UK Bruce Thomas University of South Australia, Australia iii Program for VDM@ECML/PKDD2002 Workshop Monday, 19 August 2002, Helsinki, Finland 9:00 - 9:10 Opening and Welcome 9:10 - 10:30 Session 1 – Visual Data Pre-processing • 09:10 - 09:50 CLUSTERING BY ORDERING DENSITY-BASED SUBSPACES Kan Liu, Dongru Zhou, Xiaozheng Zhou • 09:50 - 10:30 CAN HIERARCHICAL CLUSTERING IMPROVE THE EFFICIENCY OF NON-LINEAR DIMENSION REDUCTION WITH SPRING EMBEDDING Michael Schroeder and George Katopodis 10:30 - 11:00 Coffee break 11:00 - 13:00 Session 2 - Visual Environments for Data Mining and Analysis • 11:00 - 11:40 A VISUAL DATA MINING ENVIRONMENT Stephen Kimani, Tiziana Catarci and Giuseppe Santucci • 11:40 - 12:20 A POST-PROCESSING ENVIRONMENT FOR BROWSING LARGE SETS OF ASSOCIATION RULES Alipio Jorge, João Poças and Paulo Azevedo • 12:20 - 13:00 VISUAL POST-ANALYSIS OF ASSOCIATION RULES Dario Bruzzese and Cristina Davino 13:00 - 14:30 Lunch 14:30 - 16:00 Session 3 – Interactive Visual Data Mining Algorithms and Applications • 14:30 - 15:00 COOPERATION BETWEEN AUTOMATIC ALGORITHMS, INTERACTIVE ALGORITHMS AND VISUALIZATION TOOLS FOR VISUAL DATA MINING François Poulet • 15:00 - 15:30 DEFINING LIKE-MINDED AGENTS WITH THE AID OF VISUALIZATION Penny Noy and Michael Schroeder • 15:30 - 16:00 VISUAL DATA MINING OF CLINICAL DATABASES: AN APPLICATION TO THE HEMODIALYTIC TREATMENT BASED ON 3D INTERACTIVE BAR CHARTS Luca Chittaro, Carlo Combi and Giampaolo Trapasso 16:00 - 16:30 Coffee break iv 16:30 - 17:30 Session 4 – ‘Perceptual’ Data Mining • 16:30 - 17:00 SONIFICATION OF TIME DEPENDENT DATA Monique Noirhomme-Fraiture, Olivier Schöller, Christophe Demoulin, Simeon J. Simoff • 17:00 - 17:30 A SURPRISE FROM THE 3DVDM GROUP Michael H. Böhlen 17:30 - 18:00 Discussion “Where Visual Data Mining is Heading” and Closure v Table of Contents Clustering By Ordering Density-Based Subspaces K. Liu, D. Zhou, X. Zhou ……….…………………………….…………. 1 Can Hierarchical Clustering Improve The Efficiency Of Non-Linear Dimension Reduction With Spring Embedding M. Schroeder, G. Katopodis ……………………………………….……. 11 A Visual Data Mining Environment S. Kimani, T. Catarci, G. Santucci ……………………………………….. 27 A Post-Processing Environment For Browsing Large Sets Of Association Rules A. Jorge, J. Poças, P. Azevedo …………………………………………. 43 Visual Post-Analysis Of Association Rules D. Bruzzese, C. Davino ……………………………………..…………. 55 Cooperation Between Automatic Algorithms, Interactive Algorithms And Visualization Tools For Visual Data Mining F. Poulet ………………………………………………………….…. 67 Defining Like-Minded Agents With The Aid Of Visualization P. Noy, M. Schroeder ………………………………………………….. 81 Visual Data Mining Of Clinical Databases: An Application To The Hemodialytic Treatment Based On 3D Interactive Bar Charts L. Chittaro, C. Combi, G. Trapasso ………………………………..…… 97 Sonification Of Time Dependent Data M. Noirhomme-Fraiture, O. Schöller, C. Demoulin, S. J. Simoff Author Index vi ……………. 113 Clustering by Ordering Density-Based Subspaces Kan LIU*, Dongru ZHOU, Xiaozheng ZHOU School of Computer, Wuhan University 430072 Wuhan, China * lk2000@public.wh.hb.cn Abstract. Finding clusters on the basis of density distribution is a traditional approach to discover clusters with arbitrary shape. Some density-based clustering algorithms such as DBSCAN, OPTICS, DENCLUE, and CLIQUE etc have been explored in recent researches. This paper presents a new approach which is based on the ordered subspace to find clusters. The key idea is to sort the subspaces according to their density, and set a new cluster for the maximal subspace of the subspace list. Since the number of the subspaces is much less than that of the data, very large databases with high-dimensional data sets can be processed with high efficiency. We also present a new method to project high-dimensional data, and then some results of clustering with visualization are demonstrated in this paper. Keywords: Clustering, Density-based, Data visualization 1 Introduction The process of grouping the physical data or abstract data based on their similarity is called clustering. Clustering is an important analysis method in data mining, which could help people to better understand and observe the natural classification or structure of the data. Clustering algorithms are used to automatically classify data items into the relative, meaningful clusters. After clustering, the items within any cluster are highly relevant and the items across different clusters are lowly relevant. The factors listed below are always being considered when evaluating a clustering algorithm. • Scalability: A good clustering algorithm can deal with large datasets including up to millions data items. • Discovery of clusters with arbitrary shape: A cluster may have an arbitrary shape. A clustering algorithm should not only apply to the regular clusters. • Minimum parameters input: It is a heavy burden for the users to input those important parameters. In the meantime, this brings more trouble in getting good quality clustering. 2 Liu et al. • Insensitive to order of input records: Inputting data in different order should not lead to different results. • High-dimensionality: A dataset may include many attributes, clustering data in high-dimensional space is highly demanded in many applications. Current clustering algorithms can be broadly classified into two categories: hierarchical and partitional. Hierarchical algorithms, such as BIRCH [7], CURE [8] and CHAMELEON [9] etc, decompose a dataset into several levels of nested partitions. They start by placing each object in its own cluster and then merge these atomic clusters into larger and larger clusters until all objects are in a single cluster. Or they reverse the process by starting with all objects in a cluster and subdividing into smaller pieces. Partitional algorithms, such as CLARA [5] and CLARANS [6], partition the objects based on a clustering criterion. The popular methods, K-means and K-medoid, use the cluster average, or the closest object to the cluster center, to represent a cluster. New clustering algorithms have been proposed in recent researches [4]. For example, DBSCAN, OPTICS and CLIQUE are based on density; STING, WAVE CLUSTER are based on grid; and COBWEB, SOM are based on model. 2 Related Work The main idea of density-based approaches is to find regions of high-density and lowdensity, with high-density regions being separated from low-density regions. These approaches can make it easy to discover arbitrary clusters. A common way is to divide the high-dimensional space into density-based grid units. Units containing relatively high densities are the cluster centers and the boundaries between clusters fall in the regions of low-density units. For example, the CLIQUE [1] algorithm processes dense units level-by-level. It first determines 1-dimensional dense units by making a pass over the data. Having determined (k-1)-dimensional dense units, the candidates for k-dimensional units are determined. A cluster is a maximal set of connected dense units in k-dimensions. This algorithm automatically finds subspaces of the highest dimensionality such that high-density clusters exist in those subspaces, but the accuracy of the clustering result may be degraded at the expense of simplicity of the method. The alternative way is to calculate parameter ‘directly density-reachable’ or ‘reachability-distance’ of the object. For example, the DBSCAN [3] aims at discovering clusters of arbitrary shape based on the formal notion of densityreachability for k-dimensional points. OPTICS [2] solves the problem of DBSCAN, which only computes a single level clustering. OPTICS produces a special order of the database with respect to its density-based clustering structure, and according to the order of the reachability-distance of each object, it can quickly reach the high-density region. OPTICS is good for both automatic and interactive cluster analysis, including finding intrinsic clustering structure, and is not limited to one global parameter setting. But it is infeasible to apply it in its current form to a database containing several million high-dimensional objects. In this paper, we propose an integrative Clustering by Ordering Density-Based Subspaces 3 clustering method which is to partition the data space based on the density and gridbased techniques and realize the visualization of the clustered result. 3 Clustering by Ordering Dense Units 3.1 Basic statement The density-based technique is adopted in our approach, in which the data space is partitioned into non-overlapping rectangular units and the density distribution will be deduced by calculating the data volume of each rectangular unit. Suppose D is a n-dimensional dataset with m items: D={X1, X2, …, Xi, …, Xm}, where Xi=(Xi1, Xi2, …, Xij, …, Xin), (i≤m), and Xijis the value of the j th attribute of Xi. If each dimension of the dataset is equally divided into t parts, then all the items in the dataset fall into k units: U={U1, U2, …, Ui, …, Uk} (k≤tn), where Ui=(Ui1, Ui2, …, Uij, …, Uin) is the vector of each equally divided attribute. Two units U1 and U2 are defined to be adjacent only when any attribute of one unit is adjacent to that of the other unit: | U1j -U2j | = 1, U1s = U2s, (j, s ≤ k, j ≠ s). We define density peaks as those units whose densities are larger than those of the adjacent units; similarly, we define density valleys as the units whose densities are lower than those of the adjacent ones. 3.2 Algorithm When high-dimensional space is divided into k equal subspaces (units), the density peaks are regarded as the clusters centers. So the key process is how to find the density peaks. In our approach, CODBU (Clustering by Ordering Density-Based Units), the units with densities greater than threshold are ranked by the density value. The change from density peaks to density valleys is expressed by hierarchical level. The density peak is positioned in the first level of the cluster, the adjacent units are in the second level, and finally, the density valley is positioned in the last level of the cluster. First, we calculate the density values of each unit, and then rank the units whose densities are greater than threshold value. The largest-density unit will be analysed first. Each unit is compared with its adjacent units (neighbours) in density, if its density is greater than that of any other adjacent unit, it is considered as a density peak and then be set as the first level, a new cluster is emerging. If its density is less than one of adjacent units, then this unit will be grouped into the cluster of the adjacent unit; if its value is less than many of adjacent units, then this unit will be grouped into the cluster of the lowest-level unit. Fig.1 describes the process of cluster, in which only 2 basic parameters are required: the number of subdivisions for each dimension and the density threshold value. These two values are entered manually according to the size of the dataset and the required accuracy of the result. 4 Liu et al. CODBU (MinDen, t) BEGIN int cluster_no = 0; int k = 0; Divide each attribute into t equal parts, initialize U; Read data x from dataset If (x ∈ Uj) Uj.density++; For all Uj if (Uj.density >= MinDen){ U.addElement(Uj); Uj.cluster_no=0; Uj.layer_no=0; k++; } Quicksort ( U ); // sort in the descending order for (j=0; j<k; j++){ for all neighbors of Uj if (neighbor.density > Uj.density) // group into the high-density unit cluster if (Uj.layer_no = 0) { Uj.cluster_no = neighbor.cluster_no; Uj.layer_no = neighbor.layer_no+1;} else // group into the low-level unit cluster if (Uj.layer_no > neighbor.layer_no+1){ Uj.cluster_no=neighbor.cluster_no; Uj.layer_no=neighbor.layer_no+1;} // form a new cluster if (Uj.layer_no=0){ cluster_no++; Uj.cluster_no=cluster_no; Uj.layer_no=1;} } END. Fig.1 CODBU: Clustering by Ordering Density-Based Units Figure 2 shows a simple 2-dimension data set. The sequence number of each unit is shown in the unit, and ‘ * ’ stands for the spread points among them. Now sort all the squares whose density values are larger than 3, the result is (sorted by density): 4 (11), 5 (9), 8 (8), 10 (8), 1 (7), 9 (7), 11 (7), 3 (6), 6 (5), 12 (5), 7 (4), 2 (3). Based on the above result, we can find 2 clusters (sorted by level): C1={4 (1), 5 (2), 1 (2), 10 (2), 3 (2), 6 (3), 11 (3)}; C2={8 (1), 9 (2), 12 (2), 7 (2), 2 (2)}. Fig.2 Two clusters based on the density values of units Clustering by Ordering Density-Based Subspaces 5 We can see that, although the density of unit 5 is larger than that of unit 8, as it is adjacent to unit 4 which has higher density, unit 5 is still grouped into the first cluster. Since the density value of unit 8 is greater than that of any adjacent unit, it forms a new cluster. The density of unit 7 is lower than those of unit 6 and 8, but the unit 8 has a lower level than unit 6, so unit 7 is grouped into the second cluster. 3.3 Algorithm analysis The dataset is only scanned once in our approach. Suppose k is the number of subspaces whose densities are greater than threshold value, the time complexity of applying quick sort is O (k*log k). By building up a search tree, the time complexity of comparing the density value of each unit with those of its adjacent units is O(nk), and n is the number of dimensions. Since the total number of units is much less than that of the data items in the dataset, the time complexity is decreased dramatically. Furthermore, the analysis of the data space based on the density order can better reflect various clusters than the traditional approach in which the data items are simply grouped together if the densities are greater than a set threshold. Therefore our approach can cluster high-dimensional data space more quickly and accurately. In addition, the quality of the clustered result will not be influenced by the shape of the high-dimensional clusters or the order of the data input, and the parameters are easily set up and modified, so all the criteria mentioned in the Section 1 have been met. 4 Visualization Clustering high-dimensional datasets is used to help users to better understand the data structure and relationships. Visualization techniques play a very important role in displaying data and clustered results, making it more clear and reliable. The visualization techniques for high-dimensional dataset [10] [11] can be divided into two types. One type, such as “parallel coordinates” [12], is to divide the 2-d plane into several parts, each part representing an attribute. The other type is to reduce the dimensions, which is implemented through giving weights to the attributes of ndimensional data according to the relative importance and then combine them linearly. We present a new method to project high-dimensional data, which uses stimulation spectrum to project high-dimensional data on a 3-d space. The natural color is the summation of energy distribution in each wavelength in the range of visible spectrum. This energy distribution is called stimulation spectrum Φ(λ). Every stimulation spectrum can be transferred to a point in RGB color space. The quantitative relationship between stimulation spectrum Φ(λ) and RGB color coordinate are listed in the following formula: 6 Liu et al. R = k * å Φ (λ ) * r (λ ) * ∆λ λ G = k * å Φ(λ ) * g (λ ) * ∆λ (4.1) λ B = k * å Φ (λ ) * b(λ ) * ∆λ λ In this formula, k is the ratio. λ refers to wavelength of visible spectrum, ranging from 400nm to 700nm. r(λ), g(λ) and b(λ) stand for the spectrum tristimulus functions of red, green and blue, and the value of r(λ), g(λ) and b(λ) at every 5nm is measured by CIE, which is already known. Fig.3 is the spectrum tristimulus functions graph of CIE 1931 standard colorimetric system. Fig. 3 Spectrum tristimulus functions graph Each data item Xi =(Xi1, Xi2, …, Xij, …, Xin) in high-dimensional space can be viewed as a stimulation spectrum, and spreads equally in the range of visible spectrum with the wavelength from 400nm to 700nm. Here Xi can be regarded as the function of λ, and the change of the attribute values corresponds to that of the spectrum tristimulus. For example, Xi (400) = Xi1, …, Xi (700) = Xin. We can work out the 3-d coordinate of data Xi in projection space according to formula (4.2), and through adjusting the value of k can make projection space not only the RGB color space, but also any 3-d space. x = k * å Xi (λ ) * r (λ ) * ∆λ λ y = k * å Xi(λ ) * g (λ ) * ∆λ (4.2) λ z = k * å Xi(λ ) * b(λ ) * ∆λ λ We can see from the above that the process of projecting the data items as stimulation spectrums can also be viewed as a kind of weight linear combination of ndimensional data through spectrum tristimulus functions r(λ), g(λ) and b(λ). Taking advantage of spectrum tristimulus functions to convert high-dimensional data can completely project the data in projection space. This is because the fundamental Clustering by Ordering Density-Based Subspaces 7 function of r(λ), g(λ) and b(λ) is to project stimulation spectrums in color space, so it can well reflect the feature of the original data. From fig.3, we can see that the n attributes can be divided into 3 parts, and b(λ) corresponds to the attributes of the former part of the data while g(λ) and r(λ) mainly correspond to the middle and the last part of the data attributes. In this way, the projection result of a data item will be described by all of the 3 coordinate values. On the other hand, because the spectrum tristimulus functions cover the equal area, the ranges of the coordinate axes in projection space are equal, and the data will not be over-concentrated around some coordinate axes. 5 Experiments A 6-dimensional dataset containing 400 points was used in our CODBU testing experiment (it is a car dataset from http://stat.cmu.edu/datasets/). The attributes in the data set are: fuel economy in miles per U.S. gallon, number of cylinders in the engine, engine displacement in cubic inches, output of the engine in horsepower, 0 to 60 mph acceleration, and vehicle weight in U.S. pounds. Each attribute was divided into 5 parts, so there were 56=15,625 units and the 400 points scattered in 53 units. The threshold was set up as 1 which means all the points were processed. 7 clusters were obtained through linking the associated units. Two visualization techniques were explored in displaying the result: parallel coordinates and the spectrum tristimulus functions projection. Fig.4 shows the result using parallel coordinates, in which different clusters are represented by different colors but the characteristics of the clusters are not obvious. Fig. 4 Visualization of the clusters using parallel coordinates Fig.5 shows the result using the spectrum tristimulus functions projection. The side length of the projection space is 200, ∆λ=5nm, and r(λ), g(λ), b(λ) are given by CIE. In fig.5 (a), the values of the densities are reflected by color: the darker the color, the higher the density. 3 clusters are found. Fig.5 (b) displays the result of clustering using our algorithm; clusters are represented by different colors. 7 clusters are obtained based on the previous 3 big clusters. This is because that there are 7 density peaks being discovered. The clusters in white are removed due to only one data point 8 Liu et al. included. Fig.5 (c) uses symbols (such as *, +, o, ^, etc.) instead of colors to display the clusters of the data. (a) (b) (c) Fig. 5 Visualization of the clusters using the spectrum tristimulus functions projection. Shades in (a) reflect the different density, and different colors in (b) represent the different clusters, while in (c) the different clusters are represented by the symbols instead of the colors. 6 Conclusions The paper presents a new clustering approach by sorting density-based units. The basic idea is to rank the units in high-dimensional data space according to the values of the density, and start from the highest density unit to search for the density peaks in order to discover clusters. The experimental results are very promising. Clusters extend from the density peaks to density valleys and this will not be affected by the Clustering by Ordering Density-Based Subspaces 9 shape of data items. Arbitrary clusters can be obtained through our approach. We also propose the method of projecting high-dimensional data on the basis of the spectrum tristimulus functions, by which the data and its distribution can be displayed in the 3dimensional space. The combination of the data mining and the visualization techniques makes it easier for users to observe and understand data, and make better use of the data to do prediction and decision. References 1. R. Agrawal, J. Gehrke, D. Gunopulos, P. Raghavan: Automatic Subspace Clustering of High dimensional Data for Data Mining Applications. Proc. ACM SIGMOD’98 Int. Conf. on Management of Data, Seattle, WA (1998) .94~105 2. M. Ankerst, M.M. Breunig, H. Kriegel, J. Sander: OPTICS: Ordering Points To Identify the Clustreing Structure. Proc. ACM SIGMOD’99 Int. Conf. on Management of Data, Philadelphia, PA (1999) 3. M. Ester, H.P. Kriegel, J. Sander, X. Xu: A density-based algorithm for discovering clusters in large spatial databases. Proc. 1996 Int. Conf. Knowledge Discovery and Data Mining (KDD’96), Portland, OR, Aug (1996) 226-231 4. J.W. Han, M. Damber: Data Mining: Concepts and Techniques. Morgan Kaufmann Publishers (2001) 5. L. Kaufman, P.J. Rousseeuw: Finding Groups in Data: An Introduction to Cluster Analysis. New York, John Wiley & Sons (1990) 6. R. Ng, J. Han: Efficient and effective clustering method for spatial data mining. Proc. Int. Conf. Very Large Data Bases (VLDB’94) (1994) 144-155 7. T. Zhang, R. Ramakrishnan, M. Livny: BIRCH: An efficient data clustering method for very large databases. Proc. ACM-SIGMOD Int. Conf. Management of Data (SIGMOD’96), Montreal, Canada, June (1996) 103-114 8. S. Guha, R. Rastogi, K. Shim: Cure: An efficient clustering algorithm for large databases. Proc. ACM-SIGMOD Int. Conf. Management of Data (SIGMOD’98), Seattle, WA, June (1998) 73-84 9. G. Karypis, E. H. Han, V. Kumar: CHAMELEON: A hierarchical clustering algorithm using dynamic modeling. COMPUTER, 32 (1999) 68-75 10. I. Herman, G. MelancËon, M.S. Marshall: Graph Visualization and Navigation in Information Visualization: A Survey. IEEE Transactions on Visualization and Computer Graphics, Vol.6, No.1, January-March (2000) .24~43 11. A. Buja, D. Cook, D. Swayne: Interactive High-Dimensional Data Visualization. Journal of Computational and Graphical Statistics, No.5 (1996) 78-99. 12. H. Siirtola: Direct Manipulation of Parallel Coordinates. Proc. of the IEEE International Conference on Information Visualization (IV2000), London (2000) 373378          Æ                                           !"#    $!$ & & &' ( ) *!!*                                                                                               !¾ "                                                                              !¿ "                                                                                 #                                                                                            -        &'$  (   7                          ''     +**℄2    5                %  !                            0                               +**℄  )                                    &()*$      3  4       (           '' )(1'*        6 &'$               . "#$%       /              "#$%                        "( 89        :1'* (($* *% 12 Schroeder and Katopodis     ℄              !   !    "  #  $ #     %# %  #  # %!   !           %    % &        !   ' %     &  &     &    #    (   !  %  # !              %  Æ  #   # &           &!) * ! %  %  + & #   !      &                &       !     +  ,-%'./℄ '+ ! %  %%! %                    (   0          # #      +       #  (    ! #   +         '+ !      1   #     &  Æ   #      #  * # !    !             & !      # &             2  &          +     %          # %         &     %  & '    # ,( 3℄ !    # #  % & + & %    &       +      " * /$ 4    + &   # &   #   # ,  5℄      5..   ! % & + 6    #      #        7 #   & +    2   %   %    +     /5  %           4      !  %  # !   + & 5.. ! " $   /5    "%$ '+ !     %   +#  8 #                 ¾ Ê    "   $ 9  "            "   $ 9                 $          "   $ *  + ! !                %  &        + % !  !  #  % :               !      %             %# ; Can Hierarchical Clustering Improve the Efficiency       « «             «                      13                                                    !"#$$% ℄ '             (           )  *    +                             -            .   ,                -        /    0        1           0                                                    )      )  /          (      )  (     (     2                 !3  "4  544 6"57%  8%%℄ 9     /       )   0              /                     ) 0      0   0          (     ) )            0 :                        ( ;  /      )   0      0          #                       *       :'<37= * (               + !#7, ℄  .   ) * !3  "4  8%%℄++                )    >     *  !3  "4  544 8%%℄+ 1              *¾ +           -      .        *¾ +   !2%,℄ =                 ?        )    )      $               (    (         ;         #     !5@ A$4% 55 % % ==-%@ 6%  %% $"%%℄   )       *  @+ =    -.B                       #               )      ?  2 14 Schroeder Schroederand and Katopodis         ℄       A A C B C B           !"    #   $   "    "          Can Hierarchical Clustering Improve the Efficiency                            *                                                 & ' (                                                                     )                             ! "#  "$℄                                                 +         *             2                                              *       6-7 !04/℄                                                                                     +                99       8  23                                          6-7     )               1                                1  )                   *               !99.4℄    1        )            !4/℄       99          23 -7           9  9       )                    +    *  6             )                                               -                        1         1                                  1 -                                                                                                   !-./℄  !2345℄              23          , -             0            3          1)                     )                                        : 15 16 Schroeder and Katopodis                                                                                                            !" #"$%&℄ (                    )   *               +   *     ,     *                  -      #"$%&℄ )           #%%℄ +                              .           (    / /                        .                        +    ,                   -* +                         *                                 0 01 2  ,  2,    #2%&℄                   +                     +     ( +               +                              +              3  +                     !"  2,             )      #%4℄             2,    5 + +         !"                    +           &0   #"$%&℄ ,      +            6      +      -    6                    " -     #(&% 71 ! 87 88 "98% : +%%℄                   +   +      ;  :                < =½       > ¾ Ê      < = > ¾ Ê      =   > <     !     => +        9             +               + +        ?      +  .   !     $     +    3 9  A C D $       2          3      !   =         @> 9         3 9 $   * +    3 9  $  2      B           (          *          & Can Hierarchical Clustering Improve the Efficiency                            ℄! "  #$ #              #     ! %  $        &  ! '  &       &  ! (     # )(*+, --℄          $         &    !    ./      "       #  $    #        0     $      ! '   1  2       "              3''44℄$         #    &       #      $     # & !  $           $      #    #! +                   5!  ! ! + #       $            $ .           ! !        $ !!  6 )½       ,           ! 5! 7                      8 6  9  $              # 8   ¼ :!  ;     )   , 6   )   , )   ,  6   )   , 6 5  )  , ¼ ¼     6   9  8  $ !!     # <   =$    ! > 17 18 Schroeder and Katopodis          £ ¾                                             ¿   Æ                                                            !                "                           #     $                     !      ¯ % &               ¯ ' &        ( ¯ )   &             *                              ¯ )   +                        )       , "           -                                                           )        .                                            -       $               #                   /                               -           0         1         2 ¾  &       3 ¯ 4   /                    (                                 *                2  ¯ 5)& &     5   )   & 5)& 67..℄   $       *                         $    "                          9       5$     *   $                        5                  $         ¾  9 Can Hierarchical Clustering Improve the Efficiency                                                                                                 ( #                      #     ! &'     " #  #   $  %#  &' &' &¾ ' &¾ ' &¿ '   )   ,  * +       19 20 Schroeder and Katopodis                     ¯                                                               ¯     !"#      #                   $   #           #                             #            %           ¯     !"#     #                                              #                $               &      $         '    #                  (    " )  #   # *& '#      #         #                                     #                +  #         ,  #      #     -           .   #                      /             0 1  %    &   !"#  2           #         /  '    #                          3 2 #                *& ' #    /            #     #         0           #                      &   !"#  &                   4        '      #          #    $        #       #                         ' $   -#     #                     ( +    $  #    /$          %         &   !! +   /$    #  /                   /             56#     7  869#            /$#        !!# #            !" Can Hierarchical Clustering Improve the Efficiency Stress per iteration 6000 Stress Circle Spanning tree Zero 5000 Random Clustering PCoA 4000 3000 2000 1000 0 0 5 10 15 20 25 30 Iteration                             Stress per iteration Stress per iteration 2600 400 Spanning tree Zero 2400 Random Clustering PCoA 2200 350 300 Spanning tree Zero Random Clustering Stress Stress 250 2000 1800 200 150 1600 100 1400 50 1200 0 0 2 4 6 8 10 12 14 16 1 2 3 4 5 6 7 8 9 10 Iteration Iteration                              !"       "  "         21 22 Schroeder and Katopodis                                        Stress per iteration 16000 14000 40 2% 4% 6% 8% 35 Oscilation threshold 18000 Threshold by Number of Nodes Stress 12000 10000 8000 6000 30 25 20 15 4000 10 2000 5 0 0 5 101520253035404550 0 Iteration 50 100 150 200 250 300 Number of Nodes                                !"      #  $        %     &        '      (        )             * +      ,-                              .                          (          /   -    +      *%  0%          (     /   /     /           /+  +     123 4567 22!7 #7 88973 #%7  77 5: 77℄    -   /     1<! :6! 266 =2567 >77℄   /             ?@8 1<+℄     )-   /  +    +  -      +       5   ?@8                                              (  +  +     A          /      +       +   -                    +         +   /               >   '             B      '   >  +  +      * Can Hierarchical Clustering Improve the Efficiency Stress per iteration 13000 12000 fixed temp temp as % 11000 10000 Stress 9000 8000 7000 6000 5000 4000 3000 2000 5 101520253035404550 Iteration                   ¯                       ¯          ¯           ¯                             ¯                        !      "                           " #       #                 !          $     %     "                         $% &%                 '(            )*&℄ "  #              #            ,         "             -   .   (  )-.(/℄    " "      0" )0/℄         "        1   "   ,  1  -      0" "  "        " " "   "    0"2               "    0           3    " ,     4         ,   0"             3  " "        3  $ 23 24 Schroeder and Katopodis                                                    ℄   ! "    #              $           %&&'(% )   * +,--℄ . + , ,  / -   $ -                0   * 1,2#&℄ 3 1   ) ,  2     340 *&     . #        $+5℄ 6  / 7 / $    0 8 +   ,                     95&:;&(';&; *5 ,&℄ 0 ,      9:*'*;( *& ,)++&℄ # + ,  0 - )   0 8 +     1 +    3       <      =>9=?:*&;%' *&;& *& , 5&℄ + ) ,  ,   +  #3;℄ 6+          /*℄ -  .  1 . &*℄                        ! *5&  #    3/       < <       @       $         5;'&5 )   *;      , /             "   .     <  !   9*:**9'**; **   3    *&* . ;;℄ 6 3 .   )               <         #  =%:%9='%& *;; 22&℄ - 2   ) 2       $      %  *:5'*= *& 2 ;℄ 6 2 ! #  <       2 55℄ 6 2 ! -    $  #2+5℄ 2 7 #  6 - 2    0  *5 #/3(9℄  #   9((9      #        @   9:*'95 *;     A           0  *55 "         6 # + &         . /    #  3  3             4   .  * Can Hierarchical Clustering Improve the Efficiency ℄ "#$℄     % & *    " ' ( '   ) * ' #  5 6 )     '  '   ,801%4℄ (  ,   )** / 2  *  3  ' - )   **  '  /0' 8+     *  :* 4!  . ' 3    5 9 0 1' ' 3 0   +    $$!  $          ,  *  *        74!74 . % 3            +  ' & /0'  ' / 1   ,.℄      !       ),-.. &/.℄              ''           !$ 44 ,(℄ & & ,  '   3 / (   * )    '  0-         ! "   74!7   ;7℄ / ;  ;   ) *    ** *    ' 2 '  '  *    (  <0 3 ,     * , -   7 ;℄ ;.4℄ ;. ℄ 64℄ / ;  6 ; ;  0:      4 6 ; ;       50   0  * 3  0 62 = 5 #%& * 7 *     * !7  = ,*     '  (    47! 4 .4 1 2  ' 2   *  '  (     $7 !$. . ) (0 ) ( 0  7 6+℄ 5      #$ )' 2 6++   '    -/  ' 8 * / 2-  #$ ) &   ' ' # *+++ (, * $! ,*   #  44          )  '   25                                                             !   "  " # $ %          ! " %& '  "  (  # $ %    "        $ # "   (  "     &   $ "  % "       (     ( "  $ #& (     !  %% " (  ( ")" (  % (      ! "! % % * %"  (   %   & +"  % %    "      "  (       ( % " ( " (%" ( "     & +" $ #    "        ,   %  %%   -%   $" (     "  (  " (  "     % # $ %&                                                        !    "!                               #    Æ   $              %                   &'℄                               )         )     *                   *                          +     ,           #                                             $                     +" $ #  (  ! " -     ,  ./             ! 28  Kimani et  al.                                                                                                                                                        !                  "                       #   $    %    &                          '           ( )            *   +                  %                               ,   -.℄       +    0 1+ 02 3   + 0   !4 455. ,                     6    7      6                                   )8   -!℄                                      "                     )8               9       6 "     %     :;<     -5℄  +9 "  = ,                       %                              6    &      +9                9 +        -*℄            >   +           &    +                     &   9                              >>  -44℄           8                &           ?>  -4@℄ A             >>             A Visual Data Mining Environment 29                         ℄          ! #!      !      !                 # %     '      /%$#  )      #     ! (                      /%$01   %  "    !    &℄       , (    $ %    +++   - ++ # .!       '    *   ,%# )          !  %                    '           (          ,%#  $    !              !      )2 "     ! !   3              ) !     4   ß      3    1  Æ  .            ß   !     8 9   .  .     )  $        $   $    !       !   )   !           )  5 6  7     7      !         : 3      )2    2    2              %   !  )2      !            )          !   #       .          !         !       .      3                * !  (  2   (      4  ( <(                )         ! ;    <(  )    !    ! 7 3  1     $ )       !       )     8 2) ! 9        !    <(                     4           .        !           <(     !     !            2) !      !       )     1      *    1        !     1        1 30  Kimani et  al. Parametric User Interface GUI Container Metaquery user interaction component Association Rule user interaction component Clustering user interaction component Abstract Data Mining Engine Metaquery Engine Association Rule Engine Metadata Clustering Engine Target Dataset                                                                                                                   ß        !     ß                          "    # ß $     ß             $       %           ß &       %   '(    ß &                )   "  % %      #                            !           (     "      # (           !                     "# *    %    !          (      (         )         A Visual Data Mining Environment 31                                                                       ß !                          "        # $        % ß !                         &                                                 '        $         (  " % )                (                               ß !                       '                             *         (     ß !      (      (        +'               (        *         (                         ,   *    ( -.        #                   , (     ( -.       -. / #          '          '      -.           '      !  #                 ( ,                                ( -. )            (          ½              !                      (    "     )  ( "     (          (                    01℄                   34     5                 ½                                        Kimani et  al. 32                                                                                              $               "     !    %       !            #  "                                            &    "                " " #     "                              %                   %     (  "    "             &   "                       &    "  !          &                "             #                )        #                 "              "     '&  %                %                  Æ    #    %       "     "   "   Æ                                               %&      (           "            "   "                                  &   +  ,-             "                                                & &     " "  "         %           *     %        "                               %            .     "     "         "            A Visual Data Mining Environment 33                                                                                 Æ                 !    " CLUSTERING OUTPUT CLUSTERS All/Any Customer Information 6 Some 2 String Name of Relation DETAILS Number of Attributes Attribute Type 3 The customer is located 1.73 miles North East of the Main Office Numeric Any Relation Alphanumeric 4 Logical 1 Date/Time Service Customer      Specification Space Database                       CLUSTERING INPUT Relation Type                                                5 X Cluster quality Target Space Save/Export Only Export + Switch To X Export Facility Standard Tools CustID CustX CustY CustAmt SEPARATION OF THE CLUSTER FROM THE OTHERS Level of Similarity Dissimilarity Separation ServID 10% 10% Number of Clusters Attribute for Labels 2     ß ß ß ß #  %     &        $       (       &                       )                                   '  (                               )             5             & '   4 3 Cluster quality Cluster density   $                     34  Kimani et  al.                                                                           !                                                         "              #            $    %                      &  '      (                            )     "                     *                  +           )                        "                         +        '   (       ,                                                & &          &         &        )      "                                          )                                                         "   $      &                                                    .          '              (                                       '           (                   ,  -              "                       &                    & &                                /                  A Visual Data Mining Environment 35                            Æ                                           Æ                                                                                    !                                   "                 #   $         #  $                                                                 %                                     "                           &  "            '                                                                                                     &                       (              )                                                         "                             *  +        #       $          "          ,*%  +¾   #     $            #     $ (                         )                                                  ¾ 36  Kimani et al. RULE VISUALIZATION DESIGN OF METAQUERIES Specification Space Relation Type SCATTER PLOT OF METARULES ServID All/Any 2 Some Support (%) 70 Numeric ServCent Alphanumeric Logical 60 50 40 30 20 A rule with the highest confidence 10 Date/Time ServCent CustAge CustGender CustID CentID 80 String Name of Relation VISUALIZATION OF TUPLES Number of Attributes Attribute Type Database 1 2 3 4 Number of items ServCent Target Space Male Female Confidence Save/Export Only Age Export + Switch To CustServ ServCent CustID Standard Tools Export Facility ServID ServID CentID DEDICATED VIEW OF METARULES ClustOut1 CustCent IF THEN CustID CentID Manual Joins Automatic Joins Add Pattern ServCent(ServID,CentID) CustServ(CustID,ServID) IF CustServ(ServID,CustID),CustCent(CustID,CentID) THEN ServCent(ServID,CentID) IF CustCent(CustID,CentID),ServCent(CentID,ServID) THEN CustServ(CustID,ServID) IF CustServ(CustID,ServID),ServCent(ServID,CentID) THEN CustCent(CustID,CentID) CustCent(CustID,CentID) 100 50 0 Sort Facility Minimum level of Confidence Support Rules Standard visualization tools 10% 10% Support Confidence                                                                                                           !            "           #  ! "         $           % %  &                $                             $     '      " (  )         *          #                                                   "         *            +  , (       ß ß         "   *                          A Visual Data Mining Environment 37                                                                                                             !            ! ! "       #          !       $ % + (                 & &   &     ' (((()  "    ' %**)             ' ,,,,)  "    ' -.)              ' .../)  "    ' /*)  0 1 (       2         0   2                      !                                                   !                                          #         " #                     3 4                       $         Æ           5      2             2            6        2              $                                    7           2      + 8        8         9          8                               9       7                                                                   :                    38  Kimani et al.                                                                                                                                                          !                                          "                    #$ %          DESIGN OF ASSOCIATION RULES Specification Space Relation Type Number of Attributes Attribute Type All/Any Database 2 Some String Name of Relation Numeric Service Alphanumeric Logical Date/Time Service Service Target Space Service ServID Prod_ID S001 S002 S003 S004 S005 ServName Product WWW-Access Printing Scanning Faxing Telephone Manual Automatic Define Joins Skip Joins IF Service = "WWW-Access" Minimum level of Confidence Support   THEN Service = X 10% 10%                            !                                       "         &                                      #                  "                             A Visual Data Mining Environment 39                                                               !                                                                                                                                                              "                            $                                 #       #                 %           "           &   '(℄ *            +     ,      +      -&      . &   $     #            $   "     +.        1&2  #   #    &3   * &3    -         #   /    &   4           $                            '℄   1&2#       Æ      ,  5&&2 (6  +   #               1&2 ""         *           5&&2                      #             +     &00   1&2   + +     "     &   * &               &     +   & + 0    0  -&00.  /        &    &     /                              5&&2 (6                                     7     #               40  Kimani et al.                                                                                                                    !                   !                 !     "                      !         #  $    ß              $!             %   &          ' !          %                                            !          (    !    "       "      "                   "          "    "                 ' !          %      ß         %   "    !          )  #  $       !             (          !          $!    '     *  +                ! !        &       (   !        *     +           "    "        ß                    ,                                                       ß      !      !    " )         #  $      !    "             "      "    "    -      ! ß      !                #  $      !              "     "    ß                           !                          !        *      +    !  !    )  #  $    '     ! !   *                +   !          A Visual Data Mining Environment 41                                                         !℄  #   $℄ %                                         &                        '     (                          )                #            *                            -           *          ,       4    '                        (    +                                                     (     '  ,   ' .     /)01     2℄  3℄     %    ' ,         )             /     +            5                               +          '   ,          ) /                                                 !       " #       "  %&' ())( ( *  + *  % *   "  ,    - , %. - %  ,    , *     ,  '  + '   /. 0/( 1 (     '1        2    ,   / %3       ()) 0 '4 * '4  '    - 4 5   /  + ,'  ,  6   1 6& %   " / '   ..  *    + '   /. *78)9 :     778 9 +'    . % - () 1 %    4. -          ,   42      / 3 4 5    Kimani et al.                           ! " # ! $! %!  &' (   ) *  + ,-   - & .*   -  $  !   *+  -       ,-        ! / ' " 0 1      2*,    +'-                    ! 6 !  !  ! ' !  # !  1!  #  #*  - " 72    " + *8 7, !   *+  -    ,  "' ,,    9(6 * 2)  222  !   ! 6 !  !   6- ,  ,  - " 72  !   *+  - /  :"       +  "     **!  ! -+!  (  !  )! ; ! , $!  6-  * "+ ;'  * -    < "+ ;' ½ = > ? / ✁✄✂✆☎✆✝✞☎✆✟✡✠☛✟✞☞✍✌✏✎✒✑✔✓✔✠✕✟✔✖✗✟✙✘✚✟✔✛✗✜✗✢☛✣✍✓✞✘✚✜✥✤✏✂✆✟✡✁✧✦✗✎✥★✏✎✗✩✔✟✫✪ ✬✫✭✯✮✱✰✳✲✴✲✴✵✷✶✞✸✺✹✴✻✽✼✳✾✫✿✷❀✥❁✕❂✽❃❄✿✷❅✫❆✕❅✫❁❇❅✫❈❉✹❊❈✫❂✱❋✷❈✒✻✽✹❊❈✥❂✱❋✷●✫❍✗❆☛■ ❆✕❏▲❑▼❈✫❂✱❍✗❁❖◆P■◗✿✷❆✕❘☛❘❖❙ ●✫❚✯❙ ❚✫❋ ❯ ✮✱✾✯❃◗❋◗✿✷❋✷●✫❋✷❈❉❱❲❆☛❘✕✿✷❈✫✾✫❆✕■❖❅✫❁❇✸❳❃❄❋✷❆☛❋✷❨▼❃❄❋✷✿◗❘✕❆❖✻❩✹❊❈✫❂✱❋✷●✥❍✥❆✕■ ❑▼❈✥❆☛❈✯❙ ❚✫❈✫❘✕❆❖❃❬◆P✿◗✾✥❁❖❙ ❚✥❋ ❭ ✼❲✾✫✿✷❀✗❁☛❂✽❃❄✿✷❅✫❆✕❅✫❁❇❅✫❈❉❪❫✿✷✾✥❴✫❈✯✻❩✹❊❈✥❂✱❋◗●✥❍✥❆✕■ ❚❩❑✍❆❵◆✡❅✫✿▼❙ ●✥❏❛✿✷✾✥❴✫❈✯❙ ❚✫❋ ❜✳❝✞❞✍❡✱❢☛❣✗❤☛❡✱✐ ✰❥❃✍❃❄❈✫❘✕✿◗❆✕❋✷✿◗❈✥✾❦❂✱●✫■✷❁❧❁✕✾✫❍✗✿✷✾✫❁❖❃✳❋✷♠♥❚✫✿✷❘✕❆☛■✷■✷♠❫❈✥●✫❋✷❚✫●✥❋❳❆❧❀✗❁☛❂✱♠♦■✷❆☛❂✱❍✗❁♣❃◗❁☛❋▲❈✫q❲❂✱●✥■✷❁❖❃❬❙ r ❖❁ ❃❄❚✥✿◗❋✷❁❧❋✷❴✥❁❧qs❆☛❘✕❋❳❋✷❴✥❆☛❋❇❆❖❃❬❃❄❈✥❘☛✿✷❆☛❋✷✿✷❈✫✾❫❂✱●✫■✷❁❖❃❇❆☛❂✱❁✙❂✱❁✕❍✥❆✕❂✱❅✫❁✕❅t❆❖❃❇❴✫✿✷❍✥❴✥■✷♠✉❘✕❈✫❏❉❚✫❂✱❁✕❴✫❁☛✾✒❃❄✿✷✈✫■✷❁ ❆✕✾✫❅❥●✒❃❄❁☛qs●✥■❊qs❈✥❂✇❅✥❆☛❋✷❆①❏❉✿◗✾✥✿✷✾✫❍✧❆✕✾✫❅✧❅✥❁☛❘☛✿▼❃❄✿✷❈✥✾♦❃❄●✫❚✥❚✫❈✥❂✱❋✔✿✷✾✧qs✿✷❁✕■◗❅✒❃✴❃❄●✫❘✕❴✧❆❖❃❊❏❉❆☛❂✱②✗❁☛❋✷✿✷✾✫❍❊✻☛❂✱❁☛③ ❋✷❆✕✿◗■▼✻✯❏❛❁✕❅✫✿✷❘☛✿✷✾✥❁❖✻✯❅✫❁☛❏❉❈✫❍✗❂✱❆☛❚✥❴✫✿✷❘❖❃✍✻✺❆✕❏❛❈✥✾✫❍❦❈✥❋◗❴✥❁☛❂✽❃❬✻✒■✷❁☛✾✥❍✥❋✷❴✥♠t❈✫●✥❋✷❚✫●✫❋▼❃❲❏❉❆☛♠t❅✥✿▼❃❄❘☛❈✥●✫❂✱❆☛❍✗❁ ●✒❃❄❁☛❂✽❃❉qs❂✱❈✫❏④●✯❃❄✿✷✾✥❍⑤❋✷❴✥❁❫❋✷❁☛❘✕❴✫✾✥✿◗⑥✥●✫❁❵❙✚✮✱✾⑤❋◗❴✥✿▼❃❉❚✫❆✕❚✫❁✕❂✧⑦✳❁✉❚✫❂✱❈✥❚✫❈✯❃◗❁❫❆✉❚✫❈✒❃❄❋✷③✱❚✫❂✱❈✫❘✕❁❖❃❬❃◗✿◗✾✥❍ ❏❉❁☛❋✷❴✥❈✫❅✫❈✥■✷❈✫❍✗♠P❆✕✾✫❅✡❋◗❈✥❈✫■✺qs❈✥❂✴✈✫❂✱❈✫⑦✧❃❄✿✷✾✫❍✗✵✺❀✥✿▼❃❄●✥❆☛■✷✿✷⑧☛✿✷✾✫❍P■✷❆☛❂✱❍✗❁⑨❃❄❁✕❋✍❃✚❈✫q⑩❆❵❃❬❃❄❈✫❘✕✿✷❆☛❋✷✿✷❈✫✾✙❂✱●✥■✷❁❖❃❬❙ ❶ ❴✥❁✄❏❉❁☛❋✷❴✫❈✥❅✡✿▼❃✺✈✫❆❖❃◗❁☛❅P❈✫✾P❆❧❃❄❁✕❋✯❈✫q✇❈✫❚✫❁✕❂✱❆☛❋✷❈✫❂✽❃✺❋✷❴✫❆☛❋✯❋✷❂✱❆✕✾✯❃❄qs❈✥❂✱❏❷❃❄❁☛❋▼❃✚❈✥q❸❂✱●✥■◗❁❵❃✚✿✷✾✫❋✷❈t❃❄❁☛❋▼❃ ❈✥q✇❂✱●✥■✷❁❖❃❬✻❹❆☛■✷■✷❈✫⑦❳✿◗✾✥❍❥qs❈✫❘✕●✯❃❄✿✷✾✥❍❥❈✫✾❺✿✷✾✫❋✷❁☛❂✱❁❵❃❄❋✷✿◗✾✥❍❥❂✱❁☛❍✗✿✷❈✫✾✯❃✯❈✫q✚❋✷❴✫❁❻❂✱●✫■✷❁✡❃❄❚✥❆☛❘☛❁❵❙♥✸✺❆✕❘☛❴❫❃❄❁☛❋✔❈✫q ❂✱●✥■✷❁❖❃❸❘☛❆✕✾❧✈✫❁❛❋✷❴✫❁☛✾✡❅✫❁✕❚✫✿✷❘☛❋✷❁☛❅✡⑦✳✿✷❋✷❴✙❅✥✿◗q✆qs❁☛❂✱❁✕✾✫❋✺❍✥❂✱❆✕❚✫❴✥✿◗❘✕❆☛■✺❂✱❁☛❚✥❂✱❁❖❃❄❁✕✾✫❋✷❆☛❋✷✿✷❈✫✾✒❃❬❙ ❶ ❴✥❁✄❋✷❈✫❈✥■✚✿▼❃ ⑦❳❁☛✈✥③✱✈✫❆❖❃◗❁☛❅❉❆☛✾✫❅❉●✯❃◗❁❖❃❸❼❹❽①❾❥❙ ❶ ❴✫❁❇✿✷✾✫❚✫●✥❋✚❃❄❁☛❋❖❈✥q❊❆❵❃❬❃❄❈✫❘✕✿✷❆☛❋✷✿✷❈✫✾❉❂✱●✫■✷❁❖❃✏✿✍❃✏❍✥✿✷❀✗❁☛✾❉✿✷✾❛✹✔❪❫❪✉✭❇❙ ❿✧➀☛➁✥➂❉➃ ❢☛➄✏❞❹➅ r ❆☛❋✷❆❳❏❉✿✷✾✫✿✷✾✫❍❊✻✽❆❵❃❬❃❄❈✥❘☛✿✷❆☛❋✷✿✷❈✫✾❉❂✱●✫■✷❁❖❃✍✻✽❚✫❈✒❃❄❋❖❚✫❂✱❈✥❘☛❁❵❃❬❃❄✿✷✾✫❍❊✻✽❅✥❁☛❘☛✿▼❃❄✿✷❈✥✾❧❃❄●✥❚✫❚✥❈✫❂✱❋▼✻✽❀✗✿▼❃❄●✫❆✕■◗✿✷⑧✕❆☛❋✷✿✷❈✫✾✯❙ ✁✧✢✕✢✽✟✞➆✥☎❄✜✥➇❄☎✆✟✞➈✉➉①✤✏✂❄✎✡➊▼✁✧➉①➋✳✩✞☎✆✢✽➆✥✟✞★✏✎✥☞▼➌⑨➊✍✁✧✌✏☞▼✜✥➍❛✜✥✂✇✎✥➇✇✜✥✂❬➎✫➏✔➐✞➋❳☎❄✢❻➑✙✜✗➈✞➌✙➇✆☎✆➑✡✎✥✢❻✤✏✢✽✎✥✩✞✓✯➒▼✟✞☞❇✩✔✎✗➆✥☎❄✢✕☎❄✟✔➈ ✢✕✤✏✝✞✝✔✟✞☞✍➇❄✓✧☎✆➈➓✩✞✜✗➇✆✜➔➑✙☎❄➈✏☎✆➈✏✌➓✜✗✝✔✝✞✂✆☎❄➆✥✜✗➇✆☎❄✟✔➈✏✢♣✂✆☎❄→✏✎➔➑✙✜✗☞✍→✏✎✥➇❥➣✞✜✗✢✕→✏✎✗➇❥✜✥➈✏✜✗✂✆➌❖✢✽☎✆✢✽✓✄➑✙✜✗☞✍→✏✎✗➇✆☎❄➈✞✌✏✓❛☞✍✎✗➇✆✜✗☎✆✂✆✓ ✢✕➇❄✤✞✩✞➌♦✟✔➒❳➆✥✎✗➈✏✢✕✤✏✢❛✩✔✜✗➇✆✜✗✓✇✜✗➈✏✜✥✂❄➌❖✢✕☎❄✢✄✟✞➒❳➑✡✎✥✩✞☎✆➆✗✜✗✂❸✩✞✜✥➇❄✜✥✓✺✜✗➑✙✟✞➈✞✌♦✟✞➇✆↔✏✎✗☞✍✢✫➎❖↕❇↔✏☎✆✢❉➇✆➌❖✝✞✎✡✟✞➒❇→✞➈✏✟✞➍✄✂❄✎✥✩✞✌✏✎ ✩✔☎❄✢✕➆✗✟✔★✏✎✗☞✍➌♣☎❄✢✡✜✗✩✔✎✗➙✞✤✞✜✗➇✆✎✉➍✄↔✏✎✗➈➛➇✆↔✏✎❫✩✞✜✗➇✆✜✉➑✙☎✆➈✏☎❄➈✞✌➛➇❄✜✥✢✽→♣↔✏✜✥✢❺➈✏✟➛✢✕☎❄➈✞✌✏✂❄✎✉➆✗✟✞➈✞➆✗☞✍✎✗➇✆✎♦✟✞➣♥➜➝✎✥➆✗➇✆☎❄★✞✎♦➇❄✟ ➒▼✤✏✂✆➒➞☎✆✂❛➊✍✢✽✤✏➆✥↔➟✜✗✢✉↔✏✟✞➍➠➇❄✟➟✩✔☎❄✢✕➆✗☞✍☎❄➑✙☎✆➈✏✜✗➇✆✎➛✌✏✟✔✟✞✩➟➆✥✂❄☎✆✎✗➈✏➇✆✢♦➒▼☞▼✟✔➑➡➣✞✜✥✩➟✟✞➈✞✎✗✢✽➋✍✓①➆✗✟✔➈✏➇✆☞▼✜✥☞▼☎✆✂❄➌➢➇✆✟➟➍✄↔✏✜✗➇ ↔✞✜✗✝✞✝✔✎✗➈✏✢t☎✆➈➤➆✗✂✆✜✗✢✕✢✽☎✆➒➞☎✆➆✗✜✥➇❄☎✆✟✞➈➥✟✞☞❧☞▼✎✥✌✏☞✍✎✗✢✽✢✕☎❄✟✔➈✇➎✴➦✍➈✏✢✕➇❄✎✥✜✗✩✞✓❛➇❄↔✞✎➧✤✏✢✕✎➧✟✔➒✡✁✧➉➨✜✥✂❄✂✆✟✞➍✄✢✉➇✆↔✏✎⑤✩✔✎✗➆✥☎❄✢✕☎❄✟✔➈ ➑✙✜✗→✞✎✗☞✍➩❇→✏➈✏✟✔➍❛✂✆✎✗✩✔✌✏✎♦✢✕✎✗✎✗→✞✎✗☞❉➇✆✟➛↔✏✜✥★✏✎♦➑✙✜✗➈✏➌❦✩✔☎❄➒▼➒➞✎✗☞✍✎✗➈✞➇❇★✏☎✆✎✗➍✄✢❺✟✞➈♣➇✆↔✏✎♦✩✔✜✗➇✆✜❊➎❊↕❳↔✏✎✗☞✍✎♦➑✙✜✗➌❦➣✞✎♦✜ ✢✕✎✗➇✇✟✞➒▲✌✏✎✥➈✏✎✗☞✍✜✗✂✚✌✏✟✞✜✥✂❄✢①✝✔✟✞✢✽✢✕☎❄➣✔✂❄➌✙➈✏✟✔➇✺➑✡✎✥✜✗✢✕✤✏☞▼✜✥➣✞✂✆✎❥➊▼✂✆☎✆→✏✎❥➫♥➍✄↔✏✜✗➇✺➆✥↔✏✜✗☞✍✜✗➆✥➇❄✎✥☞▼☎✆✦✗✎✥✢①✜❥✌✏✟✞✟✔✩⑨➆✗✂✆☎❄✎✥➈✏➇✆➭✥➯✥✓ ✬ ❶ ✫❴ ✿▼❃✴⑦✳❈✥❂✱②⑨✿✍❃❉❃❄●✥❚✫❚✫❈✥❂✱❋✷❁☛❅♦✈✫♠❧❋✷❴✥❁P✸✺●✥❂✱❈✫❚✫❁✕❆☛✾⑨✼❲✾✥✿✷❈✫✾⑨❍✗❂✱❆☛✾✫❋✇✮✽❼ ❶ ③✽➲❩➳✫➳✥➳✫③✽➲✫➲✥❙ ➵✫➳✒➸⑨❼♥❈✥■✷③✱✸✚●✥③✱❱❲❁☛❋✇❆☛✾✥❅⑨❋◗❴✥❁ ✹✔➺❺❼❹✮✱✵◗➻✥➼✫➼✒➲✽✵✷✲✴■✷❆❖❃❬❃✯✹❊❂✱❈❩❑✍❁✕❘☛❋▲❃◗❚✫❈✫✾✒❃❄❈✫❂✱❁✕❅❺✈✫♠✧✶✞●✫✾✥❅✫❆☛➽✕➾☛❈❥✲✴✿✷➚☛✾✥❘☛✿✷❆①❁ ❶ ❁☛❘✕✾✫❈✥■◗❈✥❍✥✿✷❆❵✻❹✶✞✸ r ✸✺➪➧❁①✹✔❂✱❈✫❍✗❂✱❆☛❏❉❆ ❅✥❁❳✶✞✿✷✾✫❆✕✾✫❘☛✿✷❆✕❏❛❁✕✾✫❋✷❈❛✹✔■✷●✫❂✱✿✷❆☛✾✥●✫❆☛■❖❅✥❁❳✼✳✾✫✿✷❅✫❆✕❅✫❁❖❃✏❅✫❁❇✮✒➶ r ❙ 44 Jorge et al. ➹✗➘❛➴✏➷✆➬✗➴➥➷❄➮✙➱✞✃✔❐▼❒✆❮✗❰✞❒✄Ï✞❐▼✃✔Ð✏➱✞Ñt✃✞Ò✙➬✗Ó✆➷❄Ô✥❰✏❒✆Ñ❫Õ✔✃➥Ö✙➴✏❮✗×✞Ô✗Ø✫Ù✗Ú❻➹♥➘✄➴✏➷✆➬✗➴➟➱✔❐▼✃✔Õ✞Ð✏➬✥❒❄Ñ❫Õ✞✃➥➘❛➴✏➷✆➬✗➴➤➬✗Ó✆➷❄Ô✥❰✏❒❄Ñ ❒✆Û❖➱✞➷✆➬✗❮✥Ó❄Ó✆Û♣Ü✞Ð✞Û❖Ø✥Ù✥ÝßÞ✒à➢✃✔❐▼Ô✥✃✞×✏Ô✥❐▼Ú▲❒✆➴✏Ô❫Õ✞Ô✗➬✥➷❄Ñ✕➷❄✃✔❰➛➮✙❮✗á✏Ô✥❐❛➮✙❮✗Û♣Ô✥×✏Ô✗❰➛Ò▼➷❄❰✞Õ⑤❐✍Ô✗Ó✆Ô✗×✏❮✥❰✏❒❳➱✔❮✗❒✆❒❄Ô✥❐▼❰✞ÑP❒✆➴✏❮✗❒ Õ✔✃♦❰✏✃✔❒✚➬✗✃✔❐▼❐✍Ô✗Ñ✕➱✞✃✔❰✏Õ⑨❒❄✃⑨❮✥❰✏Û✙â✞Ð✏Ô✥Ñ✽❒✆➷❄✃✔❰✙Ò▼✃✞❐✍➮✡Ð✞Ó❄❮✥❒❄Ô✥Õ⑨Ü✞Ô✗Ò▼✃✞❐✍Ô✗➴✏❮✥❰✏Õ❸Þ✥ã❳➴✏➷❄Ñ①Ñ✕❒❄Û❖Ó✆Ô❥✃✞Ò✴Õ✔❮✗❒✆❮❥➮✡➷✆❰✏➷✆❰✏Ï❧➷❄Ñ Ñ✕✃✞➮✙Ô✗❒✆➷❄➮✙Ô✗Ñ✳➬✗❮✥Ó❄Ó✆Ô✗Õ✙➹♥Ò▼➷❄Ñ✕➴✏➷❄❰✞Ï✏Ù❛ä✍Ò➞✃✔❐⑩á✏❰✏✃✞➘✄Ó✆Ô✗Õ✞Ï✞Ô✗Ý❹Þ å Ð✞Ô❫❒✆✃⑤❒✆➴✏Ô✉Õ✔❮✗❒✆❮✉➬✥➴✏❮✗❐✍❮✗➬✥❒❄Ô✥❐▼➷✆æ✗❮✥❒❄➷✆✃✞❰➛✃✔Ü❹ç❩Ô✗➬✥❒❄➷✆×✏Ô✥ÑP✃✞Ò❛❒❄➴✞Ô✉❮✗Ñ✕Ñ✽✃✔➬✗➷✆❮✗❒✆➷❄✃✔❰➛❐▼Ð✞Ó❄Ô❫Õ✞➷✆Ñ✽➬✗✃✔×✏Ô✗❐✍Û♣❒✆❮✗Ñ✽á✞Ú è✧éêÕ✔➷❄Ñ✕➬✗✃✞×✞Ô✗❐✍Û⑨❮✗Ó✆Ï✏✃✞❐✍➷❄❒✆➴✏➮✙Ñ❉➱✔❐▼✃✔Õ✞Ð✏➬✥ÔP❮P➬✥✃✞➮✙➱✞Ó✆Ô✗❒✆ÔPÑ✕Ô✗❒❸✃✞Ò▲❐✍Ð✏Ó✆Ô✗Ñ❻❮✥Ü✞✃✞×✞Ô❺Ð✏Ñ✽Ô✥❐▼ë▼➱✞❐✍✃✞×✏➷✆Õ✞Ô✥Õ♦❒✆➴✏❐▼Ô✥Ñ✽➴✏ë ✃✔Ó❄Õ✔ÑPä✍❒❄Û❖➱✔➷❄➬✥❮✗Ó✆Ó❄Û♣➮✙➷✆❰✏➷❄➮✙❮✗Ó✳Ñ✽Ð✞➱✞➱✞✃✔❐▼❒✳❮✗❰✞Õ⑤➮✙➷❄❰✏➷✆➮✙❮✗Ó✳➬✗✃✞❰✞Ò➞➷✆Õ✞Ô✗❰✞➬✗Ô✗Ú✴Õ✔Ô✗Ò▼➷❄❰✏Ô✥Õ➛➷✆❰♣ì✺Ô✗➬✥❒❄➷✆✃✞❰♣í✔ÝßÞ❊ã❳➴✏➷❄Ñ ➷✆➮✙➱✞Ó✆➷❄Ô✥Ñ⑨❒❄➴✏❮✥❒❻❒✆➴✏Ô♣✃✞Ð✏❒✆➱✞Ð✞❒❻✃✞Ò❥Ñ✕Ð✏➬✗➴➢❮✗❰➢❮✥Ó❄Ï✏✃✔❐▼➷✆❒✆➴✏➮î➷✆Ñ⑨❮t×✞Ô✗❐✍Û⑤Ó✆❮✗❐✍Ï✏Ô❦Ñ✽Ô✗❒①✃✔Ò✧❐▼Ð✏Ó✆Ô✗Ñ✕Ú❳➘✄➴✏➷✆➬✗➴ï➬✗❮✥❰ Ô✥❮✗Ñ✕➷❄Ó✆Û⑨Ï✏Ô✗❒❸❒✆✃✉❒✆➴✏ÔP❒✆➴✏✃✞Ð✞Ñ✽❮✗❰✞Õ✞Ñ✽Ú✚✃✞×✏Ô✥❐▼➘✄➴✏Ô✗Ó✆➮✙➷❄❰✏Ï✉❒❄➴✏Ô✡Ð✏Ñ✽Ô✥❐ßÞ✫ã❳✃♦➮✙❮✗á✏Ô❺❒✆➴✏➷✆❰✏Ï✏Ñ❻➘✄✃✞❐✍Ñ✽Ô✥Ú✯❒❄➴✞Ô❺❒❄Û❖➱✔➷❄➬✥❮✗Ó ❮✥Ñ✽Ñ✕✃✞➬✗➷✆❮✗❒✆➷✆✃✞❰♣❐✍Ð✏Ó❄Ô⑨❮✥Ó❄Ï✞✃✞❐✍➷❄❒✆➴✏➮ð✃✔Ð✏❒❄➱✔Ð✏❒✆Ñ❥❒❄➴✏Ô⑨Ó✆➷✆Ñ✽❒▲✃✔Ò①❐▼Ð✏Ó✆Ô✗Ñ❥❮✥Ñ❥❮⑨Ó❄✃✔❰✏Ï❦❒❄Ô✥ñ✏❒▲ä✍Ô✗×✏Ô✥❰❦➷❄❰❦❒✆➴✏Ô⑨➬✗❮✥Ñ✽Ô⑨✃✞Ò ➬✥✃✞➮✙➮✡Ô✥❐▼➬✥➷❄❮✥Ó❳❒✆✃✞✃✔Ó❄Ñ✡Ó❄➷✆á✏Ô❫ì✺ò✚ì✺ì➤ó①Ó❄Ô✥➮✡Ô✥❰✏❒❄➷✆❰✏Ô✥Ý▼Ú✴❮✥❰✏Õ➛Ó✆❮✗➬✗á✞Ñ❺➱✞✃✔Ñ✽❒❇➱✔❐▼✃✔➬✗Ô✥Ñ✽Ñ✽➷✆❰✏Ï♣ä✍Ñ✽✃✔➮✡Ô✥❒❄➷✆➮✡Ô✥Ñ❺❮✗Ó✆Ñ✽✃ ➬✥❮✗Ó✆Ó❄Ô✥Õ✡❐✍Ð✏Ó✆Ô❛➮✙➷✆❰✏➷❄❰✞Ï✏Ý⑩Ò➞❮✗➬✥➷❄Ó✆➷✆❒❄➷✆Ô✗Ñ✳Ò➞✃✔❐❸➷✆❰✏Ñ✕➱✞Ô✗➬✥❒❄➷✆❰✏Ï✡❒❄➴✏Ô✄Ñ✽Ô✥❒✒✃✞Ò❸➱✞❐✍✃✞Õ✞Ð✞➬✗Ô✗Õ✙❐✍Ð✏Ó❄Ô✥Ñ✫Þ Ö✍❰♣❒✆➴✏➷❄ÑP➱✞❮✗➱✔Ô✗❐❛➘❛Ô♦➱✔❐▼✃✔➱✞✃✔Ñ✽Ô♦❮♦➮✙Ô✗❒✆➴✏✃✔Õ➛❮✗❰✏Õ⑤❒❄✃✔✃✞Ó❇Ò▼✃✞❐❻❒✆➴✏Ô⑨Ü✔❐▼✃✔➘❛Ñ✕➷❄❰✏Ï❦❮✥❰✏Õ♣×✏➷✆Ñ✽Ð✏❮✥Ó❄➷✆æ✗❮✥❒❄➷✆✃✞❰❦✃✞Ò ❮✥Ñ✽Ñ✕✃✞➬✗➷✆❮✗❒✆➷✆✃✞❰⑨❐✍Ð✏Ó❄Ô✥Ñ✫Þ✫ã❳➴✏Ô❺❒✆✃✞✃✞Ó✇❐▼Ô✥❮✗Õ✔Ñ❻Ñ✽Ô✥❒❄Ñ❻✃✔Ò▲❐✍Ð✏Ó❄Ô✥Ñ①❐▼Ô✥➱✞❐✍Ô✗Ñ✕Ô✗❰✏❒✆Ô✗Õ⑨➷✆❰❧❒❄➴✏Ô❥➱✔❐▼✃✔➱✞✃✔Ñ✽Ô✗Õ⑨Ñ✕❒❄❮✥❰✏Õ✞❮✥❐▼Õ⑨Ò▼✃✞❐ ➱✔❐▼Ô✥Õ✞➷✆➬✗❒✆➷❄×✞Ô♦➮✡✃✔Õ✞Ô✥Ó❄Ñ✕Ú✴ò✚à➢à➢ôõä å ❮✗❒✆❮♦à➢➷✆❰✏➷❄❰✞Ï♣ö❥❐▼✃✔Ð✏➱✞Ý❹Þ✔ã❇➴✏Ô⑨➬✥✃✞➮✙➱✞Ó✆Ô✗❒✆Ô⑨Ñ✽Ô✗❒❇✃✞Ò①❐✍Ð✏Ó❄Ô✥Ñ❥➬✗❮✥❰t❒✆➴✏Ô✥❰ Ü✔Ô♣Ü✞❐✍✃✞➘✄Ñ✽Ô✥Õ➔Ü✞Û⑤❮✥➱✞➱✔Ó❄Û❖➷✆❰✏Ïï❐▼Ð✏Ó✆Ô❦Ñ✽Ô✥❒①✃✞➱✞Ô✥❐▼❮✥❒❄✃✔❐▼Ñ❧Ü✔❮✗Ñ✽Ô✥Õ➢✃✞❰ï❒❄➴✏Ô❦Ï✏Ô✥❰✏Ô✗❐✍❮✗Ó✆➷✆❒❄Û➧❐▼Ô✥Ó❄❮✥❒❄➷✆✃✞❰ïÜ✞Ô✥❒❄➘✄Ô✗Ô✥❰ ➷✆❒✆Ô✗➮✙Ñ✽Ô✗❒✆Ñ✫Þ❖ã❇➴✏Ô✡Ñ✽Ô✗❒❸✃✞Ò❳❐▼Ð✞Ó❄Ô✥Ñ❉❐✍Ô✗Ñ✽Ð✞Ó❄❒✆➷❄❰✞Ï♦Ò➞❐✍✃✞➮÷Ô✗❮✥➬✗➴♦✃✔➱✞Ô✗❐✍❮✗❒✆➷✆✃✞❰♦➬✥❮✗❰♦Ü✔ÔP×✏➷✆Ô✗➘✄Ô✗Õ❫❮✗Ñ❉❮✡Ó❄➷✆Ñ✽❒❸✃✞❐❳➬✗❮✥❰ Ü✔Ô❛Ï✞❐▼❮✥➱✞➴✏➷✆➬✗❮✥Ó❄Ó✆Û❺Ñ✽Ð✞➮✡➮✙❮✗❐✍➷❄æ✥Ô✗Õ✙❒❄➴✞❐▼✃✔Ð✏Ï✏➴P❮✄❰✏Ð✏➮✙Ü✞Ô✥❐⑩✃✞Ò❸❒❄Ô✥➬✗➴✏❰✏➷✆â✞Ð✞Ô✗Ñ✫Þ ã❳➴✏➷✆Ñ✳➱✔❮✗➱✔Ô✗❐✴➷✆Ñ✳✃✔❐▼Ï✞❮✗❰✏➷✆æ✗Ô✥Õ✙❮✥Ñ❲Ò➞✃✞Ó✆Ó✆✃✞➘✄Ñ✫ø✫➘✄Ô✄Ñ✕❒❄❮✥❐▼❒✯Ü✔ÛP➷✆❰✏❒✆❐▼✃✔Õ✞Ð✏➬✥➷❄❰✏Ï✙❒✆➴✏Ô✧Ü✞❮✗Ñ✕➷❄➬✧❰✏✃✔❒❄➷✆✃✞❰✞Ñ✳❐✍Ô✗Ó✆❮✗❒✆Ô✗Õ ❒✆✃ù❮✥Ñ✽Ñ✽✃✔➬✗➷✆❮✗❒✆➷❄✃✔❰✡❐✍Ð✏Ó✆Ô✧Õ✞➷✆Ñ✽➬✗✃✔×✏Ô✗❐✍Û❖Ú❊❮✗❰✏Õ❧❮✗Ñ✕Ñ✽✃✔➬✗➷✆❮✗❒✆➷❄✃✔❰✡❐✍Ð✏Ó✆Ô✄Ñ✕➱✞❮✥➬✗Ô❊Þ✗ú➓Ô❛❒✆➴✏Ô✥❰PÕ✞Ô✥Ñ✽➬✥❐▼➷✆Ü✞Ô✄ò✚û▲è✧é①Ú✔❒❄➴✏Ô ➱✔✃✞Ñ✕❒✯➱✞❐✍✃✞➬✗Ô✥Ñ✽Ñ✕➷❄❰✏Ï✙Ô✗❰✞×✏➷❄❐✍✃✞❰✞➮✡Ô✥❰✏❒✯Ò➞✃✔❐❸❮✥Ñ✽Ñ✕✃✞➬✗➷✆❮✗❒✆➷✆✃✞❰P❐✍Ð✏Ó✆Ô✗Ñ✳❮✗❰✏Õ✙➷✆❒❄Ñ✳➷❄➮✙➱✞Ó✆Ô✗➮✙Ô✗❰✞❒❄❮✥❒❄➷✆✃✞❰✇Þ♥ú➓Ô❛Õ✔Ô✗Ñ✽➬✥❐▼➷✆Ü✞Ô ❒✆➴✏Ô♣Ñ✕Ô✗❒❻✃✔Ò❥✃✞➱✞Ô✥❐▼❮✥❒❄✃✔❐▼Ñ⑨➷✆❰➢➮✡✃✔❐▼Ô♣Õ✔Ô✗❒✆❮✗➷✆Ó❄Ú✳Ñ✽➴✞✃✞➘➨✃✞❰✞ÔtÔ✥ñ✏❮✗➮✙➱✞Ó✆Ô❦✃✞Ò✧❒✆➴✏Ô❦❮✗➱✞➱✔Ó❄➷✆➬✗❮✥❒❄➷✆✃✞❰ï✃✞Ò✧ò✇û▲è✧é①Ú ➬✥✃✞➮✙➱✞❮✥❐▼Ô✄➘❛➷✆❒✆➴P❐✍Ô✗Ó✆❮✗❒✆Ô✗Õ✙➘❛✃✔❐▼á✡❮✗❰✏Õ✙➬✗✃✔❰✏➬✗Ó✆Ð✏Õ✞Ô✥Ú✞❮✥Ó❄Ñ✕✃✡Ñ✕Ð✏Ï✏Ï✏Ô✥Ñ✽❒✆➷❄❰✏Ï✡❒❄➴✞Ô❛❰✏Ô✥ñ✏❒✒Ñ✽❒✆Ô✗➱✔Ñ❳✃✔Ò✇✃✔Ð✏❐⑩➘❛✃✔❐▼á✇Þ è✧❰♦❮✥Ñ✽Ñ✕✃✞➬✗➷✆❮✗❒✆➷✆✃✞❰♦❐✍Ð✏Ó✆Ô❲ü →ý ❐▼Ô✥➱✞❐✍Ô✗Ñ✕Ô✗❰✏❒✆Ñ❉❮✡❐▼Ô✥Ó❄❮✥❒❄➷✆✃✞❰✏Ñ✕➴✏➷✆➱✉Ü✞Ô✥❒❄➘✄Ô✗Ô✥❰♦❒✆➴✏Ô❺Ñ✽Ô✥❒❄Ñ❉✃✞Ò▲➷✆❒❄Ô✥➮✡Ñ⑩üþ❮✗❰✏Õ ý Þ û❇❮✗➬✗➴⑨➷✆❒✆Ô✗➮➓ÿ❥➷❄Ñ❉❮✗❰⑨❮✗❒✆✃✞➮❷❐▼Ô✥➱✞❐✍Ô✗Ñ✕Ô✗❰✏❒✆➷❄❰✞Ï⑨❒❄➴✏Ô❥➱✔❐▼Ô✥Ñ✽Ô✥❰✏➬✗Ô❥✃✔Ò✴❮❥➱✞❮✗❐✍❒✆➷❄➬✥Ð✏Ó❄❮✥❐▲✃✞Ü♥ç❩Ô✗➬✗❒➝Þ✥ã❇➴✏Ô❥❐✍Ô✗Ó✆❮✗❒✆➷✆✃✞❰❧➷❄Ñ ➬✥➴✏❮✗❐✍❮✗➬✥❒❄Ô✥❐▼➷✆æ✗Ô✥Õ♣Ü✞Ût❒❄➘✄✃♣➮✙Ô✗❮✗Ñ✕Ð✏❐✍Ô✗Ñ✫ø✯Ñ✽Ð✏➱✔➱✞✃✞❐✍❒▲❮✥❰✏Õ♣➬✗✃✔❰✏Ò➞➷✆Õ✞Ô✥❰✏➬✗Ô⑨✃✔Ò①❒❄➴✏Ô⑨❐✍Ð✏Ó✆Ô❊Þ✔ã❇➴✞Ô⑨Ñ✽Ð✏➱✔➱✞✃✞❐✍❒▲✃✔Ò✳❮ ❐✍Ð✏Ó✆✁Ô ê➘✄➷❄❒✆➴✏➷✆❰✡❮✧Õ✔❮✗❒✆❮✗Ñ✽Ô✥✄❒ ✂❫Ú✔➘✄➴✏Ô✥❐▼✁Ô ✂ ➷✆❒✆Ñ✽Ô✗Ó✆Ò❸➷❄Ñ✳❮❛➬✥✃✞Ó✆Ó❄Ô✥➬✗❒✆➷❄✃✔❰P✃✔Ò✇Ñ✕Ô✗❒✆Ñ❳✃✔Ò✇➷✆❒❄Ô✥➮✡Ñ✳ä✍✃✞❐⑩➷❄❒✆Ô✗➮✙Ñ✽Ô✥❒❄Ñ✕Ý▼Ú✔➷❄Ñ ❒✆➴✏Ô✡❰✏Ð✏➮✙Ü✞Ô✗❐✳✃✞Ò❳❒❄❐✍❮✗❰✞Ñ✽❮✗➬✥❒❄➷✆✃✞❰✞Ñ❉➷✆☎❰ ✂➠❒❄➴✞❮✗❒❸➬✗✃✞❰✞❒❄❮✥➷❄❰♦❮✥Ó❄Ó❸❒✆➴✏ÔPÔ✥Ó❄Ô✥➮✡Ô✥❰✏❒❄Ñ❛➷❄❰✧ü ∪ý Þ❵ã❇➴✞ÔP➬✗✃✔❰✏Ò➞➷✆Õ✞Ô✥❰✏➬✗Ô ✃✔Ò▲❒✆➴✏Ô❺❐✍Ð✏Ó❄ÔP➷❄Ñ❻❒✆➴✏Ô❺➱✔❐▼✃✔➱✞✃✔❐▼❒✆➷❄✃✔❰⑨✃✞Ò▲❒✆❐✍❮✗❰✏Ñ✕❮✗➬✗❒✆➷✆✃✞❰✏Ñ❻❒✆➴✏❮✥❒✚➬✗✃✔❰✏❒✆❮✗➷✆❰❛ü ➘❛➷✆❒❄➴❧❐✍Ô✗Ñ✕➱✞Ô✗➬✥❒✺❒✆✃⑨❒❄➴✏Ô❥❰✞Ð✏➮✡ë Ü✔Ô✗❐❳✃✞Ò▲❒✆❐▼❮✥❰✏Ñ✽❮✥➬✗❒✆➷❄✃✔❰✏Ñ❻❒✆➴✏❮✗❒✇➬✗✃✔❰✏❒❄❮✥➷❄❰❛ü♦Þ✫û▲❮✗➬✥➴⑨❐▼Ð✏Ó✆Ô❺❐✍Ô✗➱✔❐▼Ô✥Ñ✽Ô✗❰✞❒❄Ñ①❮❥➱✔∪❮✗ý ❒✆❒❄Ô✥❐▼❰❧➬✗❮✥➱✞❒✆Ð✏❐✍Ô✗Õ⑨➷✆❰✙❮❥Õ✔❮✗❒✆❮✗Ñ✕Ô✗❒➝Þ ã❳➴✏Ô❺Ñ✕Ð✏➱✞➱✔✃✞❐✍❒✚✃✞Ò❇❒❄➴✏ÔP❐▼Ð✏Ó✆Ô❺➷✆Ñ❻❒✆➴✏Ô❺➬✗✃✔➮✡➮✙✃✞❰✞❰✏Ô✗Ñ✕Ñ①✃✞Ò✴❒✆➴✏❮✗❒✺➱✔❮✗❒✆❒❄Ô✥❐▼❰✞Ú❊➘✄➴✏➷✆Ó❄Ô❥❒✆➴✏Ô❥➬✥✃✞❰✏Ò▼➷❄Õ✔Ô✗❰✏➬✥Ô❥➮✡Ô✥❮✗Ñ✽ë Ð✞❐▼Ô✥Ñ❳➷✆❒❄Ñ✳➱✞❐✍Ô✗Õ✔➷❄➬✥❒❄➷✆×✏Ô✄❮✗Ü✔➷❄Ó✆➷❄❒✆Û✚Þ ã❳➴✏Ô⑤➮✙✃✞Ñ✕❒❛➬✥✃✞➮✙➮✡✃✔❰➟❮✗Ó✆Ï✏✃✔❐▼➷✆❒❄➴✞➮ Ò▼✃✞❐✙Õ✞➷✆Ñ✽➬✗✃✔×✏Ô✗❐✍➷✆❰✏Ï➟è✧é➨Ò➞❐✍✃✞➮ ❮➛Õ✞❮✥❒❄❮✥Ñ✽Ô✗✆❒ ✂ ➷❄Ñ♦è✧ò✇é①✞Ö ❥ ✝ é①Ö ä✍è✧Ï✏❐✍❮✗➘✄❮✗Ó✯Ô✗❒✯❮✥Ó❬✠Þ ✟☛✡✔ÝßÞ✗ã❇➴✞➷❄Ñ❲❮✗Ó✆Ï✏✃✔❐▼➷✆❒❄➴✞➮ ➱✞❐✍✃✞Õ✞Ð✞➬✗Ô✗Ñ❲❮✥Ó❄Ó✯❒✆➴✏Ô✄❮✗Ñ✽Ñ✕✃✞➬✥➷❄❮✥❒❄➷✆✃✞❰✡❐▼Ð✏Ó✆Ô✗Ñ✳❒✆➴✏❮✗❒✒➬✗❮✥❰PÜ✔Ô❛Ò▼✃✞Ð✏❰✏Õ Ò▼❐▼✃✔➮❷❮✡Õ✞❮✥❒❄❮✥Ñ✽Ô✗☞❒ ✂➠❮✗Ü✔✃✞×✏Ô✡Ï✏➷✆×✏Ô✗❰♦×✞❮✗Ó✆Ð✏Ô✗Ñ❻✃✔Ò▲Ñ✽Ð✞➱✞➱✞✃✔❐▼❒✇❮✗❰✏Õ✉➬✗✃✞❰✞Ò➞➷✆Õ✞Ô✗❰✞➬✗Ô✗Ú✯Ð✞Ñ✽Ð✏❮✥Ó❄Ó✆Û❧❐▼Ô✥Ò➞Ô✗❐✍❐✍Ô✗Õ♦❒✆✃♦❮✗Ñ ✌✎✍✑✏☛✒✔✓✖✕ ❮✥❰✏Õ ✌✎✍✑✏☛✗✄✘☛✏✖✙ Þ✫è✧ò✚é①Ö✖✝❥é①Ö❇➴✏❮✥Ñ①➮✡❮✥❰✏Û✙×✏❮✗❐✍➷❄❮✥❰✏❒✆Ñ①➘❛➷✆❒❄➴❧➮✙✃✞❐✍Ô❥❮✗➱✔➱✞Ô✗❮✥Ó❄➷✆❰✏Ï❧➬✗✃✔➮✡➱✔Ð✏❒✆❮✗❒✆➷❄✃✔❰✏❮✗Ó ➱✔❐▼✃✔➱✞Ô✥❐▼❒✆➷❄Ô✥Ñ✽Ú✳Ñ✽Ð✏➬✥➴➧❮✥Ñ✡ò✇è✧é①ã❇Ö✍ã❇✖Ö ✝✛✚✢✜▼ì✺❮✗×✞❮✗Ñ✽Ô✥❐▼ÔtÔ✗❒❲❮✥Ó❬Þ Ý✍Ú å Ö✍óù✖ä ✣❻❐✍➷❄❰➧Ô✗❒❲❮✗Ó➝Þ Ý✧✃✞❐✧ì✺è✧à➢ò✚ô✴✤Ö ✚✉ö ä✍ã❇✃✔➷❄×✞✃✞❰✏Ô✥❰✏Ý▼Ú❛Ü✞Ð✏❒✧❒✆➴✏❮✗❒✧Ñ✕➴✏✃✞Ð✏Ó✆Õõ➱✞❐✍✃✞Õ✞Ð✞➬✗Ô⑤Ô✥ñ✏❮✗➬✥❒❄Ó✆Û➔ä✍➷❄❰➟❒✆➴✏Ô➧➬✗❮✗Ñ✕Ô⑤✃✔ÒPì✺è✧à➢ò✚ô✴Ö✤✚✉ö ➷❄❒✄➬✗❮✥❰➟Ü✞Ô ❮✥➱✞➱✔❐▼✃✔ñ✏➷❄➮✙❮✗❒✆Ô✗Ó✆Û❖Ý❉❒✆➴✏Ô♦Ñ✕❮✗➮✙Ô♦Ñ✽Ô✥❒❇✃✔Ò❻❐▼Ð✞Ó❄Ô✥Ñ❺Ñ✽➷✆❰✏➬✗Ô✉❒❄➴✏Ô✉Ô✗ñ✏❮✥➬✗❒❇Ñ✕Ô✗❒❳✃✞Ò❻❐✍Ð✏Ó✆Ô✗Ñ❺❒✆✃➛➱✞❐✍✃✞Õ✔Ð✏➬✗Ô⑨➷✆Ñ❥Õ✞Ô✥❒❄Ô✥❐▼ë ➮✙➷✆❰✏Ô✗Õ✙Ü✞Û❺❒✆➴✏Ô✄➱✞❐✍✃✞Ü✔Ó❄Ô✥➮ Õ✔Ô✗Ò▼➷❄❰✏➷✆❒✆➷❄✃✔❰P❮✗❰✞Õ✡❒✆➴✏Ô✄Õ✞❮✗❒✆❮❊Þ A Post-Processing Environment 45 ✥✧✦✩★ ✪✬✫✮✭☎✯✎✰✱✰✔✲☛✳✄✴✶✵✧✷✞✴✑✲☛✸✺✹✎✻✮✼✶✭☎✰✱✽✾✵☛✳✠✭ ✿ ❀☞❂❄❃✔❅☛❆✄❇✠❂❄❈✧❉✁❊✑❋✶❂✄●❍❃✔❂✠❋✑❃❏■❑❇✠❆✄▲◆▼☛❂❑❃✔❋✶❖✖P☞❇✄❋✶P☞❖✖❂✠◗❘❊✶▲❙❆❑❚✑❆✠❋✑❋✑❊✶❇✄❂❑❯❱❊✑❋✑❀❙❋✶❀☞❂ ❖✖❂✠❚✑❆✠❋✑❊✑❈☛▲❙▼☛❂✄❋✶❯☎❂✠❂✄▲❙❃✱❂✄❋✶❃❳❲ ❁ ✿❁❀☞❂◆❂✄●❍❅✧❋✶❨❩❊✑❋✶❂✄●❍❃✔❂✠❋ ❊✑❃❑❆✠❋❁❋✶❀☛❂◆▼☛❈✧❋✶❋✑❈☛●❬❈☛❉❭❋✑❀☞❂◆❚✑❆✠❋✑❋✶❊✑❇✠❂❘❆✠▲☞◗❪❋✑❀☞❂◆❃✔⊆❂✠❋❁❈☛❉❭❆✄❚✶❚❫❊✑❋✑❂✠●❴❃✱❂✠❋✑❃✎❆✠❋✁❋✶❀☞❂ ❋✑❈☛❅❏❲❵✿❁❀☞❂ ❖✞❂✄❚✶❆✄❋✶❊✑❈☛▲❍❆✠∅❚✑❃✱❈❴❇✠❈☛❖✖❖✖❂✠❃✱❅✧❈☛▲☞◗✧❃❁❋✑❈❴❋✶❀☞❂☎❛☞❂✄▲☞❂✠❖✖❆✠❚✑❊✶❋✑❨❑❖✖❂✠❚✑❆✠❋✑❊✶❈✧▲❄▼✧❂✠❋✑❯❱❂✄❂✠▲❄❊✑❋✑❂✠●❴❃✱❂✠❋✑❃❜❲ ✿❁❈✺❃✔❋✶❖✖P☞⊆❇✄❋✶P☞❖✖❂❴❋✶❀☞❂❴❃✱❂✄❋❏❈☛❉✛❖✞P☞❚✑❂✠❃✔❝✮❯☎❂❄▲☞❂✄❂✠◗❘❆❍▲☞P☞●❴▼☛❂✄❖❁❈✧❉✁❚✑❆✠❋✑❋✶❊✑❇✠❂✄❃✱❝✮❇✄❈☛❖✖❖✞❂✄❃✱❅☛❈✧▲☞◗☛❊✑▲☞❛◆❂✄❆✠❇✄❀◆❚✶❆✄❋✶❞ ❋✑❊✑❇✠❂❡❋✑❈❢❈✧▲☞❂❪❅✧❆✠❖✖❋✶❊✑❇✠P☞❚✑❆✠❖❍❊✶❋✑❂✠●❴❃✱❂✄❋❣❋✑❀☞❆✠❋❱❆✠❅✧❅☛❂✠❆✄❖✞❃❘❆✠❃◆❋✑❀☞❂❪❆✄▲☞❋✶❂✄❇✠❂✄◗☛❂✠▲☛❋✶❝✬❈✧❖❄❋✑❈❤❈☛▲☞❂❡❊✶❋✑❂✠●❴❃✱❂✄❋❣❋✑❀☞❆✠❋ ❈✧❇✠❇✄P☞❖✞❃❱❆✠❃❭❆❄❇✠❈☛▲☛❃✱❂✠✐✧P☞❂✠▲☛❋❥❲❜❦✮❈✧❖✁❂✄❧☞❆✠●❴❅☛❚✑❂✠❝♠❋✑❀☞❂❑❖✖P☞❚✑❂✺♥✔♦✧❝q♣r❝qs✉t☞✈✇♥✔①✧❝q②♠t③❝♠▼☛❂✠❚✑❈☛▲☛❛☞❃❭❋✑❈◆❋✶❯☎❈◆❚✑❆✠❋✑❋✶❊✑❇✠❂✄❃❜④ ❋✑❀☞❂❍❈☛▲☞❂❍❈☛❉✁❋✑❀☞❂❍❖✞P☞❚✑❂✠❃❱❯❱❊✑❋✑❀◆❆✠▲☞❋✑❂✠❇✄❂✠◗✧❂✠▲☞❋✬♥✔♦✧❝q♣r❝qs✉t③❝♠❃✔❋✶❖✖P☞❇✄❋✶P☞❖✖❂✠◗❘▼☛❨⑤❋✶❀☛❂❑❛☞❂✠▲☛❂✠❖✖❆✠❚✑❊✶❋✑❨⑤❖✞❂✄❚✶❆✄❋✶❊✑❈☛▲❙❈✧⑥☞❂✠❖ ❋✑❀☞❂◆❇✄❈☛▲☞❃✔❂✠✐✧P☞❂✠▲☞❋✑❝❏❆✠▲☞◗⑦❋✑❀☞❂❙❚✑❆✠❋✑❋✶❊✑❇✠❂❙❈✧❉✬❖✞P☛❚✶❂✄❃✎❯❱❊✑❋✶❀❤♥✔①r❝q②✉t❄❆✠❃✎❆❙❇✄❈☛▲☞❃✔❂✠✐☛P☛❂✠▲☞❋✑❝❏❃✱❋✑❖✞P☞❇✄❋✶P☛❖✞❂✄◗⑦▼☛❨⑧❋✶❀☞❂ ❛☛❂✠▲☞❂✄❖✞❆✄❚✶❊✑❋✶❨❄❖✞❂✄❚✶❆✄❋✶❊✑❈☛▲❍❈☛⑥☞❂✄❖⑨❋✑❀☞❂☎❆✠▲☞❋✑❂✠❇✄❂✠◗✧❂✠▲☞❋✑❃❁❈✧❉⑩❋✑❀☞❂☎❖✞P☛❚✶❂✄❃❜❲ ❶❷❂❑❇✄❆✠▲⑤⑥☞❊✑❂✠❯✇❋✑❀☞❊✑❃✬❇✠❈✧❚✶❚✑❂✠❇✄❋✶❊✑❈☛▲⑤❈☛❉✆❚✑❆✠❋✑❋✑❊✶❇✄❂✠❃✬❆✠❃✬❆✎❛☛❖✞❊✑◗☛❝✉❯☎❀☞❂✠❖✖❂✎❂✠❆✄❇✠❀⑤❖✞P☛❚✶❂✎▼✧❂✠❚✑❈☛▲☞❛☞❃✬❋✑❈❙❈☛▲☛❂✎❊✶▲☞❞ ❋✑❂✠❖✖❃✱❂✄❇✠❋✑❊✶❈✧▲❩❈☛❉✬❋✑❯❱❈⑦❚✑❆✠❋✑❋✶❊✑❇✠❂✄❃❜❲✧✿✁❀☛❂❙❊✶◗✧❂✠❆❙▼☛❂✄❀☞❊✑▲☞◗⑦❋✶❀☛❂❙❖✞P☞❚✑❂❙▼☛❖✖❈☛❯☎❃✱❊✑▲☞❛❩❆✠❅✧❅☛❖✖❈☛❆✄❇✠❀❩❯❱❂⑤❅✧❖✞❂✄❃✱❂✠▲☛❋✶❝⑨❊✶❃ ❋✑❀☞❆✄❋❫❋✶❀☛❂❙P☞❃✱❂✄❖❭❇✠❆✄▲⑧⑥☛❊✶❃✔❊✶❋❫❈✧▲☞❂❙❈☛❉✬❋✑❀☞❂✄❃✱❂❙❚✑❆✠❋✑❋✶❊✑❇✠❂✄❃☎❸✖❈☛❖✬❅✧❆✠❖✖❋✆❈☛❉❹❊✑❋✶❺✬❆✄❋✆❆❴❋✑❊✑●❍❂✄❝⑩❆✄▲☞◗❩❋✶❆✄❻☞❂⑤❈☛▲☞❂⑤❅☛❆✄❖✞❞ ❋✑❊✑❇✠P☞❚✑❆✠❖❏❊✑▲☞❋✶❂✄❖✞❃✔❂✠❇✄❋✶❊✑❈☛▲❍❋✶❈❴●❍❈✧⑥☞❂☎❊✶▲☞❋✑❈❴❆✠▲☞❈✧❋✶❀☞❂✄❖❏❚✶❆✄❋✶❋✑❊✶❇✄❂❱❸✖❃✱❂✄❋✉❈☛❉⑨❖✞P☞❚✑❂✠❃✔❺❼❲ ★ ④✄➀✾➁✁➂☎➃❢❃✱❇✠❖✖❂✠❂✄▲❄❃✔❀☞❈☛❯☎❊✶▲☛❛❄❃✱❈✧●❍❂☎❖✖P☞❚✶❂✄❃❜❲ ❽ ✴✶❾✧✻✮❿✠✭ ✿❁❈⑤❀☞❂✠❚✑❅⑤❋✶❀☞❂➄P☞❃✔❂✠❖✆▼✧❖✞❈✧❯❱❃✔❊✶▲☞❛❍❆❱❚✑❆✠❖✖❛☞❂☎❃✱❂✄❋r❈✧❉⑩❖✖P☞❚✑❂✠❃✛❆✠▲☞◗❴P☞❚✑❋✶❊✑●❴❆✠❋✑❂✠❚✑❨❑❉➅❊✑▲☞◗❴❋✶❀☞❂☎❃✱P☛▼☛❃✱❂✄❋✉❈☛❉⑨❊✶▲☞❋✑❂✠❖✖❂✠❃✔❋✶❞ ❊✑▲☞❛❙❖✖P☞❚✑❂✠❃✱❝♠❯☎❂❑◗✧❂✠⑥☞❂✄❚✶❈✧❅☛❂✠◗❘➀✾➁❫➂➄➃➆❸✞➀⑩❈☛❃✱❋⑩❅☛❖✖❈☛❇✄❂✠❃✱❃✔❊✶▲☛❛❙➁❫▲☞⑥☞❊✑❖✞❈✧▲☞●❴❂✠▲☞❋⑩❉➅❈☛❖✁➂☎❃✔❃✱❈✧❇✠❊✑❆✠❋✑❊✶❈✧▲❴➃✬P☛❚✶❂✄❃✱❺❵❲ ➀⑩➁❫➂➄➃❷❊✶●❴❅☛❚✑❂✠●❴❂✠▲☞❋✑❃❭❋✑❀☞❂❑❃✔❂✠❋✾❈✧❉❫❈☛❅✧❂✠❖✖❆✠❋✑❈☛❖✖❃✬◗☛❂✄❃✱❇✠❖✖❊✑▼☛❂✠◗❙▼✧❂✠❚✑❈☛❯✇❋✑❀☞❆✄❋✮❋✶❖✖❆✠▲☛❃✱❉➅❈✧❖✞●➇❈☛▲☞❂✎❃✔❂✠❋✮❈✧❉✆❖✞P☞❚✑❂✠❃ ❊✑▲☞❋✑❈❡❆✄▲☞❈☛❋✑❀☞❂✠❖✖❝❫❆✄▲☞◗❡❆✄❚✶❚✑❈☛❯☎❃❄❆✺▲☞P☞●❴▼☛❂✠❖❱❈☛❉❭⑥☛❊✶❃✔P☞❆✠❚✑❊✶➈✄❆✠❋✑❊✶❈✧▲⑦❋✶❂✄❇✠❀☞▲☛❊✶✐✧P☞❂✠❃❳❲r➀✾➁❫➂➄➃❄➉q❃❄❃✱❂✠❖✖⑥☞❂✄❖❣❊✑❃❑❖✞P☛▲ 46 Jorge et al. ➊☛➋☞➌☛➍✄➎✛➏✄➋❘➐☞➑✑➑✑➒✺➓✔➍✠➎✖➔☞➍✠➎❵→↔➣➙↕⑩➛❫➣➄➜✇➝✠➞✑➟✶➍✄➋☞➑❏➟✶➓❱➎✞➊☛➋◆➠☛➋◆➏❍➡❱➍✄➢❘➢✧➎✞➠✧➡❱➓✔➍✠➎❵→❳➣➄➞✶➑✑➐☞➠☛➊☛➤☞➐◆➋☞➠✧➑⑩➝✄➊☞➎✞➎✖➍✠➋☛➑✶➞✑➥ ➟✑➦❴➒☛➞✑➍✠➦❴➍✠➋☞➑✑➍✠➌✧➧☛➦❴➊☞➞✶➑✑➟✑➒☛➞✑➍❱➝✄➞✶➟✑➍✠➋☞➑✑➓✛➝✠➏✠➋❍➒☛➠✧➑✶➍✄➋☞➑✶➟✑➏✠➞✑➞✑➥❑➎✞➊☛➋❄➝✠➠✧➋☞➝✠➊☛➎✞➎✖➍✠➋☞➑✑➞✑➥✾→ ↕⑩➛❫➣➄➜➨➠✧➒☛➍✄➎✞➏✄➑✶➍✄➓✬➢☛➥❴➞✶➠✧➏✠➌☛➟✑➋☞➤⑤➏✎↕✾➩➫➩➫➭❡➎✖➍✠➒✧➎✞➍✄➓✱➍✄➋☞➑✶➏✄➑✶➟✑➠☛➋❴➠☛➯❏➑✑➐☞➍➄➎✞➊☛➞✶➍➄➓✱➍✄➑❥→✠➲✁➐☛➟✶➓❹➟✑➋☞➟✶➑✑➟✑➏✠➞♠➓✱➍✄➑♠➟✶➓ ➌✧➟✶➓✔➒☛➞✑➏✠➥↔➍✄➌➳➏✄➓⑤➏⑧➡☎➍✠➢➫➒✧➏✠➤☞➍❩➵✖➸✮➟✶➤☛➊☞➎✞➍❤➺➼➻❼→✾➸✮➎✖➠☛➦✢➑✑➐☞➟✶➓⑤➒✧➏✠➤☞➍❩➑✑➐☞➍❩➊☞➓✱➍✄➎✎➝✠➏✠➋➳➤☞➠➫➑✑➠➳➠✧➑✶➐☛➍✠➎✎➒☛➏✄➤☞➍✠➓ ➝✄➠☛➋☞➑✑➏✠➟✑➋☞➟✑➋☞➤❄➠✧➎✞➌✧➍✠➎✖➍✠➌❴➞✶➟✑➓✱➑✑➓✛➠☛➯⑨➎✞➊☞➞✑➍✠➓✛➡☎➟✶➑✑➐❄➓✔➊☞➒☛➒✧➠☛➎✖➑r➏✄➋☞➌❴➝✠➠☛➋☛➯➅➟✑➌☛➍✠➋☛➝✠➍r→ ➲❁➠❪➦❴➠☛➔☞➍◆➯✞➎✞➠✧➦➽➒✧➏✠➤☞➍❙➵✖➓✱➍✄➑❫➠☛➯✬➎✖➊☞➞✑➍✠➓✱➻❣➑✶➠⑦➒✧➏✠➤☞➍✄➧⑨➑✑➐☞➍❙➊☞➓✔➍✠➎❭➏✄➒☛➒✧➞✶➟✑➍✠➓✎➎✖➍✠➓✔➑✶➎✖➟✶➝✄➑✶➟✑➠☛➋☛➓✎➏✠➋☞➌⑦➠✧➒☛➍✠➎✖➏✠➾ ➑✑➠☛➎✖➓❜→r➲❁➐☞➍❙➎✖➍✠➓✱➑✑➎✖➟✶➝✄➑✶➟✑➠☛➋☞➓✎➝✄➏✠➋❩➢☛➍❙➌✧➠☛➋☞➍❙➠✧➋❩➑✶➐☞➍❙➦❴➟✑➋☞➟✶➦❴➊☞➦➚➝✄➠☛➋☞➯✞➟✶➌✧➍✠➋☞➝✄➍✠➧❏➦❴➟✶➋☞➟✑➦❴➊☞➦➚➓✱➊☛➒☛➒☛➠✧➎✞➑✑➧❏➠☛➎ ➠✧➋❩➯➅➊☞➋☞➝✄➑✶➟✑➠☛➋☛➓✎➠☛➯✬➑✑➐☞➍❙➓✱➊☞➒✧➒☛➠✧➎✞➑❫➏✄➋☞➌⑦➝✠➠✧➋☞➯➅➟✑➌☛➍✄➋☞➝✠➍❙➠✧➯✬➑✶➐☛➍❙➟✶➑✑➍✠➦❴➓✱➍✄➑✶➓✎➟✑➋❩➑✶➐☞➍⑤➎✖➊☞➞✑➍r→☛➪✎➒☛➍✠➎✖➏✠➑✑➠☛➎✖➓➄➝✠➏✄➋ ➢✧➍❄➓✔➍✠➞✑➍✠➝✠➑✑➍✠➌❘➯➅➎✖➠☛➦➇➏❄➞✶➟✑➓✱➑➶→❜➹✞➯✁➟✶➑⑩➟✶➓❭➏⑧➘➷➴❱➬☛➮✑➱✉✃ → ➘✔❐☛➱✠❒✑❮❭❰❥Ï✛➴❱➬✧➮✶➱✄❮❳✃❱➠☛➒✧➍✠➎✖➏✠➑✑➠☛➎✖➧♠➑✶➐☞➍❄➟✶➋☞➒✧➊☞➑✾➎✖➊☞➞✑➍❑➦❴➊☞➓✱➑ ➏✄➞✶➓✔➠❍➢✧➍❱➓✔➍✠➞✑➍✠➝✄➑✶➍✄➌⑨→ Ð❫Ñ✶Ò✧Ó✮Ô✠Õ☎Ö⑨× ↕⑩➛❫➣➄➜❤➒☛➞✑➠☛➑✑➑✶➟✑➋☞➤❍➓✱➊☞➒✧➒☛➠☛➎✖➑✉Ø❄➝✄➠☛➋☞➯✞➟✶➌✧➍✠➋☞➝✄➍❱➒✧➠☛➟✑➋☞➑✶➓✛➯➅➠✧➎❏➏❱➓✔➊☞➢☛➓✔➍✠➑✉➠☛➯⑨➎✞➊☛➞✶➍✄➓✱➧☛➏✄➋☞➌❴➓✱➐☞➠✧➡❱➾ ➟✑➋☞➤❄➏☎➦❴➊☞➞✶➑✑➟✑➾➅➢☛➏✄➎❏➐☞➟✶➓✔➑✶➠✧➤☞➎✖➏✠➦⑦→ ➸✾➠☛➎✆➍✠➏✄➝✠➐❴➒☛➏✄➤☞➍✠➧r➑✶➐☛➍☎➊☛➓✱➍✠➎❏➝✄➏✠➋❄➏✄➞✶➓✔➠❍➓✔➍✠➞✑➍✠➝✄➑r➏☎➤☞➎✖➏✠➒✧➐☞➟✶➝✄➏✠➞✉➔☞➟✑➓✱➊☞➏✄➞✶➟✑Ù✠➏✄➑✶➟✑➠☛➋❍➑✶➐☛➏✠➑✉➓✱➊☞➦❴➦❴➏✠➎✖➟✶Ù✄➍✠➓✛➑✶➐☞➍ ➓✔➍✠➑⑨➠☛➯❁➎✞➊☞➞✑➍✠➓❱➠☛➋◆➑✑➐☞➍❍➒☛➏✠➤☛➍r→❳Ú✬➊☞➎✖➎✖➍✠➋☞➑✑➞✶➥↔➧✾➑✶➐☞➍❄➏✠➔☞➏✄➟✶➞✑➏✠➢✧➞✶➍❑➔☛➟✶➓✔➊☞➏✠➞✑➟✶Ù✄➏✠➑✑➟✶➠✧➋☞➓❭➏✄➎✞➍❑Ú✬➠✧➋☞➯➅➟✑➌☛➍✄➋☞➝✠➍ × Û ➊☞➒☛➾ ➒✧➠☛➎✖➑❣➒✧➞✶➠✧➑❣➏✠➋☛➌❤Ú✬➠☛➋☞➯✞➟✶➌✧➍✠➋☞➝✄➍❪Ü❣➓✔➊☞➒☛➒✧➠☛➎✖➑❣➐☞➟✑➓✱➑✑➠☛➤☛➎✞➏✄➦❍➓◆➵✖➸✮➟✑➤☞➊☞➎✖➍❪Ý☛➻❵→⑩➲✁➐☞➍⑦➒✧➎✞➠✧➌☛➊☞➝✄➍✠➌Þ➝✄➐☞➏✠➎✖➑✶➓❙➏✄➎✞➍ ➟✑➋☞➑✑➍✠➎✖➏✠➝✄➑✶➟✑➔☞➍☎➏✠➋☞➌❴➟✶➋☛➌☛➟✑➝✠➏✠➑✑➍☎➑✶➐☞➍☎➎✖➊☞➞✶➍☎➑✑➐☞➏✠➑✉➝✠➠✧➎✞➎✖➍✠➓✔➒☛➠✧➋☞➌☛➓✛➑✶➠❴➑✑➐☞➍❱➒✧➠☛➟✑➋☞➑✉➊☞➋☞➌✧➍✠➎❏➑✶➐☛➍❱➦❴➠☛➊☞➓✔➍r→ ➲❁➐☞➍⑤➏✠➓✔➓✱➠☛➝✄➟✶➏✄➑✶➟✑➠☛➋⑧➎✖➊☞➞✶➍⑤➢✧➎✞➠✧➡❱➓✔➍✠➎✬➐☞➍✄➞✶➒✧➓☎➑✑➐☞➍⑤➊☞➓✔➍✠➎✬➑✑➠⑧➋☛➏✠➔☞➟✑➤☞➏✠➑✑➍⑤➑✶➐☛➎✞➠✧➊☞➤☞➐❘➑✑➐☞➍❴➓✱➒✧➏✠➝✠➍❴➠☛➯✛➎✖➊☞➞✶➍✄➓❱➢✧➥ ➔☛➟✶➍✄➡❱➟✑➋☞➤❴➠☛➋☞➍➄➓✱➍✄➑✉➠☛➯⑨➎✞➊☞➞✑➍✠➓✛➏✠➑✉➏☎➑✶➟✑➦❴➍r→❵➛✁➏✠➝✠➐❍➓✱➍✄➑r➠✧➯⑩➎✖➊☞➞✑➍✠➓✛➝✠➠✧➎✞➎✖➍✠➓✔➒☛➠☛➋☛➌☛➓✛➑✶➠❴➠☛➋☞➍☎➒☛➏✄➤☞➍r→➼➸✮➎✖➠☛➦ß➠☛➋☞➍ ➤☛➟✶➔☞➍✄➋⑦➒☛➏✄➤☞➍◆➑✑➐☞➍◆➊☞➓✔➍✠➎❣➦❴➠☛➔☛➍✠➓❑➑✑➠❪➑✑➐☞➍◆➯➅➠✧➞✶➞✑➠☛➡☎➟✑➋☞➤⑦➢☛➥❩➏✠➒✧➒☛➞✑➥↔➟✶➋☛➤⑧➏❙➓✔➍✠➞✑➍✠➝✄➑✶➍✄➌⑦➠☛➒☛➍✄➎✞➏✄➑✶➠✧➎❭➑✑➠⑦➏✠➞✑➞❫➠☛➎ ➓✔➠☛➦❴➍⑤➠☛➯❹➑✶➐☛➍❍➎✖➊☞➞✑➍✠➓☎➔☞➟✶➍✄➡❱➍✄➌✺➠✧➋❘➑✑➐☞➍❴➝✠➊☞➎✖➎✞➍✄➋☞➑❏➒☛➏✄➤☞➍r→☞➹✞➋✺➑✶➐☞➟✑➓☎➓✱➍✠➝✄➑✶➟✑➠☛➋✺➡❱➍❴➌☛➍✄➯➅➟✑➋☞➍❴➑✶➐☞➍❴➓✱➍✄➑❏➠☛➯✛➠☛➒☛➾ ➍✄➎✞➏✄➑✶➠✧➎✞➓✛➑✑➠❍➏✄➒☛➒☛➞✑➥❑➑✑➠❍➓✔➍✠➑✑➓❁➠✧➯⑩➏✄➓✱➓✔➠☛➝✠➟✑➏✠➑✑➟✑➠☛➋❄➎✖➊☞➞✑➍✠➓❳→ ➲❁➐☞➍❙➠☛➒✧➍✠➎✖➏✠➑✑➠☛➎✖➓✎➡❱➍❙➌✧➍✠➓✔➝✠➎✖➟✶➢✧➍❙➐☞➍✠➎✖➍❙➑✶➎✖➏✠➋☛➓✱➯➅➠✧➎✞➦➽➠☛➋☞➍❙➓✔➟✶➋☞➤☛➞✶➍❙➎✖➊☞➞✑➍❱➴ ∈ ➘➷➴❱➬☛➮✑➱✠❮↔✃❑➟✑➋☞➑✑➠⑧➏⑤➓✔➍✠➑✆➠☛➯ ➎✖➊☞➞✑➍✠➓❑➴❱❐ ∈ ➘✔❐✧➱✠❒✑❮✺❰❥Ï❙➴☎➬☛➮✑➱✠❮❳✃③➏✄➋☞➌à➝✠➠☛➎✖➎✖➍✠➓✱➒✧➠☛➋☞➌à➑✑➠➨➑✑➐☞➍➳➝✠➊☞➎✖➎✞➍✄➋☞➑✑➞✶➥Þ➟✑➦❍➒✧➞✶➍✄➦❍➍✄➋☞➑✑➍✠➌❢➠✧➋☞➍✠➓❳→❏➪✎➑✑➐☞➍✠➎ ➟✑➋☞➑✑➍✠➎✖➍✠➓✔➑✶➟✑➋☞➤❘➠✧➒☛➍✄➎✞➏✄➑✶➠✧➎✞➓☎➦❍➏✄➥◆➑✑➎✞➏✄➋☞➓✱➯✞➠☛➎✖➦á➠✧➋☞➍❴➓✱➍✠➑❏➠✧➯❁➎✖➊☞➞✑➍✠➓☎➟✶➋☞➑✑➠⑧➏✠➋☞➠✧➑✶➐☞➍✄➎❼→☞➹✞➋❘➑✶➐☞➍❍➯➅➠✧➞✶➞✑➠☛➡☎➟✶➋☛➤◆➡❱➍ ➌✧➍✠➓✔➝✠➎✖➟✶➢✧➍❱➑✑➐☞➍☎➠☛➒☛➍✄➎✞➏✄➑✶➠✧➎✞➓✛➠☛➯⑨➑✑➐☞➍❱➯✞➠☛➎✖➦❍➍✄➎❏➝✠➞✑➏✠➓✱➓❳→ â✎ã✾ä Õ✄å✠Õ✠æ✾Õ ã✮ä Ò☛Õ ã Õ✠Ô✄ç☛è✑Ñ✶é✄ç ä Ñ✶ê ã❘ë✖â✎ã✮ä✖ì❴í A Post-Processing Environment 47 î ï✧ð❥ñ❹ò✩î ❱ î❙÷ î❙÷☞ú✑û✛ü☛ý☛þ✑ÿ✠ú✁✄✂✆☎❍ý✞✝✟☎✠✂✆✡☛✂✄þ✶ú✁✄☞❄ü✞✄✂☎ü✠✌✎✍❍ü✞✌✏☎ ✂ ÿ✠þ✑ü✠✍❴û✛ú☛✬î✒✑ →ó➄ô⑨õÞö →óùø ü ✖✠✂✆✌✞ÿ✄þ✶ü✞✌✗✖✠✌✖ü✠☎✠✘✠✙✚✂✠û✛✌✜✘✄✡✁✠✂ û➄û✱ú✁❍ ✍ ú✁✡✑ÿ✚✌✬þ✑ü✺þ ✔ ✂✢☞ ☞ ú☛✣✠✂✚❘ü✞✄✂❴✠ ý ✘☞þ✎✤☎ú✶þ ✔ ÿ❴✥û ✝✦☞þ✑ÿ✚✙✄þ✶ú✁✙✠ÿ✆✡☛✡✁✝◆û✔ú☛✍✢✖✠✡✁✂✚✌ ✓✕✔ ú✑û☎✞ ÿ✆☞þ✁✂✚✙✚✂✆☎✠✂✆☞þ★✧ ✓✩✔ ú✑û❣✆ÿ ✡☛✡✑✠ü ✤☎û❣þ ✔ ✂❄✁ú ☎✠✂✆☞þ✑☛ú ✪✞☛ú ✙✄ÿ✠þ✑ú✶✞ü ◆ü✠✪✕✌✏✂✆✡☛✂✆✣☞ÿ✚☞þ⑨ü✠✌✛ú✁✌✏✌✜✂✚✡✁✂✚✣☞ÿ✆☞þ⑨ú✶þ✁✂✚✍❴û❣ú✁❙þ ✔ ✂✟✙✚✘✄✌✜✌✜✂✚☞þ ✌✜✘✄✡✁✂✫✧ ✓✩✔ ✂⑦✥ û ✘✄✖✞✖☛✠ü ✌✖þ✬✚ÿ ✠☎✬✄✙ ✠ü ✄✪✞☛ú ☎✞✂✚✄✙✆✂✭✡✶✁ú ✄✂✠û⑤ü✞☎ ✪ þ ✔ ✂✭✌✜✂✠✥ û ✘✠✡✶þ✑☛ú ✠☞ ✮û ✂✠þ✬✠ü ✪✛✌✜✘✄✡✁✂✠û⑤✚ÿ ✡✁✡✶✞ü ✤ þ ✔ ✂✯✣☛ú✶✮û ✘☞ÿ✚✡ ✁ú ☎✠✂✆☞þ✑☛ú ✪✞☛ú ✙✄ÿ✠þ✑ú✶✞ ü ❘ü✠✪❁ú✶✁þ ✂✚✍❴û❣þ✑✰ ü ✖✞✌✏✘✠✄✂❄✁ú ◆þ ✔ ✂❄ÿ✆☞þ☛✂✆✙✚✂✆☎✠✂✚☛þ★✧✲✱✜◆þ✁✂✚✌✜✍❍û❱✠ü ✪❁þ ✔ ✂❍✚ÿ ☞þ✁✂✚✙✆✂✚☎✞✂✚☞þ✳✶✡ ÿ✄þ✶þ✑ú✁✙✚✂✚✴✾ú✶þ ☞☛☛ú ✣✄✂✄û❁✆ ÿ ✡☛✡✉þ ✔ ✂✵✌✏✘✄✡✁✠✂ û✛✠ý ✂✆✡✶✞ü ✤❷þ ✔ ✂✵✙✚✘✠✌✏✌✜✂✚☞þ✉✞ü ✄✂✶✤☎ú✑þ ✔ þ ✔ ✂☎û✱✆ÿ ✍✷✂✵✙✠ü✞☞û✥✂✆✸✠✘✄✂✆☞þ✹✧ ✺✼✻✾✽✏✿✆❀✚✿✚❁✾✿✚✻❂✽✎❃✁✿✚❄✠❅✮✽✎❆✠✿✚✻✾✿✚❇✚❄✞❃❈❆✠✿✚✻✾✿✚❇✚❄✞❃☛❉✁❊✚❄✞✽✏❉✁❋✠✻❍●✏✺✼✻✾✽✏■✗❏❑❏❑▲ î ï✧ð ▼➄ñ✎ñ❹ò✩î ❱ î❙÷ î❙÷☛ú✶û✛ü☛ý✧þ✶ÿ✄ú☛✄✂✆☎❍ý✞✝✟☎✠✂✆✡☛✂✄þ✶ú✁✄☞❍ü✠✄✂☎ÿ✠þ✑ü✠✍ß✁ú ❹î✒✑ →ó☎ô❏õ➫ö →ó➇ø ✙ þ✁✂✚✌◆✣✠✂✚✌✖û✱ú✑✠ ü ❩ü✠✪✬þ ✔ ✂❱î❱ï✧ð❥ñP❖◗✱✖þ❘☞☞ú✁✣✄✂✠û✎ü✞✄✡☛✝⑧þ ✔ ✂✒✌✏✘✠✡☛✂✄û✎✠ü ❩þ ✔ ✓✕✔ ú✑û✎ü✠✖✠✂✆✌✞ÿ✄þ✶ü✞✌❭ú✑û✎ÿ❙û✱✁þ ✌✞ú✁✠ þ ✔ ✂☎✚ÿ ☞þ✁✂✚✙✆✂✚☎✞✂✚☞þ❈✑✡ ÿ✠þ✑þ✶✁ú ✙✚☎ ✂ ☛ú ✍✢✍✢✂✚☎☛ú✑ÿ✠✁þ ✂✚✡✁❑ ✝ ✞ ý ✂✚✡✑✠ü ✤❷þ ✔ ✂✶✙✆✘✄✌✏✌✜✂✚☛þ✫✌✜✘✄✡✁✂✫✧ ✂✢✡✁✂✚✣✠✂✚✡ ✆ü✠✪ ❙✼❋✞✻❂❅✥✿✆❚❂❯❂✿✆✻❂✽✎❆✠✿✚✻✾✿✚❇✚❄✞❃☛❉✁❊✚❄✞✽✏❉✁❋✠✻❍●✏❙✼❋✞✻❂❅✥❏❑▲ ☛ï ❜ñ❹ò✩î ❱✳❲ ✞❳ î →ó➄ô⑨õÞö ÷ →ó ø✶ó ÷☞ú✑û✛ü☛ý☛þ✑ÿ✠ú✁✄✂✆☎❍ý✞✝✟☎✠✂✆✡☛✂✄þ✶ú✁✄☞❄ÿ✄þ✶ü✞❍ ✍ û✛ú✁ ó ✑ ❙✼❋✞✻❂❅✥✿✆❚❂❯❂✿✆✻❂✽✎❃☛✿✆❄✠❅✥✽✎❆✞✿✚✻❂✿✆❇✚❄✠❃❈❆✞✿✚✻❂✿✆❇✚❄✠❃✁❉✁❊✚❄✠✽✜❉✁❋✠✻✰●✜❙✼❋✠✻✾❅✥■✗❏❑❏❑▲ ☛ï ➄ñ✎ñ❹ò✩î î ÷ ÷☛ú✶û✛ü☛ý✧þ✶ÿ✄ú☛✄✂✆☎❍ý✞✝✟☎✠✂✆✡☛✄✂ þ✶ú✁✄☞❍ü✠✄✂☎ÿ✠þ✑✠ü ✍ßú✁ ✑ ó →ó☎ô❏õ➫ö →ó ø✤ó ❩✾ú☛✍❴☛ú ✡✑✚ ÿ ✌✁þ✶ü❱î❱ï✧ð❥ñ➨ÿ✚✠☎❱î❱ï☛ð ▼➄ñ✎ñ❬✌✏✄✂ û✥✖✠✂✆✙✠þ✑ú☛✣✠✂✚✡✁✝✦✴rý✞☞✘ þ♠þ ✔ ✂➄û✱ú✁✍✷✖✞✶✡ ú✁✪➅ú✁✙✠ÿ✄þ✶ú✑ü✠❴ú✶û❭☎✧✠ü ✄✂➄✠ü ❴þ ✔ ✂✛✙✠ü✠☛û✥✂✚❪ ✸✞✘✄✂✚☛þrú✁☞û✔☛þ ✂✄✚ ÿ ☎❴✠ü ✪⑨✠ü ❄þ ✔ ✂☎✚ÿ ☞þ✁✂✚✙✆✂✚☎✞✂✚☞þ✹✧ ❱✳❲ ✞❳❨▼ ✺✼✻✾✽✏✿✆❀✚✿✚❁✾✿✚✻❂✽✎❅✥❫✾✿✚❀✚❉✁❄✠❃✁❉✁❊✚❄✠✽✜❉✁❋✠✻✰●✜✺✼✻❂✽✜❴❂▲ ✓✕✔ î◆÷ î❙÷ î❛✑ î ï✧ð☛❵ròqî ❱ →ó☎ô❏õ➫ö →ó➇ø ⊇ ú✑✕ û ✖✞✌✞ü✞☎✠✘✄✙✆✂✠P û ✌✏✘✄✡✁✠✂ ûP✤☎ú✶þ ✔ ✡✑✠ü ✤✵✂✚✌❏û✥✘✠✖✠☛✖ ü✞✌✞þ✉✞ý ✘☞þ ✔ ☛ú ☞ ✠ü ➅ú ✔ ✂✚✌✎✙ ✞✄✪ ✁☎✠✂✆✄✙✚✂ ☎þ ✔ ÿ✚❍þ ✔ ✂✵✙✚✘✠✌✏✌✜✂✚ ☞þ✉ü✞✄✂✫✧ ✺✼✻✾✽✏✿✆❀✚✿✚❁✾✿✚✻❂✽✎❃✁✿✚❄✠❅✮✽✎❅✥❫❂✿✆❀✚❉✁❜✏❉✁❀✶❅✮❫❂✿✚❀✆❉☛❄✞❃☛❉✁❊✚❄✞✽✏❉✁❋✠✻❍●✏✺✼✻✾✽✏■✗❴❂❴✾▲ î❙÷ î❙÷☞ú✑û✛ü☛ý☛þ✑ÿ✠ú✁✄✂✆☎❍ý✞❑ ✝ ÿ✚☎✞☎☛ú✁✄☞❄ü✞✄✂❱ò✖ÿ✚✠✝ ÿ✄þ✶✞ü ✍ßþ✶ü❭î✒✑ î ï✧ð ▼✕❵✞❵✧ò✩î ❱ ô →ó➄ô⑨õÞö →óùø ❝➄û❱î❱ï✧ð☛❵✫✴❹ý✠✘☞þ✬ü✞✄✡☛✝❞✪➅ü✞✌✎þ ✔ ✂❩✁ú ✍✷✍✢✂✚☎✧ú✶ÿ✄þ☛✂✭✡✁✂✚✣✄✂✆✡✬ÿ✠ý✧✠ ü ✣✄✂❩þ ✔ ✂✭✙✚✘✄✌✜✌✏✂✆☞þ✗✌✜✘✄✡☛✂❩ü✞ þ ✔ ✂❩ÿ✚☞þ✁✂✚✙✆✂✚☎✞✂✚☞þ ✡✑ÿ✠þ✑þ✑☛ú ✙✆✂✫✧ ❙✼❋✞✻❂❅✥✿✆❚❂❯❂✿✆✻❂✽✎❅✥❫❂✿✆❀✚❉✁❄✠❃✁❉☛❊✆❄✠✽✜❉☛❋✞✻✰●✜❙✼❋✠✻❂❅✮❴❂▲ ☛ï ròqî ❱✳❲ ✞❳✥❵ →ó☎ô❏õ➫ö î →ó ÷ ÷ ø✤ó ⊇ó ✑ ❙✼❋✞✻❂❅✥✿✆❚❂❯❂✿✆✻❂✽✎❃☛✿✆❄✠❅✥✽✎❅✮❫❂✿✚❀✆❉☛❜✜❉☛❀✵❅✥❫✾✿✚❀✚❉✁❄✠❃✁❉✁❊✚❄✠✽✜❉✁❋✠✻✰●✜❙✼❋✠✻✾❅✥■✗❴❂❴❂▲ ☛ï ✧ò✩î î ÷ ÷☞ú✑û✛ü☛ý☛þ✑ÿ✠ú✁✄✂✆☎❍ý✞✝❑✚ÿ ☎✞☎☛ú✁✄☞❄ü✞✄✂❱ò✖ÿ✚✠✝ ÿ✄þ✶ü✞✍ßþ✶ü →ó➄ô⑨õÞö →ó ø✹❡ ô ó ❩✾ú☛✍❴ú☛✡✑✚ ÿ ❏✌ þ✑ü✬î☎ï☛ð✁❵❴ÿ✆✄☎✬î☎ï☛ð✁❵✠❵✞✴✧ý✠✘☞þ✉✞ü ❄þ ✔ ✂✶✙✄ü✠☞û✮✂✚✸✞✘✄✂✚☞þ✹✧ ❱✳❲ ✞❳❨▼✕❵✞❵ ❢ ❋✠❀✚❯✾❅✕❋✞✻✰❄✠✻✾✽✏✿✆❀✚✿✚❁✾✿✚✻❂✽✎● ❢ ✑ ✺✼✻❂✽✜▲ î❱ï✧ð✑ò✩î î ❱ ❱⑦ú✑û✛✚ÿ ✄✝✾✑ →ó☎ô❏õ➫ö → ø ❤✎ú✁✣✄✂✄û❁✆ ÿ ✡☛✡✉þ ✔ ✂✶✌✜✘✄✡✁✠✂ P û ✤❱ú✑þ ✔ ✂✚✐☛ÿ✚✠✙ þ✁✡✁✝❑þ ✔ ✂❱û✔ÿ✚✍✢✂❱ÿ✆☞þ☛✂✆✙✚✂✆☎✠✂✚☛þ★✧ ❣ î❱ï☛ð ò❦❥ ❣ ô❏õ î❱ï✧ð❥ñ❹ò❦❥ ô ∪ î❱ï✧ð☛❵rò❧❥ ô ✧ 48 Jorge et al. ♠✩♥✠♦✚♣✾q✕♥✞r✰♦✚♥✞r❂q✥s✆t❂♣❂s✆r❂✉✎✈✏♠❘✇✼♥✞r❂q✥① ②❑③✳④✞⑤✠⑥✮⑦❦⑧ →⑨✵⑩✎❶❸❷ ③ →⑨❺❹ ③❼❻✁❽P❾✚❿✄➀✾➁ ➂ ❻✁➃✄➄✆❽✕❾✆➅☛➅❈➆✁➇✄➄✶➈✜➉✄➅✁➄✚❽P➊✶❻✁➆✁➇➋➆✁➇✄➄✶❽✮❾✚➌✢➄✶➍✆➎✠❿✄❽✮➄✚➏✞➉✄➄✚❿✄➆✹➐✏②✒③✳④✠⑤✠⑥➑⑦❧➒ ➣ ➉✠➈✗➌✷➄✆➆☛➇✄➎✞↔✠➎✞➅☛➎✞↕✄➀✰❻✁❽✛➙✠❾✚❽✮➄✚↔✭➎✠❿✯➆✁➇✄➄❑➛✠➇✄❻✁➅☛➎✞❽✥➎✞➛✠➇✄➀❍➎✠➜❭➊✶➄✆➙✭➙✠➈✜➎✠➊✵❽✥❻✁❿✄↕✄➝✳➛✠❾✚↕✠➄✷➙✞➀❛➛✠❾✆↕✄➄✷➜✏➎✠➅✁➅☛➎✞➊✶➞ ⑩✳❶ ③✳④✠⑤✞⑥➔➓❭⑦❦➒ ⑩ ∪ ③✳④✠⑤✞⑥✥→✞⑦❦➒ ⑩ ➐ ❻✁❿✄↕➟➇✠➀✦➛✠➄✚➈✜➅✁❻☛❿✄➠✠❽➔➐❈➡✕➇✄➄❛➎✠➛✞➄✚➈✜❾✚➆✁➎✠➈✜❽✟❻✁➌✷➛✞➅☛➄✆➌✷➄✆❿✄➆✩➆✁➇✄➄❛➇✄➀✦➛✞➄✚➈✜➅☛❻✁❿✄➠✄❽➋➙✠➄✚➆✁➊✵➄✚➄✚❿❼➆✁➊✵➎➟➛✠❾✆↕✄➄✚❽✲➐✫➡✩➎➟❽✮➆☛❾✆➈✏➆ ➙✞➈✏➎✞➊✶❽✮❻☛❿✠↕✄➝❂➆✁➇✄➄➋➉✄❽✮➄✚➈P❿✄➄✚➄✆↔✠❽✶❾✚❿❛❻✁❿✄↔✠➄✆➢❛➛✠❾✆↕✄➄✫➐✦➡✩➇✄❻✁❽➤❽✮➇✄➎✠➉✄➅✁↔✰❻✁❿✄➍✆➅☛➉✄↔✞➄➋❾✷❽✥➉✄➙✞❽✥➄✚➆✳➎✠➜✕➆☛➇✠➄➋➈✜➉✄➅☛➄✆❽◆➆✁➇✄❾✚➆ ❽✮➉✄➌✢➌✷❾✆➈✏❻✁➥✚➄✒➆✁➇✄➄✒➊✵➇✄➎✠➅✁➄✒❽✥➄✆➆★➐✞➦✜❿✯➆✁➄✚➈✜➌✢❽✵➎✞➜❭➊✶➄✆➙✯➙✞➈✏➎✞➊✶❽✮❻☛❿✠↕✄➝➧❻✁➆◗❽✥➇✠➎✠➉✄➅✁↔✭➙✠➄❑❾❑❽✥➌✢❾✚➅✁➅◗❽✥➄✚➆◗➎✞➜❭➈✏➉✄➅✁➄✚❽ ➆✁➇✄❾✆➆❘❾✚➅✁➅☛➎✞➊✶❽✼↕✠➄✚➆✁➆☛❻✁❿✄↕✭➆☛➎❼❾✆❿✄➀✯➛✠❾✚↕✠➄✒❻☛❿✭❾✒➅✁❻☛➌✢❻☛➆✁➄✚↔❼❿✠➉✄➌✷➙✞➄✚➈✗➎✞➜P➍✆➅☛❻✁➍✚➠✠❽➔➐✠➨❺➍✚❾✆❿✄↔✠❻✁↔✠❾✆➆☛➄❑➜➩➎✞➈✗❽✥➉✄➍✆➇❍❾ ❽✮➄✚➆◆➍✆➎✠➉✄➅✁↔➫➙✠➄❼➆✁➇✄➄✆➝P➜✏➎✠➈✟➄✆➢✄❾✚➌✢➛✠➅✁➄✚➝❭➆✁➇✄➄❼❽✥➌✢❾✚➅✁➅✁➄✚❽✥➆➤➈✏➉✄➅✁➄❼➜➩➎✞➈✟➄✚❾✆➍✚➇✬➍✚➎✞❿✄❽✥➄✆➏✠➉✄➄✆❿✄➆★➐✾➭❘❾✆➍✚➇✬➎✠➜✛➆☛➇✠➄✚❽✥➄ ➈✜➉✄➅✁➄✚❽✛➊✵➎✠➉✄➅✁↔✭➈✏➄✆➛✠➈✜➄✚❽✥➄✆❿✄➆◗➆✁➇✄➄❑➅☛❾✆➆☛➆✁❻☛➍✆➄❑➎✠❿✯➆☛➇✄➄❑❾✆❿✄➆☛➄✆➍✚➄✆↔✠➄✚❿✠➆☛❽✛➎✠➜❭➆✁➇✄➄✢➈✏➉✄➅✁➄✚❽✵➊✵❻☛➆✁➇✰➆✁➇✄➄✢❽✥❾✚➌✢➄✢➍✚➎✠❿✠❽✥➄✚➞ ➏✞➉✄➄✚❿✠➆★➐✄➯❂❻☛❿✠➍✚➄✢➆☛➇✄➄✢➅✁❾✚➆✁➆☛❻✁➍✚➄✆❽✶❻✁❿✄➆✁➄✚➈✜❽✥➄✚➍✆➆☛➝➧➊✶➄✢➍✚❾✆❿✰➍✆➇✄❾✚❿✄↕✠➄✷➆✁➎❍❾✢➜➩➎✞➍✚➉✄❽✵➎✠❿❍➆☛➇✠➄✷❾✆❿✄➆☛➄✆➍✚➄✆↔✠➄✚❿✠➆✳➎✞❿✰❾✆❿✄➀ ➈✜➉✄➅✁➄✶➙✞➀✟❾✚➛✞➛✠➅✁➀✦❻☛❿✠↕➋❾✚❿✷❾✚➛✞➛✠➈✜➎✠➛✠➈✜❻✁❾✚➆✁➄✶➎✞➛✠➄✚➈✜❾✚➆✁➎✠➈➲➐ ➯✾❻☛➌✢❻☛➅✁❾✚➈✜➅✁➀✦➝✫➊✵➄✼➍✚➎✞➉✄➅✁↔✒❽✥➆✁❾✚➈✜➆❂➊✶❻✁➆✁➇✢➆✁➇✄➄✼❽✮➄✚➆❂➎✞➜◗❽✥➌✢❾✚➅✁➅☛➄✆❽✥➆❂➈✜➉✄➅✁➄✚❽✗➜➩➎✞➈❘➄✚❾✆➍✚➇❑❾✚❿✠➆☛➄✆➍✚➄✚↔✞➄✚❿✄➆✹➐✆➨✛➅☛➆✁➄✚➈✜❿✄❾✚➞ ➆✁❻✁➃✄➄✚➅✁➀✦➝✎❻☛❿✄❽✮➆☛➄✆❾✚↔❼➎✞➜✗➆☛➇✄➄✒❽✮❻☛➥✆➄✚➝✎➊✵➄✒➍✚➎✠➉✠➅☛↔❼➍✆➎✠❿✄❽✮❻☛↔✞➄✚➈◆➆✁➇✄➄✒❽✥➉✠➛✠➛✠➎✞➈✏➆✁➝✎➍✚➎✞❿✄➜➩❻✁↔✠➄✆❿✄➍✚➄✆➝✳➎✞➈◆➎✠➆✁➇✄➄✆➈✗➌✷➄✆❾✚❽✥➞ ➉✠➈✏➄❈➐✲➨✛➅☛➅✳➆☛➇✠➄✚❽✥➄✷➛✠➎✞❽✥❽✥❻✁➙✠❻✁➅✁❻☛➆✁❻☛➄✆❽➤➌✢➉✄❽✥➆✳➙✠➄➋❽✥➆✁➉✄↔✠❻✁➄✚↔❛❾✆❿✄↔❛❽✮➎✠➌✢➄✟➎✠➜❘➆✁➇✄➄✆➌➳❻✁➌✷➛✞➅☛➄✆➌✷➄✆❿✄➆☛➄✆↔❛❻✁❿✒➎✠➉✄➈✕❽✥➀✦❽✥➞ ➆✁➄✚➌✢➝✠➊✵➇✄❻✁➍✚➇✷➍✚➉✄➈✜➈✏➄✆❿✄➆✁➅☛➀✟❽✮➇✄➎✠➊✵❽✥➝✞❾✚❽P➆☛➇✠➄✶❻✁❿✄❻✁➆☛❻✁❾✚➅❈➛✠❾✆↕✄➄✚➝✞➆☛➇✠➄✶❽✮➄✚➆❈➎✠➜✳❾✚➅✁➅❈➈✏➉✄➅✁➄✚❽✲➐ ➵➸➄✟❿✠➎✠➊➺↔✞➄✚❽✮➍✚➈✜❻☛➙✞➄✟➇✄➎✠➊➻➆☛➇✠➄✟➌✷➄✆➆☛➇✠➎✠↔❛➙✞➄✚❻✁❿✄↕❑➛✠➈✜➎✠➛✠➎✞❽✥➄✆↔✒➍✚❾✚❿❑➙✞➄✼❾✚➛✠➛✞➅☛❻✁➄✚↔✒➆✁➎✒➙✠➈✜➎✠➊✵❽✥➄✼➆✁➇✄➈✜➎✠➉✄↕✠➇✢❾ ❽✮➄✚➆✵➎✠➜✷❾✚❽✥❽✮➎✠➍✆❻☛❾✆➆☛❻✁➎✠❿➼➈✜➉✄➅✁➄✚❽✲➐✎➡✩➇✄➄❞↔✠➎✞➌✷❾✆❻☛❿➼➍✆➎✠❿✄❽✮❻☛↔✞➄✚➈✜➄✚↔❬❻☛❽❍➆☛➇✄➄❞❾✚❿✄❾✆➅☛➀✦❽✮❻☛❽❍➎✠➜✷↔✠➎✠➊✵❿✄➅✁➎✠❾✆↔✠❽✰↔✞➎✠❿✄➄ ➜✏➈✏➎✞➌➽➆☛➇✠➄➾❽✮❻☛➆✁➄➚➎✠➜❸➆✁➇✄➄➚➪✾➎✠➈✜➆✁➉✄↕✄➉✄➄✆❽✥➄❬➶✰❾✚➆✁❻☛➎✞❿✄❾✚➅❼➦✜❿✄❽✮➆☛❻✁➆☛➉✠➆☛➄➚➎✠➜❸➯❂➆✁❾✚➆✁❻✁❽✥➆✁❻☛➍✆❽➹⑦✜➦➘➶✰➭ ⑦✜➊✵➊✶➊✒➐ ❻✁❿✄➄✫➐ ➛✞➆☛➴✁❻✁❿✄➜➩➎✞➅☛❻✁❿✄➄ ➐➋➡✩➇✠❻☛❽➹❽✥❻✁➆☛➄ ⑩ ➜✏➉✄❿✄➍✚➆✁❻✁➎✠❿✄❽✵➅☛❻✁➠✄➄✢❾✚❿❍➄✚➅✁➄✚➍✚➆✁➈✜➎✠❿✄❻✁➍✷❽✮➆☛➎✞➈✏➄✆➝✾➊✵➇✄➄✚➈✜➄✢➆☛➇✄➄✢➛✠➈✜➎✠↔✞➉✄➍✚➆✁❽✵❾✚➈✜➄➋➆✁❾✚➙✠➅✁➄✚❽ ⑩ ❻✁❿➋↔✞❻☛↕✠❻☛➆✁❾✚➅❈➜➩➎✞➈✏➌✢❾✚➆❈➊✵❻☛➆✁➇➋❽✮➆☛❾✆➆☛❻✁❽✥➆✁❻✁➍✚❽P❾✚➙✠➎✞➉✄➆❈➪✾➎✞➈✏➆✁➉✄↕✄❾✆➅★➐ ➷✾➈✏➎✞➌➬➆✁➇✄➄➋➊✵➄✚➙❍❾✚➍✆➍✚➄✚❽✮❽➤➅✁➎✠↕✄❽✶➎✠➜✕➆☛➇✄➄✷❽✥❻✁➆☛➄❈➮❧❽✶➇✄➆☛➆✁➛✰❽✮➄✚➈✜➃✄➄✚➈P➊✵➄➋➛✠➈✜➎✠↔✞➉✄➍✚➄✆↔❛❾✟❽✮➄✚➆✾➎✞➜❘❾✚❽✮❽✥➎✞➍✚❻✁❾✚➆✁❻☛➎✞❿ ➈✜➉✄➅✁➄✚❽✢➈✏➄✆➅☛❾✆➆☛❻✁❿✄↕❞➆☛➇✄➄✯➌✢❾✚❻✁❿➱➆✁➇✄➄✚➌✢❾✚➆✁❻☛➍✯➍✚❾✆➆☛➄✆↕✄➎✠➈✜❻✁➄✚❽✢➎✠➜✵➆☛➇✄➄✯↔✠➎✞➊✶❿✠➅☛➎✞❾✚↔✠➄✆↔❞➆✁❾✚➙✞➅☛➄✆❽➔➐❈➡✕➇✄❻✁❽➋❻✁❽➋❾✰➈✜➄✚➅✁❾✚➞ ➆✁❻✁➃✄➄✚➅✁➀❑❽✥➌✢❾✚➅✁➅✾❽✥➄✆➆✾❽✮➄✚➆❂➎✞➜◗➈✏➉✄➅✁➄✚❽✗⑦✜✃✳❐✞❐ ❻✁❿✄➃✄➎✞➅☛➃✠❻☛❿✄↕❑❒✒❻✁➆☛➄✆➌✷❽✗➆✁➇✄❾✆➆❂❽✥➄✆➈✏➃✄➄✆❽✗❾✚❽✗❾✆❿✢❻✁➅✁➅☛➉✄❽✮➆☛➈✜❾✚➆✁❻✁➃✄➄✼➄✚➢✄❾✆➌✷➞ ⑩ ➛✞➅☛➄❈➐➧➡✩➇✄➄❼❾✆❻☛➌✢❽✒➎✠➜✼➦➘➶✰➭➻❾✆➈✏➄✭➆✁➎❸❻☛➌✢➛✠➈✜➎✠➃✄➄✭➆✁➇✄➄✭➉✄❽✥❾✆➙✠❻✁➅☛❻✁➆✁➀➱➎✠➜✛➆✁➇✄➄✭❽✥❻✁➆☛➄✭➙✞➀➱↔✠❻✁❽✥➍✆➎✠➃✄➄✆➈✏❻✁❿✄↕✬➊✶➇✠❻☛➍✆➇ ❻✁➆✁➄✚➌✢❽✗❾✚➈✜➄✼➆☛➀✦➛✞❻☛➍✆❾✚➅✁➅☛➀✷➍✚➎✠➌✢➙✠❻✁❿✄➄✆↔✢➙✞➀➋➆✁➇✄➄✛❽✥❾✆➌✷➄✛➉✄❽✮➄✚➈➲➐❮➡✕➇✄➄✛➈✏➄✆❽✥➉✄➅✁➆☛❽❭➎✞➙✠➆✁❾✚❻✁❿✄➄✚↔❑➍✚❾✆❿✷➙✞➄✛➉✄❽✥➄✆↔✢❻✁❿✢➆☛➇✄➄ ➈✜➄✚❽✮➆☛➈✜➉✄➍✆➆☛➉✄➈✜❻✁❿✄↕➟➎✠➜✶➆☛➇✠➄✰❽✥❻✁➆✁➄✰➎✠➈✵❻✁❿➟➆☛➇✠➄✰❻✁❿✄➍✚➅✁➉✄❽✥❻✁➎✠❿❼➎✞➜◆➈✜➄✚➍✚➎✞➌✷➌✢➄✚❿✠↔✠❾✚➆✁❻✁➎✠❿❼➅✁❻☛❿✄➠✠❽✟➎✠❿❼❽✮➎✠➌✢➄❛➛✠❾✆↕✄➄✚❽✲➐ ➨✛➅✁➆☛➇✠➎✠➉✄↕✄➇❸➊✵➄❼❽✥➇✄➎✞➊❰➇✄➄✆➈✏➄❼➇✄➎✞➊❰➈✜➉✄➅☛➄✆❽✒❾✚➆✗➆✁➇✄➄✭➇✄❻✁↕✄➇✄➄✚❽✮➆✗➅☛➄✆➃✄➄✚➅✗➎✞➜✛➆☛➇✄➄✭➛✞➈✏➎✞↔✠➉✄➍✆➆☛❽❑➆✁❾✚➢✄➎✞❿✄➎✠➌✢➀✦➝✕❾ ❽✮❻☛➌✢❻✁➅☛❾✆➈✳❽✮➆☛➉✠↔✠➀✟➍✆➎✠➉✄➅✁↔✷➙✞➄✶➍✆❾✚➈✜➈✏❻✁➄✚↔✢➎✠➉✠➆✫➜✏➎✠➈✎➅✁➎✠➊✵➄✚➈✎➅☛➄✆➃✄➄✚➅✁❽➔➐ Rule Economics_and_Finance <= Population_and_Social_Conditions & Industry_and_Energy & External_Commerce Commerce_Tourism_and_Services <= Economics_and_Finance & Industry_and_Energy & General_Statistics Industry_and_Energy <= Economics_and_Finance & Commerce_Tourism_and_Services & General_Statistics Territory_and_Environment <= Population_and_Social_Conditions & Industry_and_Energy & General_Statistics General_Statistics <= Commerce_Tourism_and_Services & Industry_and_Energy & Territory_and_Environment External_Commerce <= Economics_and_Finance & Industry_and_Energy & General_Statistics Agriculture_and_Fishing <= Commerce_Tourism_and_Services & Territory_and_Environment & General_Statistics ♠✩Ï☛Ð✞♣❂Ñ✚s✵Ò✳Ó✠➷❂❻✁➈✏❽✮➆❈➛✠❾✚↕✠➄✶⑦✜❻☛❿✠↔✠➄✚➢ ⑩ Sup 0,038 0,036 0,043 0,043 0,040 0,036 0,043 Conf 0,94 0,93 0,77 0,77 0,73 0,62 0,51 A Post-Processing Environment 49 Ô✕Õ✄Ö❼×✜Ø✄Ù☛Ö✆Ú✒Û☛Ü❸Ý❂Û✁Þ✄Ø✄×✜Ö❼ß➫Ú✥Õ✠à✠á❰â✁Õ✄Ö❼ã✚à✞Ü✄â✁Ö✚Ü✄â✁Ú✒à✠ä✼à✞Ü✄Ö❼Û☛Ü✠å✠Ö✚æ❸ç✞è✚Þ✄Ö✆éPá✵Û✁â☛Õ❸à✠Ü✠Ö✯×✜Ø✄Ù✁Ö✭ä➩à✠×✼Ö✆è✚ã✆Õ ã✆à✠Ü✄Ú✮Ö✚ê✞Ø✄Ö✚Ü✄â◗ë✜ä➩×✜à✠ìíâ✁Õ✄Ö❑î✭Û☛â✁Ö✚ì✢Ú✥é✳à✠Ü✠Ù☛ï❍ð✯è✆ç✠ç✞Ö✚è✚×✜ñ➲ò✄Ô✕Õ✄Ö❑Ø✄Ú✥Ö✆×✗â☛Õ✄Ö✆Ü✰ä✏Û☛Ü✄å✞Ú✶â✁Õ✄Ö✢×✏Ø✠Ù☛Ö✢à✠Ü❍ó❮Ô✩Ö✆×✏×✜Û☛ô â✁à✠×✜ï✦õ✠è✆Ü✄õ✠ö✩Ü✄÷✄Û✁×✏à✞Ü✄ì✷Ö✆Ü✄â✁ø✷×✜Ö✚Ù✁Ö✚÷✄è✆Ü✄â✎ä➩à✞×PÚ✮â☛×✜Ø✄ã✆â☛Ø✄×✜Û✁Ü✄Þ✰â✁Õ✄Ö➋ã✆è✚â✁Ö✚Þ✄à✞×✏Û✁Ö✚Ú✶à✠Ü❛â✁Õ✄Ö✷Ú✥Û✁â☛Ö❈ò✲ù◆ï✒è✚ç✞ç✠Ù✁ï✦Û☛Ü✠Þ â✁Õ✄Ö❛ú✗à✞Ü✄Ú✥ûüà✞ç✠Ö✚×✜è✚â✁à✠×✜é◗Ú✥Õ✠Ö❛ã✚è✆Ü❼å✠×✜Û☛Ù✁Ù✩å✞à✠á✵Ü✯â✁Õ✄Ö✒Ù✁è✚â✁â☛Û✁ã✚Ö✒è✆×✏à✞Ø✄Ü✄å❼â✁Õ✄è✚â✩×✏Ø✄Ù✁Ö✚é✎à✞ý✠â✁è✚Û✁Ü✄Û☛Ü✠Þ✯è✆Ù☛Ù✩â☛Õ✄Ö ×✜Ø✄Ù✁Ö✚ÚPá✶Û✁â✁Õ➋è✵Þ✄Ö✚Ü✄Ö✆×✏è✆Ù☛Û✁þ✚Ö✆å✷è✆Ü✄â✁Ö✚ã✚Ö✆å✠Ö✆Ü✄â★ò Rule Sup Conf Territory_and_Environment <= Population_and_Social_Conditions & Industry_and_Energy & General_Statistics 0,043 0,77 Territory_and_Environment <= Population_and_Social_Conditions & Industry_and_Energy 0,130 0,41 Territory_and_Environment <= Population_and_Social_Conditions & General_Statistics 0,100 0,63 Territory_and_Environment <= Industry_and_Energy & General_Statistics 0,048 0,77 Territory_and_Environment <= General_Statistics 0,140 0,54 ✁✄✂✆☎✞✝✠✟☛✡✌☞✎✍✛ç✠ç✞Ù☛ï✦Û✁Ü✄Þ➋â✁Õ✄Ö✵à✠ç✞Ö✚×✜è✚â✁à✠×✎ú✗à✠Ü✄Ú✮û ë✜ã✚à✠Ü✠Ú✥Ö✚ê✞Ø✄Ö✚Ü✠â✫Þ✠Ö✚Ü✄Ö✆×✏è✆Ù☛Û✁þ✚è✆â☛Û✁à✠Ü✄ñ➲ò Ý✾×✏à✞ì Õ✄Ö✆×✏Ö✆é✟á✵Ö ã✚è✆Ü Ú✥Ö✆Ö â✁Õ✄è✚â✷✑ ó ✾ ✏ à✞ç✠Ø✄Ù✁è✚â✁Û✁à✠Ü✄õ✞è✚Ü✄å✞õ✓✒❂à✞ã✚Û✁è✚Ù✁õ✠ú✗à✠Ü✠å✠Û✁â☛Û✁à✠Ü✄Ú✮ø➼Û✁Ú➟Ü✄à✠â✷×✏Ö✆Ù☛Ö✆÷✄è✚Ü✠â☛Ù✁ï è✆Ú✥Ú✮à✠ã✚Û✁è✚â✁Ö✚å✒â✁à✒ó❮Ô✕Ö✚×✜×✏Û✁â☛à✞×✏ï✦õ✞è✚Ü✄å✞õ✠ö❘Ü✠÷✄Û☛×✜à✠Ü✠ì✷Ö✆Ü✄â☛ø❈ò✆Ô✩Õ✄Ö✼Ø✄Ú✮Ö✚×❘ã✆è✚Ü✢Ü✄à✠á✵é✫ä➩à✠×◗Ö✆æ✄è✚ì✢ç✠Ù✁Ö✚é✫Ù✁à✠à✓✢ ✔ Û✁Ü✄â☛à ×✜Ø✄Ù✁Ö✚Ú✵á✶Û✁â✁Õ✰✑ ó ➧ ✏ à✠ç✠Ø✠Ù☛è✆â☛Û✁à✠Ü✄õ✞è✚Ü✄å✞õ✓❂✒ à✞ã✚Û✁è✚Ù✁õ✠ú✗à✞Ü✄å✠Û✁â☛Û✁à✠Ü✠Ú✥ø✢ý✠ï✒è✚ç✞ç✠Ù✁ï✦Û☛Ü✠Þ❛â☛Õ✠Ö➋Ý✞✛ ✍ Ü✄â✳ë✏ä✏à✠ã✆Ø✄Ú➤à✞Ü❛è✚Ü✄â✁Ö✚ô ÿ ã✆Ö✚å✞Ö✚Ü✄â✁ñ➋à✞ç✠Ö✚×✜è✚â✁à✠×✷ë✏×✜Ö✚Ú✮Ø✄Ù✁â☛Ú❛Ü✠à✠â◆Ú✮Õ✄à✠á✵Ü❸Õ✄Ö✚×✜Ö✚ñ➲ò➧Ý❂×✏à✞ì â✁Õ✄Ö✚×✜Ö❼Ú✥Õ✄Ö❼ã✆à✠Ø✄Ù✁å➫Ú✥Ö✆Ö❼á✶Õ✠è✚â◆â✁Õ✄Ö❼ì✢è✚Û✁Ü è✆Ú✥Ú✮à✠ã✚Û✁è✚â✁Û✁à✠Ü✄ÚPâ☛à✢â✁Õ✄Û☛ÚPÛ✁â☛Ö✆ì➾è✆×✏Ö❈ò Ô✕Õ✄Ö❼ç✠×✜à✠ã✆Ö✚Ú✮Ú✒á✶à✞Ø✄Ù☛å➫â✁Õ✄Ö✆Ü❞Û✁â☛Ö✆×✏è✆â☛Ö✆é✕è✆Ù☛Ù✁à✠á✵Û✁Ü✄Þ❞â✁Õ✄Ö✭Ø✄Ú✮Ö✚×✼â✁à✬ä✏à✠Ù✁Ù☛à✞áíç✠è✆×✏â✁Û☛ã✆Ø✄Ù✁è✚×✼Û✁Ü✄â☛Ö✆×✏Ö✆Ú✥â✁Û☛Ü✠Þ ✕✏ â✁Õ✄×✜Ö✚è✆å✠Ú✢Û☛Ü❞â☛Õ✄Ö✯×✜Ø✄Ù☛Ö✯Ú✥ç✞è✚ã✆Ö✫ò ➧Ù☛à✞â☛Ú✷è✚Ü✄å❞ý✠è✚×✵ã✚Õ✠è✚×✜â☛Ú✷Ú✥Ø✄ì✢ì✷è✆×✏Û✁þ✚Ö❍â☛Õ✠Ö✰×✜Ø✄Ù☛Ö✆Ú➋Û✁Ü➟à✠Ü✠Ö✰ç✠è✆×✏â✁Û✁ã✚Ø✄Ù✁è✚× ✑✖ ç✞è✚Þ✄Ö❈ò✲Ô✩Õ✠Ö➋Ø✄Ú✮Ö✚×✩ã✆è✚Ü✒è✆Ù☛á✵è✚ï✦Ú◆×✜Ö✚â✁Ø✄×✜Ü✒â☛à✰è✚Ü✒Û✁Ü✄å✠Ö✆æ✒ç✠è✚Þ✠Ö✫ò➔Ô✩Õ✠Ö✟à✠ý ✹Ö✆ã✚â✁Û☛÷✠Ö✟Û☛Ú➤â☛à❛Þ✠è✚Û✁Ü✒Û☛Ü✄Ú✮Û☛Þ✠Õ✄â✾à✞Ü ✗✔ ✙✘➹Õ✠è✚â➤Û✁Ú❛è✆Ü ✚✔✄Ü✄à✞á✶Ù✁Ö✚å✞Þ✄Ö❑à✠äPâ☛Õ✄Ö â✁Õ✄Ö❞×✏Ø✄Ù✁Ö➱Ú✮Ö✚â✶ë✏è✆Ü✄å➼à✠Ü➫â✁Õ✄Ö➟å✞è✚â✁è✚ñ✷ý✠ï❸Ö✚æ✄è✆ì✷Û✁Ü✄Û✁Ü✄Þ➫å✠Û✁Þ✄Ö✆Ú✥â✁Û☛ý✞Ù☛Ö➟ã✆Õ✄Ø✄Ü ✠Ú❛à✠ä✟×✜Ø✄Ù✁Ö✚Ú✲ò Û✁Ü✄â✁Ö✚×✜Ö✚Ú✮â☛Û✁Ü✄Þ✯à✠×✗Ø✄Ü✄Û✁Ü✄â✁Ö✚×✜Ö✚Ú✮â☛Û✁Ü✄Þ✯×✏Ø✄Ù✁Ö❑å✠Ö✆ç✠Ö✚Ü✠å✠Ú✛à✠Ü✯â☛Õ✠Ö✢è✆ç✠ç✞Ù☛Û✁ã✚è✆â☛Û✁à✠Ü✯è✚Ü✄å✭â✁Õ✄Ö ✛✒ ☛✜ Ø✠Ú✥Ö✚×➲ò➲Ý✾à✠×✎ì✷à✞×✏Ö✵à✠Ü✷ì✷Ö✆è✚Ú✮Ø✄×✏Ö✆Ú✕à✞ä➧Û✁Ü✄â✁Ö✚×✜Ö✚Ú✮â☛Û✁Ü✄Þ✄Ü✄Ö✆Ú✥ÚPÚ✥Ö✆Ö✶ë ❂Û✁Ù☛ý✞Ö✚×✜Ú✥Õ✄è✆â☛þ ✣✢✬Û✁ã✚×✜à✠Ú✮à✠ä✏â✾ç✠Ù✁è✚â✁ä➩à✞×✏ì✢é Ô✕à✰å✞Ö✚÷✄Ö✆Ù☛à✞ç✰â✁Õ✄Û☛Ú✶á✶Ö✆ý✰Ö✆Ü✄÷✄Û✁×✏à✞Ü✄ì✷Ö✆Ü✄â✳á✶Ö✷ã✚Õ✄à✞Ú✥Ö➋è ✠✔ Ô✩Ø✄þ✆Õ✄Û✁Ù☛Û✁Ü✄ñ➲ò å✠Ø✠Ö✟â☛à✰â☛Õ✄Ö➋å✠Ö✚÷✠Ö✚Ù✁à✠ç✠ô ì✢Ö✚Ü✠â✾ý✠è✆ã ✄Þ✠×✏à✞Ø✄Ü✄å❛à✞ä❘â☛Õ✠Ö✟â☛Ö✆è✚ì✢é è✚Ü✄å✰è✚Ù✁Ú✥à❛ý✞Ö✚ã✚è✆Ø✄Ú✥Ö➋à✠ä❘â✁Õ✄Ö✟ç✞à✠Ú✥Ú✮Û☛ý✞Û☛Ù✁Û✁â☛Û✁Ö✚Ú◆à✞ä➩ä✏Ö✚×✜Ö✚å❛Û✁Ü✒â☛Ö✆×✏ì✢Ú✗à✠ä ✤✥✢✧✦➸å✠Ö✆÷✄Ö✚Ù✁à✠ç✞ì✷Ö✆Ü✄â✹ò✫Ô✩Õ✄Û✁Ú✟à✠ç✞â☛Û✁à✠Ü❼å✞à✠Ö✆Ú✟Ü✄à✠â✕ã✚à✞ì✷ç✞×✏à✞ì✷Û✁Ú✥Ö✒à✞Ø✄×◆Þ✄à✞è✚Ù❘à✞ä✗Õ✄è✚÷✄Û✁Ü✄Þ✭è✒ý✠×✜à✠á✵Ú✥Ö✆×✏ô ä✏×✏Ö✆Ö❑â☛à✞à✠Ù✹ò✄ú✗Ø✄×✜×✜Ö✚Ü✄â✁Ù☛ï✦é✳è✚Ù✁★ Ù ✾ ✏ ö✁✍☛✩✣❦✪ Ú✛ä➩Ö✚è✆â☛Ø✠×✏Ö✆Ú✵è✆×✏Ö❑Ú✮Ø✄ç✠ç✞à✠×✜â☛Ö✆å✯Û✁Ü✯ý✠à✠â✁Õ✣✰ ✫ Ö✚â✁Ú✥ã✚è✆ç✠Ö❑è✚Ü✠å✭✜✬ Ü✄â✁Ö✚×✜Ü✄Ö✚â ö✩æ✄ç✠Ù✁à✠×✜Ö✚×➲✮ ò ✜ ✬ Ü➹â☛Õ✄Ö❸ä✏à✠Ù✁Ù☛à✞á✶Û✁Ü✄Þ➹Ú✥Ö✚ã✆â☛Û✁à✠Ü✠Ú✯á✵Ö✬å✠Ö✚Ú✮ã✚×✜Û☛ý✞Ö❞â✁Õ✄Ö✬ì✷è✆Û☛Ü❬â☛Ö✆ã✚Õ✄Ü✠à✠Ù✁à✠Þ✄Û✁Ö✚Ú✯Û☛Ü✠÷✄à✠Ù✁÷✄Ö✚å➹Û✁Ü ✏➧ö✮✍✯➋ ✩ ò➲Ô✕Õ✄Ö✶Û✁Ü✄â✁Ö✚×✜è✚ã✆â☛Û✁à✠Ü✄ÚPè✚×✜Ö✵Ú✥Ø✄ì✢ì✷è✆×✏Û✁þ✚Ö✆å✷Û✁Ü➋Ý✾Û☛Þ✄Ø✠×✏☛ Ö ✳ ✰ ò ✱✆✲✴✳ ✷✶✠✝✎✸✓✹✺✸✆✻✽✼✿✾❁❀✞✼✽✟✎✝✠❀✞✟✎✼✿✾✺❀✞✻✛✸✓✝✎❂❄❃✓✼✛✷✸✓❀❄❅❆✟✠✝✠❇✆✟✠✝ ✘➸Ö✒Ø✄Ú✮Ö❑â☛Õ✄Ö✚✢❸Û✁ã✚×✜à✠Ú✥à✞ä➩★â ✜✬ Ü✄â✁Ö✚×✜Ü✄Ö✚★â ✜✬ Ü✄ä➩à✞×✏ì✢è✚â✁Û✁à✠❈ Ü ❂ ✒ Ö✚×✜÷✄Ö✚×✗ë✛✬✛✬✽✒❂ñ✗è✆✯ Ú ✾ ✏ ö✮✍✯✩✣✪❧Ú✛Õ✄â✁â✁ç✯Ú✮Ö✚×✜÷✄Ö✚×✗â✁à✭×✏Ø✠Ü â✁Õ✄Ö❉✍✛ã✚â✁Û☛÷✠Ö❈✾ ✒ Ö✚×✜÷✄Ö✚×✥➧✏ è✚Þ✄Ö✆Ú✢ë✛✍✯✒✞✾✏ ñ✼ä✏à✠×✼Ú✮Ö✚×✜÷✄Ö✚×✜ô➩Ú✮Û☛å✞Ö❍ç✞×✏à✞Þ✄×✏è✆ì✷ì✢Û✁Ü✄Þ✄é✩è✆Ù☛Ù✁à✠á✵Û✁Ü✄Þ➱å✞è✚â✁è✚ý✠è✆Ú✥Ö✯è✚Ü✄å ✤✥✢✧➫ ✦ ì✢è✚Ü✄Û✁ç✠Ø✠Ù☛è✆â☛Û✁à✠Ü❛è✆Ü✄å✰ä✏à✠×✜ì➬å✞è✚â✁è➋Ú✮Ø✄ý✠ì✢Û☛â✁â✁Ö✚å✰ý✞ï✒â☛Õ✠Ö➋Ø✄Ú✮Ö✚×➲ò❊✾✏ ö✮✍✯✩ è✆Ù☛Ú✮à❛×✏Ø✠Ü✄Ú◆à✠ä✏ä➩Ù✁Û☛Ü✠Ö✟á✶Û✁â✁Õ Ü✠à Ù☛Û✁ì✷Û✁â✁è✚â✁Û☛à✞Ü➾Ø✠Ü✄å✠Ö✆❋ × ✢ Û☛ã✆×✏à✞Ú✥à✠ä✏â●✾✏ Ö✆×✏Ú✮à✠Ü✄è✆Ù❍✘➸Ö✚ý✓✾✒ Ö✚×✜÷✄Ö✚×❼ë✛➸ ❸ ✘ Û☛Ü✄å✞à✠á✵Ú❸î✓✰✓✁■ ✓î ❏✆■✄❑✆▲✓▲✆▲✓■✷✬ ✢ Ö✆ñ❼à✠× Ø✠Ü✄å✠Ö✆×★✬ ✢ Û✁ã✚×✜à✠Ú✮à✠ä✏▼â ✾✏ Ö✚Ö✆×★➸ ✘ Ö✚ý✚❂✒ Ö✚×✜÷✄Û✁ã✚Ö✆ÚPë✛✘➸Û☛Ü✠å✠à✠á✵Ú✌✰ ✫ Ô✧➹ ✘ à✞×✽✔✠Ú✥â✁è✚â✁Û☛à✞Ü✄ñ ò✚Ô✩Õ✠Û☛Ú❭ì✢Ö✚è✚Ü✠ÚPÛ✁â❈ã✚è✆Ü➋ý✠Ö Û✁Ü✄Ú✮â☛è✆Ù☛Ù✁Ö✚å✢Û☛Ü✷è✚Ü✄ï✣✾ ✏ ú➼á✵Û☛â✁Õ➋è☛❸ ✢ Û☛ã✆×✏à✞Ú✥à✞ä➩â❈Ú✥ï✦Ú✮â☛Ö✆ì➾Û✁Ü➋Û✁â❈✽ë ➸ ✘ Û✁Ü✄å✠à✞á✶ÚPî✓❏✆■✄✢❸Ö✎■ ✰ ✫ Ô◆■✄❑✆▲✓▲✓▲✆■✄✤✥➧✏ ñ ò ✵ ✱✆✲✴❖ ✶✎✼✽✷❇✓✟☛❅✞✟✠✝✎❇✓✟✎✝❘◗✮❃✆✂✓✟✎✹◆❃✆❀✞❙❄❚✥❯✞❅❆✶✠✝✠✷❱✞✼ P ✍✛ã✆â☛Û✁÷✄❍ Ö ✒❂Ö✆×✏÷✄Ö✆❲ × ✏✾è✆Þ✄Ö✚Ú✼ë✛✍✯✒✞✏✾ñ➤ë✽❸ ✢ Û✁ã✚×✜à✠Ú✥à✞ä➩â✁ñ◆è✚×✜Ö❑å✠ï✦Ü✄è✆ì✷Û✁ã❑è✚Ü✄å✭Û✁Ü✄â✁Ö✚×✜è✚ã✚â✁Û✁÷✄Ö❑á✶Ö✆ý✯ç✞è✚Þ✄Ö✆Ú✛ç✠×✜à✠ã✚ô Ö✆Ú✥Ú✮Ö✚å➼à✞Ü✬â✁Õ✄Ö❼Ú✮Ö✚×✜÷✄Ö✚×✜ô➩Ú✮Û☛å✞Ö✚é❭â☛Õ✠Ø✄Ú✒Ø✄Ú✥Ö✆ä➩Ø✄Ù➤â☛à➫ì✢è✚Ü✄Û✁ç✠Ø✠Ù☛è✆â☛Ö❼å✞è✚â✁è❼Ú✥Ø✄ý✞ì✷Û✁â☛â✁Ö✚å➫ý✞ï❞Ø✠Ú✥Ö✚×✜Ú✒ë✜ä➩à✠×➋Û☛Ü✄ô 50 Jorge et al. ❳❁❨✄❩✎❬✗❭✠❪✎❫◆❳❁❪✠❴✷❪✠❭✎❨✄❵✷❬✗❛❜❩❈❳✺❪✎❨❞❝✓❡☛❩✠❳✺❳❁❝✓❭✎❵✄❩✎❨✄❵✷❝✓❬❣❢✽❤✗❴✷❪✠❳✐❛✗❵✷❥✗❪✠❬❣❭✠❪✎❢✽❨✷❩✠❵✷❬❜❢✛❪✠❳❁❨✄❢✛❵✄❭✎❨✄❵✷❝✓❬✓❳❧❦✆♠❋❨✄♥✓❪✭❤✗❳❁❪✠❢✛❳✺♦✯❩✠❳ ♣ ❪✠❴✷❴✕❩✠❳rq❧❩✎❬✗❵✷s✓❤✗❴✷❩✠❨✷❵✄❬✓❛✣t✓❩✎❨✄❩✎❦✓❩✎❳✺❪☛❢✽❪✎✉✓❤✗❪✎❳✺❨✷❳✇✈ ① ❬ ①✯②✞③ s✓❩✠❛✓❪☛❵✷❬✗❨✷❪✠❛✗❢✛❩✠❨✷❪✠❳❞④✥⑤◆⑥⑧⑦❣❨✄❩✎❛✗❳ ♣ ❵✷❨✷♥❧❳❁❭✠❢✛❵✄s✆❨▼❭✠❝✓q✐q✐❩✠❬✗t✆❳✇✈✠⑤✁♥✗❪✎❳✺❪☛❳✺❭✠❢✛❵✷s✓❨✷❳◆❭✎❩✠❬✣❦✆❪❘❪✎❵✄⑨ ❨✷♥✗❪✎❢★⑩✥❦ ② ❭✎❢✽❵✷s✓❨▼❝✆❢★❶❁❳❁❭✠❢✛❵✄s✆❨▼❷✽❶❸❩✠❥✗❩ ② ❭✠❢✛❵✷s✓❨▼❳✺❵✷q✐❵✄❴✷❩✠❢✛♦❹✈✠⑥✧❵✄❭✎❢✽❝✆❳✺❝✆❡❺❨▼❶ ② ❭✠❢✛❵✄s✆❨▼❵✄❳❞❩✎❬✣❝✓s✆❪✠❬❧❵✄q✐s✓❴✷❪✠q✐❪✠❬✗❨✷❩✠⑨ ❨✷❵✷❝✓❬❻❝✓❡✙❼❄❪✎❨✄❳❁❭✠❩✎s✓❪✠❽ ❳✥❶❸❩✠❥✗❩ ② ❭✠❢✛❵✷s✓❨ ♣ ♥✗❵✷❭✠♥❉❩✠❢✛❪❍❦✓❝✆❨✄♥❉❭✠❝✆q❧s✆❴✄❵✷❩✠❬✓❨ ♣ ❵✷❨✷♥❈❨✷♥✗❪❍❾✮❤✓❢✽❝✆s✓❪✠❩✎❬❉❿➀❝✓q✐s✓❤✗❨✷❪✠❢ ⑥✧❩✎❬✗❤✗❡❺❩✎❭✠❨✷❤✗❢✛❵✄❬✗❛ ① ❳✺❳❁❝✓❭✠❵✷❩✠❨✷❵✷❝✓❬✌➁✴❳☛❾✁❿➀⑥ ①✯② ❭✎❢✽❵✷s✓❨★⑦★❩✎❬✗❛✗❤✗❩✎❛✗❪ ② s✓❪✎❭✠❵✷❡❺❵✷❭✠❩✎❨✄❵✷❝✓❬✭❷✽❾✮❿➀⑥ ① ⑨❺➂✆➃✓➂❈❳✺❨✷❩✠❬✗⑨ t✆❩✠❢✛t➅➄❺♦❞➆➇♥✗❪✠❬✭❨✄♥✓❪❧s✆❩✠❛✗❪✐❵✄❳☛t✓❝ ♣ ❬✓❴✄❝✆❩✠t✓❪✎t✓❫✌❨✄♥✗❪✎❳✺❪✐❳✺❭✎❢✽❵✷s✓❨✷❳❘❩✎❢✽❪✐❪✠➈✓❪✠❭✠❤✓❨✄❪✎t✭❝✆❬❄❨✷♥✗❪ ① ❭✎❨✄❵✷❥✗❪ ② ❪✠❢✛❥✗❪✠❢ ③ ❩✠❛✗❪☛❪✠❬✓❥✗❵✄❢✛❝✓❬✓q❧❪✎❬✗❨✕❨✄♥✗❤✓❳◆s✆❢✽❝✆t✓❤✗❭✎❵✄❬✗❛❧❨✄♥✓❪❘❡✽❵✄❬✗❩✎❴✕④✥⑤✁⑥✧⑦❋❭✠❝✆t✓❪☛❨✄❝✐❨✷♥✗❪❘❢✛❪✠✉✆❤✗❪✠❳❁❨✄❵✷❬✗❛❧❦✓❢✛❝ ♣ ❳✺❪✠❢➉✈ Ú➉Û✽Ü✛Ý Þ ß✄àÐß Û✛á❮â ã ä➷å Ý æ✠ç✑è Ü✽è✕è✕í î➷éêá❮ï➷ð✎ë❮ñÐå ✚ Ú✜✛❆ð✠éêî➷ë✙✢ ✣ Þ ß❮àÐß Û✛á❮â ã ä➷å Ý æ✑ç✑è Ú✜✛✞æ ß å ß✙✤❁ß ì ë Ú➉Û✽Ü ✚ ✱Û ✰✠í✆â ë❮ì ï☎✢ å ✣ Ú➉Û✽Ü ✚ ✱Û ✠✰ í✆â ë★✲➷ï➷ë❮ì å ✣ Ü✛â éêá✄ë❮ì ì Ü✛è➛è✕í➛î➷éêá❮ï➷ð✎ë❮ñÐå ø Ú➉Û✛Ü✽Ý Þ ß❮àÐß Û✛á❮â ã ä➷å Ý æ✑ç✑è ò❹ó✄ô❒õêöê÷✷÷ ÷✄÷✷ô❒õêù ú❒û ù ô❒ü✕ý➉þ❒ÿ öê÷ Ú➉Û✛Ü ✂✁☎✄✝✆ ✞☎✁✠✟ ✡☛✟ ☞☎✌ ✍☎✎☎✞☎☞ ✏✠✑ ✑ ✍☎✒✝✆ ✁✠✟ ✆ ✍☎✓✕✔✖✎✠✗ ✡ ✑✠✑ ✘ ✁☎✒✙✡ Ú➉Û✛Ü✽Ý Þ ß❮àÐß Û✽á✄â ã ä➷å Ý æ✑ç✑è Ú➉Û✛Ü✛Ý Û✭✬✮✩✑Ý Þ ß✄à✷ß Û✽á✄â ã ä➷å ✚ æ✑ç✑è✯✣ ✩✑â ß✙✪ ãá ß ✢ Ú➉ñ ß ✢ ✫✄ì ã ì ✥✖✦ ➷ä éêâ å✛å é❊ñÐë★✧ Ü✛è➛è✕í î➷éêá✄ï➷ð✠ë❮ñÐå ➊✁➋✄➌✆➍✞➎✠➏☛➐✙➑✓➒ ❪✠❬✓❪✠❢✛❩✠❴✕❩✠❢✛❭✠♥✗❵✷❨✷❪✠❭✠❨✷❤✗❢✛❪❘❝✆❡ ③ ❾ ①✯➓ ✈ ③ ❾ ①✯➓ ❤✗❳❁❪✠❳r⑩✥❦ ② ❭✎❢✽❵✷s✓❨✕❷✛❵✄❬ ① ❭✠❨✷❵✄❥✗❪ ② ❪✎❢✽❥✗❪✎❢ ③ ❩✠❛✓❪✠❳✺♦✿❨✷❝❧s✆❢✽❝✆❭✠❪✎❳✺❳r❨✄♥✗❪☛❳✺❪✎❨✕❝✓❡✙❩✠❳✺❳❁❝✓❭✎❵✄❩✎❨✄❵✷❝✓❬❧❢✽❤✗❴✷❪✠❳ ❢✛❪✠s✆❢✽❪✎❳✺❪✎❬✗❨✄❪✎t❄❵✷❬●❩ ③ ⑥✧⑥⑧⑦➔t✆❝✓❭✎❤✗q❧❪✎❬✗❨✙❷✽❤✗❳❁❵✄❬✓❛●→✥❝✓❭✎❤✗q❧❪✎❬✗❨✙➣✥❦➉↔↕❪✠❭✎❨✌⑥✧❝✓t✆❪✠❴✷♦✽❫❆❨✄❝✭❩✠❴✷❴✄❝ ♣ ❨✷♥✗❪➙❤✗❳❁❪✠❢ ❨✷❝❍❦✓❢✛❝ ♣ ❳✺❪✥❨✷♥✗❢✛❝✓❤✗❛✓♥❧❵✷❨✄❫➛❩✠❬✓t✐❨✷❝✚❳✺❨✷❝✓❢✛❪✯❨✄♥✗❪✯❢✛❤✗❴✷❪✠❳❞❵✄❬✐❩✯❢✽❪✎❴✄❩✎❨✄❵✷❝✓❬✓❩✠❴▼t✓❩✎❨✄❩✎❦✓❩✎❳✺❪➛✈✠⑩✥❦ ② ❭✠❢✛❵✄s✆❨▼❵✄❳❞❤✗❳❁❪✠t✚❩✠❨ ❨✷♥✗❪☛❳✺❪✎❢✽❥✗❪✎❢✽⑨✽❳✺❵✷t✓❪☛❝✓❬✗❴✷♠❆✈ ➎✎➋✄➥❆➦ ➟✆➠✓➡✓➠✆➢✞➤ ❶❸❩✠❥✓❩ ② ❭✎❢✽❵✷s✓❨✷❫✌❷✛⑥✧❵✄❭✎❢✽❝✆❳✺❝✆❡❺❨★➆➇❪✠❦ ② ❵✷❨✄❪✎♦➀❵✄❳✯❤✗❳❁❪✠t❈❡❺❝✆❢rt✆❩✠❨✷❩❧q✐❩✠❬✓❵✄s✆❤✗❴✄❩✎❨✄❵✷❝✓❬✭❝✓❬✭❨✄♥✗❪✐❭✠❴✷❵✷❪✠❬✗❨✷⑨❺❳❁❵✄t✆❪✠❫❆❡✽❝✓❢ ❵✷❨✷❳❍s✓❝✓❢✛❨✷❩✠❦✓❵✷❴✷❵✄❨✷♠❆✈✌➆➧❪❻❨✷❢✛♠❣❨✷❝➔❤✗❳✺❪❻❝✆❬✗❴✷♠❣❭✎❝✓q✐q❧❩✎❬✗t✓❳✚❨✷♥✗❩✠❨➀❩✎❢✽❪❉❭✠❝✆q❧s✆❴✄❵✷❩✠❬✓❨ ♣ ❵✷❨✄♥⑧❾✮❿➀⑥ ①✯② ❭✠❢✛❵✷s✓❨➨✈ ⑤◆♥✗❵✷❳ ♣ ❩✎♠ ③ ❾ ①✯➓ q❧❩✎♠❉❢✽❤✗❦❜❤✗❬✗t✓❪✎❢★❼❄❪✎❨✄❳❁❭✠❩✠s✆❪●❩✠❬✓t❻➩✽❬✗❨✷❪✠❢✛❬✗❪✎❨✮❾✮➈✗s✆❴✄❝✆❢✽❪✎❢❹✈✆➆➇❵✄❨✷♥❉❶❁❩✎❥✗❩ ② ❭✠❢✛❵✄s✆❨ ♣ ❪ ❭✎❢✽❪✎❩✠❨✷❪➫❩✠❬✓t➭q❧❩✎❬✗❵✷s✓❤✗❴✷❩✠❨✷❪ ③ ⑥✧⑥✧⑦➯t✓❝✓❭✎❤✗q✐❪✠❬✗❨✷❳➔❝✓❢ ② ⑩ ➒ ❷✛❦✓❝✆❨✄♥➳➲✥⑥✧⑦➵t✆❝✓❭✎❤✗q❧❪✎❬✗❨✷❳✺♦❋❤✗❳❁❵✄❬✓❛ ➜✆➝✴➞ ➄❞➸❆➺★➻❄➼➾➽✛➚✮➪❸➶❄➽✴➶✇➹✴➘❸➴➷➶✎➪❸➹✴➽✴➬✇➶✇➪❁➮★➽✴➶✇➱✇✃✕➚✄➹✴➴➷❐❍➪➅➚❒➚❮➬✇❰❸➽✴➪❁➹❮➽✴➬✎➶✭ÏÐ➬✎✃✇➶✎➱✇➘❸➱❄➽✴➶✧Ñ↕Ò✇Ó✕Ñ➙➪❁➶✇➱❄➱✇➘❸➱✎➽✴❰❸➪❸➹✴➘❁➱✭➹✴➬❄➹❮Ô✎➘❉➚❮➹❮➪❁➶✇Õ ➱✎➪❸➴➷➱✎➽❮Ö❁➪❸➹✴➽✴➬✇➶×➬✇Ï✆➽❮➶✎ÏÐ➬✇➴➷Ø×➪❸➹✴➽✴➬✇➶×➪❸➶✎➱❘❰❁➬✇Ø×Ø❘✃✎➶✇➽✴❰❸➪❁➹✴➽❮➬✎➶✚➚❮❐➛➚❮➹❮➘❁Ø✚➚✛Ù A Post-Processing Environment 51 ✳✵✴✷✶✹✸✻✺✽✼✹✾☛✿❁❀✵❂✜❃✖✼❄✶✹✿❁❅❆✴✷❇☛✼✹❈❊❉●❋■❍✹❏✻❍❄❑▲✶✹▼✭◆✙❖✷✿❁◆◗P❘❍✹❈◗P✠✴❙◆✙✺✽❖☛✴✷▼✱✿◗❍✹✾✻✿❁❚❯✴☛▼❱❇☛❍✹✿◗❍❲❏✻❍❄❈✙◆◗❇☛❍❄✿✙◆◗✴☛✾❳❍✹✾☛❇❨❚❯✴☛▼❱◆✙✾✻❩ ✿◗✼✹▼✭❍✹✶❄✿✙◆◗✴☛✾❭❬❪◆◗✿◗❫❴✿◗❫✻✼❪✸☛P✠✼✹▼❛❵✭✼✹❏✻✼❄✾✻✿✯❫✻❍✹✾☛❇☛❈◗◆✙✾✻❜☛❝❞❉ ❡✷❢❤❣ ✐✵❥✷❦✹❧▲♠♦♥✹♣▲q❛r❲s✉t✜♥✹❦❄q✇✈①❥✷②▲♥✹③ ④⑤❫✻✼❙✳✵✴✷✶✹✸✻✺✽✼✹✾☛✿❪❀✵❂✮❃❊✼❄✶✹✿❪❅⑥✴✷❇☛✼❄❈❱❵✭✳✵❀✵❅⑥❝❭◆✙P⑦❍⑧✿✙▼✭✼✹✼❙P✠✿◗▼✱✸✻✶❄✿✙✸☛▼✱✼❄❩❯❂☛❍❄P✠✼❄❇⑨❍✹❖✷❖☛❈◗◆✙✶❄❍✹✿◗◆✙✴✷✾⑩❖☛▼✭✴☛❜✻▼✭❍✹✺ ◆◗✾✻✿◗✼✹▼✭❚❯❍❄✶✹✼✵❵✭❶❸❷❺❹✱❝❁❚❯✴☛▼❁❻✵④❁❅❆❼❽❍❄✾✻❇❲❾✵❅❆❼❽❇✷✴☛✶❄✸✻✺❭✼❄✾✻✿◗P❿◆✙P✂P✠✸✻✼❄❇✽❍❄P➀❍❸➁➃➂✷➄➆➅❿✼✹✶❄✴☛✺✽✺❭✼❄✾✻❇☛❍❄✿✙◆◗✴☛✾✽◆◗✾ ❀✵✶❄✿✙✴✷❂☛✼❄▼❱➇✜➈☛➈✷➉➊❵✱➁➃➂☛➄➋✳✵❀✵❅➌❼➍✼✹❏☛✼✹❈❁➇❪P✠❖✷✼✹✶✹◆◗❚❯◆◗✶✹❍❄✿✙◆◗✴☛✾☛❝❞❉✹❹✱✿➎◆◗P➀✸✻P✠✼❄❇➊✿✙✴➊❖☛▼✭✴☛✶❄✼✹P✂P⑤❍❄✾✻❇❭✺✽❍✹✾☛◆✙❖✷✸✻❈✙❍❄✿✙✼ ❍❄✾➏❾✵❅❆❼➐❇☛✴☛✶❄✸✻✺✽✼✹✾✻✿❲❂✷➑➒❍✹✶❄✶✹✼❄P✠P✠◆◗✾✻❜➓◆✙✿◗P❆◆✙✾☛✿✙✼❄▼✱✾✻❍❄❈❲P✠✿◗▼✱✸☛✶✹✿◗✸✻▼✱✼✯❉➔✳✵❀✵❅→▼✱✼❄❖☛▼✭✼✹P✠✼❄✾✻✿◗P❽❍❄✾➒❾✵❅❆❼ ❇✷✴☛✶❄✸✻✺❭✼❄✾✻✿➣❍❄P➔❍❘✿◗▼✱✼❄✼●❉✕❹✭✿✙P❱✾✻✴☛❇✷✼✹P➔❍❄▼✱✼❘✼❄❈✙✼❄✺❭✼❄✾✻✿◗P✠↔➎✿◗✼✹↕✻✿◗↔➎❍✹✾✻❇⑦P✠✴❲✴☛✾✇❉✹✳✵❀✵❅➙✺❭❍❄➛✻✼✹P❿◆◗✿▲✶❄✴☛✾✻❏✻✼❄✾✻◆◗✼✹✾✻✿ ❚✱✴☛▼❪❍✹❖☛❖✷❈✙◆◗✶✹❍❄✿✙◆◗✴☛✾❳❖✷▼✱✴✷❜✻▼✱❍❄✺❭P✵✿◗✴❨✿◗▼✱❍❄❏✻✼✹▼✭P✠✼❲✿◗❫✻✼❲✿◗▼✱✼❄✼❲❍✹✾✻❇❨❍❄✶✹✶✹✼❄P✠P✵✿◗❫✻✼❲✶✹✴✷✾✻✿◗✼✹✾✻✿◗P✵✴☛❚❿✿◗❫✻✼❲✿◗▼✱✼❄✼●❉✷④❁❫✻✼ ✳✵✴✷✶✹✸✻✺✽✼✹✾☛✿❿❀✵❂✜❃✖✼✹✶❄✿❿❅⑥✴✷❇☛✼❄❈❿❖☛▼✭✴☛❏✻◆◗❇☛✼❄P✽❍❳P✂✿✙❍❄✾✻❇☛❍❄▼✱❇❆❖☛▼✭✴☛❜☛▼✱❍❄✺❭✺✽◆✙✾☛❜❽✺✽✴☛❇✷✼✹❈❿❚❯✴✷▼✵❬❪✴✷▼✱➛☛◆✙✾✻❜❽❬❪◆◗✿◗❫ ❾✵❅❆❼❱❉ ❷❺➜➝❶➞➅➟✸✻P✂✼✹P✽✿✙❫✻✼➠❅❆◆✙✶❄▼✱✴✷P✠✴✷❚❯✿➀❾✵❅❆❼➡❖☛❍❄▼✱P✂✼✹▼❸❖☛▼✭✴☛❏✻◆◗❇☛✼❄❇❙◆◗✾⑧❅❆◆✙✶❄▼✱✴✷P✠✴☛❚✱✿⑤❹✭✾✻✿◗✼✹▼✭✾✻✼✹✿➢➜➝↕✻❖✷❈✙✴✷▼✱✼❄▼❪➤ ❵✭❍✹✾☛❇➥❍✹❂✷✴☛❏✻✼❄❝✱↔➎❬❸❫✻◆◗✶✹❫❲◆◗✺❭❖✷❈✙✼❄✺❭✼❄✾✻✿✙P❱✿✙❫✻✼❴➁➦➂✷➄➃✳✵❀✵❅➙P✂❖☛✼✹✶❄◆✙❚✱◆✙✶❄❍✹✿◗◆✙✴✷✾❺❉✕➁➃◆✙✿◗❫❲✿◗❫✻◆✙P❱❖☛❍✹▼✭P✠✼❄▼❁❬❸✼✵✶✹❍❄✾ ✼❄❍✹P✂◆✙❈◗➑⑦❍❄✶✹✶✹✼❄P✠P➞❍✹✾☛❇♦✺✽❍✹✾✻◆◗❖☛✸☛❈✙❍❄✿✙✼✽✿✙❫☛✼❭◆◗✾✻✿◗✼✹▼✭✾✻❍✹❈❛✿◗▼✱✼❄✼✹❩❯P✂✿✙▼✭✸✻✶❄✿✙✸✻▼✭✼✽✴☛❚➢❍✹✾⑦❾✵❅❆❼⑨❇✷✴☛✶❄✸✻✺❭✼❄✾✻✿❊❉➧❹✭✾⑦❖☛❍❄▼✱❩ ✿◗◆◗✶✹✸✻❈◗❍✹▼✭↔➎❬❪✼❴✸✻P✠✼❘✿◗❫✻✼❘✳✵❀✵❅➨✿◗✴➥▼✱✼❄❍✹❇➥❍❄✾✻❇➥✺✽❍✹✾✻◆◗❖☛✸☛❈✙❍❄✿✙✼❘✿◗❫✻✼❘✴✷▼✱◆◗❜✻◆◗✾✻❍✹❈❺❷➣❅❆❅❆❼❆❇☛✴☛✶❄✸✻✺✽✼✹✾✻✿▲❵✭❾✵❅❆❼ ❇✷✴☛✶❄✸✻✺❭✼❄✾✻✿▲✿◗❫✻❍✹✿➣▼✱✼❄❖☛▼✭✼✹P✠✼❄✾✻✿◗P❿❍❸❇✷❍✹✿◗❍➞✺❭◆◗✾✻◆◗✾✻❜❭✺✽✴☛❇✷✼✹❈◗❝✱↔●✿◗✴✽✼❄↕✻❖☛✴✷▼✱✿➎❍➞✾✻✼❄❬①❷➣❅❆❅❆❼❙❇✷✴☛✶✹✸☛✺❭✼❄✾✻✿➎❍✹✾✻❇ ❍❄❈✙P✂✴❭✿◗✴✽✶✹▼✭✼✹❍✹✿◗✼❸❍✹✾✻❇✽✺❭❍❄✾✻◆◗❖☛✸✻❈◗❍✹✿◗✼❪✿◗❫✻✼❸❜✻▼✭❍✹❖☛❫☛◆✙✶❄❍✹❈✯❏✻◆◗P✠✸✻❍❄❈✙◆◗➩✹❍❄✿✙◆◗✴☛✾❭❵✱❑▲➫✵➭➒❇☛✴☛✶❄✸✻✺✽✼✹✾✻✿◗P✠❝✜❉ ❡✷❢❤➯ ➲▲❦❄➳☛③◗➳☛s▲③◗♥❸➵✵♥✹❦✹q✭❥☛➸❸➺➊➸✹➳✷➻▲➼▲➽◗❦✹➾ ❑➣✶✹❍✹❈◗❍✹❂✷❈✙✼✵➫✵✼❄✶✹✿◗✴☛▼➝➭✵▼✭❍✹❖✷❫✻◆◗✶✹P❿❵✭❑▲➫✵➭✵❝➝◆◗P➢❍❄✾❭❾✵❅❆❼➍❩✱❂☛❍✹P✂✼✹❇➊❈◗❍✹✾✻❜✻✸☛❍✹❜✻✼➞✿◗❫✻❍✹✿➎P✂❖☛✼✹✶❄◆✙❚✱◆✙✼❄P➀❍✹✾✻❇➊❇☛✼❄❚❯◆◗✾✻✼✹P ❏☛✼✹✶✹✿◗✴☛▼❴❜✻▼✱❍❄❖☛❫✻◆◗✶✹P❲✿◗❫✻❍❄✿➔✶✹❍❄✾⑥❂✷✼❨❏✻◆◗P✠✸✻❍❄❈✙◆◗➩✹✼❄❇⑩❂☛➑⑥❍➠❬❸✼✹❂❆❂✷▼✱✴✷❬❪P✂✼✹▼✜❉⑤➚★➁➃➂☛➄➪➅❿✼❄✶✹✴✷✺❭✺✽✼✹✾✻❇✷❍✹✿◗◆✙✴✷✾✻➶✜❉ ➹▲➘ ❇☛✼❄❚❯◆◗✾✻✼✹P➴✿✙❫☛✼➐❚✱✼✹❍✹✿◗✸✻▼✭✼✹P➴❍✹✾✻❇➨P✠➑➧✾☛✿✙❍❄↕➷❚❯✴✷▼➬❑▲➫✵➭✵↔➆❍➐❈✙❍❄✾✻❜✻✸✻❍❄❜✻✼➮❚✱✴☛▼➓❇☛✼✹P✂✶✹▼✭◆✙❂✷◆✙✾☛❜➌✿✙❬❸✴☛❩ ❇✷◆✙✺✽✼✹✾☛P✠◆◗✴☛✾✻❍❄❈➣❏✻✼❄✶✹✿◗✴☛▼❁❍❄✾✻❇➥✺✽◆✙↕☛✼✹❇➥❏✻✼❄✶✹✿◗✴☛▼✭➱✙▼✭❍✹P✂✿✙✼❄▼❁❜✻▼✭❍✹❖✷❫✻◆◗✶✹P➔◆◗✾❲❾✵❅❆❼➍✃➎❉✕❑➣✴☛↔➎✸✻P✂◆✙✾✻❜➊❑▲➫✵➭➓◆◗P❿❏✻✼✹▼✭➑ P✂◆✙✺✽◆◗❈✙❍❄▼❸✿◗✴⑥❬❪✴✷▼✱➛✻◆◗✾✻❜❽❬❪◆◗✿✙❫❽❍✹✾✻➑❨✴✷✿✙❫☛✼✹▼❸✾✻✴☛▼✭✺❭❍❄❈⑤❾✵❅❆❼①❇✷✴☛✶✹✸☛✺❭✼❄✾✻✿✝❉➎❶❸✾❙❑▲➫✵➭❐❇☛✴✷✶✹✸✻✺✽✼✹✾✻✿➢✺✽✸✻P✠✿ ❍❄❈✙P✂✴❨❚❯✴✷❈✙❈◗✴☛❬➟✿◗❫✻✼❒✳✵④❁✳❮❵✱✳✵❍❄✿✙❍➊④❁➑➧❖✷✼➊✳✵✼✹❚❯◆◗✾✻◆◗✿✙◆◗✴☛✾☛❝❿✿✙❫✻❍❄✿➍P✠❖✷✼✹✶✹◆◗❚❯◆◗✼✹P➞✿◗❫✻✼➊❜✻▼✭❍✹❖☛❫☛◆✙✶➊✼✹❈◗✼✹✺✽✼✹✾☛✿✙P➞✿◗❫✻❍✹✿ ✶❄❍✹✾❭❂☛✼❸❖☛▼✭✴☛❇☛✸☛✶✹✼✹❇❛❉ ❶➞❜✻❍❄◆✙✾☛↔➠❬❸✼❰✶✹❍❄✾❮✺✽❍✹✾☛◆✙❖✷✸✻❈✙❍❄✿✙✼➬❑▲➫✵➭Ï❜✻▼✭❍✹❖☛❫☛◆✙✶❄P➦❬❸◆◗✿✙❫➴➫✵❂☛❑▲✶❄▼✱◆◗❖☛✿❨✴✷▼⑩❋✂❍❄❏✻❍✹❑➣✶✹▼✭◆✙❖✷✿❨❵✱✸✻P✂◆✙✾☛❜ ✳✵✴✷✶✹✸✻✺✽✼✹✾☛✿➢❀✵❂✮❃✖✼✹✶✹✿➀❅❆✴☛❇✷✼✹❈◗❝❞❉▲➁➃◆✙✿◗❫❙❑▲➫✵➭✵↔⑤◆✙✿➢◆◗P❴✼✹❍❄P✠➑❨✿◗✴❙❖✷▼✱✴✷❇☛✸✻✶❄✼⑦❍⑦❇✷❍✹✿◗❍⑦❏✻◆◗P✠✸☛❍✹❈◗◆✙➩❄❍✹✿◗◆✙✴✷✾⑧❍✹✾✻❇ ✼❄❏✻✼✹✾✽✺✽❍✹➛✻✼➞◆◗✿➎◆✙✾✻✿◗✼✹▼✭❍✹✶❄✿✙◆◗❏✻✼➞❵✭✶✹✴☛✾☛✿✙▼✭✴☛❈◗❈✙◆◗✾✻❜✽➛✻✼✹➑➧❂✷✴☛❍❄▼✱❇➊✴☛▼➍✺✽✴☛✸☛P✠✼➞✼✹❏✻✼❄✾✻✿◗P✠❝✜❉✮❷❺➜➝❶➞➅➆❜✻✼❄✿✙P➢❇☛❍❄✿✙❍❸❚❯▼✭✴☛✺ ❷❺❅❆❅⑥❼❙❍✹✾✻❇✽❖☛▼✭✼✹P✂✼✹✾✻✿◗P⑤◆◗✿✯✸✻P✠◆◗✾✻❜❭➫✵❂☛❑▲✶❄▼✱◆◗❖☛✿✯❍✹✾☛❇❭❑▲➫✵➭➒❜✻▼✱❍❄❖☛❫✻◆◗✶✹P✉❉ ❡✷❢❤Ð ✐✵➳✷q✱➳✷s▲➳☛➾✂♥❪➳✷♣▲②⑦➲▲Ñ➊Ò ➁➃✼❲✸✻P✂✼❲❍❲▼✱✼❄❈✙❍❄✿✙◆◗✴☛✾☛❍✹❈➝❇✷❍✹✿◗❍✹❂☛❍❄P✠✼❲❵✭❅❆◆✙✶❄▼✱✴✷P✠✴☛❚✱✿➝❶➞✶✹✶❄✼✹P✠P✂❝➔✿◗✴❨P✠✿◗✴☛▼✭✼❲✿✙❫✻✼❲❷❺❅❆❅⑥❼➦✺✽✴☛❇☛✼❄❈➍❍✹✾✻❇❳✿◗❍✹➛✻✼ ❍❄❇☛❏✻❍❄✾✻✿◗❍✹❜✻✼❽✴☛❚❘✸✻P✂◆✙✾☛❜⑩❑▲✿✙▼✭✸✻✶❄✿✙✸✻▼✭✼✹❇➆Ó✵✸✻✼✹▼✭➑❆❼➍❍✹✾✻❜✻✸☛❍✹❜✻✼⑧❵✭❑▲Ó✵❼➍❝❭✿✙✴⑨✴✷❂☛✿◗❍✹◆◗✾⑩P✠✼❄✿✙P➥✴✷❚❘❍✹P✂P✠✴✷✶✹◆◗❍✹✿◗◆✙✴✷✾ ▼✭✸✻❈◗✼✹P✉❉➧➄❿✴☛✺✽❖☛❍❄▼✱✼❄❇♦✿◗✴➠✸✻P✠◆◗✾✻❜⑦✳✵❀✵❅Ô❇✷◆✙▼✭✼✹✶❄✿✙❈◗➑➥✿◗✴⑦✺✽❍✹✾✻◆◗❖☛✸☛❈✙❍❄✿✙✼❭✿✙❫✻✼❭✴☛▼✭◆✙❜☛◆✙✾✻❍❄❈❺❷❺❅❆❅⑥❼⑩❇✷✴☛✶❄✸✻✺❭✼❄✾✻✿◗↔ ❑➣Ó✵❼➡❖☛▼✭✴☛❏✻◆◗❇☛✼❄P❴❍♦❚❯❍✹P✂✿✙✼❄▼❪❍❄✾✻❇❙✼❄❍✹P✂◆✙✼❄▼❪❍❄✶✹✶✹✼❄P✠P✉❉✯❹✱✾❙❷➣➜➝❶➞➅❿↔➝❍❄❈✙❈➢❇☛❍❄✿✙❍❄❂☛❍✹P✂✼⑦✶❄✴☛✾✻✾✻✼❄✶✹✿◗◆✙✴✷✾✻P❴❍❄✾✻❇❙▼✭✼✹❩ Õ ✸✻✼✹P✂✿✙P➢❍✹▼✭✼❸❇☛✴☛✾☛✼❪❬❸◆✙✿◗❫❴❶➞✶❄✿✙◆◗❏✻✼❸❑▲✼✹▼✭❏✻✼✹▼❛❷❺❍✹❜✻✼❄P⑤✴✷✾❴✿◗❫✻✼❪P✂✼✹▼✭❏✻✼✹▼❛P✂◆✙❇✷✼●❉ ❡✷❢❤❡ Ö✵♥❄➻▲➸✹♥❄➾✠♥✹♣➣q✱➽◗♣▲×✽Ø✵➾✠➾✠❥✷❦✹➽◗➳☛q✭➽✙❥✷♣▲➾➢Ö✵❧▲③✙♥❄➾⑤Ù❴➽✙q✭➼⑦Ú➝✈①✈①Ò ❷❺▼✱✼❄❇☛◆◗✶✹✿◗◆✙❏☛✼❲❅⑥✴✷❇☛✼❄❈➝❅❆❍✹▼✭➛✻✸✻❖❨❼➍❍✹✾☛❜✻✸✻❍✹❜☛✼❲❵✱❷❺❅⑥❅❆❼➍❝➔◆◗P✵❍✹✾❳❾✵❅❆❼➍❩❯❂✷❍✹P✂✼✹❇❨❈◗❍✹✾✻❜✻✸☛❍✹❜✻✼✯❉☛❶❮❷➣❅❆❅❆❼ ❇✷✴☛✶❄✸✻✺❭✼❄✾✻✿❱❖✷▼✱✴✷❏✻◆◗❇☛✼✹P⑦❍⑧✾✻✴☛✾☛❩❯❖☛▼✭✴☛✶❄✼✹❇✷✸✻▼✱❍❄❈❱❇✷✼✹❚❯◆◗✾✻◆◗✿✙◆◗✴☛✾⑩✴✷❚❘❚❯✸☛❈✙❈◗➑❆✿✙▼✭❍✹◆◗✾✻✼✹❇➆❇☛❍✹✿◗❍❨✺✽◆✙✾✻◆◗✾✻❜❆✺✽✴☛❇☛✼❄❈✙P ❬❸◆◗✿✙❫⑥P✠✸✻❚✱❚❯◆◗✶✹◆◗✼✹✾✻✿❿◆◗✾✻❚✱✴☛▼✭✺❭❍❄✿✙◆◗✴☛✾⑥❚❯✴✷▼✵❍✹✾⑥❍✹❖☛❖✷❈✙◆◗✶✹❍❄✿✙◆◗✴☛✾❽✿✙✴⑥❇☛✼❄❖☛❈◗✴☛➑⑧✿◗❫✻✼❄✺❳❉▲❹✱✿➀❖☛▼✭✴☛❏☛◆✙❇✷✼✹P✽❍♦❬❸❍✹➑⑧❚✱✴☛▼ 52 Jorge et al. Û✷Ü✹Ý✷Û☛Þ◗Ü❘ß✙Ý⑦à✠á✻â❄ã✱Ü❘ä✽Ý☛å✷Ü✹Þ◗à➔æ☛Ü❄ß✙ç❸Ü✹Ü❄è❲å☛é◗ê❯ê✱Ü✹ã✭Ü✹è✻ß❺â✹Û☛Û✷Þ✙é◗ë✹â❄ß✙é◗Ý☛è☛à✕ì✕í➍é◗î✻Ü❘â✹è☛ï✽ð✵ñ❆í❆å☛Ý✷ë✹ò✻ä✽Ü✹è☛ß✙ó✯â✹Þ◗à✠Ý â❲ô❺ñ❆ñ⑥í➋å☛Ý✷ë✹ò✻ä✽Ü✹è✻ß➍ä✽ò✻à✂ß➍ê❯Ý☛Þ◗Þ◗Ý☛ç❰â➊õ✵â✹ß◗â➊ö❁ï➧Û☛Ü➊õ✵Ü❄ê❯é◗è✻é✙ß◗é◗Ý☛è➠÷✱õ✵ö⑤õ✵ø❿ß✙á✻â❄ß➍å☛Ü❄ê❯é◗è✻Ü✹à➞ß◗á✻Ü➊Ü✹è✻ß◗é✙ù ß◗é◗Ü✹à❿â✹è☛å✽â❄ß✙ß◗ã✭é✙æ✷ò✻ß✙Ü❄à➀ê❯Ý☛ã➍å✷Ý☛ë❄ò✻ä❭Ü❄è✻ß◗é✙è✻ú✽â➞à✠Û✷Ü✹ë✹é◗ê❯é◗ë➞å☛â❄ß✙â➞ä✽é✙è✻é◗è✻ú✽ä❭Ý✷å☛Ü❄Þ✝ì✹û▲Ý☛ã➍é◗è✻à✂ß✙â❄è✻ë✹Ü❄ó✷ß◗á✻Ü✹ã✭Ü➞é✙à ✁❱â✹ï➧Ü✹à ✁✂❸ü➏ä❭Ý✷å☛Ü❄Þ➍â✹è✻å❳à✠Ý❳Ý✷è❺ì✄✂➞è✻ï☎✂➞ü➡ä❭Ý✷å☛Ü❄Þ✇ç❸ã✭é✙ß◗ß✙Ü❄è⑦é◗è⑦ô❺ñ⑥ñ❆í⑨æ✷ï å✷é✙ê✱ê❯Ü❄ã✱Ü❄è✻ß✯Ü✹è✻ß◗é✙ß◗é◗Ü✹à➢ä❭ò✻à✂ß✯ê❯Ý☛Þ◗Þ◗Ý☛ç➃ß✙á☛Ü❪à✂â✹ä✽✆ Ü ➞ ✂ ü⑨à✂Û☛Ü✹ë❄é✙ê✱é✙ë❸õ✵ö❁õ⑦ì ✂①ä✽Ý☛å✷Ü✹Þ✯å☛Ü❄à✠ë❄ã✱é◗æ☛Ü❄å❭ò✻à✂é✙è☛ú❴ô➣ñ❆ñ❆í⑧á✻â❄à⑤ß◗á✻Ü❸ê❯Ý✷Þ✙Þ◗Ý☛ç❸é✙è☛ú❴à✠ß◗ã✭ò✻ë✹ß◗ò✻ã✭Ü✞✝ Ý✷è✻Ü❨õ✵ö❁õ❒ß◗Ý⑩à✠Û☛Ü❄ë✹é◗ê❯ï⑥â❨ü❿Ü✹ú✻ã✭Ü✹à✂à✠é◗Ý☛è❆ä✽Ý☛å☛Ü❄Þ✙ý❱â✹è✻Ý✷ß✙á✻Ü❄ã❘õ✵ö❁õ➐ß✙Ý❆ã✭Ü✹Û✷ã✱Ü❄à✠Ü✹è☛ß❿â❴þ⑦â❄é✙ÿ☛Ü ä✽Ý☛å✷Ü✹Þ◗ý➍Ý☛ß◗á✻Ü✹ã❿ß◗Ý❳å☛Ü❄ê❯é◗è✻Ü➊â✹è ✟ 1) A header, 2) A data schema, 3) A data mining schema, 4) A predictive model schema, 5) Definitions for predictive models, 6) Definitions for ensembles of models, 7) Rules for selecting and combining models and ensembles of models, 8) Rules for exception handling. ✡✠ ☞☛ Ý✷ä❭Û✷Ý☛è✻Ü❄è✻ß✯÷ ✷ø❛é✙à➢ã✱Ü ☛ò✻é◗ã✭Ü✹å✇ì✮ö❁á✻Ü❸Ý☛ß◗á✻Ü❄ã✇ë❄Ý☛ä✽Û☛Ý✷è✻Ü✹è✻ß◗à➢â✹ã✭Ü❪Ý✷Û☛ß◗é✙Ý✷è✻â✹Þ❊ì ö⑤á✻Ü⑦ä✽â✹é◗è⑧ã✭Ü✹â❄à✠Ý☛è☛à❴ß◗á✻â✹ß➢å☛ã✭Ý☛ÿ✻Ü♦ß✙á☛Ü⑦ê❯Ý✷ã✱ä✽ò✻Þ◗â✹ß◗é✙Ý✷è⑧Ý☛ê❱ß✙á✻Ü⑦ô➣ñ❆ñ❆í➦ê✱Ý☛ã❪Û☛ã✭Ü✹å☛é◗ë✹ß◗é◗ÿ✻Ü➥ä✽Ý☛å☛Ü❄Þ✙à ✍✌ ✏✎ ç❸Ü✹ã✭Ü➞ß✙á☛â✹ß➎é◗ß➎ä❭ò✻à✂ß✯æ☛Ü❸ò✻è✻é◗ÿ✻Ü✹ã✭à✠â❄Þ✙ó✷Ü ✻ß◗Ü✹è✻à✂é✙æ✷Þ✙Ü❄ó☛Û✷Ý☛ã✭ß✙â❄æ☛Þ◗Ü❪â❄è✻å❭á☛ò✻ä❭â❄è❴ã✭Ü✹â❄å☛â✹æ✷Þ✙Ü✯ì ✭ß✯â✹Þ◗Þ✙Ý✷ç❪à➢ò✻à✠Ü❄ã✱à ß◗Ý➊å☛Ü❄ÿ✻Ü✹Þ◗Ý☛Û➊ä✽Ý☛å☛Ü❄Þ✙à➀ç❸é✙ß◗á✻é◗è❭Ý✷è✻Ü➞ÿ✻Ü✹è☛å☛Ý☛ã à➀â✹Û☛Û✷Þ✙é◗ë✹â❄ß✙é◗Ý☛è☛ó☛â✹è☛å❭ò✻à✂Ü❪Ý✷ß✙á☛Ü✹ã❛ÿ✻Ü✹è☛å☛Ý☛ã✭à ✜â❄Û☛Û✷Þ✙é◗ë✹â❄ß✙é◗Ý☛è✻à ✕✔ ✒✑ ✖✔ ✓✑ ß◗Ý➥ÿ✻é◗à✠ò☛â✹Þ◗é ❄Ü✹ó➎â✹è☛â✹Þ◗ï ✹Ü✹ó➎Ü❄ÿ✻â✹Þ◗ò✻â❄ß✙Ü❘Ý✷ã❁Ý✷ß✙á✻Ü❄ã✱ç❸é◗à✠Ü❘ò✻à✂Ü❘ß◗á✻Ü❘ä✽Ý☛å☛Ü❄Þ✙à✉ì✕ô➣ã✭Ü✹ÿ✻é◗Ý☛ò☛à✠Þ◗ï➧ó➎ß✙á✻é◗à➔ç❸â✹à❿ÿ✻é◗ã✭ß✙ò✻ù ✍✌ â❄Þ✙Þ◗ï❲é◗ä❭Û✷Ý☛à✠à✂é✙æ✷Þ✙Ü❄ó▲æ☛ò☛ß❺ç❸é✙ß◗á❲ô➣ñ❆ñ❆í➍ó➎ß◗á✻Ü❘Ü ☛ë✹á✻â❄è✻ú✻Ü❘Ý✷ê➝ä❭Ý✷å☛Ü❄Þ✙à➔æ✷Ü✹ß◗ç❪Ü❄Ü✹è❲ë❄Ý☛ä✽Û☛Þ◗é✙â❄è✻ß➣â❄Û☛Û☛Þ◗é◗ë✹â✹ù ß◗é◗Ý☛è✻à❿è✻Ý✷ç➡ç❸é✙Þ◗Þ▲æ☛Ü✵à✂Ü✹â❄ä❭Þ◗Ü✹à✂à✕ì ➞ß▲ß✙á☛é✙à❿ä✽Ý☛ä✽Ü✹è✻ß◗ó✯Ý☛è✻Þ◗ï✽â✵ê❯Ü✹ç➒å☛â✹ß◗â✵ä✽é✙è✻é◗è✻ú➊ß◗Ý☛Ý☛Þ◗à➀â✹è✻å➊â❄Û☛Û☛Þ◗é◗ë✹â✹ù ☞✂ ☞✌ ß◗é◗Ý☛è✻à❭â✹Þ◗Þ✙Ý✷ç❪à❘ß◗Ý⑧Ü ✻Û☛Ý✷ã✱ß⑤ß✙á✻Ü❄é✙ã❪ä❭Ý✷å☛Ü❄Þ✙à❘ß◗Ý⑧ô❺ñ⑥ñ❆í➍ó➍æ✷ò✻ß❁é◗à❘ò✻ã✭ú✻Ü✹è☛ß❁ß◗Ý⑧é✙ä✽Û☛Þ◗Ü✹ä✽Ü✹è☛ß❁é◗ß❁é◗è❨Ý☛ß◗á✻Ü✹ã à✂Ý☛ê✱ß✙ç❸â✹ã✭Ü♦ß◗Ý☛Ý✷Þ✙à✽ß◗Ý❽à✂â✹ß◗é✙à✂ê❯ï❨å☛ã✭â✹ä✽â✹ß◗é◗ë✹â✹Þ◗Þ◗ï❨é✙è✻ë❄ã✱Ü❄â✹à✂é✙è✻ú❙ã✱Ü ☛ò✻é◗ã✱Ü❄ä❭Ü❄è✻ß◗à❴ê❯Ý✷ã❪à✂ß✙â❄ß✙é◗à✠ß◗é◗ë✹â✹Þ➢â✹è☛å❙å☛â❄ß✙â ☞☛ ä✽é◗è✻é✙è☛ú❴ä✽Ý☛å☛Ü❄Þ✙à➢é◗è❴æ☛ò☛à✠é◗è✻Ü✹à✂à⑤à✂ï➧à✠ß◗Ü✹ä✽à✕ì ✘✗✙✂ ✚✂ ✛✂ ô ➞ü❮ë❄â✹è❆ã✭Ü✹â✹å❆â❄è ➞ü➪ä✽Ý☛å✷Ü✹Þ❿à✂Û☛Ü✹ë❄é✙ê✱é✙Ü❄å❆é✙è⑥â❳ô➣ñ❆ñ❆í➒å☛Ý✷ë✹ò✻ä✽Ü✹è☛ß✝ì➣ö❁á✻Ü❳ò✻à✂Ü✹ã✵ç❸é✙Þ◗Þ❿æ☛Ü â❄æ☛Þ◗Ü➊ß✙Ý❳ä✽â✹è✻é◗Û☛ò✻Þ◗â✹ß◗Ü➊ß✙á☛Ü ➞➏ ü ä❭Ý✷å☛Ü✹Þ◗ó✇ë✹ã✭Ü✹â✹ß◗é◗è✻ú♦â➊è✻Ü❄ç❰ã✱ò☛Þ✙Ü➊à✠Û✷â✹ë❄Ü✽æ✷â✹à✂Ü✹å❳Ý☛è➠â➊à✠Ü✹ß❛Ý✷ê⑤Ý✷Û☛Ü✹ã✭â✹ù ✍✌ ß◗Ý☛ã✭à✠ó✷â✹è✻å✽Ü ☛Û☛Ý☛ã✭ß✯â❪à✂ò✻æ☛à✂Ü✹ß✯Ý☛ê✇à✠Ü❄Þ✙Ü❄ë✹ß◗Ü✹å✽ã✱ò☛Þ✙Ü❄à⑤ß◗Ý❭â❸è✻Ü❄ç➦ô❺ñ❆ñ⑥í❙å☛Ý☛ë❄ò✻ä✽Ü✹è✻ß❊ì ✕✔ ✜✔ ✢✎ ö⑤á✻Ü✹ã✭Ü❳é✙à➊à✂Ý☛ä✽Ü➠ç❸Ý☛ã✭î⑥Ý☛è⑥ß✙á✻Ü❳ÿ✻é◗à✠ò☛â✹Þ◗é ❄â✹ß◗é✙Ý✷è❽â❄è✻å❆à✠ò✻ä✽ä❭â❄ã✱é ✹â❄ß✙é◗Ý☛è⑥Ý☛ê➞â❄à✠à✠Ý✷ë✹é◗â✹ß◗é✙Ý✷è❽ã✭ò✻Þ◗Ü✹à✉ì ✭è ß◗á✻é◗à⑤à✂Ü✹ë❄ß✙é◗Ý☛è❭ç❪Ü❸ã✱Ü❄ê❯Ü❄ã✇ß◗Ý✽à✠Ü✹Þ◗Ü✹ë❄ß✙Ü❄å❭ç❸Ý☛ã✭î❴Ý✷è❴ß◗á✻Ü✹ä✽Ü●ì ö⑤á✻Ü✵à✠ï➧à✂ß✙Ü❄ä➟õ ▲ù ➦÷✭ñ❆â✵Ü✹ß▲â❄Þ✝ì ø➝ò✻à✠Ü❄à❿ß✙á☛Ü✵à✠â✹ä✽Ü✵à✠Ý✷ã✱ß➣Ý☛ê➍â✹Û✷Û☛ã✭Ý☛â❄ë✹á➊â✹à❿ß◗á✻Ü✵Ý☛è☛Ü✵ç❪Ü➞Û☛ã✭Ý☛ù ★✎ ✤✣ ✡✥✦✗✧ ✤✣ ✩✥✪✗✙ ✫✗✙✂ Û✷Ý☛à✂Ü❲á✻Ü✹ã✭Ü●ì ✭è➠ë❄Ý☛ä✽ä❭Ý✷è✻ó❛õ ▲ù ➬â❄è✻å❨ô ➞ü➓á✻â✹ÿ☛Ü❲ß✙á✻Ü❲â❄é✙ä➮Ý☛ê❿Û☛Ý✷à✠ß➝Û✷ã✱Ý✷ë✹Ü❄à✠à✠é◗è✻ú➠â➊Þ✙â❄ã✱ú✻Ü à✂Ü✹ß➎Ý✷ê ➞ü⑨ß◗á✻ã✭Ý☛ò✻ú✻á❭ç❪Ü❄æ❭æ✷ã✱Ý✷ç❪à✂é✙è☛ú❴â✹è☛å❭ÿ✻é◗à✠ò☛â✹Þ◗é ❄â✹ß◗é✙Ý✷è❺ì✜õ ▲ù ➆ã✭Ü✹Þ◗é✙Ü❄à⑤Ý✷è❴ß◗á✻Ü❸Û☛ã✭Ü✹à✠Ü❄è✻ß◗â✹ß◗é✙Ý✷è ✘✂ ✕✔ ✤✣ ✡✥✪✗✙ ✤✣ Ü☞✌✻Û☛Þ◗Ý☛ã✭Ü➠ß✙á✻Ü➠ÿ✻â❄ã✱é◗â✹ß◗é✙Ý✷è✻à✽Ý☛ê❸Ü✹â✹ë❄á❙Ý✷è✻Ü♦Ý✷ê❪ß◗á✻Ü❄à✠Ü➠õ✤✣➆ã✭ò✻Þ✙Ü❄à✕ì✬✭✎ è⑧Ý✷ò✻ã❪â❄Û☛Û✷ã✱Ý✷â✹ë✹á☛ó➝á✻Ý☛ç❸Ü✹ÿ☛Ü✹ã✭ó➝ç❪Ü ã✭Ü✹Þ◗ï❴Ý✷è❭â➞à✂Ü✹ß➎Ý☛ê❛ú✻Ü❄è✻Ü✹ã✭â✹Þ➎Ý✷Û☛Ü❄ã✱â❄ß✙Ý✷ã✱à➀ß◗á✻â✹ß➎ë❄â✹è✽æ☛Ü➞â✹Û✷Û☛Þ◗é✙Ü❄å✽ß◗Ý➊â✹è✻ï❭ã✱ò☛Þ✙Ü❄ó✷é◗è✻ë✹Þ◗ò✻å✷é✙è✻ú✽õ✤➠ ✣ ã✱ò✻Þ◗Ü✹à➢â✹à å✷Ü✹ê✱é✙è✻Ü❄å➥ê❯Ý✷ã➝õ✤✣▲ù✡✥✦✗✧❭ì❄ö❁á✻Ü✵à✂Ü✹ß▲Ý✷ê➍Ý☛Û☛Ü❄ã✱â❄ß✙Ý✷ã✱à❿ç❸Ü✵å☛Ü❄ê❯é◗è✻Ü✵é◗à❿æ☛â✹à✂Ü✹å❲Ý✷è✽à✂é✙ä✽Û☛Þ◗Ü✵ä✽â✹ß◗á✻Ü✹ä✽â✹ß◗é✙ë❄â✹Þ Û✷ã✱Ý✷Û☛Ü❄ã✱ß◗é✙Ü❄à➔Ý☛ê❁ß✙á✻Ü❴é✙ß◗Ü✹ä✽à✠Ü❄ß✙à➔â❄è✻å➥á✻â❄ÿ✻Ü❘â❘ë❄Þ✙Ü❄â✹ã⑤â✹è✻å➥é◗è✻ß◗ò✻é◗ß✙é◗ÿ✻Ü❘à✂Ü✹ä✽â✹è✻ß◗é✙ë❄à✕ì✕ô✘✗✙➞ ✂ ü➦â❄Þ✙à✂Ý➥á✻â❄à❿ß✙á✻Ü â❄å☛å✷é✙ß◗é✙Ý✷è✻â✹Þ✯Û✷Ý☛à✠à✂é✙æ✷é✙Þ◗é◗ß✙ï❘Ý✷ê❺ã✭Ü✹â❄å☛é◗è✻ú✭✂➞ü⑨ä✽Ý☛å✷Ü✹Þ◗à⑤â❄à⑤ô❺ñ❆ñ⑥í❱ì ✮ é✜✔✍✥➃é✜♦✔ é◗à❴ß◗á✻Ü♦è✻Ý☛è✻ù✱Ý☛ê✱ê❯é◗ë✹é◗â✹Þ➢è✻â✹ä✽Ü⑦ê✱Ý☛ã❸â⑦ô➣ñ❆ñ❆í①é◗è✻ß✙Ü❄ã✱â❄ë✹ß◗é✙ÿ☛Ü⑦ä✽Ý☛å☛Ü❄Þ⑤ÿ☛é✙à✂ò✻â✹Þ◗✕é ❄✔ Ü✹ã❸é✙ä✽Û☛Þ◗Ü✹ù ä✽Ü✹è☛ß✙Ü❄å⑧é✙è✰✯■â✹ÿ☛â➥÷✡➃ ✥ Ü❄ß✙ß◗à✠á✻Ü❄ã✱Ü❄ë✹î✻ø✜★ì ✱✎ ß❁ú✻ã✱â❄Û☛á✻é◗ë✹â❄Þ✙Þ◗ï➠å☛é◗à✠Û☛Þ◗â✹ï➧à✂ó❛è✻Ý☛ß➝Ý✷è✻Þ◗ï♦â❄à✠à✠Ý✷ë✹é◗â✹ß◗é✙Ý✷è❳ã✱ò✻Þ◗Ü✹à✂ó❛æ☛ò✻ß Ý✷ê❱â♦ã✱Ü❄å☛ò✻ë❄Ü✹å❙à✂Ü✹ß➢Ý☛ê❪ã✱ò✻Þ◗Ü✹à✂ó➝ë✹â❄Þ✙Þ◗Ü✹å❽å☛é◗ã✱Ü❄ë✹ß◗é✙Ý✷è⑧à✠Ü❄ß✙ß◗é◗è✻ú⑧Ý☛ã❸õ ⑨ã✭ò✻Þ◗Ü✹à✠ó❁â✹è✻å❽ß✙á✻Ü❄è⑧ß◗á✻Ü⑦ò✻à✂Ü✹ã❸ë✹â❄è A Post-Processing Environment 53 ✱✛✲✍✳✵✴✷✶✵✸✜✹✄✺☞✻✽✼✵✲✍✸✜✲✾✱❀✿✜✳✄✿✜✳✄❁✁✱❀✶★✼✵✺✍❂✜❃❅❄✵❆✧✹✵✺✛❇★✹✄✿✜❂✕✶★❃✓✶✵❇★✹✄✴✷✶★❈❊❉✪✿✜❋✍❉✪✿✜❋❀❈✡✶✵✻●✼✵✿✜❃✓❇★❂✕✲☞✴✖✿✕✳✄❁☎❍❏■❑✻✡✺☞❂✕✿✜✺✍❃ ✶★✳❀✸✜✹✄✺❏❇✵✻✒✺✍❃▲✺✍✳✄✸✜✲✍✸✜✿✜✶✵✳✛✶✵❈▼✸✕✹✵✺◆❂✜✿✜❃✓✸✢✶✵❈❖✻✒P✄❂✕✺☞❃✓◗★✲✍❂✜❂✕✶★❘✆✿✜✳✄❁✭✸✜✹✄✺◆P✄❃✓✺☞✻▼✸✕✶✛❃✓✺☞✸✞✸✜✹✄✺◆✱❀✿✜✳✄✿✜✱❀✲☞❂✬❃✓P✄❇★❇✵✶✵✻✒✸✬✲✍✳✄✼ ❙☞✶✵✳✄❈✡✿✕✼★✺✍✳✄❙☞✺✚✸✜✹✄✻✒✶✵P✄❁✵✹❯❚✄✺☞✻✡✴❱✿✜✳✄✸✜P✄✿✜✸✕✿✜❚✄✺❲❁✄✲✍P✄❁✵✺✍❃❅❄✙❳✤✿✕❋☞❉✪✿✕❋❲✲✍❂✜❃✓✶✦✲✍❙☞❙✍✶✵✱✛❇✵✲☞✳✄✿✜✺✍❃✁✸✕✹✄✺❲✼✵✿✜❃✓❇★❂✕✲☞✴❱✶✵❈ ✺☞✲✍❙☞✹✚✻✒P✄❂✜✺✁❨★✴❩❙☞✶✵❂✜✶✵✻✤❨✵✲☞✻✡❃✾✻✒✺✍❇★✻✡✺☞❃✓✺✍✳✵✸✕✿✜✳✄❁❲❃✓P✄❇★❇✵✶✵✻✒✸✽✲✍✳✵✼✚❙☞✶✵✳✄❈✡✿✕✼★✺✍✳✄❙☞✺✞❄✢❆❊✹✄✿✜❃❀❚✄✿✜❃✓P✵✲✍❂✜✿✕❋☞✺✍✻❏❙✍✲☞✳❩❨✵✺ P✵❃✓✺✍✼✛✼✵✿✜✻✒✺✍❙✍✸✜❂✜✴❬✿✕✳❀✲✆❘◆✺✍❨✛❨✵✻✒✶✵❘◆❃✓✺☞✻▼✲✍❃❭✲✙❪❫✲✍❚✵✲✆❇★❂✕P✄❁✵❴✩✿✜✳✘❄ ❵✒❛ ✺✍✳✵✸✘✺☞✸✘✲☞❂✘❜★❝✵❞❭✼✵✺✍❃▲❙✍✻✒✿✕❨★✺✭✲☞✳❡✲✍❇★❇✵✻✒✶✵✲✍❙☞✹❡✸✜✶✷✸✜✹✄✺✭❙☞❂✕P✄❃▲✸✕✺☞✻✡✿✜✳✄❁❡✶★❈✧✲☞❃✓❃✓✶★❙✍✿✜✲✍✸✜✿✕✶★✳❡✻✒P✄❂✕✺☞❃❅❄❢❆❊✹✄✺❬✲☞✿✕✱ ✿✜❃❣✸✕✶❤✼✵✺☞✻✡✿✜❚✄✺✐✿✕✳✄❈✡✶✵✻✒✱❀✲☞✸✕✿✜✶✵✳❥✲✍❨★✶✵P✄✸❦✸✜✹✄✺❧❁✄✻✒✶✵P✵❇✵✿✜✳✄❁♠✶✵❈♥✻✡P✵❂✕✺☞❃♦✶✵❨★✸✕✲☞✿✕✳✵✺✍✼♣❈✡✻✡✶★✱q❙☞❂✕P✵❃✓✸✜✺✍✻✒✿✕✳✄❁❖❄ ❍❏❃r✲s❙☞✶✵✳✄❃▲✺✍t✵P✵✺✍✳✄❙☞✺s✶✵✳✄✺✉❙✍✲☞✳❥✻✒✺✍❇★❂✕✲☞❙✍✺✉❙✍❂✜P✄❃▲✸✕✺☞✻✡✺☞✼❤✻✒P✄❂✜✺✍❃✈❨✵✴❥✶✵✳✄✺✉✱✛✶✵✻✒✺✉❁✄✺✍✳✄✺☞✻✡✲☞❂✇✻✡P✵❂✕✺✬❄ ① ✶✵✻②✲✷❁✄✿✕❚✵✺✍✳✰✲✍✸✜✸✜✻✡✿✜❨✵P✄✸✜✺❡✿✜✳✰✸✕✹✵✺❡❙✍✶★✳✄❃✓✺☞t✵P✄✺☞✳✄✸✕◗③✸✜✹✄✺❡❇★✻✡✶★❇✵✶★❃✓✺✍✼✰✲☞❂✕❁✵✶✵✻✒✿✕✸✜✹✄✱♠❙☞✶✵✳✄❃▲✸✕✻✒P✄❙✍✸✜❃✤✲⑤④✵⑥✐❁✄✻✒✿✕✼ ❘◆✹✄✺☞✻✡✺❡✺☞✲✍❙☞✹✰✲✍⑦✄✿✜❃❬❙✍✶★✻✡✻✒✺✍❃▲❇✵✶★✳✄✼✵❃❬✸✜✶⑧✲☞✳✰✲✍✸✜✸✕✻✒✿✕❨★P✄✸✜✺❡✿✕✳✰✸✜✹✄✺❡✲☞✳✄✸✜✺✍❙✍✺☞✼✵✺☞✳✄✸⑨❄★❆❊✹✄✺⑤✲✍❂✜❁✄✶✵✻✒✿✜✸✕✹✄✱♣✸✕✻✒✿✕✺☞❃✤✸✕✶ ❈✡✿✕✳✵✼❶⑩❷✸✜✹✄✺♠❨✵✺☞❃✓✸✜❸♠❙✍❂✜P✄❃▲✸✕✺☞✻✡✿✜✳✄❁❹✶✵❈♦✻✡P✄❂✜✺✍❃❧❈✩✶✵✻♦✳✄✶✵✳✄❴✡✶✵❚✄✺☞✻✡❂✜✲✍❇★❇✵✿✜✳✄❁❺✲✍✻✒✺✍✲☞❃r✶✵❈❻✸✜✹✄✺s④✵⑥❼❁✄✻✒✿✕✼▼❄ ❆❊✹✄✺◆✲✍❇✵❇★✻✡✶★✲✍❙☞✹✭✶✵✳✵❂✕✴❬❙☞✶✵✳✄❃▲✿✕✼★✺✍✻✒❃❊✻✒P✄❂✜✺✍❃❭❘✆✿✜✸✜✹✭✳✄P✄✱✛✺✍✻✒✿✜❙✆✲☞✸✕✸✜✻✡✿✜❨✵P✵✸✕✺☞❃❊✿✜✳✭✸✜✹✄✺◆✲✍✳✄✸✜✺✍❙☞✺✍✼✵✺☞✳✄✸✜❃❅❄ ❍❏❃▲❃✓✶✵❙☞✿✕✲☞✸✕✿✜✶✵✳❩✻✡P✄❂✜✺✷✺☞✳✄❁✄✿✜✳✄✺✍❃❀✲✍✻✒✺✷✶★❈✩✸✜✺✍✳⑧✻✒✿✕❁✵✹✄✸✕❂✜✴✰✲✍❙☞❙✍P✄❃▲✺✍✼✚✶✵❈②✶★❚✄✺✍✻✒❂✜✶✵✲✍✼★✿✕✳✵❁⑧✸✕✹✵✺✷P✄❃▲✺✍✻◆❘✆✿✜✸✕✹❩❚✄✺✍✻✒✴ ❂✜✲✍✻✒❁✄✺❽❃▲✺✍✸✜❃❑✶✵❈❯✻✡P✄❂✜✺✍❃❢❄✛❆✧✹✄✿✜❃❑✲☞❇✵❇✵❂✜✿✜✺✍❃❑✸✜✶♠✲✍✳✵✴✐❃▲✶✵❈✡✸✕❘◆✲✍✻✒✺❽❇✵✲☞❙✍❾✄✲☞❁✄✺✍◗✰❙☞✶✵✱✛✱❀✺☞✻✡❙☞✿✕✲☞❂⑧✶✵✻❯✳✄✶✵✳✄❴ ❙☞✶✵✱✛✱❀✺☞✻✡❙☞✿✕✲☞❂✕◗★✸✕✹✵✲✍✸✬❘✆✺◆❾✄✳✄✶★❘✾❄ ❿ ✳✾✸✕✹✵✿✕❃✽❇★✲✍❇✵✺☞✻✙❘◆✺✤✼✵✺✍❃▲❙✍✻✒✿✕❨★✺✤✲✤✻✡P✵❂✕✺✤❇★✶✵❃✓✸✫❇✵✻✒✶✵❙✍✺☞❃✓❃▲✿✕✳✄❁✾✺☞✳✄❚✄✿✜✻✡✶★✳✄✱❀✺☞✳✄✸➀✸✜✹✄✲✍✸✫✲✍❂✜❂✕✶★❘✆❃✽✸✜✹✄✺❏P✄❃✓✺☞✻③✸✕✶ ❨★✻✡✶★❘✆❃▲✺❬✸✜✹✄✺❬✻✒P✄❂✕✺✭❃✓❇✵✲☞❙✍✺☞◗✢✶✵✻✒❁✄✲✍✳✄✿✜❋✍✺☞✼❡❨✵✴✾❁✄✺☞✳✄✺✍✻✒✲✍❂✜✿✜✸✕✴✖◗✢❨★✴✛❚✵✿✕✺☞❘✆✿✜✳✄❁⑤✶✵✳✵✺✤✻✡✺☞❂✕✺☞❚✄✲✍✳✄✸✫❃✓✺✍✸✫✶✵❈③✻✒P✄❂✕✺☞❃✽✲✍✸ ✲✛✸✜✿✕✱✛✺✞❄✄❍r❃✓✺✍✸❖✶✵❈❊❃✓✿✜✱❀❇★❂✕✺❀✶✵❇★✺✍✻✒✲✍✸✜✶✵✻✒❃②✲✍❂✜❂✜✶✵❘◆❃②✸✜✹✄✺✭P✄❃▲✺✍✻❭✸✕✶☎✱❀✶★❚✄✺✭❈✡✻✡✶★✱➁✶★✳✄✺✭❃▲✺✍✸❖✶✵❈❊✻✡P✄❂✜✺✍❃✆✸✕✶☎✲✍✳✄❴ ✶★✸✕✹✵✺✍✻✏❄✄➂✙✲☞❙✍✹☎❃✓✺✍✸▼✶★❈❊✻✒P✄❂✜✺✍❃◆✿✕❃◆❇✵✻✒✺✍❃▲✺✍✳✄✸✜✺✍✼✁✿✕✳☎✲❀❇★✲✍❁✄✺✛✲✍✳✵✼☎❙☞✲✍✳✷❨★✺❀❁✵✻✡✲☞❇✵✹✄✿✜❙✍✲☞❂✕❂✜✴❡❃✓P✵✱❀✱✛✲✍✻✒✿✕❋☞✺✍✼▼❄ ❿ ✳ ✸✜✹✄✺❡❈✡✶✵❂✜❂✕✶★❘✆✿✜✳✄❁✰❘◆✺❡❃✓P✵✱❀✱✛✲✍✻✒✿✕❋☞✺❡✸✜✹✄✺❡✱✛✲✍✿✜✳✰✲✍✼✵❚✵✲✍✳✄✸✜✲✍❁✄✺☞❃✓◗▼❂✜✿✕✱✛✿✕✸✜✲✍✸✜✿✜✶✵✳✄❃✤✲☞✳✄✼✰❈✩P✄✸✜P✄✻✒✺⑤❘✆✶★✻✡❾➃✶★❈✽✸✕✹✄✺ ❇★✻✡✶★❇✵✶★❃✓✺✍✼✛✲✍❇★❇✵✻✒✶✵✲☞❙✍✹✘❄ ❆❊✹✄✺◆✱❀✲☞✿✕✳❀✲✍✼✵❚✵✲✍✳✄✸✜✲✍❁✄✺☞❃❊✲☞✻✡✺✬➄ • ➅ ➂✙❍❏■❱✺✍✳✵✲✍❨✵❂✜✺✍❃❭❃✓✺☞❂✕✺☞❙✍✸✜✿✕✶★✳✭✲☞✳✄✼❀❨★✻✡✶★❘✆❃▲✿✕✳✵❁✭✲✍❙☞✻✡✶★❃✓❃❭✸✕✹✵✺✆❃▲✺✍✸✬✶✵❈❖✼✵✺☞✻✡✿✜❚✄✺✍✼✛❍❏■✭❄ ❿ • ✸✬✺✍✳✄✲☞❨✵❂✜✺✍❃❭❇✵❂✜✶✵✸✜✸✕✿✜✳✄❁❀✳✄P✄✱✛✺✍✻✒✿✕❙◆❇✵✻✒✶✵❇★✺✍✻✒✸✕✿✜✺✍❃❭✶✵❈❖✺✍✲☞❙✍✹✭❃▲P✄❨✵❃▲✺✍✸✬✶✵❈❖✻✒P✄❂✕✺☞❃❊❈✡✶✵P✄✳✵✼❖❄ • ➆ ✻✡✶★❘✆❃▲✿✕✳✄❁⑤✿✜❃➇✼★✶✵✳✄✺❬❨★✴❀✲✤❃▲✺✍✸➀✶★❈③❘✆✺☞❂✕❂✜❴✩✼★✺✍❈✡✿✕✳✄✺☞✼⑤✶✵❇★✺✍✻✒✲✍✸✜✶✵✻✒❃✽❘✆✿✜✸✕✹✾✲✤❙☞❂✕✺☞✲✍✻✙✲☞✳✄✼⑤✿✕✳✵✸✕P✄✿✜✸✜✿✕❚✄✺ ❃▲✺✍✱✛✲✍✳✄✸✜✿✜❙✍❃❅❄ • ➈ ✺✍❂✜✺✍❙✍✸✜✿✜✶✵✳✛✶✵❈▼❍❏■❯✻✒P✄❂✕✺☞❃●❨✵✴✭✲☞✳❀P✵❃✓✺✍✻③✿✜❃●✲✍✳❀✿✕✱✛❇✵❂✜✿✕❙☞✿✕✸✬❈✡✶✵✻✒✱❣✶★❈✘❇★✻✡✶★❚✄✿✜✼✵✿✜✳✄❁➉❨★✲✍❙☞❾✄❁✄✻✒✶✵P✄✳✄✼ ❾✵✳✄✶✵❘◆❂✕✺☞✼✵❁✄✺☞◗✬✸✕✹✄✲☞✸➀❙✍✲☞✳✾❨✵✺✤❂✜✲✍✸✜✺✍✻✙P✄❃▲✺✍✼★◗✞❈✡✶✵✻✙✺☞⑦✄✲✍✱✛❇✵❂✜✺✍◗✬✿✜✳✾❃✓✺✍❂✜✺✍❙☞✸✕✿✜✳✄❁✾✻✒P✄❂✕✺☞❃✽❈✩✶★✻✙✲✤✲❏❙✍❂✜✲✍❃✓❴ ❃▲✿✕❈✡✿✕✺☞✻▼✱❀✲☞✼✵✺◆✶✵P✄✸✬✶✵❈❖✲◆❃✓P✄❨★❃✓✺✍✸✬✶★❈✘✻✒P✄❂✜✺✍❃❅❄ ❛ • ➅ ➂✙❍❏■r❇✵✻✒✺✍❃▲✺✍✳✄✸✜❃✭✲☞✳⑧✶✵❇★✺✍✳⑧❇★✹✄✿✕❂✜✶✵❃▲✶✵❇★✹✄✴✰❨✵✴✰✻✒✺✍✲☞✼✵✿✜✳✄❁⑧✸✜✹✄✺✷❃▲✺✍✸❭✶✵❈②✻✒P✄❂✜✺✍❃❀✲✍❃✭✲ ➅✫➊❦➊ ✱✛✶✵✼★✺✍❂➋❄ ❆❊✹✄✺◆✱❀✲☞✿✕✳❀❂✕✿✜✱❀✿✜✸✜✲✍✸✜✿✕✶★✳✄❃❊✲☞✻✡✺✬➄ • ❳✤✿✜❃✓P✵✲✍❂✜✿✕❋☞✲✍✸✜✿✕✶★✳⑧✸✕✺☞❙✍✹✄✳✵✿✕t★P✄✺✍❃❀✲✍✻✒✺✷✲✍❂✜❘◆✲✍✴✖❃✭✼★✿✕❈✡❈✩✿✜❙✍P✄❂✜✸✧✸✜✶⑧✺☞❚✄✲✍❂✜P✄✲✍✸✜✺✞❄✞❆❊✹✄✿✜❃❬✶✵✳✄✺✷✿✕❃❬✳✵✶⑧✺✍⑦✄❴ ❙☞✺✍❇★✸✕✿✜✶✵✳✘❄ • ❆❊✹✄✺✤❙✍P✵✻✡✻✒✺✍✳✄✸✫✿✕✱✛❇✵❂✜✺✍✱✛✺✍✳✄✸✜✲✍✸✜✿✜✶✵✳✾✻✡✺☞t✵P✄✿✜✻✒✺✍❃✓◗✬✶★✳✛✸✜✹✄✺✤❃▲✺✍✻✒❚✄✺✍✻✒❴✩❃▲✿✕✼★✺✍◗✬✸✕✹✵✺✤P✄❃✓✺✤✶★❈③✲✍✳✛✶✵❇✵✺☞✻✡✲☞✸✕❴ ✿✜✳✄❁❀❃✓✴✖❃✓✸✜✺✍✱✈❈✡✻✡✶★✱❣✶★✳✄✺◆❃✓❇✵✺☞❙✍✿✜❈✩✿✜❙✆❚✵✺✍✳✄✼★✶✵✻✏❄ ❵ • ❆❊✹✄✺◆✺✍✳✄✸✜✻✡✴✭❇✵✶✵✿✜✳✄✸ ✸✕✹✄✺◆✿✜✳✄✼✵✺☞⑦✭❇✵✲☞❁✄✺✍❞▼✿✜❃❊❃▲✸✕✿✜❂✜❂✞✻✒✺✍❂✜✲✍✸✜✿✜❚✄✺✍❂✜✴❬❘◆✺✍✲✍❾❖❄ ❆❊✹✄✺◆❚✄✿✕❃▲P✄✲✍❂✜✿✜❋✍✲✍✸✜✿✜✶✵✳✭✸✜✺✍❙☞✹✄✳✄✿✜t✵P✄✺☞❃❊✶★❈✩❈✡✺✍✻✒✺✍✼✛✲✍✻✒✺✆❚✄✺☞✻✡✴✭❂✕✿✜✱❀✿✜✸✜✺✍✼❖❄ ① P✄•✸✕P✵✻✡✺◆❘✆✶★✻✡❾❖➄ • ⑥✤✺☞❚✄✺✍❂✜✶✵❇✛✱✛✺✍✸✜✻✡✿✜❙✍❃❭✸✕✶✛✱✛✺✍✲✍❃▲P✄✻✒✺✆✸✜✹✄✺◆❁✄✲✍✿✜✳✄❃❭✶✵❈❖✸✕✹✄✿✜❃❭✲✍❇✵❇★✻✡✶★✲✍❙☞✹✘❄ 54 Jorge et al. • • • • ➌✤➍☞➎✄➍✍➏✜➐✵➑✚➒❀➍☞➓✍➔✄→☞➣✄↔✜↕✓➒✛↕✭➙✜➔✄→✍➙❭→✍➏✜➏✕➐★➛➁➙✜➔✄➍❡↔✜➣✄➓☞➐✵➜✒➑✵➐✵➜✒→✍➙✜↔✜➐✵➣✰➐✵➝②➞✄↕✓➍☞➜②➟✵➍☞➝✩↔✜➣✄➍✍➟❩➎✄↔✕↕▲➞✄→✍➏✜↔✜➠✍→✍➡ ➙✜↔✜➐✵➣✄↕⑧→☞➣✄➟❑➜✒➞✄➏✜➍❱↕✓➍☞➏✕➍☞➓✍➙✜↔✕➐★➣❻➓✍➜✒↔✕➙✜➍✍➜✒↔✕→☞➢✤↕✓➞✄➓☞➔✪→✍↕✰➝✩➐★➜❡➍✍➤✵→✍➒✛➑✵➏✜➍✍➢❏➙✕➔✵➍✇➓✍➐✵➒✛➥✵↔✜➣✄→☞➙✕↔✜➐✵➣✪➐✵➝ ➑★➜✡↔✜➒✛↔✕➙✜↔✕➎✵➍✆➐★➑✵➍✍➜✒→✍➙✜➐✵➜✒↕❅➦ ➧✧➎✄→✍➏✜➞✄→✍➙✜➍✁➙✕➔✄➍✁➓✍➞✵➜✡➜✒➍✍➣✄➙●↔✜➒❀➑★➏✕➍☞➒❀➍☞➣✄➙✜→✍➙✜↔✕➐★➣❩→✍➨✵→✍↔✜➣✄↕✓➙●➐✵➙✜➔✄➍☞➜◆→☞➏✕➙✜➍✍➜✒➣✄→☞➙✕↔✜➎✄➍✍↕❀↕✓➞✄➓☞➔⑧→✍↕➇➩❫→✍➎✄→☞➢ →☞↕❊➛◆➍✍➏✜➏✬→✍↕❭→✍➣❀→✍➏✜➙✕➍☞➜✡➣✄→☞➙✕↔✜➎✄➍◆➙✕➐✛➓✍➏✜↔✜➍✍➣✄➙✜➡✩↕▲➍✍➜✒➎✄➍✍➜✒➢✵↕▲➞✄➓✍➔❀→✍↕❭➑✵➏✜➞✄➨✄➡✩↔✜➣✘➦ ➫✒➣✄➎✄➍☞↕✓➙✜↔✕➨✵→✍➙✜➍✆→☞➣✄➟❀↔✜➒✛➑✵➏✜➍✍➒✛➍✍➣✄➙✬➐✵➙✜➔✄➍☞➜❖➎✵↔✕↕▲➞✄→✍➏✬➜✒➍✍➑✵➜✒➍✍↕▲➍✍➣✄➙✜→✍➙✜↔✜➐✵➣✄↕❭➐✵➝❖↕✓➞✄➥★↕✓➍☞➙✕↕❭➐✵➝❖➜✡➞✵➏✕➍☞↕❅➦ ➭❏➏✜➏✕➐★➛➉➙✜➔✄➍❬➟★➍✍➝✡↔✕➣✄↔✜➙✜↔✕➐★➣⑤➐✵➝✙➜✒➞✄➏✜➍❬↕✓➍☞➏✕➍☞➓✍➙✜↔✕➐★➣⑤➓✍➜✒↔✕➙✜➍✍➜✒↔✕→✭➥✵→✍↕▲➍✍➟❡➐★➣⑤➙✕➔✵➍❬↕✓➞✄➑★➑✵➐★➜✡➙✘→✍➣✄➟❡➓☞➐✵➣✄➝✡↔✕➡ ➟★➍✍➣✄➓☞➍✆➐★➝✘➙✜➔✄➍◆➜✡➞✵➏✕➍☞➢✵↔✜➙✕↕❭→✍➣✄➙✜➍✍➓☞➍✍➟★➍✍➣✄➙✬→✍➣✵➟❀↔✜➙✕↕❭➓✍➐★➣✄↕✓➍☞➯✵➞✄➍☞➣✄➙➋➦ ❅➲ ➳ ❭➵ ➸✍➺➼➻➾➽❊➻➾➚✡➪✞➶ ➳ ➪✞➹✷➻➾➘☞➘❅➴➷➚➷➻✖➪★➬ ➳ ➪▼➮❷➺➼➴➷➱✍➻➾➘☞✃✒➪✵➶ ➳ ➪★❐✢❒❅➴➷❮✍❒❅➘☞❰➾➘✢➪✵➬ ➳ ➪★Ï✽❰▲➺➼➱☞➻▲Ð✆❒✬➪★➵ ➳★Ñ✓➳ ➪★Ò★➻❢Ó✕✃✙Ô●➴✡Ó✕Õ➾❒☞❮☞❰▲➺➼Ö❡❒❅× ➳➋Ú✫Û☞Ü➾Ý☞Þ❅ß➾à▲á✽â Þ②ã③Þ❅ä❅å②æ à▲Û❅ç❅à✧è❬â á⑨ß▲ä❅Ü➾à▲é✕ê◆Ý❅Þ❅Û②è●Ý✬ë Ý②ì✁â Þ✬â Þ❅ç➀í➾î☞ï❅ð☞ñ➼î❅ò✢ó☞➳❅➲❫ô❅ô❅õ✬➳ ò✢➳ ö③✤➵ ➼➺ Ó✒➴➷Ó✕➘✢❒❅➪✢Õ▲➴Ø➮ ➻▲➳ ✃➷➪❅➴Ø❒☞➹✷➘✆❒❅➶③✃➷➽❭Ù☞➚➷➻▲❰✖➘❅Ó ➴✡➪❅➶ ➳ ➪❅÷●➚➷➚➷Ð✆➻▲➘✢➪☞ø ➳ Ô ➳ ➻▲➘❅ù✛❐✧ÓØÙ❅➺✓➪✬➮ ➳ Ô●Ö❷➘❅➻▲Ð✆➴➷Õ✆➴Ø✃➷❰▲Ð✾Ó✕❰▲✃➀Õ➾❒❅Ù☞➘❅✃➷➴➷➘❅➸✭➻➾➘☞ù✛➴➷Ð✆ú☞➚➷➴ØÕ▲➻ ñ ➳þý❅✁ÿ ✫ì✄✘✂ è✆➀☎ à➾ß▲ä❅é⑨Û✞✝ Ú✠✞✟ ìrý✁✡✬à➾ß❢â Ý✬æ☞ÿ➼Þ✢ë à▲é⑨à➾á❷☛ë ✘é⑨✌ä ☞✍✡☎ä❅Þ❬ì❀Ý☞✌Þ ✎ Ý❅✃➷➴➷ç☞❒❅✑à ➘❬✏●➺➼à▲Ù❅Þ✢➚➷❰✖ë➾Ó✢ä✓❭✒ ×û❒☞è●➺❖Ý✢Ð②ë Ý✕➻➾✔ ➺➼➪➱✍ò❅❰➾õ✗✃✬✖➼ü☞ò✌➻✖✔✓Ó✕í ➱✍ò✙❰➾✘✗✃✬✘ ù☞➪ ➻➾➲❫✃➷ô❅➻ ô☞ð✢➳✁✚ ✃➷✃Øú í ✛✜✛ Õ▲➴Ø✃➷❰❢Ó✕❰➾❰▲➺ ➳ ➘✁✢ ➳ ➘☞❰➾Õ ➳ Õ➾❒❅Ð ✛ ü❅➺➼➴➷➘ ô☞ð ù☞Ö✍➘☞➻➾Ð②➴➷Õ ➳ ✚ ✃ØÐ②➚ î✢➳ Ô●➻▲✃Ø➻✧➹✷➴Ø➘☞➴➷➘❅➸✤✣❭➺➼❒☞Ù❅ú ✖✦✥ ➹✷✞➹ ✧✾ù☞❰➾❮✍❰➾➚➷❒❅ú☞Ð✆❰▲➘❅✃ ✔ ➪ ✚ ✃➷✃Øú í ✛✜✛ ➽❊➽❭➽ ➳ ù❅Ð②➸ ➳ ❒☞➺➼➸ ✛ ★✢➳ ñ➼ò❅õ☞ò Ó✕✃➷➻▲➘❅ù❅➻▲➺➼ù ✚ ✃➷✃Øú í ✛✜✛ ➽❊➽❭➽ ➳ ❰➾Õ➾Ð②➻ ➳ Õ ✚✌✛ ❰▲Õ➾Ð②➻ ➲✁✛ ➮✏❐✢➵✭✬●Ô ✛ ✩✫✪③➹✷➵ ñ➼ò☞õ❅ò✢➳ ➬●❐✢➹ ✩✫✪③➹☎➵ ✘❅➳ ➳ ➳➪ ✰ ➳ í ✪③➚➷Ù✢Ó✕✃➷❰▲➺➼➴Ø➘☞➸➃➵✤Ó✒Ó✕❒❅Õ▲➴Ø➻▲✃➷➴Ø❒☞➘✰➶③Ù❅➚➷❰✖Ó✒➪❯➴➷➘✰➵❭➚➷❰✑✰✱✣❭➺➼➻➾Ö★➪ ✥ ❰▲➺ ✦ñ ✲ ➱✍❰ ✧✢❰➾➘☞✃✡➪➀ö ➪✙➮✏➽❭➻▲Ð✆➴✡➪✬➵ ✯ ✮ ➴➷ù❅❒☞Ð ➪✬ø ✖ ➳ ✔✓✴ í ✳➀é⑨ä☞✶ ß ❖ ✵ ✓ä ♥ ✒ ë ✷❅à✹✸✺✬ ✷ â é✏ë à▲à➾Þ✬ë ✷❲ÿ➼Þ✢ë à➾é⑨Þ☞Ý✢ë➷â ä❅Þ☞Ý✢✻ æ ✞✟ ä❅✦Þ ✒þà▲é⑨à➾Þ☞ß➾à➃ä☞Þ❦è●Ý✢ë ✽ Ý ✼✫Þ☞ç✢â Þ❅à▲à➾é❷â Þ☞✌ç ✾ ÿ✧✢✟✞➻➾✿è ➺✓ÓØ✼❁❒❅➘ ❀✯❂ ✩➀ö③ù✢➴➷➺➼Ó Ð②➴Ø➘☞➸ ✚ ➻➾Ð❱÷ ➳ ❃✤➳✓Ñ ✩✫✩✫✩❁✪③❒❅Ð②ú❅Ù❅✃➷❰▲➺✙➮❷❒☞Õ➾➴➷❰➾✃➷Ö ➲❫ô❅ô❅ð ✁ õ✢➳ ➹✷➻✖➪❅❄●➴➷Ð✆➴➷➘❅➸✞➪❅✧✢➴ØÙ✬➪❷ö③➴➷➘☞➸✬❅➪ ✮✰❒❅➘☞➸✬➪ ❃ ➴➷➻➾➘ ✖➼ò❅ï☞ï❅✗ï ✔ ➪✺✮✰❰▲ü✭×✜❒❅➺❖Ô❭➻➾✃➷➻✽➹☎➴➷➘☞➴Ø➘☞➴➷➸ ✺í ❆ ➺➼➸☞➻▲➘❅➴❈❇➾➴➷➘☞➸❏➻➾➘☞ù Ñ ➘❅✃➷❰➾➺ ñ ✚ ý☞✁ÿ ✘ã③è●❊ è ✼●❋❍✧✡ æ ä❅é✒Ý✢ë➷â ä❅Þ❅á ➪➇✠➵ ✪③➹ ú❅➺➼Ñ ❰▲✃➷➴Ø❃ ➘☞➸❑✃ ✚ ❰♥Ô❭➴✒ÓØÕ➾❒❅❮✍ò ❰➾➺➼Ñ ❰▲ù➉➶③Ù☞➲ ➚➷❰✖Ó⑤÷❬ò☞Ó✕➴➷ï❅➘❅ï❅➸❻ï✬➳ ✃ ✚ ❉ ❰ ✰ ✮ ❰▲ü✢➪✭➮❷Õ ❒❅❒☞➚✡➪ Ï✽❒❅➚➷Ù☞Ð✆❰ ✖ ➪ Ó⑨Ó✕✥☛Ù❅✔✯❰ ✚ ➪❫ø⑨Ù❅í ✛✜➚➷✛ Ö ➳ ð✢➳ ➹✷➮ ✣➴ØÕ▲➺➼❒✢Ô●Ó✕❒☞Ô✤×û✶✃ ➪✓✰ ✮ ❰➾ü✛➮✏➴Ø✃➷❰ ➵✤➮ ✃➷✃Øú Ð✛Ó✕➳ ù❅➘ Ð②➴ØÕ▲➺➼❒✢Ó✕❒☞×û✃ ➳ Õ➾❒☞Ð ✛ ➚➷➴➷ü❅➺➼➻▲➺➼Ö ✛ ù☞❰➾×û➻▲Ù❅➚➷✃ ➳ ➻✖Ó✕✯ú ■✩Ù☞➺➼✜➚ ❏ ✛ ➚➷➴Øü☞➺➼➻➾➺➼Ö ✛ ❰➾➘ ñ ✛ ✛✜✚ ✛ ó❅➳ Ù✢➹✷Ó ➴Øù☞Õ▲➺➼➘❅❒✢ü☞Ó✕❰➾❒☞➸✍×û❮☞✃ ü ✃➷✮✰Ð✆❰➾➚ ü ➻➾Õ▲✃Ø➴➷❮✍❰✖➮❷➴➷Ó✕✃➷❰▲❰ ➺➼❮☞❰▲➺➼ú❅✖ ➻▲Ô●➸☞❰❢❰❢Ó✕Ó Õ➾➺➼➻✖➴➷ÓØú☞ú ✃Ø➴➷❒☞➘✢Ó ❒☞× ø⑨➻➾❮✍➻✖➪ ø❷➮❷Õ▲➺➼➴➷ú❅✃✡➪ ➻▲➘❅ù ø⑨➻➾❮✍➻✖➮✏Õ➾➺➼➴➷ú❅✃ ✔ ✚ í ✛❈✛ ➳ ➳ ✛ ➳ ñ ✌➲ ✘✁★✯❅✘ ✗ó ✘ ô✢➳ ➮✏✃➷➻➾✃➷❮✍ú ➻✖Ó✕❰▲Ó✕➺➼Ù❅❰✖ú☞➪➾ú❅➵ ❒☞➺➼➳ ➪✃ ❆ Ð✆Ð②➴➷Õ➾➴Ø➺➼❰▲❒✬Õ➾Ó✕➴➷❒❅➘✢×✜ÓØ✃ ➱☞Õ➾➴✡❒❅◆➪ ✩ Ð ➳ ù❅➻▲➘❅❰▲×û❖ù ➻➾Ù☞✬❭➚Ø➻➾✃ ❮✍➻❢➻➾Ó✕ú✌✃ ✚ ✰☛❰❢■➾➪❢ÓØÕ➾➮ ➴➷➳ ✌ù ➪▲✵❏➵❭➱☞➘✆ü✯❰➾❑ ✩●×û×û✬ ➴➷Õ▲➴Ø÷❬❰▲➘❅➮▲✃✄❑ ▼➻➾➚➷➸✍❒❅➺➼➴➷✃ ✚ Ð❯×û❒❅➺✢Ð②➴➷➘❅➴➷➘❅➸②➻✖Ó✒Ó✕❒❅Õ▲➴Ø➻▲✃➷➴Ø❒☞➘ ✍➳ ✳➀é⑨ä❅✕ ß ✵✓✓ä ✒✭P✙◗✓á✏ë➾ÿ➼Þ✢ë➷❈æ ✵✌✞✟ ä❅Þ✓✒❘✵✓ä❅❚ Þ ❙▲à➾é✕❖ê ❯✵Ý❅é⑨ç☞à❊è●Ý✬ë ✌Ý ❅❱ Ý☞á⑨à➾✿á ✝✍❙◆❯★✭è ❲✶❳ ➲✓ô☞ô✙✘❅➳ ➲✓ï✢➳ ➺➼➮✏Ù❅➴Ø➚➷➚➷ü☞❰❢❰➾Ó✵➺✓➴➷➘✆Ó✕Õ ✚➚➷➻▲➻➾➺➼❈✃ ➸☞❇❢❰✧➪✍➵ù❅➻▲➳ ✃Ø➻▲➻➾ü❅➘☞➻❢ù❀Ó✕❰✖❐✢Ó ✗Ù ❇ ✚ ➴➷➚➷➴➷➘✢➪❷➵ ➳ ➪ ❆ ➘➃ÓØÙ❅✁ü ✢✡❰▲Õ➾✃➷➴➷❮☞❰②Ð✆❰▲➻✖Ó✕Ù☞➺➼❰✖Ó➀❒❅×❖➴➷➘☞✃Ø❰▲➺➼❰✖Ó✕✃➷➴➷➘❅➸✍➘❅❰❢Ó⑨Ó✫➴➷➘✭➱☞➘☞❒❅➪ ➽❊➚Ø❰▲ù❅➸✍❰ ➳ ✳➀é⑨ä❅ß▲à➾à▲Û✢â Þ❅ç☞á③✦ä ✒✭ë ✷☞✿à ❨❊â é⑨á❷ë✏ÿ➼Þ✢ë à➾é⑨Þ☞Ý✢ë➷â ä❅Þ☞Ý✢❘æ ✟✞ä❅✓Þ ✒✏à▲é⑨à➾Þ☞ß➾à❊ä❅Þ◆ã③Þ☞ä❅å②æ à➾Û☞ç❅à❊è❬â á⑨ß➾ä☞Ü➾à➾éØê✤Ý❅Þ☞Û ◆ è●ù❅➴✡Ý✬Ó✕Õ▲ë Ý②❒❅❮✍ì✁❰➾➺➼â Þ✬Ö â Þ❅ç ➪ ➲✓ô☞✙ô ✘ ➪ ò❅✯ð ✘✓ñ➼ò✢ó☞➲❅✍➳ ✚ ✃➷✃➷ú í ✛❈✛ Õ▲➴➷✃Ø❰❢Ó✕❰➾❰▲➺ ➳ ✁➘ ✢ ➳ ➘☞❰➾Õ ➳ Õ➾❒☞Ð ✛ Ó✕➴➷➚➷ü❅❰▲➺✓Ó✕Õ ✚ ➻➾❈✃ ❇ ✯ô ✘ Ó✕Ù☞ü✁✡✢ ❰➾Õ▲✃Ø➴➷❮✍❰ ➳ ✚ ✃➷Ð✆➚ ➲❅➲❅➳ ❐✢❒❅➴➷❮✍❒❅➘☞❰➾➘✢➪✍➬ ➳ ➪✬➮✏➻➾Ð②ú❅➚➷➴➷➘❅➸❬➚➷➻➾➺➼➸✍❰✆ù☞➻➾✃➷➻➾ü☞➻✖Ó✕❰❢Ó✫×û❒☞➺▼➻✖Ó✒Ó✕❒❅Õ▲➴Ø➻▲✃➷➴Ø❒☞➘❀➺➼Ù☞➚➷❰✖Ó ❅➳ ✳✫é✒ä❅✶ß ✵❷✦ä ✒❖P✗P❅Þ❅Û✭ÿ➼Þ✢ë➷❈æ ✵☛✟✞ä☞✦Þ ✒❘✵✍ä☞Þ ❙▲à➾é✕❖ ê ❯✵Ý❅é⑨ç☞à❊è●Ý✬ë ✌Ý ❱❅Ý☞á⑨à➾á✿✝✍❙◆★❯ è✭❲✶❳ ➲✓ô☞ô❅õ✢✍➳ ✚ í ✛❈✛ ➳ ➳ ➳ ✛ ô☞õ ➳✚ ➲✓ò✢➳ ✮ î ✪✁Ô ❆ ➹❩✢✧ ❰➾❮✍❰➾➚ ➲ Ó✕ú❅❰▲Õ➾➴➷×û➴➷Õ➾➻▲✃➷➴Ø❒☞➪ ➘ ✚ ✃➷✃Øú í ✛✜✛ ✃➷➽❊✃➷ú ➽❭➽ Õ➾➳ ➴➷➽ ✃➷❰✖î✢Ó✕➳ ❰▲❒❅❰➾➺➼➺ ➸ ✛✍➘ Ô ✢ ❆ ➘❅❰➾➹ Õ ✛ Õ➾❒☞Ð ✃➷❒❅➴➷❮✍❒❅➘☞❰➾➘ Ó✕➻▲Ð✆ú☞➚➷➴Ø➘☞➸ ✃➷Ð✆➚ ➲✓î✢➳ ✮ î ✪②➪③➮✏Õ➾➻▲➚➷➻➾ü❅➚➷❰✾Ï✽❰▲Õ➾✃➷❒❅✤➺ ✣❊➺➼➻➾ú ✚ ➴➷Õ✖Ó ✖ ➮✏❬Ï ✣ ✔❀➲❅➳ ï ➮❷ú☞❰➾Õ▲➴Ø×✜➴ØÕ▲➻➾✃➷➴➷❒❅➘✬▲➪ ✮ î ✪ ➶③❰▲Õ➾❒❅Ð②Ð②❰➾➘❅ù☞➻➾✃➷➴➷❒❅➘✬➪▼➮❷❰▲ú ñ ò❅ï☞ï✢➲ ✚ í ✛❈✛ ➳ î✬➳ ✛ ✛ ✛ ➲ ★✢➳ ✃➷✮✰❰➾Ð②❰➾✃➷ü❅✃✡❰▲Ó ✚ ➺ ❰➾➺➼❰➾Õ▲➱✬➪❅➪ Ô ✃➷✃➷➳ ú➪❅➵ ➽❭❃ ➽❭Ô●➽ Ô✤➽ ➮✺✩ ñ ❒❅➴➷➘❅➺➼➸ù☞❰➾❐✢ú❅➶ ❰▲➘❅➮❷ù☞Ï❬❰➾➘❅✣ ✃ ✥ ➹✷❭➹ ✧❩Ï✽➴✡Ó✕Ù☞➻➾➚➷❈➴ ❇➾❰▲➺✓➪❅➴Ø➘ ✳✫é✒ä❅✶ß ✵❅✓ä ✒❬ÿ➼è❭è●❪ì ✎✁❫✍P✌✾✢å③ä☞❵é ❴✑✎ ✁ ❵á ✷☞✁ ä ✡❲ä❅Þ✷ÿ➼Þ✢ë à➾ç☞é⑨Ý✢ë➷â ä❅Þ✷Ý❅❛á ✡✬à➾ß❢ë á✆ä✦✒✛è●à▲ß✖â á❷â ä❅Þ✷✌ý ☞✍✡✌✡✬ä❅é❷ë✞Ý☞Þ❅Û✷è●Ý✢ë Ý✷ì➃â Þ✢â Þ ➸✬➪ ✖ ✩✫ù✬Ó ➳ ✔ ö③❒ ✚ ➻▲➘❅❰➾Õ❢➪✵➹ ➳ ➪ ➹✷➚Ø➻▲ù❅❰▲➘❅➴➷Õ✖ò☞➪➇ï❅ï❅Ô ò✬➳ ✻➪➳ ✧✢➻▲❮☞➺➼➻▲Õ✖✤➪ ✬ ➳ ➪②➻❢Ó⑨Ó✕❒☞Õ➾➴➷➻➾✃➷❰➾ù❑✃➷❒❑✃ ✚ ❰❱Õ➾❒☞➘❅×û❰▲➺➼❰➾➘❅Õ▲❰✖❁Ó ✩✫✪③✞➹ ✧ ✛❈✥☛❃ Ô●Ô ï☞ò ➪✽➬●❰➾➚✡Ó✕➴➷➘☞➱☞➴✡➪ Ò✵➴Ø➘☞➚➷➻➾➘❅ù✬➪        ½                                    !" #           $% &$'                 (     %       )       (   %      Æ %  *  ( %            +                 +,# ( (         %    -%# $      )       -% +  %        .      $ + +  % +       +        + %  % /                                                !℄# $        %                 Æ      &      &      #           &       '     #            (℄ )℄ !*℄   &   &        &       &     &               # +           ,℄ !-℄ !.℄ !/℄ !!℄         &      %     &    # 0          &                1             '    #  2   3    4 3            &                   # 0      34    3  &   5  6  -℄                                   &        # ½ )   +   % ( 0  2!  &3   4%'       ( 1 56 Bruzzese and Davino !   5%//            &          & 0  '    0           #               # 0   -   '           &     # 7  .        '                 # 7           '     *#           8 9 ½  ¾       &           ##    & &         9 ½  ¾       &                  &   ##       : &  #               ;                          %          9 # $          &    ;  ß ß    ; <     ;      9         9        &      &                  #                  &             &       # +                         & &            &&     %       #     7  &                             # =         2     2 &    &         #        '             &  ;         &        &       &  # +   '           &  !.℄#    &            &        &     2                2         # 5             &                               '                 #                   '                    &                  Visual Post Analysis of Association Rules 57  % 3      $% 6  2 &           #                          1   #  2   3    4 3  &        '               <                             #       2  &     0  =      &         2              #     '                  &   2    # 0      34    3       *℄ &           &       5  6   '                              -'  .℄#      &          & 0 0  &              Æ      ))>(      &  0           &'  #      .) 1                    #        & &  &'            2             %  2          %     # 0               #                &                          ))>(    #                               %  2     /#/!   2    /#/!   &   .(?,.  &   #         +                       &  &  1          2        # 0                 &      2    &                       2    # @             &                2        2                    % #             2 !        &                 # 5 2 !               &        &   !    2         2 !&     &             1                 2   # 58 Bruzzese and Davino 7   5%//       Rule: A & B --> C Rule: A & B --> C Rule: A --> C (b) (c) (a)             5 %      % 8    0                                       &      &    2              ;  9        <     ! 0  ℄ !< /         &       &                <     9 /              &        2      A    %          & &     <   ℄/< !                &            % #                  '          &           # 4  '        & 0 &  !),! >℄                     '    '     #                &                &    B         # 0                                               # 7             ?℄                 &   # 0          &     &;   ß     &           &   2               <  Visual Post Analysis of Association Rules 59  % 3      $% 9                      %                < ß                  2                      &   2   < ß      2                         < ß             &     &  #  C    4 3    &         0    # 0 2 -                  !>(-       %  ;      #                                  & #    )>   &   &      &        # ß            % +    -% -%   0 2 .  &   !/,*      % %        2      /#(          & B 2  7  *#-   # 0            &                      2       2                     #                         2   2                   # 60 Bruzzese and Davino "   5%//           3    % +    -% -%      Ê  9 3    % %     .    %  Visual Post Analysis of Association Rules 61  % 3      $% : 5   *>.     2      /#(                     2 *#   &               &   0=                      %                     &  &  &   #          C4    3     4$@             &     .℄   2        2           &                          2         &                -'  #  &   &              2     %                 2      %            2  #                          &     &   &          &   #  &  &          #      &    #                    2               %               2          %    #                 &    &     %  #               '-                          2                   #    &      &                                    #                   2      2                               &    #      *℄  &      &                            2            #  C  3            &                        &    % #      2      2     2      C  3      9 / (#    C  3   &         ; ¼ ;  9   ½ ;        62  Bruzzese and Davino   5%//                    &     '  &          &   %    ' D   &    #    &    &             &    ;  ;  9 / (          &        &  ;   / (<     Æ   # 0       &;     9  ;   /(  &   /( !  /(       /(      &          /< !               .(?,.      0          &#      (%   %    5        3%    !  6 %(  % "; 6!:! "7; 9"!  % .  " ; ; 96 * % .  : 9 9  % %     * % %  9  9  9  9    2  & !/E          F    2       2     #           &   ---.-                  # =    C  3      !(>-         2   %  /#(.    ,(.>         2      /#(.          #                            2                 #  !(>-           &    &      &                                   -'     #  6         6  -℄   Visual Post Analysis of Association Rules 63  % 3      $% ;     G -          &                 &  1   &             %      <    2          # $      &  &       !                  /   # 1        G- &   6 ;            & &             2           %         2         & &    &    B      2      #               &                          %  # +   6        &                    2        &       2            & 2  &       # +     &  1 & 6                         %   # if I NT 7 6 if RBS 5 if CTS if PS A 4 then PRS if I NP then PSA 3 then RBS then PRB 2 if P RM if ITV if P RS if RRE if RCS 0 if RSV if RED then PMG 1 then LOT if DOC 1 then FCM if COM if LOT then FLM if F CM if P CI if P DI if I N C then PDI if FTV then PRM if P M if P RB Gif FL M then INC if RTL if RTG 5 4 3    2 1 0 1 )   $  0 2 (         %        2               &        2 >          #    64  Bruzzese and Davino   5%//       1 if DOC then FCM 0.5 if LOT if COM then MAN if FCM then FLM then TG then FLD then FLC then CAR if MAN C if CAR S if CON if INC if MUS then MUS then FTV then MET if FLD if FTV if MET then PCI if FLM if SCA if PCI 0 then PDI 0. 5 if PMG then INC 1 0. 6 0. 4  0. 2   0 0.2 0.4 0.6 0.8 1 1.2  /     $    2      &    &              B                Æ      #                      2   ;  2 (  2 %             2   &    #    &                          &    %    '         # 5      &          %                '      # 5    %   58                       +6   6=7           &       % #     2       2 *   '               # +       &              &   &        &                       2  # 0                                           1  %                     ' &                 B    1  %   # Visual Post Analysis of Association Rules 65  % 3      $%       0          '    &        &         2              &    &        #                              %        &   C    4 3   C4    3   #               &   '   &   ##           &'                         & 6 #               '    0=                    &                      # 5                       &B                &     &                        2    #    +# $#   , # ) 8 + # <      $% (+     4 (# 3      ;;6  = 2   # # >  # # & ;;6' !:?! " ! 5/@  # A3< 4B  @# %# 3  & ;:6' 6 5%//#  8  # < 3% # C*    % /     $%#     #  !#  7# &!' 7" ?7:!  7 5%//#  8  # <    3%        $%# %    #  " &! ' 6:?6; 9 D # D 8 >  # <        $% (     3  5  # A=#   D E# 3=  &'<  !  3    %    # 3  # D ( &!' 7;;?97 " (# <    = #    4   D#  5 4 )  $ =6!!: # 7  5 4   .  & ; ' : (# <  %    +  3  # %    #  6#  # & ;;' 7:?"7  F # #  # D# $, # 3# ) # D# 8 ,# <      %           %# 3      )            F+  F ;7# & ;;7' 7 ?7: ; 4 %# 5# D%# > 8 # G< 3%   % /         # 3       =F       F+       &F ;;'# %% 9 #  # #  & ;;;'  4 %# 5# D%# ># ># F 8  # <  %   C*       $% 3      3 .       F+       &3F ;;'   C#   !"!# 5 E  & ;;;' 66 ! Bruzzese and Davino   5%//          # # 4, # 4# $  # F 8 %  <     3%    3# >,      ;;;  = 2 $ %      F+   & ;;;' ! ) # D# F # # $, # 3 # D# F 8  # D< 3%   %         % >,      C 4;9 >,     #   4 #  F+    (# D, # = #   ;;9 & ;;9' 7:?9! 6 >(# < 2 3%        = /  H%       $% 3    F+       >, #   & ;;' 7 ># 3# > # 3# ) # A<  % /     $%  )*    > # =# F #  &'< 3    CCC     % /  B;;# CCC  3# 4  #  & ;;;'         )  $<  )< %    4<   )< ) . )< %  <   3<  +  3$5<   3<  $C <    $)=<  +  <    )4 <    2 <   2<  % 4 < . +    %  < % )< (, C)< +    3 <   %  3$ <    $5<   $$C<   % /  $)4< % I  <   )4< ) . 2<    <   4 < . 3<   + 42)<  < %  3 =<  3$<     $<    $<  %  $5<  )=< + )HJ< -% /  + Cooperation between automatic algorithms, interactive algorithms and visualization tools for Visual Data Mining François Poulet ESIEA Recherche 38, rue des Docteurs Calmette et Guérin Parc Universitaire de Laval-Changé 53000 Laval, France poulet@esiea-ouest.fr Abstract. Visual data-mining strategy lies in tightly coupling the visualizations and analytical processes into one data-mining tool that takes advantage of the strengths from multiple sources. This paper presents concrete cooperation between automatic algorithms, interactive algorithms and visualization tools. The first kind of cooperation is an interactive decision tree algorithm called CIAD+. It allows the user to be helped by an automatic algorithm based on a support vector machine (SVM) to optimize the interactive split performed at the current tree node or to compute the best split in an automatic mode. This algorithm is then modified to perform an unsupervised task, the resulting clustering algorithm has the same kind of help mechanism based on another automatic algorithm (the k-means). The last effective cooperation is a visualization algorithm used to explain the results of SVM algorithm. This visualization tool is also used to view the successive planes computed by the incremental SVM algorithm. 1 Introduction Knowledge Discovery in Databases (or KDD) can be defined [1] as the non-trivial process of identifying patterns in the data that are valid, novel, potentially useful and understandable. In most existing data mining tools, visualization is only used during two particular steps of the data mining process: in the first step to view the original data, and in the last step to view the final results. Between these two steps, an automatic algorithm is used to perform the data-mining task. The user has only to tune some parameters before running his algorithm and wait for its results. Some new methods have recently appeared [2], [3], [4], trying to involve more significantly the user in the data mining process and using more intensively the visualization [5], [6], this new kind of approach is called visual data mining. In this paper we present some tools we have developed, which integrate automatic algorithms, interactive algorithms and visualization tools. These tools are two interactive classification algorithms and a visualization tool created to show the 68 Poulet results of an automatic algorithm. The classification algorithms use both human pattern recognition facilities and computer calculus power to perform an efficient user-centered classification. This paper is organized as follows. In section 2 we briefly describe some existing interactive decision tree algorithms and then we present our new interactive algorithms, the first one is an interactive decision tree algorithm called CIAD+ (Interactive Decision Tree Construction) using support vector machine (SVM) and the second is derived from the first one and performs unsupervised classification (clustering). In section 3 we present a graphical tool used to explain the results of support vector machine algorithms. These algorithms are known to be efficient but they are used as "black boxes", there is no explanation of their results. Our visualization tool graphically explains the results of the SVM algorithm. The implemented SVM algorithm can modify an existing linear classifier by both retiring old data and adding new data. We visualize the successive separating planes computed by this algorithm. Section 4 concludes the paper and lists some future work. 2 Interactive decision tree construction Some new user-centered manual (i.e. interactive or non-automatic) algorithms inducing decision trees have appeared recently: Perception Based Classification (PBC) [7], Decision Tree Visualization (DTViz) [8], [9] or CIAD [10]. All of them try to involve the user more intensively in the data-mining process. They are intended to be used by a domain expert not the usual statistician or data-analysis expert. This new kind of approach has the following advantages: - the quality of the results is improved by the use of human pattern recognition capabilities, - using the domain knowledge during the whole process (and not only for the interpretation of the results) allows a guided search for patterns, - the confidence in the results is improved, the KDD process is not just a "black box" giving more or less comprehensible results. The technical part of these algorithms are somewhat different: PBC and DTViz use an univariate decision tree by choosing split points on numeric attributes in an interactive visualization. They use a bar visualization of the data: within a bar, the attribute values are sorted and mapped to pixels in a line-by-line fashion according to their order. Each attribute is visualized in an independent bar (cf. fig.1). The first step is to sort the pairs (attri, class) according to attribute values, and then to map to lines colored according to class values. When the data set number of items is too large, each pair (attri, class) of the data set is represented with a pixel instead of a line. Once all the bars have been created, the interactive algorithm can start. The classification algorithm performs univariate splits and allows binary splits as well as n-ary splits. CIAD is described in the next section and, in section 2.2, we present a new version of CIAD (called CIAD+) with a help tool added to the interactive algorithm allowing the user to perform an automatic computation of the best bivariate split. Cooperation Between Automatic Algorithms 69 [9] uses a two dimensional polygon or more precisely, an open-sided polygon (i.e. a polyline) in a two dimensional matrix. It is interactively drawn in the matrix. The display is made of one 2D matrix and one-dimensional bar graphs (like in PBC). Only PBC provides the user with an automatic algorithm to help him choosing the best split in a given tree node. The other algorithms can only be run in a 100% manual interactive way. Attr.1 1 5 2 9 3 6 Attr.2 5 7 1 3 2 9 Class A A B B A B Attr.1 1 2 3 5 6 9 Class A B A A B B Attr.2 1 2 3 5 7 9 Class B A B A A B Attr.1 Class A Class B Attr.2 Fig. 1. Creation of the visualization bars with PBC 2.1 CIAD CIAD uses a bivariate decision tree using line drawing in a set of two-dimensional matrices (like scatter plot matrices [11]). The first step of the algorithm is the creation of a set of (n-1)2/2 two-dimensional matrices (n being the number of attributes). These matrices are the two dimensional projections of all possible pairs of attributes, the color of the point corresponds to the class value. This is a very effective way to graphically discover relationships between two quantitative attributes. One particular matrix can be selected and displayed in a larger size in the bottom right of the view (as shown in figure 2 using the Segment data set from the UCI repository [12], it is made of 19 continuous attributes, 7 classes and 2310 instances). Then the user can start the interactive decision tree construction by drawing a line in the selected matrix and performing thus a binary, univariate or bi-variate split in the current node of the tree. The strategy used to find the best split is the following. We try to find a split giving the largest pure partition, the splitting line (parallel to the axis or oblique) is interactively drawn on the screen with the mouse. The pure partition is then removed from all the projections. If a single split is not enough to get a pure partition, each half-space created by the first split will be treated alternately in a recursive way (the alternate half-space is hidden during the current one's treatment). 70 Poulet At each step of the classification, some additional information can be provided to the user like the size of the resulting nodes, the quality of the split (purity of the resulting partition) or overall purity. Some other interactions are available to help the user: it is possible to hide / show / highlight one class, one element or a group of elements. Fig. 2. The Segment data set displayed with CIAD 2.2 CIAD+ The first version of the CIAD algorithm was only an interactive algorithm. No help was available for the user, and sometimes, it was difficult to find the best pure partition in the set of two-dimensional matrices. We have decided to provide such help. Our first intention was to use a modified OC1 (Oblique Classifier 1) algorithm [13]: OC1 performs real oblique cuts (we have a real n-dimensional hyperplane with n-dimensional data) and in CIAD, the cuts are only "oblique" in two dimensions. The plane coefficients are null in all the other dimensions. We have made another choice: Cooperation Between Automatic Algorithms 71 we use a support vector machine (SVM). This algorithm is equivalent to OC1 in its simplest use, and will allow us to benefit from all its other possibilities for further developments. 2.2.1 The SVM algorithms The SVM algorithms are kernel-based classification methods. They can be seen as a geometric problem: to find the best separating plane of a two classes data set. A lot of methods can be used to find this best plane, and a lot of algorithms have been published. A review of the different algorithms can be found in [14]. They are used in a wide range of real-world applications such as text categorization, hand-written character recognition, image classification or bioinformatics. We briefly describe here the basis of the algorithm, from the geometrical point of view. Fig. 3. Two possible separating planes The aim of the SVM algorithm is to find the best separating plane between the ndimensional elements of two classes. There are two different cases according to the nature of the data: they are linearly separable or not. In the first case, the data are linearly separable i.e. there exists a plane that correctly classifies all the points in the two sets. But there are infinitely many separating planes as shown in Fig.3. Geometrically, the best plane is defined as being furthest from both classes (i.e. small perturbations of any point would not introduce misclassification errors). The problem is to construct such a plane. It has been shown in [15] that this problem is equivalent to finding the convex hull (i.e. the smallest convex set containing the points) of each class, and then to finding the nearest two points (one from each convex hull); the best plane bisects these closest points. Fig. 4. The best separating plane bisects the closest points In the second case, the data are not linearly separable (i.e. the intersection of the two convex hulls is not empty). There is no clear definition of what is the "best" plane. The solution is to create a misclassification error, and to try to minimize this error. 72 Poulet 2.2.2 SVM in CIAD+ We use two different SVM algorithms in CIAD+. The first one is the geometric version described in section 2.2.1 for the linearly separable case. The convex hulls are computed in two dimensions with the quick hull algorithm [16], then the two closest points of the convex hulls are computed with the rotating calipers algorithm [17]. For the linearly inseparable case, a lot of solutions have been developed and compared, we have chosen one of these algorithms: the RLP (Robust Linear Programming) algorithm [18] because it is the best one when the data are not linearly separable. The RLP algorithm will compute the separating plane wx=γ minimizing the average violations: 1 m 1 k ( Bi w − γ + 1) + ∑ (− Ai w + γ + 1) + + k ∑ m i =1 i =1 (1) of points of A lying on the wrong size of the plane wx=γ+1, and of points of B lying on the wrong side of the plane wx=γ-1 as shown in Fig.5. This algorithm computes the n-dimensional hyperplane, we have modified it to compute the best two-dimensional plane. wx=γ+1 wx=γ Ai wx=γ-1 Bi Fig. 5. Optimal plane for linearly inseparable data SVMs in CIAD+ are used to help the user. The first kind of help is when the user draws interactively the separating line on the screen with a pure partition on one side, the optional help optimizes the line position to reach the best line position (furthest from both groups) with the computation of the closest points of the convex hulls. The second kind of help is the same case except there is no pure partition, the best line position is computed with the RLP algorithm in the two dimensions corresponding to the selected matrix. The last kind of help is used when the user cannot find a separating plane, the help algorithm has to compute the best separating plane among all the ones corresponding to the projections along pairs of attributes. So we compute all the separating lines in the 2D projections and we keep the best one. Cooperation Between Automatic Algorithms 73 This help mechanism can be turned on / off by the user. It slightly improves the accuracy of the results on the training sets (this result may be more significant according to the kind of user), more on the test set, and it considerably reduces the time needed to perform the classification and increases the ease of use. An example is shown on figure 6, the left part is the original line drawn interactively by the user on the screen and the right part shows the transformed line (the best separating plane computed with the convex hulls). Fig. 6. An example of the automatic best separating plane on iris data set 2.3 Clustering The interactive algorithm described in the previous section can also be used for unsupervised classification. The computation of the convex hulls and the nearest points can be computed with or without the class information. The proposed algorithm can perform either usual decision tree (supervised classification) or clustering (unsupervised classification). This kind of approach allows the user to perform clustering of the dataset easily using its pattern recognition capabilities and avoiding the usual complexity of the other algorithms. The same kind of help as for decision tree construction is provided to the user: the separating line drawn in a 2D projection can be optimized to be the furthest from the two clusters. But there is one difference with the decision tree construction algorithm, when the user does not perceive clearly a separating line, this line can be computed automatically (with a modified SVM algorithm). This SVM algorithm cannot be used without the class information, so we have chosen a k-means algorithm. All possible partitions into two clusters are searched for in each matrix and the best one is kept. Then we compute the convex hulls and nearest points to find the best separating line in the same way as for decision tree construction. 74 Poulet As shown in [19], this kind of algorithm may not be very efficient for high dimensional data sets because axis-parallel projections may lead to an important loss of information. This restriction exists anyway because of the graphical representation we use (we cannot display a very large quantity of scatter plot matrices). On the other hand, the results are more comprehensible because we only use one attribute on each axis and not a linear combination of various number of attributes for the axes. 2.4 Some results of interactive algorithms The characteristics of the different algorithms are summarized in table 1. Some results of interactive algorithms compared to automatic ones have been presented by their authors. To summarize them, we can say their results concerning efficiency, are generally at least as good as automatic decision tree algorithms such as CART (Classification And Regression Trees) [20], C4.5 [21], OC1 [13], SLIQ (Supervised Learning In Quest) [22] or LTree (Linear Tree) [23]. The main difference between the two kinds of algorithms is the tree size. Most of the time, interactive algorithms have smaller tree sizes. This tree size reduction can significantly increase the result comprehensibility. Furthermore, as the user is involved in the tree construction, his confidence in the model is increased too. This may be a little less significant for the LTree algorithm because of the open-sided polygon used in the classification: they are easy to understand during the visual construction step of the tree, but without this information, the resulting equations may be not so easy to understand. PBC Vis. technique bar graphs + 1 scatter plot bar graphs DTViz bar graphs CIAD+ set of scatter plot Ware split poly-line binary univariate n-ary univariate n-ary bivariate binary + large dataset tree size large datasets help large datasets result comprehensibility plot of qual. x qual. attr. loss of information tree size help plot of qual. x qual. attr. loss of information Table 1. Comparison of interactive decision tree algorithms If we compare the interactive algorithms, we can say PBC and DTViz are particularly interesting for large data sets (because of the pixelisation technique used), but their pixelisation technique introduces some bias in the data: for example two classes very far one from the other have the same representation as the same two classes very near one to one other. We lose the distance information in this kind of representation. Ware and CIAD+ will provide smaller trees because they can use bivariate splits, but their kind of visualization tool (scatterplot matrices) are not at all suitable for the display of two qualitative attributes (a lot of points have the same projection). Only two algorithms provide the user with a help, PBC and CIAD+, this is a significant advantage of these two algorithms because during the decision tree Cooperation Between Automatic Algorithms 75 construction, there is often a least one particular step where the best split is difficult to visually detect. The kind of cooperation between automatic and manual algorithms in PBC and CIAD+ shows the interest of mixing the human pattern recognition facilities and the computer processing power. The human pattern recognition facilities reduce the cost of the computation of the best separating plane, and the computer processing power can be used at low cost (for a single step, instead of the whole process) when the human pattern recognition fails. 3 Visualization of SVM results Another kind of cooperation is between automatic algorithms and visualization tools used to show the results. As described in the previous section, SVM are today widely used because they give high quality results, but they are used as a "black box". They give high quality results, but there is no explanation of these results. One paper [24] talks about SVM results visualization: they use projection-based tour [25] method to visualize the results. They use a visualization of the distribution of the data predicted class (by the way of histograms), a visualization of the data and the support vectors in 2D projection, and examine the weights of the plane coordinates to find the most important attributes for the classification. The authors recognize that their approach is very "labor intensive for the analyst. It cannot be automated because it relies heavily on the analyst's visual skills and patience for watching rotations and manually adjusting projection coefficients." 3.1 Visualization of the SVM separating plane Our approach is to visualize all the intersections of the 2D planes (of the scatter plot matrices) with the separating plane computed by the SVM algorithm. We have chosen to use the incremental SVM algorithm from [26]. This algorithm gives the coefficient values of the separating hyperplane and the accuracy of the algorithm. We visualize the intersection of this hyperplane with the 2D scatter plot matrices, i.e. a line in each matrix (as shown in Figure 7 with the diabetes data set, from the UCI repository). As we can see on figure 7, the resulting lines do not necessarily separate the two classes, the hyperplane does separate the data (the accuracy of the incremental SVM is 77,8% on this dataset), but not its "2D projections". It is only an approximate interpretation of the results. All 2D representations of one n-dimensional feature will lead to a loose part of the information like the lines we get here or the support vectors displayed in [24]. This kind of representation seems more comprehensible than the support vectors, but it can only be used with a linear kernel function. For large data sets, it is possible to only display the separating plane and not the data. The SVM algorithm used is able to classify 1 billion points, and such a quantity of points cannot be displayed in a reasonable time (furthermore this kind of representation is not at all suitable for such data set size). 76 Poulet For other kinds of kernel functions (not linear), this method cannot be used. Fig. 7. Visualization of the separating hyperplane of the diabetes dataset. 3.2 Visualization of the evolution of SVM separating planes A very interesting feature of the incremental support vector machine we have used is its capability to modify an existing linear classifier by both withdrawing old data and adding new data. We use our visualization tool to show the successive modifications of the separating plane projections. The first plane is calculated (and projected on the 2D matrices) and then blocks of data are successively added and withdrawn. The modification of the plane is computed and the corresponding projections are displayed in the matrices. This kind of visualization tool is very powerful to examine the variations of the separating plane according to the data evolution. Even if the projections used lose some amount of information, we know what the attributes involved in the Cooperation Between Automatic Algorithms 77 modification of the separating plane are. The evolution of the n-dimensional plane is very difficult to show in another way. The authors of the paper describing the algorithm measure the difference between planes by calculating the angle between their normals. These angle values are then displayed like circle radii. Fig. 8. Three linked tools in the FullView environment 4. Conclusion and future work Before concluding, some words about the implementation. All these tools have been developed using C/C++ and three open source libraries: OpenGL, Open-Motif and Open-Inventor. OpenGL is used to easily manipulate 3D objects, Open-Motif for the graphical user interface (menus, dialogs, buttons, etc.) and Open-Inventor to manage the 3D scene. These tools are included in a 3D environment, described in [27], where each tool can be linked to other tools and be added or removed as needed. Figure 8 shows an example with a set of 2D scatter plot matrices, a 3D matrix and parallel coordinates. The element selected in the 3D matrix appeared selected too in the set of scatter plot matrices and in the parallel coordinates. The software program can be run on any platform using X-Window, it only needs to be compiled with a standard C++ compiler. Currently, the software program is developed on SGI O2 and PCs with Linux. 78 Poulet In this paper we have presented two new interactive classification tools and a visualization tool to explain the results of an automatic SVM algorithm. The classification tools are intended to involve the user in the whole classification task in order to: - take into account the domain knowledge, - improve the result comprehensibility, and the confidence in the results (because the user has taken part in the model construction), - exploit human capabilities in graphical analysis and pattern recognition. The visualization tool was created to help the user in understanding the results of an automatic SVM algorithm. These SVM algorithms are more and more frequently used and give efficient results in various applications, but they are used as "blackboxes". Our tool gives an approximate but informative graphical interpretation of these results. A forthcoming improvement will be another kind of cooperation between SVM and visualization tools: an interactive visualization tool will be used to improve the SVM results when we have to classify more than two classes. References 1. U. Fayyad, G. Piatetsky-Shapiro, P. Smyth, R. Uthurusamy, Eds, "Advances in Knowledge Discovery and Data Mining", AAAI Press, 1996. 2. P.Wong, "Visual Data Mining", in IEEE Computer Graphics and Applications, 19(5), 20-21, 1999. 3. F. Poulet, "Visualization in data mining and knowledge discovery," in Proc. HCP'99, 10th Mini Euro Conference "Human Centered Processes" ed. P. Lenca (Brest, 1999), 183-192. 4. M.Ankerst, M.Ester, H-P.Kriegel, "Toward an Effective Cooperation of the Computer and the User for Classification" in proc. of KDD'2001, 179-188. 5. C.Aggarwal, "Towards Effective and Interpretable Data Mining by Visual Interaction", in SIKDD Explorations 3(2), 11-22, accessed from www.acm.org/sigkdd/explorations/. 6. B.Schneiderman, "Inventing Discovery Tools: Combining Information Visualization with Data Mining", in Information Visualization 1(1), 5-12, 2002. 7. M.Ankerst, "Visual Data Mining", PhD Thesis, Ludwig Maximilians University of Munich, 2000. 8. J.Han, N.Cercone, "Interactive Construction of Decision Trees" in proc. of PAKDD'2001, LNAI 2035, 575-580, 2001. 9. M.Ware, E.Franck, G.Holmes, M.Hall, I.Witten, "Interactive Machine Learning: Letting Users Build Classifiers", in International Journal of Human-Computer Studies (55), 281292, 2001. 10.F.Poulet, "CIAD: Interactive Decision Tree Construction" in proc. of VIII Rencontres de la Société Francophone de Classification", Pointe-à-Pitre, 275-282, 2001 (in french). 11.J.Chambers, W.Cleveland, B.Kleiner, P.Tukey, "Graphical Methods for Data Analysis", Wadsworth (1983). 12.C.Blake, C.Merz, UCI Repository of machine learning databases, [http://www.ics.uci.edu/~mlearn/MLRepository.html]. Irvine, CA: University of California, Department of Information and Computer Science, (1998). 13.S.Murthy, S.Kasif, S.Salzberg, "A system for induction of oblique trees", Journal of Artificial Intelligence Research 2, 1-32, 1994. Cooperation Between Automatic Algorithms 79 14.K-R.Muller, S.Mika, G.Ratsch, K.Tsuda, B.Scholkopf, "An Introduction to Kernel-Based Learning Algorithms" in IEEE Transactions on Neural Networks, 12(2), 181-201, 2001. 15.K.Bennett, E.Bredensteiner, "Duality and Geometry in SVM Classifiers", in proc of the Seventeenth International Conference on Machine Learning, Pat Langley Editor, Morgan Kaufmann, San Francisco, 57-64, 2000. 16.C.Barber, D.Dobkin, H.Huhdanpaa, "The Quickhull algorithm for convex hulls", in ACM Transactions On Mathematical Software, 22, 469-483, 1996. 17.G.Toussaint, "Solving geometric problems with the rotating calipers", in proc. of IEEE MELECON'83, Athens, Greece, pp. A10.02/1-4, 1983. 18.K.Bennett, O.Mangasarian, "Robust Linear Programming Discrimination of Two Linearly Inseparable Sets", in Optimization Methods and Software, 1, 23-34, 1992. 19.C.Aggarwal, P.Yu, "Redifining Clustering for High-Dimensional Applications", in IEEE Transactions on Knowledge and Data Engineering, 14(2), 210-225, 2002. 20.L.Breiman, J.Friedman, R.Olsen, C.Stone, "Classification and Regression Trees", Wadsworth, (1984). 21.J.Quinlan, "C4.5: Programs for Machine Learning", Morgan-Kaufman Publishers, 1993. 22.M.Metha, R.Agrawal, J.Rissanen, "SLIQ: A fast scalable classifier for data mining", in proc. of the 5th International Conference on Extending Database Technology, Avignon, France, 1996, 18-32. 23.J.Gama, P.Brazdil, "Linear Tree" in Intelligent Data Analysis, 3, 1-22 (1999). 24.D.Caragea, D.Cook, V.Honavar, "Gaining Insights into Support Vector Machine Pattern Classifiers Using Projection-Based Tour Method", in proc. of KDD'2001 Workshop on Visual Data Mining. 25.D.Asimov, "The Grand Tour: A Tool for Viewing Multidimensional Data" in SIAM Journal of Scientific and Statistical Computing, 6(1):128-143, 1985. 26. G. Fung, O. Mangasarian, "Incremental Support Vector Machine Classification" in proc. of the 2nd SIAM International Conference on Data Mining, Airlington, USA, Apr.11-13, 2002. 27.F.Poulet, "FullView: A Visual Data-Mining Environment", in International Journal of Image and Graphics, 2(1), 127-144, 2002. Defining Like-minded Agents with the Aid of Visualization Penny Noy and Michael Schroeder City University, London EC1V 0HB, UK p.a.noy, msch @soi.city.ac.uk Abstract. Profile carrying agents offer the opportunity of meeting like minds and increasing efficiency in many information search applications. Profiles can also increase the sophistication of relationships and interactions in multi-agent systems in general. Such profiles (feature lists) may be of the agent owner (interests) or of the information sought (specifications). At the same time there is a continuing increase of profiling data of increasing complexity becoming available. The paper first describes various possibilities for defining profiles and selecting similarity metrics in the agent interaction context. General and domain-specific data sets are specially constructed and used to provide a concrete view of the behaviour of the metrics with visualization. Visualization usually results in xy (or xyz) coordinates for each agent, placing them in a profile space. The use of these coordinates as a means of carrying and comparing one’s profile without revealing it to other agents is proposed. Thus this paper extends our previous work on visualizing multivariate data and proximity data in the agent scenario in two ways where visualization is not the end product: using visualization tools in the phase of similarity metric choice; proposing the use of coordinates as a profile. The paper seeks to illustrate, with these applications, an objective of visual data mining, namely the increased integration of visualization and analytic techniques. 1 Introduction Every second that passes witnesses millions of people searching for information via a myriad of means. Increasingly such means and media employed are electronic. Many people are seeking similar information and work on similar problems. Software agents carrying our profile can meet with the representatives of others and exchange important information or provide introductions. Consider the following messages from some of my (hypothetical) personal agents: myNetworkAgent: Good morning. Professor Blake in Helsinki has just started work on one of your main research areas. myShoppingAssistant: Here are the fourteen houses closest to your ideal specification. myInformationSearchAgent: Here are the four most relevant documents in your specified area ’agent matchmaking’ from the survey point of view. 82 Noy and Schroeder These agents may be providing real-time search responses or background information monitoring. The concepts are embodied in many search and classification applications. These may be ordinary searches, not explicitly involving other agents (or agents at all), but agents may assist in improving precision and speed. Of course the above scenario is far fetched in several senses: the degree to which information is searchable (i.e. specified according to agreed ontologies); the accuracy of the metrics in matching seeker with sought; the willingness of humans to allow an agent to carry their profile (and the security issues involved); privacy of information. The focus of this paper is the task of ’matching seeker with sought’. This may involve a user profile or a profile of a task or desired piece of information. It is in this general sense that the word profile is used here. To compare profiles a metric is needed. There are many examples of this in agent systems. For instance: Faratin et al use the maximum value distance in making negotiation trade-offs[10]; The Yenga system uses correlation to determine user interests and then a direct comparison to find a common joint interest [4]; Somlo and Howe use incremental clustering for profile maintenance in information gathering web agents based on term vectors of documents the user has shown interest in [14]; GRAPPA (Generic Request Architecture for Passive Provider Agents) is a configurable matchmaking framework which includes demand and supply profiles and has been applied to matching multidimensional profiles of job applicants to vacancies [18,17]; Ogston and Vassiliadis use minimal agents working without a facilitator to match services and providers [9]. The many possible application areas divide broadly into two areas - matchmaking (e.g. matching services to clients, people connector) and search. The general topic lies within organizational concepts in multi-agent systems and improving learning with communication [21]. The finding and remembering of appropriate, like-minded agents can be centralized, left as a diffusion process or engineered in a computational ecology sense [5,4,9]. In the agent domain, as in others, visualization is often seen as a separate application that one adds on to an application for a variety of purposes. It may be of value from this point of view, but it can do more. Visual datamining seeks, amongst other things, to give the human visual system a more central role in the knowledge discovery process, to increase the integration of visualization and datamining techniques. This can be approached in a variety of ways, taking advantage of the human visual system’s pattern recognition abilities, for instance, or presenting overviews of large amounts of data in novel ways. Related to this, but with a different focus, is another way, which the work presented here seeks to illustrate: the merging of the visualization and analytic processes. Here the question of whether visualization is the end result is not so important. Two examples of this are presented here: – Overview understanding and similarity metric choice merge: Metrics are used for layouts involving dimension reduction, but different metrics produce different layouts. A metric (or other transformation process) may be needed for layout, especially in creating an overview of large complex datsets, but how is the user made aware of the different possibilities and their implications? At the same time many applications use a similarity measure - how can designers choose appropriately? – Viewer and computational object (an agent) merge: Our layout creates a topic space. If such topic spaces are valid (they are usually approximations), can we Defining Like-Minded Agents 83 use this notion of spaces (or surfaces) for software agents to use when meeting or seeking other agents? In this case the visualization concept is being used to assist in agent-orientated computation. Can this help in the pursuit of the use of visualization techniques for reasoning (diagrammatic reasoning [6])? Our earlier work looked at proximity data and multivariate data in the agent domain and indicated possible metric choices for visualization[11,13]. From the agent point of view, the question is how can we apply this and how can we assist in the problem of metric choice for profiling and classification in the agent domain. How can we meaningfully identify like-minded agents and then put this to use? The agent paradigm is considered by some to be a valuable way of looking at problems and therefore of general application. Our work in the agent and visualization fields seeks to use visualization to serve the agent community, but, from the agent-orientated computation point of view, suggests other uses of agent ideas within visualization. The paper first briefly surveys visualization possibilities for different types of data matrices and presents definitions of profiles, considering also the desirability of comparing profiles without revealing them. It then looks at metric choice and suggests strategies to improve choice using visualization techniques and the designing of a classification system for evaluating the metrics. The idea of using the visualization position coordinates as a profile is introduced and an example given. The nature of the examples in this paper is illustrative and the intention is to show the merging process, where visualization is not necessarily the end product, as well as to present the two specific applications. 2 Defining Profiles In general an agent’s profile is considered to be a vector of interests and behaviours (a feature list) or a similarity measure or sets and/or combinations of these [11,13]. The purpose of our work in visualization was to find layouts (in 2D or 3D) which would satisfy (usually approximately) these data either by using mathematical transformations (effective reductions via e.g. Principal Component Analysis (PCA) or distance metrics followed by Principal Coordinates Analysis (PCoA), spring embedding [3] or Selforganizing Map (SOM)[16]) or novel representations (e.g. colour maps, hierarchical axes, ’Daisy’, parallel coordinates[1,2,15]). PCA, SOM and PCoA are described briefly here as they are used in the discussion that follows: – Principal Components Analysis: PCA is a means by which a multivariate data table is transformed into a table of factor values, the factors being ordered by importance. The two or three most important factors, the principal components, can then be displayed in 2D or 3D space. – Self Organizing Map: The SOM algorithm [16] is an unsupervised neural net that can be used as a method of dimension reduction for visualization. It automatically organizes entities onto a two-dimensional grid so that related entities appear close to each other. 84 Noy and Schroeder – Principal Coordinates Analysis: PCoA is used for proximity data, finding first a multivariate matrix which satisfies the distances, then transforming this into its principal components so that the two or three most important factors can be displayed in a similar fashion to PCA. These visual representations may provide meaningful clusters or reveal patterns from which knowledge can be gained. A key problem in this area is that different methods produce different clusters (and cluster shapes). The determination of an appropriate metric 1 is a difficult problem for which general solutions are not evident. We propose the use of constructed data in a process called signature exploration [8] to assist in this area. This process uses specially constructed data sets to increase the user’s understanding of the behaviour of visualization algorithms applied to high dimensional data. Two developments suggest themselves from aspects of our previous visualization work: a tool for metric choice; the use of layout coordinates as a profile. – Tool for metric choice: Agents need to compare profiles, i.e. when they meet they need to be able to compare themselves (or their tasks) and get a measure of similarity which they can interpret. Assuming (for the moment) that they are carrying their profile with them, they will need to apply an algorithm to calculate a similarity measure by both submitting their profiles to the algorithm, either both independently of the other, or via an intermediary. In designing a specific application a decision (by the designer) needs to made about what similarity measure is appropriate. The tool for metric choice developed in application of the principle of signature exploration provides an interactive interface which can help the designer to choose the metric. – Use the layout coordinates as a profile: For layout on the screen, the data transformation or set of similarity measures results in x/y (or x/y/z) coordinates for each entity. For complex data this usually involves a significant error (i.e. it is normally not possible to find a layout which will satisfy the similarity measurements - on the one hand - and matrix transformations and truncations to 2 or 3 attributes rely on a sharp fall off of the relevant eigenvalues, which is unusual for complex data sets on the other hand). Nevertheless such algorithms are commonly used and thus the approximations involved are often adequate. The relevant point here is that the end result is that there is an x/y (or x/y/z) coordinate pair associated with each entity and within the current space of possibilities this locates their interest position. This suggests the possibility of them carrying a much more lightweight position profile with them, that also means they can compare positions without revealing profiles. The use xy or xyz coordinates as the profile avoids revealing the profile, but the implication is that either there must be a central entity which will do the calculation (and thus that one needs to reveal one’s profile to) and then give the agent its coordinates and the bounds of the space (so that it can judge relative similarity). Also this does not deal easily with dynamic situations (i.e. reflecting changing profiles), as it would require a periodic return to base to profile updating. A possible alternative is to calculate one’s own coordinates with respect to a number of reference 1 metric is here used in a general sense to mean a means of measurement which may not result in a numerical measure directly, i.e. possibly indirectly by means of layout position derived directly from SOM or PCA Defining Like-Minded Agents 85 points, i.e. calculate one’s proximity to the reference points and then find a position in space to satisfy this reduced set of distances. For instance for a feature list of length 5, consisting of a set of five possible agent interest areas and interest values in the range 0 to 1 (say), the following is an indication of the bounds of the space. ABCDE agent1 1 0 0 0 0 agent2 0 1 0 0 0 agent3 0 0 1 0 0 agent4 0 0 0 1 0 agent5 0 0 0 0 1 It may be unwise to base the position on a computation that satisfies the similarity measures to all of these vectors (since this increases the inaccuracy of the layout), but the agent could carry the set of coordinates for certain bounds (or other reference vectors) and profile position, having the calculations made back at base. These ideas are illustrated below. 3 Choosing a Metric - Possibilities For specific applications different metrics are used, this means that often an applications area uses one metric only. Measures may be chosen because of time complexity issues, rather than that they provide the most accurate or appropriate measure. There is also a link between the creation of the feature list and the metric choice (i.e. the formulation of the feature list affects which metric provides the most appropriate clustering) which is a further complication. In general terms the choice of metric and creation of the feature list should correspond to the required classification, but in many situations the starting point is an unknown set of data and clustering indications are sought. There is no training set and no classification. It is likely that different classifications exist. In fact there are hidden classifications, that is to say, the user has a set of things they are interested in and they would like to have the entities (other users, documents..) classified according to these groupings. One of the purposes of the signature exploration process that is being developed is to explore the mapping to clusters (via various metrics) of features of interest to the user. Originating as part of work to increase comprehension and choice of algorithms for visualization of complex data, it does not focus on feature list construction but on metric choice for a given feature set. In the process of constructing data sets for evaluation of the different options the user creates an ad hoc classification system for assessment purposes (demonstrated below). 3.1 How to choose - metrics, feature selection and weighting The first issue is to specify the variables to be used in describing the profile and the ways in which pairwise similarities can be derived from the matrix formed by the set of profiles. 86 Noy and Schroeder Many different measures of pairwise similarity have been proposed [7,19]. Some are closely related to one another. Measures are usually presented that are particularly relevant for comparing objects that are described by a single type of variable. This discussion restricts itself to quantitative data type for brevity. Quantitative variable Let xik denote the value that the kth quantitative variable takes n; k 1 p . The Minkowski metric defines a family of for the ith object i 1 dissimilarity measures, indexed by the parameter . Minkowski metric p wk xik di j x jk 1 1 (1) k 1 p are non-negative weights associated with the variables, allowing where wi k 1 standardization and weighting of the original variables. Values of of 1 and 2 give the two commonly used metrics of this family. City block p wk xik di j x jk (2) k 1 Euclidean distance p w2k xik di j x jk 2 1 2 (3) k 1 These measures can be standardized, for instance so that di j is bounded by 1. p k 1 , where k denotes the range of values taken by the kth variable. If wk p maxkp 1 k , which preserves the quantitative One could also consider wk comparison between objects. Also consider not the range, but (assuming it is relp maxni 1 xik or evant to consider the possible minimum value then take wk p n wk p maxk 1 maxi 1 xik . For example:- agent1 agent2 Sport Art Music 5 1 3 4 1 5 Without weighting this gives 1.29, with the range 0.816, with max value 0.086. Sometimes it is the relative magnitudes of the different variables that is of interest the behaviour across the variables rather than the absolute values. Put another way, the variables describing the object define a vector with p components and interest is in the comparison of the directions of the vectors. In the following metric the cosine of the angle between the vectors is used. Since values are between -1 and 1, the measure can 1 si j 2. be transformed to take values between 0 and 1 by defining si j Angular separation p k 1 xik x jk (4) si j p p 2 2 1 2 k 1 xik l 1 x jl For the previous example this metric gives a value for s of 0.0465. Defining Like-Minded Agents 87 Mixed variables For profiling of objects the variables will sometimes be of different types: for example a person can be described in terms of their gender (binary variable), their age (quantitative variable, their amount of interest in a subject (ordinal variable if sectioned quantitative variable is used) and their personality classification (nominal variable). A general measure is:General similarity coefficient p p si jk ; di j si j k 1 di jk (5) k 1 where si jk (and correspondingly di jk ) denotes the contribution to the measure of similarity provided by the kth variable. The values of si jk and di jk can take definitions as appropriate to the variable type. Selection (feature extraction) and standardization (normalization) Sometimes it is clear what variables should be used to describe objects. In our case, with profiling queries, documents, specifications and personal profiles, it is likely that variables have to be selected from many possibilities. Thus the process is not straightforward. The pattern recognition literature describes the appropriate specification of variables as feature extraction. It is tempting to include a large number of variables to avoid excluding anything relevant, but the addition of irrelevant variables can mask underlying structure. Whilst the choice of relevant variables is important, there is also the possibility (particularly here - multidimensional nature of profiles themselves) that there is more than one relevant classification based on different, but possible overlapping, sets of variables. Having determined appropriate variables, there is then the question of standardizing and/or differentially weighting them, followed by the construction of measures of similarity. One aspect to the standardization is that two variables can have very different variability across the dataset. It may or may not be desirable to retain this variability. Standardization may also be with respect to the data set under consideration or with respect to a population from which the samples are drawn. In the case of quantitative variables, standardization can be made by dividing by their standard deviation or by the range of values they take in the data set. The idea of standardization lies within the larger problem of the differential weighting of variables. 4 Choosing a Metric - Visual Exploration To assist the process of metric choice the use of specially constructed data sets in an exploration of the algorithm behaviours is proposed in signature exploration. Thus, by examining known data we gain a concrete idea of the behaviour of the various possible metrics. We have suggested a number of possible constructed data types[8]: generic(provided by the application to illustrate the behaviour of the particular algorithm); constructed(determined by the user to illustrate the behaviours in the data that are of interest to them, for evaluation purposes this represents an ad hoc classification); query (by visualization or sql-type, based on an unknown dataset, to examine clustering 88 Noy and Schroeder of metric in practice); landmark (to provide marker entities in the visualization); feedback (the means to enable the user to enter their assessed similarities and find or modify the appropriate metric). This paper limits itself to the first two, generic and constructed, to illustrate. 4.1 Using generic data sets Generic data sets are those considered to illustrate the behaviour of the visualization algorithms. Simple data sets do not always give an intuitive placement after such transformations. In this examination a small matrix of 7 agents were given a randomly assigned level of interest (of 1 to 10) in 7 topics. Agent1 9 3 4 6 5 5 5 Agent2 1 10 10 1 7 2 0 Agent3 4 1 6 8 0 5 7 Agent4 2 7 8 4 0 2 0 Agent5 3 6 4 7 1 10 6 Agent6 1 7 6 5 0 2 0 Agent7 8 1 7 1 2 5 9 Subsequently three other agents were added to illustrate (a) interests identical to agent1 but scaled, (b) agent1 with the same level of interest in each topic and three other agents as in (a), (c) two of the agents showing reverse behaviours of another two. These data sets were visualized with various distance measures (using the tool SpaceExplorer [11,13,12])and comments noted. The results illustrate the similarity in behaviour of the metrics, whilst indicating the differences obtained with the two basic types - Euclidean 3), City, Euclidean and Angular Separation. The measures used were Minkowski ( and Angular Separation (equations 1-4). These were followed by Principle Coordinates Analysis to find points in 2D-space that satisfied the distances. Note that the accuracy of such layouts for visualization is an issue, since it is often very low. In the case of PCoA, the eigenvalues can be examined - the sum of the values of the first three (for 3D layout) being above 70% of the sum of all the eigenvalues accounts for 70% of the variance in the data and is thus an encouraging indicator. As an illustration of this process, figure 1 shows the three shots of City, Euclidean and Angular separation with agents a,b and c having scaled interest distribution of agent1. 4.2 User-constructed data sets Here the user constructs data sets specific to their application, explicitly or implicitly creating a classification system with which to measure the performance of the metrics in clustering their interest feature(s). This may provide a distance matrix for comparison, or such a matrix may be obtained by an informal assessment. This could be followed by feedback analysis to obtain weightings of the feature list, but here the focus is on metric choice rather than modification. Defining Like-Minded Agents 89 Fig. 1. City (top left), Euclidean (top right) and Angular Separation (bottom) measures followed by layout using Principal Coordinates Analysis. Agents a, b and c have scaled interest distribution of agent1. In the angular separation, agents a,b,c and 1 are located in the same position. Step 1 - decide features of interest The first thing to consider is what features in the data one is interested in. We suppose that the aspects are: overlap of interest; intensity of interest; joint disinterest; similar pattern of interest (irrespective of subject). These elements provide a classification system with which we can construct a system to give numerical values to differences between a pair of agents’ interests. Then these differences can be used to give a comparison measure for the behaviours of the various metrics. Statistical measures indicate the closeness of the match. The metrics may not correspond to the classification, even approximately. It could be that it is useful to use the classification system as the similarity measure itself and dispense with the metrics. However, in general, we are looking for a similarity metric that is not just a simple query, but something more subtle, something that reflects the multidimensional nature of the profiling data available. This corresponds to the scope that lies between the two questions:- Are you interested in sport? and Are you like me? Also, if you are interested in sport, it may be valuable to know if you are a specialist or a generalist and in general terms what level your interest is on. Thus other similarity measures act as discriminators in this situation. Final choice of overall similarity measure may consist of additions of different similarity metrics (which may include results of specific queries) and can be arrived at in the manner of equation 5. 90 Noy and Schroeder Fig. 2. Interests level against category for agents 1 and 2 The use of visualizations of data for pairs of agents can assist in the specification of features of interest. Simple diagrams such as bar and pie charts are helpful in designing a measure with which to make an informal assessment of the similarity between two agents. Figures 2, 3 are illustrative of this process and identify the features interest level and interest intensity that are used in step 2. Fig. 3. Interests level difference against category for agents 1 and 2 (left) and intensity of overlap (right) Step 2 - create a measure for the features to generate test data sets Suppose that types of agent similarity are chosen to examine e.g.: overlaps of three or more interests of high intensity; large overlaps irrespective of intensity with high common disinterest. Data is created for a representative member and edge member of desired clusters so that representative pairs of data can be created (or groups if required) to examine the metric behaviours both visually and by comparing distance matrices. To illustrate, a data set was created to produce examples covering the range of possibilities of overlap extent and intensity with respect to a reference agent’s interests (as suggested by the visual Defining Like-Minded Agents 91 explorations of step 1). Then the metrics were examined to see how they clustered the 3 and intensity of overlap group of similarities with number of overlap subjects 2 3. One would expect the metrics to perform badly against this criterion, which is an example where a simple query would perform better (for instance, in the Yenta system, the matching between profiles is done simply on the basis of matching a single granule, which corresponds to a single interest, the metric is used in deriving the interest categories - if you simply want to exchange information on a subject that’s ok, however such aspects as level of expertise are relevant, and finding like-minds needs greater subtlety). For metric discrimination, the distance comparison should be made by also evaluating the test criteria for a number of other features (such as joint disinterest and large overlaps irrespective of intensity) and combining the similarities. Step 3 - evaluate visually and numerically The visual evaluation consists of visualizing the constructed data set and observing how well clustered the group of interest is. However, since the layout of such visualizations is an approximation (in order to satisfy the distances), and the observations not themselves measurements, evaluation by visualization is inexact. On the other hand, numerical evaluation, based on measuring differences between the estimated differences and the differences arrived at by the metric under consideration, is precise, but relies on the ability of the designer to define or estimate similarities between the data entities. For the example above this was done by awarding points according to number and intensity of topics of joint interest. Figure 4 shows PCoA layout with City, Euclidean and Angular Separation differences, the reference agent is circled and the agents that are in my group of interest (according to the criteria in step 2) are indicated. That there is little difference between City and Euclidean indicates that it would be adequate to use city where time complexity was an issue. The three outlines traced by the points in the City and Euclidean plots correspond closely to the classification system and the group of interest is well clustered in visual terms. The Angular Separation plot does not cluster so well, misplacing three agents. The layout of the angular separation distances is actually a screenshot of a 3D representation as the layout was particularly inaccurate and needed the extra dimension to improve it (the first two eigenvalues accounted for only 38% of the variance in the data and the first three for only 48%). The inaccuracy of this layout highlights the difficulty of using visualization to assess similarity. 5 The Use of Position Instead of Vector for Profile The pictures of information spaces as maps or terrains derived from multivariate data using self-organizing maps [16] provide us with a compelling image of the profile or topic space we are exploring. The metrics discussed above generate similar conceptual spaces when visualized. Yet this is a misleading image, since the data are high dimensional and it is impossible to represent their similarities accurately in 2 or 3D space (direct mapping methods for multivariate data, such as colour maps and parallel coordinate plots, are not included in this comment). Nevertheless, as an approximation and as a representation, an overview perhaps, of a large body of entities, it is being found useful (see e.g. [20]). Suppose we assume the validity of the layout and propose that 92 Noy and Schroeder Fig. 4. Constructed data set with (left to right) PCoA and City, Euclidean and Angular differences. the agent carries with them their xy (or xyz) coordinates and uses them as their profile. When meeting a fellow agent they can ask for the agent’s xy coordinates and compute the Euclidean distance to calculate their similarity. This would be more efficient than carrying a potentially long profile vector and enable them to use their profile without revealing details or requiring encryption. Two different ways of using this idea suggest themselves - calculating back at base and on the fly. 5.1 By base calculation The agents both have the calculations done at a base point and periodically return for updates. Here the error will be that of the layout itself and the agent would be able to have details of the mean error and variance supplied with its coordinates, so that it can take this into account. Figure 5 shows the layout after City distance and PCoA of the seven agents of randomly generated data from above. Thus, if Agent1 meets Agent2 they can compare coordinates, ((-12.30, -5.20),(23.27,-8.44)), to calculate the Euclidean distance to give them the distance they are apart in this map. 5.2 By calculation on the fly Here the agent calculates its position with respect to a number of reference vectors (either dynamically or at an earlier point in time) and then compares with another agent’s position calculated similarly. Using the seven agent random data again, the reference vectors are chosen to be agents 5,6 and 7 illustrated in the generic data section. Three reference agents are the minimum since only two will create two possible arrangements when agents 1 and 2 overlay their positions. Agents 1 and 2 separately calculate their City distances to the three reference vectors and subsequently lay out these distances with PCoA as shown in figure 6. They now have xy coordinates, but in order to compare them they must be scaled (the Euclidean distance between 5 and 6 is used here), centered (here Agent 5 is placed Defining Like-Minded Agents 93 Fig. 5. Illustration of base plot, the three reference agents (5,6 and 7) and the two of interest in this measurement (1 and 2) are circled. Fig. 6. Illustration of plots calculated individually by agents 1 (left) and 2 (right) with respect to the three reference agents (5,6 and 7) as circled and numbered. at 0,0) and finally rotated to bring the agents 5,6 and 7 into position. Now the coordinates of the agent’s position are in a form that they can use for comparisons. The results of the base calculation and on-the-fly calculation of the difference between agents 1 and 2 are given in the table below. (Since these are normalized with respect to the distance between agents 5 and 6, a value of 1 would indicate that they were the same distance away from each other as agents 5 and 6 are) original city dist base dist on-the-fly dist 1.77 1.57 1.64 8%err -4%err exact 6 Conclusions and Future Work Visual datamining seeks to increase the integration of visualization with specific datamining techniques. This paper presents two applications with this in mind. Appropriate clusterings of data are sought, whilst at the same time layouts are required to present overviews. The user needs understanding of the layout algorithm to appreciate the implications of the overview, the implication of arriving at different clusterings with different algorithms needs to be understood by those seeking valid cluster- 94 Noy and Schroeder ings and classifications. These two purposes concern the same process, but are subtly different in their focus. The first application described in this paper, illustrating the use of signature exploration in making the behaviour of the similarity metrics more concrete and assisting in similarity metric choice, is an example of the merging of understanding of overview and determining appropriate metric. Visualization of pairs of data helped in the creation of an ad hoc, user-specific, classification with which to assess the overview and thus also the metrics. An obvious next step is to use feedback to select and modify metrics and features and this is another part of the signature exploration process. Continuing work lies in further developing the interface for exploration, the data construction engine and in conducting usability tests. Visualizations sometimes suggest the idea of a topic or similarity space - looking at a 2D or 3D scatterplot the closeness of entities is intuitively understood as similarity. Where dimension reduction is involved, considerable approximation or abstraction is required. If this is a valid procedure (in the sense of the considerable error sometimes incurred), and such diagrams are widely used without warnings given, then the idea of using location as a form of privacy protection (the transformation is a one-way function) must also hold on some level. The simple example for using position as a profile demonstrated in this paper - a potentially most useful mechanism - is encouraging, now evaluation for many different data sets is required to test its robustness, in terms of whether the original profile is fully protected and the tolerance of approximation in locations of the entities. Evaluation of the position-as-profile concept points to one of our most pressing problems in visualization - how valid are our visualizations when dealing with complex data and involving approximation or abstraction? How can the level of approximation be indicated to the viewer? Correspondingly, how can a measure of confidence in the agent’s location in the interest space be given to the agent? The investigation of the position-as-profile idea is the same investigation as that of the validity of layout. Thus, we begin to think in terms of transfering our picture as a viewer to the agent, so that the two can become one - a kind of viewer/agent entity. The agent thus may be a software agent or a human agent. The question now becomes, how can the boundaries or parameters of the validity be described to the viewer/agent? How can they be encoded visually and in software terms? We interchange the viewer with agent and must express what the user sees (or finds useful) in a form that the software agent can work with. Via the agent paradigm we may thus be helped toward creating programs that can use graphical elements to mimic our visual thinking. 6.1 Acknowledgements This work is supported by the EPSRC and British Telecom (CASE studentship - award number 99803052). Defining Like-Minded Agents 95 References 1. S. K. Card, J. D. Mackinlay, and B. Shneiderman. Readings in Information Visualization: Using Vision To Think. Morgan Kaufmann, 1999. 2. C. Chen. Information Visualisation and Virtual Environments. Springer, 1999. 3. G. di Battista, P. Eades, R. Tamassia, and I. G.Tollis. Graph Drawing: Algorithms for the Visualization of Graphs. Prentice Hall, 1999. 4. L. Foner. Yenta: a multi-agent, referral-based matchmaking system. In The First International Conference on Autonomous Agents, Marina del Rey, California. ACM press, 1997. 5. L. N. Foner. Clustering and information sharing in an ecology of cooperating agents. In AAAI Spring Symposium ’95 on Information Gathering in Distributed, Heterogeneous Environments, Palo Alto, 1995. 6. J. Glasgow, N. Hari Narayanan, and B. Chandrasekaran. Diagrammatic Reasoning. AAAI Press / The MIT Press, 1995. 7. A. D. Gordon. Classification. Chapman and Hall / CRC, 2nd edition edition, 1999. 8. P. Noy and M. Schroeder. Introducing signature exploration: a means to aid the comprehension and choice of visualization algorithms. In ECML-PKDD01 Visual Data Mining Workshop, Freiburg, Germany, Sept 2001. 9. E. Ogston and S. Vassiliadis. Matchmaking among minimal agents without a facilitator. In Proceedings of Autonomous Agents2001, Montreal, Canada, 2001. ACM press. 10. P.Faratin, C.Sierra, and N.R.Jennings. Using similarity criteria to make negotiation tradeoffs. In Proc. of 4th Int. Conf. on Multi-Agent Systems ICMAS-2000, pages 119–126, Boston, USA, 2000. IEEE Computer Society. 11. M. Schroeder. Using singular value decomposition to visualise relations within multi-agent systems. In Proceedings of the third Conference on Autonomous Agents, Seattle, USA, 1999. ACM Press. 12. M. Schroeder, D. Gilbert, J. van Helden, and P. Noy. Approaches to visualisation in bioinformatics: from dendrograms to Space Explorer. Information Sciences, 139:19–57, 2001. 13. M. Schroeder and P. Noy. Multi-agent visualization based on multivariate data. In Proceedings of Autonomous Agents2001, Montreal, Canada, 2001. ACM press. 14. G. Somlo and A. Howe. Incremental clustering for profile maintenance in information gathering web agents. In Proceedings of Autonomous Agents2001, Montreal, Canada, 2001. ACM press. 15. R. Spence. Information Visualization. Addison-Wesley, 2001. 16. T.Kohonen. Self-organising maps. Springer-Verlag, 2nd edition edition, 1997. 17. D. Veit. Matchmaking algorithms for autonomous agent systems. Master’s thesis, Institute of Computer Science, University of Giessen, Germany, 1999. 18. D. Veit, J. Muller, M. Schneider, and B. Fiehn. Matchmaking for autonomous agents in electronic marketplaces. In Proceedings of Autonomous Agents2001, Montreal, Canada, 2001. ACM press. 19. A. Webb. Statistical Pattern Recognition. Arnold, 1999. 20. Websom. http://websom.hut.fi/websom/. 21. G. Weiss, editor. Multiagent Systems. MIT Press, 1999. Visual Data Mining of Clinical Databases: an Application to the Hemodialytic Treatment based on 3D Interactive Bar Charts Luca Chittaro 1 , Carlo Combi2 , Giampaolo Trapasso1 1 HCI Lab, Dept of Math and Computer Science, University of Udine, via delle Scie nze 206, 33100 Udine, Italy chittaro@dimi.uniud.it 2 Department of Computer Science, University of Verona, strada le Grazie 15, 37134 Verona, Italy combi@sci.univr.it Abstract. The capability of interactively mining clinical databases is an increasingly urgent need. This paper considers a relevant medical application (i.e., hemodialysis) and proposes a system for the visualization and vis ual data mining (VDM) of the collections of time -series acquired during hemodialytic treatments. Our proposal adopts bar charts as the basic visualization technique (because it is very familiar for clinicians) and augments them with several interactive features, exploiting a 3D space to significantly increase both the number of time -series that can be simultaneously analyzed in a convenient way and the number of values associated with each series. 1 Introduction The capability of interactively mining patient clinical information is an increasingly urgent need in the clinical domain, due to the continuous growth in the number of parameters that can be automatically acquired and in the size of the databases where they accumulate [5]. This is particularly critical for the success of medical research projects which generate massive databases of patient data. Some techniques for visual data mining (VDM) of multidimensional clinical databases are illustrated in [7]. They are mainly based on 3D versions of parallel coordinate plots. Graphical connections between points in adjacent planes are drawn in such a way that each patient's case is visually represented by a line connecting individual points referring to it. This allows for VDM of interesting patterns (e.g., a group of patients with the same profile results in parallel lines). 98 Chittaro et al. A different approach is presented by [10] and is based on tables displaying records of the clinical database and their attributes in highly compressed format such that they fit onto the screen. Users directly manipulate the table (e.g., performing zoom and filter operations) that dynamically rearranges itself. To compress the tables, the system relies on visualization criteria such as (i) neighboring cells with identical values are combined into a larger cell, or (ii) if there is no space to display a numeric value in its cell, the value is substituted by a small horizontal line whose position indicates relative size. This paper explores a third possibility, especially suited to clinical databases containing time-series data. Since, historically, bar charts are a widely adopted approach to display time-series and are a very familiar representation for clinicians, we chose them as the basis of our visual approach. Unfortunately, while a bar chart allows for an easy comparison among the data values for a single time-series, when the considered task requires to compare a collection of time-series (such as a monitored signal from the same patient in different sessions of the same clinical test or treatment), traditional bar charts (as other historical approaches) become unfeasible. Therefore, we augment bar charts exploiting a 3D space and adding several interactive features. A 3D space can significantly increase both the number of time-series that can be simultaneously analyzed in a convenient way and the number of values associated with each timeseries, but poses well-known problems such as occlusions, 3D navigation, difficulties in comparing heights, proper use of space, and the need for effective interaction techniques to aid the user in the analysis of large datasets (e.g., highlighting interesting patterns, checking trends,…). The limited capabilities of commercial tools that generate 3D bar charts have led well-known researchers (e.g. [9]) to classify these visualizations as “chartjunk 3D”. However, solutions to the problems of 3D bar charts are emerging from research: e.g., Cichlid offers temporal animation capabilities of 3D stacked bar charts [2], while ADVIZOR allows one to interactively link the 3D bar chart representation with related 2D representations, compare heights with a “water level” plane (perpendicular to the bars) and use filtering tools [6]. Alternative approaches to time-series visualization have been recently proposed, e.g. drawing the timeline along spiral structures [12] is reported to allow for an easier detection of cyclic phenomena. However, we preferred to adopt bar charts, because they were familiar to clinicians. Moreover, we did not have a focus on a specific pattern such as cycles. In the following, we first introduce the real-world clinical context we are working in and motivate the need for VDM in that context. Then, we illustrate the system we have built and its main features. Finally, we show some examples of how our system is being applied to the clinical context. 2 Hemodialysis and Visual Data Mining Hemodialysis is the widely used treatment for patients with acute or chronic endstage renal failure. During an hemodialysis session, the blood passes through an extra- Visual Data Mining of Clinical Databases 99 corporeal circuit where metabolites (e.g., urea) are eliminated, the acid-base equilibrium is re-established, and water in excess is removed. In general, hemodialysis patients are treated 3 times a week and each session lasts about 4 hours. Hemodialysis treatment is very costly and extremely demanding both from an organizational viewpoint [8] and from the point of view of the patient’s quality-of-life. A medium-size hemodialysis center can manage up to 60 patients per day, i.e. more than 19000 hemodialytic sessions per year. Unfortunately, the number of patients that need hemodialysis is constantly increasing [12]. In this context, it is very important to be able to evaluate the quality of (i) each single hemodialysis session, (ii) all the sessions concerning the same patient, and (iii) sets of sessions concerning a specific hemodialyzer device or a specific day, for the early detection of problems in the quality of the hemodialytic treatment. Modern hemodialyzers are able to acquire up to 50 different parameters from the patient (e.g., heart rate, blood pressure, weight loss due to lost liquids,…) and from the process (e.g., pressures in the extra-corporeal circuit, incoming blood flow,…), with a configurable sampling time whose lower bound is 1 sec. As an average example, considering only 25 parameters with a sampling time of 30 seconds, 12000 values (4*120*25) are collected in each session, and a medium-sized center collects more than 228 millions of values per year (considering 19000 provided treatments). While the daily accumulation of huge amounts of data prompts the need for suitable techniques to detect and understand relevant patterns, hemodialysis software is more concerned with acquiring and storing data, rather than visualizing and analyzing it. Data mining applications can thus play a crucial role in this context. More specifically, visual data mining applications are of particular interest for three main reasons. First, clinicians’ abilities in recognizing interesting patterns are used suboptimally or not used at all in the current context. Visual mining of hemodialytic data would allow clinicians to take decisions affecting different important aspects such as therapy (personalizing the individual treatment of specific patients), management (assessing and improving the quality of care delivered by the whole hemodialysis centre), medical research (discovering relations and testing hypothesis in nephrology research). Second, since data mining on the considered database is (at least, at initial stages) intrinsically vague for clinicians, the adoption of VDM techniques can be more promising than fully automatic techniques, because it supports clinicians in discovering structures and finding patterns by freely exploring the datasets as they see fit. Third, the clinical context is characterized by a need for user interfaces that require minimal technical sophistication and expertise to the users, while supporting a wide variety of information intensive tasks. A proper exploitation of visual aspects and interactive techniques can greatly increase the ease of use of the provided solutions. In summary, a clinical VDM system has to achieve two possibly conflicting goals: (i) offering powerful data analysis capabilities, while (ii) minimizing the number of concepts and functions to be learned by clinicians. In the following, we illustrate how our system attempts to achieve these two goals. 100 Chittaro et al. 3 The Proposed Approach The system we have built, called IPBC (Interactive Parallel Bar Charts) connects to the hemodialysis clinical database, produces a visualization that replaces tens of separate screens used in traditional hemodialysis systems, and extends them with a set of interactive tools that will be described in detail in this section. Each hemodialysis session returns a time-series for each recorded clinical parameter. In IPBC, we visually represent each time-series in a bar chart format where the X axis is associated with time and the Y axis with the value (height of a bar) of the series at that time. Then, we layout the obtained bar charts side by side, using an additional axis to identify the single time-series, and we draw them in a 3D space, using an orthogonal view. It must be noted that also the additional axis has typically a temporal dimension, e.g. it is important to order the series by date of the hemodialysis session to analyze the evolution of a patient. An example is shown in Fig. 1, that illustrates a vis ualization of 50 time-series of 50 values each, resulting in a total of 2500 values (the axis on the right is the time axis for single sessions, while the axis on the left identifies the different time-series, ordered by date). Hereinafter, we refer to this representation as a parallel bar chart. Fig. 1. A Parallel Bar Chart. 3.1. The RoundToolbar widget In designing how the different interactive functions of IPBC should be invoked by the user, we wanted to face two different problems: • First, one well-known limitation of many 3D visualizations is the possible waste of screen space towards the corners of the screen; Visual Data Mining of Clinical Databases 101 • Second, the traditional menu bar approach would require long mouse movements from the visualization to the menu bar and vice versa. To this purpose, we designed a specific round-shaped pop-up menu (see Fig. 2), called RoundToolbar (RT), that appears where the user clicks with the right mouse button. The RT can be easily positioned in the unused screen corners, thus allowing a better usage of the screen space (e.g., see Fig. 1) and a reduction of the distance between the visualization and the menu. Moreover, to further improve selection time of functions with respect to a traditional menu, the organization of modes in the toolbar is inspired by Pie Menus [3]: in particular, the main modes are on the perimeter of the RT, and when a mode is selected, the center of the RT contains the corresponding tools (which are immediately reachable by the user, who can also quickly switch back from the tools to a different mode). Fig. 2. Viewpoint mode. Fig. 3. Viewpoint movements: A) Low; B) Rotate; C) Far. 3.2. Changing Viewpoint It is well-known that free navigation in a 3D space is difficult for the average user, 102 Chittaro et al. because (s)he has to control 6 different degrees of freedom and can follow any possible trajectory. To make 3D navigation easier, when the Viewpoint mode is selected in the RT (as in Fig. 2), the proposed controls for viewpoint movement (Rotate, HighLow and Near-Far) cause movement along limited pre-defined trajectories which can be useful to examine the visualization: in particular, Fig. 3 shows how viewpoint movement is constrained. The remaining Vertical scale control in the Viewpoint mode is used to scale the bars on the Y axis. Vertical scaling has been included in the Viewpoint mode, because it has been observed that when users scaled the bars, they typically changed the viewpoint as the following operation. Fig. 4. Dynamic Query mode. 3.3. Dynamic Queries IPBC uses color to classify time-series values into different ranges. In particular, at the beginning of a session, the user can define units of measure and her general range of interest for the values, specifying its lowest and highest value. These will be taken as the lower and upper bounds for an IPBC dynamic query control in the RT (as shown in Fig. 4) that allows the user to interactively partition the specified range into subranges of interest. Different colors are associated to the subranges and when the user moves the slider elements, colors of the affected bars in the IPBC change in realtime. Possible bars with values outside the specified general range of interest are highlighted with a proper single color. For example, Fig. 1 shows a partition that includes the three subranges corresponding to the colors shown by the slider in Fig. 4, and also some bars which are outside the user’s predefined range. The color coding scheme can be personalized by the user with the Colors mode in the RT. The dynamic query control allows the user to: • move the two slider elements independently (to change the relative size of adja- Visual Data Mining of Clinical Databases • 103 cent subranges). For example, in Fig. 4, one has been set to 130 mmHg and the other to 180 mmHg. This can be done both by dragging the edges or (more easily) the tooltips which indicate the precise value. Plus and minus signs in the tooltips also allow for a fine tuning of the value. Move the two slider elements together by clicking and dragging the area between the two bounds. This can be particularly useful (especially when the other areas are associated to the same color), because it will result in a “spotlight” effect on the vis ualization: as we move the area, our attention is immediately focused on its corresponding set of bars, highlighted in the visualization. Fig. 5. Tide mode. 3.4. Comparing data with (time-varying) thresholds A frequent need in VDM is to quickly perceive how many and which values are below or above a given threshold. This can be easily done with the previously described dynamic queries when the threshold is constant. However, the required threshold is often time-varying, e.g. one can be interested in knowing how many and which values are not consistent with an increasing or decreasing trend. For this need, IPBC offers a mode based on a tide metaphor. As it can be seen in Fig. 5, the Tide mode adds a semitransparent solid to the visualization: the solid metaphorically represents a mass of water that floods the bar chart, highlighting those bars which are above the level of water. The slope of the top side of the solid can be set by moving two tooltips shown in the RT (that specify the initial and final values for the solid height), thus determin- 104 Chittaro et al. ing the desired linearly increasing or decreasing trend. The height of the solid can be also changed without affecting the slope by clicking and dragging the blue area in the RT. An opaque/transparent control allows the user to choose how much the solid should hide what is below the threshold. When the Tide mode is activated, all the bars in the user’s range of interest are turned to a single color to allow the user to more easily perceive which bars are above or below the threshold; if multiple colors were maintained, the task would be more difficult, also because the chromatic interaction between the semitransparent surface and the parts of bars inside it adds new colors to the visualization. The Tide mode can be also used to help compare sizes of bars by selecting a zero slope and changing the height of the solid (in this special case, Tide becomes analogous to the “water level” function of other visualization systems). Fig. 5 illustrates this latter case, while Fig. 9 shows a positive slope case. Implementing a non-linear Tide would be relatively straightforward (only linear trends are anyway used by clinicians in the considered hemodialysis domain). Fig. 6. Matrix Visualization. 3.5. Managing Occlusions As any 3D visualization, IPBC can suffer from occlusion problems. To face them, the approach offers two possible solutions. First, by clicking on the 2D/3D label on the RT, the user can transform the parallel bar chart into a matrix format and vice versa. For example, Fig. 6 shows the same data as Fig. 1 in the matrix format. The transformation is simply obtained by automatically Visual Data Mining of Clinical Databases 105 moving the viewpoint over the 3D visualization (and taking it back to the previous position when the user deselects the matrix format). This can solve any occlusion problem (and the dynamic query control can still be used to affect the color of the matrix cells), but the information given by the height of the bars is lost. Transitions to matrix format and back are animated to avoid disorienting the user and allow her to keep her attention on the part of the visualization (s)he was focusing on. Second, by directly clicking on any time-series in the 3D visualization, only the timeseries which can possibly occlude the chosen one collapse into a flat representation analogous to the matrix one, as illustrated in Fig. 7. Fig. 7. Removing occlusions. 3.6. Pattern Matching When the user notices an interesting sequence of values in one of the time-series, IPBC offers her the opportunity to automatically search for and highlight occurrences of a similar pattern in all the visualization (a detailed example will be described in Section 4.4). The user selects her desired sequence of values in a time-series by simply dragging the mouse over it, then (s)he can specify how much precise the search should be, by indicating two tolerance values in the RT: (i) how much a single value can differ in percentage from the corresponding one in the given pattern, (ii) the maximum number (possibly zero) of values in a pattern that can violate the given percentage. 106 Chittaro et al. 3.7. Mining Multidimensional Data If multiple variables are associated to the considered time-series, IPBC can organize the screen into multiple windows, each one displaying a parallel bar chart for one of the variables. The visualizations in all the windows are linked together, e.g. if one selects a single time-series in one of the windows (or a specific value in a time-series), that time-series (or the corresponding value) is automatically highlighted in every other window. This (as some other features of IPBC) will be shown in more detail in the next section. 4. Mining Hemodialytic data In the following, we will show how IPBC can be used during real clinical tasks, to help physicians evaluating the quality of the hemodialytic treatments given to single patients, on the basis of the clinical parameters acquired during the sessions. Each hemodialysis session returns a time-series for each parameter; different time-series are displayed side by side in the parallel bar chart according to date (in this case, the axis on the left chronologically orders the sessions). The following examples are ordered according to the complexity of the related task: in particular, the first two tasks are relatively simple and are taken from the daily activity of clinicians, while the last two tasks are more complex and are performed by clinicians only in specific occasions (in the two considered examples, they are related to a detailed evaluation of the quality of care provided by nurses). 4.1. Mining patient signs data A first task consists in analyzing patient signs, as the systolic and diastolic blood pressures and the heart rate; indeed, these parameters are important both for the health status of the patient and for the management of device settings during the hemodialytic session. Let us consider, for example, the task of analyzing all the systolic pressures of a given patient: Fig. 8 shows a parallel bar chart (containing more than 5000 bars), representing the systolic pressure measurements (about 50 per session) during more than 100 hemodialytic sessions. In this figure, we can observe that the presence of out-ofscale values, usually related to measurement errors (e.g., the patient was mo ving; the measurement device was not properly operating), has been highlighted by specifying a proper range of interest (that highlights them in a suitable color) and hiding their height. In the specific situation represented in the figure, the presence of several outof-scale values at the beginning of each session is due to the fact that nurses activate the measurement of patient’s blood pressure with some delay with respect to the beginning of the session. In the figure, the user is focusing on a specific session, avoiding occlusion problems (as described in Section 3.5). At the same time, with a dynamic query, (s)he is able Visual Data Mining of Clinical Databases 107 to distinguish low, normal, and high blood pressures. In this case, the clinician can observe that the systolic pressure in the chosen session, after a period of low values (yellow bars), was in the range of normal values (orange bars). While the values for the chosen session correspond to a normal state, it is easy to observe that several sessions among those in the more recent half part of the collection contain several high values (in red) for the systolic blood pressure. Thus, the clinician can conclude that in those sessions the patient had some hypertension, i.e. a clinically undesired situation. Fig. 8. Analyzing systolic blood pressures. 4.2. Mining bl ood volume data Another task is related to observing the percentage of reduction of the blood volume during hemodialysis, mainly due to the removal of the water in excess. This reduction is sometimes slowed down to avoid situations in which the patient has too low blood pressures. In this case, VDM can benefit from the usage of the Tide mode. Fig. 9 shows an IPBC with more than 9000 bars, representing 36 hemodialytic sessions, containing about 250 values each. In this case, being the percentage of reduction of the blood volume increasing during a session, Tide allows the physician to distinguish those (parts of) sessions characterized by a percentage of reduction above or below the desired trend. In the figure, for example, the selected session has a first part emerging from the tide, while the last part is below. At the same time, it is possible to observe that one of the last sessions has the percentage of reduction above the tide during almost the entire session. The clinician can thus easily identify those (parts of) sessions with a satisfying reduction of the blood volume as the emerging (parts of) 108 Chittaro et al. sessions. Fig. 9. Visualizing the time-varying reduction of the blood volume in the Tide mode. 4.3. Mining related clinical parameters The next task we consider is related to the analysis of three related parameters: the systolic and diastolic blood pressures (measured on the patient) and the blood flow (QB) entering the hemodialyzer. QB is initially set by the hemodyalizer, but it can be manually set (reduced) by nurses when the patient’s blood pressures are considered too low by the medical staff. It is thus interesting to visually relate QB and blood pressures, to check whether suboptimal QBs are related to low pressures. Otherwise, suboptimal values of QB would be due to human errors during the manual setting of the hemodialyzer. Fig. 10 shows the coordinated visualization of three clinical parameters for the same patient: the diastolic blood pressure (small window in the upper left part), the systolic blood pressure (small window in the lower left part), and QB (right window). The user can freely organize the visualization, switching the different charts from the smaller to the larger windows (by clicking on the arrow in the upper right part of the smaller windows). In the figure, the clinician is focusing on a session where the QB was below the prescribed value during the first two hours of hemodialysis (yellow color for QB) and (s)he has selected a specific value (the system highlights that value and the corresponding values in the other windows with black arrows). It is easy to notice that the suboptimal QB was related to low blood pressures (yellow bars in the corresponding time-series in the two small windows); then, QB was set to the correct value by nurses (see black arrow in the right window) only after blood pressures reached normal values (orange color in the corresponding charts). In this case, the Visual Data Mining of Clinical Databases 109 physician can conclude that the suboptimal QB has been correctly set by nurses because of the patient’s hypotension. Fig. 10. Coordinated analysis of blood pressures and incoming blood flow. 4.4. Mining for similar patterns Finally, let us consider a task concerning the analysis of QB. As previously mentioned, the value of QB can be manually set by nurses and it may happen that this value is below the optimal one, due to hypotensive episodes. Fig. 11 shows a visualization where the clinician noticed a change of QB from a lower value to the correct one in a session: this means that, after a period of suboptimal treatment, the proper setting had been entered. Therefore, the clinician asks IPBC to identify QB patterns similar to the one (s)he noticed, by indicating it with the mouse, and setting the tolerance parameters (see Section 3.6). Fig. 11 shows the selected pattern (see the area near the black arrow) and the similar patterns automatically found by IPBC (two are in the lower right part of the figure, one in the upper left part): these patterns are identified by a line of a suitable color, which highlights the contours of the first and last bar of the pattern and intersects the inner bars. To avoid possible occlusion problems in visually detecting the patterns, the physician can move the viewpoint or switch to the matrix representation, where each pattern can be easily observed. 110 Chittaro et al. Fig. 11. Automatic Pattern Matching. 5. Conclusions and Future Work In this paper, we described the main features of IPBC (Interactive Parallel Bar Charts), a VDM system devoted to interactively analyze collections of time-series, and showed its application to a real clinical database of hemodialytic data. We are currently carrying out a field evaluation of IPBC with the clinical staff of the hemodialysis center at the Hospital of Mede, PV, Italy. One of the major advantages of IPBC that is emerging is that the visualization and its interactive features are very quickly learned and remembered by clinicians, the major disadvantage is that usage of screen space becomes difficult if a clinician tries to relate more than 3 collections of time-series simultaneously (Section 4.3 dealt with the analysis of 3 collections). This early feedback received from the field evaluation is helping us in identifying new research directions. Besides facing the problem of analyzing more than 3 collections in a convenient way, we aim to face another problem (that is considered very relevant by clinicians), i.e. dealing with time-series at different abstraction levels, allowing for both a fine exploration of time-series (e.g., to detect specific unusual values) and their coarse exploration (to focus on more abstract derived information). In both cases, we are working at the integration of parallel bar charts with other visualizations that can provide a synthetic view of data (e.g., the medical literature is proposing some comp utation methods to derive some quality indexes of the hemodialytic session from the time-series of that session). In particular, we are experimenting with Parallel Coordinate Plots, e.g. a trajectory in a plot could connect the quality indexes (typically, 5-7 values) of a session, and this high-level perspective would be linked to the much Visual Data Mining of Clinical Databases 111 of a session, and this high-level perspective would be linked to the much more detailed perspective of the parallel bar chart. Acknowledgements This work is partially supported by a MURST COFIN 2000 project (“Analysis, Information Visualization, and Visual Query in Databases for Clinical Monitoring”). References 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. 11. 12. 13. Ahlberg, C., Williamson, C., Shneiderman B.: Dynamic queries for information exploration: An implementation and evaluation. Proc. of the CHI ‘92 Co nference on Human Factors in Computing Systems, ACM Press, New York (1992) 619-626 Brown, J.A., McGregor, A.J., Braun H-W.: Network Performance Visualization: Insight Through Animation. Proc. of PAM2000: Passive and Active Measurement Workshop, Hamilton, New Zealand (2000) 33-41 Callahan, J., Hopkins, D., Weiser, M., Shneiderman, B.: An empirical comparison of pie vs. linear menus. Proc. of the CHI ‘88 Conference on Human Factors in Computing Systems, ACM Press, New York (1998) 95-100 Chittaro L. (ed.), Special issue on Information Visualization in Medicine, Artificial Intelligence in Medicine Journal, 22(2) (2001) Chittaro L.: Information Visualization and its Application to Medicine, in [4] 81-88 Eick, S. G.: Visualizing Multi-Dimensional Data. ACM SIGGRAPH Computer Graphics, 34(1) (2000) 61-67 Falkman, G.: Information Visualization in Clinical Odontology: Multidimensional Analysis and Interactive Data Exploration, in [4] 133-158 McFarlane, P.A., Mendelssohn, D.C.: A call to arms: economic barriers to optimal hemodialysis care. Perit Dial Int 20 (2000) 7-12. Shneiderman, B.: 3D or Not 3D: When and Why Does it Work?, invited talk at Web3D: 7th Internatio nal Conference on 3D Web Technology, Tempe, AZ (2002) Spenke, M.: Visualization and Interactive Analysis of Blood Parameters with InfoZoom, in [4] 159-172 Tufte, E.R.: The Visual Display of Quantitative Informatio n, Graphics Press (1982) USRDS, The United States Renal data system, http://www.usrds.org Weber, M., Alexa, M., Mueller, W.: Visualizing Time-Series on Spirals. Proc. of the IEEE InfoVis Symposium, IEEE Press, Los Alamitos, CA (2001) Sonification of time dependent data Monique Noirhomme-Fraiture1, Olivier Schöller1, Christophe Demoulin1, Simeon Simoff2 1 2 University of Namur, Institut d’Informatique, 5000 Namur, Belgium mno@info.fundp.ac.be University of Technology Sydney, Faculty of Information Technology, NSW2007 Sydney, Australia simeon@it.uts.edu.au Abstract. This paper presents the results of experiments with sonification of 2D and 3D time dependent data. A number of sonification means for these experiments have been implemented. An Internet Web site was created where sound sequences were presented and could be evaluated by the participants in the experiment. All participants that performed the tests also needed to fill an evaluation questionnaire. The purpose of the experimentation was to determine how the sonification of two and three-dimensional graphs can support or be an alternative to visually displayed graphs. The paper concludes discussion of the results and the issues related with the experiments. 1 Introduction Visual data mining is a part of the KDD process [1], which places an emphasis on visualisation techniques and human cognition to identify patterns in a data set. [1] identified three different scenarios for visual data mining, two of which are connected actually with the visualisation of final or intermediate results and one operates directly with visual representation of the data. The design of data visualisation techniques, in broad sense, is the formal definition of the rules for translation of data into graphics. Generally, the term ‘information visualisation’ has been related to the visualisation of large volumes of abstract data. The basic assumption is that large and normally incomprehensible amounts of data can be reduced to a form that can be understood and interpreted by a human through the use of visualisation techniques. The process of finding the appropriate visualisation is not a trivial one. A number of works offer some results that can be applied as guiding heuristics. For example, [2] defined the Proximity Compatibility Principles (PCP) for various visualization methods in terms of tasks, data and displays - if a task requires the integration of multiple data variables, they should be bundled in proximity in an integrated display. Based on this principle authors have concluded that 3D graphs do not have an advantage over 2D graphs for scientific visualisation (which may not necessarily hold for visual data mining). 114 Noirhomme-Fraiture et al. Visual data mining relies heavily on human visual processing channel and utilises human cognition overall. The visual data mining cycles are shown in Fig. 1. In most systems, visualisation is used to represent the output of conventional data mining algorithms (the path shown in Fig. 1a). Fig. 2 shows an example of visualisation of the output of an association rule mining algorithm. In this case, visualisation assists to comprehend the output of the data mining algorithms. Fig. 1b shows the visual data mining cycle when visualisation is applied to the original or pre-processed data. In this case, the discovery of the patterns and dependencies is left to the capacity of the human visual reasoning system. The success of the exercise depends on the metaphor selected to visualise the input data [3]. Knowledge Timelines Timelines Knowledge 0 100 200 300 400 500 600 700 800 U tterances 0 100 200 300 400 500 600 700 800 Utterances Visualisation of the output Visualisation of the output output output Data Mining Algorithms Data Mining Algorithms Visualisation of the input input input Data Data a. b. Fig. 1. Visualisation and visual data mining Although human visual processing system remains a powerful ‘tool’ that can be used in data mining, there are other perceptual channels that seem to be underused. Our capability to distinguish harmonies in audio sequences (not necessarily musical ones) is one possibility to complement the visual channel. Such approach can be summarised as ‘What You Hear Is What You See’. The idea of combining the visual and audio channels is illustrated in Fig. 3. The conversion of data into a sound signal is known as sonification. Similar to the application of visualisation techniques in Fig. 1b, sonification can be used both for representing the input and/or the output of the data mining algorithms. Sonification of Time Dependent Data 115 Fig. 2. Example of visualisation of the output of an association rule miner. Timelines Knowledge 0 100 200 300 400 500 600 700 800 Uttera nces Visualisation of the output output Sonification of the output output Data Mining Algorithms Sound Analysis Algorithms Visualisation of the input Sonification of the input input Data Fig. 3. Combining visual data mining and sonification In visual data mining, sonification should be synchronised with the visualisation technique. Further, in this paper we discuss the issues connected with designing such data mining techniques and present an example of a practical implementation of combined technique. We briefly discuss the characteristics of the sound that are suitable for such approach, the actual sonification process, the design of the overall combined technique and the results of the experiments conducted with proposed technique. 116 Noirhomme-Fraiture et al. 2 Characteristics of sound for time dependent data representation Several researchers have investigated the use of sound as means for data representation [4-11]. In this context, the important feature of the sound is that it has a sequential nature, having particular duration and evolving as a function of time. A sound sequence has to be heard in a given order, i.e. it is not possible to hear the end before the beginning1. Similarly, a time series depends on time and have the same sequential characteristics. Consequently sound provides good means to represent time series. 3 Sonification The easiest way to transform time dependent data into sound is to map the data to frequencies by using linear as well as chromatic scale mappings. We call this process a pitch-based mapping. We compute the minimum and maximum data values from the chosen series and map this data interval into a frequency range, chosen in advance. Each value of the series is then mapped into a frequency. To avoid too large, non-realistic intervals, we first discard outliers (see below). Another pre-treatment is the smoothing of the series. In fact, if we map all the points of a series into a sound, we will hear rather inconsistent sounds. A first treatment consists in smoothing the series by a standard mean, for example, by moving average method. After that, we map the smoothed curve into pitch. Beat drums can be used to enhance the shape of the curve (see below). 3.1 Detection of outliers To detect statistically the values of the outliers, a confidence interval is computed at each time t., based on the normal distribution. Once a data value is detected outside the confidence interval, the corresponding time value is stored and sonified at the experiment phase. 3.2 Beat drums mapping The rhythm of a beat drum increases with respect to the rate of growth of the curve (i.e. the first derivative). 1 Images and drawings do not have such constraint. Strictly speaking digital sound recording can provide access to an arbitrary section of the sound fragment, even reproduce the sound in reverse order, which is beyond the scope of this paper. Sonification of Time Dependent Data 117 3.3 Stereo panning Variation of the stereo acoustics is introduced, for example, an increase of the volume of the right speaker and decrease of the volume of the left speaker. 3.4 3D Curve When the time series is given at each time, not by a single value but by a function of values, we decide to “hear” at each discrete time the function. We can also choose to cut the surface at a certain level and to hear “continuously” the obtained curve as a function of time. We call these transformations respectively horizontal and vertical travelling. An example of a 3D data surface for sonification is shown in Fig. 4. Fig. 4. Example of a surface that can be sonified 4. Prototype implementation The prototype has been implemented in Java programming language, using the MIDI package of the Java Sound API2 [12]. The MIDI sequence is constructed before the actual playback. When the designer starts the sonification, the whole sequence is computed. Then computed sequence is sent to the MIDI sequencer for playback. 5. Experimentation The purpose of the experimentation is to determine how the sonification of two and three-dimensional graphs can complement or be an alternative to visually displayed 2 API – Application Programming Interface 118 Noirhomme-Fraiture et al. graphs. An Internet Web site has been created, where sound sequences are presented and can be evaluated by the visitors. The site contains questionnaire that has to be filled in by visitors performing the test. The structure of the questionnaire is the following one: A. Identification of the user: name, age, gender, title/position, e-mail address. These data are used to identify the subject and to validate the answer. B. Ability: field of activity, musical experience (instrument played, practicing period), self-evaluation of musical level (from ‘no experience’ to ‘expert level’). C. 2D evaluation: 2D evaluation is divided into three subtasks: Part C.a: Explanation about the four sonification techniques used: pitch based only, beat drums, stereo and extreme values detection. Each of them is briefly described and at least one example is given. Part C.b: Application test: four sequences are presented to the user. Each time precise questions are asked: Question 1: Annual sheep population in England and Wales between 1867 and 1939 (see Fig. 5). − Were there more sheep in 1867 then in 1939? − In your opinion, when (which year) did the sheep population reach the minimum? Fig. 5. Annual sheep population in England and Wales between 1867 and 1939 This question aims to evaluate if subject can perceive a global trend in the series and to understand if the relation with the time scale is done. For each sequence, beat drums and stereo mapping are added to enhance the pitch-based sonification. Question 2 aims to identify whether extreme values are detected. Question 3 aims to identify whether seasonal trend can be detected. Question 4 is focused on trend Sonification of Time Dependent Data 119 identification. For each question, the subject must specify the number of times he listened to the sonified data before answering the question. Part C.c: Subject preferences: four other questions aim to evaluate subject preference: 1. Choice of instrument for pitch mapping. The user is asked to hear the sonification of the same data, but with different MIDI instruments for the pitch mapping. The user is asked to grade on the scale 0 to 10 the different instruments (acoustic grand, steel string guitar, violin, synthstrings 2, pan flute). The instruments proposed are very different and belong to a specific MIDI group such as piano, guitar or to string group. 2. Choice of instrument for drums mapping. Different instruments for the beat drums mapping are presented and have to be marked (Celesta, Slap Bass 1, Timpani, Tinkle Bell, Woodstock). 3. Choice of sonification technique. The subject has to give his opinion on the different mapping techniques: pitch, beat drums, stereo mapping, extreme values detection and on the sonification in general. The four levels are proposed: useless, sometimes useful, always useful and essential. 4. Open question: “Please tell us what you think about our project, our applications, this web page or anything else that comes to mind”. D. 3D Evaluation. The same schema is used for 3D curves: explanation about the sonification method, the application test (3 questions), subject preferences. The sonification method is based on a cutting of the initial surface following some direction: vertical (see Fig. 6), horizontal (see Fig. 7) or diagonal. Each 2D line obtained by the cutting process can be heard, similar to the 2D case. The different lines, proposed one after the other, are separated by a specific sound. Beat drums can still be added to pitch mapping. 120 Noirhomme-Fraiture et al. Fig. 6. Data for vertical travelling Fig. 7. Data for horizontal travelling 6 Results Below we present the results of the experiments. 6.1 The sample 23 visitors answered the questions for the case of a 2D visualisation and 18 visitors for the case of 3D visualisation. A large part of the sample (9) includes people working in the computer science area, who have limited musical experience or no experience at all. To see if people with musical experience get a better score, we have compared the average score in both groups. The influence of the musical and computer science background on the results is presented in Table 1 and Table 2, respectively. The score obtained for each question is the number of good answers, normalised on a Sonification of Time Dependent Data 121 score out of 100. The average score is the mean of the scores for the different questions. Table 1. The influence of musical background on results Musical background No experience 2D sonification 3D sonification Av. score/100 Av. score/100 76 66 N 9 14 23 40 42 The same comparison was done for computer scientist. Table 2. The influence of computer science background on results Computer science Others 2D sonification 3D sonification Av. score/100 Av. score/100 72 69 N 11 12 23 41 40 Having a musical or a computer science background gives a minor advantage in using sonification of 2D curves. The differences are not significant. We do not observe any difference for the 3D sonification case. The way 3D case has been implemented is rather complex and uses a good spatial representation. Musical or computer experience has a minor influence on the result in this case. 6.2 Results in 2D The following questions were included in the questionnaire, targeting 2D sonification: Qu1 Annual sheep population in England and Wales between 1867 and 1939 1.1 Were there more sheep in 1867 than in 1939? 1.2 About which year did the sheep population reach the minimum? Qu2 Daily morning temperature of an adult woman during two months 2.1 Did she have fever during the period? 2.2 If yes, for how long did she have the fever? Qu3 Monthly electricity production in Australia between January 1956 and August 1995 3.1 Is the electricity production in Australia lower in 1956 than in 1995? 3.2 How would you categorise the evolution of electricity production in Australia: as linear or as exponential? 122 Noirhomme-Fraiture et al. 3.3 Is the evolution of electricity production in Australia characterised by seasonal trend? Qu4 Monthly Minneapolis public drunkenness intakes between January 1966 and July 1978 (151 months) 4.1 Were there more intakes in 1966 than in 1978 ? 4.2 Is the evolution of public drunkenness intakes linear? The results are summarised in Table 3. Table 3. Summary of the results for 2D Correct Qu1 Qu2 Qu3 Qu4 1.1 1.2 2.1 2.2 3.1 3.2 3.3 4.1 4.2 17 17 23 13 22 12 16 20 19 Wrong 5 6 0 10 1 11 6 2 3 No idea 1 0 0 0 0 1 1 1 6.3 Results in 3D The following questions were included in the questionnaire, targeting 2D sonification: Qu1 A 3D graph containing 2 bumps has been sonified. The selected mapping is the vertical travelling and the sonification starts from the bottom right corner. - If the grid below (3 x 3) represents the graph, where are these 2 bumps located? - Do they have the same height? Qu2 Same kind of questions with respect to horizontal travelling. Qu3 Same kind of questions with respect to diagonal travelling. The results are summarised in Table 4. Sonification of Time Dependent Data 123 Table 4. Summary of the results for 3D Correct Qu1 Qu2 Qu3 1.1 Trend 1.2 Value 2.1 Outlyers 2.2 3.1 General trend 3.2 2 correct 4 1 correct 6 11 4 6 12 1 11 10 Wrong No idea 8 - 4 8 33 6 6 3 0 0 2 7 Discussion There are some issues related to the design of the experiments that could have influenced the outcome of the experimentation: − In a graphical representation, if you want to identify a particular point, you need to find the information concerning that point on each axis. In the experiment, we provided little information about the scale (for example, see Fig. 6 and Fig. 7). The wording gives the limits for the time period. The lack of scaling information could have caused some difficulty in identifying particular points or sub-periods. − The outcomes in the case of sonification of a 3D graph are worse than in the case of 2D. It is necessary to take in account that the sonification of a 3D graphical representation is more difficult then the sonification of a 2D graphical representation. A possible reason could be that the sonification technique is based on visual representation and does not use sound properties, but surface properties, as seen in a 3 axis referential. An important issue becomes the correspondence between visual and audio representations of the data. Consistent representations should provide audio representations that allow transitions from 3D to 2D projections in terms of corresponding sound representations. 8 Conclusions Overall, the results of the experimentation on sonification of time dependent data leave optimism for further investigation of sound as medium for presenting information. The sound can be an effective complementary interface to the visual interface for data representation. Similar results were presented by Alty [5] for people with disabilities. On the other hand, the experimentation with sonification of surfaces in 124 Noirhomme-Fraiture et al. 3D space did not efficiently support the visual representation and certainly could not replace it (at least for the way it had been implemented). In general, this experimental work contributes to the research efforts on bringing other (non-visual) channels for information and data processing. This research area, that can be labelled as ‘perceptual data mining’, is focused on interactive systems that support rich perceptual – visual, audio, tactile – interaction between the human and the data representation. Such systems are expected to play significant role in assisting data understanding and supporting pattern discovery process, utilising human information processing capabilities. Acknowledgements We thank our colleagues Anne de Baenst and Florence Collot for their help in the preparation of this paper. References 1. 2. 3. 4. 5. 6. 7. 8. Ankerst, M.: Visual Data Mining: Ph.D. thesis. Faculty of Mathematics and Computer Science, University of Munich, Dissertation.de, 2000. Wickens, C.D. and Carswell, C.M.: The proximity compatibility principle: Its psychological foundation and relevance to display design. Human Factors. 37 (1995) 473-495 Simoff, S.J.: Towards the development of environments for designing visualisation support for visual data mining. Proceedings Int. Workshop on Visual Data Mining, 12th European Conference on Machine Learning and 5th European Conference on Principles and Practice of Knowledge Discovery in Databases ECML/PKDD2001. Freiburg, Germany (2001) 93-106 Anrijs, K.: The use of sound in 3d representations of symbolic objects. Namur, Facultés Universitaires Notre-Dame de la Paix, 1999. Alty, J.L., Rigas, D. and Vickers, P.: Using music as a communication medium. Proceedings of CHI'97 (1997) Brewster, S.A., Wright, P.C. and Edwards, A.D.N.: An evaluation of earcons for use in auditory human-computer interfaces. Proceedings of Inter CHI'93 (1993) Conversy, S. and Beaudouin-Lafon, M.: Le son dans les applications interactives. Paris, Laboratoire de Recherche en Informatique, Université de ParisSud, 1995. Hermann, T.: Data exploration by sonification, 1999. Available from [http://www.techfak.unibielefeld.de/techfak/ags/ni/projects/datamining/datamin_e.htm]. Sonification of Time Dependent Data 9. 10. 11. 12. 125 Kramer, G.: An introduction to auditory display, Auditory Display: Sonification, Audification, and Auditory Interfaces. Santa Fe Institute Studies in the Sciences of Complexity (1994) Noirhomme-Fraiture, M.: Le son dans les interfaces IHM: Application à la représen-tation de données multivariées complexes. Actes des 2èmes Journées Multimédia. Namur (2000) Sahyun, S.C.: A comparison of auditory and visual graphs for use in physics and mathematics, Oregon State University, 1999. Available from [http://www.physics.orst.edu/sahyun/thesis/]. Java: Javasound API programmer's guide, Sun Microsystems, 2000. Author Index Paulo Azevedo ………………………………………… 43 Dario Bruzzese ………………………………………… 55 Tizianna Catarci ………………………………………… 27 Luca Chittaro ………………………………………… 97 Carlo Combi ………………………………………… 97 Cristina. Davino ………………………………………… 55 Christophe Demoulin ………………………………………… 113 Alipio Jorge ………………………………………… 43 George Katopodis ………………………………………… 11 Stephen Kimani ………………………………………… 27 Kan Liu ………………………………………… 1 Monique Noirhomme-Fraiture ………………………………………… 113 Penny Noy ………………………………………… 81 João Poças ………………………………………… 43 François Poulet ………………………………………… 67 Giuseppe Santucci ………………………………………… 27 Olivier Schöller ………………………………………… 113 Michael Schroeder ………………………………………… 11, 81 Simeon. J. Simoff ………………………………………… 113 Giampaolo Trapasso ………………………………………… 97 Dongru Zhou ………………………………………… 1 Xiaozheng Zhou ………………………………………… 1