Lecture Notes In Computer Science 7914

2y ago
24 Views
2 Downloads
346.93 KB
15 Pages
Last View : 9d ago
Last Download : 3m ago
Upload by : Philip Renner
Transcription

Lecture Notes in Computer ScienceCommenced Publication in 1973Founding and Former Series Editors:Gerhard Goos, Juris Hartmanis, and Jan van LeeuwenEditorial BoardDavid HutchisonLancaster University, UKTakeo KanadeCarnegie Mellon University, Pittsburgh, PA, USAJosef KittlerUniversity of Surrey, Guildford, UKJon M. KleinbergCornell University, Ithaca, NY, USAAlfred KobsaUniversity of California, Irvine, CA, USAFriedemann MatternETH Zurich, SwitzerlandJohn C. MitchellStanford University, CA, USAMoni NaorWeizmann Institute of Science, Rehovot, IsraelOscar NierstraszUniversity of Bern, SwitzerlandC. Pandu RanganIndian Institute of Technology, Madras, IndiaBernhard SteffenTU Dortmund University, GermanyMadhu SudanMicrosoft Research, Cambridge, MA, USADemetri TerzopoulosUniversity of California, Los Angeles, CA, USADoug TygarUniversity of California, Berkeley, CA, USAGerhard WeikumMax Planck Institute for Informatics, Saarbruecken, Germany7914

Jesús Ariel Carrasco-OchoaJosé Francisco Martínez-TrinidadJoaquín Salas RodríguezGabriella Sanniti di Baja (Eds.)Pattern Recognition5th Mexican Conference, MCPR 2013Querétaro, Mexico, June 26-29, 2013Proceedings13

Volume EditorsJesús Ariel Carrasco-OchoaInstituto Nacional de Astrofísica, Óptica y Electrónica (INAOE)72840 Sta. Maria Tonantzintla, Puebla, MexicoE-mail: ariel@inaoep.mxJosé Francisco Martínez-TrinidadInstituto Nacional de Astrofísica, Óptica y Electrónica (INAOE)72840 Sta. Maria Tonantzintla, Puebla, MexicoE-mail: fmartine@inaoep.mxJoaquín Salas RodríguezInstituto Politécnico Nacional (IPN)76090 Colinas del Cimatario, Queretaro, MexicoE-mail: jsalasr@ipn.mxGabriella Sanniti di BajaIstituto di Cibernetica "E. Caianiello", CNR80078 Pozzuoli, Naples, ItalyE-mail: g.sannitidibaja@cib.na.cnr.itISSN 0302-9743e-ISSN 1611-3349ISBN 978-3-642-38988-7e-ISBN 978-3-642-38989-4DOI 10.1007/978-3-642-38989-4Springer Heidelberg Dordrecht London New YorkLibrary of Congress Control Number: 2013940329CR Subject Classification (1998): I.2, I.4, I.5, H.3, F.1, H.4LNCS Sublibrary: SL 6 – Image Processing, Computer Vision, Pattern Recognition,and Graphics Springer-Verlag Berlin Heidelberg 2013This work is subject to copyright. All rights are reserved by the Publisher, whether the whole or part ofthe material is concerned, specifically the rights of translation, reprinting, reuse of illustrations, recitation,broadcasting, reproduction on microfilms or in any other physical way, and transmission or informationstorage and retrieval, electronic adaptation, computer software, or by similar or dissimilar methodologynow known or hereafter developed. Exempted from this legal reservation are brief excerpts in connectionwith reviews or scholarly analysis or material supplied specifically for the purpose of being entered andexecuted on a computer system, for exclusive use by the purchaser of the work. Duplication of this publicationor parts thereof is permitted only under the provisions of the Copyright Law of the Publisher’s location,in its current version, and permission for use must always be obtained from Springer. Permissions for usemay be obtained through RightsLink at the Copyright Clearance Center. Violations are liable to prosecutionunder the respective Copyright Law.The use of general descriptive names, registered names, trademarks, service marks, etc. in this publicationdoes not imply, even in the absence of a specific statement, that such names are exempt from the relevantprotective laws and regulations and therefore free for general use.While the advice and information in this book are believed to be true and accurate at the date of publication,neither the authors nor the editors nor the publisher can accept any legal responsibility for any errors oromissions that may be made. The publisher makes no warranty, express or implied, with respect to thematerial contained herein.Typesetting: Camera-ready by author, data conversion by Scientific Publishing Services, Chennai, IndiaPrinted on acid-free paperSpringer is part of Springer Science Business Media (www.springer.com)

XIVTable of ContentsNeural NetworksAssociative Model for the Forecasting of Time Series Based on theGamma Classifier . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .Itzamá López-Yáñez, Leonid Sheremetov, andCornelio Yáñez-MárquezModified Dendrite Morphological Neural Network Applied to 3DObject Recognition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .Humberto Sossa and Elizabeth GuevaraHybrid Associative Memories for Imbalanced Data Classification:An Experimental Study . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .L. Cleofas-Sánchez, V. Garcı́a, R. Martı́n-Félez, R.M. Valdovinos,J.S. Sánchez, and O. Camacho-Nieto304314325Assessments Metrics for Multi-class Imbalance Learning: A PreliminaryStudy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .R. Alejo, J.A. Antonio, R.M. Valdovinos, and J.H. Pacheco-Sánchez335Non-conventional Control and Implementation of an Electric WheelchairDesigned to Climb Up Stairs, Controlled via Electromyography andSupported by Artificial Neural Network Processing . . . . . . . . . . . . . . . . . . .Martı́n L. Guzmán, Juan P. Pinto, Luis F. Reina, andCarlos A. Esquit344Document ProcessingA Question Answering System for Reading Comprehension Tests . . . . . . .Helena Gómez-Adorno, David Pinto, and Darnes VilariñoDetermining the Degree of Semantic Similarity Using PrototypeVectors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .Mireya Tovar, David Pinto, Azucena Montes, and Darnes Vilariño354364Single Extractive Text Summarization Based on a Genetic Algorithm . . .René Arnulfo Garcı́a-Hernández and Yulia Ledeneva374Author Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .385

Hybrid Associative Memories for Imbalanced DataClassification: An Experimental StudyL. Cleofas-Sánchez1 , V. Garcı́a2 , R. Martı́n-Félez2, R.M. Valdovinos3,J.S. Sánchez2 , and O. Camacho-Nieto41Centro de Investigación en Computación, Instituto Politécnico Nacional, Av. Juan de DiosBátiz s/n, Col. Nueva Industrial Vallejo, 07738 México D.F., México2Institute of New Imaging Technologies, Department of Computer Languages and Systems,Universitat Jaume I, Av. Vicent Sos Baynat s/n, 12071 Castellón de la Plana, Spain3Centro Universitario Valle de Chalco, Universidad Autónoma del Estado de México,Hermenegildo Galena 3, 56615 Valle de Chalco, México4Centro de Innovación y Desarrollo Tecnológico en Cómputo, Instituto Politécnico Nacional,Av. Juan de Dios Bátiz s/n, Col. Nueva Industrial Vallejo, 07700 México D.F., MéxicoAbstract. Hybrid associative memories are based on the combination of twowell-known associative networks, the lernmatrix and the linear associator, withthe aim of taking advantage of their merits and overcoming their limitations.While these models have extensively been applied to information retrieval problems, they have not been properly studied in the framework of classification andeven less with imbalanced data. Accordingly, this work intends to give a comprehensive response to some issues regarding imbalanced data classification: (i) Arethe hybrid associative models suitable for dealing with this sort of data? and, (ii)Does the degree of imbalance affect the performance of these neural classifiers?Experiments on real-world data sets demonstrate that independently of the imbalance ratio, the hybrid associative memories perform poorly in terms of area underthe ROC curve, but the hybrid associative classifier with translation appears to bethe best solution when assessing the true positive rate.Keywords: Class Imbalance, Associative Memory, Neural Network.1 IntroductionAn associative memory [1] is a type of neural network that allows to recall the previously stored training example xi that most closely resembles the one presented to thenetwork. This connectionist model has demonstrated to be very effective for information storage and retrieval [2–4], but it has not been much studied in the framework ofclassification. Among the simplest and first studied associative memory models are thelernmatrix [5] and the linear associator [6,7], which are considered as hetero-associativememories capable of producing exact recall. Both these models can also work as classifiers, but they present some drawbacks that make difficult their application to many reallife problems: the lernmatrix needs to be provided with binary input vectors xi {0, 1},whereas the linear associator requires the input vectors to be orthonormal and linearlyindependent.J.A. Carrasco-Ochoa et al. (Eds.): MCPR 2013, LNCS 7914, pp. 325–334, 2013.c Springer-Verlag Berlin Heidelberg 2013

326L. Cleofas-Sánchez et al.In order to benefit from the advantages of these associative memories and overcome their shortcomings, several extensions have been developed. These include thehybrid associative classifier (HAC) and the hybrid associative classifier with translation(HACT) [8], which combine the procedure used by the linear associator in the learningphase with the recall stage of the lernmatrix. While these two classification models havebeen used with some success in a number of applications, there still exist open questionsregarding their limitations that deserve a more thorough investigation. For example, thepresent paper addresses the issue of imbalanced data classification [9], which appearsas a much more challenging task for this type of associative memories.Many complex pattern recognition and data mining problems are characterized byimbalanced data, where at least one class is heavily under-represented as compared toothers. Following the common practice in the area [10, 11], we will here consider onlybinary classification problems where the examples from the majority class are oftenreferred to as the negative examples and those from the minority class as the positiveexamples, since these usually represent the concept of most interest.The importance of the class imbalance problem comes from the fact that in general,it hinders the performance of most standard learning algorithms because they are oftenbiased towards the majority class and have a poor performance on the minority class.Besides the classifiers are commonly built with the aim of reducing the overall error,what may lead to erroneous conclusions; for example, an algorithm that achieves anaccuracy of 99% will be worthless if it fails on classifying all positive examples.Many classifiers have been investigated in the context of class imbalance, rangingfrom the nearest neighbor rule and decision trees to support vector machines and varioustopologies of neural networks [11–15]. However, to the best of our knowledge, the useof associative memory models has not received adequate attention from researchers onthis topic. In fact, we have found only a recent work [16] that analyzes the performanceof the HACT approach after under-sampling the imbalanced data set, but it presentsseveral limitations such as the reduced number of databases used in the experiments,the lack of comparisons with other state-of-the-art classifiers and especially, the factthat it does not take care of the imbalance ratio (i.e. the ratio of the majority to theminority instances) and its effect on the HACT performance.The purpose of this paper is to gain insight into the behavior of the HAC and HACTassociative models when these are used for the classification of imbalanced data, pursuing to fully understand how the class imbalance affects the performance of theseclassifiers. To this end, we provide a large pool of experiments on 58 real-world benchmarking data sets that have different degrees of imbalance, comparing those hybridassociative memories with other well-known artificial neural networks: a Bayesian network (BNet), a multilayer perceptron (MLP) and a radial basis function (RBF). Weconducted our experiments by evaluating three performance metrics: the area under theROC curve, the true positive rate and the true negative rate.2 Two Hybrid Associative MemoriesIn this section we provide a brief introduction to the associative memory models thatwill be further experimented with, covering only the general concepts and notation

Hybrid Associative Memories for Imbalanced Data Classification327needed to understand their foundations. For a complete description of associative memories, the reader may review any of the many books on this subject (e.g. [17, 18]).In general, an associative memory can be defined as a mapping matrix M so that aninput vector xi Rn (with n components) will be transformed into an output vectoryi Rm (with m components), that isyi Mxii 1, . . . , p(1)where p denotes the number of input vectors.The stored samples will be represented in the form of pairs of associations (xi , yi )between the input and output vectors, xi and yi , and are often called fundamental pattern. The set of p pairs (fundamental patterns) constitutes the fundamental set of associations.The matrix M has to be determined through an iterative procedure in the learningphase. Afterwards, during the recall or recovery phase, an unknown pattern x0 will beapplied to the input of the matrix in order to produce the vector y0 , which is expectedto be a good approximation of the true output y.Hybrid Associative Classifier (HAC). As previously pointed out, the HAC model [8]arises from the combination of the lernmatrix and the linear associator with the aim ofovercoming the practical drawbacks of these associative neural networks. Apart fromthese obvious advantages, it is worth remarking that the HAC model presents someother interesting properties such as simplicity, requirements of low computational costand the ability to support real-valued input vectors [8].During the learning phase, the HAC memory imitates the process of the linear associator: each sample that belongs to class k is represented by a vector with zeros inall components except the k’th element that equals 1. In this way, the outer product ofvectors xi and yi gives the corresponding associations between them. Then the matrixM of size n m will be obtained as the sum of all p outer products asM p (yi )(xi )T(2)i 1After computing the mapping matrix M, the recovery of a given input sample will beperformed following the process of the lernmatrix model in order to estimate its classlabel.It has to be pointed out, however, that a practical drawback of the HAC model comesfrom the possible large differences in the magnitude of the input vectors because in sucha case, the vectors with a lower magnitude will be assigned to the class of the vectorswith a larger magnitude.Hybrid Associative Classifier with Translation (HACT). This is a modification ofthe HAC model that tries to face several of its limitations. More specifically, if the inputsamples are clustered in the same quadrant, the performance of the HAC memory will

328L. Cleofas-Sánchez et al.be affected negatively. Thus the HACT approach [8] starts with a translation of thecoordinate axes whose origin is taken to lie in the mean vector of all the input vectorsas computed byp1 ix x(3)p i 1In this way, the new coordinate axes are parallel to the original coordinate axes, buteliminates the clustering of samples in a unique quadrant. Then the input and test vectors in the new coordinate system will be obtained as xi xi x. After the corresponding translation of axes, the learning and recovery phases will be the same as thosedescribed for the HAC model.3 Experimental Set-UpAs already discussed, the aim of this work and the experiments conducted here is toinvestigate whether two models of hybrid associative memories, which are based on thelernmatrix and the linear associator, are suitable or not for imbalanced data classification, and to what extent the degree of imbalance may affect their performance.Table 1. Description of the data sets used in the experimentsData setsFeatures 81484Haberman3306Vehicle3188469214Glass-0-1-2-3 vs age-blocks01054727200Ecoli-0-3-4 vs 58514Yeast-2 vs 47222Ecoli-0-6-7 vs 3-57202Ecoli-0-2-3-4 vs 59172Glass-0-1-5 vs 28506Yeast-0-3-5-9 vs 7-881004Yeast-0-2-5-6 vs 3-7-8-981004Yeast-0-2-5-7-9 vs 3-6-86203Ecoli-0-4-6 vs 57244Ecoli-0-1 vs 2-3-57224Ecoli-0-2-6-7 vs 3-5992Glass-0-4 vs .149.159.179.189.22Data setsFeatures SamplesEcoli-0-3-4-6 vs 57205Ecoli-0-3-4-7 vs 5-67257Yeast-0-5-6-7-9 vs 48528Vowel013988Ecoli-0-6-7 vs 56220Glass-0-1-6 vs 29192Ecoli-0-1-4-7 vs 2-3-5-67336Led-0-2-4-5-6-7-8-9 vs 17443Ecoli-0-1 vs 56240Glass-0-6 vs 59108Glass-0-1-4-6 vs 29205Glass29214Ecoli-0-1-4-7 vs 5-66332Cleveland-0 vs 413177Ecoli-0-1-4-6 vs 56280Shuttle-0 vs 491829Yeast-1 vs 77459Glass49214Ecoli47336Page-blocks-1-3 vs 410472Glass-0-1-6 vs 59184Yeast-1-4-5-8 vs 78693Glass59214Yeast-2 vs 88482Yeast481484Yeast-1-2-8-9 vs 78947Yeast581484Ecoli-0-1-3-7 vs 339.1441.40

Hybrid Associative Memories for Imbalanced Data Classification329The empirical analysis has been performed over a total of 58benchmarking data sets taken from the KEEL Data Set Repository(http://www.keel.es/dataset.php) [19]. Note that all the originalmulti-class databases have firstly been transformed into two-class problems. Table 1summarizes the main characteristics of the data sets, including the imbalance ratio(IR), i.e. the number of negative examples divided by the number of positive examples.As can be seen, the databases chosen for the experiments go from a low imbalance of1.82 in Glass1 to a high/moderate imbalance of 41.40 in the case of Yeast6.In order to gain sufficient insight into the behavior of the associative memory models, three other neural networks (BNet, MLP, RBF) have been used as baselines forcomparison purposes. These were taken from the Weka toolkit [20] with their defaultparameter values. For the experiments here carried out, we have adopted a 5-fold crossvalidation method to estimate three classification performance measures commonlyused in skewed domains: the area under the ROC curve (AUC), the true positive rate(TPrate) and the true negative rate (TNrate). Each data set has been divided into fivestratified blocks of size N/5 (where N denotes the total number of samples in thedatabase), using four folds for training the connectionist classifiers and the remainingblock as an independent test set. Therefore the results reported in tables of Section 4correspond to those three measures averaged over the five runs.Table 2. Confusion matrix for a two-class problemPredicted positive Predicted negativePositive class True Positive (TP) False Negative (FN)Negative class False Positive (FP) True Negative (TN)Given a 2 2 confusion matrix as that illustrated in Table 2, the performance meaPsures used in the experiments can be calculated as follows: T P rate T PT FN,TNT P rate T N rateT N rate T N F P , and AU C . Note that the latter corresponds2to the AUC defined by a single point on the ROC curve.4 Experimental ResultsTable 3 reports the AUC values obtained by the neural network models on each database,along with the average across the whole collection of data sets. From these results, thefirst observation is that the HAC memory yields a 50% of AUC, which indicates thatall samples of one class have been misclassified while all of the other have been correctly classified. This effect has not been found in the case of the HACT model, but itsperformance in terms of AUC is lower than that achieved by the three other neural networks on most databases. When paying attention of the average values, the MLP modelclearly performs the best (80.70% of AUC), but the results of the Bayesian network andthe RBF are not too far from that of the HACT approach.

330L. Cleofas-Sánchez et al.Table 3. Experimental results using the AUCData s-0-1-2-3 vs 6Yeast3Ecoli3Page-blocks0Ecoli-0-3-4 vs 5Yeast-2 vs 4Ecoli-0-6-7 vs 3-5Ecoli-0-2-3-4 vs 5Glass-0-1-5 vs 2Yeast-0-3-5-9 vs 7-8Yeast-0-2-5-6 vs 3-7-8-9Yeast-0-2-5-7-9 vs 3-6-8Ecoli-0-4-6 vs 5Ecoli-0-1 vs 2-3-5Ecoli-0-2-6-7 vs 3-5Glass-0-4 vs 050.2461.4567.6688.8686.6979.2181.0194.41Data setEcoli-0-3-4-6 vs 5Ecoli-0-3-4-7 vs 5-6Yeast-0-5-6-7-9 vs 4Vowel0Ecoli-0-6-7 vs 5Glass-0-1-6 vs 2Ecoli-0-1-4-7 vs 2-3-5-6Led-0-2-4-5-6-7-8-9 vs 1Ecoli-0-1 vs 5Glass-0-6 vs 5Glass-0-1-4-6 vs 2Glass2Ecoli-0-1-4-7 vs 5-6Cleveland-0 vs 4Ecoli-0-1-4-6 vs 5Shuttle-0 vs 4Yeast-1 vs 7Glass4Ecoli4Page-blocks-1-3 vs 4Glass-0-1-6 vs 5Yeast-1-4-5-8 vs 7Glass5Yeast-2 vs 8Yeast4Yeast-1-2-8-9 vs 7Yeast5Ecoli-0-1-3-7 vs 6763.3084.6350.0077.05By the analysis of the behavior of these classifiers as a function of the imbalanceratio, one can guess that there is not necessarily a direct relationship between the classification performance and the degree of imbalance. For instance, the balanced accuraciesfor the Ecoli-0-1-3-7 vs 2-6 database, which has an imbalance ratio of 39.14, are significantly higher than those for Glass1, even though this presents a very low imbalanceratio of 1.82. In some sense, it appears that databases may also suffer from other intrinsic problems such as class overlapping, small disjuncts, feature noise and lack ofrepresentative data, which in turn may affect classification performance much morestrongly than the presence of class imbalance.In order to accomplish a better understanding of the performance of these neuralnetwork models, Tables 4 and 5 report the true positive and true negative rates, respectively. These measures allow to analyze the behavior of a classifier on each individualclass, thus drawing out whether it is biased towards one class or another. This is especially important in the context of imbalanced data because the examples from theminority class, which usually correspond to the most interesting cases, are more likelyto be misclassified. In addition, it is often preferable to achieve a higher true positiverate rather than a higher true negative rate and consequently, the AUC by itself is notsufficiently informative when evaluating the performance of a set of classifiers.

Hybrid Associative Memories for Imbalanced Data Classification331Table 4. Experimental results using the true positive rateData setHAC HACT BNet MLP RBF Data setHAC HACT BNet MLP RBFGlass10 80.17 47.34 59.60 50.00 Ecoli-0-3-4-6 vs 50 95.00 7080 85.00Pima0 44.36 58.22 67.18 55.20 Ecoli-0-3-4-7 vs 5-60 96.00 48.00 80.00 72.000 86.36 18.00 48.72 8.00Iris00100 100 100 100 Yeast-0-5-6-7-9 vs 40 97.78 78.86 98.88 75.56Glass00100 80.00 70.00 42.84 Vowel00 95.00 65.00 75.00 75.00Yeast10 76.68 46.14 43.84 27.28 Ecoli-0-6-7 vs 50100000Haberman0 59.26 17.52 28.20 15.98 Glass-0-1-6 vs 2Vehicle30 60.33 63.64 58.94 41.92 Ecoli-0-1-4-7 vs 2-3-5-6 0 93.33 62.66 76.00 58.660 94.00 80.18 87.74 84.36 Led-0-2-4-5-6-7-8-9 vs 1 2.50 100 78.20 81.06 67.84Glass-0-1-2-3 vs 4-5-60100 75.00 80.00 80.00Vehicle00100 95.94 90.98 80.92 Ecoli-0-1 vs 50100 70.00 100 90.00Ecoli10 94.83 83.16 76.68 91.02 Glass-0-6 vs 50100000New-thyroid20 91.43 85.70 91.42 97.14 Glass-0-1-4-6 vs 201000 6.66 0Ecoli20 96.36 77.44 82.72 87.08 Glass20100 44.00 72.00 68.00Segment00100 98.20 99.10 97.90 Ecoli-0-1-4-7 vs 5-60 33.50 26.04 78.18 71.52Glass60 96.67 86.66 72.00 78.66 Cleveland-0 vs 40100 75.00 60.00 80.00Yeast30 98.79 72.94 74.28 77.32 Ecoli-0-1-4-6 vs 50 99.20 100 99.20 98.40Ecoli30 97.14 79.98 59.98 34.30 Shuttle-0 vs 40 76.67 13.34 26.64 10.00Page-blocks00 19.15 85.32 76.92 50.84 Yeast-1 vs 7Ecoli-0-3-4 vs 50100 70.00 80.00 85.00 Glass40 90.00 33.32 76.68 76.680 90.18 76.54 66.54 78.36 Ecoli40100 65.00 80.00 80.00Yeast-2 vs 40 88.00 80.00 67.00 41.00 Page-blocks-1-3 vs 40 68.67 100 96.00 86.00Ecoli-0-6-7 vs 3-50100 75.00 80.00 80.00 Glass-0-1-6 vs 50100 90.00 60.00 80.00Ecoli-0-2-3-4 vs 50 95.00 0 13.34 5.00 Yeast-1-4-5-8 vs 70 66.67 0 3.34 0Glass-0-1-5 vs 20 86.00 20.00 34.00 24.00 Glass50100 90.00 80.00 70.00Yeast-0-3-5-9 vs 7-80 70.00 55.00 55.00 60.00Yeast-0-2-5-6 vs 3-7-8-9 0 77.68 54.36 49.42 37.32 Yeast-2 vs 80 90.18 29.28 29.46 0Yeast-0-2-5-7-9 vs 3-6-8 0 89.95 70.68 73.78 79.94 Yeast40 95.00 80.00 80.00 75.00 Yeast-1-2-8-9 vs 70 80.00 16.68 13.34 3.34Ecoli-0-4-6 vs 50 96.00 10.00 65.00 65.00 Yeast50100 86.4 68.08 26.92Ecoli-0-1 vs 2-3-50 90.00 63.00 64.00 64.00 Ecoli-0-1-3-7 vs 2-60100 70.00 70.00 70.00Ecoli-0-2-6-7 vs 3-50100 100 100 90.00 Yeast60 94.29 71.42 48.58 0Glass-0-4 vs 5Average0.04 88.96 60.16 64.75 57.42For instance, the results of HAC in Table 4 demonstrate that this hybrid associativemodel is of no value at all because it fails on the classification of all positive examples.This makes clear that the AUC of 50% reported in Table 3 is due to the awful truepositive rate of this classifier and the very high rate achieved on the negative class (seeTable 5). On the contrary, the true positive rate of HACT suggests that this can be a goodtool for the classification of data with class imbalance because it yields a true positiverate of close to 89% in average, that is, even higher than that of the best performingalgorithm (MLP) in terms of AUC.It is also interesting to note that in general, the highest differences between HACTand MLP are found in the most strongly imbalanced data sets. Unfortunately, in thesecases, the true negative rate of the HACT model is lower than that of the MLP, but weshould recall that there exist numerous real-world applications in which the minorityclass represents the concept of most interest and therefore, it will be crucial to correctlyclassify the positive examples even if this might entail a certain degradation of the truenegative rate.

332L. Cleofas-Sánchez et al.Table 5. Experimental results using the true negative rateData s-0-1-2-3 vs 6Yeast3Ecoli3Page-blocks0Ecoli-0-3-4 vs 5Yeast-2 vs 4Ecoli-0-6-7 vs 3-5Ecoli-0-2-3-4 vs 5Glass-0-1-5 vs 2Yeast-0-3-5-9 vs 7-8Yeast-0-2-5-6 vs 3-7-8-9Yeast-0-2-5-7-9 vs 3-6-8Ecoli-0-4-6 vs 5Ecoli-0-1 vs 2-3-5Ecoli-0-2-6-7 vs 3-5Glass-0-4 vs 8.0097.7898.3893.4298.0298.82Data setEcoli-0-3-4-6 vs 5Ecoli-0-3-4-7 vs 5-6Yeast-0-5-6-7-9 vs 4Vowel0Ecoli-0-6-7 vs 5Glass-0-1-6 vs 2Ecoli-0-1-4-7 vs 2-3-5-6Led-0-2-4-5-6-7-8-9 vs 1Ecoli-0-1 vs 5Glass-0-6 vs 5Glass-0-1-4-6 vs 2Glass2Ecoli-0-1-4-7 vs 5-6Cleveland-0 vs 4Ecoli-0-1-4-6 vs 5Shuttle-0 vs 4Yeast-1 vs 7Glass4Ecoli4Page-blocks-1-3 vs 4Glass-0-1-6 vs 5Yeast-1-4-5-8 vs 7Glass5Yeast-2 vs 8Yeast4Yeast-1-2-8-9 vs 7Yeast5Ecoli-0-1-3-7 vs 99.4210098.0499.5610010099.6899.2610096.685 Conclusions an

Hermenegildo Galena 3, 56615 Valle de Chalco, M exico 4 Centro de Innovaci on y Desarrollo Tecnol ogico en C omputo, Instituto Polit ecnico Nacional, Av. Juan de Dios B atiz s/n, Col. Nueva Industrial Vallejo, 07700 M exico D.F., M exico Abstract. Hybrid associative memories are based on the combination of two

Related Documents:

Introduction of Chemical Reaction Engineering Introduction about Chemical Engineering 0:31:15 0:31:09. Lecture 14 Lecture 15 Lecture 16 Lecture 17 Lecture 18 Lecture 19 Lecture 20 Lecture 21 Lecture 22 Lecture 23 Lecture 24 Lecture 25 Lecture 26 Lecture 27 Lecture 28 Lecture

GEOMETRY NOTES Lecture 1 Notes GEO001-01 GEO001-02 . 2 Lecture 2 Notes GEO002-01 GEO002-02 GEO002-03 GEO002-04 . 3 Lecture 3 Notes GEO003-01 GEO003-02 GEO003-03 GEO003-04 . 4 Lecture 4 Notes GEO004-01 GEO004-02 GEO004-03 GEO004-04 . 5 Lecture 4 Notes, Continued GEO004-05 . 6

Lecture 1: A Beginner's Guide Lecture 2: Introduction to Programming Lecture 3: Introduction to C, structure of C programming Lecture 4: Elements of C Lecture 5: Variables, Statements, Expressions Lecture 6: Input-Output in C Lecture 7: Formatted Input-Output Lecture 8: Operators Lecture 9: Operators continued

2 Lecture 1 Notes, Continued ALG2001-05 ALG2001-06 ALG2001-07 ALG2001-08 . 3 Lecture 1 Notes, Continued ALG2001-09 . 4 Lecture 2 Notes ALG2002-01 ALG2002-02 ALG2002-03 . 5 Lecture 3 Notes ALG2003-01 ALG2003-02 ALG

Artificial Intelligence COMP-424 Lecture notes by Alexandre Tomberg Prof. Joelle Pineau McGill University Winter 2009 Lecture notes Page 1 . I. History of AI 1. Uninformed Search Methods . Lecture notes Page 58 . Lecture notes Page 59 . Soft EM for a general Bayes net: Lecture notes Page 60 . Machine Learning: Clustering March-19-09

Lecture 1: Introduction and Orientation. Lecture 2: Overview of Electronic Materials . Lecture 3: Free electron Fermi gas . Lecture 4: Energy bands . Lecture 5: Carrier Concentration in Semiconductors . Lecture 6: Shallow dopants and Deep -level traps . Lecture 7: Silicon Materials . Lecture 8: Oxidation. Lecture

TOEFL Listening Lecture 35 184 TOEFL Listening Lecture 36 189 TOEFL Listening Lecture 37 194 TOEFL Listening Lecture 38 199 TOEFL Listening Lecture 39 204 TOEFL Listening Lecture 40 209 TOEFL Listening Lecture 41 214 TOEFL Listening Lecture 42 219 TOEFL Listening Lecture 43 225 COPYRIGHT 2016

Partial Di erential Equations MSO-203-B T. Muthukumar tmk@iitk.ac.in November 14, 2019 T. Muthukumar tmk@iitk.ac.in Partial Di erential EquationsMSO-203-B November 14, 2019 1/193 1 First Week Lecture One Lecture Two Lecture Three Lecture Four 2 Second Week Lecture Five Lecture Six 3 Third Week Lecture Seven Lecture Eight 4 Fourth Week Lecture .