Semi-supervised and unsupervised machine learning : novel strategies /

"This book provides a detailed and up-to-date overview on classification and data mining methods. The first part is focused on supervised classification algorithms and their applications, including recent research on the combination of classifiers. The second part deals with unsupervised data m...

Πλήρης περιγραφή

Λεπτομέρειες βιβλιογραφικής εγγραφής
Κύριος συγγραφέας: Albalate, Amparo
Άλλοι συγγραφείς: Minker, Wolfgang
Μορφή: Ηλ. βιβλίο
Γλώσσα:English
Έκδοση: London : ISTE ; 2011.
Hoboken, NJ : Wiley, 2011.
Θέματα:
Διαθέσιμο Online:Full Text via HEAL-Link
LEADER 10773nam a2200985 4500
001 ocn826657834
003 OCoLC
005 20170124070802.7
006 m o d
007 cr cnu---unuuu
008 130204s2011 enka ob 001 0 eng d
040 |a N$T  |b eng  |e pn  |c N$T  |d YDXCP  |d E7B  |d DG1  |d IDEBK  |d UMI  |d COO  |d DEBSZ  |d OCLCQ  |d OCLCF  |d OCLCQ  |d DEBBG  |d GrThAP 
019 |a 857717622 
020 |a 9781118557693  |q (electronic bk.) 
020 |a 1118557697  |q (electronic bk.) 
020 |a 9781118586334  |q (electronic bk.) 
020 |a 1118586336  |q (electronic bk.) 
020 |a 9781118586136 
020 |a 1118586131 
020 |z 9781848212039 
020 |z 1848212038 
029 1 |a AU@  |b 000050718668 
029 1 |a AU@  |b 000052007855 
029 1 |a CHBIS  |b 009914518 
029 1 |a CHNEW  |b 000605226 
029 1 |a CHVBK  |b 140082158 
029 1 |a DEBBG  |b BV041432025 
029 1 |a DEBBG  |b BV041911133 
029 1 |a DEBSZ  |b 398279144 
029 1 |a NZ1  |b 15916408 
029 1 |a DEBBG  |b BV043395352 
035 |a (OCoLC)826657834  |z (OCoLC)857717622 
037 |a CL0500000277  |b Safari Books Online 
050 4 |a QA76.9.D343  |b A3393 2011eb 
072 7 |a COM  |x 005030  |2 bisacsh 
072 7 |a COM  |x 004000  |2 bisacsh 
082 0 4 |a 006.3  |2 22 
084 |a TEC008000  |2 bisacsh 
049 |a MAIN 
100 1 |a Albalate, Amparo. 
245 1 0 |a Semi-supervised and unsupervised machine learning :  |b novel strategies /  |c Amparo Albalate, Wolfgang Minker. 
264 1 |a London :  |b ISTE ;  |c 2011. 
264 1 |a Hoboken, NJ :  |b Wiley,  |c 2011. 
300 |a 1 online resource (x, 244 pages) :  |b illustrations 
336 |a text  |b txt  |2 rdacontent 
337 |a computer  |b c  |2 rdamedia 
338 |a online resource  |b cr  |2 rdacarrier 
520 |a "This book provides a detailed and up-to-date overview on classification and data mining methods. The first part is focused on supervised classification algorithms and their applications, including recent research on the combination of classifiers. The second part deals with unsupervised data mining and knowledge discovery, with special attention to text mining. Discovering the underlying structure on a data set has been a key research topic associated to unsupervised techniques with multiple applications and challenges, from web-content mining to the inference of cancer subtypes in genomic microarray data. Among those, the book focuses on a new application for dialog systems which can be thereby made adaptable and portable to different domains. Clustering evaluation metrics and new approaches, such as the ensembles of clustering algorithms, are also described"--  |c Provided by publisher. 
504 |a Includes bibliographical references and index. 
505 0 0 |g Machine generated contents note:  |g pt. 1  |t State of the Art --  |g ch. 1  |t Introduction --  |g 1.1.  |t Organization of the book --  |g 1.2.  |t Utterance corpus --  |g 1.3.  |t Datasets from the UCI repository --  |g 1.3.1.  |t Wine dataset (wine) --  |g 1.3.2.  |t Wisconsin breast cancer dataset (breast) --  |g 1.3.3.  |t Handwritten digits dataset (Pendig) --  |g 1.3.4.  |t Pima Indians diabetes (diabetes) --  |g 1.3.5.  |t Iris dataset (Iris) --  |g 1.4.  |t Microarray dataset --  |g 1.5.  |t Simulated datasets --  |g 1.5.1.  |t Mixtures of Gaussians --  |g 1.5.2.  |t Spatial datasets with non-homogeneous inter-cluster distance --  |g ch. 2  |t State of the Art in Clustering and Semi-Supervised Techniques --  |g 2.1.  |t Introduction --  |g 2.2.  |t Unsupervised machine learning (clustering) --  |g 2.3.  |t A brief history of cluster analysis --  |g 2.4.  |t Cluster algorithms --  |g 2.4.1.  |t Hierarchical algorithms --  |g 2.4.1.1.  |t Agglomerative clustering --  |g 2.4.1.2.  |t Divisive algorithms --  |g 2.4.2.  |t Model-based clustering --  |g 2.4.2.1.  |t The expectation maximization (EM) algorithm --  |g 2.4.3.  |t Partitional competitive models. 
505 0 0 |g 2.4.3.1.  |t K-means --  |g 2.4.3.2.  |t Neural gas --  |g 2.4.3.3.  |t Partitioning around Medoids (PAM) --  |g 2.4.3.4.  |t Self-organizing maps --  |g 2.4.4.  |t Density-based clustering --  |g 2.4.4.1.  |t Direct density reachability --  |g 2.4.4.2.  |t Density reachability --  |g 2.4.4.3.  |t Density connection --  |g 2.4.4.4.  |t Border points --  |g 2.4.4.5.  |t Noise points --  |g 2.4.4.6.  |t DBSCAN algorithm --  |g 2.4.5.  |t Graph-based clustering --  |g 2.4.5.1.  |t Pole-based overlapping clustering --  |g 2.4.6.  |t Affectation stage --  |g 2.4.6.1.  |t Advantages and drawbacks --  |g 2.5.  |t Applications of cluster analysis --  |g 2.5.1.  |t Image segmentation --  |g 2.5.2.  |t Molecular biology --  |g 2.5.2.1.  |t Biological considerations --  |g 2.5.3.  |t Information retrieval and document clustering --  |g 2.5.3.1.  |t Document pre-processing --  |g 2.5.3.2.  |t Boolean model representation --  |g 2.5.3.3.  |t Vector space model --  |g 2.5.3.4.  |t Term weighting --  |g 2.5.3.5.  |t Probabilistic models --  |g 2.5.4.  |t Clustering documents in information retrieval --  |g 2.5.4.1.  |t Clustering of presented results --  |g 2.5.4.2.  |t Post-retrieval document browsing (Scatter-Gather) --  |g 2.6.  |t Evaluation methods. 
505 0 0 |g 2.7.  |t Internal cluster evaluation --  |g 2.7.1.  |t Entropy --  |g 2.7.2.  |t Purity --  |g 2.7.3.  |t Normalized mutual information --  |g 2.8.  |t External cluster validation --  |g 2.8.1.  |t Hartigan --  |g 2.8.2.  |t Davies Bouldin index --  |g 2.8.3.  |t Krzanowski and Lai index --  |g 2.8.4.  |t Silhouette --  |g 2.8.5.  |t Gap statistic --  |g 2.9.  |t Semi-supervised learning --  |g 2.9.1.  |t Self training --  |g 2.9.2.  |t Co-training --  |g 2.9.3.  |t Generative models --  |g 2.10.  |t Summary --  |g pt. 2  |t Approaches to Semi-Supervised Classification --  |g ch. 3  |t Semi-Supervised Classification Using Prior Word Clustering --  |g 3.1.  |t Introduction --  |g 3.2.  |t Dataset --  |g 3.3.  |t Utterance classification scheme --  |g 3.3.1.  |t Pre-processing --  |g 3.3.1.1.  |t Utterance vector representation --  |g 3.3.2.  |t Utterance classification --  |g 3.4.  |t Semi-supervised approach based on term clustering --  |g 3.4.1.  |t Term clustering --  |g 3.4.2.  |t Semantic term dissimilarity --  |g 3.4.2.1.  |t Term vector of lexical co-occurrences --  |g 3.4.2.2.  |t Metric of dissimilarity --  |g 3.4.3.  |t Term vector truncation --  |g 3.4.4.  |t Term clustering --  |g 3.4.5.  |t Feature extraction and utterance feature vector. 
505 0 0 |g 3.4.6.  |t Evaluation --  |g 3.5.  |t Disambiguation --  |g 3.5.1.  |t Evaluation --  |g 3.6.  |t Summary --  |g ch. 4  |t Semi-Supervised Classification Using Pattern Clustering --  |g 4.1.  |t Introduction --  |g 4.2.  |t New semi-supervised algorithm using the cluster and label strategy --  |g 4.2.1.  |t Block diagram --  |g 4.2.1.1.  |t Dataset --  |g 4.2.1.2.  |t Clustering --  |g 4.2.1.3.  |t Optimum cluster labeling --  |g 4.2.1.4.  |t Classification --  |g 4.3.  |t Optimum cluster labeling --  |g 4.3.1.  |t Problem definition --  |g 4.3.2.  |t The Hungarian algorithm --  |g 4.3.2.1.  |t Weighted complete bipartite graph --  |g 4.3.2.2.  |t Matching, perfect matching and maximum weight matching --  |g 4.3.2.3.  |t Objective of Hungarian method --  |g 4.3.2.4.  |t Complexity considerations --  |g 4.3.3.  |t Genetic algorithms --  |g 4.3.3.1.  |t Reproduction operators --  |g 4.3.3.2.  |t Forming the next generation --  |g 4.3.3.3.  |t GAs applied to optimum cluster labeling --  |g 4.3.3.4.  |t Comparison of methods --  |g 4.4.  |t Supervised classification block --  |g 4.4.1.  |t Support vector machines --  |g 4.4.1.1.  |t The kernel trick for nonlinearly separable classes --  |g 4.4.1.2.  |t Multi-class classification --  |g 4.4.2.  |t Example. 
505 0 0 |g 4.5.  |t Datasets --  |g 4.5.1.  |t Mixtures of Gaussians --  |g 4.5.2.  |t Datasets from the UCI repository --  |g 4.5.2.1.  |t Iris dataset (Iris) --  |g 4.5.2.2.  |t Wine dataset (wine) --  |g 4.5.2.3.  |t Wisconsin breast cancer dataset (breast) --  |g 4.5.2.4.  |t Handwritten digits dataset (Pendig) --  |g 4.5.2.5.  |t Pima Indians diabetes (diabetes) --  |g 4.5.3.  |t Utterance dataset --  |g 4.6.  |t An analysis of the bounds for the cluster and label approaches --  |g 4.7.  |t Extension through cluster pruning --  |g 4.7.1.  |t Determination of silhouette thresholds --  |g 4.7.2.  |t Evaluation of the cluster pruning approach --  |g 4.8.  |t Simulations and results --  |g 4.9.  |t Summary --  |g pt. 3  |t Contributions to Unsupervised Classification -- Algorithms to Detect the Optimal Number of Clusters --  |g ch. 5  |t Detection of the Number of Clusters through Non-Parametric Clustering Algorithms --  |g 5.1.  |t Introduction --  |g 5.2.  |t New hierarchical pole-based clustering algorithm --  |g 5.2.1.  |t Pole-based clustering basis module --  |g 5.2.2.  |t Hierarchical pole-based clustering --  |g 5.3.  |t Evaluation --  |g 5.3.1.  |t Cluster evaluation metrics --  |g 5.4.  |t Datasets. 
505 0 0 |g 5.4.1.  |t Results --  |g 5.4.2.  |t Complexity considerations for large databases --  |g 5.5.  |t Summary --  |g ch. 6  |t Detecting the Number of Clusters through Cluster Validation --  |g 6.1.  |t Introduction --  |g 6.2.  |t Cluster validation methods --  |g 6.2.1.  |t Dunn index --  |g 6.2.2.  |t Hartigan --  |g 6.2.3.  |t Davies Bouldin index --  |g 6.2.4.  |t Krzanowski and Lai index --  |g 6.2.5.  |t Silhouette --  |g 6.2.6.  |t Hubert's & gamma; --  |g 6.2.7.  |t Gap statistic --  |g 6.3.  |t Combination approach based on quantiles --  |g 6.4.  |t Datasets --  |g 6.4.1.  |t Mixtures of Gaussians --  |g 6.4.2.  |t Cancer DNA-microarray dataset --  |g 6.4.3.  |t Iris dataset --  |g 6.5.  |t Results --  |g 6.5.1.  |t Validation results of the five Gaussian dataset --  |g 6.5.2.  |t Validation results of the mixture of seven Gaussians --  |g 6.5.3.  |t Validation results of the NCI60 dataset --  |g 6.5.4.  |t Validation results of the Iris dataset --  |g 6.5.5.  |t Discussion --  |g 6.6.  |t Application of speech utterances --  |g 6.7.  |t Summary. 
588 0 |a Print version record. 
650 0 |a Data mining. 
650 0 |a Discourse analysis  |x Statistical methods. 
650 0 |a Speech processing systems. 
650 0 |a Computational intelligence. 
650 4 |a Machine learning. 
650 4 |a Computational learning theory. 
650 4 |a Supervised learning (Machine learning) 
650 4 |a Data mining. 
650 4 |a Discourse analysis  |x Statistical methods. 
650 4 |a Speech processing systems. 
650 4 |a Computational intelligence. 
650 7 |a COMPUTERS  |x Enterprise Applications  |x Business Intelligence Tools.  |2 bisacsh 
650 7 |a COMPUTERS  |x Intelligence (AI) & Semantics.  |2 bisacsh 
650 7 |a Computational intelligence.  |2 fast  |0 (OCoLC)fst00871995 
650 7 |a Data mining.  |2 fast  |0 (OCoLC)fst00887946 
650 7 |a Discourse analysis  |x Statistical methods.  |2 fast  |0 (OCoLC)fst00894942 
650 7 |a Speech processing systems.  |2 fast  |0 (OCoLC)fst01129243 
650 7 |a Machine learning.  |2 local 
650 7 |a Computational learning theory.  |2 local 
650 7 |a Supervised learning (Machine learning)  |2 local 
650 7 |a Data mining.  |2 local 
650 7 |a Discourse analysis / Statistical methods.  |2 local 
650 7 |a Speech processing systems.  |2 local 
650 7 |a Computational intelligence.  |2 local 
655 4 |a Electronic books. 
700 1 |a Minker, Wolfgang. 
776 0 8 |i Print version:  |a Albalate, Amparo.  |t Semi-supervised and unsupervised machine learning.  |d London : ISTE ; Hoboken, NJ : Wiley, 2011  |z 9781848212039  |w (DLC) 2010040730  |w (OCoLC)700509842 
856 4 0 |u https://doi.org/10.1002/9781118557693  |z Full Text via HEAL-Link 
994 |a 92  |b DG1