Data Mining for Business Intelligence, Second Edition uses real data and actual cases to illustrate the applicability of data mining (DM) intelligence in the development of successful business models. Featuring complimentary access to XLMiner, the Microsoft Office Excel add-in, this book allows readers to follow along and implement algorithms at their own speed, with a minimal learning curve. In addition, students and practitioners of DM techniques are presented with hands-on, business-oriented applications. An abundant amount of exercises and examples, now doubled in number in the second edition, are provided to motivate learning and understanding. This book helps readers understand the beneficial relationship that can be established between DM and smart business practices, and is an excellent learning tool for creating valuable strategies and making wiser business decisions. New topics include detailed coverage of visualization (enhanced by Spotfire subroutines) and time series forecasting, among a host of other subject matter.
Foreword by Daryl Pregibon (Google, Inc.).
PART I PRELIMINARIES.
1.1 What Is Data Mining?
1.2 Where Is Data Mining Used?
1.3 The Origins of Data Mining.
1.4 The Rapid Growth of Data Mining.
1.5 Why Are There So Many Different Methods?
1.6 Terminology and Notation.
1.7 Road Maps to This Book.
2 Overview of the Data Mining Process.
2.2 Core Ideas in Data Mining.
2.3 Supervised and Unsupervised Learning.
2.4 The Steps in Data Mining.
2.5 Preliminary Steps.
2.6 Building a Model: Example with Linear Regression.
2.7 Using Excel for Data Mining.
PART II DATA EXPLORATION AND DIMENSION REDUCTION.
3 Data Visualization.
3.1 Uses of Data Visualization.
3.2 Data Examples.
Example 1: Boston Housing Data.
Example 2: Ridership on Amtrak Trains.
3.3 Basic Charts: bar charts, line graphs, and scatterplots.
Heatmaps: visualizing correlations and missing values.
3.4 MultiDimensional Visualization.
Adding Variables: color, hue, size, shape, multiple panels, animation.
Manipulations: rescaling,aggregation and hierarchies, zooming and panning, filtering.
Reference: trend line and labels.
Scaling up: large datasets.
Multivariate plot: parallel coordinates plot.
3.5 Specialized Visualizations.
Visualizing networked data.
Visualizing hierarchical data: treemaps.
Visualizing geographical data: maps.
3.6 Summary of major visualizations and operations, according to data mining goal.
Time series forecasting.
4 Dimension Reduction.
4.2 Practical Considerations.
Example 1: House Prices in Boston.
4.3 Data Summaries.
4.4 Correlation Analysis.
4.5 Reducing the Number of Categories in Categorical Variables.
4.6 Converting A Categorical Variable to A Numerical Variable.
4.7 Principal Components Analysis.
Example 2: Breakfast Cereals.
Normalizing the Data.
Using Principal Components for Classification and Prediction.
4.8 Dimension Reduction Using Regression Models.
4.9 Dimension Reduction Using Classification and Regression Trees.
PART III PERFORMANCE EVALUATION.
5 Evaluating Classification and Predictive Performance.
5.2 Judging Classification Performance.
Benchmark: The Naive Rule.
The Classification Matrix.
Using the Validation Data.
Cutoff for Classification.
Performance in Unequal Importance of Classes.
Asymmetric Misclassification Costs.
Oversampling and Asymmetric Costs.
Classification Using a Triage Strategy.
5.3 Evaluating Predictive Performance.
Benchmark: The Average.
Prediction Accuracy Measures.
PART IV PREDICTION AND CLASSIFICATION METHODS.
6 Multiple Linear Regression.
6.2 Explanatory vs. Predictive Modeling.
6.3 Estimating the Regression Equation and Prediction.
Example: Predicting the Price of Used Toyota Corolla Automobiles.
6.4 Variable Selection in Linear Regression.
Reducing the Number of Predictors.
How to Reduce the Number of Predictors.
7.1 The kNN.
Classifier (categorical outcome).
Example: Riding Mowers.
Setting the Cutoff Value.
With More Than 2 Classes.
for a Numerical Response.
7.3 Advantages and Shortcomings of kNN.
8 Naive Bayes.
Example 1: Predicting Fraudulent Financial Reporting.
The Practical Difficulty with the Complete (Exact) Bayes Procedure.
The Solution: Na*ve Bayes.
Example 2: Predicting Fraudulent Financial Reports, 2 Predictors.
Example 3: Predicting Delayed Flights.
Advantages and Shortcomings of the naive Bayes Classifier.
9 Classification and Regression Trees.
9.2 Classification Trees.
Example 1: Riding Mowers.
9.3 Measures of Impurity.
9.4 Evaluating the Performance of a Classification Tree.
Example 2: Acceptance of Personal Loan.
9.5 Avoiding Overfitting.
Stopping Tree Growth: CHAID.
Pruning the Tree.
9.6 Classification Rules from Trees.
9.7 Classification Trees for More Than 2 Classes.
9.8 Regression Trees.
9.9 Advantages, Weaknesses, and Extensions.
10 Logistic Regression.
10.2 The Logistic Regression Model.
Example: Acceptance of Personal Loan.
Model with a Single Predictor.
Estimating the Logistic Model from Data: Computing Parameter
Interpreting Results in Terms of Odds.
10.3 Evaluating Classification Performance.
10.4 Example of Complete Analysis: Predicting Delayed Flights.
Model Fitting and Estimation.
Appendix: Logistic Regression for Profiling.
Appendix: Logistic regression for profiling.
Appendix: B: Evaluating Goodness of Fit.
Appendix B: Evaluating Goodness of Fit.
Appendix: C: Logistic Regression for More Than Two Classes.
Appendix C: Logistic Regression for More Than Two Classes.
11 Neural Nets.
11.2 Concept and Structure of a Neural Network.
11.3 Fitting a Network to Data.
Example 1: Tiny Dataset.
Computing Output of Nodes.
Preprocessing the Data.
Training the Model.
Example 2: Classifying Accident Severity.
Using the Output for Prediction and Classification.
11.4 Required User Input.
11.5 Exploring the Relationship Between Predictors and Response.
11.6 Advantages and Weaknesses of Neural Networks.
12 Discriminant Analysis.
Example 1: Riding Mowers.
Example 2: Personal Loan Acceptance.
12.2 Distance of an Observation from a Class.
12.3 Fisher's Linear Classification Functions.
12.4 Classification Performance of Discriminant Analysis.
12.5 Prior Probabilities.
12.6 Unequal Misclassification Costs.
12.7 Classifying More Than Two Classes.
Example 3: Medical Dispatch to Accident Scenes.
12.8 Advantages and Weaknesses.
PART V MINING RELATIONSHIPS AMONG RECORDS.
13 Association Rules.
13.2 Discovering Association Rules in Transaction Databases.
Example 1: Synthetic Data on Purchases of Phone Faceplates.
13.3 Generating Candidate Rules.
The Apriori Algorithm.
13.4 Selecting Strong Rules.
Support and Confidence.
The Process of Rule Selection.
Interpreting the Results.
Statistical Significance of Rules.
Example 2: Rules for Similar Book Purchases.
14 Cluster Analysis.
Example: Public Utilities.
14.2 Measuring Distance Between Two Records.
Normalizing Numerical Measurements.
Other Distance Measures for Numerical Data.
Distance Measures for Categorical Data.
Distance Measures for Mixed Data.
14.3 Measuring Distance Between Two Clusters.
14.4 Hierarchical (Agglomerative) Clustering.
Minimum Distance (Single Linkage).
Maximum Distance (Complete Linkage).
Average Distance (Average Linkage).
Dendrograms: Displaying Clustering Process and Results.
Limitations of Hierarchical Clustering.
14.5 Nonhierarchical Clustering: The kMeans Algorithm.
Initial Partition into k Clusters.
PART VI FORECASTING TIME SERIES.
15 Handling Time Series.
15.2 Explanatory vs. Predictive Modeling.
15.3 Popular Forecasting Methods in Business.
15.4 Time Series Components.
Example: Ridership on Amtrak Trains.
15.5 Data Partitioning.
16 Regression Based Forecasting.
16.1 A Model with Trend.
16.2 A Model with Seasonality.
16.3 A model with trend and seasonality.
16.4 Autocorrelation and ARIMA Models.
16.4.1 Computing Autocorrelation.
16.4.2 Improving Forecasts by Integrating Autocorrelation Information.
Improving Forecasts by Integrating Autocorrelation Information.
16.4.3 Evaluating Predictability.
17 Smoothing Methods.
17.2 Moving Average.
Centered Moving Average for Visualization.
Trailing Moving Average for Forecasting.
Choosing Window Width.
17.3 Simple Exponential Smoothing.
Choosing Smoothing Parameter.
Relation Between Moving Average and Simple Exponential
17.4 Advanced Exponential Smoothing.
Series with a trend.
Series with a trend and seasonality.
Series with seasonality (no trend).
PART VII CASES.
18.1 Charles Book Club.
18.2 German Credit.
18.3 Tayko Software Cataloger.
18.4 Segmenting Consumers of Bath Soap.
18.5 DirectMail Fundraising.
18.6 Catalog CrossSelling.
18.7 Predicting Bankruptcy.
18.8 Time Series Case: Forecasting Public Transportation Demand.