Difference between revisions of "Feature Exploration/Learning"
m |
m |
||
| Line 87: | Line 87: | ||
|| | || | ||
<youtube>V0u6bxQOUJ8</youtube> | <youtube>V0u6bxQOUJ8</youtube> | ||
| − | <b> | + | <b>Pre-Modeling: Data Preprocessing and Feature Exploration in [[Python]] |
| − | </b><br> | + | </b><br>April Chen Data preprocessing and feature exploration are crucial steps in a modeling workflow. In this tutorial, I will demonstrate how to use [[Python]] libraries such as scikit-learn, statsmodels, and matplotlib to perform pre-modeling steps. Topics that will be covered include: missing values, variable types, outlier detection, multicollinearity, interaction terms, and visualizing variable distributions. Finally, I will show the impact of utilizing these techniques on model performance. Interactive [[Jupyter]] notebooks will be provided. |
|} | |} | ||
|<!-- M --> | |<!-- M --> | ||
| Line 95: | Line 95: | ||
|| | || | ||
<youtube>_XOKz5VlTQY</youtube> | <youtube>_XOKz5VlTQY</youtube> | ||
| − | <b> | + | <b>Recent Advances in Feature Selection: A Data Perspective part 1 |
| − | </b><br> | + | </b><br>Authors: Huan Liu, Department of Computer Science and Engineering, Arizona State University Jundong Li, School of Computing, Informatics and Decision Systems Engineering, Arizona State University Jiliang Tang, Department of Computer Science and Engineering, Michigan State University Feature selection, as a data preprocessing strategy, is imperative in preparing high-dimensional data for myriad of data mining and machine learning tasks. By selecting a subset of features of high quality, feature selection can help build simpler and more comprehensive models, improve data mining performance, and prepare clean and understandable data. The proliferation of big data in recent years has presented substantial challenges and opportunities for feature selection research. In this tutorial, we provide a comprehensive overview of recent advances in feature selection research from a data perspective. After we introduce some basic concepts, we review state-of-the-art feature selection algorithms and recent techniques of feature selection for structured, social, heterogeneous, and streaming data. In particular, we also discuss what the role of feature selection is in the context of deep learning and how feature selection is related to feature engineering. To facilitate and promote the research in this community, we present an open-source feature selection repository scikit-feature that consists of most of the popular feature selection algorithms. We conclude our discussion with some open problems and pressing issues in future research. |
|} | |} | ||
|}<!-- B --> | |}<!-- B --> | ||
| Line 104: | Line 104: | ||
|| | || | ||
<youtube>arhdVDsPLVI</youtube> | <youtube>arhdVDsPLVI</youtube> | ||
| − | <b> | + | <b>Alexandru Agachi - Introductory tutorial on data exploration and statistical models |
| − | </b><br> | + | </b><br>This tutorial will focus on analyzing a dataset and building statistical models from it. We will describe and visualize the data. We will then build and analyze statistical models, including linear and logistic regression, as well as chi-square tests of independence. We will then apply 4 machine learning techniques to the dataset: decision trees, random forests, lasso regression, and clustering. I would be happy to conduct an introductory level tutorial on exploring a dataset with the pandas/StatsModels/scikit-learn framework: 1. Descriptive statistics. Here we will describe each variable depending on its type, as well as the dataset overall. 2. Visualization for categorical and quantitative variables. We will learn effective visualization techniques for each type of variable in the dataset. 3. Statistical modeling for quantitative and categorical, explanatory and response variables: chi-square tests of independence, linear regression and logistic regression. We will learn to test hypotheses, and to interpret our models, their strengths, and their limitations. 4. I will then expand to the application of machine learning techniques, including decision trees, random forests, lasso regression, and clustering. Here we will explore the advantages and disadvantages of each of these techniques, as well as apply them to the dataset. This would be a very applied, introductory tutorial, to the statistical exploration of a dataset and the building of statistical models from it. I would be happy to send you the ipython notebook for this tutorial as well. www.pydata.org |
|} | |} | ||
|<!-- M --> | |<!-- M --> | ||
| Line 112: | Line 112: | ||
|| | || | ||
<youtube>TsqTuwTKFSs</youtube> | <youtube>TsqTuwTKFSs</youtube> | ||
| − | <b> | + | <b>Feature Selection in Machine learning| Variable selection| Dimension Reduction |
| − | </b><br> | + | </b><br>Feature selection is an important step in machine learning model building process. The performance of models depends in the following : Choice of algorithm Feature Selection |
|} | |} | ||
|}<!-- B --> | |}<!-- B --> | ||
| Line 121: | Line 121: | ||
|| | || | ||
<youtube>YaKMeAlHgqQ</youtube> | <youtube>YaKMeAlHgqQ</youtube> | ||
| − | <b> | + | <b>How do I select features for Machine Learning? |
| − | </b><br> | + | </b><br>Selecting the "best" features for your Machine Learning model will result in a better performing, easier to understand, and faster running model. But how do you know which features to select? In this video, I'll discuss 7 feature selection tactics used by the pros that you can apply to your own model. At the end, I'll give you my top 3 tips for effective feature selection. |
|} | |} | ||
|<!-- M --> | |<!-- M --> | ||
| Line 128: | Line 128: | ||
{| class="wikitable" style="width: 550px;" | {| class="wikitable" style="width: 550px;" | ||
|| | || | ||
| − | <youtube> | + | <youtube>ZKaOfJIjMRg</youtube> |
| − | <b> | + | <b>Lecture 15.6 — Anomaly Detection | Choosing What Features To Use — [[Creatives#Andrew Ng|Andrew Ng]] |
| − | </b><br> | + | </b><br>Artificial Intelligence - All in One |
|} | |} | ||
|}<!-- B --> | |}<!-- B --> | ||
Revision as of 21:00, 21 September 2020
YouTube search... ...Google search
- Feature selection | Wikipedia
- Notes on Feature Preprocessing: The What, the Why, and the How | Matthew Mayo - KDnuggets
- Evaluating Machine Learning Models
- Automated Machine Learning (AML) - AutoML
- Recursive Feature Elimination (RFE)
- Principal Component Analysis (PCA)
- Representation Learning
- Feature Engineering and Selection: A Practical Approach for Predictive Models | Max Kuhn and Kjell Johnson
- Jon Tupitza's Famous Jupyter Notebooks:
- AI Governance
- Visualization
- Tools: Paxata, Trifacta, alteryx, databricks, Qubole
A feature is an individual measurable property or characteristic of a phenomenon being observed. The concept of a “feature” is related to that of an explanatory variable, which is used in statistical techniques such as linear regression. Feature vectors combine all of the features for a single row into a numerical vector. Part of the art of choosing features is to pick a minimum set of independent variables that explain the problem. If two variables are highly correlated, either they need to be combined into a single feature, or one should be dropped. Sometimes people perform principal component analysis to convert correlated variables into a set of linearly uncorrelated variables. Some of the transformations that people use to construct new features or reduce the dimensionality of feature vectors are simple. For example, subtract Year of Birth from Year of Death and you construct Age at Death, which is a prime independent variable for lifetime and mortality analysis. In other cases, feature construction may not be so obvious. Machine learning algorithms explained | Martin Heller - InfoWorld
|
|
|
|
Feature Selection
YouTube search... ...Google search
|
|
|
|
|
|
Sparse Coding - Feature Extraction
|
|