Difference between revisions of "Feature Exploration/Learning"
m |
m |
||
| Line 198: | Line 198: | ||
{| class="wikitable" style="width: 550px;" | {| class="wikitable" style="width: 550px;" | ||
|| | || | ||
| − | + | ||
| − | |||
| − | |||
|} | |} | ||
|}<!-- B --> | |}<!-- B --> | ||
Revision as of 20:47, 18 September 2023
YouTube ... Quora ...Google search ...Google News ...Bing News
- Data Science ... Governance ... Preprocessing ... Exploration ... Interoperability ... Master Data Management (MDM) ... Bias and Variances ... Benchmarks ... Datasets
- Data Quality ...validity, accuracy, cleaning, completeness, consistency, encoding, padding, augmentation, labeling, auto-tagging, normalization, standardization, and imbalanced data
- Evaluating Machine Learning Models
- Artificial General Intelligence (AGI) to Singularity ... Curious Reasoning ... Emergence ... Moonshots ... Explainable AI ... Automated Learning
- Recursive Feature Elimination (RFE)
- Principal Component Analysis (PCA)
- Representation Learning
- Managed Vocabularies
- Excel ... Documents ... Database; Vector & Relational ... Graph ... LlamaIndex
- Analytics ... Visualization ... Graphical Tools ... Diagrams & Business Analysis ... Requirements ... Loop ... Bayes ... Network Pattern
- Development ... Notebooks ... AI Pair Programming ... Codeless, Generators, Drag n' Drop ... AIOps/MLOps ... AIaaS/MLaaS
- Feature selection | Wikipedia
- Notes on Feature Preprocessing: The What, the Why, and the How | Matthew Mayo - KDnuggets
- Feature Engineering and Selection: A Practical Approach for Predictive Models | Max Kuhn and Kjell Johnson
- Jon Tupitza's Famous Jupyter Notebooks:
- AI Governance / Algorithm Administration
- Tools:
A feature is an individual measurable property or characteristic of a phenomenon being observed. The concept of a “feature” is related to that of an explanatory variable, which is used in statistical techniques such as linear regression. Feature vectors combine all of the features for a single row into a numerical vector. Part of the art of choosing features is to pick a minimum set of independent variables that explain the problem. If two variables are highly correlated, either they need to be combined into a single feature, or one should be dropped. Sometimes people perform principal component analysis to convert correlated variables into a set of linearly uncorrelated variables. Some of the transformations that people use to construct new features or reduce the dimensionality of feature vectors are simple. For example, subtract Year of Birth from Year of Death and you construct Age at Death, which is a prime independent variable for lifetime and mortality analysis. In other cases, feature construction may not be so obvious. Machine learning algorithms explained | Martin Heller - InfoWorld
Contents
Feature Examples
For example, if you were building a machine learning model to predict whether someone would like a particular movie, you might use features like the person's age, gender, and favorite genres of movies. You might also use features about the movie itself, such as the genre, director, and rating. Features are important because they allow machine learning models to learn about the world. By providing models with features, we can teach them to identify patterns and make predictions.
Here is an example of a feature in AI that a 7th grader might understand:
Imagine you are building a machine learning model to predict whether a student will pass or fail a math test. You might use the following features:
- The student's grades on previous math tests
- The student's attendance record in math class
- The student's homework completion rate
- The student's score on the math portion of the standardized test
Your machine learning model would learn to identify patterns in this data. For example, the model might learn that students who have high grades on previous math tests and good attendance are more likely to pass the test. The model could also learn that students who miss a lot of class or have incomplete homework are more likely to fail the test. Once your machine learning model is trained, you can use it to predict whether a new student is likely to pass or fail the math test. You can do this by providing the model with the student's features, such as their grades on previous math tests and their attendance record. The model will then use this information to make a prediction.
Feature Store
A feature store in AI is a system for managing and serving features to machine learning models. Features are measurable pieces of data that can be used to train and evaluate models. Feature stores provide a central repository for features, making them easier to discover, reuse, and manage. Feature stores are important because they can help to improve the quality, efficiency, and scalability of machine learning development and deployment. For example, feature stores can help to:
- Reduce the time and effort required to develop and maintain machine learning models
- Improve the performance and accuracy of machine learning models
- Make machine learning models more reproducible and scalable
- Ensure that machine learning models are using consistent and up-to-date data
- FEAST: is a feature store that is open-source and cloud-native. It is designed to be scalable and performant, and it can be deployed on a variety of cloud platforms. FEAST provides a number of features, including:
- A central repository for features, with support for versioning and lineage tracking
- A variety of data processing and transformation capabilities
- A low-latency serving API for online inference
- A batch API for offline training and batch predictions
- FeatureBase: is a commercial feature store that is offered by Google Cloud. It is designed to be easy to use and manage, and it provides a number of features, including:
- A central repository for features, with support for versioning and lineage tracking
- Automatic feature discovery and generation
- A low-latency serving API for online inference
- A batch API for offline training and batch predictions
- feathr: is a commercial feature store that is offered by Amazon Web Services (AWS). It is designed to be scalable and performant, and it provides a number of features, including:
- A central repository for features, with support for versioning and lineage tracking
- Automatic feature generation and transformation
- A low-latency serving API for online inference
- A batch API for offline training and batch predictions
How each fits:
- FEAST is a good choice for organizations that want a flexible and scalable feature store that they can deploy and manage on their own.
- FeatureBase is a good choice for organizations that want a feature store that is easy to use and manage, and that provides integration with Google Cloud Platform services.
- feathr is a good choice for organizations that want a feature store that is scalable and performant, and that provides integration with Amazon Web Services (AWS) services.
|
|
|
|
Feature Selection
YouTube search... ...Google search
- Beginner's Guide to Feature Selection in Python | Sayak Paul ...Learn about the basics of feature selection and how to implement and investigate various feature selection techniques in Python
- Feature Selection For Machine Learning in Python | Jason Brownlee
- How to Perform Feature Selection with Categorical Data | Jason Brownlee
|
|
|
|
|
|
Sparse Coding - Feature Extraction
|
|