Difference between revisions of "Memory"
m |
m |
||
| Line 20: | Line 20: | ||
[https://www.bing.com/news/search?q=memory+ai&qft=interval%3d%228%22 ...Bing News] | [https://www.bing.com/news/search?q=memory+ai&qft=interval%3d%228%22 ...Bing News] | ||
| + | * [[Memory Networks]] | ||
| + | * [[State Space Model (SSM)]] ... [[Mamba]] ... [[Sequence to Sequence (Seq2Seq)]] ... [[Recurrent Neural Network (RNN)]] ... [[(Deep) Convolutional Neural Network (DCNN/CNN)|Convolutional Neural Network (CNN)]] | ||
| + | * [[Hierarchical Temporal Memory (HTM)]] | ||
| + | * [[Recurrent Neural Network (RNN)]] Variants: | ||
| + | ** [[Long Short-Term Memory (LSTM)]] | ||
| + | ** [[Manhattan LSTM (MaLSTM)]] — a Siamese architecture based on recurrent neural network | ||
| + | ** [[Gated Recurrent Unit (GRU)]] | ||
| + | ** [[Bidirectional Long Short-Term Memory (BI-LSTM)]] | ||
| + | ** [[Bidirectional Long Short-Term Memory (BI-LSTM) with Attention Mechanism]] | ||
| + | ** [[Average-Stochastic Gradient Descent (SGD) Weight-Dropped LSTM (AWD-LSTM)]] | ||
| + | ** [[Hopfield Network (HN)]] | ||
| + | * [[Lifelong Learning]] | ||
| + | * [[Decentralized: Federated & Distributed]] | ||
* [[Assistants]] ... [[Personal Companions]] ... [[Agents]] ... [[Negotiation]] ... [[LangChain]] | * [[Assistants]] ... [[Personal Companions]] ... [[Agents]] ... [[Negotiation]] ... [[LangChain]] | ||
| − | + | * [[Context]] | |
| − | |||
| − | |||
| − | * [[ | ||
| − | |||
= Catastrophic Forgetting and Mitigation Strategies = | = Catastrophic Forgetting and Mitigation Strategies = | ||
Revision as of 22:07, 1 March 2024
Youtube ... Quora ...Google search ...Google News ...Bing News
- Memory Networks
- State Space Model (SSM) ... Mamba ... Sequence to Sequence (Seq2Seq) ... Recurrent Neural Network (RNN) ... Convolutional Neural Network (CNN)
- Hierarchical Temporal Memory (HTM)
- Recurrent Neural Network (RNN) Variants:
- Long Short-Term Memory (LSTM)
- Manhattan LSTM (MaLSTM) — a Siamese architecture based on recurrent neural network
- Gated Recurrent Unit (GRU)
- Bidirectional Long Short-Term Memory (BI-LSTM)
- Bidirectional Long Short-Term Memory (BI-LSTM) with Attention Mechanism
- Average-Stochastic Gradient Descent (SGD) Weight-Dropped LSTM (AWD-LSTM)
- Hopfield Network (HN)
- Lifelong Learning
- Decentralized: Federated & Distributed
- Assistants ... Personal Companions ... Agents ... Negotiation ... LangChain
- Context
Contents
Catastrophic Forgetting and Mitigation Strategies
Catastrophic forgetting is a significant challenge in AI, where neural networks overwrite old information when learning new data, akin to digital amnesia. This issue is particularly problematic for autonomous systems operating in dynamic environments, as it limits their ability to acquire new competencies over time. To address this, researchers have developed various techniques:
- Regularization and Weight Consolidation: Methods like Elastic Weight Consolidation (EWC) and Synaptic Intelligence (SI) aim to preserve important weight parameters and minimize changes to critical weights during new learning.
- Replay Methods: These involve retraining neural networks on old datasets to refresh memories, with Memory Replay using subsets of old data and Generative Replay employing generative models to create synthetic samples.
- Dynamic Networks: Instead of combating forgetting within fixed structures, dynamic networks expand their architecture to accommodate new tasks, such as Progressive Neural Networks and Expert Gate Modules.
Despite these efforts, catastrophic forgetting remains a significant obstacle, necessitating ongoing research to enhance AI's memory capacity and learning abilities.
Controlled Forgetting and Trustworthy AI
Controlled forgetting in AI is an emerging field focusing on enabling AI systems to forget specific data efficiently without complete retraining. This is crucial for creating robust AI systems that can adaptively manage their knowledge and comply with privacy regulations like the "right to be forgotten" under GDPR. The Neuralyzer algorithm is an example of a technique that adjusts the logits or raw prediction scores generated by the model to facilitate controlled forgetting.
Sleep and Memory Consolidation in AI
Research has shown that incorporating sleep-like phases in neural networks can help overcome catastrophic forgetting, drawing inspiration from the human brain's ability to consolidate memory during sleep. This approach has been detailed in scientific publications and is considered a promising direction for future AI memory research.
Forgetting as a Feature in AI
Simulating human forgetting is gaining attention in AI research, as it can help manage computational resources by prioritizing relevant data and discarding less useful information. Techniques like neural network pruning and regularization, such as dropout, are forms of induced forgetting that help AI models adapt to new information without being overwhelmed. Advanced AI systems that evolve and self-modify their rules are also exploring mechanisms of 'forgetting' less effective strategies.
Memory Enhancements in AI Products
OpenAI's ChatGPT is an example of a product incorporating memory to remember user-specific information and preferences over time. This feature allows for a more personalized interaction, with mechanisms in place to avoid retaining sensitive information. Users can also opt for a temporary chat mode for conversations that won't affect the AI's memory of them.
Memory Storage and State Management
The memory market is experiencing a resurgence, driven by the demand for server memory, especially for AI servers, which necessitates DDR and high bandwidth memory (HBM). Cloud service providers are customizing chips to optimize costs and energy efficiency, which is pivotal for the semiconductor industry's trajectory.
Impact on the Field
The latest research and products in memory AI are reshaping the field by addressing the challenges of catastrophic forgetting and controlled forgetting. These advancements are crucial for the development of AI systems capable of lifelong learning, trustworthy AI, and personalized user experiences. The semiconductor industry is also adapting to these changes, with a focus on memory enhancements to support the growing needs of AI servers and applications.