Difference between revisions of "Hugging Face"
m |
m |
||
Line 18: | Line 18: | ||
Hugging Face is an American company that develops tools for building applications using machine learning. It is most notable for its transformers library built for natural language processing applications and its platform that allows users to share machine learning models and datasets. Hugging Face is a community and a platform for artificial intelligence and data science that aims to democratize AI knowledge and assets used in AI models. The platform allows users to build, train and deploy state of the art models powered by open source machine learning. It also provides a place where a broad community of data scientists, researchers, and ML engineers can come together and share ideas, get support and contribute to open source projects. Is there anything else you would like to know? - [https://en.wikipedia.org/wiki/Hugging_Face Wikipedia] | Hugging Face is an American company that develops tools for building applications using machine learning. It is most notable for its transformers library built for natural language processing applications and its platform that allows users to share machine learning models and datasets. Hugging Face is a community and a platform for artificial intelligence and data science that aims to democratize AI knowledge and assets used in AI models. The platform allows users to build, train and deploy state of the art models powered by open source machine learning. It also provides a place where a broad community of data scientists, researchers, and ML engineers can come together and share ideas, get support and contribute to open source projects. Is there anything else you would like to know? - [https://en.wikipedia.org/wiki/Hugging_Face Wikipedia] | ||
− | |||
− | |||
<youtube>agwbNgxwkHc</youtube> | <youtube>agwbNgxwkHc</youtube> | ||
<youtube>QEaBAZQCtwE</youtube> | <youtube>QEaBAZQCtwE</youtube> | ||
+ | = <span id="Hugging Face's Research"></span>Hugging Face's Research = | ||
+ | <youtube>eqOSQeQNqaw</youtube> | ||
− | + | = Hugging Face Community = | |
* [https://towardsdatascience.com/whats-hugging-face-122f4e7eb11a What's Hugging Face? An AI community for sharing ML models and datasets] | * [https://towardsdatascience.com/whats-hugging-face-122f4e7eb11a What's Hugging Face? An AI community for sharing ML models and datasets] | ||
* [[Agents#HuggingGPT|HuggingGPT]] ... in partnership with [[Microsoft]] | * [[Agents#HuggingGPT|HuggingGPT]] ... in partnership with [[Microsoft]] | ||
Line 37: | Line 37: | ||
Their platform is home to a large community of developers and researchers who work together to solve problems in audio, vision, and language with AI. | Their platform is home to a large community of developers and researchers who work together to solve problems in audio, vision, and language with AI. | ||
− | + | == <span id="Hugging Face's Open-source Library"></span>Hugging Face's Open-source Library == | |
Hugging Face's open-source library, Transformers, is widely used for natural language processing tasks. The company also offers an Inference API that allows developers to serve their models directly from Hugging Face infrastructure and run large scale NLP models in milliseconds with just a few lines of code. Hugging Face offers a wide range of machine learning models and datasets, as well as tools for building, training, and deploying state-of-the-art models. | Hugging Face's open-source library, Transformers, is widely used for natural language processing tasks. The company also offers an Inference API that allows developers to serve their models directly from Hugging Face infrastructure and run large scale NLP models in milliseconds with just a few lines of code. Hugging Face offers a wide range of machine learning models and datasets, as well as tools for building, training, and deploying state-of-the-art models. | ||
− | + | == <span id="Private Hub"></span>Private Hub == | |
* [https://huggingface.co/platform Private Hub] | * [https://huggingface.co/platform Private Hub] | ||
** [https://huggingface.co/docs/hub/main Hugging Face Hub documentation] | ** [https://huggingface.co/docs/hub/main Hugging Face Hub documentation] | ||
− | + | == <span id="LightGPT"></span>LightGPT == | |
* [https://huggingface.co/amazon/LightGPT amazon/LightGPT] | * [https://huggingface.co/amazon/LightGPT amazon/LightGPT] | ||
* [https://huggingface.co/amazon/LightGPT/blob/main/README.md README.md · amazon/LightGPT] | * [https://huggingface.co/amazon/LightGPT/blob/main/README.md README.md · amazon/LightGPT] | ||
Line 56: | Line 56: | ||
GPT-J 6B is a transformer model trained using Ben Wang's Mesh Transformer JAX. "GPT-J" refers to the class of model, while "6B" represents the number of trainable parameters. The model consists of 28 layers with a model dimension of 4096, and a feedforward dimension of 16384. The model dimension is split into 16 heads, each with a dimension of 256. Rotary Position Embedding (RoPE) is applied to 64 dimensions of each head. The model is trained with a tokenization vocabulary of 50257, using the same set of BPEs as GPT-2/GPT-3. GPT-J learns an inner representation of the English language that can be used to extract features useful for downstream tasks. The model is best at what it was pretrained for however, which is generating text from a prompt. GPT-J-6B is not intended for deployment without fine-tuning, supervision, and/or moderation. It is not a product in itself and cannot be used for human-facing interactions. For example, the model may generate harmful or offensive text. Please evaluate the risks associated with your particular use case. | GPT-J 6B is a transformer model trained using Ben Wang's Mesh Transformer JAX. "GPT-J" refers to the class of model, while "6B" represents the number of trainable parameters. The model consists of 28 layers with a model dimension of 4096, and a feedforward dimension of 16384. The model dimension is split into 16 heads, each with a dimension of 256. Rotary Position Embedding (RoPE) is applied to 64 dimensions of each head. The model is trained with a tokenization vocabulary of 50257, using the same set of BPEs as GPT-2/GPT-3. GPT-J learns an inner representation of the English language that can be used to extract features useful for downstream tasks. The model is best at what it was pretrained for however, which is generating text from a prompt. GPT-J-6B is not intended for deployment without fine-tuning, supervision, and/or moderation. It is not a product in itself and cannot be used for human-facing interactions. For example, the model may generate harmful or offensive text. Please evaluate the risks associated with your particular use case. | ||
− | + | == <span id="Whisper"></span>Whisper == | |
<youtube>8xYYvO7LGBw</youtube> | <youtube>8xYYvO7LGBw</youtube> |
Revision as of 05:34, 15 June 2023
YouTube ... Quora ...Google search ...Google News ...Bing News
- Hugging Face ... The AI community building the future
- Platforms: AI/Machine Learning as a Service (AIaaS/MLaaS)
- Hugging Face course
- What is Hugging Face - A Beginner's Guide | ByteXD ... allows users to share machine learning models and datasets
Hugging Face is an American company that develops tools for building applications using machine learning. It is most notable for its transformers library built for natural language processing applications and its platform that allows users to share machine learning models and datasets. Hugging Face is a community and a platform for artificial intelligence and data science that aims to democratize AI knowledge and assets used in AI models. The platform allows users to build, train and deploy state of the art models powered by open source machine learning. It also provides a place where a broad community of data scientists, researchers, and ML engineers can come together and share ideas, get support and contribute to open source projects. Is there anything else you would like to know? - Wikipedia
Contents
Hugging Face's Research
Hugging Face Community
- What's Hugging Face? An AI community for sharing ML models and datasets
- HuggingGPT ... in partnership with Microsoft
- Pretrain Transformers Models in PyTorch Using Hugging Face Transformers | George Mihaila - TOPBOTS
- OpenChatKit | TogetherCompute ... The first open-source ChatGPT alternative released; a 20B chat-GPT model under the Apache-2.0 license, which is available for free on Hugging Face.
- LAION
- Ontocord
- Wolfram ChatGPT
- FLAN-T5 LLM
Their platform is home to a large community of developers and researchers who work together to solve problems in audio, vision, and language with AI.
Hugging Face's Open-source Library
Hugging Face's open-source library, Transformers, is widely used for natural language processing tasks. The company also offers an Inference API that allows developers to serve their models directly from Hugging Face infrastructure and run large scale NLP models in milliseconds with just a few lines of code. Hugging Face offers a wide range of machine learning models and datasets, as well as tools for building, training, and deploying state-of-the-art models.
Private Hub
LightGPT
- amazon/LightGPT
- README.md · amazon/LightGPT
- EleutherAI/gpt-j-6b
- GPT-J | Wikipedia
- Deploy GPT-J 6B for inference using Hugging Face Transformers
- Fine-tuning GPT-J 6B on Google Colab or Equivalent Desktop or Server
LightGPT is a language model developed by AWS Contributors. It is based on GPT-J 6B and was instruction fine-tuned on the high-quality, Apache-2.0 licensed OIG-small-chip instruction dataset with ~200K training examples. The model is designed to generate text based on a given instruction, and it can be deployed to Amazon SageMaker
GPT-J 6B is a transformer model trained using Ben Wang's Mesh Transformer JAX. "GPT-J" refers to the class of model, while "6B" represents the number of trainable parameters. The model consists of 28 layers with a model dimension of 4096, and a feedforward dimension of 16384. The model dimension is split into 16 heads, each with a dimension of 256. Rotary Position Embedding (RoPE) is applied to 64 dimensions of each head. The model is trained with a tokenization vocabulary of 50257, using the same set of BPEs as GPT-2/GPT-3. GPT-J learns an inner representation of the English language that can be used to extract features useful for downstream tasks. The model is best at what it was pretrained for however, which is generating text from a prompt. GPT-J-6B is not intended for deployment without fine-tuning, supervision, and/or moderation. It is not a product in itself and cannot be used for human-facing interactions. For example, the model may generate harmful or offensive text. Please evaluate the risks associated with your particular use case.
Whisper