


The star mark exceeded 100,000! After Auto-GPT, Transformer reaches new milestone
In 2017, the Google team proposed the groundbreaking NLP architecture Transformer in the paper "Attention Is All You Need" and has been cheating ever since.
Over the years, this architecture has been popular with large technology companies such as Microsoft, Google, and Meta. Even ChatGPT, which has swept the world, was developed based on Transformer.
And just today, Transformer’s star rating on GitHub exceeded 100,000!
Hugging Face, which started out as a chatbot program, rose to fame as the centerpiece of the Transformer model, becoming The world-famous open source community.
To celebrate this milestone, Hugging Face also summarized 100 projects based on the Transformer architecture.
Transformer detonated the machine learning circle
In June 2017, when Google released the "Attention Is All You Need" paper, perhaps no one thought of this deep learning architecture Transformer How many surprises can it bring.
Since its birth, Transformer has become the cornerstone king in the AI field. In 2019, Google also applied for a patent specifically for it.
As Transformer occupies a mainstream position in the NLP field, it has also begun to cross-border into other fields, and more and more work has begun. Try to steer it into the CV realm.
Many netizens were very excited to see Transformer break through this milestone.
"I have been a contributor to many popular open source projects, but seeing Transformer reach 10 on GitHub Ten thousand stars, it’s still very special!”
Some time ago, the number of GitHub stars of Auto-GPT exceeded that of pytorch. caused a big stir.
Netizens can’t help but wonder how Auto-GPT compares with Transformer?
In fact, Auto-GPT far surpasses Transformer and already has 130,000 stars.
Currently, Tensorflow has more than 170,000 stars. It can be seen that Transformer is the third machine learning library with a star rating of over 100,000 after these two projects.
Some netizens recalled that when they first used the Transformers library, it was called "pytorch-pretrained-BERT".
50 Awesome Projects Based on Transformers
Transformers is not only a toolkit that uses pre-trained models, it is also a project built around Transformers and Hugging Face Hub Community.
In the following list, Hugging Face summarizes 100 amazing and novel projects based on Transformer.
Below, we have selected the first 50 projects for introduction:
gpt4all
gpt4all is an open source chatbot ecosystem. It is trained on a large collection of clean assistant data, including code, stories, and conversations. It provides open source large-scale language models, such as LLaMA and GPT-J, for training in an assistant manner.
Keywords: open source, LLaMa, GPT-J, instructions, assistant
This repository contains examples and best practices for building recommender systems, provided in the form of Jupiter notebooks. It covers several aspects needed to build an effective recommendation system: data preparation, modeling, evaluation, model selection and optimization, and operationalization.
Keywords: recommendation system, AzureML
lama-cleanerImage repair tool based on Stable Diffusion technology. You can erase any unwanted objects, defects, or even people from the image and replace anything on the image.
Keywords: patch, SD, Stable Diffusion
Keywords: NLP, text embedding, document embedding, biomedicine, NER, PoS, sentiment analysis
##mindsdb
MindsDB is a low-code machine learning platform. It automatically integrates several ML frameworks into the data stack as "AI tables" to simplify the integration of AI in applications and make them accessible to developers of all skill levels.
Keywords: database, low code, AI table
langchain
Langchain is designed to assist in the development of compatible LLM and others Knowledge source applications. The library allows chaining calls to applications, creating a sequence in many tools.
Keywords: LLM, large language model, agent, chain
ParlAI
ParlAI is a platform for sharing , a Python framework for training and testing dialogue models, from open-domain chat, to task-oriented dialogue, to visual question answering. It provides over 100 datasets, many pre-trained models, a set of agents, and several integrations under the same API.
Keywords: dialogue, chatbot, VQA, data set, agent
sentence-transformers
This framework Provides a simple way to compute dense vector representations of sentences, paragraphs and images. These models are based on Transformer-based networks such as BERT/RoBERTa/XLM-RoBERTa and have achieved SOTA in various tasks. Text is embedded in a vector space such that similar texts are close and can be found efficiently through cosine similarity.
Keywords: dense vector representation, text embedding, sentence embedding
ludwig
Ludwig is a declarative machine Learning Framework, which makes it easy to define machine learning pipelines using a simple and flexible data-driven configuration system. Ludwig targets various AI tasks and provides a data-driven configuration system, training, prediction and evaluation scripts, and a programming API.
Keywords: declarative, data-driven, ML framework
InvokeAI
InvokeAI is an engine for the Stable Diffusion model, aimed at professionals, artists and enthusiasts. It leverages the latest AI-driven technology through CLI as well as WebUI.
Keywords: Stable Diffusion, WebUI, CLI
PaddleNLP
PaddleNLP is an easy-to-use and powerful NLP library, especially for the Chinese language. It supports multiple pre-trained model zoos and supports a wide range of NLP tasks from research to industrial applications.
Keywords: Natural Language Processing, Chinese, Research, Industry
stanza
Official Python of the Stanford University NLP Group NLP library. It supports running a wide range of precise natural language processing tools in more than 60 languages, and supports access to Java Stanford CoreNLP software from Python.
Keywords: NLP, multi-language, CoreNLP
DeepPavlov
DeepPavlov is an open source conversational artificial intelligence library . It is designed for the development of production-ready chatbots, and complex dialogue systems, as well as research in the field of NLP, specifically dialogue systems.
Keywords: dialogue, chatbot
alpaca-lora
Alpaca-lora includes the use of low-rank adaptation ( LoRA) code to reproduce Stanford Alpaca results. This repository provides training (fine-tuning) and generation scripts.
Keywords: LoRA, efficient fine-tuning of parameters
imagen-pytorch
An open source implementation of Imagen, Google’s closed Source text-to-image neural network beats DALL-E2. imagen-pytorch is the new SOTA for text-to-image synthesis.
Keywords: Imagen, Wenshengtu
adapter-transformers is an extension of the Transformers library that integrates adapters into state-of-the-art language models by incorporating AdapterHub, a central repository of pre-trained adapter modules. It is a drop-in replacement for Transformers and is updated regularly to keep pace with Transformers developments.
Keywords: Adapter, LoRA, parameter efficient fine-tuning, Hub
NeMoNVIDIA NeMo is designed for automatic speech A conversational AI toolkit built by researchers in recognition (ASR), text-to-speech synthesis (TTS), large language models, and natural language processing. The main goal of NeMo is to help researchers from industry and academia repurpose previous work (code and pre-trained models) and make it easier to create new projects.
Keywords: dialogue, ASR, TTS, LLM, NLP
RunhouseRunhouse allows you to combine code with Python Send data to any computer or data underlying and continue to interact with them normally from existing code and environments. Runhouse developers mentioned:
#You can think of it as an extension package for the Python interpreter, which can bypass remote machines or operate remote data.
Keywords: MLOps, infrastructure, data storage, modeling
MONAI
MONAI is part of the PyTorch ecosystem and is an open source framework based on PyTorch for deep learning in the field of medical imaging. Its objectives are:
- To develop a collaborative community of academic, industrial and clinical researchers on a common basis;
- To contribute to the medical Imaging creates SOTA, end-to-end training workflow;
- Provides an optimized and standardized method for the establishment and evaluation of deep learning models.
Keywords: medical imaging, training, evaluation
simpletransformers
Simple Transformers allows you to quickly train and evaluate Transformer models . Only 3 lines of code are needed to initialize, train and evaluate the model. It supports a wide variety of NLP tasks.
Keywords: framework, simplicity, NLP
JARVIS
JARVIS is a GPT-4, etc. The LLM system merges with other models from the open source machine learning community, leveraging up to 60 downstream models to perform tasks identified by LLM.
Keywords: LLM, agent, HF Hub
transformers.js
transformers.js is a JavaScript library that aims to run models from transformers directly in the browser.
Keywords: Transformers, JavaScript, browser
bumblebee
Bumblebee provides pre-trained on top of Axon Neural network model, Axon is a neural network library for the Elixir language. It includes integration with models, allowing anyone to download and perform machine learning tasks with just a few lines of code.
Keywords: Elixir, Axon
argilla
Argilla is a tool that provides advanced NLP labeling, monitoring and workspaces Open source platform. It is compatible with many open source ecosystems such as Hugging Face, Stanza, FLAIR, etc.
Keywords: NLP, labeling, monitoring, workspace
haystack
Haystack is an open source NLP framework that can interact with data using Transformer models and LLM. It provides production-ready tools for quickly building complex decision making, question answering, semantic search, text generation applications, and more.
Keywords: NLP, Framework, LLM
SpaCy is a library for advanced natural language processing in Python and Cython. It is built on the latest research and designed from the ground up for use in real products. It provides support for the Transformers model through its third-party package spacy-transformers.
Keywords: NLP, architecture
speechbrain
SpeechBrain is an open source, integrated conversational AI toolkit based on PyTorch. Our goal is to create a single, flexible, user-friendly toolkit that can be used to easily develop state-of-the-art speech technologies, including speech recognition, speaker identification, speech enhancement, speech separation, speech recognition, multi-microphone signal processing and other systems.
Keywords: dialogue, speech
skorch
Skorch is a wrapper for PyTorch with scikit-learn compatibility Neural network library. It supports models in Transformers, as well as tokenizers from tokenizers.
Keywords: Scikit-Learning, PyTorch
bertviz
BertViz is an interactive tool used in applications such as Visualize attention in Transformer language models like BERT, GPT2, or T5. It can be run in Jupiter or Colab notebooks via a simple Python API that supports most Huggingface models.
Keywords: Visualization, Transformers
mesh-transformer-jax is a Haiku library that implements Transformers model parallelism using xmap/pjit operators in JAX.
This library is designed to scale to approximately 40B parameters on TPUv3. It is a library used to train GPT-J models.
Keywords: Haiku, model parallelism, LLM, TPUdeepchem
OpenNREA method for neural relationship extraction Open Source Packages (NRE). It targets a wide range of users, from novices, to developers, researchers or students.
Keywords: neural relationship extraction, framework
pycorrectorA Chinese text error correction tool. This method utilizes language model detection errors, pinyin features, and shape features to correct Chinese text errors. Can be used for Chinese Pinyin and stroke input methods.
Keywords: Chinese, error correction tools, language model, Pinyin
Keywords: data augmentation, synthetic data generation, audio, natural language processing
dream-textures
dream- textures is a library designed to bring stable diffusion support to Blender. It supports multiple use cases such as image generation, texture projection, in/out painting, ControlNet and upgrades.Keywords: Stable-Diffusion, Blender
##seldon-core
Seldon core turns your ML models (Tensorflow, Pytorch, H2o, etc.) or language wrappers (Python, Java, etc.) into production REST/GRPC microservices. Seldon can handle scaling to thousands of production machine learning models and provides advanced machine learning features including advanced metrics, request logs, interpreters, outlier detectors, A/B testing, canaries, and more.
Keywords: microservices, modeling, language packaging
open_model_zoo
This library includes optimized deep learning models and a set of demos to accelerate the development of high-performance deep learning inference applications. Use these free pre-trained models instead of training your own to speed up development and production deployment processes.
Keywords: optimization model, demonstration
ml-stable-diffusion
ML-Stable-Diffusion is Apple’s A repository that brings Stable Diffusion support to Core ML on Apple silicon devices. It supports stable diffusion checkpoints hosted on Hugging Face Hub.
Keywords: Stable Diffusion, Apple chip, Core ML
Stable-Dreamfusion is a pytorch implementation of text to 3D model Dreamfusion, powered by Stable Diffusion text to 2D model.
Keywords: Text to 3D, Stable Diffusion
Keywords: semantic search, LLM
##djl
Deep Java Library (DJL) is an open source, high-level, engine-agnostic Java framework for deep learning that is easy for developers to use. DJL provides native Java development experience and functions like other regular Java libraries. DJL provides Java bindings for HuggingFace Tokenizer and a simple conversion toolkit for deploying HuggingFace models in Java.
Keywords: Java, architecture
lm-evaluation-harness
This project provides a unified framework to test generative language models on a large number of different evaluation tasks. It supports more than 200 tasks and supports different ecosystems: HF Transformers, GPT-NeoX, DeepSpeed, and OpenAI API.
Keywords: LLM, evaluation, few samples
gpt-neox
This resource library records the use of EleutherAI A library for training large-scale language models on GPUs. The framework is based on NVIDIA's Megatron language model and is enhanced with DeepSpeed's technology and some new optimizations. Its focus is on training models with billions of parameters.
Keywords: training, LLM, Megatron, DeepSpeed
muzic
Muzic is a research on artificial intelligence music project, which is able to understand and generate music through deep learning and artificial intelligence. Muzic was created by researchers at Microsoft Research Asia.
Keywords: music understanding, music generation
dalle-flow
DALL · E Flow is an interactive workflow for generating high-definition images from text prompts. It uses DALL·E-Mega, GLID-3 XL and Stable Diffusion to generate candidate images, and then calls CLIP-as-service to prompt sort the candidate images. Preferred candidates are fed to GLID-3 XL for diffusion, which often enriches textures and backgrounds. Finally, the candidate is extended to 1024x1024 via SwinIR.
Keywords: High-definition image generation, Stable Diffusion, DALL-E Mega, GLID-3 XL, CLIP, SwinIR
lightseq
LightSeq is a high-performance training and inference library implemented in CUDA for sequence processing and generation. It is capable of efficiently computing modern NLP and CV models such as BERT, GPT, Transformer, etc. Therefore, it is useful for machine translation, text generation, image classification, and other sequence-related tasks.
Keywords: training, inference, sequence processing, sequence generation
LaTeX- OCR
The goal of this project is to create a learning-based system that takes an image of a mathematical formula and returns the corresponding LaTeX code.
Keywords: OCR, LaTeX, mathematical formulas
open_clip
OpenCLIP is an open source implementation of OpenAI’s CLIP.
The goal of this repository is to enable the training of models with contrastive image-text supervision and to study their properties such as robustness to distribution shifts. The starting point of the project is an implementation of CLIP that matches the accuracy of the original CLIP model when trained on the same dataset.
Specifically, a ResNet-50 model trained on OpenAI’s 15 million image subset YFCC as the code base achieved the highest accuracy of 32.7% on ImageNet.
Keywords: CLIP, open source, comparison, image text
dalle-playground
A playground to generate images from any text prompt using Stable Diffusion and Dall-E mini.
Keywords: WebUI, Stable Diffusion, Dall-E mini
FedML is a federated learning and analytics library that enables secure and collaborative machine learning on distributed data anywhere and at any scale.
Keywords: federated learning, analysis, collaborative machine learning, decentralized
The above is the detailed content of The star mark exceeded 100,000! After Auto-GPT, Transformer reaches new milestone. For more information, please follow other related articles on the PHP Chinese website!

This article explores the growing concern of "AI agency decay"—the gradual decline in our ability to think and decide independently. This is especially crucial for business leaders navigating the increasingly automated world while retainin

Ever wondered how AI agents like Siri and Alexa work? These intelligent systems are becoming more important in our daily lives. This article introduces the ReAct pattern, a method that enhances AI agents by combining reasoning an

"I think AI tools are changing the learning opportunities for college students. We believe in developing students in core courses, but more and more people also want to get a perspective of computational and statistical thinking," said University of Chicago President Paul Alivisatos in an interview with Deloitte Nitin Mittal at the Davos Forum in January. He believes that people will have to become creators and co-creators of AI, which means that learning and other aspects need to adapt to some major changes. Digital intelligence and critical thinking Professor Alexa Joubin of George Washington University described artificial intelligence as a “heuristic tool” in the humanities and explores how it changes

LangChain is a powerful toolkit for building sophisticated AI applications. Its agent architecture is particularly noteworthy, allowing developers to create intelligent systems capable of independent reasoning, decision-making, and action. This expl

Radial Basis Function Neural Networks (RBFNNs): A Comprehensive Guide Radial Basis Function Neural Networks (RBFNNs) are a powerful type of neural network architecture that leverages radial basis functions for activation. Their unique structure make

Brain-computer interfaces (BCIs) directly link the brain to external devices, translating brain impulses into actions without physical movement. This technology utilizes implanted sensors to capture brain signals, converting them into digital comman

This "Leading with Data" episode features Ines Montani, co-founder and CEO of Explosion AI, and co-developer of spaCy and Prodigy. Ines offers expert insights into the evolution of these tools, Explosion's unique business model, and the tr

This article explores Retrieval Augmented Generation (RAG) systems and how AI agents can enhance their capabilities. Traditional RAG systems, while useful for leveraging custom enterprise data, suffer from limitations such as a lack of real-time dat


Hot AI Tools

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Undress AI Tool
Undress images for free

Clothoff.io
AI clothes remover

Video Face Swap
Swap faces in any video effortlessly with our completely free AI face swap tool!

Hot Article

Hot Tools

MantisBT
Mantis is an easy-to-deploy web-based defect tracking tool designed to aid in product defect tracking. It requires PHP, MySQL and a web server. Check out our demo and hosting services.

Dreamweaver Mac version
Visual web development tools

SublimeText3 Mac version
God-level code editing software (SublimeText3)

PhpStorm Mac version
The latest (2018.2.1) professional PHP integrated development tool

WebStorm Mac version
Useful JavaScript development tools