What does HackerNews think of transformers?

🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

Language: Python

#14 in Deep learning
#125 in Hacktoberfest
#7 in R
#102 in Python
#8 in Tensorflow
You need to use the transformers from the main branch instead of the pypi version, because the llama support is recently added. According to the readme of the repo, you need to install transformers with: pip3 install git+https://github.com/huggingface/transformers
I know nothing, but have heard Hugging Face is in that direction.

https://github.com/huggingface/transformers

>Transformers provides thousands of pretrained models to perform tasks on different modalities such as text, vision, and audio.

> These models can be applied on:

> - Text, for tasks like text classification, information extraction, question answering, summarization, translation, text generation, in over 100 languages.

> - Images, for tasks like image classification, object detection, and segmentation.

> - Audio, for tasks like speech recognition and audio classification.

---

Also I read about GPT-J, whose capability is comparable with GTP-3.

https://github.com/kingoflolz/mesh-transformer-jax

But I believe it requires buying or renting GPUs.

A lot of SOTA models seem to be gravitating towards transformer based models. Obviously, I can't speak for the entire field, but you can just go take a look at the most popular HuggingFace repos and see what I mean. They started out focused on language, but because transformers have become so popular, they're expanding into the audio and vision domains quickly. Their library called 'transformers' is, outside of research, most peoples go to high level framework as it largely abstracts away a lot of the boilerplate that writing in pure TF, PyTorch, Jax requires.

See:

https://huggingface.co/spaces

https://github.com/huggingface/transformers

The GPT-2 weights were released by OpenAI when GPT-2 was released. (https://github.com/openai/gpt-2)

Around that time (since no one else was doing it) I released a wrapper to streamline that code and make it much easier to finetune on your own data. (https://github.com/minimaxir/gpt-2-simple)

Nowadays, the easiest way to interact with GPT-2 is to use the transformers library (https://github.com/huggingface/transformers), of which I've created a much better library for GPT-2 that leverages it. (https://github.com/minimaxir/aitextgen)

Try spacy alpha 3.0 it integrate the https://github.com/huggingface/transformers library You should almost always use XLnet large in order to achieve the best accuracy