Import Transformers into Spark NLP 🚀 #5669
Locked
maziyarpanahi
announced in
Announcement
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Import Models into Spark NLP
Table of Contents
Overview
Since version 3.1.0, Spark NLP 🚀 has supported importing pretrained models from Hugging Face 🤗 and TensorFlow Hub into equivalent Spark NLP annotators.
This means you can bring your favorite Transformer architectures such as BERT, RoBERTa, DistilBERT, DeBERTa, XLM-RoBERTa, Longformer, CamemBERT, XLNet, and many others directly into Spark NLP pipelines for tasks like:
With every release, we extend this compatibility to cover more architectures and runtimes.
Quick Start
Basic Model Import
👉 Explore runnable end-to-end examples in our Notebook Gallery Repository. You’ll find Colab/Jupyter notebooks for each annotator and runtime (TensorFlow, ONNX, OpenVINO, Llama.cpp).
Compatibility Matrix
Text Embeddings
Sequence Classification
Token Classification
Question Answering
Text Generation
Computer Vision
Speech Processing
Large Language Models
Vision-Language Models
Importing Pretrained Models to Spark NLP
We provide a comprehensive collection of end-to-end notebooks for importing and converting pretrained models into Spark NLP. These resources cover all major annotators and runtimes:
HuggingFace to Spark NLP (TensorFlow)
HuggingFace to Spark NLP (ONNX)
HuggingFace to Spark NLP (OpenVINO)
HuggingFace to Spark NLP (Llama.cpp)
Beta Was this translation helpful? Give feedback.
All reactions