You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The Question Answering Model presented in this repository is built on BERT (Bidirectional Encoder Representations from Transformers), a state-of-the-art transformer-based architecture tailored for natural language processing tasks.
DistilBERT, a distilled version of BERT (Bidirectional Encoder Representations from Transformers). DistilBERT is known for its efficiency and reduced computational requirements while retaining significant language understanding capabilities.