This resource contains a Jupyter Notebook that walks though a streamlined approach to training, evaluating, and testing models for Question Answering (QA) tasks, leveraging NVIDIA’s NeMo Framework. Within this notebook, users can explore three QA model types: BERT-like models for Extractive Question Answering, Sequence-to-Sequence (S2S) models like T5/BART for Generative Question Answering, and GPT-like models for advanced generative responses. This adaptation focuses on practical application, guiding users through the process of generating answers from given contexts and queries, with detailed explanations on how to use NeMo.
To streamline your experience and jump directly into a GPU-accelerated environment with this notebook and NeMo pre-installed, click the badge below. Our 1-click deploys are powered by Brev.dev.
Use the 1-click deploy link above to set up a machine with NeMO installed. Once the VM is ready, use the Access Notebook button to enter the Jupyter Lab instance
For this notebook, we use two types of question-answering paradigms and three different models:
We will be using the SQuAD dataset to showcase the training and inference. We train, test, and deploy all three models for inference and evaluate performance of all three architectures.
NVIDIA NeMo Framework is a generative AI framework built for researchers and pytorch developers working on large language models (LLMs), multimodal models (MM), automatic speech recognition (ASR), and text-to-speech synthesis (TTS). NeMO provides a scalable framework to easily design, implement, and scale new AI models using existing pre-trained models and a simple API for configuration.