@@ -324,13 +324,13 @@ that are included with NeMo:
324324- `Language Modeling (BERT Pretraining) <https://github.com/NVIDIA/NeMo/blob/v1.0.0b1/tutorials/nlp/01_Pretrained_Language_Models_for_Downstream_Tasks.ipynb >`_
325325- `Question Answering <https://github.com/NVIDIA/NeMo/blob/v1.0.0b1/tutorials/nlp/Question_Answering_Squad.ipynb >`_
326326- `Text Classification <https://github.com/NVIDIA/NeMo/tree/v1.0.0b1/examples/nlp/text_classification >`_ (including Sentiment Analysis)
327- - `Token Classifcation <https://github.com/NVIDIA/NeMo/tree/v1.0.0b1/examples/nlp/token_classification >`_ (including Named Entity Recognition)
327+ - `Token Classification <https://github.com/NVIDIA/NeMo/tree/v1.0.0b1/examples/nlp/token_classification >`_ (including Named Entity Recognition)
328328- `Punctuation and Capitalization <https://github.com/NVIDIA/NeMo/blob/v1.0.0b1/tutorials/nlp/Punctuation_and_Capitalization.ipynb >`_
329329
330330Named Entity Recognition (NER)
331331------------------------------
332332
333- NER (or more generally token classifcation ) is the NLP task of detecting and classifying key information (entities) in text.
333+ NER (or more generally token classification ) is the NLP task of detecting and classifying key information (entities) in text.
334334This task is very popular in Healthcare and Finance. In finance, for example, it can be important to identify
335335geographical, geopolitical, organizational, persons, events, and natural phenomenon entities.
336336See this `NER notebook <https://github.com/NVIDIA/NeMo/blob/v1.0.0b1/tutorials/nlp/Token_Classification_Named_Entity_Recognition.ipynb >`_
@@ -435,7 +435,7 @@ Hydra makes every aspect of the NeMo model, including the PyTorch Lightning Trai
435435Tokenizers
436436----------
437437
438- Tokenization is the process of converting natural langauge text into integer arrays
438+ Tokenization is the process of converting natural language text into integer arrays
439439which can be used for machine learning.
440440For NLP tasks, tokenization is an essential part of data preprocessing.
441441NeMo supports all BERT-like model tokenizers from
@@ -462,7 +462,7 @@ Much of the state-of-the-art in natural language processing is achieved
462462by fine-tuning pretrained language models on the downstream task.
463463
464464With NeMo, you can either `pretrain <https://github.com/NVIDIA/NeMo/blob/v1.0.0b1/examples/nlp/language_modeling/bert_pretraining.py >`_
465- a BERT model on your data or use a pretrained lanugage model from `HuggingFace Transformers <https://github.com/huggingface/transformers >`_
465+ a BERT model on your data or use a pretrained language model from `HuggingFace Transformers <https://github.com/huggingface/transformers >`_
466466or `NVIDIA Megatron-LM <https://github.com/NVIDIA/Megatron-LM >`_.
467467
468468To see the list of language models available in NeMo:
0 commit comments