Skip to content

Commit 4c2214e

Browse files
authored
* Update the `external/morpheus-visualizations` submodule to adopt fixes from nv-morpheus/morpheus-visualizations#50 ## By Submitting this PR I confirm: - I am familiar with the [Contributing Guidelines](https://github.com/nv-morpheus/Morpheus/blob/main/docs/source/developer_guide/contributing.md). - When the PR is ready for review, new or existing tests cover these changes. - When the PR is ready for review, the documentation is up to date with these changes. Authors: - David Gardner (https://github.com/dagardner-nv) Approvers: - Yuchen Zhang (https://github.com/yczhang-nv) URL: #2186
1 parent 4f2ed4f commit 4c2214e

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

docs/source/developer_guide/guides/2_real_world_phishing.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -413,7 +413,7 @@ else:
413413
pipeline.add_stage(RecipientFeaturesStage(config))
414414
```
415415

416-
To tokenize the input data we will use Morpheus' `PreprocessNLPStage`. This stage uses the [cuDF subword tokenizer](https://docs.rapids.ai/api/cudf/legacy/user_guide/api_docs/subword_tokenize/#subwordtokenizer) to transform strings into a tensor of numbers to be fed into the neural network model. Rather than split the string by characters or whitespaces, we split them into meaningful subwords based upon the occurrence of the subwords in a large training corpus. You can find more details here: [https://arxiv.org/abs/1810.04805v2](https://arxiv.org/abs/1810.04805v2). All we need to know for now is that the text will be converted to subword token ids based on the vocabulary file that we provide (`vocab_hash_file=vocab file`).
416+
To tokenize the input data we will use Morpheus' `PreprocessNLPStage`. This stage uses the [cuDF subword tokenizer](https://docs.rapids.ai/api/cudf/legacy/user_guide/api_docs/pylibcudf/nvtext/subword_tokenize/#module-pylibcudf.nvtext.subword_tokenize) to transform strings into a tensor of numbers to be fed into the neural network model. Rather than split the string by characters or whitespaces, we split them into meaningful subwords based upon the occurrence of the subwords in a large training corpus. You can find more details here: [https://arxiv.org/abs/1810.04805v2](https://arxiv.org/abs/1810.04805v2). All we need to know for now is that the text will be converted to subword token ids based on the vocabulary file that we provide (`vocab_hash_file=vocab file`).
417417

418418
Let's go ahead and instantiate our `PreprocessNLPStage` and add it to the pipeline:
419419

0 commit comments

Comments
 (0)