From 4e6881f6a0f681c21f964f6646391ea832062e2f Mon Sep 17 00:00:00 2001 From: Landon Wilkins Date: Sat, 30 Mar 2024 04:16:29 -0600 Subject: [PATCH] fix typo in C3_W2_Lab_2_sarcasm_classifier.ipynb --- C3/W2/ungraded_labs/C3_W2_Lab_2_sarcasm_classifier.ipynb | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/C3/W2/ungraded_labs/C3_W2_Lab_2_sarcasm_classifier.ipynb b/C3/W2/ungraded_labs/C3_W2_Lab_2_sarcasm_classifier.ipynb index fce6cf15..b347cbc0 100755 --- a/C3/W2/ungraded_labs/C3_W2_Lab_2_sarcasm_classifier.ipynb +++ b/C3/W2/ungraded_labs/C3_W2_Lab_2_sarcasm_classifier.ipynb @@ -420,7 +420,7 @@ "\n", "In this lab, you were able to build a binary classifier to detect sarcasm. You saw some overfitting in the initial attempt and hopefully, you were able to arrive at a better set of hyperparameters.\n", "\n", - "So far, you've been tokenizing datasets from scratch and you're treating the vocab size as a hyperparameter. Furthermore, you're tokenizing the texts by building a vocabulary of full words. In the next lab, you will make use of a pre-tokenized dataset that uses a vocabulary of *subwords*. For instance, instead of having a uniqe token for the word `Tensorflow`, it will instead have a token each for `Ten`, `sor`, and `flow`. You will see the motivation and implications of having this design in the next exercise. See you there!" + "So far, you've been tokenizing datasets from scratch and you're treating the vocab size as a hyperparameter. Furthermore, you're tokenizing the texts by building a vocabulary of full words. In the next lab, you will make use of a pre-tokenized dataset that uses a vocabulary of *subwords*. For instance, instead of having a unique token for the word `Tensorflow`, it will instead have a token each for `Ten`, `sor`, and `flow`. You will see the motivation and implications of having this design in the next exercise. See you there!" ] } ],