diff --git a/notebooks/data_loader.ipynb b/notebooks/data_loader.ipynb index 3e6ea4275ee9b855c84e4fa6d5f3a5b0b693460e..c1a1880d78963eb25f83483bc6a494b879a2f738 100644 --- a/notebooks/data_loader.ipynb +++ b/notebooks/data_loader.ipynb @@ -35,19 +35,42 @@ }, { "cell_type": "code", - "execution_count": 29, + "execution_count": 3, "metadata": { "ExecuteTime": { - "end_time": "2020-05-23T21:04:37.477348Z", - "start_time": "2020-05-23T21:04:37.475048Z" + "end_time": "2020-05-23T23:20:55.848368Z", + "start_time": "2020-05-23T23:20:54.138572Z" } }, - "outputs": [], + "outputs": [ + { + "name": "stderr", + "output_type": "stream", + "text": [ + "[nltk_data] Downloading package punkt to /home/water/nltk_data...\n", + "[nltk_data] Package punkt is already up-to-date!\n" + ] + }, + { + "data": { + "text/plain": [ + "True" + ] + }, + "execution_count": 3, + "metadata": {}, + "output_type": "execute_result" + } + ], "source": [ "import os\n", "import pandas as pd\n", "import numpy as np\n", - "from nltk.tokenize import word_tokenize" + "import nltk\n", + "from nltk.tokenize import word_tokenize\n", + "\n", + "\n", + "nltk.download(\"punkt\")" ] }, { @@ -280,7 +303,7 @@ "cell_type": "markdown", "metadata": {}, "source": [ - "At least, we go through all the sentences in the dataset and create the token dictionary." + "Finally, we go through all the sentences in the dataset and create the token dictionary." ] }, {