diff --git a/gonito.yaml b/gonito.yaml index 104ea91..82485c8 100644 --- a/gonito.yaml +++ b/gonito.yaml @@ -1,14 +1,16 @@ -description: zad8, trigram with left/right context embeddings +description: zad 8, trigram with right context embeddings (embeddings =100,200,300) tags: - neural-network - - left-to-right + - right-context params: epochs: 3 - learning-rate: 0.0003 - vocab-size: 20000 - batch_s: 3200 + embed_size: 300 + learning-rate: 0.001 + vocab-size: 30000 + batch_s: 1600 top_k_words: 20 -param-files: - - config/*.yaml + minimal_wildcard: 0.02 links: - - repo: "https://git.wmi.amu.edu.pl/s470618/challenging-america-word-gap-prediction" + -title: "repo" + url: + - repo: https://git.wmi.amu.edu.pl/s470618/challenging-america-word-gap-prediction-s470618.git diff --git a/zad7.ipynb b/zad7.ipynb deleted file mode 100644 index e23b29f..0000000 --- a/zad7.ipynb +++ /dev/null @@ -1,13640 +0,0 @@ -{ - "cells": [ - { - "cell_type": "code", - "execution_count": 7, - "metadata": { - "collapsed": true - }, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Defaulting to user installation because normal site-packages is not writeable\n", - "Collecting torchtext\n", - " Downloading torchtext-0.15.2-cp310-cp310-manylinux1_x86_64.whl (2.0 MB)\n", - "\u001b[2K \u001b[38;2;114;156;31m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m2.0/2.0 MB\u001b[0m \u001b[31m1.5 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m[36m0:00:01\u001b[0m[36m0:00:01\u001b[0m:01\u001b[0m\n", - "\u001b[?25hCollecting tqdm\n", - " Using cached tqdm-4.65.0-py3-none-any.whl (77 kB)\n", - "Requirement already satisfied: numpy in /home/gedin/.local/lib/python3.10/site-packages (from torchtext) (1.24.3)\n", - "Collecting torchdata==0.6.1\n", - " Downloading torchdata-0.6.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (4.6 MB)\n", - "\u001b[2K \u001b[38;2;114;156;31m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m4.6/4.6 MB\u001b[0m \u001b[31m1.5 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0mm eta \u001b[36m0:00:01\u001b[0m[36m0:00:01\u001b[0m\n", - "\u001b[?25hRequirement already satisfied: requests in /usr/lib/python3/dist-packages (from torchtext) (2.25.1)\n", - "Collecting torch==2.0.1\n", - " Downloading torch-2.0.1-cp310-cp310-manylinux1_x86_64.whl (619.9 MB)\n", - "\u001b[2K \u001b[38;2;114;156;31m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m619.9/619.9 MB\u001b[0m \u001b[31m1.1 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0mm eta \u001b[36m0:00:01\u001b[0m[36m0:00:09\u001b[0m\n", - "\u001b[?25hCollecting sympy\n", - " Downloading sympy-1.12-py3-none-any.whl (5.7 MB)\n", - "\u001b[2K \u001b[38;2;114;156;31m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m5.7/5.7 MB\u001b[0m \u001b[31m1.7 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0mm eta \u001b[36m0:00:01\u001b[0m[36m0:00:01\u001b[0m\n", - "\u001b[?25hCollecting nvidia-cudnn-cu11==8.5.0.96\n", - " Using cached nvidia_cudnn_cu11-8.5.0.96-2-py3-none-manylinux1_x86_64.whl (557.1 MB)\n", - "Collecting nvidia-cuda-cupti-cu11==11.7.101\n", - " Using cached nvidia_cuda_cupti_cu11-11.7.101-py3-none-manylinux1_x86_64.whl (11.8 MB)\n", - "Collecting nvidia-cusparse-cu11==11.7.4.91\n", - " Using cached nvidia_cusparse_cu11-11.7.4.91-py3-none-manylinux1_x86_64.whl (173.2 MB)\n", - "Collecting networkx\n", - " Using cached networkx-3.1-py3-none-any.whl (2.1 MB)\n", - "Collecting nvidia-cufft-cu11==10.9.0.58\n", - " Using cached nvidia_cufft_cu11-10.9.0.58-py3-none-manylinux1_x86_64.whl (168.4 MB)\n", - "Collecting filelock\n", - " Downloading filelock-3.12.0-py3-none-any.whl (10 kB)\n", - "Collecting nvidia-cuda-runtime-cu11==11.7.99\n", - " Using cached nvidia_cuda_runtime_cu11-11.7.99-py3-none-manylinux1_x86_64.whl (849 kB)\n", - "Collecting triton==2.0.0\n", - " Downloading triton-2.0.0-1-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (63.3 MB)\n", - "\u001b[2K \u001b[38;2;114;156;31m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m63.3/63.3 MB\u001b[0m \u001b[31m1.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0mm eta \u001b[36m0:00:01\u001b[0m[36m0:00:02\u001b[0m\n", - "\u001b[?25hCollecting nvidia-cusolver-cu11==11.4.0.1\n", - " Using cached nvidia_cusolver_cu11-11.4.0.1-2-py3-none-manylinux1_x86_64.whl (102.6 MB)\n", - "Requirement already satisfied: jinja2 in /home/gedin/.local/lib/python3.10/site-packages (from torch==2.0.1->torchtext) (3.1.2)\n", - "Collecting nvidia-cublas-cu11==11.10.3.66\n", - " Using cached nvidia_cublas_cu11-11.10.3.66-py3-none-manylinux1_x86_64.whl (317.1 MB)\n", - "Collecting typing-extensions\n", - " Downloading typing_extensions-4.6.3-py3-none-any.whl (31 kB)\n", - "Collecting nvidia-nccl-cu11==2.14.3\n", - " Using cached nvidia_nccl_cu11-2.14.3-py3-none-manylinux1_x86_64.whl (177.1 MB)\n", - "Collecting nvidia-cuda-nvrtc-cu11==11.7.99\n", - " Using cached nvidia_cuda_nvrtc_cu11-11.7.99-2-py3-none-manylinux1_x86_64.whl (21.0 MB)\n", - "Collecting nvidia-curand-cu11==10.2.10.91\n", - " Using cached nvidia_curand_cu11-10.2.10.91-py3-none-manylinux1_x86_64.whl (54.6 MB)\n", - "Collecting nvidia-nvtx-cu11==11.7.91\n", - " Using cached nvidia_nvtx_cu11-11.7.91-py3-none-manylinux1_x86_64.whl (98 kB)\n", - "Requirement already satisfied: urllib3>=1.25 in /usr/lib/python3/dist-packages (from torchdata==0.6.1->torchtext) (1.26.5)\n", - "Requirement already satisfied: wheel in /usr/lib/python3/dist-packages (from nvidia-cublas-cu11==11.10.3.66->torch==2.0.1->torchtext) (0.37.1)\n", - "Requirement already satisfied: setuptools in /usr/lib/python3/dist-packages (from nvidia-cublas-cu11==11.10.3.66->torch==2.0.1->torchtext) (59.6.0)\n", - "Collecting lit\n", - " Downloading lit-16.0.5.tar.gz (138 kB)\n", - "\u001b[2K \u001b[38;2;114;156;31m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m138.0/138.0 KB\u001b[0m \u001b[31m1.7 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m[31m1.6 MB/s\u001b[0m eta \u001b[36m0:00:01\u001b[0m\n", - "\u001b[?25h Preparing metadata (setup.py) ... \u001b[?25ldone\n", - "\u001b[?25hCollecting cmake\n", - " Using cached cmake-3.26.3-py2.py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (24.0 MB)\n", - "Requirement already satisfied: MarkupSafe>=2.0 in /usr/lib/python3/dist-packages (from jinja2->torch==2.0.1->torchtext) (2.0.1)\n", - "Collecting mpmath>=0.19\n", - " Using cached mpmath-1.3.0-py3-none-any.whl (536 kB)\n", - "Building wheels for collected packages: lit\n", - " Building wheel for lit (setup.py) ... \u001b[?25ldone\n", - "\u001b[?25h Created wheel for lit: filename=lit-16.0.5-py3-none-any.whl size=88192 sha256=f6c57a31a147cbfe0af3d6bf4b856390ad14c28a9ddb38c8044ec29331b35c26\n", - " Stored in directory: /home/gedin/.cache/pip/wheels/eb/02/84/d82f0b1a6098209edf7e3607be6cc592ebbc015a8a3127c68d\n", - "Successfully built lit\n", - "Installing collected packages: mpmath, lit, cmake, typing-extensions, tqdm, sympy, nvidia-nvtx-cu11, nvidia-nccl-cu11, nvidia-cusparse-cu11, nvidia-curand-cu11, nvidia-cufft-cu11, nvidia-cuda-runtime-cu11, nvidia-cuda-nvrtc-cu11, nvidia-cuda-cupti-cu11, nvidia-cublas-cu11, networkx, filelock, nvidia-cusolver-cu11, nvidia-cudnn-cu11, triton, torch, torchdata, torchtext\n", - "Successfully installed cmake-3.26.3 filelock-3.12.0 lit-16.0.5 mpmath-1.3.0 networkx-3.1 nvidia-cublas-cu11-11.10.3.66 nvidia-cuda-cupti-cu11-11.7.101 nvidia-cuda-nvrtc-cu11-11.7.99 nvidia-cuda-runtime-cu11-11.7.99 nvidia-cudnn-cu11-8.5.0.96 nvidia-cufft-cu11-10.9.0.58 nvidia-curand-cu11-10.2.10.91 nvidia-cusolver-cu11-11.4.0.1 nvidia-cusparse-cu11-11.7.4.91 nvidia-nccl-cu11-2.14.3 nvidia-nvtx-cu11-11.7.91 sympy-1.12 torch-2.0.1 torchdata-0.6.1 torchtext-0.15.2 tqdm-4.65.0 triton-2.0.0 typing-extensions-4.6.3\n" - ] - } - ], - "source": [ - "!pip install torchtext" - ] - }, - { - "cell_type": "code", - "execution_count": 50, - "metadata": {}, - "outputs": [], - "source": [ - "train_file ='train/in.tsv.xz'\n", - "test_file = 'test-A/in.tsv.xz'\n", - "out_file = 'test-A/out.tsv'" - ] - }, - { - "cell_type": "code", - "execution_count": 10, - "metadata": {}, - "outputs": [], - "source": [ - "from itertools import islice\n", - "import regex as re\n", - "import sys\n", - "from torchtext.vocab import build_vocab_from_iterator\n", - "import lzma\n", - "import pickle\n", - "import re\n", - "import torch\n", - "from torch import nn\n", - "from torch.utils.data import IterableDataset\n", - "import itertools\n", - "from torch.utils.data import DataLoader\n", - "import gc" - ] - }, - { - "cell_type": "code", - "execution_count": 51, - "metadata": {}, - "outputs": [], - "source": [ - "embed_size = 300\n", - "device = 'cuda'\n", - "vocab_size = 25000\n", - "batch_s = 3200\n", - "learning_rate = 0.0001\n", - "epochs = 4\n", - "k = 20 #top k words\n", - "wildcard_minweight = 0.001" - ] - }, - { - "cell_type": "code", - "execution_count": 12, - "metadata": { - "scrolled": true - }, - "outputs": [], - "source": [ - "###preprocessing\n", - "def preprocess(line):\n", - " line = get_rid_of_header(line)\n", - " line = replace_endline(line)\n", - " return line\n", - "\n", - "def get_rid_of_header(line):\n", - " line = line.split('\\t')[6:]\n", - " return \"\".join(line)\n", - " \n", - "def replace_endline(line):\n", - " line = line.replace(\"\\\\n\", \" \")\n", - " return line\n", - "\n", - "\n", - "def get_last_word(text):\n", - " \"\"\"Return the last word of a string.\"\"\"\n", - " last_word = \"\"\n", - " for i in range(len(text)-1, -1, -1):\n", - " if text[i] == ' ':\n", - " return last_word[::-1].rstrip()\n", - " else:\n", - " last_word += text[i]\n", - " return last_word[::-1].rstrip()\n", - "\n", - "def get_first_word(text):\n", - " \"\"\"Return the first word of a string.\"\"\"\n", - " word = \"\"\n", - " for i in range(len(text)-1):\n", - " if text[i] == ' ':\n", - " return word\n", - " else:\n", - " word += text[i]\n", - " return word\n", - "\n", - "\n", - "def get_words_from_line(line):\n", - " line = line.rstrip()\n", - " yield ''\n", - " line = preprocess(line)\n", - " for t in line.split(' '):\n", - " yield t\n", - " yield ''\n", - "\n", - "\n", - "def get_word_lines_from_file(file_name):\n", - " n = 0\n", - " with lzma.open(file_name, 'r') as fh:\n", - " for line in fh:\n", - " n+=1\n", - " if n%1000==0:\n", - " print(n)\n", - " yield get_words_from_line(line.decode('utf-8'))" - ] - }, - { - "cell_type": "code", - "execution_count": 5, - "metadata": {}, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "1000\n", - "2000\n", - "3000\n", - "4000\n", - "5000\n", - "6000\n", - "7000\n", - "8000\n", - "9000\n", - "10000\n", - "11000\n", - "12000\n", - "13000\n", - "14000\n", - "15000\n", - "16000\n", - "17000\n", - "18000\n", - "19000\n", - "20000\n", - "21000\n", - "22000\n", - "23000\n", - "24000\n", - "25000\n", - "26000\n", - "27000\n", - "28000\n", - "29000\n", - "30000\n", - "31000\n", - "32000\n", - "33000\n", - "34000\n", - "35000\n", - "36000\n", - "37000\n", - "38000\n", - "39000\n", - "40000\n", - "41000\n", - "42000\n", - "43000\n", - "44000\n", - "45000\n", - "46000\n", - "47000\n", - "48000\n", - "49000\n", - "50000\n", - "51000\n", - "52000\n", - "53000\n", - "54000\n", - "55000\n", - "56000\n", - "57000\n", - "58000\n", - "59000\n", - "60000\n", - "61000\n", - "62000\n", - "63000\n", - "64000\n", - "65000\n", - "66000\n", - "67000\n", - "68000\n", - "69000\n", - "70000\n", - "71000\n", - "72000\n", - "73000\n", - "74000\n", - "75000\n", - "76000\n", - "77000\n", - "78000\n", - "79000\n", - "80000\n", - "81000\n", - "82000\n", - "83000\n", - "84000\n", - "85000\n", - "86000\n", - "87000\n", - "88000\n", - "89000\n", - "90000\n", - "91000\n", - "92000\n", - "93000\n", - "94000\n", - "95000\n", - "96000\n", - "97000\n", - "98000\n", - "99000\n", - "100000\n", - "101000\n", - "102000\n", - "103000\n", - "104000\n", - "105000\n", - "106000\n", - "107000\n", - "108000\n", - "109000\n", - "110000\n", - "111000\n", - "112000\n", - "113000\n", - "114000\n", - "115000\n", - "116000\n", - "117000\n", - "118000\n", - "119000\n", - "120000\n", - "121000\n", - "122000\n", - "123000\n", - "124000\n", - "125000\n", - "126000\n", - "127000\n", - "128000\n", - "129000\n", - "130000\n", - "131000\n", - "132000\n", - "133000\n", - "134000\n", - "135000\n", - "136000\n", - "137000\n", - "138000\n", - "139000\n", - "140000\n", - "141000\n", - "142000\n", - "143000\n", - "144000\n", - "145000\n", - "146000\n", - "147000\n", - "148000\n", - "149000\n", - "150000\n", - "151000\n", - "152000\n", - "153000\n", - "154000\n", - "155000\n", - "156000\n", - "157000\n", - "158000\n", - "159000\n", - "160000\n", - "161000\n", - "162000\n", - "163000\n", - "164000\n", - "165000\n", - "166000\n", - "167000\n", - "168000\n", - "169000\n", - "170000\n", - "171000\n", - "172000\n", - "173000\n", - "174000\n", - "175000\n", - "176000\n", - "177000\n", - "178000\n", - "179000\n", - "180000\n", - "181000\n", - "182000\n", - "183000\n", - "184000\n", - "185000\n", - "186000\n", - "187000\n", - "188000\n", - "189000\n", - "190000\n", - "191000\n", - "192000\n", - "193000\n", - "194000\n", - "195000\n", - "196000\n", - "197000\n", - "198000\n", - "199000\n", - "200000\n", - "201000\n", - "202000\n", - "203000\n", - "204000\n", - "205000\n", - "206000\n", - "207000\n", - "208000\n", - "209000\n", - "210000\n", - "211000\n", - "212000\n", - "213000\n", - "214000\n", - "215000\n", - "216000\n", - "217000\n", - "218000\n", - "219000\n", - "220000\n", - "221000\n", - "222000\n", - "223000\n", - "224000\n", - "225000\n", - "226000\n", - "227000\n", - "228000\n", - "229000\n", - "230000\n", - "231000\n", - "232000\n", - "233000\n", - "234000\n", - "235000\n", - "236000\n", - "237000\n", - "238000\n", - "239000\n", - "240000\n", - "241000\n", - "242000\n", - "243000\n", - "244000\n", - "245000\n", - "246000\n", - "247000\n", - "248000\n", - "249000\n", - "250000\n", - "251000\n", - "252000\n", - "253000\n", - "254000\n", - "255000\n", - "256000\n", - "257000\n", - "258000\n", - "259000\n", - "260000\n", - "261000\n", - "262000\n", - "263000\n", - "264000\n", - "265000\n", - "266000\n", - "267000\n", - "268000\n", - "269000\n", - "270000\n", - "271000\n", - "272000\n", - "273000\n", - "274000\n", - "275000\n", - "276000\n", - "277000\n", - "278000\n", - "279000\n", - "280000\n", - "281000\n", - "282000\n", - "283000\n", - "284000\n", - "285000\n", - "286000\n", - "287000\n", - "288000\n", - "289000\n", - "290000\n", - "291000\n", - "292000\n", - "293000\n", - "294000\n", - "295000\n", - "296000\n", - "297000\n", - "298000\n", - "299000\n", - "300000\n", - "301000\n", - "302000\n", - "303000\n", - "304000\n", - "305000\n", - "306000\n", - "307000\n", - "308000\n", - "309000\n", - "310000\n", - "311000\n", - "312000\n", - "313000\n", - "314000\n", - "315000\n", - "316000\n", - "317000\n", - "318000\n", - "319000\n", - "320000\n", - "321000\n", - "322000\n", - "323000\n", - "324000\n", - "325000\n", - "326000\n", - "327000\n", - "328000\n", - "329000\n", - "330000\n", - "331000\n", - "332000\n", - "333000\n", - "334000\n", - "335000\n", - "336000\n", - "337000\n", - "338000\n", - "339000\n", - "340000\n", - "341000\n", - "342000\n", - "343000\n", - "344000\n", - "345000\n", - "346000\n", - "347000\n", - "348000\n", - "349000\n", - "350000\n", - "351000\n", - "352000\n", - "353000\n", - "354000\n", - "355000\n", - "356000\n", - "357000\n", - "358000\n", - "359000\n", - "360000\n", - "361000\n", - "362000\n", - "363000\n", - "364000\n", - "365000\n", - "366000\n", - "367000\n", - "368000\n", - "369000\n", - "370000\n", - "371000\n", - "372000\n", - "373000\n", - "374000\n", - "375000\n", - "376000\n", - "377000\n", - "378000\n", - "379000\n", - "380000\n", - "381000\n", - "382000\n", - "383000\n", - "384000\n", - "385000\n", - "386000\n", - "387000\n", - "388000\n", - "389000\n", - "390000\n", - "391000\n", - "392000\n", - "393000\n", - "394000\n", - "395000\n", - "396000\n", - "397000\n", - "398000\n", - "399000\n", - "400000\n", - "401000\n", - "402000\n", - "403000\n", - "404000\n", - "405000\n", - "406000\n", - "407000\n", - "408000\n", - "409000\n", - "410000\n", - "411000\n", - "412000\n", - "413000\n", - "414000\n", - "415000\n", - "416000\n", - "417000\n", - "418000\n", - "419000\n", - "420000\n", - "421000\n", - "422000\n", - "423000\n", - "424000\n", - "425000\n", - "426000\n", - "427000\n", - "428000\n", - "429000\n", - "430000\n", - "431000\n", - "432000\n" - ] - } - ], - "source": [ - "vocab = build_vocab_from_iterator(\n", - " get_word_lines_from_file(train_file),\n", - " max_tokens = vocab_size,\n", - " specials = [''])\n", - "\n", - "with open('filename.pickle', 'wb') as handle:\n", - " pickle.dump(vocab, handle, protocol=pickle.HIGHEST_PROTOCOL)" - ] - }, - { - "cell_type": "code", - "execution_count": 5, - "metadata": {}, - "outputs": [ - { - "ename": "NameError", - "evalue": "name 'vocab' is not defined", - "output_type": "error", - "traceback": [ - "\u001b[0;31m---------------------------------------------------------------------------\u001b[0m", - "\u001b[0;31mNameError\u001b[0m Traceback (most recent call last)", - "\u001b[0;32m/tmp/ipykernel_20466/3224446201.py\u001b[0m in \u001b[0;36m\u001b[0;34m\u001b[0m\n\u001b[0;32m----> 1\u001b[0;31m \u001b[0mvocab\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mlookup_tokens\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m[\u001b[0m\u001b[0;36m0\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;36m1\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;36m2\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;36m10\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;36m2000\u001b[0m\u001b[0;34m]\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m", - "\u001b[0;31mNameError\u001b[0m: name 'vocab' is not defined" - ] - } - ], - "source": [ - "vocab.lookup_tokens([0, 1, 2, 10, 2000])" - ] - }, - { - "cell_type": "markdown", - "metadata": {}, - "source": [ - "#### Definicja sieci\n", - "\n" - ] - }, - { - "cell_type": "markdown", - "metadata": {}, - "source": [ - "Naszą prostą sieć neuronową zaimplementujemy używając frameworku PyTorch.\n", - "\n" - ] - }, - { - "cell_type": "code", - "execution_count": 13, - "metadata": {}, - "outputs": [], - "source": [ - "class SimpleBigramNeuralLanguageModel(nn.Module):\n", - " def __init__(self, vocabulary_size, embedding_size):\n", - " super(SimpleBigramNeuralLanguageModel, self).__init__()\n", - " self.model = nn.Sequential(\n", - " nn.Embedding(vocabulary_size, embedding_size),\n", - " nn.Linear(embedding_size, vocabulary_size),\n", - " nn.Softmax()\n", - " )\n", - " \n", - " def forward(self, x):\n", - " return self.model(x)\n", - "\n", - "with open('filename.pickle','rb') as handle:\n", - " vocab = pickle.load(handle)\n", - "\n", - "vocab.set_default_index(vocab[''])" - ] - }, - { - "cell_type": "code", - "execution_count": 6, - "metadata": {}, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Help on Vocab in module torchtext.vocab.vocab object:\n", - "\n", - "class Vocab(torch.nn.modules.module.Module)\n", - " | Vocab(vocab) -> None\n", - " | \n", - " | Base class for all neural network modules.\n", - " | \n", - " | Your models should also subclass this class.\n", - " | \n", - " | Modules can also contain other Modules, allowing to nest them in\n", - " | a tree structure. You can assign the submodules as regular attributes::\n", - " | \n", - " | import torch.nn as nn\n", - " | import torch.nn.functional as F\n", - " | \n", - " | class Model(nn.Module):\n", - " | def __init__(self):\n", - " | super().__init__()\n", - " | self.conv1 = nn.Conv2d(1, 20, 5)\n", - " | self.conv2 = nn.Conv2d(20, 20, 5)\n", - " | \n", - " | def forward(self, x):\n", - " | x = F.relu(self.conv1(x))\n", - " | return F.relu(self.conv2(x))\n", - " | \n", - " | Submodules assigned in this way will be registered, and will have their\n", - " | parameters converted too when you call :meth:`to`, etc.\n", - " | \n", - " | .. note::\n", - " | As per the example above, an ``__init__()`` call to the parent class\n", - " | must be made before assignment on the child.\n", - " | \n", - " | :ivar training: Boolean represents whether this module is in training or\n", - " | evaluation mode.\n", - " | :vartype training: bool\n", - " | \n", - " | Method resolution order:\n", - " | Vocab\n", - " | torch.nn.modules.module.Module\n", - " | builtins.object\n", - " | \n", - " | Methods defined here:\n", - " | \n", - " | __contains__(self, token: str) -> bool\n", - " | Args:\n", - " | token: The token for which to check the membership.\n", - " | \n", - " | Returns:\n", - " | Whether the token is member of vocab or not.\n", - " | \n", - " | __getitem__(self, token: str) -> int\n", - " | Args:\n", - " | token: The token used to lookup the corresponding index.\n", - " | \n", - " | Returns:\n", - " | The index corresponding to the associated token.\n", - " | \n", - " | __init__(self, vocab) -> None\n", - " | Initializes internal Module state, shared by both nn.Module and ScriptModule.\n", - " | \n", - " | __len__(self) -> int\n", - " | Returns:\n", - " | The length of the vocab.\n", - " | \n", - " | __prepare_scriptable__(self)\n", - " | Return a JITable Vocab.\n", - " | \n", - " | append_token(self, token: str) -> None\n", - " | Args:\n", - " | token: The token used to lookup the corresponding index.\n", - " | \n", - " | Raises:\n", - " | RuntimeError: If `token` already exists in the vocab\n", - " | \n", - " | forward(self, tokens: List[str]) -> List[int]\n", - " | Calls the `lookup_indices` method\n", - " | \n", - " | Args:\n", - " | tokens: a list of tokens used to lookup their corresponding `indices`.\n", - " | \n", - " | Returns:\n", - " | The indices associated with a list of `tokens`.\n", - " | \n", - " | get_default_index(self) -> Union[int, NoneType]\n", - " | Returns:\n", - " | Value of default index if it is set.\n", - " | \n", - " | get_itos(self) -> List[str]\n", - " | Returns:\n", - " | List mapping indices to tokens.\n", - " | \n", - " | get_stoi(self) -> Dict[str, int]\n", - " | Returns:\n", - " | Dictionary mapping tokens to indices.\n", - " | \n", - " | insert_token(self, token: str, index: int) -> None\n", - " | Args:\n", - " | token: The token used to lookup the corresponding index.\n", - " | index: The index corresponding to the associated token.\n", - " | Raises:\n", - " | RuntimeError: If `index` is not in range [0, Vocab.size()] or if `token` already exists in the vocab.\n", - " | \n", - " | lookup_indices(self, tokens: List[str]) -> List[int]\n", - " | Args:\n", - " | tokens: the tokens used to lookup their corresponding `indices`.\n", - " | \n", - " | Returns:\n", - " | The 'indices` associated with `tokens`.\n", - " | \n", - " | lookup_token(self, index: int) -> str\n", - " | Args:\n", - " | index: The index corresponding to the associated token.\n", - " | \n", - " | Returns:\n", - " | token: The token used to lookup the corresponding index.\n", - " | \n", - " | Raises:\n", - " | RuntimeError: If `index` not in range [0, itos.size()).\n", - " | \n", - " | lookup_tokens(self, indices: List[int]) -> List[str]\n", - " | Args:\n", - " | indices: The `indices` used to lookup their corresponding`tokens`.\n", - " | \n", - " | Returns:\n", - " | The `tokens` associated with `indices`.\n", - " | \n", - " | Raises:\n", - " | RuntimeError: If an index within `indices` is not int range [0, itos.size()).\n", - " | \n", - " | set_default_index(self, index: Union[int, NoneType]) -> None\n", - " | Args:\n", - " | index: Value of default index. This index will be returned when OOV token is queried.\n", - " | \n", - " | ----------------------------------------------------------------------\n", - " | Readonly properties defined here:\n", - " | \n", - " | is_jitable\n", - " | \n", - " | ----------------------------------------------------------------------\n", - " | Data and other attributes defined here:\n", - " | \n", - " | __jit_unused_properties__ = ['is_jitable']\n", - " | \n", - " | ----------------------------------------------------------------------\n", - " | Methods inherited from torch.nn.modules.module.Module:\n", - " | \n", - " | __call__ = _call_impl(self, *args, **kwargs)\n", - " | \n", - " | __delattr__(self, name)\n", - " | Implement delattr(self, name).\n", - " | \n", - " | __dir__(self)\n", - " | Default dir() implementation.\n", - " | \n", - " | __getattr__(self, name: str) -> Union[torch.Tensor, ForwardRef('Module')]\n", - " | \n", - " | __repr__(self)\n", - " | Return repr(self).\n", - " | \n", - " | __setattr__(self, name: str, value: Union[torch.Tensor, ForwardRef('Module')]) -> None\n", - " | Implement setattr(self, name, value).\n", - " | \n", - " | __setstate__(self, state)\n", - " | \n", - " | add_module(self, name: str, module: Union[ForwardRef('Module'), NoneType]) -> None\n", - " | Adds a child module to the current module.\n", - " | \n", - " | The module can be accessed as an attribute using the given name.\n", - " | \n", - " | Args:\n", - " | name (str): name of the child module. The child module can be\n", - " | accessed from this module using the given name\n", - " | module (Module): child module to be added to the module.\n", - " | \n", - " | apply(self: ~T, fn: Callable[[ForwardRef('Module')], NoneType]) -> ~T\n", - " | Applies ``fn`` recursively to every submodule (as returned by ``.children()``)\n", - " | as well as self. Typical use includes initializing the parameters of a model\n", - " | (see also :ref:`nn-init-doc`).\n", - " | \n", - " | Args:\n", - " | fn (:class:`Module` -> None): function to be applied to each submodule\n", - " | \n", - " | Returns:\n", - " | Module: self\n", - " | \n", - " | Example::\n", - " | \n", - " | >>> @torch.no_grad()\n", - " | >>> def init_weights(m):\n", - " | >>> print(m)\n", - " | >>> if type(m) == nn.Linear:\n", - " | >>> m.weight.fill_(1.0)\n", - " | >>> print(m.weight)\n", - " | >>> net = nn.Sequential(nn.Linear(2, 2), nn.Linear(2, 2))\n", - " | >>> net.apply(init_weights)\n", - " | Linear(in_features=2, out_features=2, bias=True)\n", - " | Parameter containing:\n", - " | tensor([[1., 1.],\n", - " | [1., 1.]], requires_grad=True)\n", - " | Linear(in_features=2, out_features=2, bias=True)\n", - " | Parameter containing:\n", - " | tensor([[1., 1.],\n", - " | [1., 1.]], requires_grad=True)\n", - " | Sequential(\n", - " | (0): Linear(in_features=2, out_features=2, bias=True)\n", - " | (1): Linear(in_features=2, out_features=2, bias=True)\n", - " | )\n", - " | \n", - " | bfloat16(self: ~T) -> ~T\n", - " | Casts all floating point parameters and buffers to ``bfloat16`` datatype.\n", - " | \n", - " | .. note::\n", - " | This method modifies the module in-place.\n", - " | \n", - " | Returns:\n", - " | Module: self\n", - " | \n", - " | buffers(self, recurse: bool = True) -> Iterator[torch.Tensor]\n", - " | Returns an iterator over module buffers.\n", - " | \n", - " | Args:\n", - " | recurse (bool): if True, then yields buffers of this module\n", - " | and all submodules. Otherwise, yields only buffers that\n", - " | are direct members of this module.\n", - " | \n", - " | Yields:\n", - " | torch.Tensor: module buffer\n", - " | \n", - " | Example::\n", - " | \n", - " | >>> # xdoctest: +SKIP(\"undefined vars\")\n", - " | >>> for buf in model.buffers():\n", - " | >>> print(type(buf), buf.size())\n", - " | (20L,)\n", - " | (20L, 1L, 5L, 5L)\n", - " | \n", - " | children(self) -> Iterator[ForwardRef('Module')]\n", - " | Returns an iterator over immediate children modules.\n", - " | \n", - " | Yields:\n", - " | Module: a child module\n", - " | \n", - " | cpu(self: ~T) -> ~T\n", - " | Moves all model parameters and buffers to the CPU.\n", - " | \n", - " | .. note::\n", - " | This method modifies the module in-place.\n", - " | \n", - " | Returns:\n", - " | Module: self\n", - " | \n", - " | cuda(self: ~T, device: Union[int, torch.device, NoneType] = None) -> ~T\n", - " | Moves all model parameters and buffers to the GPU.\n", - " | \n", - " | This also makes associated parameters and buffers different objects. So\n", - " | it should be called before constructing optimizer if the module will\n", - " | live on GPU while being optimized.\n", - " | \n", - " | .. note::\n", - " | This method modifies the module in-place.\n", - " | \n", - " | Args:\n", - " | device (int, optional): if specified, all parameters will be\n", - " | copied to that device\n", - " | \n", - " | Returns:\n", - " | Module: self\n", - " | \n", - " | double(self: ~T) -> ~T\n", - " | Casts all floating point parameters and buffers to ``double`` datatype.\n", - " | \n", - " | .. note::\n", - " | This method modifies the module in-place.\n", - " | \n", - " | Returns:\n", - " | Module: self\n", - " | \n", - " | eval(self: ~T) -> ~T\n", - " | Sets the module in evaluation mode.\n", - " | \n", - " | This has any effect only on certain modules. See documentations of\n", - " | particular modules for details of their behaviors in training/evaluation\n", - " | mode, if they are affected, e.g. :class:`Dropout`, :class:`BatchNorm`,\n", - " | etc.\n", - " | \n", - " | This is equivalent with :meth:`self.train(False) `.\n", - " | \n", - " | See :ref:`locally-disable-grad-doc` for a comparison between\n", - " | `.eval()` and several similar mechanisms that may be confused with it.\n", - " | \n", - " | Returns:\n", - " | Module: self\n", - " | \n", - " | extra_repr(self) -> str\n", - " | Set the extra representation of the module\n", - " | \n", - " | To print customized extra information, you should re-implement\n", - " | this method in your own modules. Both single-line and multi-line\n", - " | strings are acceptable.\n", - " | \n", - " | float(self: ~T) -> ~T\n", - " | Casts all floating point parameters and buffers to ``float`` datatype.\n", - " | \n", - " | .. note::\n", - " | This method modifies the module in-place.\n", - " | \n", - " | Returns:\n", - " | Module: self\n", - " | \n", - " | get_buffer(self, target: str) -> 'Tensor'\n", - " | Returns the buffer given by ``target`` if it exists,\n", - " | otherwise throws an error.\n", - " | \n", - " | See the docstring for ``get_submodule`` for a more detailed\n", - " | explanation of this method's functionality as well as how to\n", - " | correctly specify ``target``.\n", - " | \n", - " | Args:\n", - " | target: The fully-qualified string name of the buffer\n", - " | to look for. (See ``get_submodule`` for how to specify a\n", - " | fully-qualified string.)\n", - " | \n", - " | Returns:\n", - " | torch.Tensor: The buffer referenced by ``target``\n", - " | \n", - " | Raises:\n", - " | AttributeError: If the target string references an invalid\n", - " | path or resolves to something that is not a\n", - " | buffer\n", - " | \n", - " | get_extra_state(self) -> Any\n", - " | Returns any extra state to include in the module's state_dict.\n", - " | Implement this and a corresponding :func:`set_extra_state` for your module\n", - " | if you need to store extra state. This function is called when building the\n", - " | module's `state_dict()`.\n", - " | \n", - " | Note that extra state should be picklable to ensure working serialization\n", - " | of the state_dict. We only provide provide backwards compatibility guarantees\n", - " | for serializing Tensors; other objects may break backwards compatibility if\n", - " | their serialized pickled form changes.\n", - " | \n", - " | Returns:\n", - " | object: Any extra state to store in the module's state_dict\n", - " | \n", - " | get_parameter(self, target: str) -> 'Parameter'\n", - " | Returns the parameter given by ``target`` if it exists,\n", - " | otherwise throws an error.\n", - " | \n", - " | See the docstring for ``get_submodule`` for a more detailed\n", - " | explanation of this method's functionality as well as how to\n", - " | correctly specify ``target``.\n", - " | \n", - " | Args:\n", - " | target: The fully-qualified string name of the Parameter\n", - " | to look for. (See ``get_submodule`` for how to specify a\n", - " | fully-qualified string.)\n", - " | \n", - " | Returns:\n", - " | torch.nn.Parameter: The Parameter referenced by ``target``\n", - " | \n", - " | Raises:\n", - " | AttributeError: If the target string references an invalid\n", - " | path or resolves to something that is not an\n", - " | ``nn.Parameter``\n", - " | \n", - " | get_submodule(self, target: str) -> 'Module'\n", - " | Returns the submodule given by ``target`` if it exists,\n", - " | otherwise throws an error.\n", - " | \n", - " | For example, let's say you have an ``nn.Module`` ``A`` that\n", - " | looks like this:\n", - " | \n", - " | .. code-block:: text\n", - " | \n", - " | A(\n", - " | (net_b): Module(\n", - " | (net_c): Module(\n", - " | (conv): Conv2d(16, 33, kernel_size=(3, 3), stride=(2, 2))\n", - " | )\n", - " | (linear): Linear(in_features=100, out_features=200, bias=True)\n", - " | )\n", - " | )\n", - " | \n", - " | (The diagram shows an ``nn.Module`` ``A``. ``A`` has a nested\n", - " | submodule ``net_b``, which itself has two submodules ``net_c``\n", - " | and ``linear``. ``net_c`` then has a submodule ``conv``.)\n", - " | \n", - " | To check whether or not we have the ``linear`` submodule, we\n", - " | would call ``get_submodule(\"net_b.linear\")``. To check whether\n", - " | we have the ``conv`` submodule, we would call\n", - " | ``get_submodule(\"net_b.net_c.conv\")``.\n", - " | \n", - " | The runtime of ``get_submodule`` is bounded by the degree\n", - " | of module nesting in ``target``. A query against\n", - " | ``named_modules`` achieves the same result, but it is O(N) in\n", - " | the number of transitive modules. So, for a simple check to see\n", - " | if some submodule exists, ``get_submodule`` should always be\n", - " | used.\n", - " | \n", - " | Args:\n", - " | target: The fully-qualified string name of the submodule\n", - " | to look for. (See above example for how to specify a\n", - " | fully-qualified string.)\n", - " | \n", - " | Returns:\n", - " | torch.nn.Module: The submodule referenced by ``target``\n", - " | \n", - " | Raises:\n", - " | AttributeError: If the target string references an invalid\n", - " | path or resolves to something that is not an\n", - " | ``nn.Module``\n", - " | \n", - " | half(self: ~T) -> ~T\n", - " | Casts all floating point parameters and buffers to ``half`` datatype.\n", - " | \n", - " | .. note::\n", - " | This method modifies the module in-place.\n", - " | \n", - " | Returns:\n", - " | Module: self\n", - " | \n", - " | ipu(self: ~T, device: Union[int, torch.device, NoneType] = None) -> ~T\n", - " | Moves all model parameters and buffers to the IPU.\n", - " | \n", - " | This also makes associated parameters and buffers different objects. So\n", - " | it should be called before constructing optimizer if the module will\n", - " | live on IPU while being optimized.\n", - " | \n", - " | .. note::\n", - " | This method modifies the module in-place.\n", - " | \n", - " | Arguments:\n", - " | device (int, optional): if specified, all parameters will be\n", - " | copied to that device\n", - " | \n", - " | Returns:\n", - " | Module: self\n", - " | \n", - " | load_state_dict(self, state_dict: Mapping[str, Any], strict: bool = True)\n", - " | Copies parameters and buffers from :attr:`state_dict` into\n", - " | this module and its descendants. If :attr:`strict` is ``True``, then\n", - " | the keys of :attr:`state_dict` must exactly match the keys returned\n", - " | by this module's :meth:`~torch.nn.Module.state_dict` function.\n", - " | \n", - " | Args:\n", - " | state_dict (dict): a dict containing parameters and\n", - " | persistent buffers.\n", - " | strict (bool, optional): whether to strictly enforce that the keys\n", - " | in :attr:`state_dict` match the keys returned by this module's\n", - " | :meth:`~torch.nn.Module.state_dict` function. Default: ``True``\n", - " | \n", - " | Returns:\n", - " | ``NamedTuple`` with ``missing_keys`` and ``unexpected_keys`` fields:\n", - " | * **missing_keys** is a list of str containing the missing keys\n", - " | * **unexpected_keys** is a list of str containing the unexpected keys\n", - " | \n", - " | Note:\n", - " | If a parameter or buffer is registered as ``None`` and its corresponding key\n", - " | exists in :attr:`state_dict`, :meth:`load_state_dict` will raise a\n", - " | ``RuntimeError``.\n", - " | \n", - " | modules(self) -> Iterator[ForwardRef('Module')]\n", - " | Returns an iterator over all modules in the network.\n", - " | \n", - " | Yields:\n", - " | Module: a module in the network\n", - " | \n", - " | Note:\n", - " | Duplicate modules are returned only once. In the following\n", - " | example, ``l`` will be returned only once.\n", - " | \n", - " | Example::\n", - " | \n", - " | >>> l = nn.Linear(2, 2)\n", - " | >>> net = nn.Sequential(l, l)\n", - " | >>> for idx, m in enumerate(net.modules()):\n", - " | ... print(idx, '->', m)\n", - " | \n", - " | 0 -> Sequential(\n", - " | (0): Linear(in_features=2, out_features=2, bias=True)\n", - " | (1): Linear(in_features=2, out_features=2, bias=True)\n", - " | )\n", - " | 1 -> Linear(in_features=2, out_features=2, bias=True)\n", - " | \n", - " | named_buffers(self, prefix: str = '', recurse: bool = True, remove_duplicate: bool = True) -> Iterator[Tuple[str, torch.Tensor]]\n", - " | Returns an iterator over module buffers, yielding both the\n", - " | name of the buffer as well as the buffer itself.\n", - " | \n", - " | Args:\n", - " | prefix (str): prefix to prepend to all buffer names.\n", - " | recurse (bool, optional): if True, then yields buffers of this module\n", - " | and all submodules. Otherwise, yields only buffers that\n", - " | are direct members of this module. Defaults to True.\n", - " | remove_duplicate (bool, optional): whether to remove the duplicated buffers in the result. Defaults to True.\n", - " | \n", - " | Yields:\n", - " | (str, torch.Tensor): Tuple containing the name and buffer\n", - " | \n", - " | Example::\n", - " | \n", - " | >>> # xdoctest: +SKIP(\"undefined vars\")\n", - " | >>> for name, buf in self.named_buffers():\n", - " | >>> if name in ['running_var']:\n", - " | >>> print(buf.size())\n", - " | \n", - " | named_children(self) -> Iterator[Tuple[str, ForwardRef('Module')]]\n", - " | Returns an iterator over immediate children modules, yielding both\n", - " | the name of the module as well as the module itself.\n", - " | \n", - " | Yields:\n", - " | (str, Module): Tuple containing a name and child module\n", - " | \n", - " | Example::\n", - " | \n", - " | >>> # xdoctest: +SKIP(\"undefined vars\")\n", - " | >>> for name, module in model.named_children():\n", - " | >>> if name in ['conv4', 'conv5']:\n", - " | >>> print(module)\n", - " | \n", - " | named_modules(self, memo: Union[Set[ForwardRef('Module')], NoneType] = None, prefix: str = '', remove_duplicate: bool = True)\n", - " | Returns an iterator over all modules in the network, yielding\n", - " | both the name of the module as well as the module itself.\n", - " | \n", - " | Args:\n", - " | memo: a memo to store the set of modules already added to the result\n", - " | prefix: a prefix that will be added to the name of the module\n", - " | remove_duplicate: whether to remove the duplicated module instances in the result\n", - " | or not\n", - " | \n", - " | Yields:\n", - " | (str, Module): Tuple of name and module\n", - " | \n", - " | Note:\n", - " | Duplicate modules are returned only once. In the following\n", - " | example, ``l`` will be returned only once.\n", - " | \n", - " | Example::\n", - " | \n", - " | >>> l = nn.Linear(2, 2)\n", - " | >>> net = nn.Sequential(l, l)\n", - " | >>> for idx, m in enumerate(net.named_modules()):\n", - " | ... print(idx, '->', m)\n", - " | \n", - " | 0 -> ('', Sequential(\n", - " | (0): Linear(in_features=2, out_features=2, bias=True)\n", - " | (1): Linear(in_features=2, out_features=2, bias=True)\n", - " | ))\n", - " | 1 -> ('0', Linear(in_features=2, out_features=2, bias=True))\n", - " | \n", - " | named_parameters(self, prefix: str = '', recurse: bool = True, remove_duplicate: bool = True) -> Iterator[Tuple[str, torch.nn.parameter.Parameter]]\n", - " | Returns an iterator over module parameters, yielding both the\n", - " | name of the parameter as well as the parameter itself.\n", - " | \n", - " | Args:\n", - " | prefix (str): prefix to prepend to all parameter names.\n", - " | recurse (bool): if True, then yields parameters of this module\n", - " | and all submodules. Otherwise, yields only parameters that\n", - " | are direct members of this module.\n", - " | remove_duplicate (bool, optional): whether to remove the duplicated\n", - " | parameters in the result. Defaults to True.\n", - " | \n", - " | Yields:\n", - " | (str, Parameter): Tuple containing the name and parameter\n", - " | \n", - " | Example::\n", - " | \n", - " | >>> # xdoctest: +SKIP(\"undefined vars\")\n", - " | >>> for name, param in self.named_parameters():\n", - " | >>> if name in ['bias']:\n", - " | >>> print(param.size())\n", - " | \n", - " | parameters(self, recurse: bool = True) -> Iterator[torch.nn.parameter.Parameter]\n", - " | Returns an iterator over module parameters.\n", - " | \n", - " | This is typically passed to an optimizer.\n", - " | \n", - " | Args:\n", - " | recurse (bool): if True, then yields parameters of this module\n", - " | and all submodules. Otherwise, yields only parameters that\n", - " | are direct members of this module.\n", - " | \n", - " | Yields:\n", - " | Parameter: module parameter\n", - " | \n", - " | Example::\n", - " | \n", - " | >>> # xdoctest: +SKIP(\"undefined vars\")\n", - " | >>> for param in model.parameters():\n", - " | >>> print(type(param), param.size())\n", - " | (20L,)\n", - " | (20L, 1L, 5L, 5L)\n", - " | \n", - " | register_backward_hook(self, hook: Callable[[ForwardRef('Module'), Union[Tuple[torch.Tensor, ...], torch.Tensor], Union[Tuple[torch.Tensor, ...], torch.Tensor]], Union[NoneType, Tuple[torch.Tensor, ...], torch.Tensor]]) -> torch.utils.hooks.RemovableHandle\n", - " | Registers a backward hook on the module.\n", - " | \n", - " | This function is deprecated in favor of :meth:`~torch.nn.Module.register_full_backward_hook` and\n", - " | the behavior of this function will change in future versions.\n", - " | \n", - " | Returns:\n", - " | :class:`torch.utils.hooks.RemovableHandle`:\n", - " | a handle that can be used to remove the added hook by calling\n", - " | ``handle.remove()``\n", - " | \n", - " | register_buffer(self, name: str, tensor: Union[torch.Tensor, NoneType], persistent: bool = True) -> None\n", - " | Adds a buffer to the module.\n", - " | \n", - " | This is typically used to register a buffer that should not to be\n", - " | considered a model parameter. For example, BatchNorm's ``running_mean``\n", - " | is not a parameter, but is part of the module's state. Buffers, by\n", - " | default, are persistent and will be saved alongside parameters. This\n", - " | behavior can be changed by setting :attr:`persistent` to ``False``. The\n", - " | only difference between a persistent buffer and a non-persistent buffer\n", - " | is that the latter will not be a part of this module's\n", - " | :attr:`state_dict`.\n", - " | \n", - " | Buffers can be accessed as attributes using given names.\n", - " | \n", - " | Args:\n", - " | name (str): name of the buffer. The buffer can be accessed\n", - " | from this module using the given name\n", - " | tensor (Tensor or None): buffer to be registered. If ``None``, then operations\n", - " | that run on buffers, such as :attr:`cuda`, are ignored. If ``None``,\n", - " | the buffer is **not** included in the module's :attr:`state_dict`.\n", - " | persistent (bool): whether the buffer is part of this module's\n", - " | :attr:`state_dict`.\n", - " | \n", - " | Example::\n", - " | \n", - " | >>> # xdoctest: +SKIP(\"undefined vars\")\n", - " | >>> self.register_buffer('running_mean', torch.zeros(num_features))\n", - " | \n", - " | register_forward_hook(self, hook: Union[Callable[[~T, Tuple[Any, ...], Any], Union[Any, NoneType]], Callable[[~T, Tuple[Any, ...], Dict[str, Any], Any], Union[Any, NoneType]]], *, prepend: bool = False, with_kwargs: bool = False) -> torch.utils.hooks.RemovableHandle\n", - " | Registers a forward hook on the module.\n", - " | \n", - " | The hook will be called every time after :func:`forward` has computed an output.\n", - " | \n", - " | If ``with_kwargs`` is ``False`` or not specified, the input contains only\n", - " | the positional arguments given to the module. Keyword arguments won't be\n", - " | passed to the hooks and only to the ``forward``. The hook can modify the\n", - " | output. It can modify the input inplace but it will not have effect on\n", - " | forward since this is called after :func:`forward` is called. The hook\n", - " | should have the following signature::\n", - " | \n", - " | hook(module, args, output) -> None or modified output\n", - " | \n", - " | If ``with_kwargs`` is ``True``, the forward hook will be passed the\n", - " | ``kwargs`` given to the forward function and be expected to return the\n", - " | output possibly modified. The hook should have the following signature::\n", - " | \n", - " | hook(module, args, kwargs, output) -> None or modified output\n", - " | \n", - " | Args:\n", - " | hook (Callable): The user defined hook to be registered.\n", - " | prepend (bool): If ``True``, the provided ``hook`` will be fired\n", - " | before all existing ``forward`` hooks on this\n", - " | :class:`torch.nn.modules.Module`. Otherwise, the provided\n", - " | ``hook`` will be fired after all existing ``forward`` hooks on\n", - " | this :class:`torch.nn.modules.Module`. Note that global\n", - " | ``forward`` hooks registered with\n", - " | :func:`register_module_forward_hook` will fire before all hooks\n", - " | registered by this method.\n", - " | Default: ``False``\n", - " | with_kwargs (bool): If ``True``, the ``hook`` will be passed the\n", - " | kwargs given to the forward function.\n", - " | Default: ``False``\n", - " | \n", - " | Returns:\n", - " | :class:`torch.utils.hooks.RemovableHandle`:\n", - " | a handle that can be used to remove the added hook by calling\n", - " | ``handle.remove()``\n", - " | \n", - " | register_forward_pre_hook(self, hook: Union[Callable[[~T, Tuple[Any, ...]], Union[Any, NoneType]], Callable[[~T, Tuple[Any, ...], Dict[str, Any]], Union[Tuple[Any, Dict[str, Any]], NoneType]]], *, prepend: bool = False, with_kwargs: bool = False) -> torch.utils.hooks.RemovableHandle\n", - " | Registers a forward pre-hook on the module.\n", - " | \n", - " | The hook will be called every time before :func:`forward` is invoked.\n", - " | \n", - " | \n", - " | If ``with_kwargs`` is false or not specified, the input contains only\n", - " | the positional arguments given to the module. Keyword arguments won't be\n", - " | passed to the hooks and only to the ``forward``. The hook can modify the\n", - " | input. User can either return a tuple or a single modified value in the\n", - " | hook. We will wrap the value into a tuple if a single value is returned\n", - " | (unless that value is already a tuple). The hook should have the\n", - " | following signature::\n", - " | \n", - " | hook(module, args) -> None or modified input\n", - " | \n", - " | If ``with_kwargs`` is true, the forward pre-hook will be passed the\n", - " | kwargs given to the forward function. And if the hook modifies the\n", - " | input, both the args and kwargs should be returned. The hook should have\n", - " | the following signature::\n", - " | \n", - " | hook(module, args, kwargs) -> None or a tuple of modified input and kwargs\n", - " | \n", - " | Args:\n", - " | hook (Callable): The user defined hook to be registered.\n", - " | prepend (bool): If true, the provided ``hook`` will be fired before\n", - " | all existing ``forward_pre`` hooks on this\n", - " | :class:`torch.nn.modules.Module`. Otherwise, the provided\n", - " | ``hook`` will be fired after all existing ``forward_pre`` hooks\n", - " | on this :class:`torch.nn.modules.Module`. Note that global\n", - " | ``forward_pre`` hooks registered with\n", - " | :func:`register_module_forward_pre_hook` will fire before all\n", - " | hooks registered by this method.\n", - " | Default: ``False``\n", - " | with_kwargs (bool): If true, the ``hook`` will be passed the kwargs\n", - " | given to the forward function.\n", - " | Default: ``False``\n", - " | \n", - " | Returns:\n", - " | :class:`torch.utils.hooks.RemovableHandle`:\n", - " | a handle that can be used to remove the added hook by calling\n", - " | ``handle.remove()``\n", - " | \n", - " | register_full_backward_hook(self, hook: Callable[[ForwardRef('Module'), Union[Tuple[torch.Tensor, ...], torch.Tensor], Union[Tuple[torch.Tensor, ...], torch.Tensor]], Union[NoneType, Tuple[torch.Tensor, ...], torch.Tensor]], prepend: bool = False) -> torch.utils.hooks.RemovableHandle\n", - " | Registers a backward hook on the module.\n", - " | \n", - " | The hook will be called every time the gradients with respect to a module\n", - " | are computed, i.e. the hook will execute if and only if the gradients with\n", - " | respect to module outputs are computed. The hook should have the following\n", - " | signature::\n", - " | \n", - " | hook(module, grad_input, grad_output) -> tuple(Tensor) or None\n", - " | \n", - " | The :attr:`grad_input` and :attr:`grad_output` are tuples that contain the gradients\n", - " | with respect to the inputs and outputs respectively. The hook should\n", - " | not modify its arguments, but it can optionally return a new gradient with\n", - " | respect to the input that will be used in place of :attr:`grad_input` in\n", - " | subsequent computations. :attr:`grad_input` will only correspond to the inputs given\n", - " | as positional arguments and all kwarg arguments are ignored. Entries\n", - " | in :attr:`grad_input` and :attr:`grad_output` will be ``None`` for all non-Tensor\n", - " | arguments.\n", - " | \n", - " | For technical reasons, when this hook is applied to a Module, its forward function will\n", - " | receive a view of each Tensor passed to the Module. Similarly the caller will receive a view\n", - " | of each Tensor returned by the Module's forward function.\n", - " | \n", - " | .. warning ::\n", - " | Modifying inputs or outputs inplace is not allowed when using backward hooks and\n", - " | will raise an error.\n", - " | \n", - " | Args:\n", - " | hook (Callable): The user-defined hook to be registered.\n", - " | prepend (bool): If true, the provided ``hook`` will be fired before\n", - " | all existing ``backward`` hooks on this\n", - " | :class:`torch.nn.modules.Module`. Otherwise, the provided\n", - " | ``hook`` will be fired after all existing ``backward`` hooks on\n", - " | this :class:`torch.nn.modules.Module`. Note that global\n", - " | ``backward`` hooks registered with\n", - " | :func:`register_module_full_backward_hook` will fire before\n", - " | all hooks registered by this method.\n", - " | \n", - " | Returns:\n", - " | :class:`torch.utils.hooks.RemovableHandle`:\n", - " | a handle that can be used to remove the added hook by calling\n", - " | ``handle.remove()``\n", - " | \n", - " | register_full_backward_pre_hook(self, hook: Callable[[ForwardRef('Module'), Union[Tuple[torch.Tensor, ...], torch.Tensor]], Union[NoneType, Tuple[torch.Tensor, ...], torch.Tensor]], prepend: bool = False) -> torch.utils.hooks.RemovableHandle\n", - " | Registers a backward pre-hook on the module.\n", - " | \n", - " | The hook will be called every time the gradients for the module are computed.\n", - " | The hook should have the following signature::\n", - " | \n", - " | hook(module, grad_output) -> Tensor or None\n", - " | \n", - " | The :attr:`grad_output` is a tuple. The hook should\n", - " | not modify its arguments, but it can optionally return a new gradient with\n", - " | respect to the output that will be used in place of :attr:`grad_output` in\n", - " | subsequent computations. Entries in :attr:`grad_output` will be ``None`` for\n", - " | all non-Tensor arguments.\n", - " | \n", - " | For technical reasons, when this hook is applied to a Module, its forward function will\n", - " | receive a view of each Tensor passed to the Module. Similarly the caller will receive a view\n", - " | of each Tensor returned by the Module's forward function.\n", - " | \n", - " | .. warning ::\n", - " | Modifying inputs inplace is not allowed when using backward hooks and\n", - " | will raise an error.\n", - " | \n", - " | Args:\n", - " | hook (Callable): The user-defined hook to be registered.\n", - " | prepend (bool): If true, the provided ``hook`` will be fired before\n", - " | all existing ``backward_pre`` hooks on this\n", - " | :class:`torch.nn.modules.Module`. Otherwise, the provided\n", - " | ``hook`` will be fired after all existing ``backward_pre`` hooks\n", - " | on this :class:`torch.nn.modules.Module`. Note that global\n", - " | ``backward_pre`` hooks registered with\n", - " | :func:`register_module_full_backward_pre_hook` will fire before\n", - " | all hooks registered by this method.\n", - " | \n", - " | Returns:\n", - " | :class:`torch.utils.hooks.RemovableHandle`:\n", - " | a handle that can be used to remove the added hook by calling\n", - " | ``handle.remove()``\n", - " | \n", - " | register_load_state_dict_post_hook(self, hook)\n", - " | Registers a post hook to be run after module's ``load_state_dict``\n", - " | is called.\n", - " | \n", - " | It should have the following signature::\n", - " | hook(module, incompatible_keys) -> None\n", - " | \n", - " | The ``module`` argument is the current module that this hook is registered\n", - " | on, and the ``incompatible_keys`` argument is a ``NamedTuple`` consisting\n", - " | of attributes ``missing_keys`` and ``unexpected_keys``. ``missing_keys``\n", - " | is a ``list`` of ``str`` containing the missing keys and\n", - " | ``unexpected_keys`` is a ``list`` of ``str`` containing the unexpected keys.\n", - " | \n", - " | The given incompatible_keys can be modified inplace if needed.\n", - " | \n", - " | Note that the checks performed when calling :func:`load_state_dict` with\n", - " | ``strict=True`` are affected by modifications the hook makes to\n", - " | ``missing_keys`` or ``unexpected_keys``, as expected. Additions to either\n", - " | set of keys will result in an error being thrown when ``strict=True``, and\n", - " | clearing out both missing and unexpected keys will avoid an error.\n", - " | \n", - " | Returns:\n", - " | :class:`torch.utils.hooks.RemovableHandle`:\n", - " | a handle that can be used to remove the added hook by calling\n", - " | ``handle.remove()``\n", - " | \n", - " | register_module(self, name: str, module: Union[ForwardRef('Module'), NoneType]) -> None\n", - " | Alias for :func:`add_module`.\n", - " | \n", - " | register_parameter(self, name: str, param: Union[torch.nn.parameter.Parameter, NoneType]) -> None\n", - " | Adds a parameter to the module.\n", - " | \n", - " | The parameter can be accessed as an attribute using given name.\n", - " | \n", - " | Args:\n", - " | name (str): name of the parameter. The parameter can be accessed\n", - " | from this module using the given name\n", - " | param (Parameter or None): parameter to be added to the module. If\n", - " | ``None``, then operations that run on parameters, such as :attr:`cuda`,\n", - " | are ignored. If ``None``, the parameter is **not** included in the\n", - " | module's :attr:`state_dict`.\n", - " | \n", - " | register_state_dict_pre_hook(self, hook)\n", - " | These hooks will be called with arguments: ``self``, ``prefix``,\n", - " | and ``keep_vars`` before calling ``state_dict`` on ``self``. The registered\n", - " | hooks can be used to perform pre-processing before the ``state_dict``\n", - " | call is made.\n", - " | \n", - " | requires_grad_(self: ~T, requires_grad: bool = True) -> ~T\n", - " | Change if autograd should record operations on parameters in this\n", - " | module.\n", - " | \n", - " | This method sets the parameters' :attr:`requires_grad` attributes\n", - " | in-place.\n", - " | \n", - " | This method is helpful for freezing part of the module for finetuning\n", - " | or training parts of a model individually (e.g., GAN training).\n", - " | \n", - " | See :ref:`locally-disable-grad-doc` for a comparison between\n", - " | `.requires_grad_()` and several similar mechanisms that may be confused with it.\n", - " | \n", - " | Args:\n", - " | requires_grad (bool): whether autograd should record operations on\n", - " | parameters in this module. Default: ``True``.\n", - " | \n", - " | Returns:\n", - " | Module: self\n", - " | \n", - " | set_extra_state(self, state: Any)\n", - " | This function is called from :func:`load_state_dict` to handle any extra state\n", - " | found within the `state_dict`. Implement this function and a corresponding\n", - " | :func:`get_extra_state` for your module if you need to store extra state within its\n", - " | `state_dict`.\n", - " | \n", - " | Args:\n", - " | state (dict): Extra state from the `state_dict`\n", - " | \n", - " | share_memory(self: ~T) -> ~T\n", - " | See :meth:`torch.Tensor.share_memory_`\n", - " | \n", - " | state_dict(self, *args, destination=None, prefix='', keep_vars=False)\n", - " | Returns a dictionary containing references to the whole state of the module.\n", - " | \n", - " | Both parameters and persistent buffers (e.g. running averages) are\n", - " | included. Keys are corresponding parameter and buffer names.\n", - " | Parameters and buffers set to ``None`` are not included.\n", - " | \n", - " | .. note::\n", - " | The returned object is a shallow copy. It contains references\n", - " | to the module's parameters and buffers.\n", - " | \n", - " | .. warning::\n", - " | Currently ``state_dict()`` also accepts positional arguments for\n", - " | ``destination``, ``prefix`` and ``keep_vars`` in order. However,\n", - " | this is being deprecated and keyword arguments will be enforced in\n", - " | future releases.\n", - " | \n", - " | .. warning::\n", - " | Please avoid the use of argument ``destination`` as it is not\n", - " | designed for end-users.\n", - " | \n", - " | Args:\n", - " | destination (dict, optional): If provided, the state of module will\n", - " | be updated into the dict and the same object is returned.\n", - " | Otherwise, an ``OrderedDict`` will be created and returned.\n", - " | Default: ``None``.\n", - " | prefix (str, optional): a prefix added to parameter and buffer\n", - " | names to compose the keys in state_dict. Default: ``''``.\n", - " | keep_vars (bool, optional): by default the :class:`~torch.Tensor` s\n", - " | returned in the state dict are detached from autograd. If it's\n", - " | set to ``True``, detaching will not be performed.\n", - " | Default: ``False``.\n", - " | \n", - " | Returns:\n", - " | dict:\n", - " | a dictionary containing a whole state of the module\n", - " | \n", - " | Example::\n", - " | \n", - " | >>> # xdoctest: +SKIP(\"undefined vars\")\n", - " | >>> module.state_dict().keys()\n", - " | ['bias', 'weight']\n", - " | \n", - " | to(self, *args, **kwargs)\n", - " | Moves and/or casts the parameters and buffers.\n", - " | \n", - " | This can be called as\n", - " | \n", - " | .. function:: to(device=None, dtype=None, non_blocking=False)\n", - " | :noindex:\n", - " | \n", - " | .. function:: to(dtype, non_blocking=False)\n", - " | :noindex:\n", - " | \n", - " | .. function:: to(tensor, non_blocking=False)\n", - " | :noindex:\n", - " | \n", - " | .. function:: to(memory_format=torch.channels_last)\n", - " | :noindex:\n", - " | \n", - " | Its signature is similar to :meth:`torch.Tensor.to`, but only accepts\n", - " | floating point or complex :attr:`dtype`\\ s. In addition, this method will\n", - " | only cast the floating point or complex parameters and buffers to :attr:`dtype`\n", - " | (if given). The integral parameters and buffers will be moved\n", - " | :attr:`device`, if that is given, but with dtypes unchanged. When\n", - " | :attr:`non_blocking` is set, it tries to convert/move asynchronously\n", - " | with respect to the host if possible, e.g., moving CPU Tensors with\n", - " | pinned memory to CUDA devices.\n", - " | \n", - " | See below for examples.\n", - " | \n", - " | .. note::\n", - " | This method modifies the module in-place.\n", - " | \n", - " | Args:\n", - " | device (:class:`torch.device`): the desired device of the parameters\n", - " | and buffers in this module\n", - " | dtype (:class:`torch.dtype`): the desired floating point or complex dtype of\n", - " | the parameters and buffers in this module\n", - " | tensor (torch.Tensor): Tensor whose dtype and device are the desired\n", - " | dtype and device for all parameters and buffers in this module\n", - " | memory_format (:class:`torch.memory_format`): the desired memory\n", - " | format for 4D parameters and buffers in this module (keyword\n", - " | only argument)\n", - " | \n", - " | Returns:\n", - " | Module: self\n", - " | \n", - " | Examples::\n", - " | \n", - " | >>> # xdoctest: +IGNORE_WANT(\"non-deterministic\")\n", - " | >>> linear = nn.Linear(2, 2)\n", - " | >>> linear.weight\n", - " | Parameter containing:\n", - " | tensor([[ 0.1913, -0.3420],\n", - " | [-0.5113, -0.2325]])\n", - " | >>> linear.to(torch.double)\n", - " | Linear(in_features=2, out_features=2, bias=True)\n", - " | >>> linear.weight\n", - " | Parameter containing:\n", - " | tensor([[ 0.1913, -0.3420],\n", - " | [-0.5113, -0.2325]], dtype=torch.float64)\n", - " | >>> # xdoctest: +REQUIRES(env:TORCH_DOCTEST_CUDA1)\n", - " | >>> gpu1 = torch.device(\"cuda:1\")\n", - " | >>> linear.to(gpu1, dtype=torch.half, non_blocking=True)\n", - " | Linear(in_features=2, out_features=2, bias=True)\n", - " | >>> linear.weight\n", - " | Parameter containing:\n", - " | tensor([[ 0.1914, -0.3420],\n", - " | [-0.5112, -0.2324]], dtype=torch.float16, device='cuda:1')\n", - " | >>> cpu = torch.device(\"cpu\")\n", - " | >>> linear.to(cpu)\n", - " | Linear(in_features=2, out_features=2, bias=True)\n", - " | >>> linear.weight\n", - " | Parameter containing:\n", - " | tensor([[ 0.1914, -0.3420],\n", - " | [-0.5112, -0.2324]], dtype=torch.float16)\n", - " | \n", - " | >>> linear = nn.Linear(2, 2, bias=None).to(torch.cdouble)\n", - " | >>> linear.weight\n", - " | Parameter containing:\n", - " | tensor([[ 0.3741+0.j, 0.2382+0.j],\n", - " | [ 0.5593+0.j, -0.4443+0.j]], dtype=torch.complex128)\n", - " | >>> linear(torch.ones(3, 2, dtype=torch.cdouble))\n", - " | tensor([[0.6122+0.j, 0.1150+0.j],\n", - " | [0.6122+0.j, 0.1150+0.j],\n", - " | [0.6122+0.j, 0.1150+0.j]], dtype=torch.complex128)\n", - " | \n", - " | to_empty(self: ~T, *, device: Union[str, torch.device]) -> ~T\n", - " | Moves the parameters and buffers to the specified device without copying storage.\n", - " | \n", - " | Args:\n", - " | device (:class:`torch.device`): The desired device of the parameters\n", - " | and buffers in this module.\n", - " | \n", - " | Returns:\n", - " | Module: self\n", - " | \n", - " | train(self: ~T, mode: bool = True) -> ~T\n", - " | Sets the module in training mode.\n", - " | \n", - " | This has any effect only on certain modules. See documentations of\n", - " | particular modules for details of their behaviors in training/evaluation\n", - " | mode, if they are affected, e.g. :class:`Dropout`, :class:`BatchNorm`,\n", - " | etc.\n", - " | \n", - " | Args:\n", - " | mode (bool): whether to set training mode (``True``) or evaluation\n", - " | mode (``False``). Default: ``True``.\n", - " | \n", - " | Returns:\n", - " | Module: self\n", - " | \n", - " | type(self: ~T, dst_type: Union[torch.dtype, str]) -> ~T\n", - " | Casts all parameters and buffers to :attr:`dst_type`.\n", - " | \n", - " | .. note::\n", - " | This method modifies the module in-place.\n", - " | \n", - " | Args:\n", - " | dst_type (type or string): the desired type\n", - " | \n", - " | Returns:\n", - " | Module: self\n", - " | \n", - " | xpu(self: ~T, device: Union[int, torch.device, NoneType] = None) -> ~T\n", - " | Moves all model parameters and buffers to the XPU.\n", - " | \n", - " | This also makes associated parameters and buffers different objects. So\n", - " | it should be called before constructing optimizer if the module will\n", - " | live on XPU while being optimized.\n", - " | \n", - " | .. note::\n", - " | This method modifies the module in-place.\n", - " | \n", - " | Arguments:\n", - " | device (int, optional): if specified, all parameters will be\n", - " | copied to that device\n", - " | \n", - " | Returns:\n", - " | Module: self\n", - " | \n", - " | zero_grad(self, set_to_none: bool = True) -> None\n", - " | Sets gradients of all model parameters to zero. See similar function\n", - " | under :class:`torch.optim.Optimizer` for more context.\n", - " | \n", - " | Args:\n", - " | set_to_none (bool): instead of setting to zero, set the grads to None.\n", - " | See :meth:`torch.optim.Optimizer.zero_grad` for details.\n", - " | \n", - " | ----------------------------------------------------------------------\n", - " | Data descriptors inherited from torch.nn.modules.module.Module:\n", - " | \n", - " | __dict__\n", - " | dictionary for instance variables (if defined)\n", - " | \n", - " | __weakref__\n", - " | list of weak references to the object (if defined)\n", - " | \n", - " | ----------------------------------------------------------------------\n", - " | Data and other attributes inherited from torch.nn.modules.module.Module:\n", - " | \n", - " | T_destination = ~T_destination\n", - " | \n", - " | __annotations__ = {'__call__': typing.Callable[..., typing.Any], '_bac...\n", - " | \n", - " | call_super_init = False\n", - " | \n", - " | dump_patches = False\n", - "\n" - ] - } - ], - "source": [ - "help(vocab)" - ] - }, - { - "cell_type": "code", - "execution_count": 14, - "metadata": {}, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "1000\n", - "2000\n", - "3000\n", - "4000\n", - "5000\n", - "6000\n", - "7000\n", - "8000\n", - "9000\n", - "10000\n", - "11000\n", - "12000\n", - "13000\n", - "14000\n", - "15000\n", - "16000\n", - "17000\n", - "18000\n", - "19000\n", - "20000\n", - "21000\n", - "22000\n", - "23000\n", - "24000\n", - "25000\n", - "26000\n", - "27000\n", - "28000\n", - "29000\n", - "30000\n", - "31000\n", - "32000\n", - "33000\n", - "34000\n", - "35000\n", - "36000\n", - "37000\n", - "38000\n", - "39000\n", - "40000\n", - "41000\n", - "42000\n", - "43000\n", - "44000\n", - "45000\n", - "46000\n", - "47000\n", - "48000\n", - "49000\n", - "50000\n", - "51000\n", - "52000\n", - "53000\n", - "54000\n", - "55000\n", - "56000\n", - "57000\n", - "58000\n", - "59000\n", - "60000\n", - "61000\n", - "62000\n", - "63000\n", - "64000\n", - "65000\n", - "66000\n", - "67000\n", - "68000\n", - "69000\n", - "70000\n", - "71000\n", - "72000\n", - "73000\n", - "74000\n", - "75000\n", - "76000\n", - "77000\n", - "78000\n", - "79000\n", - "80000\n", - "81000\n", - "82000\n", - "83000\n", - "84000\n", - "85000\n", - "86000\n", - "87000\n", - "88000\n", - "89000\n", - "90000\n", - "91000\n", - "92000\n", - "93000\n", - "94000\n", - "95000\n", - "96000\n", - "97000\n", - "98000\n", - "99000\n", - "100000\n", - "101000\n", - "102000\n", - "103000\n", - "104000\n", - "105000\n", - "106000\n", - "107000\n", - "108000\n", - "109000\n", - "110000\n", - "111000\n", - "112000\n", - "113000\n", - "114000\n", - "115000\n", - "116000\n", - "117000\n", - "118000\n", - "119000\n", - "120000\n", - "121000\n", - "122000\n", - "123000\n", - "124000\n", - "125000\n", - "126000\n", - "127000\n", - "128000\n", - "129000\n", - "130000\n", - "131000\n", - "132000\n", - "133000\n", - "134000\n", - "135000\n", - "136000\n", - "137000\n", - "138000\n", - "139000\n", - "140000\n", - "141000\n", - "142000\n", - "143000\n", - "144000\n", - "145000\n", - "146000\n", - "147000\n", - "148000\n", - "149000\n", - "150000\n", - "151000\n", - "152000\n", - "153000\n", - "154000\n", - "155000\n", - "156000\n", - "157000\n", - "158000\n", - "159000\n", - "160000\n", - "161000\n", - "162000\n", - "163000\n", - "164000\n", - "165000\n", - "166000\n", - "167000\n", - "168000\n", - "169000\n", - "170000\n", - "171000\n", - "172000\n", - "173000\n", - "174000\n", - "175000\n", - "176000\n", - "177000\n", - "178000\n", - "179000\n", - "180000\n", - "181000\n", - "182000\n", - "183000\n", - "184000\n", - "185000\n", - "186000\n", - "187000\n", - "188000\n", - "189000\n", - "190000\n", - "191000\n", - "192000\n", - "193000\n", - "194000\n", - "195000\n", - "196000\n", - "197000\n", - "198000\n", - "199000\n", - "200000\n", - "201000\n", - "202000\n", - "203000\n", - "204000\n", - "205000\n", - "206000\n", - "207000\n", - "208000\n", - "209000\n", - "210000\n", - "211000\n", - "212000\n", - "213000\n", - "214000\n", - "215000\n", - "216000\n", - "217000\n", - "218000\n", - "219000\n", - "220000\n", - "221000\n", - "222000\n", - "223000\n", - "224000\n", - "225000\n", - "226000\n", - "227000\n", - "228000\n", - "229000\n", - "230000\n", - "231000\n", - "232000\n", - "233000\n", - "234000\n", - "235000\n", - "236000\n", - "237000\n", - "238000\n", - "239000\n", - "240000\n", - "241000\n", - "242000\n", - "243000\n", - "244000\n", - "245000\n", - "246000\n", - "247000\n", - "248000\n", - "249000\n", - "250000\n", - "251000\n", - "252000\n", - "253000\n", - "254000\n", - "255000\n", - "256000\n", - "257000\n", - "258000\n", - "259000\n", - "260000\n", - "261000\n", - "262000\n", - "263000\n", - "264000\n", - "265000\n", - "266000\n", - "267000\n", - "268000\n", - "269000\n", - "270000\n", - "271000\n", - "272000\n", - "273000\n", - "274000\n", - "275000\n", - "276000\n", - "277000\n", - "278000\n", - "279000\n", - "280000\n", - "281000\n", - "282000\n", - "283000\n", - "284000\n", - "285000\n", - "286000\n", - "287000\n", - "288000\n", - "289000\n", - "290000\n", - "291000\n", - "292000\n", - "293000\n", - "294000\n", - "295000\n", - "296000\n", - "297000\n", - "298000\n", - "299000\n", - "300000\n", - "301000\n", - "302000\n", - "303000\n", - "304000\n", - "305000\n", - "306000\n", - "307000\n", - "308000\n", - "309000\n", - "310000\n", - "311000\n", - "312000\n", - "313000\n", - "314000\n", - "315000\n", - "316000\n", - "317000\n", - "318000\n", - "319000\n", - "320000\n", - "321000\n", - "322000\n", - "323000\n", - "324000\n", - "325000\n", - "326000\n", - "327000\n", - "328000\n", - "329000\n", - "330000\n", - "331000\n", - "332000\n", - "333000\n", - "334000\n", - "335000\n", - "336000\n", - "337000\n", - "338000\n", - "339000\n", - "340000\n", - "341000\n", - "342000\n", - "343000\n", - "344000\n", - "345000\n", - "346000\n", - "347000\n", - "348000\n", - "349000\n", - "350000\n", - "351000\n", - "352000\n", - "353000\n", - "354000\n", - "355000\n", - "356000\n", - "357000\n", - "358000\n", - "359000\n", - "360000\n", - "361000\n", - "362000\n", - "363000\n", - "364000\n", - "365000\n", - "366000\n", - "367000\n", - "368000\n", - "369000\n", - "370000\n", - "371000\n", - "372000\n", - "373000\n", - "374000\n", - "375000\n", - "376000\n", - "377000\n", - "378000\n", - "379000\n", - "380000\n", - "381000\n", - "382000\n", - "383000\n", - "384000\n", - "385000\n", - "386000\n", - "387000\n", - "388000\n", - "389000\n", - "390000\n", - "391000\n", - "392000\n", - "393000\n", - "394000\n", - "395000\n", - "396000\n", - "397000\n", - "398000\n", - "399000\n", - "400000\n", - "401000\n", - "402000\n", - "403000\n", - "404000\n", - "405000\n", - "406000\n", - "407000\n", - "408000\n", - "409000\n", - "410000\n", - "411000\n", - "412000\n", - "413000\n", - "414000\n", - "415000\n", - "416000\n", - "417000\n", - "418000\n", - "419000\n", - "420000\n", - "421000\n", - "422000\n", - "423000\n", - "424000\n", - "425000\n", - "426000\n", - "427000\n", - "428000\n", - "429000\n", - "430000\n", - "431000\n", - "432000\n" - ] - } - ], - "source": [ - "def look_ahead_iterator(gen):\n", - " prev = None\n", - " for item in gen:\n", - " if prev is not None:\n", - " yield (prev, item)\n", - " prev = item\n", - "\n", - "class Bigrams(IterableDataset):\n", - " def __init__(self, text_file, vocabulary_size):\n", - " self.vocab = build_vocab_from_iterator(\n", - " get_word_lines_from_file(text_file),\n", - " max_tokens = vocabulary_size,\n", - " specials = [''])\n", - " self.vocab.set_default_index(self.vocab[''])\n", - " self.vocabulary_size = vocabulary_size\n", - " self.text_file = text_file\n", - "\n", - " def __iter__(self):\n", - " return look_ahead_iterator(\n", - " (self.vocab[t] for t in itertools.chain.from_iterable(get_word_lines_from_file(self.text_file))))\n", - "\n", - "train_dataset = Bigrams(train_file, vocab_size)" - ] - }, - { - "cell_type": "code", - "execution_count": 15, - "metadata": {}, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "<__main__.Bigrams object at 0x7fdd26d23940>\n" - ] - } - ], - "source": [ - "print(train_dataset)" - ] - }, - { - "cell_type": "code", - "execution_count": 17, - "metadata": { - "collapsed": true - }, - "outputs": [ - { - "data": { - "text/plain": [ - "'|===========================================================================|\\n| PyTorch CUDA memory summary, device ID 0 |\\n|---------------------------------------------------------------------------|\\n| CUDA OOMs: 1 | cudaMalloc retries: 1 |\\n|===========================================================================|\\n| Metric | Cur Usage | Peak Usage | Tot Alloc | Tot Freed |\\n|---------------------------------------------------------------------------|\\n| Allocated memory | 699613 KiB | 1903 MiB | 3735 MiB | 3052 MiB |\\n| from large pool | 699414 KiB | 1903 MiB | 3734 MiB | 3051 MiB |\\n| from small pool | 199 KiB | 1 MiB | 1 MiB | 1 MiB |\\n|---------------------------------------------------------------------------|\\n| Active memory | 699613 KiB | 1903 MiB | 3735 MiB | 3052 MiB |\\n| from large pool | 699414 KiB | 1903 MiB | 3734 MiB | 3051 MiB |\\n| from small pool | 199 KiB | 1 MiB | 1 MiB | 1 MiB |\\n|---------------------------------------------------------------------------|\\n| Requested memory | 699611 KiB | 1903 MiB | 3735 MiB | 3052 MiB |\\n| from large pool | 699413 KiB | 1903 MiB | 3734 MiB | 3051 MiB |\\n| from small pool | 197 KiB | 1 MiB | 1 MiB | 1 MiB |\\n|---------------------------------------------------------------------------|\\n| GPU reserved memory | 710656 KiB | 1918 MiB | 1918 MiB | 1224 MiB |\\n| from large pool | 708608 KiB | 1916 MiB | 1916 MiB | 1224 MiB |\\n| from small pool | 2048 KiB | 2 MiB | 2 MiB | 0 MiB |\\n|---------------------------------------------------------------------------|\\n| Non-releasable memory | 11043 KiB | 19364 KiB | 28939 KiB | 17896 KiB |\\n| from large pool | 9194 KiB | 17514 KiB | 25954 KiB | 16760 KiB |\\n| from small pool | 1849 KiB | 1950 KiB | 2985 KiB | 1136 KiB |\\n|---------------------------------------------------------------------------|\\n| Allocations | 10 | 17 | 38 | 28 |\\n| from large pool | 5 | 7 | 10 | 5 |\\n| from small pool | 5 | 11 | 28 | 23 |\\n|---------------------------------------------------------------------------|\\n| Active allocs | 10 | 17 | 38 | 28 |\\n| from large pool | 5 | 7 | 10 | 5 |\\n| from small pool | 5 | 11 | 28 | 23 |\\n|---------------------------------------------------------------------------|\\n| GPU reserved segments | 5 | 7 | 7 | 2 |\\n| from large pool | 4 | 6 | 6 | 2 |\\n| from small pool | 1 | 1 | 1 | 0 |\\n|---------------------------------------------------------------------------|\\n| Non-releasable allocs | 6 | 8 | 20 | 14 |\\n| from large pool | 4 | 6 | 9 | 5 |\\n| from small pool | 2 | 3 | 11 | 9 |\\n|---------------------------------------------------------------------------|\\n| Oversize allocations | 0 | 0 | 0 | 0 |\\n|---------------------------------------------------------------------------|\\n| Oversize GPU segments | 0 | 0 | 0 | 0 |\\n|===========================================================================|\\n'" - ] - }, - "execution_count": 17, - "metadata": {}, - "output_type": "execute_result" - } - ], - "source": [ - "torch.cuda.memory_summary(device=None, abbreviated=False)" - ] - }, - { - "cell_type": "code", - "execution_count": 10, - "metadata": {}, - "outputs": [], - "source": [ - "import os\n", - "os.environ[\"PYTORCH_CUDA_ALLOC_CONF\"] = \"max_split_size_mb:256\"" - ] - }, - { - "cell_type": "code", - "execution_count": 12, - "metadata": {}, - "outputs": [], - "source": [ - "device = 'cuda'\n", - "model = SimpleBigramNeuralLanguageModel(vocab_size, embed_size).to(device)" - ] - }, - { - "cell_type": "code", - "execution_count": 13, - "metadata": {}, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "epoch: = 1\n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "/home/gedin/.local/lib/python3.8/site-packages/torch/nn/modules/container.py:217: UserWarning: Implicit dimension choice for softmax has been deprecated. Change the call to include dim=X as an argument.\n", - " input = module(input)\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "0 tensor(5.9599, device='cuda:0', grad_fn=)\n", - "1000\n", - "100 tensor(6.1015, device='cuda:0', grad_fn=)\n", - "200 tensor(5.9708, device='cuda:0', grad_fn=)\n", - "2000\n", - "300 tensor(6.2176, device='cuda:0', grad_fn=)\n", - "3000\n", - "400 tensor(5.9401, device='cuda:0', grad_fn=)\n", - "4000\n", - "500 tensor(6.2084, device='cuda:0', grad_fn=)\n", - "5000\n", - "600 tensor(5.9736, device='cuda:0', grad_fn=)\n", - "6000\n", - "700 tensor(6.1423, device='cuda:0', grad_fn=)\n", - "7000\n", - "800 tensor(5.7344, device='cuda:0', grad_fn=)\n", - "8000\n", - "900 tensor(6.0950, device='cuda:0', grad_fn=)\n", - "9000\n", - "1000 tensor(5.8473, device='cuda:0', grad_fn=)\n", - "10000\n", - "1100 tensor(6.0612, device='cuda:0', grad_fn=)\n", - "11000\n", - "1200 tensor(6.1509, device='cuda:0', grad_fn=)\n", - "12000\n", - "1300 tensor(6.0760, device='cuda:0', grad_fn=)\n", - "13000\n", - "1400 tensor(6.2047, device='cuda:0', grad_fn=)\n", - "14000\n", - "1500 tensor(6.1186, device='cuda:0', grad_fn=)\n", - "15000\n", - "1600 tensor(5.8722, device='cuda:0', grad_fn=)\n", - "16000\n", - "1700 tensor(5.8741, device='cuda:0', grad_fn=)\n", - "17000\n", - "1800 tensor(5.8971, device='cuda:0', grad_fn=)\n", - "18000\n", - "1900 tensor(5.8521, device='cuda:0', grad_fn=)\n", - "19000\n", - "2000 tensor(5.9434, device='cuda:0', grad_fn=)\n", - "20000\n", - "2100 tensor(6.0348, device='cuda:0', grad_fn=)\n", - "21000\n", - "2200 tensor(5.8840, device='cuda:0', grad_fn=)\n", - "22000\n", - "2300 tensor(5.8641, device='cuda:0', grad_fn=)\n", - "23000\n", - "2400 tensor(5.9068, device='cuda:0', grad_fn=)\n", - "24000\n", - "2500 tensor(5.9170, device='cuda:0', grad_fn=)\n", - "25000\n", - "2600 tensor(5.9812, device='cuda:0', grad_fn=)\n", - "26000\n", - "2700 tensor(5.8985, device='cuda:0', grad_fn=)\n", - "27000\n", - "2800 tensor(6.0008, device='cuda:0', grad_fn=)\n", - "28000\n", - "2900 tensor(6.1230, device='cuda:0', grad_fn=)\n", - "29000\n", - "3000 tensor(5.8770, device='cuda:0', grad_fn=)\n", - "30000\n", - "3100 tensor(5.9268, device='cuda:0', grad_fn=)\n", - "31000\n", - "3200 tensor(5.8530, device='cuda:0', grad_fn=)\n", - "32000\n", - "3300 tensor(5.8436, device='cuda:0', grad_fn=)\n", - "33000\n", - "3400 tensor(5.7692, device='cuda:0', grad_fn=)\n", - "34000\n", - "3500 tensor(5.8909, device='cuda:0', grad_fn=)\n", - "35000\n", - "3600 tensor(5.8325, device='cuda:0', grad_fn=)\n", - "36000\n", - "3700 tensor(5.8082, device='cuda:0', grad_fn=)\n", - "37000\n", - "3800 tensor(5.8106, device='cuda:0', grad_fn=)\n", - "38000\n", - "3900 tensor(5.6382, device='cuda:0', grad_fn=)\n", - "39000\n", - "4000 tensor(5.6596, device='cuda:0', grad_fn=)\n", - "40000\n", - "4100 tensor(5.9587, device='cuda:0', grad_fn=)\n", - "41000\n", - "4200 tensor(5.8862, device='cuda:0', grad_fn=)\n", - "42000\n", - "4300 tensor(5.9541, device='cuda:0', grad_fn=)\n", - "43000\n", - "4400 tensor(5.8681, device='cuda:0', grad_fn=)\n", - "44000\n", - "4500 tensor(5.6963, device='cuda:0', grad_fn=)\n", - "45000\n", - "4600 tensor(6.0707, device='cuda:0', grad_fn=)\n", - "46000\n", - "4700 tensor(5.7091, device='cuda:0', grad_fn=)\n", - "47000\n", - "4800 tensor(5.8139, device='cuda:0', grad_fn=)\n", - "48000\n", - "4900 tensor(5.8696, device='cuda:0', grad_fn=)\n", - "49000\n", - "5000 tensor(5.8844, device='cuda:0', grad_fn=)\n", - "50000\n", - "5100 tensor(5.9806, device='cuda:0', grad_fn=)\n", - "51000\n", - "5200 tensor(6.0075, device='cuda:0', grad_fn=)\n", - "52000\n", - "5300 tensor(6.0588, device='cuda:0', grad_fn=)\n", - "53000\n", - "5400 tensor(5.8456, device='cuda:0', grad_fn=)\n", - "54000\n", - "5500 tensor(5.9166, device='cuda:0', grad_fn=)\n", - "55000\n", - "5600 tensor(5.6528, device='cuda:0', grad_fn=)\n", - "56000\n", - "5700 tensor(5.8988, device='cuda:0', grad_fn=)\n", - "57000\n", - "5800 tensor(5.9132, device='cuda:0', grad_fn=)\n", - "58000\n", - "5900 tensor(5.9460, device='cuda:0', grad_fn=)\n", - "59000\n", - "6000 tensor(5.7543, device='cuda:0', grad_fn=)\n", - "60000\n", - "6100 tensor(5.8256, device='cuda:0', grad_fn=)\n", - "61000\n", - "6200 tensor(5.9448, device='cuda:0', grad_fn=)\n", - "62000\n", - "6300 tensor(5.7601, device='cuda:0', grad_fn=)\n", - "63000\n", - "6400 tensor(5.7091, device='cuda:0', grad_fn=)\n", - "64000\n", - "6500 tensor(5.5621, device='cuda:0', grad_fn=)\n", - "65000\n", - "6600 tensor(5.7094, device='cuda:0', grad_fn=)\n", - "66000\n", - "6700 tensor(5.6785, device='cuda:0', grad_fn=)\n", - "67000\n", - "6800 tensor(5.9249, device='cuda:0', grad_fn=)\n", - "68000\n", - "6900 tensor(5.8775, device='cuda:0', grad_fn=)\n", - "69000\n", - "7000 tensor(5.8075, device='cuda:0', grad_fn=)\n", - "70000\n", - "7100 tensor(5.5748, device='cuda:0', grad_fn=)\n", - "71000\n", - "7200 tensor(5.7217, device='cuda:0', grad_fn=)\n", - "72000\n", - "7300 tensor(5.9124, device='cuda:0', grad_fn=)\n", - "73000\n", - "7400 tensor(5.7197, device='cuda:0', grad_fn=)\n", - "74000\n", - "7500 tensor(5.6429, device='cuda:0', grad_fn=)\n", - "75000\n", - "7600 tensor(5.6847, device='cuda:0', grad_fn=)\n", - "76000\n", - "7700 tensor(5.7197, device='cuda:0', grad_fn=)\n", - "77000\n", - "7800 tensor(5.8559, device='cuda:0', grad_fn=)\n", - "78000\n", - "7900 tensor(5.5600, device='cuda:0', grad_fn=)\n", - "79000\n", - "8000 tensor(5.6288, device='cuda:0', grad_fn=)\n", - "80000\n", - "8100 tensor(5.7767, device='cuda:0', grad_fn=)\n", - "81000\n", - "8200 tensor(5.8037, device='cuda:0', grad_fn=)\n", - "82000\n", - "8300 tensor(5.7344, device='cuda:0', grad_fn=)\n", - "83000\n", - "8400 tensor(5.8092, device='cuda:0', grad_fn=)\n", - "84000\n", - "8500 tensor(5.8847, device='cuda:0', grad_fn=)\n", - "85000\n", - "8600 tensor(5.8754, device='cuda:0', grad_fn=)\n", - "86000\n", - "8700 tensor(5.9227, device='cuda:0', grad_fn=)\n", - "87000\n", - "8800 tensor(5.8028, device='cuda:0', grad_fn=)\n", - "88000\n", - "8900 tensor(5.6476, device='cuda:0', grad_fn=)\n", - "89000\n", - "9000 tensor(5.7656, device='cuda:0', grad_fn=)\n", - "90000\n", - "9100 tensor(5.7805, device='cuda:0', grad_fn=)\n", - "91000\n", - "9200 tensor(5.6879, device='cuda:0', grad_fn=)\n", - "92000\n", - "9300 tensor(5.7098, device='cuda:0', grad_fn=)\n", - "93000\n", - "9400 tensor(5.5631, device='cuda:0', grad_fn=)\n", - "94000\n", - "9500 tensor(5.6497, device='cuda:0', grad_fn=)\n", - "95000\n", - "9600 tensor(5.7500, device='cuda:0', grad_fn=)\n", - "96000\n", - "9700 tensor(5.6607, device='cuda:0', grad_fn=)\n", - "97000\n", - "9800 tensor(5.7196, device='cuda:0', grad_fn=)\n", - "9900 tensor(5.5987, device='cuda:0', grad_fn=)\n", - "98000\n", - "10000 tensor(5.7795, device='cuda:0', grad_fn=)\n", - "99000\n", - "10100 tensor(5.6980, device='cuda:0', grad_fn=)\n", - "100000\n", - "10200 tensor(5.6093, device='cuda:0', grad_fn=)\n", - "101000\n", - "10300 tensor(5.6792, device='cuda:0', grad_fn=)\n", - "102000\n", - "10400 tensor(5.7035, device='cuda:0', grad_fn=)\n", - "103000\n", - "10500 tensor(5.8282, device='cuda:0', grad_fn=)\n", - "104000\n", - "10600 tensor(5.8605, device='cuda:0', grad_fn=)\n", - "105000\n", - "10700 tensor(5.7354, device='cuda:0', grad_fn=)\n", - "106000\n", - "10800 tensor(5.8034, device='cuda:0', grad_fn=)\n", - "107000\n", - "10900 tensor(5.6194, device='cuda:0', grad_fn=)\n", - "108000\n", - "11000 tensor(5.8502, device='cuda:0', grad_fn=)\n", - "109000\n", - "11100 tensor(5.4406, device='cuda:0', grad_fn=)\n", - "110000\n", - "11200 tensor(5.6379, device='cuda:0', grad_fn=)\n", - "111000\n", - "11300 tensor(5.6668, device='cuda:0', grad_fn=)\n", - "112000\n", - "11400 tensor(5.6140, device='cuda:0', grad_fn=)\n", - "113000\n", - "11500 tensor(5.6565, device='cuda:0', grad_fn=)\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "114000\n", - "11600 tensor(5.6308, device='cuda:0', grad_fn=)\n", - "115000\n", - "11700 tensor(5.5680, device='cuda:0', grad_fn=)\n", - "116000\n", - "11800 tensor(5.7604, device='cuda:0', grad_fn=)\n", - "117000\n", - "11900 tensor(5.5792, device='cuda:0', grad_fn=)\n", - "118000\n", - "12000 tensor(5.7329, device='cuda:0', grad_fn=)\n", - "119000\n", - "12100 tensor(5.7726, device='cuda:0', grad_fn=)\n", - "120000\n", - "12200 tensor(5.7151, device='cuda:0', grad_fn=)\n", - "121000\n", - "12300 tensor(5.8561, device='cuda:0', grad_fn=)\n", - "122000\n", - "12400 tensor(5.6791, device='cuda:0', grad_fn=)\n", - "123000\n", - "12500 tensor(5.5574, device='cuda:0', grad_fn=)\n", - "124000\n", - "12600 tensor(5.6817, device='cuda:0', grad_fn=)\n", - "125000\n", - "12700 tensor(5.5375, device='cuda:0', grad_fn=)\n", - "126000\n", - "12800 tensor(5.7270, device='cuda:0', grad_fn=)\n", - "127000\n", - "12900 tensor(5.6252, device='cuda:0', grad_fn=)\n", - "128000\n", - "13000 tensor(5.4536, device='cuda:0', grad_fn=)\n", - "129000\n", - "13100 tensor(5.6091, device='cuda:0', grad_fn=)\n", - "130000\n", - "13200 tensor(5.7324, device='cuda:0', grad_fn=)\n", - "131000\n", - "13300 tensor(5.5253, device='cuda:0', grad_fn=)\n", - "132000\n", - "13400 tensor(5.6491, device='cuda:0', grad_fn=)\n", - "133000\n", - "13500 tensor(5.5728, device='cuda:0', grad_fn=)\n", - "134000\n", - "13600 tensor(5.6632, device='cuda:0', grad_fn=)\n", - "135000\n", - "13700 tensor(5.6678, device='cuda:0', grad_fn=)\n", - "136000\n", - "13800 tensor(5.6112, device='cuda:0', grad_fn=)\n", - "137000\n", - "13900 tensor(5.4884, device='cuda:0', grad_fn=)\n", - "138000\n", - "14000 tensor(5.7304, device='cuda:0', grad_fn=)\n", - "139000\n", - "14100 tensor(5.4326, device='cuda:0', grad_fn=)\n", - "140000\n", - "14200 tensor(5.7188, device='cuda:0', grad_fn=)\n", - "141000\n", - "14300 tensor(5.6519, device='cuda:0', grad_fn=)\n", - "142000\n", - "14400 tensor(5.5892, device='cuda:0', grad_fn=)\n", - "143000\n", - "14500 tensor(5.7225, device='cuda:0', grad_fn=)\n", - "144000\n", - "14600 tensor(5.7216, device='cuda:0', grad_fn=)\n", - "145000\n", - "14700 tensor(5.5748, device='cuda:0', grad_fn=)\n", - "146000\n", - "14800 tensor(6.0184, device='cuda:0', grad_fn=)\n", - "147000\n", - "14900 tensor(5.6781, device='cuda:0', grad_fn=)\n", - "148000\n", - "15000 tensor(5.6038, device='cuda:0', grad_fn=)\n", - "149000\n", - "15100 tensor(5.7875, device='cuda:0', grad_fn=)\n", - "150000\n", - "15200 tensor(5.6485, device='cuda:0', grad_fn=)\n", - "151000\n", - "15300 tensor(5.5927, device='cuda:0', grad_fn=)\n", - "152000\n", - "15400 tensor(5.5156, device='cuda:0', grad_fn=)\n", - "153000\n", - "15500 tensor(5.6556, device='cuda:0', grad_fn=)\n", - "154000\n", - "15600 tensor(5.6485, device='cuda:0', grad_fn=)\n", - "155000\n", - "15700 tensor(5.5904, device='cuda:0', grad_fn=)\n", - "156000\n", - "15800 tensor(5.4613, device='cuda:0', grad_fn=)\n", - "157000\n", - "15900 tensor(5.6254, device='cuda:0', grad_fn=)\n", - "158000\n", - "16000 tensor(5.4349, device='cuda:0', grad_fn=)\n", - "159000\n", - "16100 tensor(5.5205, device='cuda:0', grad_fn=)\n", - "160000\n", - "16200 tensor(5.8051, device='cuda:0', grad_fn=)\n", - "161000\n", - "16300 tensor(5.6452, device='cuda:0', grad_fn=)\n", - "162000\n", - "16400 tensor(5.6071, device='cuda:0', grad_fn=)\n", - "163000\n", - "16500 tensor(5.7237, device='cuda:0', grad_fn=)\n", - "164000\n", - "16600 tensor(5.5771, device='cuda:0', grad_fn=)\n", - "165000\n", - "16700 tensor(5.5355, device='cuda:0', grad_fn=)\n", - "166000\n", - "16800 tensor(5.6363, device='cuda:0', grad_fn=)\n", - "167000\n", - "16900 tensor(5.3746, device='cuda:0', grad_fn=)\n", - "168000\n", - "17000 tensor(5.6707, device='cuda:0', grad_fn=)\n", - "169000\n", - "17100 tensor(5.5359, device='cuda:0', grad_fn=)\n", - "170000\n", - "17200 tensor(5.6118, device='cuda:0', grad_fn=)\n", - "171000\n", - "17300 tensor(5.6740, device='cuda:0', grad_fn=)\n", - "172000\n", - "17400 tensor(5.4438, device='cuda:0', grad_fn=)\n", - "173000\n", - "17500 tensor(5.5001, device='cuda:0', grad_fn=)\n", - "174000\n", - "17600 tensor(5.4953, device='cuda:0', grad_fn=)\n", - "175000\n", - "17700 tensor(5.5398, device='cuda:0', grad_fn=)\n", - "176000\n", - "17800 tensor(5.6053, device='cuda:0', grad_fn=)\n", - "177000\n", - "17900 tensor(5.4726, device='cuda:0', grad_fn=)\n", - "178000\n", - "18000 tensor(5.6747, device='cuda:0', grad_fn=)\n", - "179000\n", - "18100 tensor(5.6238, device='cuda:0', grad_fn=)\n", - "180000\n", - "18200 tensor(5.5469, device='cuda:0', grad_fn=)\n", - "181000\n", - "18300 tensor(5.5299, device='cuda:0', grad_fn=)\n", - "182000\n", - "18400 tensor(5.6323, device='cuda:0', grad_fn=)\n", - "183000\n", - "18500 tensor(5.5893, device='cuda:0', grad_fn=)\n", - "184000\n", - "18600 tensor(5.7452, device='cuda:0', grad_fn=)\n", - "185000\n", - "18700 tensor(5.5576, device='cuda:0', grad_fn=)\n", - "186000\n", - "18800 tensor(5.7439, device='cuda:0', grad_fn=)\n", - "187000\n", - "18900 tensor(5.6106, device='cuda:0', grad_fn=)\n", - "188000\n", - "19000 tensor(5.6647, device='cuda:0', grad_fn=)\n", - "189000\n", - "19100 tensor(5.7728, device='cuda:0', grad_fn=)\n", - "190000\n", - "19200 tensor(5.6169, device='cuda:0', grad_fn=)\n", - "191000\n", - "19300 tensor(5.7852, device='cuda:0', grad_fn=)\n", - "192000\n", - "19400 tensor(5.5627, device='cuda:0', grad_fn=)\n", - "193000\n", - "19500 tensor(5.5682, device='cuda:0', grad_fn=)\n", - "194000\n", - "19600 tensor(5.5978, device='cuda:0', grad_fn=)\n", - "195000\n", - "19700 tensor(5.6453, device='cuda:0', grad_fn=)\n", - "196000\n", - "19800 tensor(5.4786, device='cuda:0', grad_fn=)\n", - "197000\n", - "19900 tensor(5.4894, device='cuda:0', grad_fn=)\n", - "198000\n", - "20000 tensor(5.4999, device='cuda:0', grad_fn=)\n", - "199000\n", - "20100 tensor(5.4881, device='cuda:0', grad_fn=)\n", - "200000\n", - "20200 tensor(5.3915, device='cuda:0', grad_fn=)\n", - "201000\n", - "20300 tensor(5.5216, device='cuda:0', grad_fn=)\n", - "20400 tensor(5.5761, device='cuda:0', grad_fn=)\n", - "202000\n", - "20500 tensor(5.5586, device='cuda:0', grad_fn=)\n", - "203000\n", - "20600 tensor(5.7870, device='cuda:0', grad_fn=)\n", - "204000\n", - "20700 tensor(5.5776, device='cuda:0', grad_fn=)\n", - "205000\n", - "20800 tensor(5.4417, device='cuda:0', grad_fn=)\n", - "206000\n", - "20900 tensor(5.7186, device='cuda:0', grad_fn=)\n", - "207000\n", - "21000 tensor(5.5415, device='cuda:0', grad_fn=)\n", - "208000\n", - "21100 tensor(5.5141, device='cuda:0', grad_fn=)\n", - "209000\n", - "21200 tensor(5.4401, device='cuda:0', grad_fn=)\n", - "210000\n", - "21300 tensor(5.6511, device='cuda:0', grad_fn=)\n", - "211000\n", - "21400 tensor(5.6474, device='cuda:0', grad_fn=)\n", - "212000\n", - "21500 tensor(5.3946, device='cuda:0', grad_fn=)\n", - "213000\n", - "21600 tensor(5.3958, device='cuda:0', grad_fn=)\n", - "214000\n", - "21700 tensor(5.4040, device='cuda:0', grad_fn=)\n", - "215000\n", - "21800 tensor(5.5745, device='cuda:0', grad_fn=)\n", - "216000\n", - "21900 tensor(5.4996, device='cuda:0', grad_fn=)\n", - "217000\n", - "22000 tensor(5.5234, device='cuda:0', grad_fn=)\n", - "218000\n", - "22100 tensor(5.3870, device='cuda:0', grad_fn=)\n", - "219000\n", - "22200 tensor(5.2661, device='cuda:0', grad_fn=)\n", - "220000\n", - "22300 tensor(5.7031, device='cuda:0', grad_fn=)\n", - "221000\n", - "22400 tensor(5.3633, device='cuda:0', grad_fn=)\n", - "222000\n", - "22500 tensor(5.4404, device='cuda:0', grad_fn=)\n", - "223000\n", - "22600 tensor(5.5951, device='cuda:0', grad_fn=)\n", - "224000\n", - "22700 tensor(5.3901, device='cuda:0', grad_fn=)\n", - "225000\n", - "22800 tensor(5.6404, device='cuda:0', grad_fn=)\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "226000\n", - "22900 tensor(5.6646, device='cuda:0', grad_fn=)\n", - "227000\n", - "23000 tensor(5.5949, device='cuda:0', grad_fn=)\n", - "228000\n", - "23100 tensor(5.5284, device='cuda:0', grad_fn=)\n", - "229000\n", - "23200 tensor(5.5617, device='cuda:0', grad_fn=)\n", - "230000\n", - "23300 tensor(5.6426, device='cuda:0', grad_fn=)\n", - "231000\n", - "23400 tensor(5.7283, device='cuda:0', grad_fn=)\n", - "232000\n", - "23500 tensor(5.4558, device='cuda:0', grad_fn=)\n", - "233000\n", - "23600 tensor(5.4600, device='cuda:0', grad_fn=)\n", - "234000\n", - "23700 tensor(5.4961, device='cuda:0', grad_fn=)\n", - "235000\n", - "23800 tensor(5.3373, device='cuda:0', grad_fn=)\n", - "236000\n", - "23900 tensor(5.4470, device='cuda:0', grad_fn=)\n", - "237000\n", - "24000 tensor(5.4346, device='cuda:0', grad_fn=)\n", - "238000\n", - "24100 tensor(5.5112, device='cuda:0', grad_fn=)\n", - "239000\n", - "24200 tensor(5.6918, device='cuda:0', grad_fn=)\n", - "240000\n", - "24300 tensor(5.6115, device='cuda:0', grad_fn=)\n", - "241000\n", - "24400 tensor(5.7404, device='cuda:0', grad_fn=)\n", - "242000\n", - "24500 tensor(5.4982, device='cuda:0', grad_fn=)\n", - "243000\n", - "24600 tensor(5.6136, device='cuda:0', grad_fn=)\n", - "244000\n", - "24700 tensor(5.5225, device='cuda:0', grad_fn=)\n", - "245000\n", - "24800 tensor(5.5563, device='cuda:0', grad_fn=)\n", - "246000\n", - "24900 tensor(5.6283, device='cuda:0', grad_fn=)\n", - "247000\n", - "25000 tensor(5.6176, device='cuda:0', grad_fn=)\n", - "248000\n", - "25100 tensor(5.5795, device='cuda:0', grad_fn=)\n", - "249000\n", - "25200 tensor(5.5831, device='cuda:0', grad_fn=)\n", - "250000\n", - "25300 tensor(5.5894, device='cuda:0', grad_fn=)\n", - "251000\n", - "25400 tensor(5.5670, device='cuda:0', grad_fn=)\n", - "252000\n", - "25500 tensor(5.5016, device='cuda:0', grad_fn=)\n", - "253000\n", - "25600 tensor(5.7909, device='cuda:0', grad_fn=)\n", - "254000\n", - "25700 tensor(5.5229, device='cuda:0', grad_fn=)\n", - "255000\n", - "25800 tensor(5.6035, device='cuda:0', grad_fn=)\n", - "256000\n", - "25900 tensor(5.5293, device='cuda:0', grad_fn=)\n", - "257000\n", - "26000 tensor(5.5553, device='cuda:0', grad_fn=)\n", - "258000\n", - "26100 tensor(5.4476, device='cuda:0', grad_fn=)\n", - "259000\n", - "26200 tensor(5.3721, device='cuda:0', grad_fn=)\n", - "260000\n", - "26300 tensor(5.6142, device='cuda:0', grad_fn=)\n", - "261000\n", - "26400 tensor(5.6202, device='cuda:0', grad_fn=)\n", - "262000\n", - "26500 tensor(5.3529, device='cuda:0', grad_fn=)\n", - "263000\n", - "26600 tensor(5.7148, device='cuda:0', grad_fn=)\n", - "264000\n", - "26700 tensor(5.5755, device='cuda:0', grad_fn=)\n", - "265000\n", - "26800 tensor(5.7480, device='cuda:0', grad_fn=)\n", - "266000\n", - "26900 tensor(5.5025, device='cuda:0', grad_fn=)\n", - "267000\n", - "27000 tensor(5.4017, device='cuda:0', grad_fn=)\n", - "268000\n", - "27100 tensor(5.3996, device='cuda:0', grad_fn=)\n", - "269000\n", - "27200 tensor(5.4862, device='cuda:0', grad_fn=)\n", - "270000\n", - "27300 tensor(5.6392, device='cuda:0', grad_fn=)\n", - "271000\n", - "27400 tensor(5.5634, device='cuda:0', grad_fn=)\n", - "272000\n", - "27500 tensor(5.4420, device='cuda:0', grad_fn=)\n", - "273000\n", - "27600 tensor(5.7835, device='cuda:0', grad_fn=)\n", - "274000\n", - "27700 tensor(5.5555, device='cuda:0', grad_fn=)\n", - "275000\n", - "27800 tensor(5.5381, device='cuda:0', grad_fn=)\n", - "276000\n", - "27900 tensor(5.6515, device='cuda:0', grad_fn=)\n", - "277000\n", - "28000 tensor(5.5254, device='cuda:0', grad_fn=)\n", - "278000\n", - "28100 tensor(5.4929, device='cuda:0', grad_fn=)\n", - "279000\n", - "28200 tensor(5.6218, device='cuda:0', grad_fn=)\n", - "280000\n", - "28300 tensor(5.2878, device='cuda:0', grad_fn=)\n", - "281000\n", - "28400 tensor(5.7112, device='cuda:0', grad_fn=)\n", - "282000\n", - "28500 tensor(5.5490, device='cuda:0', grad_fn=)\n", - "283000\n", - "28600 tensor(5.4572, device='cuda:0', grad_fn=)\n", - "284000\n", - "28700 tensor(5.6349, device='cuda:0', grad_fn=)\n", - "285000\n", - "28800 tensor(5.6607, device='cuda:0', grad_fn=)\n", - "286000\n", - "28900 tensor(5.5422, device='cuda:0', grad_fn=)\n", - "287000\n", - "29000 tensor(5.4277, device='cuda:0', grad_fn=)\n", - "288000\n", - "29100 tensor(5.1870, device='cuda:0', grad_fn=)\n", - "289000\n", - "29200 tensor(5.3593, device='cuda:0', grad_fn=)\n", - "290000\n", - "29300 tensor(5.6512, device='cuda:0', grad_fn=)\n", - "291000\n", - "29400 tensor(5.8051, device='cuda:0', grad_fn=)\n", - "292000\n", - "29500 tensor(5.5308, device='cuda:0', grad_fn=)\n", - "293000\n", - "29600 tensor(5.3791, device='cuda:0', grad_fn=)\n", - "294000\n", - "29700 tensor(5.6108, device='cuda:0', grad_fn=)\n", - "295000\n", - "29800 tensor(5.4015, device='cuda:0', grad_fn=)\n", - "296000\n", - "29900 tensor(5.6953, device='cuda:0', grad_fn=)\n", - "297000\n", - "30000 tensor(5.3925, device='cuda:0', grad_fn=)\n", - "298000\n", - "30100 tensor(5.4241, device='cuda:0', grad_fn=)\n", - "299000\n", - "30200 tensor(5.4216, device='cuda:0', grad_fn=)\n", - "300000\n", - "30300 tensor(5.5074, device='cuda:0', grad_fn=)\n", - "301000\n", - "30400 tensor(5.3631, device='cuda:0', grad_fn=)\n", - "302000\n", - "30500 tensor(5.5690, device='cuda:0', grad_fn=)\n", - "30600 tensor(5.4734, device='cuda:0', grad_fn=)\n", - "303000\n", - "30700 tensor(5.5061, device='cuda:0', grad_fn=)\n", - "304000\n", - "30800 tensor(5.5709, device='cuda:0', grad_fn=)\n", - "305000\n", - "30900 tensor(5.5478, device='cuda:0', grad_fn=)\n", - "306000\n", - "31000 tensor(5.6687, device='cuda:0', grad_fn=)\n", - "307000\n", - "31100 tensor(5.2899, device='cuda:0', grad_fn=)\n", - "308000\n", - "31200 tensor(5.3663, device='cuda:0', grad_fn=)\n", - "309000\n", - "31300 tensor(5.6274, device='cuda:0', grad_fn=)\n", - "310000\n", - "31400 tensor(5.4358, device='cuda:0', grad_fn=)\n", - "311000\n", - "31500 tensor(5.5738, device='cuda:0', grad_fn=)\n", - "312000\n", - "31600 tensor(5.5612, device='cuda:0', grad_fn=)\n", - "313000\n", - "31700 tensor(5.5104, device='cuda:0', grad_fn=)\n", - "314000\n", - "31800 tensor(5.6343, device='cuda:0', grad_fn=)\n", - "315000\n", - "31900 tensor(5.2243, device='cuda:0', grad_fn=)\n", - "316000\n", - "32000 tensor(5.4320, device='cuda:0', grad_fn=)\n", - "317000\n", - "32100 tensor(5.3344, device='cuda:0', grad_fn=)\n", - "318000\n", - "32200 tensor(5.6543, device='cuda:0', grad_fn=)\n", - "319000\n", - "32300 tensor(5.6512, device='cuda:0', grad_fn=)\n", - "320000\n", - "32400 tensor(5.6237, device='cuda:0', grad_fn=)\n", - "321000\n", - "32500 tensor(5.4246, device='cuda:0', grad_fn=)\n", - "322000\n", - "32600 tensor(5.5469, device='cuda:0', grad_fn=)\n", - "323000\n", - "32700 tensor(5.5338, device='cuda:0', grad_fn=)\n", - "324000\n", - "32800 tensor(5.6954, device='cuda:0', grad_fn=)\n", - "325000\n", - "32900 tensor(5.5754, device='cuda:0', grad_fn=)\n", - "326000\n", - "33000 tensor(5.3334, device='cuda:0', grad_fn=)\n", - "327000\n", - "33100 tensor(5.5284, device='cuda:0', grad_fn=)\n", - "328000\n", - "33200 tensor(5.6350, device='cuda:0', grad_fn=)\n", - "329000\n", - "33300 tensor(5.4312, device='cuda:0', grad_fn=)\n", - "330000\n", - "33400 tensor(5.6854, device='cuda:0', grad_fn=)\n", - "331000\n", - "33500 tensor(5.4921, device='cuda:0', grad_fn=)\n", - "332000\n", - "33600 tensor(5.4345, device='cuda:0', grad_fn=)\n", - "333000\n", - "33700 tensor(5.4950, device='cuda:0', grad_fn=)\n", - "334000\n", - "33800 tensor(5.5757, device='cuda:0', grad_fn=)\n", - "335000\n", - "33900 tensor(5.3466, device='cuda:0', grad_fn=)\n", - "336000\n", - "34000 tensor(5.5373, device='cuda:0', grad_fn=)\n", - "337000\n", - "34100 tensor(5.5144, device='cuda:0', grad_fn=)\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "338000\n", - "34200 tensor(5.5543, device='cuda:0', grad_fn=)\n", - "339000\n", - "34300 tensor(5.3564, device='cuda:0', grad_fn=)\n", - "340000\n", - "34400 tensor(5.8091, device='cuda:0', grad_fn=)\n", - "341000\n", - "34500 tensor(5.6699, device='cuda:0', grad_fn=)\n", - "342000\n", - "34600 tensor(5.5536, device='cuda:0', grad_fn=)\n", - "343000\n", - "34700 tensor(5.6261, device='cuda:0', grad_fn=)\n", - "344000\n", - "34800 tensor(5.6504, device='cuda:0', grad_fn=)\n", - "345000\n", - "34900 tensor(5.7067, device='cuda:0', grad_fn=)\n", - "346000\n", - "35000 tensor(5.7307, device='cuda:0', grad_fn=)\n", - "347000\n", - "35100 tensor(5.4831, device='cuda:0', grad_fn=)\n", - "348000\n", - "35200 tensor(5.4367, device='cuda:0', grad_fn=)\n", - "349000\n", - "35300 tensor(5.6503, device='cuda:0', grad_fn=)\n", - "350000\n", - "35400 tensor(5.2892, device='cuda:0', grad_fn=)\n", - "351000\n", - "35500 tensor(5.4198, device='cuda:0', grad_fn=)\n", - "352000\n", - "35600 tensor(5.4870, device='cuda:0', grad_fn=)\n", - "353000\n", - "35700 tensor(5.4489, device='cuda:0', grad_fn=)\n", - "354000\n", - "35800 tensor(5.5170, device='cuda:0', grad_fn=)\n", - "355000\n", - "35900 tensor(5.4699, device='cuda:0', grad_fn=)\n", - "356000\n", - "36000 tensor(5.2451, device='cuda:0', grad_fn=)\n", - "357000\n", - "36100 tensor(5.6311, device='cuda:0', grad_fn=)\n", - "358000\n", - "36200 tensor(5.5157, device='cuda:0', grad_fn=)\n", - "359000\n", - "36300 tensor(5.7751, device='cuda:0', grad_fn=)\n", - "360000\n", - "36400 tensor(5.4740, device='cuda:0', grad_fn=)\n", - "361000\n", - "36500 tensor(5.4746, device='cuda:0', grad_fn=)\n", - "362000\n", - "36600 tensor(5.5244, device='cuda:0', grad_fn=)\n", - "363000\n", - "36700 tensor(5.3037, device='cuda:0', grad_fn=)\n", - "364000\n", - "36800 tensor(5.4238, device='cuda:0', grad_fn=)\n", - "365000\n", - "36900 tensor(5.5203, device='cuda:0', grad_fn=)\n", - "366000\n", - "37000 tensor(5.4431, device='cuda:0', grad_fn=)\n", - "367000\n", - "37100 tensor(5.4286, device='cuda:0', grad_fn=)\n", - "368000\n", - "37200 tensor(5.5108, device='cuda:0', grad_fn=)\n", - "369000\n", - "37300 tensor(5.4229, device='cuda:0', grad_fn=)\n", - "370000\n", - "37400 tensor(5.8406, device='cuda:0', grad_fn=)\n", - "371000\n", - "37500 tensor(5.4602, device='cuda:0', grad_fn=)\n", - "372000\n", - "37600 tensor(5.4417, device='cuda:0', grad_fn=)\n", - "373000\n", - "37700 tensor(5.6200, device='cuda:0', grad_fn=)\n", - "374000\n", - "37800 tensor(5.4527, device='cuda:0', grad_fn=)\n", - "375000\n", - "37900 tensor(5.4631, device='cuda:0', grad_fn=)\n", - "376000\n", - "38000 tensor(5.5196, device='cuda:0', grad_fn=)\n", - "377000\n", - "38100 tensor(5.5436, device='cuda:0', grad_fn=)\n", - "378000\n", - "38200 tensor(5.5269, device='cuda:0', grad_fn=)\n", - "379000\n", - "38300 tensor(5.4716, device='cuda:0', grad_fn=)\n", - "380000\n", - "38400 tensor(5.5081, device='cuda:0', grad_fn=)\n", - "381000\n", - "38500 tensor(5.5249, device='cuda:0', grad_fn=)\n", - "382000\n", - "38600 tensor(5.5018, device='cuda:0', grad_fn=)\n", - "383000\n", - "38700 tensor(5.4845, device='cuda:0', grad_fn=)\n", - "384000\n", - "38800 tensor(5.5505, device='cuda:0', grad_fn=)\n", - "385000\n", - "38900 tensor(5.6658, device='cuda:0', grad_fn=)\n", - "386000\n", - "39000 tensor(5.3333, device='cuda:0', grad_fn=)\n", - "387000\n", - "39100 tensor(5.5598, device='cuda:0', grad_fn=)\n", - "388000\n", - "39200 tensor(5.6624, device='cuda:0', grad_fn=)\n", - "389000\n", - "39300 tensor(5.4714, device='cuda:0', grad_fn=)\n", - "390000\n", - "39400 tensor(5.5470, device='cuda:0', grad_fn=)\n", - "391000\n", - "39500 tensor(5.6905, device='cuda:0', grad_fn=)\n", - "392000\n", - "39600 tensor(5.3592, device='cuda:0', grad_fn=)\n", - "393000\n", - "39700 tensor(5.3170, device='cuda:0', grad_fn=)\n", - "394000\n", - "39800 tensor(5.4491, device='cuda:0', grad_fn=)\n", - "395000\n", - "39900 tensor(5.2872, device='cuda:0', grad_fn=)\n", - "396000\n", - "40000 tensor(5.3865, device='cuda:0', grad_fn=)\n", - "397000\n", - "40100 tensor(5.4536, device='cuda:0', grad_fn=)\n", - "398000\n", - "40200 tensor(5.4382, device='cuda:0', grad_fn=)\n", - "399000\n", - "40300 tensor(5.4819, device='cuda:0', grad_fn=)\n", - "40400 tensor(5.5250, device='cuda:0', grad_fn=)\n", - "400000\n", - "40500 tensor(5.4396, device='cuda:0', grad_fn=)\n", - "401000\n", - "40600 tensor(5.5062, device='cuda:0', grad_fn=)\n", - "402000\n", - "40700 tensor(5.5362, device='cuda:0', grad_fn=)\n", - "403000\n", - "40800 tensor(5.5015, device='cuda:0', grad_fn=)\n", - "404000\n", - "40900 tensor(5.4610, device='cuda:0', grad_fn=)\n", - "405000\n", - "41000 tensor(5.5083, device='cuda:0', grad_fn=)\n", - "406000\n", - "41100 tensor(5.4346, device='cuda:0', grad_fn=)\n", - "407000\n", - "41200 tensor(5.3340, device='cuda:0', grad_fn=)\n", - "408000\n", - "41300 tensor(5.4608, device='cuda:0', grad_fn=)\n", - "409000\n", - "41400 tensor(5.3758, device='cuda:0', grad_fn=)\n", - "410000\n", - "41500 tensor(5.5160, device='cuda:0', grad_fn=)\n", - "411000\n", - "41600 tensor(5.4290, device='cuda:0', grad_fn=)\n", - "412000\n", - "41700 tensor(5.4426, device='cuda:0', grad_fn=)\n", - "413000\n", - "41800 tensor(5.4764, device='cuda:0', grad_fn=)\n", - "414000\n", - "41900 tensor(5.4730, device='cuda:0', grad_fn=)\n", - "415000\n", - "42000 tensor(5.6150, device='cuda:0', grad_fn=)\n", - "416000\n", - "42100 tensor(5.3622, device='cuda:0', grad_fn=)\n", - "417000\n", - "42200 tensor(5.4380, device='cuda:0', grad_fn=)\n", - "418000\n", - "42300 tensor(5.5031, device='cuda:0', grad_fn=)\n", - "419000\n", - "42400 tensor(5.3124, device='cuda:0', grad_fn=)\n", - "420000\n", - "42500 tensor(5.4812, device='cuda:0', grad_fn=)\n", - "421000\n", - "42600 tensor(5.2723, device='cuda:0', grad_fn=)\n", - "422000\n", - "42700 tensor(5.5998, device='cuda:0', grad_fn=)\n", - "423000\n", - "42800 tensor(5.5254, device='cuda:0', grad_fn=)\n", - "424000\n", - "42900 tensor(5.3716, device='cuda:0', grad_fn=)\n", - "425000\n", - "43000 tensor(5.5020, device='cuda:0', grad_fn=)\n", - "426000\n", - "43100 tensor(5.5091, device='cuda:0', grad_fn=)\n", - "427000\n", - "43200 tensor(5.3182, device='cuda:0', grad_fn=)\n", - "428000\n", - "43300 tensor(5.4001, device='cuda:0', grad_fn=)\n", - "429000\n", - "43400 tensor(5.5150, device='cuda:0', grad_fn=)\n", - "430000\n", - "43500 tensor(5.2440, device='cuda:0', grad_fn=)\n", - "431000\n", - "43600 tensor(5.4439, device='cuda:0', grad_fn=)\n", - "432000\n", - "epoch: = 2\n", - "0 tensor(5.3953, device='cuda:0', grad_fn=)\n", - "1000\n", - "100 tensor(5.4847, device='cuda:0', grad_fn=)\n", - "200 tensor(5.3626, device='cuda:0', grad_fn=)\n", - "2000\n", - "300 tensor(5.4127, device='cuda:0', grad_fn=)\n", - "3000\n", - "400 tensor(5.3734, device='cuda:0', grad_fn=)\n", - "4000\n", - "500 tensor(5.5564, device='cuda:0', grad_fn=)\n", - "5000\n", - "600 tensor(5.3391, device='cuda:0', grad_fn=)\n", - "6000\n", - "700 tensor(5.6198, device='cuda:0', grad_fn=)\n", - "7000\n", - "800 tensor(5.2255, device='cuda:0', grad_fn=)\n", - "8000\n", - "900 tensor(5.5161, device='cuda:0', grad_fn=)\n", - "9000\n", - "1000 tensor(5.3517, device='cuda:0', grad_fn=)\n", - "10000\n", - "1100 tensor(5.5420, device='cuda:0', grad_fn=)\n", - "11000\n", - "1200 tensor(5.6031, device='cuda:0', grad_fn=)\n", - "12000\n", - "1300 tensor(5.5343, device='cuda:0', grad_fn=)\n", - "13000\n", - "1400 tensor(5.5547, device='cuda:0', grad_fn=)\n", - "14000\n", - "1500 tensor(5.6080, device='cuda:0', grad_fn=)\n", - "15000\n", - "1600 tensor(5.2940, device='cuda:0', grad_fn=)\n", - "16000\n", - "1700 tensor(5.3671, device='cuda:0', grad_fn=)\n", - "17000\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "1800 tensor(5.3777, device='cuda:0', grad_fn=)\n", - "18000\n", - "1900 tensor(5.3593, device='cuda:0', grad_fn=)\n", - "19000\n", - "2000 tensor(5.4348, device='cuda:0', grad_fn=)\n", - "20000\n", - "2100 tensor(5.5513, device='cuda:0', grad_fn=)\n", - "21000\n", - "2200 tensor(5.3939, device='cuda:0', grad_fn=)\n", - "22000\n", - "2300 tensor(5.4063, device='cuda:0', grad_fn=)\n", - "23000\n", - "2400 tensor(5.4092, device='cuda:0', grad_fn=)\n", - "24000\n", - "2500 tensor(5.4460, device='cuda:0', grad_fn=)\n", - "25000\n", - "2600 tensor(5.4738, device='cuda:0', grad_fn=)\n", - "26000\n", - "2700 tensor(5.4848, device='cuda:0', grad_fn=)\n", - "27000\n", - "2800 tensor(5.5244, device='cuda:0', grad_fn=)\n", - "28000\n", - "2900 tensor(5.6711, device='cuda:0', grad_fn=)\n", - "29000\n", - "3000 tensor(5.4024, device='cuda:0', grad_fn=)\n", - "30000\n", - "3100 tensor(5.4842, device='cuda:0', grad_fn=)\n", - "31000\n", - "3200 tensor(5.4863, device='cuda:0', grad_fn=)\n", - "32000\n", - "3300 tensor(5.4114, device='cuda:0', grad_fn=)\n", - "33000\n", - "3400 tensor(5.3231, device='cuda:0', grad_fn=)\n", - "34000\n", - "3500 tensor(5.4598, device='cuda:0', grad_fn=)\n", - "35000\n", - "3600 tensor(5.4579, device='cuda:0', grad_fn=)\n", - "36000\n", - "3700 tensor(5.3890, device='cuda:0', grad_fn=)\n", - "37000\n", - "3800 tensor(5.4162, device='cuda:0', grad_fn=)\n", - "38000\n", - "3900 tensor(5.2854, device='cuda:0', grad_fn=)\n", - "39000\n", - "4000 tensor(5.3370, device='cuda:0', grad_fn=)\n", - "40000\n", - "4100 tensor(5.5078, device='cuda:0', grad_fn=)\n", - "41000\n", - "4200 tensor(5.5341, device='cuda:0', grad_fn=)\n", - "42000\n", - "4300 tensor(5.4704, device='cuda:0', grad_fn=)\n", - "43000\n", - "4400 tensor(5.4990, device='cuda:0', grad_fn=)\n", - "44000\n", - "4500 tensor(5.3300, device='cuda:0', grad_fn=)\n", - "45000\n", - "4600 tensor(5.6674, device='cuda:0', grad_fn=)\n", - "46000\n", - "4700 tensor(5.3622, device='cuda:0', grad_fn=)\n", - "47000\n", - "4800 tensor(5.4762, device='cuda:0', grad_fn=)\n", - "48000\n", - "4900 tensor(5.5403, device='cuda:0', grad_fn=)\n", - "49000\n", - "5000 tensor(5.5359, device='cuda:0', grad_fn=)\n", - "50000\n", - "5100 tensor(5.6058, device='cuda:0', grad_fn=)\n", - "51000\n", - "5200 tensor(5.6209, device='cuda:0', grad_fn=)\n", - "52000\n", - "5300 tensor(5.6273, device='cuda:0', grad_fn=)\n", - "53000\n", - "5400 tensor(5.4695, device='cuda:0', grad_fn=)\n", - "54000\n", - "5500 tensor(5.5771, device='cuda:0', grad_fn=)\n", - "55000\n", - "5600 tensor(5.3552, device='cuda:0', grad_fn=)\n", - "56000\n", - "5700 tensor(5.5957, device='cuda:0', grad_fn=)\n", - "57000\n", - "5800 tensor(5.5952, device='cuda:0', grad_fn=)\n", - "58000\n", - "5900 tensor(5.5643, device='cuda:0', grad_fn=)\n", - "59000\n", - "6000 tensor(5.4346, device='cuda:0', grad_fn=)\n", - "60000\n", - "6100 tensor(5.4620, device='cuda:0', grad_fn=)\n", - "61000\n", - "6200 tensor(5.6256, device='cuda:0', grad_fn=)\n", - "62000\n", - "6300 tensor(5.4832, device='cuda:0', grad_fn=)\n", - "63000\n", - "6400 tensor(5.4063, device='cuda:0', grad_fn=)\n", - "64000\n", - "6500 tensor(5.2587, device='cuda:0', grad_fn=)\n", - "65000\n", - "6600 tensor(5.4320, device='cuda:0', grad_fn=)\n", - "66000\n", - "6700 tensor(5.3770, device='cuda:0', grad_fn=)\n", - "67000\n", - "6800 tensor(5.6077, device='cuda:0', grad_fn=)\n", - "68000\n", - "6900 tensor(5.5788, device='cuda:0', grad_fn=)\n", - "69000\n", - "7000 tensor(5.4929, device='cuda:0', grad_fn=)\n", - "70000\n", - "7100 tensor(5.2828, device='cuda:0', grad_fn=)\n", - "71000\n", - "7200 tensor(5.3992, device='cuda:0', grad_fn=)\n", - "72000\n", - "7300 tensor(5.6273, device='cuda:0', grad_fn=)\n", - "73000\n", - "7400 tensor(5.4385, device='cuda:0', grad_fn=)\n", - "74000\n", - "7500 tensor(5.3176, device='cuda:0', grad_fn=)\n", - "75000\n", - "7600 tensor(5.3834, device='cuda:0', grad_fn=)\n", - "76000\n", - "7700 tensor(5.4532, device='cuda:0', grad_fn=)\n", - "77000\n", - "7800 tensor(5.5669, device='cuda:0', grad_fn=)\n", - "78000\n", - "7900 tensor(5.2508, device='cuda:0', grad_fn=)\n", - "79000\n", - "8000 tensor(5.3027, device='cuda:0', grad_fn=)\n", - "80000\n", - "8100 tensor(5.4813, device='cuda:0', grad_fn=)\n", - "81000\n", - "8200 tensor(5.4822, device='cuda:0', grad_fn=)\n", - "82000\n", - "8300 tensor(5.4510, device='cuda:0', grad_fn=)\n", - "83000\n", - "8400 tensor(5.5712, device='cuda:0', grad_fn=)\n", - "84000\n", - "8500 tensor(5.5634, device='cuda:0', grad_fn=)\n", - "85000\n", - "8600 tensor(5.5616, device='cuda:0', grad_fn=)\n", - "86000\n", - "8700 tensor(5.6568, device='cuda:0', grad_fn=)\n", - "87000\n", - "8800 tensor(5.5397, device='cuda:0', grad_fn=)\n", - "88000\n", - "8900 tensor(5.3852, device='cuda:0', grad_fn=)\n", - "89000\n", - "9000 tensor(5.5022, device='cuda:0', grad_fn=)\n", - "90000\n", - "9100 tensor(5.5088, device='cuda:0', grad_fn=)\n", - "91000\n", - "9200 tensor(5.4214, device='cuda:0', grad_fn=)\n", - "92000\n", - "9300 tensor(5.4641, device='cuda:0', grad_fn=)\n", - "93000\n", - "9400 tensor(5.3085, device='cuda:0', grad_fn=)\n", - "94000\n", - "9500 tensor(5.3852, device='cuda:0', grad_fn=)\n", - "95000\n", - "9600 tensor(5.5097, device='cuda:0', grad_fn=)\n", - "96000\n", - "9700 tensor(5.4373, device='cuda:0', grad_fn=)\n", - "97000\n", - "9800 tensor(5.4786, device='cuda:0', grad_fn=)\n", - "9900 tensor(5.3198, device='cuda:0', grad_fn=)\n", - "98000\n", - "10000 tensor(5.5310, device='cuda:0', grad_fn=)\n", - "99000\n", - "10100 tensor(5.4341, device='cuda:0', grad_fn=)\n", - "100000\n", - "10200 tensor(5.3571, device='cuda:0', grad_fn=)\n", - "101000\n", - "10300 tensor(5.4712, device='cuda:0', grad_fn=)\n", - "102000\n", - "10400 tensor(5.4810, device='cuda:0', grad_fn=)\n", - "103000\n", - "10500 tensor(5.5463, device='cuda:0', grad_fn=)\n", - "104000\n", - "10600 tensor(5.6233, device='cuda:0', grad_fn=)\n", - "105000\n", - "10700 tensor(5.4678, device='cuda:0', grad_fn=)\n", - "106000\n", - "10800 tensor(5.5040, device='cuda:0', grad_fn=)\n", - "107000\n", - "10900 tensor(5.3963, device='cuda:0', grad_fn=)\n", - "108000\n", - "11000 tensor(5.6295, device='cuda:0', grad_fn=)\n", - "109000\n", - "11100 tensor(5.2378, device='cuda:0', grad_fn=)\n", - "110000\n", - "11200 tensor(5.4184, device='cuda:0', grad_fn=)\n", - "111000\n", - "11300 tensor(5.4404, device='cuda:0', grad_fn=)\n", - "112000\n", - "11400 tensor(5.3875, device='cuda:0', grad_fn=)\n", - "113000\n", - "11500 tensor(5.4523, device='cuda:0', grad_fn=)\n", - "114000\n", - "11600 tensor(5.4418, device='cuda:0', grad_fn=)\n", - "115000\n", - "11700 tensor(5.3604, device='cuda:0', grad_fn=)\n", - "116000\n", - "11800 tensor(5.5647, device='cuda:0', grad_fn=)\n", - "117000\n", - "11900 tensor(5.3936, device='cuda:0', grad_fn=)\n", - "118000\n", - "12000 tensor(5.4823, device='cuda:0', grad_fn=)\n", - "119000\n", - "12100 tensor(5.5069, device='cuda:0', grad_fn=)\n", - "120000\n", - "12200 tensor(5.4983, device='cuda:0', grad_fn=)\n", - "121000\n", - "12300 tensor(5.6030, device='cuda:0', grad_fn=)\n", - "122000\n", - "12400 tensor(5.4763, device='cuda:0', grad_fn=)\n", - "123000\n", - "12500 tensor(5.3718, device='cuda:0', grad_fn=)\n", - "124000\n", - "12600 tensor(5.4416, device='cuda:0', grad_fn=)\n", - "125000\n", - "12700 tensor(5.3554, device='cuda:0', grad_fn=)\n", - "126000\n", - "12800 tensor(5.5392, device='cuda:0', grad_fn=)\n", - "127000\n", - "12900 tensor(5.4164, device='cuda:0', grad_fn=)\n", - "128000\n", - "13000 tensor(5.2286, device='cuda:0', grad_fn=)\n", - "129000\n", - "13100 tensor(5.4288, device='cuda:0', grad_fn=)\n", - "130000\n", - "13200 tensor(5.4770, device='cuda:0', grad_fn=)\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "131000\n", - "13300 tensor(5.3352, device='cuda:0', grad_fn=)\n", - "132000\n", - "13400 tensor(5.4349, device='cuda:0', grad_fn=)\n", - "133000\n", - "13500 tensor(5.3860, device='cuda:0', grad_fn=)\n", - "134000\n", - "13600 tensor(5.4648, device='cuda:0', grad_fn=)\n", - "135000\n", - "13700 tensor(5.4444, device='cuda:0', grad_fn=)\n", - "136000\n", - "13800 tensor(5.4320, device='cuda:0', grad_fn=)\n", - "137000\n", - "13900 tensor(5.2935, device='cuda:0', grad_fn=)\n", - "138000\n", - "14000 tensor(5.5387, device='cuda:0', grad_fn=)\n", - "139000\n", - "14100 tensor(5.2424, device='cuda:0', grad_fn=)\n", - "140000\n", - "14200 tensor(5.5177, device='cuda:0', grad_fn=)\n", - "141000\n", - "14300 tensor(5.4831, device='cuda:0', grad_fn=)\n", - "142000\n", - "14400 tensor(5.3877, device='cuda:0', grad_fn=)\n", - "143000\n", - "14500 tensor(5.4919, device='cuda:0', grad_fn=)\n", - "144000\n", - "14600 tensor(5.5253, device='cuda:0', grad_fn=)\n", - "145000\n", - "14700 tensor(5.3948, device='cuda:0', grad_fn=)\n", - "146000\n", - "14800 tensor(5.8442, device='cuda:0', grad_fn=)\n", - "147000\n", - "14900 tensor(5.4967, device='cuda:0', grad_fn=)\n", - "148000\n", - "15000 tensor(5.3788, device='cuda:0', grad_fn=)\n", - "149000\n", - "15100 tensor(5.5832, device='cuda:0', grad_fn=)\n", - "150000\n", - "15200 tensor(5.4482, device='cuda:0', grad_fn=)\n", - "151000\n", - "15300 tensor(5.4260, device='cuda:0', grad_fn=)\n", - "152000\n", - "15400 tensor(5.3273, device='cuda:0', grad_fn=)\n", - "153000\n", - "15500 tensor(5.4840, device='cuda:0', grad_fn=)\n", - "154000\n", - "15600 tensor(5.4851, device='cuda:0', grad_fn=)\n", - "155000\n", - "15700 tensor(5.3871, device='cuda:0', grad_fn=)\n", - "156000\n", - "15800 tensor(5.2933, device='cuda:0', grad_fn=)\n", - "157000\n", - "15900 tensor(5.4374, device='cuda:0', grad_fn=)\n", - "158000\n", - "16000 tensor(5.2555, device='cuda:0', grad_fn=)\n", - "159000\n", - "16100 tensor(5.3127, device='cuda:0', grad_fn=)\n", - "160000\n", - "16200 tensor(5.6423, device='cuda:0', grad_fn=)\n", - "161000\n", - "16300 tensor(5.4702, device='cuda:0', grad_fn=)\n", - "162000\n", - "16400 tensor(5.4419, device='cuda:0', grad_fn=)\n", - "163000\n", - "16500 tensor(5.5640, device='cuda:0', grad_fn=)\n", - "164000\n", - "16600 tensor(5.4099, device='cuda:0', grad_fn=)\n", - "165000\n", - "16700 tensor(5.3822, device='cuda:0', grad_fn=)\n", - "166000\n", - "16800 tensor(5.4643, device='cuda:0', grad_fn=)\n", - "167000\n", - "16900 tensor(5.2234, device='cuda:0', grad_fn=)\n", - "168000\n", - "17000 tensor(5.5021, device='cuda:0', grad_fn=)\n", - "169000\n", - "17100 tensor(5.3524, device='cuda:0', grad_fn=)\n", - "170000\n", - "17200 tensor(5.4725, device='cuda:0', grad_fn=)\n", - "171000\n", - "17300 tensor(5.5034, device='cuda:0', grad_fn=)\n", - "172000\n", - "17400 tensor(5.2911, device='cuda:0', grad_fn=)\n", - "173000\n", - "17500 tensor(5.3147, device='cuda:0', grad_fn=)\n", - "174000\n", - "17600 tensor(5.3426, device='cuda:0', grad_fn=)\n", - "175000\n", - "17700 tensor(5.3414, device='cuda:0', grad_fn=)\n", - "176000\n", - "17800 tensor(5.3991, device='cuda:0', grad_fn=)\n", - "177000\n", - "17900 tensor(5.2936, device='cuda:0', grad_fn=)\n", - "178000\n", - "18000 tensor(5.5238, device='cuda:0', grad_fn=)\n", - "179000\n", - "18100 tensor(5.4684, device='cuda:0', grad_fn=)\n", - "180000\n", - "18200 tensor(5.3916, device='cuda:0', grad_fn=)\n", - "181000\n", - "18300 tensor(5.3888, device='cuda:0', grad_fn=)\n", - "182000\n", - "18400 tensor(5.4299, device='cuda:0', grad_fn=)\n", - "183000\n", - "18500 tensor(5.4103, device='cuda:0', grad_fn=)\n", - "184000\n", - "18600 tensor(5.5980, device='cuda:0', grad_fn=)\n", - "185000\n", - "18700 tensor(5.4135, device='cuda:0', grad_fn=)\n", - "186000\n", - "18800 tensor(5.5855, device='cuda:0', grad_fn=)\n", - "187000\n", - "18900 tensor(5.4583, device='cuda:0', grad_fn=)\n", - "188000\n", - "19000 tensor(5.4854, device='cuda:0', grad_fn=)\n", - "189000\n", - "19100 tensor(5.5879, device='cuda:0', grad_fn=)\n", - "190000\n", - "19200 tensor(5.4675, device='cuda:0', grad_fn=)\n", - "191000\n", - "19300 tensor(5.5741, device='cuda:0', grad_fn=)\n", - "192000\n", - "19400 tensor(5.3977, device='cuda:0', grad_fn=)\n", - "193000\n", - "19500 tensor(5.4042, device='cuda:0', grad_fn=)\n", - "194000\n", - "19600 tensor(5.4364, device='cuda:0', grad_fn=)\n", - "195000\n", - "19700 tensor(5.4868, device='cuda:0', grad_fn=)\n", - "196000\n", - "19800 tensor(5.3476, device='cuda:0', grad_fn=)\n", - "197000\n", - "19900 tensor(5.3553, device='cuda:0', grad_fn=)\n", - "198000\n", - "20000 tensor(5.3707, device='cuda:0', grad_fn=)\n", - "199000\n", - "20100 tensor(5.3226, device='cuda:0', grad_fn=)\n", - "200000\n", - "20200 tensor(5.2488, device='cuda:0', grad_fn=)\n", - "201000\n", - "20300 tensor(5.3648, device='cuda:0', grad_fn=)\n", - "20400 tensor(5.4156, device='cuda:0', grad_fn=)\n", - "202000\n", - "20500 tensor(5.4102, device='cuda:0', grad_fn=)\n", - "203000\n", - "20600 tensor(5.6109, device='cuda:0', grad_fn=)\n", - "204000\n", - "20700 tensor(5.4335, device='cuda:0', grad_fn=)\n", - "205000\n", - "20800 tensor(5.2795, device='cuda:0', grad_fn=)\n", - "206000\n", - "20900 tensor(5.5609, device='cuda:0', grad_fn=)\n", - "207000\n", - "21000 tensor(5.3918, device='cuda:0', grad_fn=)\n", - "208000\n", - "21100 tensor(5.3831, device='cuda:0', grad_fn=)\n", - "209000\n", - "21200 tensor(5.2790, device='cuda:0', grad_fn=)\n", - "210000\n", - "21300 tensor(5.4710, device='cuda:0', grad_fn=)\n", - "211000\n", - "21400 tensor(5.5050, device='cuda:0', grad_fn=)\n", - "212000\n", - "21500 tensor(5.2692, device='cuda:0', grad_fn=)\n", - "213000\n", - "21600 tensor(5.2668, device='cuda:0', grad_fn=)\n", - "214000\n", - "21700 tensor(5.2633, device='cuda:0', grad_fn=)\n", - "215000\n", - "21800 tensor(5.4067, device='cuda:0', grad_fn=)\n", - "216000\n", - "21900 tensor(5.3829, device='cuda:0', grad_fn=)\n", - "217000\n", - "22000 tensor(5.3773, device='cuda:0', grad_fn=)\n", - "218000\n", - "22100 tensor(5.2472, device='cuda:0', grad_fn=)\n", - "219000\n", - "22200 tensor(5.1171, device='cuda:0', grad_fn=)\n", - "220000\n", - "22300 tensor(5.5545, device='cuda:0', grad_fn=)\n", - "221000\n", - "22400 tensor(5.2499, device='cuda:0', grad_fn=)\n", - "222000\n", - "22500 tensor(5.2943, device='cuda:0', grad_fn=)\n", - "223000\n", - "22600 tensor(5.4748, device='cuda:0', grad_fn=)\n", - "224000\n", - "22700 tensor(5.2436, device='cuda:0', grad_fn=)\n", - "225000\n", - "22800 tensor(5.5053, device='cuda:0', grad_fn=)\n", - "226000\n", - "22900 tensor(5.5519, device='cuda:0', grad_fn=)\n", - "227000\n", - "23000 tensor(5.4541, device='cuda:0', grad_fn=)\n", - "228000\n", - "23100 tensor(5.4279, device='cuda:0', grad_fn=)\n", - "229000\n", - "23200 tensor(5.4286, device='cuda:0', grad_fn=)\n", - "230000\n", - "23300 tensor(5.5179, device='cuda:0', grad_fn=)\n", - "231000\n", - "23400 tensor(5.5355, device='cuda:0', grad_fn=)\n", - "232000\n", - "23500 tensor(5.3505, device='cuda:0', grad_fn=)\n", - "233000\n", - "23600 tensor(5.3313, device='cuda:0', grad_fn=)\n", - "234000\n", - "23700 tensor(5.3509, device='cuda:0', grad_fn=)\n", - "235000\n", - "23800 tensor(5.2170, device='cuda:0', grad_fn=)\n", - "236000\n", - "23900 tensor(5.3101, device='cuda:0', grad_fn=)\n", - "237000\n", - "24000 tensor(5.2962, device='cuda:0', grad_fn=)\n", - "238000\n", - "24100 tensor(5.3882, device='cuda:0', grad_fn=)\n", - "239000\n", - "24200 tensor(5.5633, device='cuda:0', grad_fn=)\n", - "240000\n", - "24300 tensor(5.4595, device='cuda:0', grad_fn=)\n", - "241000\n", - "24400 tensor(5.5932, device='cuda:0', grad_fn=)\n", - "242000\n", - "24500 tensor(5.3717, device='cuda:0', grad_fn=)\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "243000\n", - "24600 tensor(5.4943, device='cuda:0', grad_fn=)\n", - "244000\n", - "24700 tensor(5.3985, device='cuda:0', grad_fn=)\n", - "245000\n", - "24800 tensor(5.4347, device='cuda:0', grad_fn=)\n", - "246000\n", - "24900 tensor(5.5008, device='cuda:0', grad_fn=)\n", - "247000\n", - "25000 tensor(5.5100, device='cuda:0', grad_fn=)\n", - "248000\n", - "25100 tensor(5.4427, device='cuda:0', grad_fn=)\n", - "249000\n", - "25200 tensor(5.4508, device='cuda:0', grad_fn=)\n", - "250000\n", - "25300 tensor(5.4724, device='cuda:0', grad_fn=)\n", - "251000\n", - "25400 tensor(5.4525, device='cuda:0', grad_fn=)\n", - "252000\n", - "25500 tensor(5.3620, device='cuda:0', grad_fn=)\n", - "253000\n", - "25600 tensor(5.6446, device='cuda:0', grad_fn=)\n", - "254000\n", - "25700 tensor(5.3966, device='cuda:0', grad_fn=)\n", - "255000\n", - "25800 tensor(5.4889, device='cuda:0', grad_fn=)\n", - "256000\n", - "25900 tensor(5.4251, device='cuda:0', grad_fn=)\n", - "257000\n", - "26000 tensor(5.4346, device='cuda:0', grad_fn=)\n", - "258000\n", - "26100 tensor(5.3395, device='cuda:0', grad_fn=)\n", - "259000\n", - "26200 tensor(5.2695, device='cuda:0', grad_fn=)\n", - "260000\n", - "26300 tensor(5.4767, device='cuda:0', grad_fn=)\n", - "261000\n", - "26400 tensor(5.5083, device='cuda:0', grad_fn=)\n", - "262000\n", - "26500 tensor(5.2347, device='cuda:0', grad_fn=)\n", - "263000\n", - "26600 tensor(5.5761, device='cuda:0', grad_fn=)\n", - "264000\n", - "26700 tensor(5.4402, device='cuda:0', grad_fn=)\n", - "265000\n", - "26800 tensor(5.6173, device='cuda:0', grad_fn=)\n", - "266000\n", - "26900 tensor(5.3775, device='cuda:0', grad_fn=)\n", - "267000\n", - "27000 tensor(5.2863, device='cuda:0', grad_fn=)\n", - "268000\n", - "27100 tensor(5.3007, device='cuda:0', grad_fn=)\n", - "269000\n", - "27200 tensor(5.3551, device='cuda:0', grad_fn=)\n", - "270000\n", - "27300 tensor(5.5439, device='cuda:0', grad_fn=)\n", - "271000\n", - "27400 tensor(5.4334, device='cuda:0', grad_fn=)\n", - "272000\n", - "27500 tensor(5.3266, device='cuda:0', grad_fn=)\n", - "273000\n", - "27600 tensor(5.6412, device='cuda:0', grad_fn=)\n", - "274000\n", - "27700 tensor(5.4420, device='cuda:0', grad_fn=)\n", - "275000\n", - "27800 tensor(5.4381, device='cuda:0', grad_fn=)\n", - "276000\n", - "27900 tensor(5.5550, device='cuda:0', grad_fn=)\n", - "277000\n", - "28000 tensor(5.4154, device='cuda:0', grad_fn=)\n", - "278000\n", - "28100 tensor(5.3823, device='cuda:0', grad_fn=)\n", - "279000\n", - "28200 tensor(5.5344, device='cuda:0', grad_fn=)\n", - "280000\n", - "28300 tensor(5.1615, device='cuda:0', grad_fn=)\n", - "281000\n", - "28400 tensor(5.6069, device='cuda:0', grad_fn=)\n", - "282000\n", - "28500 tensor(5.4426, device='cuda:0', grad_fn=)\n", - "283000\n", - "28600 tensor(5.3672, device='cuda:0', grad_fn=)\n", - "284000\n", - "28700 tensor(5.5133, device='cuda:0', grad_fn=)\n", - "285000\n", - "28800 tensor(5.5556, device='cuda:0', grad_fn=)\n", - "286000\n", - "28900 tensor(5.4294, device='cuda:0', grad_fn=)\n", - "287000\n", - "29000 tensor(5.3359, device='cuda:0', grad_fn=)\n", - "288000\n", - "29100 tensor(5.0951, device='cuda:0', grad_fn=)\n", - "289000\n", - "29200 tensor(5.2511, device='cuda:0', grad_fn=)\n", - "290000\n", - "29300 tensor(5.5364, device='cuda:0', grad_fn=)\n", - "291000\n", - "29400 tensor(5.6708, device='cuda:0', grad_fn=)\n", - "292000\n", - "29500 tensor(5.4371, device='cuda:0', grad_fn=)\n", - "293000\n", - "29600 tensor(5.2942, device='cuda:0', grad_fn=)\n", - "294000\n", - "29700 tensor(5.4637, device='cuda:0', grad_fn=)\n", - "295000\n", - "29800 tensor(5.2914, device='cuda:0', grad_fn=)\n", - "296000\n", - "29900 tensor(5.5562, device='cuda:0', grad_fn=)\n", - "297000\n", - "30000 tensor(5.2833, device='cuda:0', grad_fn=)\n", - "298000\n", - "30100 tensor(5.3481, device='cuda:0', grad_fn=)\n", - "299000\n", - "30200 tensor(5.3122, device='cuda:0', grad_fn=)\n", - "300000\n", - "30300 tensor(5.4103, device='cuda:0', grad_fn=)\n", - "301000\n", - "30400 tensor(5.2480, device='cuda:0', grad_fn=)\n", - "302000\n", - "30500 tensor(5.4258, device='cuda:0', grad_fn=)\n", - "30600 tensor(5.3835, device='cuda:0', grad_fn=)\n", - "303000\n", - "30700 tensor(5.4193, device='cuda:0', grad_fn=)\n", - "304000\n", - "30800 tensor(5.4438, device='cuda:0', grad_fn=)\n", - "305000\n", - "30900 tensor(5.4518, device='cuda:0', grad_fn=)\n", - "306000\n", - "31000 tensor(5.5607, device='cuda:0', grad_fn=)\n", - "307000\n", - "31100 tensor(5.2059, device='cuda:0', grad_fn=)\n", - "308000\n", - "31200 tensor(5.2571, device='cuda:0', grad_fn=)\n", - "309000\n", - "31300 tensor(5.5208, device='cuda:0', grad_fn=)\n", - "310000\n", - "31400 tensor(5.3061, device='cuda:0', grad_fn=)\n", - "311000\n", - "31500 tensor(5.4834, device='cuda:0', grad_fn=)\n", - "312000\n", - "31600 tensor(5.4653, device='cuda:0', grad_fn=)\n", - "313000\n", - "31700 tensor(5.4308, device='cuda:0', grad_fn=)\n", - "314000\n", - "31800 tensor(5.5400, device='cuda:0', grad_fn=)\n", - "315000\n", - "31900 tensor(5.1536, device='cuda:0', grad_fn=)\n", - "316000\n", - "32000 tensor(5.3460, device='cuda:0', grad_fn=)\n", - "317000\n", - "32100 tensor(5.2300, device='cuda:0', grad_fn=)\n", - "318000\n", - "32200 tensor(5.5511, device='cuda:0', grad_fn=)\n", - "319000\n", - "32300 tensor(5.5391, device='cuda:0', grad_fn=)\n", - "320000\n", - "32400 tensor(5.5157, device='cuda:0', grad_fn=)\n", - "321000\n", - "32500 tensor(5.3336, device='cuda:0', grad_fn=)\n", - "322000\n", - "32600 tensor(5.4475, device='cuda:0', grad_fn=)\n", - "323000\n", - "32700 tensor(5.3894, device='cuda:0', grad_fn=)\n", - "324000\n", - "32800 tensor(5.6022, device='cuda:0', grad_fn=)\n", - "325000\n", - "32900 tensor(5.4663, device='cuda:0', grad_fn=)\n", - "326000\n", - "33000 tensor(5.2387, device='cuda:0', grad_fn=)\n", - "327000\n", - "33100 tensor(5.4446, device='cuda:0', grad_fn=)\n", - "328000\n", - "33200 tensor(5.5450, device='cuda:0', grad_fn=)\n", - "329000\n", - "33300 tensor(5.3179, device='cuda:0', grad_fn=)\n", - "330000\n", - "33400 tensor(5.5905, device='cuda:0', grad_fn=)\n", - "331000\n", - "33500 tensor(5.4066, device='cuda:0', grad_fn=)\n", - "332000\n", - "33600 tensor(5.3542, device='cuda:0', grad_fn=)\n", - "333000\n", - "33700 tensor(5.4097, device='cuda:0', grad_fn=)\n", - "334000\n", - "33800 tensor(5.4912, device='cuda:0', grad_fn=)\n", - "335000\n", - "33900 tensor(5.2358, device='cuda:0', grad_fn=)\n", - "336000\n", - "34000 tensor(5.4470, device='cuda:0', grad_fn=)\n", - "337000\n", - "34100 tensor(5.4207, device='cuda:0', grad_fn=)\n", - "338000\n", - "34200 tensor(5.4651, device='cuda:0', grad_fn=)\n", - "339000\n", - "34300 tensor(5.2545, device='cuda:0', grad_fn=)\n", - "340000\n", - "34400 tensor(5.7106, device='cuda:0', grad_fn=)\n", - "341000\n", - "34500 tensor(5.5699, device='cuda:0', grad_fn=)\n", - "342000\n", - "34600 tensor(5.4638, device='cuda:0', grad_fn=)\n", - "343000\n", - "34700 tensor(5.5382, device='cuda:0', grad_fn=)\n", - "344000\n", - "34800 tensor(5.5603, device='cuda:0', grad_fn=)\n", - "345000\n", - "34900 tensor(5.6072, device='cuda:0', grad_fn=)\n", - "346000\n", - "35000 tensor(5.6037, device='cuda:0', grad_fn=)\n", - "347000\n", - "35100 tensor(5.4069, device='cuda:0', grad_fn=)\n", - "348000\n", - "35200 tensor(5.3398, device='cuda:0', grad_fn=)\n", - "349000\n", - "35300 tensor(5.5607, device='cuda:0', grad_fn=)\n", - "350000\n", - "35400 tensor(5.2068, device='cuda:0', grad_fn=)\n", - "351000\n", - "35500 tensor(5.3112, device='cuda:0', grad_fn=)\n", - "352000\n", - "35600 tensor(5.4126, device='cuda:0', grad_fn=)\n", - "353000\n", - "35700 tensor(5.3091, device='cuda:0', grad_fn=)\n", - "354000\n", - "35800 tensor(5.4252, device='cuda:0', grad_fn=)\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "355000\n", - "35900 tensor(5.3956, device='cuda:0', grad_fn=)\n", - "356000\n", - "36000 tensor(5.1705, device='cuda:0', grad_fn=)\n", - "357000\n", - "36100 tensor(5.5497, device='cuda:0', grad_fn=)\n", - "358000\n", - "36200 tensor(5.4066, device='cuda:0', grad_fn=)\n", - "359000\n", - "36300 tensor(5.6858, device='cuda:0', grad_fn=)\n", - "360000\n", - "36400 tensor(5.3812, device='cuda:0', grad_fn=)\n", - "361000\n", - "36500 tensor(5.3990, device='cuda:0', grad_fn=)\n", - "362000\n", - "36600 tensor(5.4302, device='cuda:0', grad_fn=)\n", - "363000\n", - "36700 tensor(5.2253, device='cuda:0', grad_fn=)\n", - "364000\n", - "36800 tensor(5.3347, device='cuda:0', grad_fn=)\n", - "365000\n", - "36900 tensor(5.4426, device='cuda:0', grad_fn=)\n", - "366000\n", - "37000 tensor(5.3419, device='cuda:0', grad_fn=)\n", - "367000\n", - "37100 tensor(5.3579, device='cuda:0', grad_fn=)\n", - "368000\n", - "37200 tensor(5.4332, device='cuda:0', grad_fn=)\n", - "369000\n", - "37300 tensor(5.3362, device='cuda:0', grad_fn=)\n", - "370000\n", - "37400 tensor(5.7100, device='cuda:0', grad_fn=)\n", - "371000\n", - "37500 tensor(5.3742, device='cuda:0', grad_fn=)\n", - "372000\n", - "37600 tensor(5.3615, device='cuda:0', grad_fn=)\n", - "373000\n", - "37700 tensor(5.5402, device='cuda:0', grad_fn=)\n", - "374000\n", - "37800 tensor(5.3734, device='cuda:0', grad_fn=)\n", - "375000\n", - "37900 tensor(5.3621, device='cuda:0', grad_fn=)\n", - "376000\n", - "38000 tensor(5.4380, device='cuda:0', grad_fn=)\n", - "377000\n", - "38100 tensor(5.4513, device='cuda:0', grad_fn=)\n", - "378000\n", - "38200 tensor(5.4554, device='cuda:0', grad_fn=)\n", - "379000\n", - "38300 tensor(5.3735, device='cuda:0', grad_fn=)\n", - "380000\n", - "38400 tensor(5.4297, device='cuda:0', grad_fn=)\n", - "381000\n", - "38500 tensor(5.4561, device='cuda:0', grad_fn=)\n", - "382000\n", - "38600 tensor(5.4118, device='cuda:0', grad_fn=)\n", - "383000\n", - "38700 tensor(5.3996, device='cuda:0', grad_fn=)\n", - "384000\n", - "38800 tensor(5.4825, device='cuda:0', grad_fn=)\n", - "385000\n", - "38900 tensor(5.5692, device='cuda:0', grad_fn=)\n", - "386000\n", - "39000 tensor(5.2573, device='cuda:0', grad_fn=)\n", - "387000\n", - "39100 tensor(5.4847, device='cuda:0', grad_fn=)\n", - "388000\n", - "39200 tensor(5.5802, device='cuda:0', grad_fn=)\n", - "389000\n", - "39300 tensor(5.3968, device='cuda:0', grad_fn=)\n", - "390000\n", - "39400 tensor(5.4666, device='cuda:0', grad_fn=)\n", - "391000\n", - "39500 tensor(5.5847, device='cuda:0', grad_fn=)\n", - "392000\n", - "39600 tensor(5.2648, device='cuda:0', grad_fn=)\n", - "393000\n", - "39700 tensor(5.2423, device='cuda:0', grad_fn=)\n", - "394000\n", - "39800 tensor(5.3731, device='cuda:0', grad_fn=)\n", - "395000\n", - "39900 tensor(5.2014, device='cuda:0', grad_fn=)\n", - "396000\n", - "40000 tensor(5.2903, device='cuda:0', grad_fn=)\n", - "397000\n", - "40100 tensor(5.3712, device='cuda:0', grad_fn=)\n", - "398000\n", - "40200 tensor(5.3557, device='cuda:0', grad_fn=)\n", - "399000\n", - "40300 tensor(5.4151, device='cuda:0', grad_fn=)\n", - "40400 tensor(5.4358, device='cuda:0', grad_fn=)\n", - "400000\n", - "40500 tensor(5.3498, device='cuda:0', grad_fn=)\n", - "401000\n", - "40600 tensor(5.4152, device='cuda:0', grad_fn=)\n", - "402000\n", - "40700 tensor(5.4551, device='cuda:0', grad_fn=)\n", - "403000\n", - "40800 tensor(5.4138, device='cuda:0', grad_fn=)\n", - "404000\n", - "40900 tensor(5.3628, device='cuda:0', grad_fn=)\n", - "405000\n", - "41000 tensor(5.4124, device='cuda:0', grad_fn=)\n", - "406000\n", - "41100 tensor(5.3750, device='cuda:0', grad_fn=)\n", - "407000\n", - "41200 tensor(5.2687, device='cuda:0', grad_fn=)\n", - "408000\n", - "41300 tensor(5.3987, device='cuda:0', grad_fn=)\n", - "409000\n", - "41400 tensor(5.2976, device='cuda:0', grad_fn=)\n", - "410000\n", - "41500 tensor(5.4418, device='cuda:0', grad_fn=)\n", - "411000\n", - "41600 tensor(5.3558, device='cuda:0', grad_fn=)\n", - "412000\n", - "41700 tensor(5.3767, device='cuda:0', grad_fn=)\n", - "413000\n", - "41800 tensor(5.3836, device='cuda:0', grad_fn=)\n", - "414000\n", - "41900 tensor(5.3904, device='cuda:0', grad_fn=)\n", - "415000\n", - "42000 tensor(5.5445, device='cuda:0', grad_fn=)\n", - "416000\n", - "42100 tensor(5.2890, device='cuda:0', grad_fn=)\n", - "417000\n", - "42200 tensor(5.3691, device='cuda:0', grad_fn=)\n", - "418000\n", - "42300 tensor(5.4364, device='cuda:0', grad_fn=)\n", - "419000\n", - "42400 tensor(5.2507, device='cuda:0', grad_fn=)\n", - "420000\n", - "42500 tensor(5.4215, device='cuda:0', grad_fn=)\n", - "421000\n", - "42600 tensor(5.2136, device='cuda:0', grad_fn=)\n", - "422000\n", - "42700 tensor(5.5296, device='cuda:0', grad_fn=)\n", - "423000\n", - "42800 tensor(5.4544, device='cuda:0', grad_fn=)\n", - "424000\n", - "42900 tensor(5.3009, device='cuda:0', grad_fn=)\n", - "425000\n", - "43000 tensor(5.4403, device='cuda:0', grad_fn=)\n", - "426000\n", - "43100 tensor(5.4384, device='cuda:0', grad_fn=)\n", - "427000\n", - "43200 tensor(5.2520, device='cuda:0', grad_fn=)\n", - "428000\n", - "43300 tensor(5.2945, device='cuda:0', grad_fn=)\n", - "429000\n", - "43400 tensor(5.4455, device='cuda:0', grad_fn=)\n", - "430000\n", - "43500 tensor(5.1633, device='cuda:0', grad_fn=)\n", - "431000\n", - "43600 tensor(5.3649, device='cuda:0', grad_fn=)\n", - "432000\n", - "epoch: = 3\n", - "0 tensor(5.3427, device='cuda:0', grad_fn=)\n", - "1000\n", - "100 tensor(5.4180, device='cuda:0', grad_fn=)\n", - "200 tensor(5.2939, device='cuda:0', grad_fn=)\n", - "2000\n", - "300 tensor(5.3083, device='cuda:0', grad_fn=)\n", - "3000\n", - "400 tensor(5.3086, device='cuda:0', grad_fn=)\n", - "4000\n", - "500 tensor(5.4733, device='cuda:0', grad_fn=)\n", - "5000\n", - "600 tensor(5.2627, device='cuda:0', grad_fn=)\n", - "6000\n", - "700 tensor(5.5664, device='cuda:0', grad_fn=)\n", - "7000\n", - "800 tensor(5.1641, device='cuda:0', grad_fn=)\n", - "8000\n", - "900 tensor(5.4272, device='cuda:0', grad_fn=)\n", - "9000\n", - "1000 tensor(5.2926, device='cuda:0', grad_fn=)\n", - "10000\n", - "1100 tensor(5.4848, device='cuda:0', grad_fn=)\n", - "11000\n", - "1200 tensor(5.5283, device='cuda:0', grad_fn=)\n", - "12000\n", - "1300 tensor(5.4635, device='cuda:0', grad_fn=)\n", - "13000\n", - "1400 tensor(5.4590, device='cuda:0', grad_fn=)\n", - "14000\n", - "1500 tensor(5.5386, device='cuda:0', grad_fn=)\n", - "15000\n", - "1600 tensor(5.2150, device='cuda:0', grad_fn=)\n", - "16000\n", - "1700 tensor(5.3116, device='cuda:0', grad_fn=)\n", - "17000\n", - "1800 tensor(5.3130, device='cuda:0', grad_fn=)\n", - "18000\n", - "1900 tensor(5.2889, device='cuda:0', grad_fn=)\n", - "19000\n", - "2000 tensor(5.3574, device='cuda:0', grad_fn=)\n", - "20000\n", - "2100 tensor(5.4860, device='cuda:0', grad_fn=)\n", - "21000\n", - "2200 tensor(5.3206, device='cuda:0', grad_fn=)\n", - "22000\n", - "2300 tensor(5.3447, device='cuda:0', grad_fn=)\n", - "23000\n", - "2400 tensor(5.3333, device='cuda:0', grad_fn=)\n", - "24000\n", - "2500 tensor(5.3822, device='cuda:0', grad_fn=)\n", - "25000\n", - "2600 tensor(5.4039, device='cuda:0', grad_fn=)\n", - "26000\n", - "2700 tensor(5.4280, device='cuda:0', grad_fn=)\n", - "27000\n", - "2800 tensor(5.4575, device='cuda:0', grad_fn=)\n", - "28000\n", - "2900 tensor(5.5878, device='cuda:0', grad_fn=)\n", - "29000\n", - "3000 tensor(5.3311, device='cuda:0', grad_fn=)\n", - "30000\n", - "3100 tensor(5.4103, device='cuda:0', grad_fn=)\n", - "31000\n", - "3200 tensor(5.4323, device='cuda:0', grad_fn=)\n", - "32000\n", - "3300 tensor(5.3521, device='cuda:0', grad_fn=)\n", - "33000\n", - "3400 tensor(5.2512, device='cuda:0', grad_fn=)\n", - "34000\n", - "3500 tensor(5.3813, device='cuda:0', grad_fn=)\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "35000\n", - "3600 tensor(5.4000, device='cuda:0', grad_fn=)\n", - "36000\n", - "3700 tensor(5.3312, device='cuda:0', grad_fn=)\n", - "37000\n", - "3800 tensor(5.3553, device='cuda:0', grad_fn=)\n", - "38000\n", - "3900 tensor(5.2275, device='cuda:0', grad_fn=)\n", - "39000\n", - "4000 tensor(5.2883, device='cuda:0', grad_fn=)\n", - "40000\n", - "4100 tensor(5.4294, device='cuda:0', grad_fn=)\n", - "41000\n", - "4200 tensor(5.4801, device='cuda:0', grad_fn=)\n", - "42000\n", - "4300 tensor(5.3863, device='cuda:0', grad_fn=)\n", - "43000\n", - "4400 tensor(5.4470, device='cuda:0', grad_fn=)\n", - "44000\n", - "4500 tensor(5.2610, device='cuda:0', grad_fn=)\n", - "45000\n", - "4600 tensor(5.5962, device='cuda:0', grad_fn=)\n", - "46000\n", - "4700 tensor(5.3029, device='cuda:0', grad_fn=)\n", - "47000\n", - "4800 tensor(5.4265, device='cuda:0', grad_fn=)\n", - "48000\n", - "4900 tensor(5.4823, device='cuda:0', grad_fn=)\n", - "49000\n", - "5000 tensor(5.4749, device='cuda:0', grad_fn=)\n", - "50000\n", - "5100 tensor(5.5356, device='cuda:0', grad_fn=)\n", - "51000\n", - "5200 tensor(5.5513, device='cuda:0', grad_fn=)\n", - "52000\n", - "5300 tensor(5.5476, device='cuda:0', grad_fn=)\n", - "53000\n", - "5400 tensor(5.4039, device='cuda:0', grad_fn=)\n", - "54000\n", - "5500 tensor(5.5156, device='cuda:0', grad_fn=)\n", - "55000\n", - "5600 tensor(5.2975, device='cuda:0', grad_fn=)\n", - "56000\n", - "5700 tensor(5.5492, device='cuda:0', grad_fn=)\n", - "57000\n", - "5800 tensor(5.5379, device='cuda:0', grad_fn=)\n", - "58000\n", - "5900 tensor(5.4874, device='cuda:0', grad_fn=)\n", - "59000\n", - "6000 tensor(5.3808, device='cuda:0', grad_fn=)\n", - "60000\n", - "6100 tensor(5.3932, device='cuda:0', grad_fn=)\n", - "61000\n", - "6200 tensor(5.5657, device='cuda:0', grad_fn=)\n", - "62000\n", - "6300 tensor(5.4233, device='cuda:0', grad_fn=)\n", - "63000\n", - "6400 tensor(5.3438, device='cuda:0', grad_fn=)\n", - "64000\n", - "6500 tensor(5.2002, device='cuda:0', grad_fn=)\n", - "65000\n", - "6600 tensor(5.3774, device='cuda:0', grad_fn=)\n", - "66000\n", - "6700 tensor(5.3193, device='cuda:0', grad_fn=)\n", - "67000\n", - "6800 tensor(5.5394, device='cuda:0', grad_fn=)\n", - "68000\n", - "6900 tensor(5.5196, device='cuda:0', grad_fn=)\n", - "69000\n", - "7000 tensor(5.4282, device='cuda:0', grad_fn=)\n", - "70000\n", - "7100 tensor(5.2296, device='cuda:0', grad_fn=)\n", - "71000\n", - "7200 tensor(5.3175, device='cuda:0', grad_fn=)\n", - "72000\n", - "7300 tensor(5.5642, device='cuda:0', grad_fn=)\n", - "73000\n", - "7400 tensor(5.3784, device='cuda:0', grad_fn=)\n", - "74000\n", - "7500 tensor(5.2475, device='cuda:0', grad_fn=)\n", - "75000\n", - "7600 tensor(5.3194, device='cuda:0', grad_fn=)\n", - "76000\n", - "7700 tensor(5.3934, device='cuda:0', grad_fn=)\n", - "77000\n", - "7800 tensor(5.5041, device='cuda:0', grad_fn=)\n", - "78000\n", - "7900 tensor(5.1814, device='cuda:0', grad_fn=)\n", - "79000\n", - "8000 tensor(5.2426, device='cuda:0', grad_fn=)\n", - "80000\n", - "8100 tensor(5.4104, device='cuda:0', grad_fn=)\n", - "81000\n", - "8200 tensor(5.4198, device='cuda:0', grad_fn=)\n", - "82000\n", - "8300 tensor(5.3854, device='cuda:0', grad_fn=)\n", - "83000\n", - "8400 tensor(5.5128, device='cuda:0', grad_fn=)\n", - "84000\n", - "8500 tensor(5.4898, device='cuda:0', grad_fn=)\n", - "85000\n", - "8600 tensor(5.4943, device='cuda:0', grad_fn=)\n", - "86000\n", - "8700 tensor(5.6012, device='cuda:0', grad_fn=)\n", - "87000\n", - "8800 tensor(5.4790, device='cuda:0', grad_fn=)\n", - "88000\n", - "8900 tensor(5.3312, device='cuda:0', grad_fn=)\n", - "89000\n", - "9000 tensor(5.4456, device='cuda:0', grad_fn=)\n", - "90000\n", - "9100 tensor(5.4537, device='cuda:0', grad_fn=)\n", - "91000\n", - "9200 tensor(5.3643, device='cuda:0', grad_fn=)\n", - "92000\n", - "9300 tensor(5.4085, device='cuda:0', grad_fn=)\n", - "93000\n", - "9400 tensor(5.2527, device='cuda:0', grad_fn=)\n", - "94000\n", - "9500 tensor(5.3289, device='cuda:0', grad_fn=)\n", - "95000\n", - "9600 tensor(5.4516, device='cuda:0', grad_fn=)\n", - "96000\n", - "9700 tensor(5.3881, device='cuda:0', grad_fn=)\n", - "97000\n", - "9800 tensor(5.4321, device='cuda:0', grad_fn=)\n", - "9900 tensor(5.2532, device='cuda:0', grad_fn=)\n", - "98000\n", - "10000 tensor(5.4727, device='cuda:0', grad_fn=)\n", - "99000\n", - "10100 tensor(5.3607, device='cuda:0', grad_fn=)\n", - "100000\n", - "10200 tensor(5.2989, device='cuda:0', grad_fn=)\n", - "101000\n", - "10300 tensor(5.4168, device='cuda:0', grad_fn=)\n", - "102000\n", - "10400 tensor(5.4272, device='cuda:0', grad_fn=)\n", - "103000\n", - "10500 tensor(5.4838, device='cuda:0', grad_fn=)\n", - "104000\n", - "10600 tensor(5.5675, device='cuda:0', grad_fn=)\n", - "105000\n", - "10700 tensor(5.4027, device='cuda:0', grad_fn=)\n", - "106000\n", - "10800 tensor(5.4252, device='cuda:0', grad_fn=)\n", - "107000\n", - "10900 tensor(5.3408, device='cuda:0', grad_fn=)\n", - "108000\n", - "11000 tensor(5.5754, device='cuda:0', grad_fn=)\n", - "109000\n", - "11100 tensor(5.1920, device='cuda:0', grad_fn=)\n", - "110000\n", - "11200 tensor(5.3604, device='cuda:0', grad_fn=)\n", - "111000\n", - "11300 tensor(5.3836, device='cuda:0', grad_fn=)\n", - "112000\n", - "11400 tensor(5.3330, device='cuda:0', grad_fn=)\n", - "113000\n", - "11500 tensor(5.4023, device='cuda:0', grad_fn=)\n", - "114000\n", - "11600 tensor(5.3923, device='cuda:0', grad_fn=)\n", - "115000\n", - "11700 tensor(5.3145, device='cuda:0', grad_fn=)\n", - "116000\n", - "11800 tensor(5.5174, device='cuda:0', grad_fn=)\n", - "117000\n", - "11900 tensor(5.3522, device='cuda:0', grad_fn=)\n", - "118000\n", - "12000 tensor(5.4232, device='cuda:0', grad_fn=)\n", - "119000\n", - "12100 tensor(5.4382, device='cuda:0', grad_fn=)\n", - "120000\n", - "12200 tensor(5.4488, device='cuda:0', grad_fn=)\n", - "121000\n", - "12300 tensor(5.5409, device='cuda:0', grad_fn=)\n", - "122000\n", - "12400 tensor(5.4200, device='cuda:0', grad_fn=)\n", - "123000\n", - "12500 tensor(5.3292, device='cuda:0', grad_fn=)\n", - "124000\n", - "12600 tensor(5.3788, device='cuda:0', grad_fn=)\n", - "125000\n", - "12700 tensor(5.3116, device='cuda:0', grad_fn=)\n", - "126000\n", - "12800 tensor(5.4948, device='cuda:0', grad_fn=)\n", - "127000\n", - "12900 tensor(5.3557, device='cuda:0', grad_fn=)\n", - "128000\n", - "13000 tensor(5.1732, device='cuda:0', grad_fn=)\n", - "129000\n", - "13100 tensor(5.3782, device='cuda:0', grad_fn=)\n", - "130000\n", - "13200 tensor(5.4178, device='cuda:0', grad_fn=)\n", - "131000\n", - "13300 tensor(5.2929, device='cuda:0', grad_fn=)\n", - "132000\n", - "13400 tensor(5.3806, device='cuda:0', grad_fn=)\n", - "133000\n", - "13500 tensor(5.3394, device='cuda:0', grad_fn=)\n", - "134000\n", - "13600 tensor(5.4191, device='cuda:0', grad_fn=)\n", - "135000\n", - "13700 tensor(5.3856, device='cuda:0', grad_fn=)\n", - "136000\n", - "13800 tensor(5.3839, device='cuda:0', grad_fn=)\n", - "137000\n", - "13900 tensor(5.2391, device='cuda:0', grad_fn=)\n", - "138000\n", - "14000 tensor(5.4865, device='cuda:0', grad_fn=)\n", - "139000\n", - "14100 tensor(5.1952, device='cuda:0', grad_fn=)\n", - "140000\n", - "14200 tensor(5.4670, device='cuda:0', grad_fn=)\n", - "141000\n", - "14300 tensor(5.4385, device='cuda:0', grad_fn=)\n", - "142000\n", - "14400 tensor(5.3347, device='cuda:0', grad_fn=)\n", - "143000\n", - "14500 tensor(5.4370, device='cuda:0', grad_fn=)\n", - "144000\n", - "14600 tensor(5.4695, device='cuda:0', grad_fn=)\n", - "145000\n", - "14700 tensor(5.3453, device='cuda:0', grad_fn=)\n", - "146000\n", - "14800 tensor(5.7928, device='cuda:0', grad_fn=)\n", - "147000\n", - "14900 tensor(5.4451, device='cuda:0', grad_fn=)\n", - "148000\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "15000 tensor(5.3087, device='cuda:0', grad_fn=)\n", - "149000\n", - "15100 tensor(5.5241, device='cuda:0', grad_fn=)\n", - "150000\n", - "15200 tensor(5.3894, device='cuda:0', grad_fn=)\n", - "151000\n", - "15300 tensor(5.3809, device='cuda:0', grad_fn=)\n", - "152000\n", - "15400 tensor(5.2696, device='cuda:0', grad_fn=)\n", - "153000\n", - "15500 tensor(5.4343, device='cuda:0', grad_fn=)\n", - "154000\n", - "15600 tensor(5.4322, device='cuda:0', grad_fn=)\n", - "155000\n", - "15700 tensor(5.3296, device='cuda:0', grad_fn=)\n", - "156000\n", - "15800 tensor(5.2456, device='cuda:0', grad_fn=)\n", - "157000\n", - "15900 tensor(5.3806, device='cuda:0', grad_fn=)\n", - "158000\n", - "16000 tensor(5.2008, device='cuda:0', grad_fn=)\n", - "159000\n", - "16100 tensor(5.2489, device='cuda:0', grad_fn=)\n", - "160000\n", - "16200 tensor(5.5902, device='cuda:0', grad_fn=)\n", - "161000\n", - "16300 tensor(5.4159, device='cuda:0', grad_fn=)\n", - "162000\n", - "16400 tensor(5.3966, device='cuda:0', grad_fn=)\n", - "163000\n", - "16500 tensor(5.5113, device='cuda:0', grad_fn=)\n", - "164000\n", - "16600 tensor(5.3599, device='cuda:0', grad_fn=)\n", - "165000\n", - "16700 tensor(5.3372, device='cuda:0', grad_fn=)\n", - "166000\n", - "16800 tensor(5.4158, device='cuda:0', grad_fn=)\n", - "167000\n", - "16900 tensor(5.1788, device='cuda:0', grad_fn=)\n", - "168000\n", - "17000 tensor(5.4497, device='cuda:0', grad_fn=)\n", - "169000\n", - "17100 tensor(5.2981, device='cuda:0', grad_fn=)\n", - "170000\n", - "17200 tensor(5.4330, device='cuda:0', grad_fn=)\n", - "171000\n", - "17300 tensor(5.4495, device='cuda:0', grad_fn=)\n", - "172000\n", - "17400 tensor(5.2431, device='cuda:0', grad_fn=)\n", - "173000\n", - "17500 tensor(5.2652, device='cuda:0', grad_fn=)\n", - "174000\n", - "17600 tensor(5.3007, device='cuda:0', grad_fn=)\n", - "175000\n", - "17700 tensor(5.2852, device='cuda:0', grad_fn=)\n", - "176000\n", - "17800 tensor(5.3431, device='cuda:0', grad_fn=)\n", - "177000\n", - "17900 tensor(5.2395, device='cuda:0', grad_fn=)\n", - "178000\n", - "18000 tensor(5.4841, device='cuda:0', grad_fn=)\n", - "179000\n", - "18100 tensor(5.4218, device='cuda:0', grad_fn=)\n", - "180000\n", - "18200 tensor(5.3397, device='cuda:0', grad_fn=)\n", - "181000\n", - "18300 tensor(5.3426, device='cuda:0', grad_fn=)\n", - "182000\n", - "18400 tensor(5.3654, device='cuda:0', grad_fn=)\n", - "183000\n", - "18500 tensor(5.3484, device='cuda:0', grad_fn=)\n", - "184000\n", - "18600 tensor(5.5509, device='cuda:0', grad_fn=)\n", - "185000\n", - "18700 tensor(5.3702, device='cuda:0', grad_fn=)\n", - "186000\n", - "18800 tensor(5.5361, device='cuda:0', grad_fn=)\n", - "187000\n", - "18900 tensor(5.4132, device='cuda:0', grad_fn=)\n", - "188000\n", - "19000 tensor(5.4235, device='cuda:0', grad_fn=)\n", - "189000\n", - "19100 tensor(5.5318, device='cuda:0', grad_fn=)\n", - "190000\n", - "19200 tensor(5.4136, device='cuda:0', grad_fn=)\n", - "191000\n", - "19300 tensor(5.5053, device='cuda:0', grad_fn=)\n", - "192000\n", - "19400 tensor(5.3472, device='cuda:0', grad_fn=)\n", - "193000\n", - "19500 tensor(5.3511, device='cuda:0', grad_fn=)\n", - "194000\n", - "19600 tensor(5.3861, device='cuda:0', grad_fn=)\n", - "195000\n", - "19700 tensor(5.4345, device='cuda:0', grad_fn=)\n", - "196000\n", - "19800 tensor(5.3067, device='cuda:0', grad_fn=)\n", - "197000\n", - "19900 tensor(5.3079, device='cuda:0', grad_fn=)\n", - "198000\n", - "20000 tensor(5.3268, device='cuda:0', grad_fn=)\n", - "199000\n", - "20100 tensor(5.2668, device='cuda:0', grad_fn=)\n", - "200000\n", - "20200 tensor(5.1998, device='cuda:0', grad_fn=)\n", - "201000\n", - "20300 tensor(5.3105, device='cuda:0', grad_fn=)\n", - "20400 tensor(5.3584, device='cuda:0', grad_fn=)\n", - "202000\n", - "20500 tensor(5.3580, device='cuda:0', grad_fn=)\n", - "203000\n", - "20600 tensor(5.5528, device='cuda:0', grad_fn=)\n", - "204000\n", - "20700 tensor(5.3871, device='cuda:0', grad_fn=)\n", - "205000\n", - "20800 tensor(5.2208, device='cuda:0', grad_fn=)\n", - "206000\n", - "20900 tensor(5.5007, device='cuda:0', grad_fn=)\n", - "207000\n", - "21000 tensor(5.3396, device='cuda:0', grad_fn=)\n", - "208000\n", - "21100 tensor(5.3407, device='cuda:0', grad_fn=)\n", - "209000\n", - "21200 tensor(5.2243, device='cuda:0', grad_fn=)\n", - "210000\n", - "21300 tensor(5.4206, device='cuda:0', grad_fn=)\n", - "211000\n", - "21400 tensor(5.4574, device='cuda:0', grad_fn=)\n", - "212000\n", - "21500 tensor(5.2328, device='cuda:0', grad_fn=)\n", - "213000\n", - "21600 tensor(5.2233, device='cuda:0', grad_fn=)\n", - "214000\n", - "21700 tensor(5.2152, device='cuda:0', grad_fn=)\n", - "215000\n", - "21800 tensor(5.3497, device='cuda:0', grad_fn=)\n", - "216000\n", - "21900 tensor(5.3425, device='cuda:0', grad_fn=)\n", - "217000\n", - "22000 tensor(5.3277, device='cuda:0', grad_fn=)\n", - "218000\n", - "22100 tensor(5.2012, device='cuda:0', grad_fn=)\n", - "219000\n", - "22200 tensor(5.0736, device='cuda:0', grad_fn=)\n", - "220000\n", - "22300 tensor(5.5070, device='cuda:0', grad_fn=)\n", - "221000\n", - "22400 tensor(5.2190, device='cuda:0', grad_fn=)\n", - "222000\n", - "22500 tensor(5.2434, device='cuda:0', grad_fn=)\n", - "223000\n", - "22600 tensor(5.4325, device='cuda:0', grad_fn=)\n", - "224000\n", - "22700 tensor(5.1909, device='cuda:0', grad_fn=)\n", - "225000\n", - "22800 tensor(5.4576, device='cuda:0', grad_fn=)\n", - "226000\n", - "22900 tensor(5.5069, device='cuda:0', grad_fn=)\n", - "227000\n", - "23000 tensor(5.4041, device='cuda:0', grad_fn=)\n", - "228000\n", - "23100 tensor(5.3908, device='cuda:0', grad_fn=)\n", - "229000\n", - "23200 tensor(5.3866, device='cuda:0', grad_fn=)\n", - "230000\n", - "23300 tensor(5.4714, device='cuda:0', grad_fn=)\n", - "231000\n", - "23400 tensor(5.4781, device='cuda:0', grad_fn=)\n", - "232000\n", - "23500 tensor(5.3154, device='cuda:0', grad_fn=)\n", - "233000\n", - "23600 tensor(5.2854, device='cuda:0', grad_fn=)\n", - "234000\n", - "23700 tensor(5.3050, device='cuda:0', grad_fn=)\n", - "235000\n", - "23800 tensor(5.1721, device='cuda:0', grad_fn=)\n", - "236000\n", - "23900 tensor(5.2637, device='cuda:0', grad_fn=)\n", - "237000\n", - "24000 tensor(5.2519, device='cuda:0', grad_fn=)\n", - "238000\n", - "24100 tensor(5.3407, device='cuda:0', grad_fn=)\n", - "239000\n", - "24200 tensor(5.5137, device='cuda:0', grad_fn=)\n", - "240000\n", - "24300 tensor(5.4080, device='cuda:0', grad_fn=)\n", - "241000\n", - "24400 tensor(5.5379, device='cuda:0', grad_fn=)\n", - "242000\n", - "24500 tensor(5.3255, device='cuda:0', grad_fn=)\n", - "243000\n", - "24600 tensor(5.4515, device='cuda:0', grad_fn=)\n", - "244000\n", - "24700 tensor(5.3535, device='cuda:0', grad_fn=)\n", - "245000\n", - "24800 tensor(5.3935, device='cuda:0', grad_fn=)\n", - "246000\n", - "24900 tensor(5.4553, device='cuda:0', grad_fn=)\n", - "247000\n", - "25000 tensor(5.4708, device='cuda:0', grad_fn=)\n", - "248000\n", - "25100 tensor(5.3920, device='cuda:0', grad_fn=)\n", - "249000\n", - "25200 tensor(5.4083, device='cuda:0', grad_fn=)\n", - "250000\n", - "25300 tensor(5.4332, device='cuda:0', grad_fn=)\n", - "251000\n", - "25400 tensor(5.4136, device='cuda:0', grad_fn=)\n", - "252000\n", - "25500 tensor(5.3147, device='cuda:0', grad_fn=)\n", - "253000\n", - "25600 tensor(5.5860, device='cuda:0', grad_fn=)\n", - "254000\n", - "25700 tensor(5.3490, device='cuda:0', grad_fn=)\n", - "255000\n", - "25800 tensor(5.4464, device='cuda:0', grad_fn=)\n", - "256000\n", - "25900 tensor(5.3857, device='cuda:0', grad_fn=)\n", - "257000\n", - "26000 tensor(5.3893, device='cuda:0', grad_fn=)\n", - "258000\n", - "26100 tensor(5.3041, device='cuda:0', grad_fn=)\n", - "259000\n", - "26200 tensor(5.2321, device='cuda:0', grad_fn=)\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "260000\n", - "26300 tensor(5.4289, device='cuda:0', grad_fn=)\n", - "261000\n", - "26400 tensor(5.4663, device='cuda:0', grad_fn=)\n", - "262000\n", - "26500 tensor(5.1922, device='cuda:0', grad_fn=)\n", - "263000\n", - "26600 tensor(5.5283, device='cuda:0', grad_fn=)\n", - "264000\n", - "26700 tensor(5.3933, device='cuda:0', grad_fn=)\n", - "265000\n", - "26800 tensor(5.5680, device='cuda:0', grad_fn=)\n", - "266000\n", - "26900 tensor(5.3281, device='cuda:0', grad_fn=)\n", - "267000\n", - "27000 tensor(5.2408, device='cuda:0', grad_fn=)\n", - "268000\n", - "27100 tensor(5.2671, device='cuda:0', grad_fn=)\n", - "269000\n", - "27200 tensor(5.3099, device='cuda:0', grad_fn=)\n", - "270000\n", - "27300 tensor(5.5049, device='cuda:0', grad_fn=)\n", - "271000\n", - "27400 tensor(5.3850, device='cuda:0', grad_fn=)\n", - "272000\n", - "27500 tensor(5.2843, device='cuda:0', grad_fn=)\n", - "273000\n", - "27600 tensor(5.5777, device='cuda:0', grad_fn=)\n", - "274000\n", - "27700 tensor(5.4017, device='cuda:0', grad_fn=)\n", - "275000\n", - "27800 tensor(5.3994, device='cuda:0', grad_fn=)\n", - "276000\n", - "27900 tensor(5.5128, device='cuda:0', grad_fn=)\n", - "277000\n", - "28000 tensor(5.3708, device='cuda:0', grad_fn=)\n", - "278000\n", - "28100 tensor(5.3382, device='cuda:0', grad_fn=)\n", - "279000\n", - "28200 tensor(5.4996, device='cuda:0', grad_fn=)\n", - "280000\n", - "28300 tensor(5.1214, device='cuda:0', grad_fn=)\n", - "281000\n", - "28400 tensor(5.5647, device='cuda:0', grad_fn=)\n", - "282000\n", - "28500 tensor(5.3959, device='cuda:0', grad_fn=)\n", - "283000\n", - "28600 tensor(5.3312, device='cuda:0', grad_fn=)\n", - "284000\n", - "28700 tensor(5.4663, device='cuda:0', grad_fn=)\n", - "285000\n", - "28800 tensor(5.5155, device='cuda:0', grad_fn=)\n", - "286000\n", - "28900 tensor(5.3872, device='cuda:0', grad_fn=)\n", - "287000\n", - "29000 tensor(5.3017, device='cuda:0', grad_fn=)\n", - "288000\n", - "29100 tensor(5.0583, device='cuda:0', grad_fn=)\n", - "289000\n", - "29200 tensor(5.2099, device='cuda:0', grad_fn=)\n", - "290000\n", - "29300 tensor(5.4934, device='cuda:0', grad_fn=)\n", - "291000\n", - "29400 tensor(5.6202, device='cuda:0', grad_fn=)\n", - "292000\n", - "29500 tensor(5.4016, device='cuda:0', grad_fn=)\n", - "293000\n", - "29600 tensor(5.2601, device='cuda:0', grad_fn=)\n", - "294000\n", - "29700 tensor(5.4038, device='cuda:0', grad_fn=)\n", - "295000\n", - "29800 tensor(5.2475, device='cuda:0', grad_fn=)\n", - "296000\n", - "29900 tensor(5.4960, device='cuda:0', grad_fn=)\n", - "297000\n", - "30000 tensor(5.2438, device='cuda:0', grad_fn=)\n", - "298000\n", - "30100 tensor(5.3221, device='cuda:0', grad_fn=)\n", - "299000\n", - "30200 tensor(5.2686, device='cuda:0', grad_fn=)\n", - "300000\n", - "30300 tensor(5.3735, device='cuda:0', grad_fn=)\n", - "301000\n", - "30400 tensor(5.2057, device='cuda:0', grad_fn=)\n", - "302000\n", - "30500 tensor(5.3767, device='cuda:0', grad_fn=)\n", - "30600 tensor(5.3515, device='cuda:0', grad_fn=)\n", - "303000\n", - "30700 tensor(5.3841, device='cuda:0', grad_fn=)\n", - "304000\n", - "30800 tensor(5.3889, device='cuda:0', grad_fn=)\n", - "305000\n", - "30900 tensor(5.4117, device='cuda:0', grad_fn=)\n", - "306000\n", - "31000 tensor(5.5205, device='cuda:0', grad_fn=)\n", - "307000\n", - "31100 tensor(5.1742, device='cuda:0', grad_fn=)\n", - "308000\n", - "31200 tensor(5.2173, device='cuda:0', grad_fn=)\n", - "309000\n", - "31300 tensor(5.4785, device='cuda:0', grad_fn=)\n", - "310000\n", - "31400 tensor(5.2577, device='cuda:0', grad_fn=)\n", - "311000\n", - "31500 tensor(5.4429, device='cuda:0', grad_fn=)\n", - "312000\n", - "31600 tensor(5.4289, device='cuda:0', grad_fn=)\n", - "313000\n", - "31700 tensor(5.3961, device='cuda:0', grad_fn=)\n", - "314000\n", - "31800 tensor(5.4999, device='cuda:0', grad_fn=)\n", - "315000\n", - "31900 tensor(5.1248, device='cuda:0', grad_fn=)\n", - "316000\n", - "32000 tensor(5.3122, device='cuda:0', grad_fn=)\n", - "317000\n", - "32100 tensor(5.1931, device='cuda:0', grad_fn=)\n", - "318000\n", - "32200 tensor(5.5096, device='cuda:0', grad_fn=)\n", - "319000\n", - "32300 tensor(5.4973, device='cuda:0', grad_fn=)\n", - "320000\n", - "32400 tensor(5.4742, device='cuda:0', grad_fn=)\n", - "321000\n", - "32500 tensor(5.2964, device='cuda:0', grad_fn=)\n", - "322000\n", - "32600 tensor(5.4063, device='cuda:0', grad_fn=)\n", - "323000\n", - "32700 tensor(5.3369, device='cuda:0', grad_fn=)\n", - "324000\n", - "32800 tensor(5.5636, device='cuda:0', grad_fn=)\n", - "325000\n", - "32900 tensor(5.4245, device='cuda:0', grad_fn=)\n", - "326000\n", - "33000 tensor(5.2032, device='cuda:0', grad_fn=)\n", - "327000\n", - "33100 tensor(5.4095, device='cuda:0', grad_fn=)\n", - "328000\n", - "33200 tensor(5.5071, device='cuda:0', grad_fn=)\n", - "329000\n", - "33300 tensor(5.2729, device='cuda:0', grad_fn=)\n", - "330000\n", - "33400 tensor(5.5492, device='cuda:0', grad_fn=)\n", - "331000\n", - "33500 tensor(5.3701, device='cuda:0', grad_fn=)\n", - "332000\n", - "33600 tensor(5.3223, device='cuda:0', grad_fn=)\n", - "333000\n", - "33700 tensor(5.3725, device='cuda:0', grad_fn=)\n", - "334000\n", - "33800 tensor(5.4572, device='cuda:0', grad_fn=)\n", - "335000\n", - "33900 tensor(5.1889, device='cuda:0', grad_fn=)\n", - "336000\n", - "34000 tensor(5.4090, device='cuda:0', grad_fn=)\n", - "337000\n", - "34100 tensor(5.3798, device='cuda:0', grad_fn=)\n", - "338000\n", - "34200 tensor(5.4259, device='cuda:0', grad_fn=)\n", - "339000\n", - "34300 tensor(5.2132, device='cuda:0', grad_fn=)\n", - "340000\n", - "34400 tensor(5.6692, device='cuda:0', grad_fn=)\n", - "341000\n", - "34500 tensor(5.5324, device='cuda:0', grad_fn=)\n", - "342000\n", - "34600 tensor(5.4271, device='cuda:0', grad_fn=)\n", - "343000\n", - "34700 tensor(5.4978, device='cuda:0', grad_fn=)\n", - "344000\n", - "34800 tensor(5.5230, device='cuda:0', grad_fn=)\n", - "345000\n", - "34900 tensor(5.5652, device='cuda:0', grad_fn=)\n", - "346000\n", - "35000 tensor(5.5478, device='cuda:0', grad_fn=)\n", - "347000\n", - "35100 tensor(5.3700, device='cuda:0', grad_fn=)\n", - "348000\n", - "35200 tensor(5.2958, device='cuda:0', grad_fn=)\n", - "349000\n", - "35300 tensor(5.5219, device='cuda:0', grad_fn=)\n", - "350000\n", - "35400 tensor(5.1702, device='cuda:0', grad_fn=)\n", - "351000\n", - "35500 tensor(5.2604, device='cuda:0', grad_fn=)\n", - "352000\n", - "35600 tensor(5.3821, device='cuda:0', grad_fn=)\n", - "353000\n", - "35700 tensor(5.2551, device='cuda:0', grad_fn=)\n", - "354000\n", - "35800 tensor(5.3840, device='cuda:0', grad_fn=)\n", - "355000\n", - "35900 tensor(5.3635, device='cuda:0', grad_fn=)\n", - "356000\n", - "36000 tensor(5.1400, device='cuda:0', grad_fn=)\n", - "357000\n", - "36100 tensor(5.5134, device='cuda:0', grad_fn=)\n", - "358000\n", - "36200 tensor(5.3632, device='cuda:0', grad_fn=)\n", - "359000\n", - "36300 tensor(5.6461, device='cuda:0', grad_fn=)\n", - "360000\n", - "36400 tensor(5.3415, device='cuda:0', grad_fn=)\n", - "361000\n", - "36500 tensor(5.3659, device='cuda:0', grad_fn=)\n", - "362000\n", - "36600 tensor(5.3874, device='cuda:0', grad_fn=)\n", - "363000\n", - "36700 tensor(5.1886, device='cuda:0', grad_fn=)\n", - "364000\n", - "36800 tensor(5.2958, device='cuda:0', grad_fn=)\n", - "365000\n", - "36900 tensor(5.4094, device='cuda:0', grad_fn=)\n", - "366000\n", - "37000 tensor(5.3023, device='cuda:0', grad_fn=)\n", - "367000\n", - "37100 tensor(5.3287, device='cuda:0', grad_fn=)\n", - "368000\n", - "37200 tensor(5.3996, device='cuda:0', grad_fn=)\n", - "369000\n", - "37300 tensor(5.3001, device='cuda:0', grad_fn=)\n", - "370000\n", - "37400 tensor(5.6516, device='cuda:0', grad_fn=)\n", - "371000\n", - "37500 tensor(5.3366, device='cuda:0', grad_fn=)\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "372000\n", - "37600 tensor(5.3282, device='cuda:0', grad_fn=)\n", - "373000\n", - "37700 tensor(5.5061, device='cuda:0', grad_fn=)\n", - "374000\n", - "37800 tensor(5.3408, device='cuda:0', grad_fn=)\n", - "375000\n", - "37900 tensor(5.3203, device='cuda:0', grad_fn=)\n", - "376000\n", - "38000 tensor(5.3996, device='cuda:0', grad_fn=)\n", - "377000\n", - "38100 tensor(5.4133, device='cuda:0', grad_fn=)\n", - "378000\n", - "38200 tensor(5.4262, device='cuda:0', grad_fn=)\n", - "379000\n", - "38300 tensor(5.3305, device='cuda:0', grad_fn=)\n", - "380000\n", - "38400 tensor(5.3983, device='cuda:0', grad_fn=)\n", - "381000\n", - "38500 tensor(5.4246, device='cuda:0', grad_fn=)\n", - "382000\n", - "38600 tensor(5.3713, device='cuda:0', grad_fn=)\n", - "383000\n", - "38700 tensor(5.3634, device='cuda:0', grad_fn=)\n", - "384000\n", - "38800 tensor(5.4504, device='cuda:0', grad_fn=)\n", - "385000\n", - "38900 tensor(5.5273, device='cuda:0', grad_fn=)\n", - "386000\n", - "39000 tensor(5.2229, device='cuda:0', grad_fn=)\n", - "387000\n", - "39100 tensor(5.4503, device='cuda:0', grad_fn=)\n", - "388000\n", - "39200 tensor(5.5406, device='cuda:0', grad_fn=)\n", - "389000\n", - "39300 tensor(5.3640, device='cuda:0', grad_fn=)\n", - "390000\n", - "39400 tensor(5.4311, device='cuda:0', grad_fn=)\n", - "391000\n", - "39500 tensor(5.5292, device='cuda:0', grad_fn=)\n", - "392000\n", - "39600 tensor(5.2217, device='cuda:0', grad_fn=)\n", - "393000\n", - "39700 tensor(5.2121, device='cuda:0', grad_fn=)\n", - "394000\n", - "39800 tensor(5.3415, device='cuda:0', grad_fn=)\n", - "395000\n", - "39900 tensor(5.1605, device='cuda:0', grad_fn=)\n", - "396000\n", - "40000 tensor(5.2472, device='cuda:0', grad_fn=)\n", - "397000\n", - "40100 tensor(5.3351, device='cuda:0', grad_fn=)\n", - "398000\n", - "40200 tensor(5.3198, device='cuda:0', grad_fn=)\n", - "399000\n", - "40300 tensor(5.3862, device='cuda:0', grad_fn=)\n", - "40400 tensor(5.3946, device='cuda:0', grad_fn=)\n", - "400000\n", - "40500 tensor(5.3120, device='cuda:0', grad_fn=)\n", - "401000\n", - "40600 tensor(5.3741, device='cuda:0', grad_fn=)\n", - "402000\n", - "40700 tensor(5.4199, device='cuda:0', grad_fn=)\n", - "403000\n", - "40800 tensor(5.3702, device='cuda:0', grad_fn=)\n", - "404000\n", - "40900 tensor(5.3212, device='cuda:0', grad_fn=)\n", - "405000\n", - "41000 tensor(5.3683, device='cuda:0', grad_fn=)\n", - "406000\n", - "41100 tensor(5.3491, device='cuda:0', grad_fn=)\n", - "407000\n", - "41200 tensor(5.2400, device='cuda:0', grad_fn=)\n", - "408000\n", - "41300 tensor(5.3728, device='cuda:0', grad_fn=)\n", - "409000\n", - "41400 tensor(5.2643, device='cuda:0', grad_fn=)\n", - "410000\n", - "41500 tensor(5.4064, device='cuda:0', grad_fn=)\n", - "411000\n", - "41600 tensor(5.3238, device='cuda:0', grad_fn=)\n", - "412000\n", - "41700 tensor(5.3469, device='cuda:0', grad_fn=)\n", - "413000\n", - "41800 tensor(5.3432, device='cuda:0', grad_fn=)\n", - "414000\n", - "41900 tensor(5.3521, device='cuda:0', grad_fn=)\n", - "415000\n", - "42000 tensor(5.5087, device='cuda:0', grad_fn=)\n", - "416000\n", - "42100 tensor(5.2556, device='cuda:0', grad_fn=)\n", - "417000\n", - "42200 tensor(5.3407, device='cuda:0', grad_fn=)\n", - "418000\n", - "42300 tensor(5.4058, device='cuda:0', grad_fn=)\n", - "419000\n", - "42400 tensor(5.2231, device='cuda:0', grad_fn=)\n", - "420000\n", - "42500 tensor(5.3912, device='cuda:0', grad_fn=)\n", - "421000\n", - "42600 tensor(5.1878, device='cuda:0', grad_fn=)\n", - "422000\n", - "42700 tensor(5.4955, device='cuda:0', grad_fn=)\n", - "423000\n", - "42800 tensor(5.4193, device='cuda:0', grad_fn=)\n", - "424000\n", - "42900 tensor(5.2662, device='cuda:0', grad_fn=)\n", - "425000\n", - "43000 tensor(5.4093, device='cuda:0', grad_fn=)\n", - "426000\n", - "43100 tensor(5.4089, device='cuda:0', grad_fn=)\n", - "427000\n", - "43200 tensor(5.2223, device='cuda:0', grad_fn=)\n", - "428000\n", - "43300 tensor(5.2456, device='cuda:0', grad_fn=)\n", - "429000\n", - "43400 tensor(5.4129, device='cuda:0', grad_fn=)\n", - "430000\n", - "43500 tensor(5.1283, device='cuda:0', grad_fn=)\n", - "431000\n", - "43600 tensor(5.3275, device='cuda:0', grad_fn=)\n", - "432000\n", - "epoch: = 4\n", - "0 tensor(5.3172, device='cuda:0', grad_fn=)\n", - "1000\n", - "100 tensor(5.3864, device='cuda:0', grad_fn=)\n", - "200 tensor(5.2618, device='cuda:0', grad_fn=)\n", - "2000\n", - "300 tensor(5.2652, device='cuda:0', grad_fn=)\n", - "3000\n", - "400 tensor(5.2749, device='cuda:0', grad_fn=)\n", - "4000\n", - "500 tensor(5.4347, device='cuda:0', grad_fn=)\n", - "5000\n", - "600 tensor(5.2271, device='cuda:0', grad_fn=)\n", - "6000\n", - "700 tensor(5.5396, device='cuda:0', grad_fn=)\n", - "7000\n", - "800 tensor(5.1379, device='cuda:0', grad_fn=)\n", - "8000\n", - "900 tensor(5.3861, device='cuda:0', grad_fn=)\n", - "9000\n", - "1000 tensor(5.2629, device='cuda:0', grad_fn=)\n", - "10000\n", - "1100 tensor(5.4575, device='cuda:0', grad_fn=)\n", - "11000\n", - "1200 tensor(5.4936, device='cuda:0', grad_fn=)\n", - "12000\n", - "1300 tensor(5.4281, device='cuda:0', grad_fn=)\n", - "13000\n", - "1400 tensor(5.4186, device='cuda:0', grad_fn=)\n", - "14000\n", - "1500 tensor(5.5070, device='cuda:0', grad_fn=)\n", - "15000\n", - "1600 tensor(5.1769, device='cuda:0', grad_fn=)\n", - "16000\n", - "1700 tensor(5.2856, device='cuda:0', grad_fn=)\n", - "17000\n", - "1800 tensor(5.2827, device='cuda:0', grad_fn=)\n", - "18000\n", - "1900 tensor(5.2544, device='cuda:0', grad_fn=)\n", - "19000\n", - "2000 tensor(5.3218, device='cuda:0', grad_fn=)\n", - "20000\n", - "2100 tensor(5.4549, device='cuda:0', grad_fn=)\n", - "21000\n", - "2200 tensor(5.2864, device='cuda:0', grad_fn=)\n", - "22000\n", - "2300 tensor(5.3145, device='cuda:0', grad_fn=)\n", - "23000\n", - "2400 tensor(5.2987, device='cuda:0', grad_fn=)\n", - "24000\n", - "2500 tensor(5.3498, device='cuda:0', grad_fn=)\n", - "25000\n", - "2600 tensor(5.3730, device='cuda:0', grad_fn=)\n", - "26000\n", - "2700 tensor(5.4017, device='cuda:0', grad_fn=)\n", - "27000\n", - "2800 tensor(5.4255, device='cuda:0', grad_fn=)\n", - "28000\n", - "2900 tensor(5.5475, device='cuda:0', grad_fn=)\n", - "29000\n", - "3000 tensor(5.2988, device='cuda:0', grad_fn=)\n", - "30000\n", - "3100 tensor(5.3753, device='cuda:0', grad_fn=)\n", - "31000\n", - "3200 tensor(5.4049, device='cuda:0', grad_fn=)\n", - "32000\n", - "3300 tensor(5.3206, device='cuda:0', grad_fn=)\n", - "33000\n", - "3400 tensor(5.2159, device='cuda:0', grad_fn=)\n", - "34000\n", - "3500 tensor(5.3423, device='cuda:0', grad_fn=)\n", - "35000\n", - "3600 tensor(5.3717, device='cuda:0', grad_fn=)\n", - "36000\n", - "3700 tensor(5.3042, device='cuda:0', grad_fn=)\n", - "37000\n", - "3800 tensor(5.3258, device='cuda:0', grad_fn=)\n", - "38000\n", - "3900 tensor(5.1989, device='cuda:0', grad_fn=)\n", - "39000\n", - "4000 tensor(5.2650, device='cuda:0', grad_fn=)\n", - "40000\n", - "4100 tensor(5.3953, device='cuda:0', grad_fn=)\n", - "41000\n", - "4200 tensor(5.4542, device='cuda:0', grad_fn=)\n", - "42000\n", - "4300 tensor(5.3466, device='cuda:0', grad_fn=)\n", - "43000\n", - "4400 tensor(5.4222, device='cuda:0', grad_fn=)\n", - "44000\n", - "4500 tensor(5.2254, device='cuda:0', grad_fn=)\n", - "45000\n", - "4600 tensor(5.5610, device='cuda:0', grad_fn=)\n", - "46000\n", - "4700 tensor(5.2753, device='cuda:0', grad_fn=)\n", - "47000\n", - "4800 tensor(5.4028, device='cuda:0', grad_fn=)\n", - "48000\n", - "4900 tensor(5.4516, device='cuda:0', grad_fn=)\n", - "49000\n", - "5000 tensor(5.4464, device='cuda:0', grad_fn=)\n", - "50000\n", - "5100 tensor(5.5018, device='cuda:0', grad_fn=)\n", - "51000\n", - "5200 tensor(5.5194, device='cuda:0', grad_fn=)\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "52000\n", - "5300 tensor(5.5077, device='cuda:0', grad_fn=)\n", - "53000\n", - "5400 tensor(5.3746, device='cuda:0', grad_fn=)\n", - "54000\n", - "5500 tensor(5.4847, device='cuda:0', grad_fn=)\n", - "55000\n", - "5600 tensor(5.2664, device='cuda:0', grad_fn=)\n", - "56000\n", - "5700 tensor(5.5265, device='cuda:0', grad_fn=)\n", - "57000\n", - "5800 tensor(5.5101, device='cuda:0', grad_fn=)\n", - "58000\n", - "5900 tensor(5.4513, device='cuda:0', grad_fn=)\n", - "59000\n", - "6000 tensor(5.3554, device='cuda:0', grad_fn=)\n", - "60000\n", - "6100 tensor(5.3616, device='cuda:0', grad_fn=)\n", - "61000\n", - "6200 tensor(5.5360, device='cuda:0', grad_fn=)\n", - "62000\n", - "6300 tensor(5.3952, device='cuda:0', grad_fn=)\n", - "63000\n", - "6400 tensor(5.3132, device='cuda:0', grad_fn=)\n", - "64000\n", - "6500 tensor(5.1732, device='cuda:0', grad_fn=)\n", - "65000\n", - "6600 tensor(5.3505, device='cuda:0', grad_fn=)\n", - "66000\n", - "6700 tensor(5.2919, device='cuda:0', grad_fn=)\n", - "67000\n", - "6800 tensor(5.5064, device='cuda:0', grad_fn=)\n", - "68000\n", - "6900 tensor(5.4881, device='cuda:0', grad_fn=)\n", - "69000\n", - "7000 tensor(5.3978, device='cuda:0', grad_fn=)\n", - "70000\n", - "7100 tensor(5.2030, device='cuda:0', grad_fn=)\n", - "71000\n", - "7200 tensor(5.2738, device='cuda:0', grad_fn=)\n", - "72000\n", - "7300 tensor(5.5317, device='cuda:0', grad_fn=)\n", - "73000\n", - "7400 tensor(5.3487, device='cuda:0', grad_fn=)\n", - "74000\n", - "7500 tensor(5.2133, device='cuda:0', grad_fn=)\n", - "75000\n", - "7600 tensor(5.2878, device='cuda:0', grad_fn=)\n", - "76000\n", - "7700 tensor(5.3644, device='cuda:0', grad_fn=)\n", - "77000\n", - "7800 tensor(5.4711, device='cuda:0', grad_fn=)\n", - "78000\n", - "7900 tensor(5.1445, device='cuda:0', grad_fn=)\n", - "79000\n", - "8000 tensor(5.2138, device='cuda:0', grad_fn=)\n", - "80000\n", - "8100 tensor(5.3741, device='cuda:0', grad_fn=)\n", - "81000\n", - "8200 tensor(5.3893, device='cuda:0', grad_fn=)\n", - "82000\n", - "8300 tensor(5.3492, device='cuda:0', grad_fn=)\n", - "83000\n", - "8400 tensor(5.4797, device='cuda:0', grad_fn=)\n", - "84000\n", - "8500 tensor(5.4501, device='cuda:0', grad_fn=)\n", - "85000\n", - "8600 tensor(5.4600, device='cuda:0', grad_fn=)\n", - "86000\n", - "8700 tensor(5.5758, device='cuda:0', grad_fn=)\n", - "87000\n", - "8800 tensor(5.4493, device='cuda:0', grad_fn=)\n", - "88000\n", - "8900 tensor(5.3035, device='cuda:0', grad_fn=)\n", - "89000\n", - "9000 tensor(5.4164, device='cuda:0', grad_fn=)\n", - "90000\n", - "9100 tensor(5.4273, device='cuda:0', grad_fn=)\n", - "91000\n", - "9200 tensor(5.3343, device='cuda:0', grad_fn=)\n", - "92000\n", - "9300 tensor(5.3797, device='cuda:0', grad_fn=)\n", - "93000\n", - "9400 tensor(5.2260, device='cuda:0', grad_fn=)\n", - "94000\n", - "9500 tensor(5.3006, device='cuda:0', grad_fn=)\n", - "95000\n", - "9600 tensor(5.4211, device='cuda:0', grad_fn=)\n", - "96000\n", - "9700 tensor(5.3615, device='cuda:0', grad_fn=)\n", - "97000\n", - "9800 tensor(5.4089, device='cuda:0', grad_fn=)\n", - "9900 tensor(5.2200, device='cuda:0', grad_fn=)\n", - "98000\n", - "10000 tensor(5.4428, device='cuda:0', grad_fn=)\n", - "99000\n", - "10100 tensor(5.3219, device='cuda:0', grad_fn=)\n", - "100000\n", - "10200 tensor(5.2692, device='cuda:0', grad_fn=)\n", - "101000\n", - "10300 tensor(5.3854, device='cuda:0', grad_fn=)\n", - "102000\n", - "10400 tensor(5.3984, device='cuda:0', grad_fn=)\n", - "103000\n", - "10500 tensor(5.4516, device='cuda:0', grad_fn=)\n", - "104000\n", - "10600 tensor(5.5380, device='cuda:0', grad_fn=)\n", - "105000\n", - "10700 tensor(5.3724, device='cuda:0', grad_fn=)\n", - "106000\n", - "10800 tensor(5.3862, device='cuda:0', grad_fn=)\n", - "107000\n", - "10900 tensor(5.3102, device='cuda:0', grad_fn=)\n", - "108000\n", - "11000 tensor(5.5487, device='cuda:0', grad_fn=)\n", - "109000\n", - "11100 tensor(5.1684, device='cuda:0', grad_fn=)\n", - "110000\n", - "11200 tensor(5.3303, device='cuda:0', grad_fn=)\n", - "111000\n", - "11300 tensor(5.3537, device='cuda:0', grad_fn=)\n", - "112000\n", - "11400 tensor(5.3064, device='cuda:0', grad_fn=)\n", - "113000\n", - "11500 tensor(5.3775, device='cuda:0', grad_fn=)\n", - "114000\n", - "11600 tensor(5.3649, device='cuda:0', grad_fn=)\n", - "115000\n", - "11700 tensor(5.2920, device='cuda:0', grad_fn=)\n", - "116000\n", - "11800 tensor(5.4908, device='cuda:0', grad_fn=)\n", - "117000\n", - "11900 tensor(5.3293, device='cuda:0', grad_fn=)\n", - "118000\n", - "12000 tensor(5.3926, device='cuda:0', grad_fn=)\n", - "119000\n", - "12100 tensor(5.4045, device='cuda:0', grad_fn=)\n", - "120000\n", - "12200 tensor(5.4246, device='cuda:0', grad_fn=)\n", - "121000\n", - "12300 tensor(5.5096, device='cuda:0', grad_fn=)\n", - "122000\n", - "12400 tensor(5.3884, device='cuda:0', grad_fn=)\n", - "123000\n", - "12500 tensor(5.3057, device='cuda:0', grad_fn=)\n", - "124000\n", - "12600 tensor(5.3466, device='cuda:0', grad_fn=)\n", - "125000\n", - "12700 tensor(5.2898, device='cuda:0', grad_fn=)\n", - "126000\n", - "12800 tensor(5.4714, device='cuda:0', grad_fn=)\n", - "127000\n", - "12900 tensor(5.3255, device='cuda:0', grad_fn=)\n", - "128000\n", - "13000 tensor(5.1438, device='cuda:0', grad_fn=)\n", - "129000\n", - "13100 tensor(5.3498, device='cuda:0', grad_fn=)\n", - "130000\n", - "13200 tensor(5.3890, device='cuda:0', grad_fn=)\n", - "131000\n", - "13300 tensor(5.2710, device='cuda:0', grad_fn=)\n", - "132000\n", - "13400 tensor(5.3541, device='cuda:0', grad_fn=)\n", - "133000\n", - "13500 tensor(5.3156, device='cuda:0', grad_fn=)\n", - "134000\n", - "13600 tensor(5.3957, device='cuda:0', grad_fn=)\n", - "135000\n", - "13700 tensor(5.3548, device='cuda:0', grad_fn=)\n", - "136000\n", - "13800 tensor(5.3577, device='cuda:0', grad_fn=)\n", - "137000\n", - "13900 tensor(5.2122, device='cuda:0', grad_fn=)\n", - "138000\n", - "14000 tensor(5.4587, device='cuda:0', grad_fn=)\n", - "139000\n", - "14100 tensor(5.1704, device='cuda:0', grad_fn=)\n", - "140000\n", - "14200 tensor(5.4419, device='cuda:0', grad_fn=)\n", - "141000\n", - "14300 tensor(5.4142, device='cuda:0', grad_fn=)\n", - "142000\n", - "14400 tensor(5.3058, device='cuda:0', grad_fn=)\n", - "143000\n", - "14500 tensor(5.4082, device='cuda:0', grad_fn=)\n", - "144000\n", - "14600 tensor(5.4414, device='cuda:0', grad_fn=)\n", - "145000\n", - "14700 tensor(5.3177, device='cuda:0', grad_fn=)\n", - "146000\n", - "14800 tensor(5.7665, device='cuda:0', grad_fn=)\n", - "147000\n", - "14900 tensor(5.4171, device='cuda:0', grad_fn=)\n", - "148000\n", - "15000 tensor(5.2698, device='cuda:0', grad_fn=)\n", - "149000\n", - "15100 tensor(5.4915, device='cuda:0', grad_fn=)\n", - "150000\n", - "15200 tensor(5.3576, device='cuda:0', grad_fn=)\n", - "151000\n", - "15300 tensor(5.3567, device='cuda:0', grad_fn=)\n", - "152000\n", - "15400 tensor(5.2379, device='cuda:0', grad_fn=)\n", - "153000\n", - "15500 tensor(5.4092, device='cuda:0', grad_fn=)\n", - "154000\n", - "15600 tensor(5.4042, device='cuda:0', grad_fn=)\n", - "155000\n", - "15700 tensor(5.3017, device='cuda:0', grad_fn=)\n", - "156000\n", - "15800 tensor(5.2188, device='cuda:0', grad_fn=)\n", - "157000\n", - "15900 tensor(5.3497, device='cuda:0', grad_fn=)\n", - "158000\n", - "16000 tensor(5.1718, device='cuda:0', grad_fn=)\n", - "159000\n", - "16100 tensor(5.2145, device='cuda:0', grad_fn=)\n", - "160000\n", - "16200 tensor(5.5591, device='cuda:0', grad_fn=)\n", - "161000\n", - "16300 tensor(5.3864, device='cuda:0', grad_fn=)\n", - "162000\n", - "16400 tensor(5.3719, device='cuda:0', grad_fn=)\n", - "163000\n", - "16500 tensor(5.4842, device='cuda:0', grad_fn=)\n", - "164000\n", - "16600 tensor(5.3329, device='cuda:0', grad_fn=)\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "165000\n", - "16700 tensor(5.3130, device='cuda:0', grad_fn=)\n", - "166000\n", - "16800 tensor(5.3903, device='cuda:0', grad_fn=)\n", - "167000\n", - "16900 tensor(5.1551, device='cuda:0', grad_fn=)\n", - "168000\n", - "17000 tensor(5.4229, device='cuda:0', grad_fn=)\n", - "169000\n", - "17100 tensor(5.2686, device='cuda:0', grad_fn=)\n", - "170000\n", - "17200 tensor(5.4099, device='cuda:0', grad_fn=)\n", - "171000\n", - "17300 tensor(5.4198, device='cuda:0', grad_fn=)\n", - "172000\n", - "17400 tensor(5.2162, device='cuda:0', grad_fn=)\n", - "173000\n", - "17500 tensor(5.2385, device='cuda:0', grad_fn=)\n", - "174000\n", - "17600 tensor(5.2786, device='cuda:0', grad_fn=)\n", - "175000\n", - "17700 tensor(5.2576, device='cuda:0', grad_fn=)\n", - "176000\n", - "17800 tensor(5.3158, device='cuda:0', grad_fn=)\n", - "177000\n", - "17900 tensor(5.2105, device='cuda:0', grad_fn=)\n", - "178000\n", - "18000 tensor(5.4627, device='cuda:0', grad_fn=)\n", - "179000\n", - "18100 tensor(5.3966, device='cuda:0', grad_fn=)\n", - "180000\n", - "18200 tensor(5.3108, device='cuda:0', grad_fn=)\n", - "181000\n", - "18300 tensor(5.3148, device='cuda:0', grad_fn=)\n", - "182000\n", - "18400 tensor(5.3321, device='cuda:0', grad_fn=)\n", - "183000\n", - "18500 tensor(5.3171, device='cuda:0', grad_fn=)\n", - "184000\n", - "18600 tensor(5.5247, device='cuda:0', grad_fn=)\n", - "185000\n", - "18700 tensor(5.3469, device='cuda:0', grad_fn=)\n", - "186000\n", - "18800 tensor(5.5092, device='cuda:0', grad_fn=)\n", - "187000\n", - "18900 tensor(5.3902, device='cuda:0', grad_fn=)\n", - "188000\n", - "19000 tensor(5.3904, device='cuda:0', grad_fn=)\n", - "189000\n", - "19100 tensor(5.5019, device='cuda:0', grad_fn=)\n", - "190000\n", - "19200 tensor(5.3838, device='cuda:0', grad_fn=)\n", - "191000\n", - "19300 tensor(5.4674, device='cuda:0', grad_fn=)\n", - "192000\n", - "19400 tensor(5.3223, device='cuda:0', grad_fn=)\n", - "193000\n", - "19500 tensor(5.3235, device='cuda:0', grad_fn=)\n", - "194000\n", - "19600 tensor(5.3589, device='cuda:0', grad_fn=)\n", - "195000\n", - "19700 tensor(5.4063, device='cuda:0', grad_fn=)\n", - "196000\n", - "19800 tensor(5.2838, device='cuda:0', grad_fn=)\n", - "197000\n", - "19900 tensor(5.2807, device='cuda:0', grad_fn=)\n", - "198000\n", - "20000 tensor(5.3038, device='cuda:0', grad_fn=)\n", - "199000\n", - "20100 tensor(5.2397, device='cuda:0', grad_fn=)\n", - "200000\n", - "20200 tensor(5.1723, device='cuda:0', grad_fn=)\n", - "201000\n", - "20300 tensor(5.2827, device='cuda:0', grad_fn=)\n", - "20400 tensor(5.3245, device='cuda:0', grad_fn=)\n", - "202000\n", - "20500 tensor(5.3303, device='cuda:0', grad_fn=)\n", - "203000\n", - "20600 tensor(5.5211, device='cuda:0', grad_fn=)\n", - "204000\n", - "20700 tensor(5.3629, device='cuda:0', grad_fn=)\n", - "205000\n", - "20800 tensor(5.1882, device='cuda:0', grad_fn=)\n", - "206000\n", - "20900 tensor(5.4671, device='cuda:0', grad_fn=)\n", - "207000\n", - "21000 tensor(5.3110, device='cuda:0', grad_fn=)\n", - "208000\n", - "21100 tensor(5.3181, device='cuda:0', grad_fn=)\n", - "209000\n", - "21200 tensor(5.1968, device='cuda:0', grad_fn=)\n", - "210000\n", - "21300 tensor(5.3940, device='cuda:0', grad_fn=)\n", - "211000\n", - "21400 tensor(5.4308, device='cuda:0', grad_fn=)\n", - "212000\n", - "21500 tensor(5.2127, device='cuda:0', grad_fn=)\n", - "213000\n", - "21600 tensor(5.2003, device='cuda:0', grad_fn=)\n", - "214000\n", - "21700 tensor(5.1881, device='cuda:0', grad_fn=)\n", - "215000\n", - "21800 tensor(5.3180, device='cuda:0', grad_fn=)\n", - "216000\n", - "21900 tensor(5.3197, device='cuda:0', grad_fn=)\n", - "217000\n", - "22000 tensor(5.3005, device='cuda:0', grad_fn=)\n", - "218000\n", - "22100 tensor(5.1776, device='cuda:0', grad_fn=)\n", - "219000\n", - "22200 tensor(5.0509, device='cuda:0', grad_fn=)\n", - "220000\n", - "22300 tensor(5.4807, device='cuda:0', grad_fn=)\n", - "221000\n", - "22400 tensor(5.2040, device='cuda:0', grad_fn=)\n", - "222000\n", - "22500 tensor(5.2161, device='cuda:0', grad_fn=)\n", - "223000\n", - "22600 tensor(5.4083, device='cuda:0', grad_fn=)\n", - "224000\n", - "22700 tensor(5.1619, device='cuda:0', grad_fn=)\n", - "225000\n", - "22800 tensor(5.4301, device='cuda:0', grad_fn=)\n", - "226000\n", - "22900 tensor(5.4791, device='cuda:0', grad_fn=)\n", - "227000\n", - "23000 tensor(5.3785, device='cuda:0', grad_fn=)\n", - "228000\n", - "23100 tensor(5.3705, device='cuda:0', grad_fn=)\n", - "229000\n", - "23200 tensor(5.3633, device='cuda:0', grad_fn=)\n", - "230000\n", - "23300 tensor(5.4443, device='cuda:0', grad_fn=)\n", - "231000\n", - "23400 tensor(5.4496, device='cuda:0', grad_fn=)\n", - "232000\n", - "23500 tensor(5.2961, device='cuda:0', grad_fn=)\n", - "233000\n", - "23600 tensor(5.2603, device='cuda:0', grad_fn=)\n", - "234000\n", - "23700 tensor(5.2793, device='cuda:0', grad_fn=)\n", - "235000\n", - "23800 tensor(5.1461, device='cuda:0', grad_fn=)\n", - "236000\n", - "23900 tensor(5.2376, device='cuda:0', grad_fn=)\n", - "237000\n", - "24000 tensor(5.2269, device='cuda:0', grad_fn=)\n", - "238000\n", - "24100 tensor(5.3154, device='cuda:0', grad_fn=)\n", - "239000\n", - "24200 tensor(5.4852, device='cuda:0', grad_fn=)\n", - "240000\n", - "24300 tensor(5.3785, device='cuda:0', grad_fn=)\n", - "241000\n", - "24400 tensor(5.5053, device='cuda:0', grad_fn=)\n", - "242000\n", - "24500 tensor(5.2987, device='cuda:0', grad_fn=)\n", - "243000\n", - "24600 tensor(5.4275, device='cuda:0', grad_fn=)\n", - "244000\n", - "24700 tensor(5.3283, device='cuda:0', grad_fn=)\n", - "245000\n", - "24800 tensor(5.3707, device='cuda:0', grad_fn=)\n", - "246000\n", - "24900 tensor(5.4294, device='cuda:0', grad_fn=)\n", - "247000\n", - "25000 tensor(5.4479, device='cuda:0', grad_fn=)\n", - "248000\n", - "25100 tensor(5.3629, device='cuda:0', grad_fn=)\n", - "249000\n", - "25200 tensor(5.3849, device='cuda:0', grad_fn=)\n", - "250000\n", - "25300 tensor(5.4124, device='cuda:0', grad_fn=)\n", - "251000\n", - "25400 tensor(5.3932, device='cuda:0', grad_fn=)\n", - "252000\n", - "25500 tensor(5.2893, device='cuda:0', grad_fn=)\n", - "253000\n", - "25600 tensor(5.5512, device='cuda:0', grad_fn=)\n", - "254000\n", - "25700 tensor(5.3227, device='cuda:0', grad_fn=)\n", - "255000\n", - "25800 tensor(5.4217, device='cuda:0', grad_fn=)\n", - "256000\n", - "25900 tensor(5.3637, device='cuda:0', grad_fn=)\n", - "257000\n", - "26000 tensor(5.3632, device='cuda:0', grad_fn=)\n", - "258000\n", - "26100 tensor(5.2841, device='cuda:0', grad_fn=)\n", - "259000\n", - "26200 tensor(5.2107, device='cuda:0', grad_fn=)\n", - "260000\n", - "26300 tensor(5.4024, device='cuda:0', grad_fn=)\n", - "261000\n", - "26400 tensor(5.4410, device='cuda:0', grad_fn=)\n", - "262000\n", - "26500 tensor(5.1685, device='cuda:0', grad_fn=)\n", - "263000\n", - "26600 tensor(5.5023, device='cuda:0', grad_fn=)\n", - "264000\n", - "26700 tensor(5.3654, device='cuda:0', grad_fn=)\n", - "265000\n", - "26800 tensor(5.5407, device='cuda:0', grad_fn=)\n", - "266000\n", - "26900 tensor(5.3000, device='cuda:0', grad_fn=)\n", - "267000\n", - "27000 tensor(5.2141, device='cuda:0', grad_fn=)\n", - "268000\n", - "27100 tensor(5.2490, device='cuda:0', grad_fn=)\n", - "269000\n", - "27200 tensor(5.2850, device='cuda:0', grad_fn=)\n", - "270000\n", - "27300 tensor(5.4811, device='cuda:0', grad_fn=)\n", - "271000\n", - "27400 tensor(5.3561, device='cuda:0', grad_fn=)\n", - "272000\n", - "27500 tensor(5.2602, device='cuda:0', grad_fn=)\n", - "273000\n", - "27600 tensor(5.5429, device='cuda:0', grad_fn=)\n", - "274000\n", - "27700 tensor(5.3794, device='cuda:0', grad_fn=)\n", - "275000\n", - "27800 tensor(5.3792, device='cuda:0', grad_fn=)\n", - "276000\n", - "27900 tensor(5.4873, device='cuda:0', grad_fn=)\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "277000\n", - "28000 tensor(5.3454, device='cuda:0', grad_fn=)\n", - "278000\n", - "28100 tensor(5.3113, device='cuda:0', grad_fn=)\n", - "279000\n", - "28200 tensor(5.4785, device='cuda:0', grad_fn=)\n", - "280000\n", - "28300 tensor(5.1013, device='cuda:0', grad_fn=)\n", - "281000\n", - "28400 tensor(5.5403, device='cuda:0', grad_fn=)\n", - "282000\n", - "28500 tensor(5.3676, device='cuda:0', grad_fn=)\n", - "283000\n", - "28600 tensor(5.3108, device='cuda:0', grad_fn=)\n", - "284000\n", - "28700 tensor(5.4403, device='cuda:0', grad_fn=)\n", - "285000\n", - "28800 tensor(5.4926, device='cuda:0', grad_fn=)\n", - "286000\n", - "28900 tensor(5.3638, device='cuda:0', grad_fn=)\n", - "287000\n", - "29000 tensor(5.2819, device='cuda:0', grad_fn=)\n", - "288000\n", - "29100 tensor(5.0362, device='cuda:0', grad_fn=)\n", - "289000\n", - "29200 tensor(5.1871, device='cuda:0', grad_fn=)\n", - "290000\n", - "29300 tensor(5.4697, device='cuda:0', grad_fn=)\n", - "291000\n", - "29400 tensor(5.5909, device='cuda:0', grad_fn=)\n", - "292000\n", - "29500 tensor(5.3807, device='cuda:0', grad_fn=)\n", - "293000\n", - "29600 tensor(5.2398, device='cuda:0', grad_fn=)\n", - "294000\n", - "29700 tensor(5.3690, device='cuda:0', grad_fn=)\n", - "295000\n", - "29800 tensor(5.2220, device='cuda:0', grad_fn=)\n", - "296000\n", - "29900 tensor(5.4597, device='cuda:0', grad_fn=)\n", - "297000\n", - "30000 tensor(5.2205, device='cuda:0', grad_fn=)\n", - "298000\n", - "30100 tensor(5.3061, device='cuda:0', grad_fn=)\n", - "299000\n", - "30200 tensor(5.2432, device='cuda:0', grad_fn=)\n", - "300000\n", - "30300 tensor(5.3527, device='cuda:0', grad_fn=)\n", - "301000\n", - "30400 tensor(5.1823, device='cuda:0', grad_fn=)\n", - "302000\n", - "30500 tensor(5.3526, device='cuda:0', grad_fn=)\n", - "30600 tensor(5.3318, device='cuda:0', grad_fn=)\n", - "303000\n", - "30700 tensor(5.3634, device='cuda:0', grad_fn=)\n", - "304000\n", - "30800 tensor(5.3571, device='cuda:0', grad_fn=)\n", - "305000\n", - "30900 tensor(5.3875, device='cuda:0', grad_fn=)\n", - "306000\n", - "31000 tensor(5.4983, device='cuda:0', grad_fn=)\n", - "307000\n", - "31100 tensor(5.1554, device='cuda:0', grad_fn=)\n", - "308000\n", - "31200 tensor(5.1952, device='cuda:0', grad_fn=)\n", - "309000\n", - "31300 tensor(5.4546, device='cuda:0', grad_fn=)\n", - "310000\n", - "31400 tensor(5.2307, device='cuda:0', grad_fn=)\n", - "311000\n", - "31500 tensor(5.4188, device='cuda:0', grad_fn=)\n", - "312000\n", - "31600 tensor(5.4085, device='cuda:0', grad_fn=)\n", - "313000\n", - "31700 tensor(5.3744, device='cuda:0', grad_fn=)\n", - "314000\n", - "31800 tensor(5.4766, device='cuda:0', grad_fn=)\n", - "315000\n", - "31900 tensor(5.1062, device='cuda:0', grad_fn=)\n", - "316000\n", - "32000 tensor(5.2924, device='cuda:0', grad_fn=)\n", - "317000\n", - "32100 tensor(5.1728, device='cuda:0', grad_fn=)\n", - "318000\n", - "32200 tensor(5.4863, device='cuda:0', grad_fn=)\n", - "319000\n", - "32300 tensor(5.4748, device='cuda:0', grad_fn=)\n", - "320000\n", - "32400 tensor(5.4518, device='cuda:0', grad_fn=)\n", - "321000\n", - "32500 tensor(5.2752, device='cuda:0', grad_fn=)\n", - "322000\n", - "32600 tensor(5.3822, device='cuda:0', grad_fn=)\n", - "323000\n", - "32700 tensor(5.3088, device='cuda:0', grad_fn=)\n", - "324000\n", - "32800 tensor(5.5403, device='cuda:0', grad_fn=)\n", - "325000\n", - "32900 tensor(5.4000, device='cuda:0', grad_fn=)\n", - "326000\n", - "33000 tensor(5.1837, device='cuda:0', grad_fn=)\n", - "327000\n", - "33100 tensor(5.3888, device='cuda:0', grad_fn=)\n", - "328000\n", - "33200 tensor(5.4849, device='cuda:0', grad_fn=)\n", - "329000\n", - "33300 tensor(5.2471, device='cuda:0', grad_fn=)\n", - "330000\n", - "33400 tensor(5.5246, device='cuda:0', grad_fn=)\n", - "331000\n", - "33500 tensor(5.3479, device='cuda:0', grad_fn=)\n", - "332000\n", - "33600 tensor(5.3043, device='cuda:0', grad_fn=)\n", - "333000\n", - "33700 tensor(5.3487, device='cuda:0', grad_fn=)\n", - "334000\n", - "33800 tensor(5.4368, device='cuda:0', grad_fn=)\n", - "335000\n", - "33900 tensor(5.1620, device='cuda:0', grad_fn=)\n", - "336000\n", - "34000 tensor(5.3873, device='cuda:0', grad_fn=)\n", - "337000\n", - "34100 tensor(5.3545, device='cuda:0', grad_fn=)\n", - "338000\n", - "34200 tensor(5.4001, device='cuda:0', grad_fn=)\n", - "339000\n", - "34300 tensor(5.1902, device='cuda:0', grad_fn=)\n", - "340000\n", - "34400 tensor(5.6453, device='cuda:0', grad_fn=)\n", - "341000\n", - "34500 tensor(5.5124, device='cuda:0', grad_fn=)\n", - "342000\n", - "34600 tensor(5.4069, device='cuda:0', grad_fn=)\n", - "343000\n", - "34700 tensor(5.4734, device='cuda:0', grad_fn=)\n", - "344000\n", - "34800 tensor(5.5014, device='cuda:0', grad_fn=)\n", - "345000\n", - "34900 tensor(5.5412, device='cuda:0', grad_fn=)\n", - "346000\n", - "35000 tensor(5.5132, device='cuda:0', grad_fn=)\n", - "347000\n", - "35100 tensor(5.3455, device='cuda:0', grad_fn=)\n", - "348000\n", - "35200 tensor(5.2694, device='cuda:0', grad_fn=)\n", - "349000\n", - "35300 tensor(5.4988, device='cuda:0', grad_fn=)\n", - "350000\n", - "35400 tensor(5.1485, device='cuda:0', grad_fn=)\n", - "351000\n", - "35500 tensor(5.2299, device='cuda:0', grad_fn=)\n", - "352000\n", - "35600 tensor(5.3643, device='cuda:0', grad_fn=)\n", - "353000\n", - "35700 tensor(5.2247, device='cuda:0', grad_fn=)\n", - "354000\n", - "35800 tensor(5.3615, device='cuda:0', grad_fn=)\n", - "355000\n", - "35900 tensor(5.3453, device='cuda:0', grad_fn=)\n", - "356000\n", - "36000 tensor(5.1217, device='cuda:0', grad_fn=)\n", - "357000\n", - "36100 tensor(5.4909, device='cuda:0', grad_fn=)\n", - "358000\n", - "36200 tensor(5.3382, device='cuda:0', grad_fn=)\n", - "359000\n", - "36300 tensor(5.6225, device='cuda:0', grad_fn=)\n", - "360000\n", - "36400 tensor(5.3167, device='cuda:0', grad_fn=)\n", - "361000\n", - "36500 tensor(5.3458, device='cuda:0', grad_fn=)\n", - "362000\n", - "36600 tensor(5.3608, device='cuda:0', grad_fn=)\n", - "363000\n", - "36700 tensor(5.1660, device='cuda:0', grad_fn=)\n", - "364000\n", - "36800 tensor(5.2737, device='cuda:0', grad_fn=)\n", - "365000\n", - "36900 tensor(5.3883, device='cuda:0', grad_fn=)\n", - "366000\n", - "37000 tensor(5.2783, device='cuda:0', grad_fn=)\n", - "367000\n", - "37100 tensor(5.3110, device='cuda:0', grad_fn=)\n", - "368000\n", - "37200 tensor(5.3794, device='cuda:0', grad_fn=)\n", - "369000\n", - "37300 tensor(5.2802, device='cuda:0', grad_fn=)\n", - "370000\n", - "37400 tensor(5.6133, device='cuda:0', grad_fn=)\n", - "371000\n", - "37500 tensor(5.3138, device='cuda:0', grad_fn=)\n", - "372000\n", - "37600 tensor(5.3083, device='cuda:0', grad_fn=)\n", - "373000\n", - "37700 tensor(5.4860, device='cuda:0', grad_fn=)\n", - "374000\n", - "37800 tensor(5.3216, device='cuda:0', grad_fn=)\n", - "375000\n", - "37900 tensor(5.2969, device='cuda:0', grad_fn=)\n", - "376000\n", - "38000 tensor(5.3759, device='cuda:0', grad_fn=)\n", - "377000\n", - "38100 tensor(5.3914, device='cuda:0', grad_fn=)\n", - "378000\n", - "38200 tensor(5.4089, device='cuda:0', grad_fn=)\n", - "379000\n", - "38300 tensor(5.3068, device='cuda:0', grad_fn=)\n", - "380000\n", - "38400 tensor(5.3798, device='cuda:0', grad_fn=)\n", - "381000\n", - "38500 tensor(5.4051, device='cuda:0', grad_fn=)\n", - "382000\n", - "38600 tensor(5.3471, device='cuda:0', grad_fn=)\n", - "383000\n", - "38700 tensor(5.3415, device='cuda:0', grad_fn=)\n", - "384000\n", - "38800 tensor(5.4310, device='cuda:0', grad_fn=)\n", - "385000\n", - "38900 tensor(5.5029, device='cuda:0', grad_fn=)\n", - "386000\n", - "39000 tensor(5.2021, device='cuda:0', grad_fn=)\n", - "387000\n", - "39100 tensor(5.4283, device='cuda:0', grad_fn=)\n", - "388000\n", - "39200 tensor(5.5158, device='cuda:0', grad_fn=)\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "389000\n", - "39300 tensor(5.3452, device='cuda:0', grad_fn=)\n", - "390000\n", - "39400 tensor(5.4111, device='cuda:0', grad_fn=)\n", - "391000\n", - "39500 tensor(5.4969, device='cuda:0', grad_fn=)\n", - "392000\n", - "39600 tensor(5.1952, device='cuda:0', grad_fn=)\n", - "393000\n", - "39700 tensor(5.1946, device='cuda:0', grad_fn=)\n", - "394000\n", - "39800 tensor(5.3234, device='cuda:0', grad_fn=)\n", - "395000\n", - "39900 tensor(5.1354, device='cuda:0', grad_fn=)\n", - "396000\n", - "40000 tensor(5.2210, device='cuda:0', grad_fn=)\n", - "397000\n", - "40100 tensor(5.3133, device='cuda:0', grad_fn=)\n", - "398000\n", - "40200 tensor(5.2990, device='cuda:0', grad_fn=)\n", - "399000\n", - "40300 tensor(5.3684, device='cuda:0', grad_fn=)\n", - "40400 tensor(5.3700, device='cuda:0', grad_fn=)\n", - "400000\n", - "40500 tensor(5.2911, device='cuda:0', grad_fn=)\n", - "401000\n", - "40600 tensor(5.3497, device='cuda:0', grad_fn=)\n", - "402000\n", - "40700 tensor(5.3981, device='cuda:0', grad_fn=)\n", - "403000\n", - "40800 tensor(5.3436, device='cuda:0', grad_fn=)\n", - "404000\n", - "40900 tensor(5.2978, device='cuda:0', grad_fn=)\n", - "405000\n", - "41000 tensor(5.3420, device='cuda:0', grad_fn=)\n", - "406000\n", - "41100 tensor(5.3342, device='cuda:0', grad_fn=)\n", - "407000\n", - "41200 tensor(5.2226, device='cuda:0', grad_fn=)\n", - "408000\n", - "41300 tensor(5.3573, device='cuda:0', grad_fn=)\n", - "409000\n", - "41400 tensor(5.2448, device='cuda:0', grad_fn=)\n", - "410000\n", - "41500 tensor(5.3863, device='cuda:0', grad_fn=)\n", - "411000\n", - "41600 tensor(5.3051, device='cuda:0', grad_fn=)\n", - "412000\n", - "41700 tensor(5.3294, device='cuda:0', grad_fn=)\n", - "413000\n", - "41800 tensor(5.3191, device='cuda:0', grad_fn=)\n", - "414000\n", - "41900 tensor(5.3289, device='cuda:0', grad_fn=)\n", - "415000\n", - "42000 tensor(5.4860, device='cuda:0', grad_fn=)\n", - "416000\n", - "42100 tensor(5.2358, device='cuda:0', grad_fn=)\n", - "417000\n", - "42200 tensor(5.3253, device='cuda:0', grad_fn=)\n", - "418000\n", - "42300 tensor(5.3869, device='cuda:0', grad_fn=)\n", - "419000\n", - "42400 tensor(5.2062, device='cuda:0', grad_fn=)\n", - "420000\n", - "42500 tensor(5.3712, device='cuda:0', grad_fn=)\n", - "421000\n", - "42600 tensor(5.1718, device='cuda:0', grad_fn=)\n", - "422000\n", - "42700 tensor(5.4735, device='cuda:0', grad_fn=)\n", - "423000\n", - "42800 tensor(5.3973, device='cuda:0', grad_fn=)\n", - "424000\n", - "42900 tensor(5.2447, device='cuda:0', grad_fn=)\n", - "425000\n", - "43000 tensor(5.3896, device='cuda:0', grad_fn=)\n", - "426000\n", - "43100 tensor(5.3916, device='cuda:0', grad_fn=)\n", - "427000\n", - "43200 tensor(5.2044, device='cuda:0', grad_fn=)\n", - "428000\n", - "43300 tensor(5.2167, device='cuda:0', grad_fn=)\n", - "429000\n", - "43400 tensor(5.3933, device='cuda:0', grad_fn=)\n", - "430000\n", - "43500 tensor(5.1078, device='cuda:0', grad_fn=)\n", - "431000\n", - "43600 tensor(5.3045, device='cuda:0', grad_fn=)\n", - "432000\n" - ] - } - ], - "source": [ - "data = DataLoader(train_dataset, batch_size=batch_s)\n", - "optimizer = torch.optim.Adam(model.parameters(), lr=learning_rate)\n", - "criterion = torch.nn.NLLLoss()\n", - "torch.cuda.empty_cache()\n", - "gc.collect()\n", - "\n", - "model.load_state_dict(torch.load('model-bigram_final.bin'))\n", - "for i in range(1, epochs+1):\n", - " print('epoch: =', i)\n", - " model.train()\n", - " step = 0\n", - " for x, y in data: # prev, predicting, following words\n", - " x = x.to(device)\n", - " y = y.to(device)\n", - " optimizer.zero_grad()\n", - " ypredicted = model(x) #previous, following word\n", - " loss = criterion(torch.log(ypredicted), y)\n", - " if step % 100 == 0:\n", - " print(step, loss)\n", - " step += 1\n", - " loss.backward()\n", - " optimizer.step()\n", - " torch.save(model.state_dict(), f'model-bigram_2nd-run{i}.bin') \n", - "torch.save(model.state_dict(), f'model-bigram_final.bin') " - ] - }, - { - "cell_type": "code", - "execution_count": 17, - "metadata": {}, - "outputs": [ - { - "data": { - "text/plain": [ - "[('be', 11, 0.2570849657058716),\n", - " ('', 0, 0.07411641627550125),\n", - " ('not', 22, 0.05940083786845207),\n", - " ('have', 28, 0.02751326560974121),\n", - " ('bo', 167, 0.014936885796487331),\n", - " ('make', 116, 0.013943656347692013),\n", - " ('give', 193, 0.011286991648375988),\n", - " ('take', 153, 0.011171611957252026),\n", - " ('do', 86, 0.010088067501783371),\n", - " ('he', 20, 0.009703895077109337)]" - ] - }, - "execution_count": 17, - "metadata": {}, - "output_type": "execute_result" - } - ], - "source": [ - "device = 'cuda'\n", - "torch.cuda.empty_cache()\n", - "model = SimpleBigramNeuralLanguageModel(vocab_size, embed_size).to(device)\n", - "model.load_state_dict(torch.load(f'model-bigram_final.bin'))\n", - "model.eval()\n", - "\n", - "ixs = torch.tensor(vocab.forward(['will'])).to(device)\n", - "\n", - "out = model(ixs)\n", - "top = torch.topk(out[0], 10)\n", - "top_indices = top.indices.tolist()\n", - "top_probs = top.values.tolist()\n", - "top_words = vocab.lookup_tokens(top_indices)\n", - "list(zip(top_words, top_indices, top_probs))" - ] - }, - { - "cell_type": "code", - "execution_count": 34, - "metadata": {}, - "outputs": [ - { - "data": { - "text/plain": [ - "[('', 0, 0.19996878504753113),\n", - " ('and', 3, 0.05288130044937134),\n", - " ('of', 2, 0.042051784694194794),\n", - " ('the', 1, 0.026572922244668007),\n", - " ('to', 4, 0.022689413279294968),\n", - " ('in', 6, 0.015904497355222702),\n", - " ('The', 17, 0.012827681377530098),\n", - " ('a', 5, 0.00961760152131319),\n", - " ('for', 8, 0.008938422426581383),\n", - " ('', 32, 0.00840282253921032)]" - ] - }, - "execution_count": 34, - "metadata": {}, - "output_type": "execute_result" - } - ], - "source": [ - "vocab = train_dataset.vocab\n", - "ixs = torch.tensor(vocab.forward(['cerned.'])).to(device)\n", - "\n", - "out = model(ixs)\n", - "top = torch.topk(out[0], 10)\n", - "top_indices = top.indices.tolist()\n", - "top_probs = top.values.tolist()\n", - "top_words = vocab.lookup_tokens(top_indices)\n", - "list(zip(top_words, top_indices, top_probs))" - ] - }, - { - "cell_type": "code", - "execution_count": 33, - "metadata": {}, - "outputs": [ - { - "data": { - "text/plain": [ - "[('', 0, 1.0),\n", - " ('particular,', 14538, 0.24527804553508759),\n", - " ('revolution.', 20446, 0.23776617646217346),\n", - " ('Territory.', 14189, 0.23417341709136963),\n", - " ('or-', 2261, 0.22888363897800446),\n", - " ('3', 479, 0.2288265973329544),\n", - " ('speak.', 13722, 0.2252315878868103),\n", - " ('attend.', 19397, 0.22110989689826965),\n", - " ('say,', 1455, 0.22106117010116577),\n", - " ('Lee.', 15326, 0.21764159202575684)]" - ] - }, - "execution_count": 33, - "metadata": {}, - "output_type": "execute_result" - } - ], - "source": [ - "cos = nn.CosineSimilarity(dim=1, eps=1e-6)\n", - "\n", - "embeddings = model.model[0].weight\n", - "\n", - "vec = embeddings[vocab['cerned.']]\n", - "\n", - "similarities = cos(vec, embeddings)\n", - "\n", - "top = torch.topk(similarities, 10)\n", - "\n", - "top_indices = top.indices.tolist()\n", - "top_probs = top.values.tolist()\n", - "top_words = vocab.lookup_tokens(top_indices)\n", - "list(zip(top_words, top_indices, top_probs))" - ] - }, - { - "cell_type": "code", - "execution_count": null, - "metadata": { - "scrolled": true - }, - "outputs": [], - "source": [ - "\n", - "vocab = train_dataset.vocab\n", - "# ixs = torch.tensor(vocab.forward(['a'])).to(device)\n", - "ixs = torch.tensor(vocab.forward(['of'])).to(device)\n", - "# ixs = torch.tensor(vocab.forward(['that'])).to(device)\n", - "# ixs = torch.tensor(vocab.forward(['church'])).to(device)\n", - "# ixs = torch.tensor(vocab.forward(['wait'])).to(device)\n", - "\n", - "out = model(ixs)\n", - "top = torch.topk(out[0], 10)\n", - "top_indices = top.indices.tolist()\n", - "top_probs = top.values.tolist()\n", - "top_words = vocab.lookup_tokens(top_indices)\n", - "list(zip(top_words, top_indices, top_probs))" - ] - }, - { - "cell_type": "code", - "execution_count": 52, - "metadata": {}, - "outputs": [], - "source": [ - "def get_values_from_model(presc_word, model, vocab, k):\n", - " ixs = torch.tensor(vocab.forward([presc_word])).to(device)\n", - " out = model(ixs)\n", - " top = torch.topk(out[0], k)\n", - " top_indices = top.indices.tolist()\n", - " top_probs = top.values.tolist()\n", - " top_words = vocab.lookup_tokens(top_indices)\n", - " return list(zip(top_words, top_probs))\n", - "\n", - "def gonito_format(dic):\n", - " tab = summarize_probs_unk(dic)\n", - " result = ''\n", - " for element in tab[:-1]:\n", - " result+=str(element[0])+':'+str(element[1])+'\\t'\n", - " result+=':'+ str(tab[-1][1])+'\\n'\n", - " return result\n", - "\n", - "def summarize_probs_unk(dic):\n", - " if '' in dic.keys():\n", - " del dic['']\n", - " probsum = sum(float(val) for key, val in dic.items())\n", - "# if \"\" in dic.keys():\n", - "# for key in dic:\n", - "# dic[key] = dic[key]/probsum #normalize to 1\n", - "# wildcard = dic[''] \n", - "# else: #no unk in first 10 entries\n", - " for key in dic:\n", - " dic[key] = dic[key]/probsum*(1-wildcard_minweight) ###leave some space for wildcard\n", - " wildcard = wildcard_minweight\n", - " tab = [(key, val) for key, val in dic.items()]\n", - " tab.append(('', wildcard))\n", - " return tab\n", - "\n" - ] - }, - { - "cell_type": "code", - "execution_count": 54, - "metadata": {}, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Present,\n", - "maintenance.\n", - "-\n", - "on\n", - "after.\n", - "included.\n", - "Delaware.\n", - "body.\n", - "route.\n", - "under\n", - "bership.\n", - "call*\n", - "waterscape.\n", - "?\n", - "soles.\n", - "simple.\"\n", - "public.\n", - "them.-\n", - "months.\n", - "icy.\"\n", - "E\n", - "level.\n", - "4t—36\n", - "financial\n", - "steady.\n", - "mlnicars.\n", - "his\n", - "circumstances.\n", - "corner.\n", - "science.\n", - "insurance.\"\n", - "fair.\n", - "scalded.\n", - "ria.\n", - "it,\n", - "mlngtonlan.\n", - "possession.\n", - "of\n", - "pray!”\n", - "occurred.\n", - "euro,\n", - "way.\n", - "185.\n", - "destruction.\n", - "conditions.\n", - "Hood.\n", - "s.\n", - "case.\n", - "Roy\n", - "country.\n", - "ccjulinalf\n", - "side.\n", - "against\n", - "servation.\n", - "Detroit.\n", - "North\n", - "country.\n", - "\"\n", - "bones.\n", - "sold.\n", - "Brown\n", - "bill.\n", - "Springer,\n", - "gaged.\n", - "their\n", - "character.\n", - "cu\n", - "former\n", - "cities.\n", - "accumulated\n", - "corporal.\n", - "^\n", - "-\n", - "o\n", - ".(.\n", - "descendants\n", - "notes.\n", - "Courier.\n", - "town\n", - "E\n", - "awake,\n", - "d\n", - "circumstances.\n", - "board.\n", - "-\n", - "union.\n", - "the\n", - "en-\n", - "children\n", - "Tre--\n", - "date.\n", - "Monday.\n", - "been\n", - "now\n", - "slaughter.”\n", - "art.\n", - "shape.\n", - "ai\n", - "Magazine.\n", - "License.\n", - "K\n", - "P.iU--\n", - "failures\n", - "v\n", - "cause.\n", - "Russian\n", - "prayed.\n", - "tariff.\n", - "action.\n", - "direction.\n", - "Tacomans\n", - "to-\n", - "cellence.\n", - "power\n", - "25\n", - "alty.\n", - "July.\"\n", - "crime.\n", - "of\n", - "conviction.\n", - "been\n", - "d\n", - "unnatural\n", - "s,\n", - "-\n", - "scccdo.\n", - "e-\n", - "Harrisburg.\n", - "cash.\n", - "Peter\n", - "OTTirn\n", - "lor\n", - "MacAuley.\n", - "O\n", - "style.\n", - "road.\n", - "YOU.\n", - "nations.\"\n", - "r.o\n", - "made.\n", - "Democratic.\n", - "......\n", - "Ocean.\n", - "r\n", - "and\n", - "bless\n", - "jaSl\n", - "Krapt.\n", - "states.\n", - "It.\n", - "river.\n", - "fag-\n", - "ShorteU.\n", - "fabric.\"\n", - "returned.\n", - "tra.\n", - "member.\n", - "both.\n", - "I\n", - "place.\"\n", - "our\n", - ".»\n", - "17\n", - "1911.\n", - "an\n", - "Age.\n", - "It\n", - "Gland--\n", - "tl\n", - "-\n", - "be-\n", - "air.\n", - "crude.\n", - "remedy.\n", - ":\n", - "Uny\n", - "tunnel.\n", - "n\n", - "Gulf.”\n", - "war.\n", - "...\n", - "contest\n", - "History.\n", - "aear.\n", - "tees.\n", - "residence.\n", - "friend,\n", - "h\n", - "\"\n", - "him.\n", - "government.\n", - "facilities.\"\n", - "remedy.\n", - "r\n", - "eent\n", - "law.\n", - "the\n", - "Representative\n", - "w\n", - "n\n", - "1S44.\n", - "adopted.\n", - "favor.\n", - "Sciurier.\n", - "bein\n", - "market.\n", - ".1\n", - "marshes.\n", - "Newark.\n", - "years\n", - "Preeton.\"\n", - "lti-\n", - "Trumpet\n", - "said.\n", - "appoint\n", - "oi\n", - "tbe\n", - "Federahi,\n", - "1858.\n", - "hereof.\n", - "treatment.\n", - "dawdling\n", - "dence.\n", - "L\n", - "requirements.\n", - "in\n", - "e\n", - "projects?\n", - "goes:\n", - "W\n", - "w\n", - "thereof.\n", - "It\n", - "injury.\"\n", - "water.\"\n", - "played\n", - "famoti'\n", - "nnd\n", - "there.'\n", - "win.\"\n", - "used.\n", - ".by\n", - "'\n", - "Woman.\n", - "viz:\n", - "causing\n", - "quarter.\n", - "i.\n", - "police.\n", - "run.\n", - "children.'\n", - "debate.\n", - "Blannerhasset\n", - "pa-\n", - "with\n", - "elbow.\n", - "J\n", - "aid.\n", - "States.\n", - "graii\n", - "terday.\n", - "was\n", - "mlngo\n", - "night.\n", - "reason.\n", - "Thlinget.\n", - "111.\n", - "'\n", - "anarled.\n", - "personally.\n", - "man.\n", - "nonresidents\n", - "man.\n", - "sently.\n", - "st\n", - "time.\n", - "shrine.\n", - "communicable\n", - "anf\n", - "Th.-\n", - "lose\n", - "thereon.\n", - "application.\n", - "disappear.\n", - "forfeited.\n", - "collis-\n", - "sirl\n", - "(\n", - "reconstructed.\n", - "(ban.)\n", - "the\n", - "tion.\n", - "afte\n", - "co-urts.\"\n", - "18W2.\n", - "attuck.\n", - "too\n", - "void.\n", - "presented,\n", - "shallton.\n", - "pnst\n", - "here.\n", - "mentioned.\n", - "session.\n", - "b\n", - "s\n", - "river.\n", - "capac\n", - "firm.\n", - "bombardment,\n", - "other.\n", - "worship.\n", - "The\n", - "liberties.\n", - "respousihiliiics,\n", - "Times.\n", - "other.\n", - "following:\n", - "consid­\n", - "arroya.\n", - "owno-\n", - "is\n", - "14.\n", - "name.\n", - "night\n", - "adjourned.\n", - "morning.^\n", - "paint.\n", - "post-nat-\n", - "union.\n", - "or-\n", - "Jiuili\n", - "said.\n", - "splurge.\n", - "him.\n", - "work.\"\n", - "-te-\n", - "tfo\n", - "there\n", - ".\n", - "rising.\n", - "Gen.\n", - "11\n", - "yours,\n", - "Farmer.\n", - "acquaintanceship,\n", - "hand.\n", - "high-cla-\n", - "nights.\n", - "AfcC-i-\n", - "elections.\n", - "way.\"\n", - "once.\n", - ".\n", - "comfortable.\n", - "rality.\n", - ".be-\n", - "approval,\n", - "sto-\n", - "cure.\n", - "27,000.\n", - "spoils.”\n", - "deeds.\n", - "now\n", - "1.\n", - "da\n", - "...11.11-\n", - "d\n", - "berry.\n", - "congratulations.\n", - "coast\n", - "Wednesday\n", - "ferences.\n", - "point.\"\n", - "preservation.\n", - "gowns.\n", - "curanhalted.\n", - "JBarmsT.\n", - "surroundings.\n", - "treat\n", - "ted.\n", - "tfrith\n", - "blossoms.\n", - "here.\"\n", - "Minnesota.\n", - "rope.\n", - "—\n", - "him:\n", - "*\n", - "is”\n", - "feet\n", - "music-\n", - "Newark.\n", - "given\n", - "country\n", - "sick.\n", - "black.\n", - "so.\n", - "archdiocese.\n", - "form.\"\n", - "k\n", - "propose.\n", - "ed.\n", - ".\n", - "healthy.\n", - "cannons.\n", - "auxiliaries\n", - "by\n", - "to\n", - "shoute,\n", - "thel.-\n", - "\"Short\n", - "side.\n", - "Koine.\n", - "thanks.\n", - "le.\n", - "communicable.\n", - "[Sigued]\n", - "ion.\n", - "remedy.\n", - "Monthly.\n", - "enterprises.\"\n", - "nlieht**.i\n", - "harmed.\"\n", - "retreat.\n", - "what\n", - "d\n", - "over\n", - "wounded.\n", - "-\n", - ".\n", - "mobiles.\"\n", - "mill.\n", - "college.\n", - "B.\n", - "Manassas.\n", - "fea-\n", - "fatigue.\n", - "\"Quartermaster\n", - "Williams.'\n", - "visible.\n", - "Cox,\n", - "bo\n", - "stitution.\n", - "said\n", - "c-\n", - "be\n", - "hat.\n", - "Sing.\n", - "penalty.\n", - "at\n", - "prison!\n", - "salt.\n", - "consider.\"\n", - "mo\n", - "TWENTY-EIGHT-\n", - "Hazen's\n", - "F\n", - "Pennsylvania\n", - "mand.\n", - "Class.”\n", - "years.\n", - "shore.\n", - "posterity.\n", - "1000.\n", - "tern-\n", - "meeting.\n", - "logical.\n", - "said:\n", - "stamps.\n", - "trol.\"\n", - "misdeeds.\n", - "Middletown.\n", - "nnd\n", - "law.\n", - "sure\n", - "recover.\n", - "e\n", - "move.\n", - "remedy.\n", - "days.\n", - "efficiency.\n", - "provement.\n", - "be\n", - "i\n", - "fall.\n", - "East\n", - "really\n", - "-\n", - "uever\n", - "leg-\n", - "deny\n", - "reason\n", - "ous\n", - "of\n", - "agencies.\n", - "eloped.\"\n", - "scale.\n", - "Colonel\n", - "company.\n", - "watera\n", - "|\n", - "enterprises,\n", - "to-day.\n", - "In\n", - "enuffforme.\n", - "justly\n", - "finitum.\"\n", - "negan.\n", - "OTTO\n", - "interests.\n", - "mo-\n", - "determined.\n", - "iasu\n", - "void.\n", - "nights\n", - "come.\"\n", - ".\n", - "pild.\n", - "(\n", - "her\n", - "from\n", - "Russians.\n", - "things.\n", - "talk.\n", - "constituted,\n", - "lows:\n", - "dustry.\n", - "state.\n", - "sections.\n", - "cent..\n", - "cm'\n", - "mile*.\n", - "liberty,\n", - "judgment.\n", - "a-t\n", - "illusion.\n", - "With\n", - "States.\n", - "and\n", - "kind.\n", - "preference\n", - "lunch.\n", - "tary.\n", - "”\n", - "the\n", - "Democratic\n", - "»ball\n", - "ribbon.\n", - "and\n", - "of\n", - ".\n", - "route.\n", - "prayed.\n", - "The\n", - "live.\n", - "train.\n", - "holi-\n", - "follow.\n", - "a\n", - "quail.”\n", - "uuu\n", - "oath.\n", - "Cohce.\n", - "',\n", - "record.\n", - "it.'\n", - "the\n", - "have\n", - "Kinley.\n", - "purchase.\n", - "violcuce.\n", - "gates-wor-e\n", - "placed\n", - "taste.\n", - "at-\n", - "servation.\n", - "\"\n", - "views.\n", - "one.\n", - "description.\n", - "lized\n", - "egress.\n", - "flames.\n", - "e\n", - "704c.\n", - "height.\n", - "office.\n", - "Pa.,\n", - "preserves\n", - "voice.\n", - "Ledger.\n", - "later.\n", - "eminent\n", - "powerfully.\"\n", - "d\n", - "g\n", - "Iremain\n", - "mi-\n", - "Constitu-\n", - "charges.\n", - "llnllrond\n", - "llWi\n", - "n,\n", - "wel-\n", - "eternity.\n", - "of\n", - "it.\n", - "on-\n", - "n-\n", - "sure.\"\n", - "n-\n", - "feform.\n", - "it.\n", - "man.\n", - "popularity.\n", - "PUGH.\n", - "dollars.\n", - "w-elfare.\n", - "motionless.\n", - "ago\n", - "call.\n", - "emergencies.\n", - "45.\n", - "again.\n", - "to\n", - "Smith,\n", - "a\n", - "h\n", - "Client\n", - "for.\n", - "gone.\n", - "thereof.\n", - "per-\n", - "herlivin?.\n", - "d\n", - "Th-\n", - "them,\n", - "nbllWiM\n", - "Measures,\n", - "dreary,\n", - "crime.\"\n", - "^ffsouJ.\n", - "obser-\n", - "home.\n", - "him.\n", - "bcr.\n", - "C.\n", - "were\n", - "purchaser.\n", - "the\n", - "tu-\n", - "policy.\n", - "Dalty!\"\n", - "enl-\n", - "win.\"\n", - "dissolution.\n", - "health.\n", - "exception.\n", - "16'h.\n", - "1859.\n", - "boat.\n", - "'\n", - "home.\n", - "eausees.\n", - "Congress.\"\n", - "few.\n", - "’'\n", - "cost:\n", - "demand.\n", - "them.\n", - "Vlr\n", - "at\n", - "Saint\n", - "which\n", - "all.\n", - "consider.\n", - "list.\n", - "on\n", - "menu.\n", - "V.rk.\n", - "onstrate.\n", - "nor\n", - "as\n", - "county.\"\n", - "Iowu,\n", - "grade.\n", - "organization.\n", - "Wyman.\n", - "Unsportsmanlike\n", - "pioneers.\n", - "spring.\n", - "appertaining.\n", - "porch.\n", - "James\n", - "we\n", - "\"\n", - "it.\n", - "hand.\"\n", - "farew-ell.\"\n", - "advantages.\n", - "all\n", - "-\n", - "secured\n", - "attention\n", - "water\n", - "igoV\n", - "State.\n", - "cattle\n", - "othor.\n", - ".more.\n", - "-\n", - ".\n", - ",\n", - "rest.\n", - "been,-\n", - "a»\n", - "rei\n", - "cases.\n", - "is.\n", - "d\n", - "aforesaid.\n", - "determine.\n", - "him.\n", - "comic,\n", - "Inquiry.\n", - "girls,\n", - "rations.\n", - "living.\n", - "delinquents.\n", - "performed.\n", - "or\n", - "Mr,\n", - "plaint.\n", - "ters\n", - "e\n", - "constructed\n", - "slaughterinj\n", - "cent.\n", - "Commerce.\n", - "trated.\n", - "Tfrnoney.\n", - "factions.\n", - "rejected.\n", - "s\n", - "brave.\"\n", - "statute.\n", - "flight\n", - "kbrsf\n", - "future.\n", - ":\n", - "oflloe.\n", - "’08.\"\n", - "dustries?\"\n", - "at\n", - "parly.\n", - "insured.\n", - "tluics.\n", - "court.\n", - "lows.”\n", - "woman.\n", - "7\n", - "complaint.\n", - "expected.\n", - "anthem\n", - "the\n", - "sundown.\n", - "same.\n", - ".\n", - "yen.\"\n", - "copy.\n", - "strong.\n", - "Government,\n", - "nlng.\n", - "prloes,\n", - "Astoria.\n", - "?\n", - "They\n", - "Figaro.\n", - "buy-\n", - "it.\n", - "plane.\n", - "Nelson,\n", - "decision.\n", - "injunction,\n", - "peace.\n", - "wid\n", - "give\n", - "departments.\"\n", - "costs.\n", - "sufficient\n", - "DOCK\n", - "incurred.\n", - "out\n", - "j\n", - "g\n", - "his\n", - "tho-\n", - "mnlns.\n", - "rder.\n", - "trade.\n", - "-\n", - "Regis.\n", - "ages.\n", - "them.\n", - "Mcllvalne.\n", - "newspaper.\n", - "matter.\n", - "er\n", - "improved\n", - "be\n", - "pastures.\n", - "himself.\n", - ".\n", - "shapes.\n", - "adopted,\n", - "and\n", - ":\n", - "Robert\n", - "the\n", - "sheep.\n", - "lasttwelrty8®*'\n", - "Miss\n", - "~\n", - "Down\n", - "Clerk\n", - "circumstances.\n", - "tomorrow's\n", - "ncr*\n", - "would.\n", - "Union.\n", - "street.\n", - "tru.--\n", - "n\n", - "little\n", - "the\n", - "speed.\n", - "be\n", - "husband.,\n", - "obtained.\n", - "able.\n", - "ards.\"\n", - "situation.\n", - "accussed\n", - "sleep.\n", - "«1-\n", - "support.\n", - "home.\n", - "else.\n", - "spread\n", - "ploying.\n", - "and\n", - "the\n", - "Coun-\n", - "shot-makin-\n", - "guarded.\n", - "an\n", - "Hoursfrom9a.M..to5e.m.\n", - "Lake\n", - "1\n", - "rip-rapping.\n", - "publication.\n", - "Stanton's\n", - "Invigorated.\n", - "rendered.\n", - "strike,\n", - "cnn-\n", - "instruction.\n", - "territory.\n", - "ly.\n", - "the\n", - "ascertained.\n", - "service.\n", - "unrelenting.\n", - "York.\n", - "to-\n", - "pursuits.\n", - "for­\n", - "ever\n", - "plates.\n", - "Meade.\n", - "superior\n", - "Dusinesa.\n", - "ho-\n", - "locomotives.\n", - "derson.\n", - "like\n", - "Flag.\n", - "night.\n", - "hla\n", - "May.\n", - "r\n", - "Health.\n", - "*\n", - "term\n", - "provinces.\n", - "off.\n", - "transao-\n", - "ferine.\n", - "the\n", - "preserves\n", - "father\n", - "them.\n", - "hopeful.\n", - "\"\n", - "it.\n", - "jfciow.\n", - "stopped.\n", - "other,\n", - "sentiments.\n", - "action,\"\n", - "harlequinade.\n", - "water.\n", - "\"detective.\"\n", - "tutlon.\n", - "institutions.(12)\n", - "dreary.\n", - "was\n", - "say\n", - "maker.\n", - "nies.\n", - "vagabond.\n", - "here.\n", - "fault.\n", - "and\n", - "EU.\n", - ";\n", - "fur.\n", - "agreement.\n", - "11,012\n", - "provide.\n", - "Sovereign.\n", - "life.\n", - "law.\n", - "immeiatly\n", - "ity.\n", - "law.\n", - "I\n", - "faction.\n", - "nificance.\n", - "mac\n", - "7p.m.\n", - "James\n", - "individuals.\n", - "sure.\n", - ">n»v.\n", - "city.\n", - "brane.\n", - "relations.\n", - "if\n", - "Moka\n", - "Press.\n", - "during\n", - "relatives\n", - "colors.\n", - "n-\n", - "Arm.\n", - "it-\n", - "from\n", - "J'e\n", - "un-\n", - "er\n", - "2\"\n", - "vote?\n", - "disease.\n", - "feetlothepUceof\n", - "postponed.\n", - "and\n", - "JyTs\n", - "Vlmndcaler.\n", - "supervision.\n", - "PANY,\n", - "-\n", - "ceptance:\n", - "constituents.\n", - "e,\n", - "interest.\n", - "ult.\n", - "the\n", - "attention.\n", - "cans\n", - ".\n", - "s\n", - "pt\n", - "day.\n", - "p,\n", - "reply.\n", - "war.\n", - "purpose\n", - "ann,\n", - "Mass.\n", - ";\n", - "form­\n", - "o\n", - "d\n", - "t:\n", - "ou\n", - "basis.”\n", - "comforter.\n", - "of.\"\n", - "State.\n", - "search.\n", - "us.\n", - "cir-\n", - "Itnddmus,\n", - "was\n", - "invited.\n", - "damageB.\n", - "study.\n", - "cussion.\n", - "Afpeal-\n", - "orders.\n", - "the\n", - "organ-laa-\n", - "1\n", - "service.\n", - "damage.\n", - "zens.\n", - "Science.\n", - "superior.\n", - "directions.\n", - "known\n", - "unsuccessful.\n", - "'\n", - "village.\n", - "Uw\n", - "pay.\n", - "In\n", - "o’clock.\n", - "ruin.\n", - "mil.\n", - "sale\n", - "o\n", - "simply\n", - "the\n", - "Sun.\n", - "—,\n", - "statute.\n", - "each.\n", - "viciously.\n", - "Small's\n", - "Galveston\n", - "Gazelle\n", - "two\n", - "passed.\n", - "cent.\n", - "death.\n", - "ex-\n", - "proof\n", - "low-c-\n", - "arbitration.\n", - "times.\n", - "deans.\n", - "directed.\n", - "saloon.\n", - "The-\n", - "me.\n", - "home.\n", - "men.\n", - "Rogers.\n", - "oootomj\n", - "x.\n", - "from\n", - "fire.\n", - "plies.\n", - "I,.\n", - "ones.\n", - "location.\n", - "e.\n", - "an\n", - "sake.\n", - "s\n", - "shall\n", - "fire.\n", - "Herald,\n", - "soldbv\n", - "citizen.\n", - "ordinance.\n", - "1902,\n", - "whose\n", - "dry\"\n", - "election.\n", - "pack.\"\n", - "side.\n", - "course\n", - "and\n", - "the\n", - "b-\n", - "day,\n", - "e-\n", - "outlaws.\n", - "roots.\n", - "work.\n", - "physicians.\n", - "however.\n", - "Kulp,\n", - "teachings\n", - "!)ile,\n", - "10%<\n", - "cemetery.\n", - "name?’\n", - "course.\n", - "oiu.\n", - "clergy.\n", - "peace\n", - "to\n", - "saved.\n", - "government.\n", - "commonplaces.\"\n", - "week\n", - "meet.\n", - "1858.\n", - "Air«.\n", - "Mass.,\n", - "county.\n", - "associations.\n", - "up.\n", - "eouth.\n", - "a-\n", - "de-\n", - "Oroville.\n", - "o\n", - "tc\n", - "evening.\n", - "state.\n", - "oDservor.\n", - "witdom\n", - "circumstances.\n", - "snags\n", - "shown,\n", - "farm¬\n", - "way\n", - "hat\n", - "plause.)\n", - "on\n", - "Grahamhes.\n", - "opponent's\n", - "the\n", - "fying.\n", - "contractor.\n", - ".iV\"\n", - "and\n", - "his\n", - "1\n", - "brother-in-la-\n", - "engaged.\n", - "inspection.\n", - "Cemetery.\n", - "country\n", - "tale.\n", - "stops.\"\n", - "rocks.\n", - "medicine.\"\n", - "men.\n", - "Zipf.\n", - "Socle'y.\n", - "erased\n", - "Mountains.-\n", - "30.\n", - "aeiil\n", - "no\n", - "Opinion.\n", - "it.\n", - "have\n", - "Robbery.—\n", - "Brnc#.\n", - "discussions.\n", - "day?\n", - "not\n", - "ewYork.\n", - "measure.\n", - "and\n", - ".M\n", - "president.\n", - "prim\n", - "drops.\n", - "hell.\n", - "can.\n", - "de=\n", - "hum.\n", - "im-a-\n", - "aucceesluily\n", - "Milton\n", - "Southwestern.\n", - "dollars.\n", - "poles.\n", - "by\n", - "Mr.\n", - "babl\n", - "1862.\n", - "action.\n", - "order.\n", - "once.\n", - "yield.\n", - "in.\n", - "l'.\n", - "God.\n", - "capital.\n", - "missionaries\n", - "plastered.\n", - "caution.”\n", - "No.\n", - "matter.\n", - "licked.\")\n", - "body.\n", - "votes.\n", - "relation.\n", - "I\n", - "them.\n", - ".'s.\"\n", - "joint.\n", - "suffrage.\n", - "list.\n", - "i\n", - "i\n", - "strengtn.\n", - "slain.\n", - "•l\n", - "plaint.\n", - "School,.\n", - "thir\n", - "Uepubllcnn-stcel-\n", - "ablte.\n", - "amendment.\n", - "water.\n", - "climb.\n", - "do.\n", - "coffee.\n", - "impossible.\n", - "feats.\n", - "rights,\n", - "detectives.\n", - "two\n", - "nalinns.\n", - "home.\n", - "le-intlon.\n", - "It.\n", - "weeks.\n", - "sncrarT\n", - "it.\n", - "reads.\n", - "at\n", - "r;\n", - "d\n", - "hcriffs\n", - "New\n", - "levied.\n", - "S.\n", - "consideration.\n", - "of.\n", - "old\n", - "inquiry.\n", - "clerk9.\n", - "both.\n", - "inn-ha-\n", - "earth.-\n", - "of\n", - "ace.\n", - "Astor.\n", - "dertaking.\n", - "Canada.\n", - "trim.\n", - "qualiiica-\n", - "case.\n", - "If\n", - "m\n", - "throats.\n", - "hero.\n", - "infidels.\n", - "was:\n", - "section.\n", - "preserves.\n", - "mine.\n", - "road.\n", - "-\n", - "*s7.\n", - "combination.\n", - "loco-\n", - "elae.'\n", - "Stat*.\n", - "follows:\n", - "windpipe.\n", - "e-\n", - "Church.\n", - "November,\n", - "shall\n", - "streets.\n", - "dlea\n", - "d\n", - "5:18).\n", - "Journal.\n", - "usual.\n", - "remedy.\n", - "guilt.\n", - "the\n", - "Jr.,\n", - "of\n", - "for.\n", - "information.\n", - "experiences.\n", - "cerned.\n", - "certain-.\n", - "10,0(58.'\n", - "Republic.\n", - "hundred\n", - "and\n", - "citizens.\"\n", - "weaken.\"\n", - "tion.\n", - "facte\n", - "y,\n", - "zation.\"\n", - "us.\n", - "will-trem-ble\n", - "e\n", - "York.\n", - "copy.\n", - "Jer-\n", - "the\n", - "Mrs.\n", - "White.\n", - "feet\n", - "If\n", - "charges.\n", - "country.\"\n", - "suffer.\n", - "existence.\n", - "flag.\n", - "two.\n", - "at\n", - "act.\n", - "city.\n", - "rita'es,\n", - "A\n", - "answer.\n", - "ner.\n", - "camp.\n", - "remedy.\n", - "(13)\n", - "trouble\n", - "pastures.\n", - "and\n", - "street.\n", - "Jersey.\n", - "honor.\"\n", - "decision?\n", - "hours.\n", - "bunker.\n", - "neighbors.\n", - "more.\n", - "“\n", - "It\n", - "Polynesian.\n", - "\"\n", - "the\n", - "around.\n", - "uithsea.\n", - "forovor.\"\n", - "the\n", - "home.\n", - "Page's\n", - "Jr.\n", - "situate--\n", - ":\n", - "of\n", - "llnotoTur\n", - "are\n", - "t.\n", - "thus:\n", - "breath:\n", - "18W-\n", - "Madeline\n", - "States.\n", - "davs.\n", - "donations.\n", - "Curtis-Wright.\n", - "time!\"\n", - "armies.\n", - "hot\n", - "terday:\n", - "never\n", - "treasury.\n", - "health.\n", - "^\n", - "sanitariums.\n", - "it.\"\n", - "house.\n", - "s\n", - "old.\n", - "life.\n", - "arrest..\n", - ".\n", - "Co.,\n", - "of\n", - "iuet\n", - "coirered.\"\n", - "the\n", - "friends\n", - "order.\n", - "t\n", - "district,\n", - "protection.\n", - "pulpit\n", - "lost.\n", - "grown.\n", - "of\n", - ",k.\n", - "Judges.\n", - "recorder.\n", - "Mr.\n", - "gunpowder.\n", - "I'arlors.\n", - "d\n", - "system.\n", - "a\n", - "horizon.\n", - "now.\n", - "dear.\"\n", - "day.\n", - "winners\n", - "Herald.\n", - "washes!\n", - "full.\n", - "blood.\n", - "there\n", - "leaaantly,\n", - "plating.\n", - "wholesale.\n", - "wti-x\n", - "c\n", - "to\n", - "and\n", - "boost.\n", - "wire.\n", - "morality.\n", - "beginning.\n", - "cneap.\n", - "Leroux,\n", - "propriations.\n", - "union.’’\n", - "Dlspatcl.\n", - "Conference.\n", - "Sart\n", - ",\n", - "tured.\n", - "o\n", - "d\n", - "-\n", - "or\n", - "pitable.\n", - "uay.\n", - "bales.\n", - "Emanuel.\n", - "change,\n", - "happiness.\n", - "e\n", - "Pott.\n", - "a\n", - "male.\n", - "see\n", - "cf\n", - "*1.392,180,294.\n", - "expenses.\n", - "h\n", - "al\n", - "be\n", - "out.\n", - "hear\n", - "publics.\n", - "1909.\n", - "dusi-\n", - "wrought.\"\n", - "862,000.\n", - "him.\n", - "o\n", - "multitude.'\n", - "South.\n", - "d-\n", - "pany.\n", - "Raleiffh\n", - "$1t)0.\n", - "girl.'\n", - "days;\n", - "feet.\n", - "ago:\n", - "proof.\n", - "Union.\n", - "of\n", - "tests.\n", - "injuries.\n", - "determined.\n", - "the\n", - "-\n", - "1916.\n", - "Dr.\n", - ",\n", - "persons.\n", - "each.\n", - "bu-\n", - "gerous?\n", - "yesterday.\n", - "winter.\n", - "22\n", - "tion,\n", - "increases.\n", - "fash-\n", - "contemplate.\n", - "needless.\n", - ".\n", - "J\n", - "Preference.\n", - "adjourned.\n", - "per.\n", - "by\n", - "planets.\n", - "turn.\n", - "way,\n", - "rd\n", - "women\n", - "the\n", - "the\n", - "with\n", - "ishnient.\n", - "it\n", - "e\n", - "state.\n", - "year.\n", - "donor.\n", - "r\n", - "executed.\n", - "Judge.\n", - "effected.\n", - "cruisers.\n", - "Bhreveport.\n", - "4c.”\n", - "to-da-\n", - "bosom\n", - "hers.\n", - "us.\n", - "renown.\n", - "Island.\n", - "d\n", - "Jlouse.\n", - "Ineversawaman\n", - "could\n", - "wet.\n", - "everything's\n", - "Louis.\n", - "present\n", - "returns.\n", - ".OOO\n", - "nil\n", - "!...<>\n", - "it.\n", - "prisoner,\n", - "1903.\n", - "Class.”\n", - "...\n", - "Weekly.\n", - "made\n", - "other\n", - "outstretched\n", - "$43,000\n", - "case.\n", - ":\n", - "rice.\"\n", - "ob|ec?.CS.,°\n", - "American.\n", - "(4)\n", - "dinner.\n", - "place.\n", - "tive.\n", - "ud-\n", - "months.\n", - "days.\n", - "place.\n", - "accession\n", - "to\n", - "boUhcvism\n", - "friend.\n", - "re-union.\n", - "ter\n", - "street.\n", - "1938.\n", - "hits.\n", - "29.\n", - "comparison.\n", - "is.\n", - "[\n", - "“\n", - "demonstrate.\n", - "thereto.\n", - "offspring.\n", - "the-\n", - "1\n", - "most\n", - "said\n", - "colonics.\n", - "lives.\n", - "tively.\n", - "sure.\n", - "sonslderatlon\n", - "committed.\n", - "witli\n", - "transport.\n", - "(Nev.)Enterprise.\n", - "BARNABE.\n", - "discharged.\n", - "to\n", - "e,\n", - "nearly\n", - "t\n", - "for.\n", - "cash.\n", - "nr-iK-\n", - "giave.\n", - "wifcrpoi\n", - "*eBult.\n", - "pose.\n", - "sun.\n", - "spakc:\n", - "remain\n", - "rights.\n", - "if\n", - "comment.\n", - "attention.\n", - "enormou\n", - "2.'\n", - "woodman\n", - "servation.\n", - "Matter.\n", - "Md.\n", - "hand.\n", - "cago,\n", - "desired.\n", - "so\n", - "County\n", - "best\n", - ".\n", - "'\n", - "affair.\n", - "confidence.\n", - "accepted.\n", - "sight.\n", - "settled.\n", - "way.\n", - "Missouri.\n", - "r,\n", - "ordinances\n", - "1\n", - "avenue.\n", - "isalts.\n", - ".\n", - "direction.\n", - "trucks.\n", - "»*>\n", - "In\n", - "John.\n", - "Department.\n", - "oyes.\n", - "?,.\n", - "as\n", - "l.s\n", - "-I\n", - "worth\n", - "promised.\n", - "hum-dru-\n", - "man.\n", - "endorsed\n", - "J.,\n", - "b\n", - "Mich.\n", - "eater.\n", - "welfare.\n", - "misery\n", - "i«*\n", - "game.\n", - "reduced.\n", - "Moses.\n", - "future.\n", - "advance?\n", - "service,\n", - "crops:\n", - "t\n", - "procured.\n", - "pendent.\n", - "making.\n", - "unknown.\n", - "market.\n", - "Theatre.\n", - "d\n", - "M.\n", - "settlement:\n", - "Press.\n", - "IMS.\n", - "mass\n", - "'localV*\n", - "mentioned.\n", - "73%c.\n", - "those.\n", - "today.\n", - "school\n", - "agriculture,\n", - "unwise.\n", - "one.'\n", - "millions.\n", - "made.\n", - "decisively.\n", - "earth.\n", - "thereor,\n", - "Atkin«m\n", - "precisely\n", - ".\n", - "inspection.\n", - "Pl'tff\n", - "people.\n", - "death.\n", - "States.\n", - "law;'\n", - "Herald:\n", - "respect.\n", - ".\n", - "Italian.\n", - "lot*.\n", - "coal\n", - "of\n", - "attendant.\n", - "rule.\n", - "\"nam?.\"\n", - "basket.\"\n", - "program.\n", - "said:\n", - "acquired\n", - "Journal.\n", - "M\n", - "dead.\n", - "$2;\n", - "done.\n", - "company.\n", - "ment,\n", - "WITH\n", - "ased\n", - "through\n", - "clubs.\n", - "wave\n", - "18\n", - "for\n", - "point.\n", - "S5.ni;:.\n", - "views.\n", - "bill.\n", - "evi-\n", - "again.\n", - "s-\n", - "work.\n", - "I'm\"\n", - "Congress.\n", - "^\n", - "ity.\"\n", - "ginner.\"\n", - "r\n", - "place.\n", - "witnesses,\n", - "church-\n", - "ers.\n", - "head.\n", - "J.,\n", - ".,\n", - "-\n", - "men.\n", - "thoughtful*\n", - "UM,Ul«a>\n", - "deficit.\n", - "amount\n", - "parade.\n", - "move.\"\n", - "Democrat.\n", - "and\n", - "ion.\n", - "surgery.\n", - "crows.\n", - "to.\n", - "in\n", - "Kansas.\n", - "Roy\n", - "remedy.\n", - "act.\n", - "shall\n", - "strikes.\n", - "continued.\n", - "table.\n", - "be.\"\n", - "navy.\n", - "a\n", - "ieed\n", - "tlierefbro,\n", - "of\n", - "to.\n", - "system.\n", - "ernoon.\n", - "Interests.\n", - "basis.\n", - "d\n", - "there.\n", - "tion.\n", - "another.\n", - "Advt.\n", - "trains.\n", - "a!\n", - "passenger.\n", - "Commlaaloner*.\n", - "lor-.\n", - "concolved.\n", - "shirt\n", - "uncover\n", - "district\n", - "energy,\n", - "perience\n", - "Shortell.\n", - "prospect.\"\n", - "Monday.\n", - "once.\n", - "forcements.\n", - "Allegheny.\n", - "$5.\n", - "1919.\n", - "-rupte-\n", - "«.\n", - "Illinois.\n", - "notea,\n", - "quarrels\n", - "Wade.\n", - "subject\n", - "Mrs\n", - "Acosta\n", - "spoliation.\n", - "ble.\n", - "g\n", - "~\n", - "-\n", - "yours.\n", - "niece.\n", - "umbrella.\n", - "dians.\n", - "return.\n", - "camps.\n", - "ownercl\n", - "dCalh*\n", - "scheme.\n", - "stadium.\n", - "Congress.\n", - "fwi«i.\n", - "travelingaisen-al-\n", - "W.\n", - "survives.\n", - "dur\n", - "u\n", - "story.\n", - "tent\n", - "would\n", - "duPont.\n", - "trouble.\n", - "stealing.\n", - "ago.\n", - "days.\n", - "regiment.\n", - ".'i?\n", - "others.\n", - "taw.\n", - "us.\n", - "'\n", - "right.\n", - "oramanheIs.\n", - "fatal.\n", - "military.\n", - "which\n", - "hymyownhand.\n", - "debt.\n", - "right.\n", - "kbUL\n", - "absurdity.\n", - ".\"\n", - "him.\n", - "Central.\n", - "soil.\n", - "ended.\n", - "passatrc.\n", - ".\n", - "done.\n", - "d\n", - "price,\n", - "urgent.\n", - "then\n", - "town!\n", - "S3.1\n", - "water.\n", - "reality.\n", - "-\n", - "country.\n", - "n-\n", - "Saulnier,\n", - ".\n", - "descending.\n", - "rtcnoaswd.\n", - "select.\n", - "praise.\n", - "expccta-\n", - "and.\n", - "proceed\n", - "s\n", - "extension.\n", - "jubilee.\n", - "lilteen.\n", - "counter.\n", - "one.\n", - "importance.\n", - ".\n", - "street.\n", - "years.\n", - "Hospital.\n", - "public,\n", - "Atlantic.\n", - "year*.\n", - "master:\n", - "ll'lSe.\n", - "UUO.UUU.\n", - "*9.\n", - "Mink.\n", - "\"\n", - "s\n", - "escaped.\n", - "PhoA\n", - "you.\"\n", - "covers.\n", - "flowers.\n", - "facilities.\"\n", - "rule.\"\n", - "facts.\n", - "tho\n", - "1891,\n", - "anxiety.\n", - "to\n", - "com\n", - ".\n", - "tbe\n", - "appeal.\n", - "the\n", - "Me.\"\n", - "rnther\n", - "livelihood.\n", - "?\"\n", - "floors.\n", - "IL\n", - "instances.\n", - "Moor.”\n", - "Detroit.\n", - "on.\n", - "uri\n", - "hi\n", - "colored\n", - "S.\n", - "pavement.\n", - "cneap.\n", - "work.\n", - "river.\n", - "confusion.\n", - "buildings.\n", - "Angnstly-beautiful-\n", - "cants.\n", - "d\n", - "iibtrology.\n", - "viz.:\n", - "distribution\n", - "wherein\n", - "June]!).\n", - "y\n", - "missioner.\n", - "command.\n", - "greet-\n", - "’’\n", - "-\n", - "Auditorium.\n", - "on\n", - "bler.\n", - "Cultivator.\n", - "they\n", - "letter\n", - "northwest.\n", - "ment.\n", - "-\n", - "quie\n", - "him.\n", - "them.\n", - "raised.\n", - "shaft.\n", - "it\n", - "therein.\n", - "can\n", - "Agent,\n", - "denizen.\n", - "tablished.\n", - "v\n", - "purchased,\n", - "him.\n", - "course.\"\n", - "war\n", - "kissed.\n", - "ing:\n", - "may\n", - "this\n", - "effort.\n", - "capstone.\n", - "minister.\n", - "laud.\n", - "future.\n", - "doomed.\n", - "philosopher.\n", - "roads.\n", - "people.\n", - "debt.\"\n", - "although\n", - "demanded\n", - "hands.\n", - "over,\n", - "r\n", - "quired,-'\n", - "Fitzsimmons-Pritchar-\n", - "i?j,d.\n", - "remedy.\n", - ".\n", - "growth.\n", - "as\n", - "Sr.u.\n", - ".\n", - "(lellglilcd\n", - ".\n", - "triple.\n", - "Sangston.\n", - "per-\n", - "it\n", - "g\n", - "driven\n", - "seen.\n", - "diaries.\n", - "corporations.\n", - "them.\n", - "ments.\n", - "postpaid.\n", - "tions.\n", - "extermination.\n", - "drowned.\n", - "t\n", - "gallows.\n", - "to\n", - "pan.\n", - "financed.\n", - "Century.\n", - "I\n", - "themeans\n", - "-\n", - "m.\n", - "o'clock.\n", - "1802.\n", - "inaugural.\n", - "utlons.\n", - "be\n", - "y.\n", - "effect\n", - "•v;.\n", - "o\n", - "Church.\n", - "Smith.\n", - "belt.\n", - "sailor.,\n", - "Medlfer-\n", - "again.\n", - "both.\n", - "cemetery.\n", - "time.\n", - "tbe\n", - "rules.\n", - "them.\n", - "groceries.\n", - "c\n", - "me.\n", - "umphantly.\n", - "sunbeam.\n", - "DallaaCo.,\n", - "officers\n", - "covered.\n", - "laws.\n", - "risen.\n", - "within.\n", - "rnment.\n", - "best.\n", - "MARTYR\n", - "which\n", - "supply.\n", - "road.\n", - "vain.”\n", - "good.\n", - "Nar-I\n", - "Sts.\n", - "was\n", - "second.\n", - "available.\n", - "evening.\n", - "uway.\"\n", - "Kitty!\n", - "one.\n", - "55\n", - "Incumbents.\n", - "of\n", - "girl.\n", - "fertilizer.\n", - "people\n", - "interior.\n", - "95.\n", - "pair,\n", - "sons\n", - "sweat,\n", - "e\n", - "lf\n", - "therein.\n", - "it\n", - "to\n", - "gentleman\n", - "it.\n", - "bottle.\n", - "officials.\n", - "question.\n", - "waters.\n", - "State.\n", - "laws.\n", - "Herald\n", - "had\n", - "West.\n", - "work\n", - "Vermonler.\n", - "resolution:\n", - "desirable.\n", - "21st.\n", - "Lincoln\n", - "Store.\n", - "months.\n", - "ordinance.\n", - "school.\n", - "It\n", - "state\n", - "Wilmington.\n", - "cnmpcm-ntioi-\n", - "possoselon.\n", - "11\n", - "l'\n", - "classes.\n", - "41\n", - "occurrence.\n", - "breeze.\n", - "war.\n", - "purposes.\n", - "aim.\n", - "ness.\n", - "{140,000,000\n", - "reply.\n", - "birth.\n", - "par\n", - "enduring.'\"\n", - "out\n", - "finished.\n", - "but\n", - "which\n", - "barrel.\n", - "teams.\n", - "s.nVI\n", - "the\n", - "complaint.\n", - "sorrow\n", - "Hoffecker.\n", - "lit.\n", - "by\n", - "was\n", - "possible.\n", - "safe-\n", - "instructions.\n", - "afterward.\n", - "seemed\n", - "about.\n", - "serious.\n", - "fighting.\n", - "deemed,\n", - "ages.\n", - "rule.\n", - "1860.\n", - "Jdn.\n", - "the\n", - "Hadlej.\n", - "ng\n", - "idea.\n", - "interest.\n", - "beginning.\n", - "remain.\n", - "vices.\n", - "perhaps,\n", - "bankers.\n", - "1)\n", - "extension.\n", - "tight.\"\n", - "only.\n", - "“\n", - "didate.\n", - "th\n", - "son*.\n", - "paper.\n", - "his\n", - "music.\n", - "disin-teime-\n", - "his\n", - "re\n", - "England.\n", - "route.\n", - "[Alta.\n", - "ordinance.\n", - "grade.\n", - "t\n", - "round.\n", - "Leonardt.\n", - "mischief,\n", - "grave.\n", - "$24.\n", - "merchandise.\n", - "action.\n", - "aforesaid.\n", - "Century.\n", - "did\n", - "Impies-slvo-\n", - "yet.\n", - "faith.\n", - "ton.\"\n", - "October\n", - "digestion.\n", - "death.\n", - "method.\n", - "waitress.\n", - "offences.\n", - "the\n", - "nour.\"\n", - "Rapublicaoa\n", - ":\n", - "Navy.\n", - "appear\n", - "vantages\n", - "them.\n", - "for\n", - "policy.\n", - "contract.”\n", - "adjourn.\n", - "crime.\n", - "a\n", - "speculators.\n", - "Cooper.\n", - "work.\n", - "shrine.\n", - "1877.\n", - "construction,\n", - "represent.\n", - "night.\n", - "damaged.\n", - "Christ.\"\n", - "chairman\n", - "stolen\n", - "~a.\n", - "past\n", - "delivery.\n", - "wound.\n", - "otherwise.\n", - "prosper\n", - "examination.\n", - "Hitters.\n", - "instruction.\n", - "fsalad-\n", - "remedy.\n", - "friends\n", - "World,\n", - "nutritious\n", - "Chapter,\n", - "war.\"\n", - "o\n", - "t'\n", - "certlflcales.\n", - "a\n", - "issued\n", - "wildcats.\n", - "advertising\n", - "century\n", - "school.\n", - "er.\n", - "an\n", - "Court.\n", - "thither.\n", - "that\n", - "politicians\n", - "understood.\n", - "Brigham,\n", - "^ashiiiKtou.\n", - "(SEAL)\n", - "raa-\n", - "fort.\n", - "things.\"\n", - "old.\n", - "rates.\n", - "line.\"\n", - "dcMdtw\n", - "rule.\n", - "stories.\n", - "Issue.\n", - "I\n", - "barbarism.\"\n", - "Montgomery\n", - "...\n", - ".\n", - "Home.\n", - "Congress,\n", - "esteem,\n", - "amirnach.\n", - "help.\n", - "nature\n", - "seems\n", - "cause.\n", - "Barker\n", - "i\n", - "the\n", - "occasion,\n", - "g\n", - "street;\n", - "ment:\n", - "death.\n", - "ladies!\n", - "men.\n", - "it\n", - "above.\n", - "is\n", - "so.\n", - "twlE.\n", - "brances.\n", - "Advocate.\n", - "simulation.\n", - "\"\n", - "e\n", - "?\n", - "I\n", - "canal.\n", - "blis\n", - "here.\n", - "Tuesday.\n", - "tory.\n", - "be\n", - "appoint­\n", - "has-\n", - "curds.”\n", - "author­\n", - "tim.\n", - "(¦mis\n", - "financially.\n", - "bard.,\n", - "order,\n", - "days.\n", - "\"\n", - "cor-\n", - "n\n", - "fo\n", - "tooth\n", - "common\n", - "edy.\"\n", - "r\n", - "exit.\n", - "be\n", - "Senate.\n", - ".\n", - "2\n", - "pose.\n", - "out.'\n", - "Congress.\n", - "Glory.\"\n", - "Chronicle\n", - "It\n", - "to\n", - "Sev-\n", - "them.\n", - "wrench.\n", - "leisure.\n", - "be.\n", - "evidences.\n", - "navy.\n", - "to-day.\n", - "gTMtlT.\n", - "dissipation.\n", - "date\n", - "m.\n", - "cirtificates\n", - "weal\n", - "west.\n", - "Appomattox.\n", - "miles.\n", - "utility.\n", - "up.\n", - "j\n", - "canyon.\n", - "error.\n", - "left.\n", - "Bolshevik\n", - "Transportation,\n", - "tho\n", - "living.\n", - "trial.\n", - "k-.\n", - "go\n", - "bounded.\n", - "churches.\n", - "o’clock.\n", - "resolvents.\n", - "the\n", - "Fislihigcrock..-\n", - "L\n", - "ISIMI\n", - "labor.\n", - "said:\n", - "aid.\n", - "order.\n", - "enormously.\n", - "reason.\n", - "year.\n", - "pos-\n", - "\\\\\n", - "mnn.\n", - "tinues.\n", - "board\n", - "mili\n", - "ones.\"\n", - "acquaintance\n", - "promised.\n", - "parts.\n", - "Tnecranwra\n", - "bidder.\n", - "vigor.\n", - "write.\n", - "property.\n", - "action.\n", - "suit.\n", - "gentlemen,\n", - "by.”\n", - "the\n", - "was,-\n", - "of\n", - "Agnes,\n", - "Racine.\n", - "Mamer.\n", - "heavy,\n", - "eto.\n", - "storv.\n", - "b\n", - "fitness.\n", - "College,\n", - "Intellectuals\n", - "maximum\n", - "man.\n", - "lasts.\n", - "the\n", - "from.\"\n", - "*\n", - "program.\n", - "cause.\n", - "fering.\n", - "Sun.\n", - "two\n", - "that\n", - "givci\n", - "carried\n", - "Adams\n", - "-\n", - "faces,\n", - "meeting.\n", - "boats.\n", - "installments\n", - "vantage.\"\n", - "tance.\n", - "her.\n", - "this\n", - "here.\n", - "in*.\"\n", - "Gazette.\n", - "s\n", - "it\n", - "meeting.\n", - "rooster\n", - "brick.\"\n", - "Agent,\n", - "World.\n", - "the\n", - "flesh.\"\n", - "Mifflin.\n", - "con-\n", - "speedily.\n", - "also.\n", - "etubjrrae\n", - "months.\n", - "1912,\n", - "groups.\n", - "them.\n", - "were\n", - "gov-\n", - "dent.\n", - "charges.\n", - "like.\n", - "bo-\n", - "acre.\n", - "1866.\n", - "painful.\n", - "domination.\n", - "county.\n", - "this\n", - "admirably.\n", - "work\n", - "Graphic.\n", - "tion.\n", - "cure.\n", - "be\n", - "ner.\n", - "y-tw-\n", - "employed.\n", - "here.\"\n", - "God.\n", - "»\n", - "means.\n", - "testation.\n", - "that?\n", - "noon.\n", - "games.\n", - "years.\n", - "(14)\n", - "crat.\n", - "jail.\n", - "both.\n", - "o\n", - "re—\n", - "were\n", - "!?\".\n", - "therefor.\n", - "fruitfulness.\n", - "nitnii,\n", - "shrritf.\n", - "Addlcks.\n", - "dol­\n", - "DEPARTMENT.\n", - "Job,\n", - "December,\n", - "a\n", - "despots,\n", - "Times.\n", - "time.\n", - "iMimiitfli\n", - "country,\n", - "museum.\n", - "reform.\n", - "t:\n", - "brother.\n", - "ence.\n", - "come.\n", - "(hem.\n", - "of\n", - "tervenes.\n", - "bins.\n", - "by\n", - "employed\n", - "laborer\n", - "January\n", - "north.\n", - "ored.\n", - "ores.\n", - "church.\n", - "bht\n", - "Palmer.\n", - "which\n", - "Sn\n", - "that\n", - "tales.\n", - "facilities.\"\n", - "times\n", - "environment.\n", - "tiser.\n", - "earth.\n", - "wl»?l«\n", - "emoluments.\n", - "which\n", - "sluice-\n", - "will\n", - "l\n", - "cop).\n", - "attendance.\"\n", - "fruits.\n", - "by\n", - "Conkllng.\n", - "noise.\n", - "croam.\n", - "political\n", - "de\n", - "sttat-\n", - "restaurant\n", - "which\n", - "him.\n", - "clear.\n", - "them.\n", - "test.\n", - "the\n", - "Kihn.\"\n", - "o\n", - "overwhelming.\n", - "streets.\n", - "Brown.\n", - "for\n", - ".\n", - "1917.\n", - "r\n", - "matter.\n", - "school.\n", - "now\n", - "'\n", - "merit.;\n", - "bettor.\n", - "hid.\n", - "officers\n", - "tomorrow\n", - "the\n", - "the\n", - "country.\n", - "demand.-\n", - "States.\n", - ";\n", - "structure.\n", - "p.\n", - "bliss.\n", - "ately.\n", - "de\n", - ",\n", - "the\n", - "camp.\n", - "$3,000,000.\n", - "business\n", - "Tribune:\n", - "cp.\n", - "city.\n", - "cure.\n", - "Jury.\n", - "roundings.\n", - "me.\n", - "duty.\n", - "politics.\"\n", - "putation\n", - "Max.\"\n", - "this.\n", - "interest\n", - "-\n", - "'\n", - "cattle.\n", - "all.\n", - "canal.\n", - "public\n", - "T.\n", - "*«\n", - "time.\n", - "in\n", - "aver\n", - "ruin.\n", - "I\n", - "and\n", - "them.\n", - "?\n", - "subject.\n", - "of\n", - "In.”\n", - "ramble.\"\n", - "d\n", - "mended,\n", - "explained.\n", - "husking.\n", - "rA\n", - "much.\n", - "destroyed.\n", - "the\n", - "-\n", - "many.\n", - "trw-Dr-\n", - "tnauuer.\n", - ".A»i,n\n", - "M.\n", - "delightful.\n", - "in\n", - "their-merr-\n", - "July,\n", - "-\n", - "them.\"\n", - "Mas­\n", - "tonishing.\n", - "uratiou.\"\n", - "erally.\n", - "values.\n", - "theoountry.\n", - "right.\n", - "thereof.\n", - "dollars.\n", - "house.\n", - "Josefa.\n", - "adopted.\n", - "elimination\n", - "Makpm,\n", - "b\n", - "morning\n", - "enter?\n", - "s\n", - "which\n", - "Philadelphia.\n", - "prosperity.\n", - "carried.out.\n", - "waru.\n", - "there.\n", - "1891,\n", - "4.\n", - "Congress\n", - "throughout\n", - "whisper.\n", - "developments\n", - "country.\n", - "con-\n", - "cc\n", - "into\n", - "consumption.\n", - "you.\n", - "time.\n", - "said\n", - "pound.\n", - "disorganized\n", - "o'clock.\n", - "taxed.\n", - "companions\n", - "success\n", - "means.\n", - "so\n", - "abadauLt.\n", - "State.\n", - "excursions.\n", - "All\n", - "band.\n", - "reside.\n", - "doctors.\"\n", - "management.\n", - "and\n", - "health.\"\n", - "haven.\n", - "d\n", - "operation.\n", - "suit.\n", - "with\n", - "and\n", - "system,\n", - "skill*.\n", - "office.\n", - "law.\n", - "A\n", - "rope.\n", - "Leady.\n", - "spent.\n", - "compared\n", - "a\n", - "most\n", - "--\n", - "surface.\n", - "-\n", - "'wife.\n", - "assumption.\n", - "instances.”\n", - "ight.\"\n", - "favor.\n", - "advance.\n", - "t\n", - "fering.\n", - "exercises.\n", - "detstood.\n", - "hand.\n", - "Britain\n", - "tai\n", - "cneap.\n", - "grain.\n", - "nt\n", - "»\n", - "iine.\n", - "rest.\n", - "J24-3t\n", - "permit.\n", - "avenue:\n", - "Hon.\n", - "inink\n", - "fast.\n", - "tor.”\n", - "was\n", - "action.\n", - "Ion.\n", - "follows:\n", - "infamy.\n", - "exceptions,\n", - "board.\n", - "Ac,\n", - "fession\n", - "Fairmont.\n", - "hearings.\n", - "speculation.\n", - "Lords.\n", - "bridge.\n", - "manner.'\n", - "signature.\n", - "by\n", - "will.\n", - "W\n", - "Borough.\n", - "re-\n", - "en\n", - "prepared\n", - "Europe\n", - "Commonwealth.\n", - "examina-\n", - "Shafer.\n", - "hungry.\n", - "UIUU\n", - "1\n", - "con-\n", - "pleased.\n", - "bad.\n", - "affair.\"\n", - "destroyed.”\n", - "dealings.\n", - "cess.\n", - "bouse.\n", - "Garrett,\n", - "«1.782\n", - "imported.\n", - "f-\n", - "sending-\n", - ",\n", - "granted.\n", - "society.\n", - "applause.]\n", - "river.\n", - "assignable.\n", - "Advocate.\n", - "field.”\n", - "Hamburg.\n", - "classified.\n", - "advantage.\n", - "the\n", - "Union.\n", - ".\"\n", - "witness-\n", - "session.\n", - "sleep.\n", - "prelcrred,\n", - "brain.\n", - "list.\n", - "army.\n", - "map.\n", - "them.\"\n", - "details.\n", - "htar-in-\n", - "town.\n", - "hue.\"\n", - "was\n", - "doubt.\n", - "painter.\n", - "6.\n", - "said,\n", - "Basin.\n", - "G:iz'tle\n", - "tho\n", - "appears-\n", - "1*.M*\n", - "later.\n", - "aim.\n", - "safe.\"\n", - "Weekly.\n", - "sense.\n", - "N.,\n", - "work.\n", - "circuinstan\n", - "Ridge,\n", - "few.\n", - "100,000.\n", - "they\n", - "substitute.\n", - "divorce.\n", - "Paper.\n", - "Piepenbrtng,\n", - "lorever.\n", - "needy.\n", - "Russia.\n", - "\"Five-Twenty-\n", - "•\n", - "music.\n", - "Ataine.)\n", - "possible.\n", - "hint.\n", - "work.\n", - "column.\n", - "S.\n", - "accounts,\n", - "tho\n", - "of\n", - "penalty.\n", - "old\n", - "two-thir-\n", - "to-day.\n", - "shrine.\n", - "believing.\n", - "rapidly.\n", - "eternity\n", - "Breckinridge.\n", - "wide.\n", - "papers\n", - "representative.\n", - "[Oheera.]\n", - "young.\"\n", - "C\n", - "documents.\n", - "meet-\n", - "s.\n", - "k\n", - "co.,\n", - "If\n", - "r\n", - "action.\n", - "veranda.\n", - "—\n", - "well.\n", - "power.\n", - "teresting.\n", - "nights.\n", - "1930.\n", - "fair-mind-\n", - "about\n", - "\"\n", - "lUtnamlnl,\n", - "to.\n", - "hai\n", - "on.\n", - "14%.:\n", - "atockhold\n", - "lives.\n", - "Inailments\n", - "so\n", - "consumption.\n", - "said\n", - "Services.\"\n", - "Polk!\n", - "a\n", - "appearance.\n", - "scene.\"\n", - "-,\n", - "llelknap.\n", - "1\n", - "cash.\n", - "prosperous.\n", - "g\n", - "superintendent's\n", - "stars.\n", - "county.\n", - "the\n", - "Ami\n", - "ven\n", - "Htraid.\n", - "time.\"\n", - "Unite\n", - "so.\n", - "\"\n", - "hibited.\n", - "\"\"\n", - "M.\n", - "\"\n", - "bosses.\n", - "that\n", - "d\n", - "settees,\n", - "American.\n", - "given.\n", - "sight.\n", - "the\n", - "hearing.\n", - "Skc.\n", - "w?»t-\n", - "Coinylaiut.\n", - "government.\n", - "great\n", - "triumphani.\n", - "friends.\n", - "minutes.\n", - "notes.\n", - "pelerfcy.\n", - "rearino-\n", - "me:\n", - "hv\n", - "1012\n", - "fall.\n", - "arooad\n", - "work\n", - "Bradley.\n", - "wel\n", - "fire.\n", - "people.\n", - "finely.\n", - "-\n", - "follows:\n", - "vandal.\n", - "time.\n", - "place.\"\n", - "knowledge.\n", - "woman\n", - "thence\n", - "t\n", - "it\n", - "seven.\n", - "prices.\n", - "prevail.\n", - "a\n", - "toilow.\n", - "wine.\n", - "thorn.\n", - "the\n", - "whereabouts.\n", - "pasted\n", - "2S°-.,\n", - "impregnable.\n", - "then.\n", - "-\n", - "widt\n", - "d\n", - "creditable\n", - "agent.\n", - "deavor.\n", - "collected.\n", - "severely.\n", - "forthcoming.\n", - "appear.\n", - "a\n", - "practice.\n", - "you.\n", - "brances.\n", - "Kenuett\n", - "1912.\n", - "Union.\"\n", - "possible.\n", - "ComImloner*.\n", - "«fcc.\n", - "was\n", - "advertisement\n", - "pig-\n", - "light.\n", - "M.\n", - "work.\n", - "stone.\n", - "bjdrug-\n", - "Clerk\n", - "Eagle,\n", - "Messrs.\n", - "000.\n", - "largest\n", - "night.\n", - "-\n", - "passed.\n", - "eagerly\n", - "e\n", - "required.\n", - "k\n", - "Castle.\n", - "$3.25?\n", - "such\n", - "possibilities,\n", - "Journal.\n", - "elevation.\n", - "renovate,\n", - "the\n", - "Tho\n", - ".\n", - "the\n", - "cemetery\n", - "n-\n", - "the\n", - "ard.\"\n", - "beyond?\n", - ".\n", - ".\n", - "thm\n", - "most\n", - "feet.\n", - "cheek.\n", - "known.\n", - "n\n", - "useless,\n", - "Marr\n", - ",7\n", - "Tribune\n", - "over.\n", - "slightly\n", - "Saturday\n", - "execution.\n", - "burn-\n", - "N.\n", - "autmal\n", - "holes.\n", - "acted,\n", - "Oregon.\n", - "servation.\n", - "be-\n", - "prevent\n", - "sort.\n", - "Mary\n", - "offer.\n", - "feet.\n", - ".\n", - "bushel.\"\n", - "for\n", - "Montgomery\n", - "pered:\n", - "he\n", - "Wheeling,\n", - "co-owner\n", - ":\n", - "railway.\n", - "es\n", - "gage.\n", - "il\n", - "(father-in-la-\n", - "Gundelflnger.\n", - "III.\n", - "ng\n", - "our\n", - "Jlotmil.\n", - "admiration.\n", - "him.”\n", - "thorn\n", - "professions.\n", - "felt.\n", - "¬\n", - "ih»\n", - "out-\n", - "her,\n", - "jecture.\n", - "service.\n", - "wood.\"\n", - "resort.\n", - "Butt**\n", - "shrine.\n", - "Instruction,\n", - "f\n", - "non.\n", - "up.”\n", - "WTIC,\n", - "sec.”\n", - "rt\n", - "flora.\"\n", - "business.\n", - "s\n", - "for\n", - "Hour*-\n", - "proteet\n", - "little\n", - "cheated\n", - "Result\n", - "»ked.\n", - "Telegram.\n", - "should\n", - "it\n", - "to8p.m.\n", - "v\n", - "wrong.\n", - "continent.\n", - "j\n", - "fire.\n", - "Cradock.\n", - "beer-garde-\n", - "automobile.\n", - "provid\n", - "tired.\n", - "readmittei\n", - "applause.\n", - "removed.\n", - "owners.\n", - "tl\n", - "their\n", - "have.\n", - "Pleas.\n", - "tions.\n", - "is\n", - "while,\n", - "Dawson.\"\n", - "Marne'.\n", - "call.\n", - "weeks.\n", - "M.\n", - "at\n", - "-\n", - "respect.\n", - "court.\"\n", - "Gazette.\n", - "wall.\n", - "water.\n", - "i,\n", - "1902.\n", - "employed.\n", - "excitement.\n", - "declined.\n", - "paid.\n", - "freedom.\n", - "toguardthr\n", - "as\n", - "position.\n", - "boxes.\n", - ".struggle.\n", - "Constitution.\n", - "lUalUc.\n", - "as\n", - "breakfast.\n", - "r-\n", - "lows:\n", - "slave-trade-\n", - "tenni.\n", - "to\n", - "more.\n", - ".\n", - "deserved.\n", - "reference.\n", - "world.\n", - "States\n", - "stantia-\n", - "expedient.\n", - "welcome.\n", - "bright\n", - "and\n", - "fice.\n", - "s\n", - "toe,\n", - "started.\n", - "Cuba.\n", - "taken.\n", - "Union.\n", - "that.\"\n", - "Nicaragua.\n", - "dition.\n", - "Newark.\n", - "andItoldhimsoandhesaidhe\n", - "d-\n", - "Ac.,\n", - "obligation.\n", - "Rivers.\n", - "happened.\n", - "possible.\n", - "loyally.\n", - "Union.\n", - "vessel.\"\n", - "me.\n", - "saries.\n", - "tliu\n", - "-\n", - "n\n", - "at\n", - "people.\n", - "brother,\n", - "civilization.\n", - "would\n", - "distress.\n", - "simplicity.\n", - "gentle\n", - "week.\n", - "state.\"\n", - "exactly.\n", - "Island\n", - "he\n", - "Ward.\n", - "our\n", - "-\n", - "cents!\n", - "cused.\n", - "s\n", - "s,\n", - "security.\n", - "school.\n", - "broad\n", - "to\n", - "Nov\n", - "Delaware,\n", - "Bonds.\n", - "o.\n", - "reality.\n", - "sound;\n", - "written.\n", - "5005peg\n", - "himself.\n", - "peace,\n", - "bereavement.\n", - "iiresular,\n", - "be\n", - "promises.\n", - "home.\n", - "over.\n", - "itself.\n", - "I).,\n", - "unfortunate\n", - "gogue.\n", - "the\n", - "contract\n", - "lrienu,\n", - "reported.\n", - "price.\n", - "enterprise.\n", - "of\n", - "game.\n", - "baud*,\n", - "American\n", - "this\n", - "want.\n", - "Jcpson\n", - "attempts.\n", - "however.\n", - "problem.\n", - "ilianki.\"\n", - "73\n", - "ago.\n", - "noon.\n", - "with\n", - "privilege.\n", - "railroads.\n", - "a\n", - "!\n", - "law.\n", - "muon.\n", - "place.\n", - "season.\n", - "grind.\n", - "ever\n", - "ground\n", - "Esq.\n", - "a\n", - "California.\n", - "of.\n", - "public.\n", - "weather.\n", - "vestigation.\n", - "summary:\n", - "feet.\"\n", - "heads.\n", - "'\n", - "nhaae.\n", - "ruWKiM\n", - "■,,'\n", - "for\n", - "by.\n", - "Calvin\n", - "charge.\n", - "mountains.\n", - "arc\n", - "of\n", - "larger,\n", - "debater.”\n", - "war.\n", - "said:\n", - "1370,\n", - "Chesapeake.\n", - ".\n", - "fort.\n", - "by\n", - "e\n", - "Is.\n", - "1903.\n", - "high.\n", - "UK,\n", - "majority.\n", - "and\n", - "result.\n", - "er\n", - "?\"\n", - "Lancaster,\n", - "slam.\"\n", - "(Signed)\n", - "Weekly.\n", - "'\n", - "tables.\"\n", - "Moines.\n", - "ObBorver,\n", - "one.\n", - "thought.\n", - "forfeited.\n", - "to\n", - "It\n", - "is\n", - "Chronicle.\n", - "it.\n", - "World.\n", - "d\n", - "way\n", - "night.\n", - "\"Congress.\"\n", - "hkeecape.\n", - "Calhoun',\n", - ".\n", - "should\n", - "the\n", - "and\n", - "1\n", - "elsewhere.\n", - "ion.\n", - "vonom\n", - "climate.\n", - "both.\n", - "Lawrence.\n", - "iionne-\n", - "co!\n", - "expedition.\n", - "contains.\n", - "and\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Burgess)\n", - "cheap.\n", - "ENOUGH\n", - ",\n", - "lor\n", - "by\n", - "over.\n", - "JAA41»\n", - "tory\n", - "belonging,\n", - "on\n", - "style:\n", - "L'XJK'IulitUrfH.\n", - "bill!\n", - "subscription\n", - "in-\n", - "crats.\n", - "them.\n", - "tho\n", - "velt\n", - "furnished.\n", - "acts.\n", - "ers.\n", - "wasted\n", - "Nevada,\n", - "harbor.\n", - "surrendered\n", - "see\n", - "knives.\n", - "evei\n", - "world:\n", - ".\n", - "Promotions.\n", - "\"\n", - "ways.\n", - "buried\n", - "extra.\n", - "J.\n", - "promote.\n", - "nay.\n", - "dr.\n", - "less.\n", - "'coop.'\n", - "Country.\n", - "wife.\n", - "rot.\"\n", - "calculated.\n", - "education.\n", - "person,\n", - "»\n", - "children.\n", - "ILLY,\n", - "Bear.\n", - "ways.\n", - "!«-\n", - "and\n", - "management.\n", - "ants.\n", - "commodity.\n", - "scenes\n", - "tribune.\n", - "wide-sprea-\n", - "costs.\n", - "success.\"\n", - "ter\n", - "indopen-deuce-\n", - "a\n", - "La'jle.\n", - "after\n", - "pose.\n", - "lauding.\n", - "llowers,\n", - "Hoursfrom9a.m..todr.m.\n", - "year.\n", - "Keswick's\n", - "forever\n", - "us,\n", - "witness\n", - "Spetoh\n", - "living-situation.\n", - "e\n", - ".\n", - "other.\n", - "helpful.\n", - "another.\n", - "exhib.:\n", - "best\n", - "so\n", - "animated.\n", - "says\n", - "arrests.\n", - "abiding\n", - "America.\n", - "¬\n", - "t\n", - "says:\n", - "efface.\n", - "hereafter.\n", - "fit.”\n", - "Iwentv-tdne\n", - "eorners\n", - "I.\n", - "the\n", - "night.\n", - "immediately.\n", - "law.\n", - "club.'\n", - "editors.\n", - "McDonough\n", - "Stomp,\"\n", - "?\n", - "as\n", - "erty.\"\n", - "evening,\n", - "manafar-\n", - "C.\n", - "distress.\n", - "(SKAL.I\n", - "States.\n", - "without.\n", - "e\n", - "school.\n", - "dealers.\n", - "gs\n", - "thing.\n", - "names.\n", - "fo\n", - "opposed.\n", - "parent\n", - "greatly\n", - "ber\n", - "lce.\n", - "their\n", - "privilege.\n", - "Cost*.\n", - "Wlikes-Barr-\n", - "Sentinel.\n", - "13.\n", - "Patriot.\n", - "night.\n", - "under\n", - "war.\n", - "com-\n", - "tion.\n", - "y\n", - ".\n", - "victim\n", - "sharks.\n", - "by\n", - "day.\n", - "graved.\n", - "other\n", - "one\n", - "Department.\n", - "the\n", - "Trado\n", - "monuments\n", - "accident.\n", - "port.\n", - "anticipated.\n", - "through\n", - "society.\n", - "soon.\n", - "llechuanas.\n", - "holidays\n", - "rr\n", - "insisted\n", - "snry,\n", - "Tribune.\n", - "can\n", - "father-in-la-\n", - "hoards.\n", - "Assembly.\n", - "e\n", - "fortunate.\n", - "East.\n", - "Gregory,\n", - "floners\n", - "more\n", - "These\n", - "stove.\n", - "overabundance\n", - "a\n", - "your-icou-\n", - "period.\n", - "ago.\n", - "them.\n", - "tug.\n", - "them.\n", - "work.\n", - "tbo\n", - "s\n", - "here.\n", - "semi-annuall-\n", - "tion.\n", - "desirable.\n", - "country.\n", - "board\n", - "cincts.\n", - "President.\n", - "lf\n", - "period.\n", - "pit.\n", - "sleep.\n", - "here.\"\n", - "do,\n", - "a\n", - "cousin\n", - "lookout.\n", - "days.\n", - "wages?\"'\n", - "DosAngcs,\n", - "t\n", - "female,\n", - "him.\n", - "Secretary\n", - "Sampson.\n", - "A\n", - "time.\n", - "President?\n", - "lost.”\n", - "mistross.\n", - "vote.\n", - "tearful.\n", - "Juneau.\n", - "you?\n", - "concerned.\n", - "perseverence.\n", - "a\n", - "Amrrian.\n", - "flattery.\"\n", - "market.\n", - "love.\n", - "years.\n", - "battle:\n", - "oi\n", - "without\n", - "-\n", - "papers\n", - "wealthy.\n", - "Newt.\n", - "tin.\n", - "Interest\n", - "fashion\n", - "togetner,\n", - "lution^\n", - "shore.\n", - "re-\n", - ".\n", - "lectators.\n", - "death.\n", - "such\n", - "betteer.\n", - "control.\n", - "thereof.\n", - "standing\n", - "nt\n", - "crystal.\n", - "payroll.\n", - "opinion.\n", - "is\n", - "’\n", - "membership.\n", - "Keehn.\n", - "premises.\n", - "people,\n", - "suspended\n", - "name.\n", - "deserted,Ibrlio\n", - "consumption.\n", - "not\n", - "accruing\n", - "Hepnblio,\n", - "iury.\n", - "as\n", - "shrine.\n", - "H\n", - "world/'\n", - "revoked,\n", - "points.\n", - "hand.\n", - "Porcupine\n", - "ago.\n", - "reasonable.\n", - "were\n", - "geese.\n", - "there,\n", - "geography.\n", - "fering.\n", - "etc.,\n", - "-w-\n", - "\"cause.\"\n", - "ble\n", - "in\n", - "Observer.\n", - "15,1888.\n", - "fire.\n", - "standpoint.\n", - "and\n", - "iut-.\n", - "accident\n", - "contented.\n", - "she\n", - "s\n", - "Dukes.\n", - "fur.\n", - "s,\n", - "rifle.\n", - "some\n", - "panies\n", - "drive.\n", - "i\n", - "roads.\n", - "the\n", - "it.\n", - "Science.\n", - "lion.\n", - "f,\n", - "to\n", - "be-\n", - "ty\n", - "u-\n", - "In.\"\n", - "-\n", - "to\n", - "reach\n", - "provided-\n", - "convention.\n", - "line.\n", - "4th\n", - "achieve¬\n", - "Gazette.\n", - "*50.000.\n", - "methods.\n", - "nodded.\n", - "two\n", - "vassals\n", - "origin.\n", - "sword.\n", - "Loos.\n", - "Paris\n", - "6\n", - "Record.\n", - "'2o.\n", - "<\n", - "ear.\"\n", - "stated.\n", - "people\".\n", - "investigation,\n", - "attraction.\n", - "highways.\n", - "court.\n", - "senseless!\n", - "Magazine.\n", - "ashes.\n", - "once\n", - "quences.\n", - "corners.\n", - "are\n", - "city.\n", - "school\n", - "dis-\n", - "leader.\n", - "Slates\n", - "spring.\n", - "potior.\n", - "the\n", - "beginning.\n", - "ford*.\n", - "14.\n", - "us.\n", - "disgust.\n", - "fork\n", - "launched.\n", - "cc\n", - "market.\n", - ".\n", - "fee\n", - "etc.\n", - "now.\"\n", - "rate\n", - "it.\n", - "4.\n", - "oliM-rvi-\n", - "notwithstanding.\"\n", - "High.\n", - "back.\n", - "habit.\n", - "taller.\n", - "candidate.\n", - "doctors.\"\n", - "just.\n", - "work.\n", - "hi-\n", - ".\n", - "We\n", - "fair\n", - "cards.\n", - "banks\n", - "X.\n", - "terrible.\n", - "-\n", - "tlon.\n", - "ago,\n", - "city.\n", - "water.\n", - "when\n", - "anybody.\n", - "cease.\n", - "telescopes\n", - "said\n", - "circumstance,\n", - "coun­\n", - "s\n", - "Iceland\n", - "advancing.\n", - "Prineviile.\n", - "remedy.\n", - "say.\n", - "on-\n", - "Mr.\n", - "carried.\n", - "paid,\n", - "future.\"\n", - "graced.\n", - "truth.\n", - "yeat--\n", - "?”\n", - "preserves\n", - "such\n", - "home.\n", - "d\n", - "waa\n", - "phy.\"\n", - "Times:\n", - "in-\n", - "footprints\n", - "funeral.\n", - "present.\n", - "crossing\n", - "robber,\n", - "diamonds\n", - "sort.\n", - "friends\n", - ":\n", - "Times.\n", - "Post.\n", - "societies.\n", - "hunger,\n", - "avarice.\n", - "it.\n", - "right*\n", - "e\n", - "taken.\n", - "\"liome-tralner-\n", - "again.\n", - "conception.\n", - "good.\n", - "Science.\n", - "feeling.\n", - "remarked,\n", - "Smith.\n", - "present.\n", - "ingredients.\n", - "pre-\n", - "ju\n", - "sold:\n", - "580pm.\n", - "ten-\n", - "somo-\n", - "lf\n", - "shall\n", - "Mexican\n", - "life.\n", - "cure.\n", - "married\n", - "high-cla-\n", - "day.\n", - "a\n", - ",\n", - "me.\n", - "lino\n", - "spot.\n", - "at-\n", - "by\n", - "aeason.\n", - "accepted.\n", - "justice.\n", - "The\n", - "nations.\n", - "away.\n", - "to.\n", - "Cal.,\n", - "country.\n", - "confessed.,\n", - "issue.\n", - "end.\n", - "al\n", - "Christ.\n", - "Auguit\n", - "nations.\n", - "c\n", - "trackage.\n", - "mine-\n", - "desperado.\n", - "won.\n", - "thereby.\n", - "ravenous.\n", - "shot.\n", - "newspapers\n", - "b62.\n", - "suf-\n", - "are\n", - "friends\n", - "defrauded.\n", - "contracted.\n", - "lesson.\n", - "remunerative\n", - "Kempls:\n", - "institutions.\n", - "icate.\n", - "y\n", - "pleasant.'\n", - "continuously.\n", - "tell.\n", - "goest.\n", - "producer.\n", - "thank\n", - "•\n", - "States.\n", - "cured.\"\n", - "Navy\n", - "M.\n", - "asser\n", - "Judgment\n", - "course.\n", - "1862.\n", - "township\n", - "case.\n", - ".\n", - "y\n", - "a\n", - "made.\n", - "made.\n", - "n\n", - "acres,\"\n", - "them.\n", - "wonderlngs.\n", - "mind\n", - "collts-tlv-\n", - "ter.\n", - "nature.\n", - "tion.\n", - "improved.\n", - "the\n", - "\"\n", - "to-w-\n", - "Paper,\n", - "broidery.\n", - "the\n", - "X'llt.\n", - "an\n", - "often.\n", - "that\n", - "of\n", - "cemetery.\n", - "depression\n", - "\"\n", - "it.\n", - "hotel.\n", - "IS\n", - "a\n", - "m.\n", - "render.\n", - "back-\n", - "themselves.\n", - "c-\n", - "a\n", - "*ore’\n", - "companies.\n", - "the\n", - "”\n", - "manner.\n", - "tha\n", - "Granger.\n", - "difficol'7>\n", - "Lake\n", - "placed\n", - "lf\n", - "for\n", - "lllll\n", - ".\n", - "learn.\n", - "so.\n", - "the\n", - "-\n", - "again.\n", - "sec-\n", - "tlmo.\n", - "isfied.”\n", - "'\n", - "GrIL\n", - "government.\n", - "democrats,\n", - "pit.\n", - "even.\n", - "McCausland.\n", - "consideration.\n", - "fragments.\n", - "the-\n", - "Colombia\n", - "defendant.\n", - "like\n", - "outrage-\n", - "yours.\n", - "Adjourned.\n", - "1\n", - "church.\"\n", - ",\n", - "People.\n", - "STORY\n", - "revolver\n", - "contract\n", - "-\n", - "nttii-\n", - "Adjourned.\n", - "him.\"\n", - "dee.\"\n", - "mous.\n", - "bars\n", - "24:10.\n", - "inde-\n", - "road.\"\n", - "is.\n", - "a.\n", - "-\n", - "remedy.\n", - "in\n", - "s\n", - "match.”\n", - "Laughter.\n", - "that\n", - "aforeiaii’,\n", - "accident.\n", - "—Exchange.\n", - "and\n", - "whiter.\n", - "frc--\n", - "week.\n", - "party.\n", - "Collins\n", - "st\n", - "•\n", - ".\n", - "departure.\n", - ";\n", - "stock-raisin-\n", - "of\n", - "large.\n", - "completed\n", - "wrong.\n", - "mines.\n", - "outside.\n", - "lower\n", - "avoided.\n", - "me\n", - "means.\n", - "Hcdjuz,\n", - "tier?\n", - "towns—wht'je\n", - "quality.\n", - "cure.\n", - "which,\n", - "ever.\n", - "of\n", - "shall\n", - "\"\n", - "—Am,\n", - "bellovo\n", - "14.\n", - "im-\n", - "expenses.\n", - "called\n", - "43,4■\n", - "Herald.\n", - "rain,\n", - "mo-\n", - "money-mak-\n", - "celebration.\n", - "es\n", - "-\n", - "miles.\n", - "friends.\n", - "m.\n", - "their\n", - "dustry,\n", - "opposition.-\n", - "damage.\n", - "atten-\n", - "ance\n", - "bargain,?\n", - "17.\n", - "yourself.\n", - "no\n", - "M\n", - "allirmod.\n", - "yours,\n", - "the-\n", - "enjoyed.\n", - "Taj.\n", - "largely.\"\n", - "d\n", - "pl-\n", - "vore.\n", - "march.\n", - "houses.\n", - "onth■\n", - "nfitasfwl\n", - "1876.\n", - "44.\n", - "tense\n", - "times.\n", - "or\n", - "1878.\n", - "23\n", - "accommodations.\n", - "Tops-ha-\n", - "such\n", - "hats.\n", - "doc-\n", - "plant.\"\n", - "e\n", - "he\n", - "truth.\n", - "him\n", - "PUUH,\n", - "SOUTH\n", - "it.\"\n", - "-ce\n", - "at\n", - "are\n", - "granjes.\n", - "co-couuaei.\n", - "former.\n", - "of\n", - "it\n", - "prayed.\n", - "mentioned,\n", - "them.\n", - "1890.\n", - "m,\n", - "ery.\n", - "full.\n", - "votes.\n", - ">\n", - "flna\n", - "Inent.\n", - "revc-\n", - "75\n", - "lefL\n", - "repealed.\n", - "toQO.\n", - "camps.\n", - "herder.\n", - "Centre.\n", - "error.\n", - "mommy\n", - "School.\n", - "Week.\n", - "sea.\n", - "night.\n", - "troops.\n", - "work,and\n", - "serv't\n", - "production.\"\n", - "clBe.\n", - "majority.\n", - "holidays.\n", - "year.\n", - "d\n", - "cients.\n", - "&c.,\n", - "1893.\"\n", - "chargo\n", - "cattle.\"\n", - "Now\n", - ".\n", - "on.\"\n", - "executive\n", - "goods.\n", - "Asam.\n", - "property.\n", - "p\n", - "have\n", - "analgin*.\n", - "the\n", - "kc.\n", - "$700,000,000.\n", - "taklrg\n", - "it,\n", - "wan\n", - "ioint.\n", - "cause.\n", - ".!\n", - "f<\n", - "h\n", - "quantities.\n", - "services\n", - "when.\n", - "*t\n", - "tea,\n", - "ments.\n", - "ngovern-\n", - "state.\n", - "Post.\n", - "city.\n", - "In-\n", - "charges.\"\n", - "trip.\n", - "2.\n", - "a\n", - "tho\n", - "yet.\n", - "race.\n", - ",\n", - "postage.\n", - "thanksgiving\n", - "that;,\n", - "was\n", - "190D.\n", - "York.\n", - "abdomen,\n", - "conventions\n", - "back.\n", - "Affiux.\n", - "baken,\n", - ":\n", - "sites.\n", - "wa-\n", - "io-\n", - "msgizines.\n", - "time.\n", - "alone,\n", - "craiy.\"\n", - "drugs.\n", - "liking.\n", - "season.\"\n", - "tower.\n", - ".\n", - "silenced.\n", - "from.\n", - "court;\n", - "Alaska.\"\n", - "effective.\n", - "neck.\n", - "not\n", - "paid.\n", - "short\n", - "Esq.\n", - "herein.\n", - "thorti.\n", - "l\n", - "void.\n", - "over.\n", - "difficulty.\n", - "it.\n", - "cities.\n", - "Investigate\n", - "confusion.\n", - "insurance\n", - "otm.\n", - "t\n", - "since.\n", - ".\n", - "away.\n", - "bers.\n", - "Harding,\n", - "fear.”\n", - "persevering,\n", - "Philippines.\n", - "alley,\n", - "house.\n", - "ores.\n", - ".\n", - "easier\n", - "Wot\n", - "him.\n", - "treatment\n", - "tbe\n", - "flub;\n", - "patriotism.\n", - "cheap.\n", - "boa\n", - "....\n", - "for\n", - "cordingly.\n", - "1\n", - "lAdtfree..\n", - "land.\n", - "a,\n", - "glory.\n", - "ao\n", - "aloe\n", - "grain.\n", - "him.\n", - "demanded.\n", - "John\n", - "cash.\n", - ".\n", - "Bussians.\n", - "expenditures.\n", - "forenoon.\n", - "weak.\n", - "sarcastically.\n", - "state.\n", - "of\n", - "N.Y.\n", - "Ms.\n", - "s\n", - "on\n", - "him\n", - "lake.\n", - "per-\n", - "cial.\n", - "Lorraine.\n", - "th\n", - "husband\n", - "\"Deaconesses.\"\n", - "—I'hicai/o\n", - "$75.\n", - "o\n", - "it.\n", - "!\n", - "them.\n", - "satisfied.\n", - "rifle.\n", - "trfct.\n", - "liberal\n", - "rules!\n", - "printer.\n", - "spring.\n", - "four\n", - "are\n", - "that\n", - "fair.\n", - "governor-elect.\n", - "advantages.\n", - "specifications\n", - "localities.\n", - "midnight.\n", - "explanatory\n", - "1852.\n", - "ino.\n", - "presents.\n", - "wharfage.\n", - "rs\n", - "cu\n", - "of\n", - "15\n", - "paper.\n", - "it.\n", - "experience.\n", - "vacation.\n", - "are.\"\n", - "goods.\n", - "home\n", - "bear-tra-\n", - "themselves,\n", - "besides.\n", - "Labor..\n", - "I\n", - "hits.\n", - "admitted\n", - "full\n", - "trT\"\"\n", - "to\n", - ">\n", - "rough.\"\n", - "verse\n", - "approve.\n", - "intheh11.\n", - "Pickett.\n", - "on\n", - "doubtful\n", - "revive^ller,\n", - "pan\n", - "busi¬\n", - "remarks:\n", - "fail\n", - "hand.\n", - "thcui.\n", - "ago.\n", - "g\n", - "defeated.\n", - "Wo\n", - "-\n", - "investigating\n", - "pounds.\n", - "!\n", - "commissioners.\n", - "inten-\n", - "Rankin\n", - "single-dro-\n", - "Shorten.\n", - "five\n", - "seven\n", - "schools.\n", - "stored.\n", - "conveyed.\n", - "to\n", - "mission\n", - "up.\"\n", - "timbers.\n", - "remedy.\n", - "a\n", - "copy.\n", - "followsi\n", - "Home.\n", - "beginning.\n", - "\"Darling,\n", - "Comuion-\n", - "above.\n", - "Taxes\n", - "gloom\n", - "In.\n", - "writing.\"\n", - "day.\n", - "corrected.\"\n", - "Gregg,\n", - "tho\n", - "vested.\n", - "him.\n", - "havo\n", - "again.\n", - "$1,155,000.\n", - "Newark.\n", - "profess.\n", - "follows:\n", - "having.\n", - "members?\n", - "of\n", - "It\n", - "Men.\"\n", - "copy.\n", - "truly,\n", - "integrity.\n", - "system.\n", - "follows:\n", - "now\n", - "army.\n", - ")\n", - "went.\n", - "expected.\n", - "2\n", - "center.\n", - "llrvi\n", - "Secretary.\n", - "tho\n", - "virtues.\n", - "1840.\n", - "God-Saviour.\n", - "God.\n", - "eat.\n", - "Oroville,\n", - "county.\n", - "live.\n", - "party,\n", - "globe\n", - "ar*\n", - "congregation.\n", - "bearing\n", - "!».\n", - "above\n", - "ideas\n", - "prayed.\n", - "Bluffs.\n", - "investigation.\n", - "traveler.\n", - "oil\n", - "fines.\n", - "oiteus.\n", - "disposal.\n", - "fol-\n", - ";\n", - "w&y\n", - "blood\n", - "dance\n", - "protection.\n", - "Californian.\n", - "almost\n", - "them.\n", - "rot-\n", - ".Mr.\n", - "pag.--\n", - "purchaser.\n", - "each\n", - "penalties.\n", - "dismissed.\n", - "denoted\n", - "excuse.\n", - "o\n", - "Herald.\n", - "Gentleman.\n", - "success.\n", - "it.\n", - "methods.\n", - "factory.\n", - "St.\n", - "tion.,\n", - "an\n", - "P.\n", - "cause.\n", - "Septembers.\n", - "log.\n", - "the\n", - "70^\n", - "them.\n", - "ninth.\n", - "ation.\n", - "beat.\n", - "situations.\n", - "for\n", - "oi\n", - "Ghost.\"\n", - "with\n", - "Cross.\n", - "troops.\n", - "avoided.\n", - ",\n", - "sacrifice\n", - "former.\n", - "\"Othello\"\n", - "Philadelphia.\n", - "power.\n", - "would\n", - "gb\n", - "y\n", - ".\n", - "chief.\n", - "Nicholaieviteh.\n", - "tho\n", - "tiling.\n", - "brnshlng.\n", - "city.\"\n", - "consum\n", - "reform.\n", - "inteuiioti.\n", - "bricklaying.\n", - "forfeited.\n", - "mate.\n", - "them.\n", - "elected\n", - "Oroville,\n", - "terms.\n", - "e\n", - "order.\n", - "old\n", - "favor.\n", - "them.\n", - "out.\n", - "consumption.\n", - "a.\n", - "e\n", - "°r\n", - "trade.\"\n", - "$1.\n", - "just.\n", - "two\n", - "charge.\n", - "adopted.\n", - "nation.\n", - "tion.”\n", - "lew\n", - "beginning.\n", - "worth.\n", - "Wood.\n", - "business.\n", - "IbOo.\n", - "l.\n", - "Iowa..\n", - "destruction\n", - "to\n", - "Interment\n", - "a\n", - "Country,\n", - "says.\n", - "these\n", - "field.\n", - "on.\n", - ",\n", - "water.\n", - "gether.\n", - "Trustees.\n", - "office.\n", - "said:\n", - "arrived.-\n", - "wares?\"\n", - "operations.\n", - "going.\n", - "a\n", - "thl\n", - "consideration.\n", - "r.\n", - "are!\n", - "done.\n", - "rules.\n", - "specimen,\"\n", - "altetnoon.\n", - "\"\n", - "aud\n", - "mentioned.\n", - "How?\n", - "cured.\"\n", - "mention.\n", - "Tur\n", - "shrieking,\n", - "And\n", - "city.\n", - "to-day.\n", - "politics,\n", - "an\n", - "Newark.\n", - "ordeal.\n", - "primary.\n", - "men,\"\n", - "d\n", - "them.\n", - "...\n", - "wit:\n", - "?\n", - "mind\n", - "plasters.\n", - "there.\n", - "resource*.\n", - "Yours,\n", - "him.\n", - "district\n", - "ican.\n", - "west.'''\n", - "opportunities\n", - "infinitum.\n", - "by.”\n", - "devices.\n", - "health.\n", - "course,\n", - "voters.\n", - "nlfi'clnd.\n", - "a\n", - "of\n", - "ss\n", - "Passaic\n", - "Express.\n", - "he\n", - "wh\n", - "prisonment.\n", - "Trustees.\n", - "j,\n", - "appurtenances.\n", - "p-\n", - "color.\n", - "tentiary.\n", - "commerce.\n", - "named.\n", - "occurred.\n", - "behalf.\n", - "Science.\n", - ".\n", - "branch\n", - "as\n", - "attended.\n", - "legislation.\n", - "\"effective.\"\n", - "sil\n", - "I\n", - "nt,\n", - "luth.\n", - "friends\n", - "Hannah.\n", - "Department.\n", - "banks.\n", - ",0c\n", - "him*\n", - "Edith,\n", - "law.\n", - "the\n", - "their\n", - "y,\n", - "harm.\n", - "ins.'\n", - "nominated.\n", - "answer.\n", - "degree.\n", - "drifted.\n", - "Ostf-ta-\n", - "nationalities.\n", - "n.e\n", - "n»A>..J!.\n", - "sides.\n", - "$2,250,000.\n", - "lodge.\n", - "described:\n", - "Fos\n", - "M.\n", - "state\n", - "e\n", - "the\n", - "or\n", - "--\n", - "da\n", - "indigestion.\n", - "continued:\n", - "K\"venuuonl.\n", - "court.\n", - "tion.\n", - "Kllevlts\n", - "man.\n", - "told.\n", - "numb'--\n", - "were\n", - "carpet-baggers.-\n", - "quality.\n", - "on\n", - "events\n", - "more.\n", - "'twos.\"\n", - "Idler.\n", - "roots.\n", - "ginning.\n", - "world\n", - "in\n", - "remedv.\n", - ">.\n", - "Bi.ll.\n", - "Presldeot-\n", - "died?\"\n", - "supervisors.\n", - "Cascades.\n", - "twelve\n", - "Atlantic.\n", - "Glentworth.\n", - "nationality.\n", - "up?\n", - "r(\n", - "one.\n", - "to-morrow.\n", - "News.\n", - "vance.\n", - "e\n", - "tustc.\n", - "Herald.\n", - "show.\n", - "the\n", - "de-\n", - "Bottomley.\n", - "estate.\n", - "bargain.\n", - "He\n", - "list.\n", - "strikes,\n", - "No.\"\n", - "beginning.\n", - "mio-\n", - "did.\n", - "the\n", - "1860.\n", - "right.\n", - "renown.\n", - "co.,\n", - "relense.\n", - "Micbigau.\n", - "5\n", - "country.\n", - "tor\n", - "house.\"\n", - "avenue.\n", - "conference.\n", - "o\n", - "undischarged.\n", - "gamuts.\n", - "time.\n", - "supporters.\n", - "copy.\n", - "conipauies.\n", - "minutet.\"\n", - "Number.\n", - "the\n", - "again,\n", - "lea.\n", - "1:474.\n", - "regretted\n", - "thence\n", - "district.\n", - "cation.\n", - "d\n", - "statements.\n", - "Lyra,\n", - "short-live-\n", - "f\n", - "s-.\n", - "a-\n", - "cats.\n", - "d,\n", - "product,\n", - "work.\n", - "absurdity\n", - "SOft.\n", - "city.\n", - "Times.\n", - "the\n", - "hut.\n", - "Veterans.\n", - "income!\n", - "y\n", - "year.\n", - "minister.\n", - "filing.\n", - "Press.\n", - "them.\n", - "sad\n", - "S.\n", - "to!d\n", - "known.\n", - "to.\"\n", - "often\n", - "tbe\n", - "dealers.\n", - "Jcnn\n", - "less.\n", - "pression.\n", - "expected.\n", - "members\n", - "sustain\n", - "r-\n", - "claimants.\n", - "P.\n", - "order.\n", - "had\n", - "'\n", - "seventies.\n", - "authorizing\n", - "notice.\n", - "self-government.\n", - "Street.\n", - "redressed.\n", - ",\n", - "Appleby\n", - "seed.\n", - "places.\n", - "wisdom.\n", - "removed.\n", - "kinds\n", - "Virginia.\n", - "7a9%c.\n", - "away.\n", - "tin\n", - "Center.\n", - "clerk.\n", - "loving,\n", - "y\n", - "now.\n", - "it\n", - "continued.\n", - "tli\n", - "vision,\n", - "Int.\n", - "Investigations,\n", - "the\n", - "nillllu.\"\n", - "b-\n", - "pioneer.\n", - "ed.\n", - "Tobeauroliolathoh\n", - "pjndlcutc.\n", - "others.\n", - "cusable.\n", - "other\n", - "con*\n", - "stomachs.\"\n", - "borer.\n", - "ships.\n", - "are\n", - "youngsters.\n", - "the-\n", - "Record.\n", - "Tiki\n", - "on.—Southron\n", - "complaint.\n", - "\\\\\n", - "her.\n", - "remedy.\n", - "neighbors.\n", - "own\n", - "children.\"\n", - "Tobolsk.\n", - "our\n", - "U.I\n", - "from.\n", - "centuries.\n", - "pigeons.\n", - "Brown's.\n", - "emetic.\n", - "plays.\n", - "disclosed.\n", - "itr\n", - "money.\n", - "ways.\n", - "discovered,\n", - "man\n", - "Iniallmetita\n", - "calls.\n", - "election.\n", - "exclaiming.\n", - "BarrlngtonBrown\n", - "said,\n", - "was\n", - "nityasafeme’\n", - "\"Gee,\n", - "lazy.\n", - "story.\n", - "era.\"\n", - "again,\n", - "Harrl-\n", - "Sim.\n", - "walls.\n", - "Jr.,\n", - "scandal.\n", - "ers\n", - "ringing.\n", - "police.\n", - "ones.\n", - "the\n", - "er\n", - "nence.\n", - "them.\n", - "tr\n", - "trap.\n", - "people.\n", - ".\n", - "hlr\n", - "ployment.\n", - "hills.\n", - "•\n", - "aforesaid\n", - "E.VzS.E.\n", - "1917.\n", - "it.\n", - "Thurmont\n", - "harm.\n", - "afford.\n", - "life.\n", - "moments.\n", - "home!”\n", - "purpose.\"\n", - "earth.\n", - "purpose.\n", - "In.\n", - "While\n", - "perlenced.\n", - "salo.\n", - "of.\n", - "it.\n", - "on\n", - "minimum.\"\n", - "Burgess)\n", - "Territory.\n", - "Lothian\n", - "hand.\n", - "life.\n", - "yachtsmen\n", - "taken.\n", - "person.\n", - "M.\n", - "motions.\n", - "rivalry\n", - ",\n", - "other.\n", - "ous\n", - "prosjierity.\n", - "Abbott.\n", - "gone.\n", - "claims.\n", - "morality.\n", - "de\n", - "Co.\n", - "foothills.\n", - "ult.\n", - "Proctor.\n", - "ii.\n", - "End.\n", - "heed.\n", - "T\n", - "Commissioner\n", - ".\n", - "others.\n", - "able-bodi-\n", - "lOWL.\n", - "filing.\n", - "transfer.\n", - "pig.\n", - "life\n", - "1,000.\n", - "g\n", - "crime.\n", - "brothers.\n", - "atTVVc\n", - "the\n", - "little.\n", - "dc-\n", - "s\n", - "he\n", - "cure.\n", - "fund.\n", - "of\n", - "manner\n", - "7\n", - "Post.\n", - "avoided.\n", - "it.\n", - "over.\n", - "Advocate\n", - "population.”\n", - "box.\n", - "Patriot.\n", - "in\n", - "failure.\n", - "backers.\n", - "affection.\n", - "Office.\n", - "have.\n", - "--\n", - "setting.\n", - "ailment.\n", - "long.\n", - "ta\n", - "justified.\n", - "Con\n", - "In\n", - "taW\n", - "yet.\n", - "fruit.\n", - "diseases.\n", - "exposed.\n", - "them.\n", - "servant,\n", - "geiu'ratlou.\n", - "Justice.\n", - "spotless,\n", - "air.\n", - "this.\"\n", - "silver\n", - "L................................\n", - "needed.\n", - "rate.\n", - "Intelligencer.\n", - "have\n", - "aostroyed.\n", - "the\n", - "ten\n", - "meanness.”\n", - "at\n", - "figures.\n", - "strawberries,\n", - "follow\n", - "system\n", - "remain\n", - "man.\n", - "prosecutions\n", - "registry-\n", - ".\"\n", - "Hevlew.\n", - "is\n", - "of\n", - "ge-\n", - "alone.\n", - "uaryL\n", - "the\n", - "him.\n", - "n\n", - "like\n", - "artists.\n", - "Americans.”\n", - "interest:\n", - "Delaware.\n", - "loa's.\n", - "fering.\n", - "winner.\"\n", - "look.\n", - "up.\n", - "California.\n", - ".\n", - "the\n", - "respect.\n", - "fed.\n", - "Lyman.\n", - "quiredi\n", - "-\n", - "tobacco\n", - "height\n", - "dinner.\"\n", - "tors.\n", - "basin.\n", - "$400.-\n", - "ad-\n", - "Chronicle.\n", - "practices.\n", - "rality.\n", - "-\n", - "people.\n", - "believe\n", - ".\n", - "decline.\n", - "Liverpool,\n", - "onlj\n", - "sort.\n", - "course.\n", - "Vou\n", - "In\n", - "why.\n", - "since.\n", - "plication.\n", - "Tesreau.\n", - "association.\n", - "-\n", - "parlors.\n", - "across\n", - "tippurunning.\n", - "the\n", - "Post.\n", - "1880.\n", - "lows:\n", - "party.\n", - "bejth.\n", - "now.\n", - "10CO\n", - "fulminations.\n", - "says:\n", - "ago.\n", - "employ.\n", - ".\n", - "haste.\n", - "dispensation.\n", - "might\n", - "the\n", - "seat\n", - "erful\n", - "convenience\n", - "tion.\n", - "possessed.\n", - "write.\n", - "lata.\n", - "was-no-\n", - "miner.\n", - "g\n", - "replied.\n", - "handsome.\n", - "th\"ir\n", - "enter.\n", - "i\n", - "that?\n", - "r.\n", - "both.\n", - "inherited\n", - "place.\n", - "g\n", - "ccptance.\n", - "hue.\n", - "nope\n", - "my\n", - "heaven.\n", - "feet.\n", - "Egj-pt-\n", - "from\n", - "another.\n", - "baggers.\n", - "75(380c.\n", - "to\n", - "Causey,\n", - "^\n", - "13.\n", - "most\n", - "destiny-\n", - "week\n", - "the\n", - "?\n", - "principal.\n", - "drugs.\n", - "14-th.\n", - "do.\n", - "tii\n", - "service.\n", - "tives.\n", - "follows\n", - "Darien.\n", - "Ui\n", - "y\n", - "Benwood.\n", - "greenbacks.\n", - "artillery.\n", - "died.\n", - "R«ncy.\n", - "tube.\n", - "once.'\n", - "high.\"\n", - "house.\"\n", - "pcuuiles\n", - "explorer.\n", - "Appeal.\n", - "development.\n", - "game?’’\n", - "Journal.\n", - "him.”\n", - "^y\n", - "the\n", - "Amerlrnn.\n", - "lands.\n", - "jl\n", - "HI\"\n", - "came\n", - "customers.\n", - "aforesaid.\n", - "them.\n", - "them.\n", - "Matthews.\n", - "Babbitt.\"\n", - "officers.\n", - "show.\n", - "pursuit.\n", - "there\n", - "it.\n", - "ever.\n", - "opin-\n", - "fnrnishlnj-\n", - "haud.\n", - "Solnt.\n", - "iu\n", - "payments.\n", - "wage*.\n", - "lawlessness.\n", - "removed.\n", - "census-taker-s\n", - "worst--\n", - "today',\n", - "captivity.\n", - "Ot.\n", - "sum.\n", - "is.\"\n", - "therefore,\n", - "up\n", - "dimensions.\n", - "spot.\n", - "of\n", - "Celluloid.\n", - "tion,\n", - "fools.\n", - "Saturdays.\n", - "attention.\n", - "something\n", - "forks.\n", - "amy.\n", - "?\n", - "har­\n", - "Medical\n", - "living\n", - "perienced.\n", - "er\n", - "hut\n", - "istrar*.\n", - "of\n", - "expo-\n", - "t\n", - "New-Cen-\n", - "nnuoyc-d-\n", - "dry\n", - "scandal.\"\n", - "pass.\n", - "A\n", - "make\n", - "President.\n", - "team.\n", - "ts\n", - "s\n", - "discussion.\n", - "MiningCompan-,\n", - "M»5.\n", - "‘ltion\n", - "be-\n", - "Asia\n", - "Thejr\n", - "him.\n", - "and\n", - "fine-looki-\n", - "Involved.\n", - "land.\n", - "weeks.\n", - "State.\n", - "missionary\n", - "heart.\n", - "yours,\n", - "liberty.\n", - "g\"\n", - "life-an-\n", - "anew.\n", - "line.!\n", - "ninjf.\n", - "day.\n", - "that\n", - "my\n", - "logs.'\n", - "Tress.\n", - "matter.\n", - "they\n", - "a\n", - "wires\n", - "months.\n", - "¬\n", - "SECRETS,\n", - "recover.\n", - "ifian.\n", - "admin-\n", - "isa\n", - "consequence.\n", - "50.\n", - ",\n", - "for\n", - "pray.\n", - "banner.\n", - "those\n", - "Beecher,\n", - "“\n", - "oi\n", - "ph.\"\n", - "says:\n", - ".\n", - "commissioners.\n", - "repair.\n", - "conjecture.\n", - "-\n", - "ley.\n", - "Duffy,\n", - "time.\n", - "evenings.\n", - "shot\n", - "boor.\n", - "operate\n", - "\"gibbet,\"\n", - "ing.\n", - "murdoror.\n", - "theud-\n", - "°l\n", - "nomination.\n", - "unflinchingly.\n", - "suddenness.\n", - "aimnch\n", - "cup.\n", - "bnJge.\n", - "road.\n", - "clients.\n", - "division.\n", - "hat\n", - "-\n", - "tirm.\n", - "Age,\n", - "tunate,\n", - "ju-Jifice.\n", - "be\n", - "particular.\n", - "enterprise.\n", - "tl..-\n", - "Bhortatl.\n", - "influence.\n", - "preserves.\n", - "-\n", - "vlllago\n", - "respon-dbl-\n", - "survivors.\n", - "8,000.\n", - "position.\n", - "Railroad-:\n", - "Khln.\n", - "printed.\n", - "State,\n", - "eat.\n", - "appertaining.\n", - "match.\n", - "Ii\n", - "mines.\n", - "or\n", - "them.”\n", - "afternoons.\n", - "directed\n", - "claims.\n", - "d,\n", - "D.\n", - "with\n", - "demanded.\n", - "Brown,\n", - "for\n", - "iron.\n", - "law.\n", - "out\n", - "interest.\n", - "44\n", - ".1\n", - "Donehoo.\n", - "onto-\n", - "health.\n", - "labor.'\n", - "o:\n", - "vehicles.\n", - "II...\n", - "solid.\n", - "term.\n", - "Germany.\n", - "telling\n", - "church.\n", - "inoffeusive.\n", - "performed,\n", - "folly.\n", - "outing.\n", - "gency.\n", - "within.\n", - "of\n", - "Hanker.\n", - "leaders.\n", - "be,\n", - "car.\n", - "then\n", - "made.\n", - "enco.\n", - "sunshine.\n", - "Hamilton.\n", - "Prospect\n", - "lay.\n", - "aueatinn\n", - "cneap.\n", - "member­\n", - "Catholio-ity-\n", - "peace.\n", - "air.\n", - "its\n", - "preachers.\"\n", - "license.\n", - "wick,\"\n", - "blight.\n", - "of\n", - "Duncan,\n", - "provisions.\"\n", - "in\n", - "on.\n", - "reclamation\n", - "Jnly,\n", - "made.\n", - "joke?\n", - "Agency.\n", - "clut-\n", - "-poli-\n", - "passage.\n", - "system.\n", - "|kj1.uo\n", - "troops.”\n", - "work.\n", - "estab­\n", - "disease.\n", - "y\n", - "vots.\n", - "yet.\n", - "regular\n", - "Ptttl\n", - "murder.\n", - "lap.\n", - "Science.\n", - "gentleman.\n", - "tions.\n", - "anywhere.\n", - "therf-\n", - "Congress.\n", - "reduction\n", - "of\n", - "Senators.\n", - "C;\n", - "garchy.\n", - "lunatics.\n", - "sale.\n", - "work.\n", - "1860.\n", - "Toledo.\n", - "-I\n", - "her.\n", - "Magazine.\n", - "mind.\n", - "State.\n", - "McClurc.\n", - "view.\n", - "It.\n", - "laid\n", - "to-day.\n", - "coals.\"\n", - "health.”\n", - "forfeited.\n", - "people.\n", - "sioner.\n", - "tion.\n", - "Tehtiacan.\n", - "jear.\n", - "provisions.\n", - "prostration.\n", - "mented\n", - "neau.\n", - "rates.\n", - "TTVe-\n", - "properties.”\n", - "is\n", - "favo.-\n", - "Philadelphia.\n", - "ing\n", - "prices.\n", - "accident.\n", - "send\n", - "music.\n", - "$1200.\n", - "him.\n", - "appetites\n", - "shoes.\n", - "existence.\n", - "beginning.\n", - "fact.\n", - "food.\n", - "transportation.\n", - "forever!\n", - "otTer.\n", - "Plalnville-Farmingto-\n", - "3--\n", - "collision.\n", - "sentiment,\n", - "At^et\n", - "anoth\n", - "direction.\n", - "reduction,\n", - "service.\n", - "America,-\n", - "health.\n", - "sums.\n", - "good.\"\n", - "Plant\n", - "thing.\n", - "News.\n", - "no\n", - "Washington.\n", - "Times.\n", - "Democratic\n", - "prevailed.\n", - "agreement.\n", - "specially\n", - "allowed.\n", - "we\n", - "treated.\n", - "d«ir\n", - "connection.\n", - "irvnn.\n", - "offer.\n", - "ways.\n", - "st\n", - "summit.\n", - "men.\n", - "trade.\n", - "injured.\n", - "branch\n", - "case.\n", - "night.\n", - "ibute.\n", - "e\n", - "lf\n", - "power.\n", - "simple\n", - "ths\n", - "world.''\n", - "ini-\n", - "with\n", - ".\n", - "surface.\"\n", - "men.\"\n", - "Nome.\n", - "things.\"\n", - "Hie\n", - "Mountain.\n", - "1858.\n", - "wishes\n", - "damages.\n", - "settlement.\"\n", - "________\n", - "Y.\n", - "sermons.\n", - "costs.\n", - "contain.\n", - "uncertain.\n", - "delay.\n", - "cheese\n", - "we\n", - "A\n", - "homes.\n", - "-\n", - "button.\n", - "fate.\n", - "n\n", - "district.\n", - "up.\n", - "said:\n", - "other.\n", - "latter.\n", - "lat*\n", - "de-\n", - "c\n", - "off.\"\n", - "1874.\"\n", - "t.\n", - "on\n", - "the-\n", - "proposed.\n", - "#\n", - "distance.\n", - "will.\n", - "rooms.\n", - "paid\n", - "suggest-th-\n", - "oligarchy.\n", - "sea.\n", - "rt\n", - "eleven\n", - "tion.\n", - "recent\n", - "information.\n", - "otllcliils.\n", - "shock.\n", - "home.\n", - "Kensington.\n", - "heart.\n", - "by\n", - "place.\n", - "Equalization.\n", - "Parker.\n", - "come.\n", - "upon.\n", - "yours,\n", - "placed\n", - "and\n", - "elections.\n", - "tices.\n", - "Louis,\n", - "supply.\n", - "Taft.\n", - "days.\n", - "Tuttle\n", - "common\n", - "voices\n", - ".\n", - "avail.\n", - "siasm.\n", - "<\n", - "sup*\n", - "navies\n", - "things.\"\n", - "safety.\n", - "my-\n", - "-\n", - "follows:\n", - "rub.\n", - "effected.\n", - "nation.”\n", - "her.\n", - "street,\n", - "ena-\n", - "date.\n", - "nnd\n", - "Informants.\n", - "Porter,\n", - "mad.\n", - "vengeance.\n", - "l._\n", - "you.\n", - "railroad\n", - "iuy\n", - "considered.\n", - "storm.\n", - "qualities.\n", - "Monday.\n", - "let-\n", - "and\n", - "Newark.\n", - "at-\n", - "vote.\n", - "all.\n", - "industries.\n", - "ama.\n", - "nations,\n", - "e\n", - "city.\n", - "the.GMeeï\n", - "don't\n", - "w\n", - "started\n", - "effected.\n", - "Ohio.\n", - "systun.\n", - "i\n", - "armies\n", - "and\n", - "family.\n", - "Is\n", - "sleep.\n", - "Bazar.\n", - "the\n", - "Lodge,\n", - "consumption.\n", - "no\n", - "St\n", - "puzzler.\n", - "repaired.\n", - "\"\n", - "organs\n", - "long\n", - "time.\n", - "important.\n", - "terms.\n", - "n-\n", - "can\n", - "\\\\\n", - "his\n", - "som--\n", - "street.\n", - "a\n", - "law.\n", - ";\n", - "pa;\n", - "\"setback.\"\n", - "waters.\n", - "naked\n", - "unskilled\n", - "generally.\n", - "bier.\"\n", - "Us.”\n", - "land\n", - "rroiirl«torofferi«.\n", - "values.\n", - "left.\n", - "3\n", - "file.\n", - "observed.''\n", - "ring.\n", - "president.\n", - "i3.\n", - "them.\n", - "wbicn\n", - "year.\n", - "us.\n", - "phalanx\n", - "bow.\n", - "j\n", - "afterward.\"\n", - "Liens.\n", - "Congress.\n", - "yore.\n", - ":\n", - "HUH.\n", - "ger.\n", - "intellectual\n", - "onions.\n", - "possible.'\n", - "be.\n", - "»SS».\n", - "\"\n", - "Titus.\n", - "conspirators.\n", - "toys,\n", - "de-\n", - "7414\n" - ] - } - ], - "source": [ - "\n", - "with lzma.open(test_file, 'rt') as file:\n", - " predict_words = []\n", - " results = []\n", - " for line in file:\n", - "# print(line)\n", - " line = preprocess(line) #get only relevant\n", - " split = line.split('\\t')\n", - " print(get_last_word(split[0]))\n", - " predict_words.append(get_last_word(split[0])) #get_first_word(split[1])\n", - " print(len(predict_words))\n", - " vocab = train_dataset.vocab\n", - " for presc_word in predict_words:\n", - " results.append(dict(get_values_from_model(presc_word, model, vocab, k=k)))\n", - " \n", - " \n", - " with open(out_file, 'w') as outfile:\n", - " for elem in results:\n", - " \n", - " outfile.write(gonito_format(elem))\n", - "\n" - ] - }, - { - "cell_type": "code", - "execution_count": null, - "metadata": {}, - "outputs": [], - "source": [] - } - ], - "metadata": { - "kernelspec": { - "display_name": "Python 3 (ipykernel)", - "language": "python", - "name": "python3" - }, - "language_info": { - "codemirror_mode": { - "name": "ipython", - "version": 3 - }, - "file_extension": ".py", - "mimetype": "text/x-python", - "name": "python", - "nbconvert_exporter": "python", - "pygments_lexer": "ipython3", - "version": "3.10.6" - }, - "org": null - }, - "nbformat": 4, - "nbformat_minor": 1 -}