Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- Initialized
- ('Loss :', 9.2027139663696289)
- ('Batch_input :', array([9971, 9972, 9974, 9975, 9976, 9980, 9981, 9982, 9983, 9984, 9986,
- 9987, 9988, 9989, 9991, 9992, 9993, 9994, 9995, 9996, 9997, 9998,
- 9999, 2, 9256, 1, 3, 72, 393, 33, 2133, 0, 146,
- 19, 6, 9207, 276, 407, 3, 2, 23, 1, 13, 141,
- 4, 1, 5465, 0, 3081, 1596, 96, 2, 7682, 1, 3,
- 72, 393, 8, 337, 141, 4, 2477, 657, 2170], dtype=int32))
- ('Batch_labels :', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- Average loss at step 0: 0.092027 learning rate: 1.000000
- ('Label: ', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- ('Predicted:', array([[-0.36508381, -0.25612 , -0.26035795, ..., -0.42688274,
- -0.4078168 , -0.36345699],
- [-0.46035308, -0.27282876, -0.34078932, ..., -0.50623679,
- -0.47014061, -0.43237451],
- [-0.14694197, -0.07506246, -0.10392818, ..., -0.1128526 ,
- -0.12404554, -0.13495158],
- ...,
- [-0.07286638, -0.04560997, -0.05932444, ..., -0.08352474,
- -0.07679331, -0.07829094],
- [-0.13576414, -0.07057529, -0.1017022 , ..., -0.11192483,
- -0.14713599, -0.11757012],
- [-0.05446544, -0.02738103, -0.03401792, ..., -0.05073205,
- -0.03746928, -0.05750648]], dtype=float32))
- ================================================================================
- [[ 0. 0. 0. ..., 0. 0. 0.]]
- 8605
- ('f', u'altman')
- ('as', u'altman')
- ('feed', array([8605]))
- ('Sentence :', u'altman rake years regatta memotec pierre <unk> nonexecutive as will <eos> ssangyong director nahb group the cluett rubens snack-food fromstein calloway and memotec a board years regatta publishing fields rake group group rake cluett ssangyong pierre calloway memotec gitano gold rubens as as director sim is publishing gitano punts join <unk> and a old punts years memotec a rake is guterman cluett ssangyong will berlitz nahb <eos> of group join <unk> board join and pierre consolidated board cluett dutch gold as ipo ssangyong guterman a kia will dutch and director centrust consolidated rudolph guterman guterman cluett years n.v. old board rubens ')
- ================================================================================
- ('Loss :', 496.78199882507323)
- ('Batch_input :', array([4115, 5, 14, 45, 55, 3, 72, 195, 1244, 220, 2,
- 0, 3150, 7426, 1, 13, 4052, 1, 496, 14, 6885, 0,
- 1, 22, 113, 2652, 8068, 5, 14, 2474, 5250, 10, 464,
- 52, 3004, 466, 1244, 15, 2, 1, 80, 0, 167, 4,
- 35, 2645, 1, 65, 10, 558, 6092, 3574, 1898, 666, 1,
- 7, 27, 1, 4241, 6036, 7, 3, 2, 366], dtype=int32))
- ('Batch_labels :', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 1., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- Average loss at step 100: 4.967820 learning rate: 1.000000
- ('Label: ', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 1., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- ('Predicted:', array([[ 4.41551352e+00, 9.98007679e+00, 1.75690575e+01, ...,
- 6.83443546e+00, -2.30797195e+00, 1.73750782e+00],
- [ 1.26826172e+01, 5.96618652e-03, 1.18247871e+01, ...,
- -3.70885038e+00, -8.55356884e+00, -9.16959190e+00],
- [ 1.44652233e+01, 5.12977028e+00, 9.42045784e+00, ...,
- 1.39444172e+00, 1.95213389e+00, -4.00810099e+00],
- ...,
- [ 2.93052626e+00, 9.41266441e+00, 1.79130135e+01, ...,
- 4.24245834e+00, -1.46551771e+01, -3.35697136e+01],
- [ 2.48945675e+01, 2.32091904e+01, 2.47276134e+01, ...,
- -6.39845896e+00, -2.66628218e+00, -4.59843445e+00],
- [ 1.34414902e+01, 4.80197811e+00, 1.89214745e+01, ...,
- -5.91268682e+00, -8.80736637e+00, -6.49542713e+00]], dtype=float32))
- ================================================================================
- [[ 0. 0. 0. ..., 0. 0. 0.]]
- 3619
- ('f', u'officially')
- ('as', u'officially')
- ('feed', array([3619]))
- ('Sentence :', u'officially <unk> to <eos> filters ago cigarettes is that cigarette stopped to <eos> researchers <unk> to <eos> filters ago cigarettes asbestos the filters ago cigarettes asbestos the filters ago cigarettes is that cigarette up the <eos> researchers to <eos> researchers <unk> to <eos> filters ago cigarettes asbestos the filters ago cigarettes asbestos <eos> filters ago cigarettes asbestos the filters ago cigarettes is that cigarette up the <eos> researchers <unk> to <eos> researchers <unk> to <eos> filters ago cigarettes asbestos of percentage years the the the <eos> researchers <unk> to <eos> filters ago cigarettes asbestos the filters ago cigarettes asbestos the filters ')
- ================================================================================
- ('Loss :', 581.9651535148621)
- ('Batch_input :', array([ 39, 13, 31, 393, 1366, 2, 64, 275, 1921, 43, 72,
- 195, 157, 1442, 2395, 4, 3150, 718, 106, 5791, 1304, 2,
- 83, 13, 102, 3150, 7, 228, 189, 99, 2, 1400, 1,
- 1415, 0, 1244, 56, 4375, 0, 431, 46, 2647, 4, 106,
- 319, 16, 8421, 4, 0, 6092, 3574, 2, 64, 34, 102,
- 4213, 529, 16, 369, 2365, 26, 18, 645, 15], dtype=int32))
- ('Batch_labels :', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- Average loss at step 200: 5.819652 learning rate: 1.000000
- ('Label: ', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- ('Predicted:', array([[ 1.11496143e+01, 2.52034187e+00, 1.96437721e+01, ...,
- -1.50473619e+00, -8.62816811e+00, 1.24034882e-02],
- [ 7.78047562e+00, 1.48387527e+00, 1.96783009e+01, ...,
- -7.75834751e+00, -2.06198444e+01, -2.16614780e+01],
- [ 2.49285393e+01, 9.25475502e+00, 1.58422909e+01, ...,
- -6.17387116e-01, -7.49549580e+00, -3.64058924e+00],
- ...,
- [ 1.52376080e+01, 2.91777039e+00, 5.62801323e+01, ...,
- 1.27006912e+01, -6.71812582e+00, -1.61099453e+01],
- [ 1.31439075e+01, 2.54877434e+01, 2.91178341e+01, ...,
- 4.45230913e+00, -4.46597481e+00, -7.17428303e+00],
- [ 2.69684029e+00, 1.76576977e+01, 2.29938049e+01, ...,
- -7.32700920e+00, -1.84341221e+01, 1.32137108e+00]], dtype=float32))
- ================================================================================
- [[ 0. 0. 0. ..., 0. 0. 0.]]
- 1146
- ('f', u'improve')
- ('as', u'improve')
- ('feed', array([1146]))
- ('Sentence :', u'improve were smokers about <eos> the to <unk> questionable researchers any to of smokers about <eos> the questionable <eos> the to <unk> questionable researchers now any any any asbestos on smokers to questionable to <unk> questionable researchers now any any any nor asbestos on the questionable <eos> the questionable <eos> the questionable <eos> the questionable <eos> the questionable <eos> the questionable <eos> the questionable <eos> the questionable <eos> the questionable <eos> the questionable <eos> the questionable <eos> the questionable <eos> the questionable <eos> the questionable <eos> the questionable <eos> the questionable <eos> the questionable <eos> the questionable <eos> the questionable <eos> ')
- ================================================================================
- ('Loss :', 657.95355551540376)
- ('Batch_input :', array([ 15, 3150, 24, 263, 7, 253, 1451, 1351, 7, 423, 398,
- 11, 0, 6036, 7, 0, 266, 5278, 8, 3122, 22, 6,
- 769, 2155, 4, 1, 7, 3, 2, 20, 3, 5, 3,
- 3, 48, 6092, 3574, 22, 0, 6036, 46, 238, 0, 37,
- 15, 2, 211, 3, 945, 56, 1389, 1101, 22, 0, 5545,
- 3, 34, 2214, 45, 55, 132, 421, 0, 169], dtype=int32))
- ('Batch_labels :', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 1., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- Average loss at step 300: 6.579536 learning rate: 1.000000
- ('Label: ', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 1., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- ('Predicted:', array([[ 27.79744148, 23.37840462, 38.46372986, ..., -1.67567897,
- 1.32131767, -5.94825935],
- [ 18.3799057 , 16.12502098, 22.62628365, ..., -1.73421931,
- -7.86558867, -4.07993984],
- [ 21.18688202, 19.58936119, 37.37857056, ..., 2.72022462,
- 4.95587444, -3.86315012],
- ...,
- [ 57.44161606, 5.75437927, -3.88923836, ..., 9.1703558 ,
- -3.17029953, -12.2054348 ],
- [ 25.20575714, 21.54052734, 17.81223869, ..., -0.76101571,
- -12.47051144, -25.800457 ],
- [ 20.22606277, 29.38560104, 28.21241188, ..., -3.76599765,
- -12.07288361, -8.28661823]], dtype=float32))
- ================================================================================
- [[ 0. 0. 0. ..., 0. 0. 0.]]
- 1663
- ('f', u'recovery')
- ('as', u'recovery')
- ('feed', array([1663]))
- ('Sentence :', u'recovery replaced N N N modest the from and 1950s <eos> schools billion paper from from and 1950s <eos> schools billion paper from from and 1950s <eos> whether said N N N national in used a researchers risk the from and 1950s <eos> whether said N N N national in used <eos> schools billion paper from from and 1950s <eos> whether said N N N national led said N N N modest with type and N harvard N N N harvard <eos> schools billion paper from from and 1950s <eos> whether said N N N national in used a researchers early ')
- ================================================================================
- ('Loss :', 715.17567856243249)
- ('Batch_input :', array([1244, 15, 2, 0, 1, 158, 13, 6, 4196, 1883, 211,
- 150, 4, 505, 56, 938, 1, 4464, 15, 978, 1, 2,
- 0, 760, 4, 5172, 1015, 2786, 211, 0, 431, 18, 0,
- 379, 1, 1713, 398, 2025, 1310, 5, 25, 0, 1590, 11,
- 106, 3150, 431, 4375, 7, 650, 6080, 672, 28, 15, 2,
- 0, 402, 41, 13, 1023, 17, 1, 82, 1], dtype=int32))
- ('Batch_labels :', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 1., ..., 0., 0., 0.],
- [ 1., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 1., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- Average loss at step 400: 7.151757 learning rate: 1.000000
- ('Label: ', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 1., ..., 0., 0., 0.],
- [ 1., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 1., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- ('Predicted:', array([[ 21.06395531, 41.48188782, 14.76991463, ..., 0.90518486,
- -9.15206528, -18.00244331],
- [ 27.82922745, 31.63118172, 58.02881622, ..., 7.33755016,
- 12.11525059, -2.95855141],
- [ 66.02002716, 48.96756744, -5.2470417 , ..., -11.50299168,
- -2.58890057, -9.010849 ],
- ...,
- [ -3.29854012, 25.49062538, 47.43249512, ..., -2.6510427 ,
- -16.02886581, -19.52365685],
- [ 42.7742424 , 38.42037201, 23.18925476, ..., 5.25420237,
- -9.42901707, -7.64269066],
- [ 0.7444973 , 25.17520905, 47.63552094, ..., -2.6350472 ,
- -14.94020557, -17.693964 ]], dtype=float32))
- ================================================================================
- [[ 0. 0. 0. ..., 0. 0. 0.]]
- 7237
- ('f', u'abbie')
- ('as', u'abbie')
- ('feed', array([7237]))
- ('Sentence :', u'abbie among industrialized study deaths cancer said <unk> workers workers N of lung at including among rate <unk> the for said lung N of lung mass. paper said than <unk> workers deaths cancer N cancer said <unk> workers workers N of lung at including <unk> researchers among cancer recently workers he lung <unk> workers workers of lung N among at <unk> research factory any western whether sold deaths cancer said <unk> workers workers of cancer said N he workers he lung <unk> workers <unk> deaths cancer said <unk> workers workers <unk> workers workers <unk> workers workers <unk> workers workers N of ')
- ================================================================================
- ('Loss :', 839.59790327381404)
- ('Batch_input :', array([ 944, 4, 3150, 1, 551, 7, 90, 1647, 8, 61, 1636,
- 978, 1, 15, 2, 0, 53, 13, 54, 4, 0, 261,
- 6080, 1618, 10, 175, 32, 34, 6, 206, 739, 4, 3119,
- 11, 0, 5533, 1, 5394, 88, 19, 1, 10, 26, 6349,
- 19, 1, 214, 5, 1, 2239, 1, 6, 2035, 4, 1,
- 18, 0, 596, 4, 5042, 1488, 4, 3199, 2], dtype=int32))
- ('Batch_labels :', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 1., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 1., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- Average loss at step 500: 8.395979 learning rate: 1.000000
- ('Label: ', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 1., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 1., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- ('Predicted:', array([[ 24.61105537, 40.75873184, 14.10590649, ..., -4.77243757,
- -8.44593239, -7.23684168],
- [ 50.75180054, 54.97891617, 18.49082565, ..., 2.8976028 ,
- -6.811059 , -4.75290489],
- [ 50.92717743, 73.37365723, -29.40761757, ..., 7.97057295,
- -21.74515915, -15.22813034],
- ...,
- [ 81.55072021, 60.01564026, 11.3656292 , ..., -3.01502323,
- -15.37229347, -6.71199465],
- [ 56.06694794, 64.74343872, -11.93397331, ..., -16.31550217,
- -31.67933273, -9.8679533 ],
- [ 70.32337952, 55.06775665, -3.79033089, ..., -5.36194277,
- -1.45014143, -4.41785622]], dtype=float32))
- ================================================================================
- [[ 0. 0. 0. ..., 0. 0. 0.]]
- 825
- ('f', u'wants')
- ('as', u'wants')
- ('feed', array([825]))
- ('Sentence :', u'wants is smooth the among have among have among a among among have among have among have among have among have among have among have among have among have among have among have among a among among have among have among have among a among among have among a among among a among among have among a among among have among have among have among have among have among a among among have among have among have among a among among have among have among have among a among among a among among a among among have among a among among ')
- ================================================================================
- ('Loss :', 933.77549325011478)
- ('Batch_input :', array([ 3, 511, 73, 1435, 1870, 4, 1, 3150, 33, 25, 9231,
- 2, 43, 3, 431, 18, 6, 2025, 10, 159, 398, 11,
- 0, 6092, 6036, 46, 4115, 5, 3150, 7, 0, 5278, 2,
- 874, 4, 0, 2025, 46, 782, 9673, 251, 0, 1, 24,
- 263, 2, 431, 3820, 380, 1, 1, 4, 0, 3021, 1681,
- 91, 6, 828, 1, 7070, 7, 3260, 8, 1], dtype=int32))
- ('Batch_labels :', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 1., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- Average loss at step 600: 9.337755 learning rate: 1.000000
- ('Label: ', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 1., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- ('Predicted:', array([[ 10.18803215, 47.19699097, 19.30631447, ..., -4.61812544,
- -9.48687172, -8.65189266],
- [ 34.51325989, 52.13234329, 38.19073486, ..., 2.01813602,
- -3.91805673, -1.44879436],
- [ 56.49404144, 67.29128265, 46.57570267, ..., 4.87337399,
- -0.90795827, -0.1669569 ],
- ...,
- [ 34.7301178 , 46.05511093, 28.7182045 , ..., 0.60070062,
- -10.29673481, -5.18640995],
- [ 49.51641083, 63.66670227, 25.37597084, ..., -1.07206571,
- 8.09779453, -15.13547134],
- [ 43.36688232, 36.55143356, 39.08946228, ..., -2.59661937,
- -12.32478523, -13.29827309]], dtype=float32))
- ================================================================================
- [[ 0. 0. 0. ..., 0. 0. 0.]]
- 8326
- ('f', u'prevention')
- ('as', u'prevention')
- ('feed', array([8326]))
- ('Sentence :', u'prevention that by workers used the <unk> all the on a a a a a used the <unk> all at <unk> kent <eos> made imposed <eos> made imposed in were <unk> all the on a a a a a used the <unk> all at <unk> kent filters <eos> made imposed <eos> made imposed in were of all be <unk> kent <eos> asbestos <unk> all the on a a a a a used the of ban into <unk> of all N were of all N were of all N were of all N were of all N were of all N were ')
- ================================================================================
- ('Loss :', 1062.2104308281921)
- ('Batch_input :', array([ 83, 9, 102, 798, 10, 57, 4, 150, 431, 8, 586,
- 5351, 1, 4464, 15, 1, 1890, 232, 70, 4, 1222, 1063,
- 11, 1, 82, 1, 2, 29, 110, 34, 5, 4173, 10,
- 144, 1466, 453, 665, 3, 72, 195, 2, 14, 30, 102,
- 5596, 16, 228, 221, 534, 325, 2, 927, 16, 3730, 1359,
- 193, 608, 5, 2276, 1633, 1692, 10, 683, 586], dtype=int32))
- ('Batch_labels :', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- Average loss at step 700: 10.622104 learning rate: 1.000000
- ('Label: ', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- ('Predicted:', array([[ 26.62189865, 74.89472961, 77.30410767, ..., 1.25362027,
- 7.7031641 , -21.29500771],
- [ 29.20170784, 60.07447052, 38.95781708, ..., 8.34551048,
- -7.05633163, -11.66267395],
- [ 26.43227196, 66.50967407, 12.55805016, ..., 7.04926825,
- 0.87124491, -20.38945961],
- ...,
- [ 48.20450974, 55.11436462, 64.52549744, ..., -0.12630704,
- 7.41949749, -15.82486343],
- [ 27.50131607, 48.1931572 , 39.01482391, ..., 2.25321031,
- -5.42439842, -5.85474253],
- [ 37.04358673, 42.9875946 , 27.22852898, ..., -0.11088732,
- -6.61701012, -5.53644276]], dtype=float32))
- ================================================================================
- [[ 0. 0. 0. ..., 0. 0. 0.]]
- 94
- ('f', u'over')
- ('as', u'over')
- ('feed', array([94]))
- ('Sentence :', u'over our you work in no <eos> ago N all N all it events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events events ')
- ================================================================================
- ('Loss :', 1048.0584556075678)
- ('Batch_input :', array([ 3, 11, 0, 123, 327, 475, 2, 4096, 927, 2288, 6191,
- 4, 1422, 8, 10, 0, 264, 236, 1049, 11, 6, 40,
- 2, 196, 1531, 4, 0, 193, 134, 744, 1, 17, 6,
- 272, 5, 3, 171, 0, 8210, 155, 266, 397, 214, 5,
- 8024, 9, 2, 917, 4320, 26, 948, 5, 2176, 1607, 131,
- 172, 76, 38, 3304, 683, 586, 5, 2457, 1308], dtype=int32))
- ('Batch_labels :', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 1., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- Average loss at step 800: 10.480585 learning rate: 1.000000
- ('Label: ', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 1., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- ('Predicted:', array([[ 51.68177795, 49.3968544 , 44.19010162, ..., -0.96025121,
- -2.75785208, -4.37986326],
- [ 74.0589447 , 51.08234024, 46.29760361, ..., 1.14275515,
- -3.16398978, -1.2056998 ],
- [ 73.70262146, 32.66469193, 39.23232269, ..., 3.43524694,
- -1.98378968, 0.50856781],
- ...,
- [ 58.23159027, 42.74452972, 25.42888641, ..., -1.53912532,
- 7.33349323, -9.04616547],
- [ 8.14913368, 36.12153625, 48.75240326, ..., 10.54703522,
- 3.37933898, -17.40168953],
- [ 31.39354706, 23.83557129, 57.65667725, ..., -10.08233261,
- -11.59241199, -10.97598267]], dtype=float32))
- ================================================================================
- [[ 0. 0. 0. ..., 0. 0. 0.]]
- 3604
- ('f', u'inventory')
- ('as', u'inventory')
- ('feed', array([3604]))
- ('Sentence :', u'inventory funds funds funds our <unk> reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment reinvestment ')
- ================================================================================
- ('Loss :', 1159.6224331202236)
- ('Batch_input :', array([ 57, 5, 25, 6, 1947, 4922, 76, 150, 586, 2245, 0,
- 47, 1101, 697, 6, 241, 374, 11, 0, 40, 3, 171,
- 2, 2937, 15, 1, 1, 1, 1704, 4, 161, 293, 237,
- 927, 93, 1, 52, 554, 157, 38, 1, 118, 76, 4,
- 181, 4783, 7, 966, 131, 172, 2, 0, 236, 16, 2015,
- 281, 613, 238, 18, 381, 9, 960, 11, 471], dtype=int32))
- ('Batch_labels :', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- Average loss at step 900: 11.596224 learning rate: 1.000000
- ('Label: ', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- ('Predicted:', array([[ 7.78842468e+01, 6.70051498e+01, 5.68985748e+01, ...,
- 9.97701585e-02, 1.53537297e+00, -2.83592224e+01],
- [ 5.26693001e+01, 8.03840942e+01, 5.94459763e+01, ...,
- -9.41697407e+00, -5.38884020e+00, -2.27148266e+01],
- [ 7.19977875e+01, 7.00528336e+01, 4.83365097e+01, ...,
- 3.95029211e+00, -1.83616962e+01, -1.87646961e+01],
- ...,
- [ 6.56173630e+01, 3.64674530e+01, 1.02826248e+02, ...,
- -3.28653312e+00, -3.12544656e+00, -1.22953815e+01],
- [ 6.99354324e+01, 3.96749344e+01, 1.00339859e+02, ...,
- -4.53111172e+00, -1.20459199e+00, -8.92706871e+00],
- [ 4.16675301e+01, 2.65603065e+01, 5.89054337e+01, ...,
- -7.48283148e-01, -4.27502155e+00, -5.28357267e+00]], dtype=float32))
- ================================================================================
- [[ 0. 0. 0. ..., 0. 0. 0.]]
- 9980
- ('f', u'fromstein')
- ('as', u'fromstein')
- ('feed', array([9980]))
- ('Sentence :', u'fromstein short-term interest fund reinvestment money <eos> they they before rates <eos> they they they they before rates <eos> they they before rates <eos> they they before rates <eos> they they before rates <eos> they they before to in on because interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest interest ')
- ================================================================================
- ('Loss :', 1257.1257801287791)
- ('Batch_input :', array([ 123, 5, 12, 3, 48, 2, 1461, 9175, 927, 2770, 1991,
- 966, 744, 76, 683, 586, 89, 3238, 4320, 8, 386, 78,
- 0, 1590, 172, 2, 0, 524, 161, 193, 26, 488, 4406,
- 200, 94, 3, 3, 2, 6010, 1408, 332, 0, 1, 293,
- 50, 6, 5227, 4096, 236, 4, 3, 3, 198, 0, 467,
- 123, 118, 20, 3, 3, 6, 123, 133, 2], dtype=int32))
- ('Batch_labels :', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 1., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- Average loss at step 1000: 12.571258 learning rate: 1.000000
- ('Label: ', array([[ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.],
- ...,
- [ 0., 0., 0., ..., 0., 0., 0.],
- [ 0., 0., 1., ..., 0., 0., 0.],
- [ 0., 0., 0., ..., 0., 0., 0.]], dtype=float32))
- ('Predicted:', array([[ 81.48632812, 72.03902435, 46.43121719, ..., 3.6133337 ,
- -2.32356 , -15.21035957],
- [ 83.48338318, 42.52036285, 58.53778839, ..., -10.0191946 ,
- -5.18699551, -22.0647583 ],
- [ 119.05833435, 67.42713928, 74.83621216, ..., 2.84793997,
- 3.17879272, -26.87680054],
- ...,
- [ 77.65213776, 44.25524139, 71.1973114 , ..., 1.99433267,
- -5.10602999, -17.80796051],
- [ 56.73459244, 6.60482407, 109.00973511, ..., -4.26551771,
- -10.72382736, -7.26380682],
- [ 30.19046211, 30.17685699, 32.82802582, ..., 0.76188219,
- -9.53625393, -5.97190285]], dtype=float32))
- ================================================================================
- [[ 0. 0. 0. ..., 0. 0. 0.]]
- 2151
- ('f', u'shipping')
- ('as', u'shipping')
- ('feed', array([2151]))
- ('Sentence :', u'shipping N managers over <eos> short-term <eos> a the a a short-term <eos> a funds over investments N managers over on comparable rises funds rates yield example latest support currently point over the funds fund recent dollar for typically fund recent dollar for typically dollar for typically dollar for typically fund recent fund recent fund recent fund recent fund recent dollar for typically fund recent dollar for typically dollar for typically dollar for typically dollar for typically fund recent fund recent dollar for typically dollar for typically dollar for typically dollar for typically fund recent fund recent fund recent dollar for ')
- ================================================================================
Add Comment
Please, Sign In to add comment