Advertisement
Guest User

Untitled

a guest
Jun 26th, 2019
86
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 10.24 KB | None | 0 0
  1. ecg
  2. 0 0.1912
  3. 1 0.3597
  4. 2 0.3597
  5. 3 0.3597
  6. 4 0.3597
  7. 5 0.3597
  8. 6 0.2739
  9. 7 0.1641
  10. 8 0.0776
  11. 9 0.0005
  12. 10 -0.0375
  13. 11 -0.0676
  14. 12 -0.1071
  15. 13 -0.1197
  16. .. .......
  17. .. .......
  18. .. .......
  19. 5616000 0.0226
  20.  
  21. emotion
  22. 0 0
  23. 1 0
  24. 2 0
  25. 3 0
  26. 4 0
  27. . .
  28. . .
  29. . .
  30. 18001 1
  31. 18002 1
  32. 18003 1
  33. . .
  34. . .
  35. . .
  36. 360001 2
  37. 360002 2
  38. 360003 2
  39. . .
  40. . .
  41. . .
  42. . .
  43. 5616000 5
  44.  
  45. train_x = train_x.values.reshape(312,18000,1)
  46. train_y = train_y.values.reshape(312,18000)
  47. train_y = train_y[:,:1] # truncated train_y to have single corresponding value to a complete signal.
  48. train_y = pd.DataFrame(train_y)
  49. train_y = pd.get_dummies(train_y[0]) #one hot encoded labels
  50.  
  51. [[[0.60399908]
  52. [0.79763273]
  53. [0.79763273]
  54. ...
  55. [0.09779361]
  56. [0.09779361]
  57. [0.14732245]]
  58.  
  59. [[0.70386905]
  60. [0.95101687]
  61. [0.95101687]
  62. ...
  63. [0.41530258]
  64. [0.41728671]
  65. [0.42261905]]
  66.  
  67. [[0.75008021]
  68. [1. ]
  69. [1. ]
  70. ...
  71. [0.46412148]
  72. [0.46412148]
  73. [0.46412148]]
  74.  
  75. ...
  76.  
  77. [[0.60977509]
  78. [0.7756791 ]
  79. [0.7756791 ]
  80. ...
  81. [0.12725148]
  82. [0.02755331]
  83. [0.02755331]]
  84.  
  85. [[0.59939494]
  86. [0.75514785]
  87. [0.75514785]
  88. ...
  89. [0.0391334 ]
  90. [0.0391334 ]
  91. [0.0578706 ]]
  92.  
  93. [[0.5786066 ]
  94. [0.71539303]
  95. [0.71539303]
  96. ...
  97. [0.41355098]
  98. [0.41355098]
  99. [0.4112712 ]]]
  100.  
  101. 0 1 2 3 4 5
  102. 0 1 0 0 0 0 0
  103. 1 1 0 0 0 0 0
  104. 2 0 1 0 0 0 0
  105. 3 0 1 0 0 0 0
  106. 4 0 0 0 0 0 1
  107. 5 0 0 0 0 0 1
  108. 6 0 0 1 0 0 0
  109. 7 0 0 1 0 0 0
  110. 8 0 0 0 1 0 0
  111. 9 0 0 0 1 0 0
  112. 10 0 0 0 0 1 0
  113. 11 0 0 0 0 1 0
  114. 12 0 0 0 1 0 0
  115. 13 0 0 0 1 0 0
  116. 14 0 1 0 0 0 0
  117. 15 0 1 0 0 0 0
  118. 16 1 0 0 0 0 0
  119. 17 1 0 0 0 0 0
  120. 18 0 0 1 0 0 0
  121. 19 0 0 1 0 0 0
  122. 20 0 0 0 0 1 0
  123. 21 0 0 0 0 1 0
  124. 22 0 0 0 0 0 1
  125. 23 0 0 0 0 0 1
  126. 24 0 0 0 0 0 1
  127. 25 0 0 0 0 0 1
  128. 26 0 0 1 0 0 0
  129. 27 0 0 1 0 0 0
  130. 28 0 1 0 0 0 0
  131. 29 0 1 0 0 0 0
  132. .. .. .. .. .. .. ..
  133. 282 0 0 0 1 0 0
  134. 283 0 0 0 1 0 0
  135. 284 1 0 0 0 0 0
  136. 285 1 0 0 0 0 0
  137. 286 0 0 0 0 1 0
  138. 287 0 0 0 0 1 0
  139. 288 1 0 0 0 0 0
  140. 289 1 0 0 0 0 0
  141. 290 0 1 0 0 0 0
  142. 291 0 1 0 0 0 0
  143. 292 0 0 0 1 0 0
  144. 293 0 0 0 1 0 0
  145. 294 0 0 1 0 0 0
  146. 295 0 0 1 0 0 0
  147. 296 0 0 0 0 0 1
  148. 297 0 0 0 0 0 1
  149. 298 0 0 0 0 1 0
  150. 299 0 0 0 0 1 0
  151. 300 0 0 0 1 0 0
  152. 301 0 0 0 1 0 0
  153. 302 0 0 1 0 0 0
  154. 303 0 0 1 0 0 0
  155. 304 0 0 0 0 0 1
  156. 305 0 0 0 0 0 1
  157. 306 0 1 0 0 0 0
  158. 307 0 1 0 0 0 0
  159. 308 0 0 0 0 1 0
  160. 309 0 0 0 0 1 0
  161. 310 1 0 0 0 0 0
  162. 311 1 0 0 0 0 0
  163.  
  164. [312 rows x 6 columns]
  165.  
  166. model = Sequential()
  167. model.add(Conv1D(100,700,activation='relu',input_shape=(18000,1))) #kernel_size is 700 because 18000 rows = 60 seconds so 700 rows = ~2.33 seconds and there is two heart beat peak in every 2 second for ecg signal.
  168. model.add(Conv1D(50,700))
  169. model.add(Dropout(0.5))
  170. model.add(BatchNormalization())
  171. model.add(Activation('relu'))
  172. model.add(MaxPooling1D(4))
  173. model.add(Flatten())
  174. model.add(Dense(6,activation='softmax'))
  175.  
  176. adam = keras.optimizers.Adam(lr=0.0001, beta_1=0.9, beta_2=0.999, epsilon=None, decay=0.0, amsgrad=False)
  177.  
  178. model.compile(optimizer = adam, loss = 'categorical_crossentropy', metrics = ['acc'])
  179. model.fit(train_x,train_y,epochs = 50, batch_size = 32, validation_split=0.33, shuffle=False)
  180.  
  181. Epoch 1/80
  182. 249/249 [==============================] - 24s 96ms/step - loss: 2.3118 - acc: 0.1406 - val_loss: 1.7989 - val_acc: 0.1587
  183. Epoch 2/80
  184. 249/249 [==============================] - 19s 76ms/step - loss: 2.0468 - acc: 0.1647 - val_loss: 1.8605 - val_acc: 0.2222
  185. Epoch 3/80
  186. 249/249 [==============================] - 19s 76ms/step - loss: 1.9562 - acc: 0.1767 - val_loss: 1.8203 - val_acc: 0.2063
  187. Epoch 4/80
  188. 249/249 [==============================] - 19s 75ms/step - loss: 1.9361 - acc: 0.2169 - val_loss: 1.8033 - val_acc: 0.1905
  189. Epoch 5/80
  190. 249/249 [==============================] - 19s 74ms/step - loss: 1.8834 - acc: 0.1847 - val_loss: 1.8198 - val_acc: 0.2222
  191. Epoch 6/80
  192. 249/249 [==============================] - 19s 75ms/step - loss: 1.8278 - acc: 0.2410 - val_loss: 1.7961 - val_acc: 0.1905
  193. Epoch 7/80
  194. 249/249 [==============================] - 19s 75ms/step - loss: 1.8022 - acc: 0.2450 - val_loss: 1.8092 - val_acc: 0.2063
  195. Epoch 8/80
  196. 249/249 [==============================] - 19s 75ms/step - loss: 1.7959 - acc: 0.2369 - val_loss: 1.8005 - val_acc: 0.2222
  197. Epoch 9/80
  198. 249/249 [==============================] - 19s 75ms/step - loss: 1.7234 - acc: 0.2610 - val_loss: 1.7871 - val_acc: 0.2381
  199. Epoch 10/80
  200. 249/249 [==============================] - 19s 75ms/step - loss: 1.6861 - acc: 0.2972 - val_loss: 1.8017 - val_acc: 0.1905
  201. Epoch 11/80
  202. 249/249 [==============================] - 19s 75ms/step - loss: 1.6696 - acc: 0.3173 - val_loss: 1.7878 - val_acc: 0.1905
  203. Epoch 12/80
  204. 249/249 [==============================] - 19s 75ms/step - loss: 1.5868 - acc: 0.3655 - val_loss: 1.7771 - val_acc: 0.1270
  205. Epoch 13/80
  206. 249/249 [==============================] - 19s 75ms/step - loss: 1.5751 - acc: 0.3936 - val_loss: 1.7818 - val_acc: 0.1270
  207. Epoch 14/80
  208. 249/249 [==============================] - 19s 75ms/step - loss: 1.5647 - acc: 0.3735 - val_loss: 1.7733 - val_acc: 0.1429
  209. Epoch 15/80
  210. 249/249 [==============================] - 19s 75ms/step - loss: 1.4621 - acc: 0.4177 - val_loss: 1.7759 - val_acc: 0.1270
  211. Epoch 16/80
  212. 249/249 [==============================] - 19s 75ms/step - loss: 1.4519 - acc: 0.4498 - val_loss: 1.8005 - val_acc: 0.1746
  213. Epoch 17/80
  214. 249/249 [==============================] - 19s 75ms/step - loss: 1.4489 - acc: 0.4378 - val_loss: 1.8020 - val_acc: 0.1270
  215. Epoch 18/80
  216. 249/249 [==============================] - 19s 75ms/step - loss: 1.4449 - acc: 0.4297 - val_loss: 1.7852 - val_acc: 0.1587
  217. Epoch 19/80
  218. 249/249 [==============================] - 19s 75ms/step - loss: 1.3600 - acc: 0.5301 - val_loss: 1.7922 - val_acc: 0.1429
  219. Epoch 20/80
  220. 249/249 [==============================] - 19s 75ms/step - loss: 1.3349 - acc: 0.5422 - val_loss: 1.8061 - val_acc: 0.2222
  221. Epoch 21/80
  222. 249/249 [==============================] - 19s 75ms/step - loss: 1.2885 - acc: 0.5622 - val_loss: 1.8235 - val_acc: 0.1746
  223. Epoch 22/80
  224. 249/249 [==============================] - 19s 75ms/step - loss: 1.2291 - acc: 0.5823 - val_loss: 1.8173 - val_acc: 0.1905
  225. Epoch 23/80
  226. 249/249 [==============================] - 19s 75ms/step - loss: 1.1890 - acc: 0.6506 - val_loss: 1.8293 - val_acc: 0.1905
  227. Epoch 24/80
  228. 249/249 [==============================] - 19s 75ms/step - loss: 1.1473 - acc: 0.6627 - val_loss: 1.8274 - val_acc: 0.1746
  229. Epoch 25/80
  230. 249/249 [==============================] - 19s 75ms/step - loss: 1.1060 - acc: 0.6747 - val_loss: 1.8142 - val_acc: 0.1587
  231. Epoch 26/80
  232. 249/249 [==============================] - 19s 75ms/step - loss: 1.0210 - acc: 0.7510 - val_loss: 1.8126 - val_acc: 0.1905
  233. Epoch 27/80
  234. 249/249 [==============================] - 19s 75ms/step - loss: 0.9699 - acc: 0.7631 - val_loss: 1.8094 - val_acc: 0.1746
  235. Epoch 28/80
  236. 249/249 [==============================] - 19s 75ms/step - loss: 0.9127 - acc: 0.8193 - val_loss: 1.8012 - val_acc: 0.1746
  237. Epoch 29/80
  238. 249/249 [==============================] - 19s 75ms/step - loss: 0.9176 - acc: 0.7871 - val_loss: 1.8371 - val_acc: 0.1746
  239. Epoch 30/80
  240. 249/249 [==============================] - 19s 75ms/step - loss: 0.8725 - acc: 0.8233 - val_loss: 1.8215 - val_acc: 0.1587
  241. Epoch 31/80
  242. 249/249 [==============================] - 19s 75ms/step - loss: 0.8316 - acc: 0.8514 - val_loss: 1.8010 - val_acc: 0.1429
  243. Epoch 32/80
  244. 249/249 [==============================] - 19s 75ms/step - loss: 0.7958 - acc: 0.8474 - val_loss: 1.8594 - val_acc: 0.1270
  245. Epoch 33/80
  246. 249/249 [==============================] - 19s 75ms/step - loss: 0.7452 - acc: 0.8795 - val_loss: 1.8260 - val_acc: 0.1587
  247. Epoch 34/80
  248. 249/249 [==============================] - 19s 75ms/step - loss: 0.7395 - acc: 0.8916 - val_loss: 1.8191 - val_acc: 0.1587
  249. Epoch 35/80
  250. 249/249 [==============================] - 19s 75ms/step - loss: 0.6794 - acc: 0.9357 - val_loss: 1.8344 - val_acc: 0.1429
  251. Epoch 36/80
  252. 249/249 [==============================] - 19s 75ms/step - loss: 0.6106 - acc: 0.9357 - val_loss: 1.7903 - val_acc: 0.1111
  253. Epoch 37/80
  254. 249/249 [==============================] - 19s 75ms/step - loss: 0.5609 - acc: 0.9598 - val_loss: 1.7882 - val_acc: 0.1429
  255. Epoch 38/80
  256. 249/249 [==============================] - 19s 75ms/step - loss: 0.5788 - acc: 0.9478 - val_loss: 1.8036 - val_acc: 0.1905
  257. Epoch 39/80
  258. 249/249 [==============================] - 19s 75ms/step - loss: 0.5693 - acc: 0.9398 - val_loss: 1.7712 - val_acc: 0.1746
  259. Epoch 40/80
  260. 249/249 [==============================] - 19s 75ms/step - loss: 0.4911 - acc: 0.9598 - val_loss: 1.8497 - val_acc: 0.1429
  261. Epoch 41/80
  262. 249/249 [==============================] - 19s 75ms/step - loss: 0.4824 - acc: 0.9518 - val_loss: 1.8105 - val_acc: 0.1429
  263. Epoch 42/80
  264. 249/249 [==============================] - 19s 75ms/step - loss: 0.4198 - acc: 0.9759 - val_loss: 1.8332 - val_acc: 0.1111
  265. Epoch 43/80
  266. 249/249 [==============================] - 19s 75ms/step - loss: 0.3890 - acc: 0.9880 - val_loss: 1.9316 - val_acc: 0.1111
  267. Epoch 44/80
  268. 249/249 [==============================] - 19s 75ms/step - loss: 0.3762 - acc: 0.9920 - val_loss: 1.8333 - val_acc: 0.1746
  269. Epoch 45/80
  270. 249/249 [==============================] - 19s 75ms/step - loss: 0.3510 - acc: 0.9880 - val_loss: 1.8090 - val_acc: 0.1587
  271. Epoch 46/80
  272. 249/249 [==============================] - 19s 75ms/step - loss: 0.3306 - acc: 0.9880 - val_loss: 1.8230 - val_acc: 0.1587
  273. Epoch 47/80
  274. 249/249 [==============================] - 19s 75ms/step - loss: 0.2814 - acc: 1.0000 - val_loss: 1.7843 - val_acc: 0.2222
  275. Epoch 48/80
  276. 249/249 [==============================] - 19s 75ms/step - loss: 0.2794 - acc: 1.0000 - val_loss: 1.8147 - val_acc: 0.2063
  277. Epoch 49/80
  278. 249/249 [==============================] - 19s 75ms/step - loss: 0.2430 - acc: 1.0000 - val_loss: 1.8488 - val_acc: 0.1587
  279. Epoch 50/80
  280. 249/249 [==============================] - 19s 75ms/step - loss: 0.2216 - acc: 1.0000 - val_loss: 1.8215 - val_acc: 0.1587
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement