Advertisement
nordlaender

Theano error message

Oct 21st, 2016
621
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
Bash 9.54 KB | None | 0 0
  1. ERROR (theano.gof.opt): Optimization failure due to: local_abstractconv_check
  2. ERROR (theano.gof.opt): node: AbstractConv2d{border_mode='half', subsample=(1, 1), filter_flip=True, imshp=(None, None, None, None), kshp=(32, 3, 3, 3)}(DimShuffle{0,3,1,2}.0, DimShuffle{3,2,0,1}.0)
  3. ERROR (theano.gof.opt): TRACEBACK:
  4. ERROR (theano.gof.opt): Traceback (most recent call last):
  5.   File "C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\theano\gof\opt.py", line 1772, in process_node
  6.     replacements = lopt.transform(node)
  7.   File "C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\theano\tensor\nnet\opt.py", line 402, in local_abstractconv_check
  8.     node.op.__class__.__name__)
  9. AssertionError: AbstractConv2d Theano optimization failed: there is no implementation available supporting the requested options. Did you exclude both "conv_dnn" and "conv_gemm" from the optimizer? If on GPU, is cuDNN available and does the GPU support it? If on CPU, do you have a BLAS library installed Theano can link against?
  10.  
  11. ---------------------------------------------------------------------------
  12. AssertionError                            Traceback (most recent call last)
  13. <ipython-input-5-19331022925d> in <module>()
  14.       6               nb_epoch=nb_epoch,
  15.       7               validation_data=(X_test, Y_test),
  16. ----> 8               shuffle=True)
  17.       9 else:
  18.      10     print('Using real-time data augmentation.')
  19.  
  20. C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\keras\models.py in fit(self, x, y, batch_size, nb_epoch, verbose, callbacks, validation_split, validation_data, shuffle, class_weight, sample_weight, **kwargs)
  21.     618                               shuffle=shuffle,
  22.     619                               class_weight=class_weight,
  23. --> 620                               sample_weight=sample_weight)
  24.     621
  25.     622     def evaluate(self, x, y, batch_size=32, verbose=1,
  26.  
  27. C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\keras\engine\training.py in fit(self, x, y, batch_size, nb_epoch, verbose, callbacks, validation_split, validation_data, shuffle, class_weight, sample_weight)
  28.    1047                                                                            check_batch_dim=False,
  29.    1048                                                                            batch_size=batch_size)
  30. -> 1049             self._make_test_function()
  31.    1050             val_f = self.test_function
  32.    1051             if self.uses_learning_phase and type(K.learning_phase()) is not int:
  33.  
  34. C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\keras\engine\training.py in _make_test_function(self)
  35.     716                                             [self.total_loss] + self.metrics_tensors,
  36.     717                                             updates=self.state_updates,
  37. --> 718                                             **self._function_kwargs)
  38.     719
  39.     720     def _make_predict_function(self):
  40.  
  41. C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\keras\backend\theano_backend.py in function(inputs, outputs, updates, **kwargs)
  42.     725                 msg = "Invalid argument '%s' passed to K.function" % key
  43.     726                 raise ValueError(msg)
  44. --> 727     return Function(inputs, outputs, updates=updates, **kwargs)
  45.     728
  46.     729
  47.  
  48. C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\keras\backend\theano_backend.py in __init__(self, inputs, outputs, updates, **kwargs)
  49.     711                                         allow_input_downcast=True,
  50.     712                                         on_unused_input='ignore',
  51. --> 713                                         **kwargs)
  52.     714
  53.     715     def __call__(self, inputs):
  54.  
  55. C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\theano\compile\function.py in function(inputs, outputs, mode, updates, givens, no_default_updates, accept_inplace, name, rebuild_strict, allow_input_downcast, profile, on_unused_input)
  56.     318                    on_unused_input=on_unused_input,
  57.     319                    profile=profile,
  58. --> 320                    output_keys=output_keys)
  59.     321     # We need to add the flag check_aliased inputs if we have any mutable or
  60.     322     # borrowed used defined inputs
  61.  
  62. C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\theano\compile\pfunc.py in pfunc(params, outputs, mode, updates, givens, no_default_updates, accept_inplace, name, rebuild_strict, allow_input_downcast, profile, on_unused_input, output_keys)
  63.     477                          accept_inplace=accept_inplace, name=name,
  64.     478                          profile=profile, on_unused_input=on_unused_input,
  65. --> 479                          output_keys=output_keys)
  66.     480
  67.     481
  68.  
  69. C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\theano\compile\function_module.py in orig_function(inputs, outputs, mode, accept_inplace, name, profile, on_unused_input, output_keys)
  70.    1774                    profile=profile,
  71.    1775                    on_unused_input=on_unused_input,
  72. -> 1776                    output_keys=output_keys).create(
  73.    1777             defaults)
  74.    1778
  75.  
  76. C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\theano\compile\function_module.py in __init__(self, inputs, outputs, mode, accept_inplace, function_builder, profile, on_unused_input, fgraph, output_keys)
  77.    1454                         optimizer, inputs, outputs)
  78.    1455                 else:
  79. -> 1456                     optimizer_profile = optimizer(fgraph)
  80.    1457
  81.    1458                 end_optimizer = time.time()
  82.  
  83. C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\theano\gof\opt.py in __call__(self, fgraph)
  84.      99
  85.     100         """
  86. --> 101         return self.optimize(fgraph)
  87.    102
  88.    103     def add_requirements(self, fgraph):
  89.  
  90. C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\theano\gof\opt.py in optimize(self, fgraph, *args, **kwargs)
  91.     87             orig = theano.tensor.basic.constant.enable
  92.     88             theano.tensor.basic.constant.enable = False
  93. ---> 89             ret = self.apply(fgraph, *args, **kwargs)
  94.     90         finally:
  95.     91             theano.tensor.basic.constant.enable = orig
  96.  
  97. C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\theano\gof\opt.py in apply(self, fgraph)
  98.    228             try:
  99.    229                 t0 = time.time()
  100. --> 230                 sub_prof = optimizer.optimize(fgraph)
  101.    231                 l.append(float(time.time() - t0))
  102.    232                 sub_profs.append(sub_prof)
  103.  
  104. C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\theano\gof\opt.py in optimize(self, fgraph, *args, **kwargs)
  105.     87             orig = theano.tensor.basic.constant.enable
  106.     88             theano.tensor.basic.constant.enable = False
  107. ---> 89             ret = self.apply(fgraph, *args, **kwargs)
  108.     90         finally:
  109.     91             theano.tensor.basic.constant.enable = orig
  110.  
  111. C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\theano\gof\opt.py in apply(self, fgraph, start_from)
  112.   1877                     node = q.popleft()
  113.   1878                 current_node = node
  114. -> 1879                 nb += self.process_node(fgraph, node)
  115.   1880             loop_t = time.time() - t0
  116.   1881         except Exception:
  117.  
  118. C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\theano\gof\opt.py in process_node(self, fgraph, node, lopt)
  119.   1775                 self.failure_callback(e, self,
  120.   1776                                       [(x, None) for x in node.outputs],
  121. -> 1777                                       lopt, node)
  122.   1778                 return False
  123.   1779             else:
  124.  
  125. C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\theano\gof\opt.py in warn_inplace(exc, nav, repl_pairs, local_opt, node)
  126.   1671         if isinstance(exc, InconsistencyError):
  127.   1672             return
  128. -> 1673         return NavigatorOptimizer.warn(exc, nav, repl_pairs, local_opt, node)
  129.   1674
  130.   1675     @staticmethod
  131.  
  132. C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\theano\gof\opt.py in warn(exc, nav, repl_pairs, local_opt, node)
  133.   1657             # We always crash on AssertionError because something may be
  134.   1658             # seriously wrong if such an exception is raised.
  135. -> 1659             raise exc
  136.   1660
  137.   1661     @staticmethod
  138.  
  139. C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\theano\gof\opt.py in process_node(self, fgraph, node, lopt)
  140.   1770         lopt = lopt or self.local_opt
  141.   1771         try:
  142. -> 1772             replacements = lopt.transform(node)
  143.   1773         except Exception as e:
  144.   1774             if self.failure_callback is not None:
  145.  
  146. C:\Users\212590884\AppData\Local\Continuum\Anaconda3\lib\site-packages\theano\tensor\nnet\opt.py in local_abstractconv_check(node)
  147.    400             'is cuDNN available and does the GPU support it? If on CPU, '
  148.    401             'do you have a BLAS library installed Theano can link against?' %
  149. --> 402             node.op.__class__.__name__)
  150.    403
  151.    404 optdb.register('AbstractConvCheck',
  152.  
  153. AssertionError: AbstractConv2d Theano optimization failed: there is no implementation available supporting the requested options. Did you exclude both "conv_dnn" and "conv_gemm" from the optimizer? If on GPU, is cuDNN available and does the GPU support it? If on CPU, do you have a BLAS library installed Theano can link against?
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement