neural-network

Normalizing to [0,1] vs [-1,1]

跟風遠走 提交于 2019-12-12 07:36:07
问题 I've been going through a few tutorials on using neural networks for key points detection. I've noticed that for the inputs (images) it's very common to divide by 255 (normalizing to [0,1] since values fall between 0 and 255). But for the targets (X/Y) coordinates I've noticed it's more common to normalize to [-1,1]. Any reason for this disparity. Example: http://danielnouri.org/notes/2014/12/17/using-convolutional-neural-nets-to-detect-facial-keypoints-tutorial/ X = np.vstack(df['Image']

How to run pretrained models of Pixel Objectness

浪尽此生 提交于 2019-12-12 06:47:17
问题 I can't run pretrained models of Pixel Objectness published by Suyog Dutt Jain et al in 2017. I referred section Using the pretrained models in README.md on GitHub, but couldn't understand the following procedures well. ・Setup: Download and install Deeplab-v1 from here ・Refer to demo.py for step-by-step instruction on how to run the code. I tryed to install and run "demo.py", but I got some errors. Though I referred solutions on Website, unsolvable error remained. /usr/bin/python2.7 /mnt/C6BF

Caffe error: no field named “net”

孤街醉人 提交于 2019-12-12 06:14:26
问题 I had the Caffe C++ example program working on my computer, but after recently recompiling Caffe, I've encountered this error when I try to run the program: [libprotobuf ERROR google/protobuf/text_format.cc:245] Error parsing text-format caffe.NetParameter: 2:4: Message type "caffe.NetParameter" has no field named "net". upgrade_proto.cpp:928] Check failed: ReadProtoFromTextFile(param_file, param) Failed to parse NetParameter file: /home/jack/Desktop/beeshiny/deploy.prototxt Am I missing

RNN: Back-propagation through time when output is taken only at final timestep

血红的双手。 提交于 2019-12-12 05:59:28
问题 In this blog on Recurrent Neural Networks by Denny Britz. Author states that, " The above diagram has outputs at each time step, but depending on the task this may not be necessary. For example, when predicting the sentiment of a sentence we may only care about the final output, not the sentiment after each word. Similarly, we may not need inputs at each time step. " In the case when we take output only at the final timestep: How will backpropogation change, if there are no outputs at each

Tensorflow CUDA GTX 1070 import error

爷,独闯天下 提交于 2019-12-12 05:58:58
问题 I'm trying to install Tensorflow with CUDA support. Here are my specs: NVIDIA GTX 1070 CUDA 7.5 Cudnn v5.0 I have installed Tensorflow via the pip installation -- so I'm picturing your answer being to install from source, but I want to make sure there isn't a quick fix. The error is: volcart@volcart-Precision-Tower-7910:~$ python Python 2.7.10 (default, Oct 14 2015, 16:09:02) [GCC 5.2.1 20151010] on linux2 Type "help", "copyright", "credits" or "license" for more information. >>> import

Neural Network Becomes Unruly with Large Layers

匆匆过客 提交于 2019-12-12 05:27:20
问题 This is a higher-level question about the performance of a neural network. The issue I'm having is that with larger numbers of neurons per layer, the network has frequent rounds of complete stupidity. They are not consistent; it seems that the probability of general success vs failure is about 50/50 when layers get larger than 60 neurons (always 3 layers). I tested this by teaching the same function to networks with input and hidden layers of sizes from 10-200. The success rate is either 0-1%

Value Error while feeding in Neural Network [closed]

|▌冷眼眸甩不掉的悲伤 提交于 2019-12-12 05:01:28
问题 Closed. This question is off-topic. It is not currently accepting answers. Want to improve this question? Update the question so it's on-topic for Stack Overflow. Closed 2 years ago . I have programmed a multi-layer neural network but I'm getting an error while feeding my dimension into it. I'm getting a Value Error. Here is The Code: import pandas as pd import tensorflow as tf import matplotlib.pyplot as plt import numpy as np from sklearn import datasets from sklearn import metrics from

Theano Cost Function, TypeError: Unknown parameter type: <class 'numpy.ndarray'>

做~自己de王妃 提交于 2019-12-12 04:48:26
问题 I'm new to Theano, just learning it. I have a ANN in python that I'm implementing in Theano as learning process. I'm using Spyder. And Theano throws out an error: TypeError: Unknown parameter type: class 'numpy.ndarray' I'm not sure where the error is. Is it in the cost function or in the gradient descent? And what is the typical reason for it? Here is my code: X = T.dmatrix() y = T.dmatrix() X_input = np.genfromtxt('X.csv',delimiter=',') #5000x195 y_input = np.genfromtxt('y.csv',delimiter=',

ValueError: Cannot feed value of shape (0,) for Tensor 'x:0', which has shape '(?, 400, 600, 1)'

空扰寡人 提交于 2019-12-12 04:48:19
问题 I get the error above when running my Neural net: print('Checking the Training on a Single Batch...') with tf.Session() as sess: # Initializing the variables sess.run(tf.global_variables_initializer()) # Training cycle for epoch in range(epochs): batch_i = 1 for batch_features, batch_labels in (input_data, input_labels): train_neural_network(sess, optimizer, keep_probability, batch_features, batch_labels) print('Epoch {:>2}, Batch {}: '.format(epoch + 1, batch_i), end='') print_stats(sess,

How to get predictions for each set of parameters using GridSearchCV?

拜拜、爱过 提交于 2019-12-12 04:39:55
问题 I'm trying to find the best parameters for NN regression model using GridSearchCV with following code: param_grid = dict(optimizer=optimizer, epochs=epochs, batch_size=batches, init=init grid = GridSearchCV(estimator=model, param_grid=param_grid, scoring='neg_mean_squared_error') grid_result = grid.fit(input_train, target_train) pred = grid.predict(input_test) As I understand, grid.predict(input_test) uses best parameters to predict the given input set. Is there any way to evaluate