Related
So, I was doing my code like the one below. I want to Fit the ANN into the Training set, but the error occurred like this. I was confused about how to solve it, I already try several suggestions on Google, but it still came out an error. I also tried several codes to display the result, but most of the errors occurred is because of the fitting like this one. So, I was thinking that the main problem my code can't run is because of the fitting model.
#Importing necessary Libraries
import numpy as np
import pandas as pd
import tensorflow as tf
import keras
#dataset
data = pd.read_excel("E:\\MATKUL LUV\\THESIS\\DATASETS\\DPM1 1052.xlsx")
#print (data.head)
# Separate Target Variable and Predictor Variables
TargetVariable=['DPM1Fault']
Predictors=['DPM1Cx', 'DPM1Cy']
X=data[Predictors].values
y=data[TargetVariable].values
# Split the data into training and testing set
from sklearn.model_selection import train_test_split
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.3, random_state=42)
# Quick sanity check with the shapes of Training and testing datasets
print(X_train.shape)
print(y_train.shape)
print(X_test.shape)
print(y_test.shape)
# importing the libraries
from keras.models import Sequential
from keras.layers import Dense
# create ANN model
model = Sequential()
# Defining the Input layer and FIRST hidden layer, both are same!
model.add(Dense(units=5, input_dim=2, kernel_initializer='normal', activation='relu'))
# Defining the Second layer of the model
# after the first layer we don't have to specify input_dim as keras configure it automatically
model.add(Dense(units=4, kernel_initializer='normal', activation='relu'))
# The output neuron is a single fully connected node
# Since we will be predicting a single number
model.add(Dense(1, kernel_initializer='normal'))
# Compiling the model
model.compile(loss='mean_squared_error', optimizer='adam')
# Fitting the ANN to the Training set
model.fit(X_train, y_train ,batch_size = 20, epochs = 50, verbose=1)
and the result like this
Epoch 1/50
---------------------------------------------------------------------------
UnimplementedError Traceback (most recent call last)
Input In [52], in <cell line: 23>()
20 model.compile(loss='mean_squared_error', optimizer='adam')
22 # Fitting the ANN to the Training set
---> 23 model.fit(X_train, y_train ,batch_size = 20, epochs = 50, verbose=1)
File ~\anaconda3\lib\site-packages\keras\utils\traceback_utils.py:70, in filter_traceback.<locals>.error_handler(*args, **kwargs)
67 filtered_tb = _process_traceback_frames(e.__traceback__)
68 # To get the full stack trace, call:
69 # `tf.debugging.disable_traceback_filtering()`
---> 70 raise e.with_traceback(filtered_tb) from None
71 finally:
72 del filtered_tb
File ~\anaconda3\lib\site-packages\tensorflow\python\eager\execute.py:52, in quick_execute(op_name, num_outputs, inputs, attrs, ctx, name)
50 try:
51 ctx.ensure_initialized()
---> 52 tensors = pywrap_tfe.TFE_Py_Execute(ctx._handle, device_name, op_name,
53 inputs, attrs, num_outputs)
54 except core._NotOkStatusException as e:
55 if name is not None:
UnimplementedError: Graph execution error:
Detected at node 'mean_squared_error/Cast' defined at (most recent call last):
File "C:\Users\16agn\anaconda3\lib\runpy.py", line 197, in _run_module_as_main
return _run_code(code, main_globals, None,
File "C:\Users\16agn\anaconda3\lib\runpy.py", line 87, in _run_code
exec(code, run_globals)
File "C:\Users\16agn\anaconda3\lib\site-packages\ipykernel_launcher.py", line 16, in <module>
app.launch_new_instance()
File "C:\Users\16agn\anaconda3\lib\site-packages\traitlets\config\application.py", line 846, in launch_instance
app.start()
File "C:\Users\16agn\anaconda3\lib\site-packages\ipykernel\kernelapp.py", line 677, in start
self.io_loop.start()
File "C:\Users\16agn\anaconda3\lib\site-packages\tornado\platform\asyncio.py", line 199, in start
self.asyncio_loop.run_forever()
File "C:\Users\16agn\anaconda3\lib\asyncio\base_events.py", line 601, in run_forever
self._run_once()
File "C:\Users\16agn\anaconda3\lib\asyncio\base_events.py", line 1905, in _run_once
handle._run()
File "C:\Users\16agn\anaconda3\lib\asyncio\events.py", line 80, in _run
self._context.run(self._callback, *self._args)
File "C:\Users\16agn\anaconda3\lib\site-packages\ipykernel\kernelbase.py", line 471, in dispatch_queue
await self.process_one()
File "C:\Users\16agn\anaconda3\lib\site-packages\ipykernel\kernelbase.py", line 460, in process_one
await dispatch(*args)
File "C:\Users\16agn\anaconda3\lib\site-packages\ipykernel\kernelbase.py", line 367, in dispatch_shell
await result
File "C:\Users\16agn\anaconda3\lib\site-packages\ipykernel\kernelbase.py", line 662, in execute_request
reply_content = await reply_content
File "C:\Users\16agn\anaconda3\lib\site-packages\ipykernel\ipkernel.py", line 360, in do_execute
res = shell.run_cell(code, store_history=store_history, silent=silent)
File "C:\Users\16agn\anaconda3\lib\site-packages\ipykernel\zmqshell.py", line 532, in run_cell
return super().run_cell(*args, **kwargs)
File "C:\Users\16agn\anaconda3\lib\site-packages\IPython\core\interactiveshell.py", line 2863, in run_cell
result = self._run_cell(
File "C:\Users\16agn\anaconda3\lib\site-packages\IPython\core\interactiveshell.py", line 2909, in _run_cell
return runner(coro)
File "C:\Users\16agn\anaconda3\lib\site-packages\IPython\core\async_helpers.py", line 129, in _pseudo_sync_runner
coro.send(None)
File "C:\Users\16agn\anaconda3\lib\site-packages\IPython\core\interactiveshell.py", line 3106, in run_cell_async
has_raised = await self.run_ast_nodes(code_ast.body, cell_name,
File "C:\Users\16agn\anaconda3\lib\site-packages\IPython\core\interactiveshell.py", line 3309, in run_ast_nodes
if await self.run_code(code, result, async_=asy):
File "C:\Users\16agn\anaconda3\lib\site-packages\IPython\core\interactiveshell.py", line 3369, in run_code
exec(code_obj, self.user_global_ns, self.user_ns)
File "C:\Users\16agn\AppData\Local\Temp\ipykernel_33292\314088425.py", line 23, in <cell line: 23>
model.fit(X_train, y_train ,batch_size = 20, epochs = 50, verbose=1)
File "C:\Users\16agn\anaconda3\lib\site-packages\keras\utils\traceback_utils.py", line 65, in error_handler
return fn(*args, **kwargs)
File "C:\Users\16agn\anaconda3\lib\site-packages\keras\engine\training.py", line 1650, in fit
tmp_logs = self.train_function(iterator)
File "C:\Users\16agn\anaconda3\lib\site-packages\keras\engine\training.py", line 1249, in train_function
return step_function(self, iterator)
File "C:\Users\16agn\anaconda3\lib\site-packages\keras\engine\training.py", line 1233, in step_function
outputs = model.distribute_strategy.run(run_step, args=(data,))
File "C:\Users\16agn\anaconda3\lib\site-packages\keras\engine\training.py", line 1222, in run_step
outputs = model.train_step(data)
File "C:\Users\16agn\anaconda3\lib\site-packages\keras\engine\training.py", line 1024, in train_step
loss = self.compute_loss(x, y, y_pred, sample_weight)
File "C:\Users\16agn\anaconda3\lib\site-packages\keras\engine\training.py", line 1082, in compute_loss
return self.compiled_loss(
File "C:\Users\16agn\anaconda3\lib\site-packages\keras\engine\compile_utils.py", line 265, in __call__
loss_value = loss_obj(y_t, y_p, sample_weight=sw)
File "C:\Users\16agn\anaconda3\lib\site-packages\keras\losses.py", line 152, in __call__
losses = call_fn(y_true, y_pred)
File "C:\Users\16agn\anaconda3\lib\site-packages\keras\losses.py", line 284, in call
return ag_fn(y_true, y_pred, **self._fn_kwargs)
File "C:\Users\16agn\anaconda3\lib\site-packages\keras\losses.py", line 1499, in mean_squared_error
y_true = tf.cast(y_true, y_pred.dtype)
Node: 'mean_squared_error/Cast'
Cast string to float is not supported
[[{{node mean_squared_error/Cast}}]] [Op:__inference_train_function_11279]
Please help me to solve my problem, I don't know how to make the program works
The error is due to a mismatch between the target data and the output layer in the model. The target data is of shape (N, 1), but the output layer of the model has shape (N,), which means the model is expecting a 1D array instead of a 2D array. You can resolve the issue by reshaping the target data to a 1D array using the reshape method from numpy:
y_train = y_train.reshape(-1,)
y_test = y_test.reshape(-1,)
Before fitting the model:
model.fit(X_train, y_train ,batch_size = 20, epochs = 50, verbose=1)
I am building a datascience model using tensorflow and i got this error and i can't figure out how to resolve it.
I am using ipynb.
import numpy as np
import tensorflow as tf
npz = np.load('Audiobooks_data_train.npz')
train_inputs = npz['inputs'].astype(np.single)
train_targets = npz['targets'].astype(np.intc)
npz = np.load('Audiobooks_data_validation.npz')
validation_inputs, validation_targets = npz['inputs'].astype(np.single), npz['targets'].astype(np.intc)
npz = np.load('Audiobooks_data_test.npz')
test_inputs, test_targets = npz['inputs'].astype(np.single), npz['targets'].astype(np.intc)
input_size = 10
output_size = 2
hidden_layer_size = 50
model = tf.keras.Sequential([
tf.keras.layers.Dense(hidden_layer_size, activation='relu'),
tf.keras.layers.Dense(hidden_layer_size, activation='relu'),
tf.keras.layers.Dense(output_size, activation='softmax')
])
model.compile(optimizer='adam', loss='sparse_categorical_crossentropy', metrics=['accuracy'])
batch_size = 100
max_epochs = 100
early_stopping = tf.keras.callbacks.EarlyStopping(patience=2)
model.fit(train_inputs,
train_targets,
batch_size=batch_size,
epochs=max_epochs,
callbacks=[early_stopping],
validation_data=(validation_inputs, validation_targets),
verbose = 2
)
Epoch 1/100
--------------------------------------------------------------------------- InvalidArgumentError Traceback (most recent call
last) d:\data
science\TensorFlow_Audiobooks_Machine_Learning_with_comments.ipynb
Cell 9 in <cell line: 40>()
36 early_stopping = tf.keras.callbacks.EarlyStopping(patience=2)
38 # fit the model
39 # note that this time the train, validation and test data are not iterable
---> 40 model.fit(train_inputs, # train inputs
41 train_targets, # train targets
42 batch_size=batch_size, # batch size
43 epochs=max_epochs, # epochs that we will train for (assuming early stopping doesn't kick in)
44 # callbacks are functions called by a task when a task is completed
45 # task here is to check if val_loss is increasing
46 callbacks=[early_stopping], # early stopping
47 validation_data=(validation_inputs, validation_targets), # validation data
48 verbose = 2 # making sure we get enough information about the training process
49 )
File
c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\keras\utils\traceback_utils.py:70,
in filter_traceback..error_handler(*args, **kwargs)
67 filtered_tb = _process_traceback_frames(e.traceback)
68 # To get the full stack trace, call:
69 # tf.debugging.disable_traceback_filtering()
---> 70 raise e.with_traceback(filtered_tb) from None
71 finally:
72 del filtered_tb
File
c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\tensorflow\python\eager\execute.py:52,
in quick_execute(op_name, num_outputs, inputs, attrs, ctx, name)
50 try:
51 ctx.ensure_initialized()
---> 52 tensors = pywrap_tfe.TFE_Py_Execute(ctx._handle, device_name, op_name,
53 inputs, attrs, num_outputs)
54 except core._NotOkStatusException as e:
55 if name is not None:
InvalidArgumentError: Graph execution error:
Detected at node
'sparse_categorical_crossentropy/SparseSoftmaxCrossEntropyWithLogits/SparseSoftmaxCrossEntropyWithLogits'
defined at (most recent call last):
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\runpy.py",
line 196, in _run_module_as_main
return _run_code(code, main_globals, None,
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\runpy.py",
line 86, in _run_code
exec(code, run_globals)
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\ipykernel_launcher.py",
line 17, in
app.launch_new_instance()
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\traitlets\config\application.py",
line 846, in launch_instance
app.start()
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\ipykernel\kernelapp.py",
line 712, in start
self.io_loop.start()
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\tornado\platform\asyncio.py",
line 199, in start
self.asyncio_loop.run_forever()
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\asyncio\base_events.py",
line 600, in run_forever
self._run_once()
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\asyncio\base_events.py",
line 1896, in _run_once
handle._run()
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\asyncio\events.py",
line 80, in _run
self._context.run(self._callback, *self._args)
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\ipykernel\kernelbase.py",
line 504, in dispatch_queue
await self.process_one()
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\ipykernel\kernelbase.py",
line 493, in process_one
await dispatch(*args)
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\ipykernel\kernelbase.py",
line 400, in dispatch_shell
await result
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\ipykernel\kernelbase.py",
line 724, in execute_request
reply_content = await reply_content
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\ipykernel\ipkernel.py",
line 383, in do_execute
res = shell.run_cell(
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\ipykernel\zmqshell.py",
line 528, in run_cell
return super().run_cell(*args, **kwargs)
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\IPython\core\interactiveshell.py",
line 2880, in run_cell
result = self._run_cell(
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\IPython\core\interactiveshell.py",
line 2935, in _run_cell
return runner(coro)
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\IPython\core\async_helpers.py",
line 129, in pseudo_sync_runner
coro.send(None)
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\IPython\core\interactiveshell.py",
line 3134, in run_cell_async
has_raised = await self.run_ast_nodes(code_ast.body, cell_name,
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\IPython\core\interactiveshell.py",
line 3337, in run_ast_nodes
if await self.run_code(code, result, async=asy):
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\IPython\core\interactiveshell.py",
line 3397, in run_code
exec(code_obj, self.user_global_ns, self.user_ns)
File "C:\Users\nakul\AppData\Local\Temp\ipykernel_10980\600390782.py", line
40, in <cell line: 40>
model.fit(train_inputs, # train inputs
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\keras\utils\traceback_utils.py",
line 65, in error_handler
return fn(*args, **kwargs)
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\keras\engine\training.py",
line 1650, in fit
tmp_logs = self.train_function(iterator)
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\keras\engine\training.py",
line 1249, in train_function
return step_function(self, iterator)
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\keras\engine\training.py",
line 1233, in step_function
outputs = model.distribute_strategy.run(run_step, args=(data,))
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\keras\engine\training.py",
line 1222, in run_step
outputs = model.train_step(data)
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\keras\engine\training.py",
line 1024, in train_step
loss = self.compute_loss(x, y, y_pred, sample_weight)
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\keras\engine\training.py",
line 1082, in compute_loss
return self.compiled_loss(
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\keras\engine\compile_utils.py",
line 265, in call
loss_value = loss_obj(y_t, y_p, sample_weight=sw)
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\keras\losses.py",
line 152, in call
losses = call_fn(y_true, y_pred)
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\keras\losses.py",
line 284, in call
return ag_fn(y_true, y_pred, **self._fn_kwargs)
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\keras\losses.py",
line 2098, in sparse_categorical_crossentropy
return backend.sparse_categorical_crossentropy(
File "c:\Users\nakul\AppData\Local\Programs\Python\Python310\lib\site-packages\keras\backend.py",
line 5633, in sparse_categorical_crossentropy
res = tf.nn.sparse_softmax_cross_entropy_with_logits( Node: 'sparse_categorical_crossentropy/SparseSoftmaxCrossEntropyWithLogits/SparseSoftmaxCrossEntropyWithLogits'
logits and labels must have the same first dimension, got logits shape
[100,2] and labels shape [1000] [[{{node
sparse_categorical_crossentropy/SparseSoftmaxCrossEntropyWithLogits/SparseSoftmaxCrossEntropyWithLogits}}]]
[Op:__inference_train_function_974]
i couldn't figure out the cause of error
So, I'm having a problem with fitting UNet model in Tensorflow. From this line of code
hist = unet.fit(train,
validation_data=val,
steps_per_epoch=STEPS_PER_EPOCH,
validation_steps=VALIDATION_STEPS,
epochs=50)
This is the error message I get
Traceback (most recent call last):
File "C:\Users\Fedor\OneDrive\Рабочий стол\Проект\Херня.py", line 178, in <module>
hist = unet.fit(train,
File "C:\Python\lib\site-packages\keras\utils\traceback_utils.py", line 70, in error_handler
raise e.with_traceback(filtered_tb) from None
File "C:\Python\lib\site-packages\tensorflow\python\eager\execute.py", line 54, in quick_execute
tensors = pywrap_tfe.TFE_Py_Execute(ctx._handle, device_name, op_name,
tensorflow.python.framework.errors_impl.InvalidArgumentError: Graph execution error:
Detected at node 'sparse_categorical_crossentropy/SparseSoftmaxCrossEntropyWithLogits/SparseSoftmaxCrossEntropyWithLogits' defined at (most recent call last):
File "C:\Users\Fedor\OneDrive\Рабочий стол\Проект\Херня.py", line 178, in <module>
hist = unet.fit(train,
File "C:\Python\lib\site-packages\keras\utils\traceback_utils.py", line 65, in error_handler
return fn(*args, **kwargs)
File "C:\Python\lib\site-packages\keras\engine\training.py", line 1564, in fit
tmp_logs = self.train_function(iterator)
File "C:\Python\lib\site-packages\keras\engine\training.py", line 1160, in train_function
return step_function(self, iterator)
File "C:\Python\lib\site-packages\keras\engine\training.py", line 1146, in step_function
outputs = model.distribute_strategy.run(run_step, args=(data,))
File "C:\Python\lib\site-packages\keras\engine\training.py", line 1135, in run_step
outputs = model.train_step(data)
File "C:\Python\lib\site-packages\keras\engine\training.py", line 994, in train_step
loss = self.compute_loss(x, y, y_pred, sample_weight)
File "C:\Python\lib\site-packages\keras\engine\training.py", line 1052, in compute_loss
return self.compiled_loss(
File "C:\Python\lib\site-packages\keras\engine\compile_utils.py", line 265, in __call__
loss_value = loss_obj(y_t, y_p, sample_weight=sw)
File "C:\Python\lib\site-packages\keras\losses.py", line 152, in __call__
losses = call_fn(y_true, y_pred)
File "C:\Python\lib\site-packages\keras\losses.py", line 272, in call
return ag_fn(y_true, y_pred, **self._fn_kwargs)
File "C:\Python\lib\site-packages\keras\losses.py", line 2084, in sparse_categorical_crossentropy
return backend.sparse_categorical_crossentropy(
File "C:\Python\lib\site-packages\keras\backend.py", line 5630, in sparse_categorical_crossentropy
res = tf.nn.sparse_softmax_cross_entropy_with_logits(
Node: 'sparse_categorical_crossentropy/SparseSoftmaxCrossEntropyWithLogits/SparseSoftmaxCrossEntropyWithLogits'
logits and labels must have the same first dimension, got logits shape [16384,59] and labels shape [49152]
[[{{node sparse_categorical_crossentropy/SparseSoftmaxCrossEntropyWithLogits/SparseSoftmaxCrossEntropyWithLogits}}]] [Op:__inference_train_function_23608]
I've viewed questions with same error, but they didn't solve my problem
I have created a transfer learning model using Resnet50. I want to perform K-fold cross-validation on my model after which I want to find the average AUC value and standard deviation. However, I am getting an error message while performing the task. I have created a separate Files.csv file which contains the image names and their corresponding labels. I am not sure if this is the correct method or not. Please let me know if there is any other process. Please find my code below:
from tensorflow.keras.applications import ResNet50
from tensorflow.keras.applications.resnet50 import preprocess_input
from tensorflow.keras import Model, layers
from tensorflow.keras.models import load_model, model_from_json
from tensorflow.keras.layers import GlobalAveragePooling2D, Dropout, Dense, Input
import numpy as np
import pandas as pd
import os
from sklearn.model_selection import KFold, StratifiedKFold
import tensorflow as tf
from tensorflow.keras.preprocessing.image import ImageDataGenerator
train_data = pd.read_csv('Files.csv')
Y = train_data[['label']]
kf = KFold(n_splits = 5)
from tensorflow.keras.preprocessing.image import ImageDataGenerator
idg = ImageDataGenerator(rescale = 1./255,
horizontal_flip=True,
rotation_range=40,
zoom_range= 0.2,
shear_range=0.2,
width_shift_range=0.2,
height_shift_range=0.2,)
validation_datagen = ImageDataGenerator(rescale = 1./255)
def get_model_name(k):
return 'model_'+str(k)+'.h5'
from keras import models
from keras.layers import Dense, Flatten
from tensorflow.keras import optimizers
from keras.applications.vgg16 import VGG16
from tensorflow.keras.applications import ResNet50
image_dir=r'D:/regionGrowing_MLT/NewSavedRGBImages/Training'
VALIDATION_ACCURACY = []
VALIDAITON_LOSS = []
save_dir = 'C:/Users/warid'
fold_var = 1
for train_index, val_index in kf.split(np.zeros(n),Y):
training_data = train_data.iloc[train_index]
validation_data = train_data.iloc[val_index]
train_data_generator = idg.flow_from_dataframe(training_data, directory = image_dir,
x_col = "filename", y_col = "label",
class_mode = "categorical", shuffle = True)
valid_data_generator = idg.flow_from_dataframe(validation_data, directory = image_dir,
x_col = "filename", y_col = "label",
class_mode = "categorical", shuffle = True)
# CREATE NEW MODEL
model = models.Sequential()
model.add(ResNet50(weights='imagenet', include_top=False, input_shape=(224,224,3)))
model.add(Flatten())
model.add(ChannelAttention(32, 8))
model.add(SpatialAttention(7))
model.add(Dense(256, activation='relu', name='fc1'))
model.add(Dense(128, activation='relu', name='fc2'))
model.add(layers.Dropout(0.5)) #### used for regularization (to aviod overfitting)
model.add(Dense(2, activation='sigmoid'))
# model.summary()
model.compile(optimizer=optimizers.Adam(learning_rate=2e-5),
loss='binary_crossentropy',
metrics=['accuracy'])
# COMPILE NEW MODEL
# CREATE CALLBACKS
checkpoint = tf.keras.callbacks.ModelCheckpoint(save_dir+get_model_name(fold_var),
monitor='val_accuracy', verbose=1,
save_best_only=True, mode='max')
callbacks_list = [checkpoint]
# There can be other callbacks, but just showing one because it involves the model name
# This saves the best model
# FIT THE MODEL
history = model.fit(train_data_generator,
epochs=num_epochs,
callbacks=callbacks_list,
validation_data=valid_data_generator)
#PLOT HISTORY
# :
# :
# LOAD BEST MODEL to evaluate the performance of the model
model.load_weights("/saved_models/model_"+str(fold_var)+".h5")
results = model.evaluate(valid_data_generator)
results = dict(zip(model.metrics_names,results))
VALIDATION_ACCURACY.append(results['accuracy'])
VALIDATION_LOSS.append(results['loss'])
tf.keras.backend.clear_session()
fold_var += 1
After running this code, I am getting the following error message:
Found 3076 validated image filenames belonging to 2 classes.
Found 769 validated image filenames belonging to 1 classes.
Epoch 1/5
---------------------------------------------------------------------------
InvalidArgumentError Traceback (most recent call last)
Input In [75], in <cell line: 7>()
39 callbacks_list = [checkpoint]
40 # There can be other callbacks, but just showing one because it involves the model name
41 # This saves the best model
42 # FIT THE MODEL
---> 43 history = model.fit(train_data_generator,
44 epochs=num_epochs,
45 callbacks=callbacks_list,
46 validation_data=valid_data_generator)
47 #PLOT HISTORY
48 # :
49 # :
50
51 # LOAD BEST MODEL to evaluate the performance of the model
52 model.load_weights("/saved_models/model_"+str(fold_var)+".h5")
File ~\anaconda3\lib\site-packages\keras\utils\traceback_utils.py:67, in filter_traceback.<locals>.error_handler(*args, **kwargs)
65 except Exception as e: # pylint: disable=broad-except
66 filtered_tb = _process_traceback_frames(e.__traceback__)
---> 67 raise e.with_traceback(filtered_tb) from None
68 finally:
69 del filtered_tb
File ~\anaconda3\lib\site-packages\tensorflow\python\eager\execute.py:54, in quick_execute(op_name, num_outputs, inputs, attrs, ctx, name)
52 try:
53 ctx.ensure_initialized()
---> 54 tensors = pywrap_tfe.TFE_Py_Execute(ctx._handle, device_name, op_name,
55 inputs, attrs, num_outputs)
56 except core._NotOkStatusException as e:
57 if name is not None:
InvalidArgumentError: Graph execution error:
Detected at node 'sequential_2/flatten_2/Reshape' defined at (most recent call last):
File "C:\Users\warid\anaconda3\lib\runpy.py", line 197, in _run_module_as_main
return _run_code(code, main_globals, None,
File "C:\Users\warid\anaconda3\lib\runpy.py", line 87, in _run_code
exec(code, run_globals)
File "C:\Users\warid\anaconda3\lib\site-packages\ipykernel_launcher.py", line 16, in <module>
app.launch_new_instance()
File "C:\Users\warid\anaconda3\lib\site-packages\traitlets\config\application.py", line 846, in launch_instance
app.start()
File "C:\Users\warid\anaconda3\lib\site-packages\ipykernel\kernelapp.py", line 677, in start
self.io_loop.start()
File "C:\Users\warid\anaconda3\lib\site-packages\tornado\platform\asyncio.py", line 199, in start
self.asyncio_loop.run_forever()
File "C:\Users\warid\anaconda3\lib\asyncio\base_events.py", line 601, in run_forever
self._run_once()
File "C:\Users\warid\anaconda3\lib\asyncio\base_events.py", line 1905, in _run_once
handle._run()
File "C:\Users\warid\anaconda3\lib\asyncio\events.py", line 80, in _run
self._context.run(self._callback, *self._args)
File "C:\Users\warid\anaconda3\lib\site-packages\ipykernel\kernelbase.py", line 471, in dispatch_queue
await self.process_one()
File "C:\Users\warid\anaconda3\lib\site-packages\ipykernel\kernelbase.py", line 460, in process_one
await dispatch(*args)
File "C:\Users\warid\anaconda3\lib\site-packages\ipykernel\kernelbase.py", line 367, in dispatch_shell
await result
File "C:\Users\warid\anaconda3\lib\site-packages\ipykernel\kernelbase.py", line 662, in execute_request
reply_content = await reply_content
File "C:\Users\warid\anaconda3\lib\site-packages\ipykernel\ipkernel.py", line 360, in do_execute
res = shell.run_cell(code, store_history=store_history, silent=silent)
File "C:\Users\warid\anaconda3\lib\site-packages\ipykernel\zmqshell.py", line 532, in run_cell
return super().run_cell(*args, **kwargs)
File "C:\Users\warid\anaconda3\lib\site-packages\IPython\core\interactiveshell.py", line 2863, in run_cell
result = self._run_cell(
File "C:\Users\warid\anaconda3\lib\site-packages\IPython\core\interactiveshell.py", line 2909, in _run_cell
return runner(coro)
File "C:\Users\warid\anaconda3\lib\site-packages\IPython\core\async_helpers.py", line 129, in _pseudo_sync_runner
coro.send(None)
File "C:\Users\warid\anaconda3\lib\site-packages\IPython\core\interactiveshell.py", line 3106, in run_cell_async
has_raised = await self.run_ast_nodes(code_ast.body, cell_name,
File "C:\Users\warid\anaconda3\lib\site-packages\IPython\core\interactiveshell.py", line 3309, in run_ast_nodes
if await self.run_code(code, result, async_=asy):
File "C:\Users\warid\anaconda3\lib\site-packages\IPython\core\interactiveshell.py", line 3369, in run_code
exec(code_obj, self.user_global_ns, self.user_ns)
File "C:\Users\warid\AppData\Local\Temp\ipykernel_37076\2928028949.py", line 43, in <cell line: 7>
history = model.fit(train_data_generator,
File "C:\Users\warid\anaconda3\lib\site-packages\keras\utils\traceback_utils.py", line 64, in error_handler
return fn(*args, **kwargs)
File "C:\Users\warid\anaconda3\lib\site-packages\keras\engine\training.py", line 1409, in fit
tmp_logs = self.train_function(iterator)
File "C:\Users\warid\anaconda3\lib\site-packages\keras\engine\training.py", line 1051, in train_function
return step_function(self, iterator)
File "C:\Users\warid\anaconda3\lib\site-packages\keras\engine\training.py", line 1040, in step_function
outputs = model.distribute_strategy.run(run_step, args=(data,))
File "C:\Users\warid\anaconda3\lib\site-packages\keras\engine\training.py", line 1030, in run_step
outputs = model.train_step(data)
File "C:\Users\warid\anaconda3\lib\site-packages\keras\engine\training.py", line 889, in train_step
y_pred = self(x, training=True)
File "C:\Users\warid\anaconda3\lib\site-packages\keras\utils\traceback_utils.py", line 64, in error_handler
return fn(*args, **kwargs)
File "C:\Users\warid\anaconda3\lib\site-packages\keras\engine\training.py", line 490, in __call__
return super().__call__(*args, **kwargs)
File "C:\Users\warid\anaconda3\lib\site-packages\keras\utils\traceback_utils.py", line 64, in error_handler
return fn(*args, **kwargs)
File "C:\Users\warid\anaconda3\lib\site-packages\keras\engine\base_layer.py", line 1014, in __call__
outputs = call_fn(inputs, *args, **kwargs)
File "C:\Users\warid\anaconda3\lib\site-packages\keras\utils\traceback_utils.py", line 92, in error_handler
return fn(*args, **kwargs)
File "C:\Users\warid\anaconda3\lib\site-packages\keras\engine\sequential.py", line 374, in call
return super(Sequential, self).call(inputs, training=training, mask=mask)
File "C:\Users\warid\anaconda3\lib\site-packages\keras\engine\functional.py", line 458, in call
return self._run_internal_graph(
File "C:\Users\warid\anaconda3\lib\site-packages\keras\engine\functional.py", line 596, in _run_internal_graph
outputs = node.layer(*args, **kwargs)
File "C:\Users\warid\anaconda3\lib\site-packages\keras\utils\traceback_utils.py", line 64, in error_handler
return fn(*args, **kwargs)
File "C:\Users\warid\anaconda3\lib\site-packages\keras\engine\base_layer.py", line 1014, in __call__
outputs = call_fn(inputs, *args, **kwargs)
File "C:\Users\warid\anaconda3\lib\site-packages\keras\utils\traceback_utils.py", line 92, in error_handler
return fn(*args, **kwargs)
File "C:\Users\warid\anaconda3\lib\site-packages\keras\layers\reshaping\flatten.py", line 98, in call
return tf.reshape(inputs, flattened_shape)
Node: 'sequential_2/flatten_2/Reshape'
Input to reshape is a tensor with 4194304 values, but the requested shape requires a multiple of 100352
[[{{node sequential_2/flatten_2/Reshape}}]] [Op:__inference_train_function_37893]
I am trying to create a model with two inputs. The model is very simple containing only one lstm layer for each input. The problem is that I want to provide lists of different length as inputs. For that, I am using ragged tensors, but the training process fails.
ds = pd.DataFrame({"col_1":[[0],[0,0],[0,0,0],[0,0,0,0],[0,0,0,0,0],[0,0,0,0,0,0],[0,0,0,0,0,0,0],[0,0,0,0,0,0,0,0]],"col_2":[8*[0],7*[1],6*[2],5*[3],4*[4],3*[5],2*[6],1*[7]]})
ds = ds.loc[ds.index.repeat(1250)].reset_index(drop=True)
ds = ds.sample(frac=1, random_state=43).reset_index(drop=True)
feat_1_inputs = [tf.keras.layers.Input(batch_shape=(None,None,1),ragged=True,name="col_1")]
feat_1 = tf.keras.layers.LSTM(10, return_sequences=True, return_state=False, stateful=False)(feat_1_inputs[0])
feat_2_inputs = [tf.keras.layers.Input(batch_shape=(None,None,1),ragged=True,name="col_2")]
feat_2 = tf.keras.layers.LSTM(10, return_sequences=True, return_state=False, stateful=False)(feat_2_inputs[0])
concat_inputs = tf.keras.layers.Concatenate()([feat_1, feat_2])
output = tf.keras.layers.Dense(10, activation='relu',kernel_initializer=glorot_uniform())(concat_inputs)
output = tf.keras.layers.Dense(10, kernel_initializer=glorot_uniform())(output)
output = tf.keras.layers.Activation(activation='softmax', dtype='float32')(output)
model = tf.keras.Model(feat_1_inputs + feat_2_inputs, output)
model.compile(optimizer=tf.keras.optimizers.Adam(learning_rate=0.001), loss=tf.keras.losses.sparse_categorical_crossentropy)
col_1_data = [tf.expand_dims(tf.ragged.constant(ds['col_1'].values,dtype=np.int64),axis=-1)]
col_2_data = tf.expand_dims(tf.ragged.constant(ds['col_2'].values,dtype=np.int64),axis=-1)
col_1_data.append(col_2_data)
model.fit(x=col_1_data,y=col_2_data,epochs=10)
Error:
Epoch 1/10
Traceback (most recent call last):
File "/home/user/.config/JetBrains/PyCharmCE2021.2/scratches/scratch_19.py", line 33, in <module>
model.fit(x=col_1_data,y=col_2_data,epochs=10)
File "/home/user/miniconda3/envs/model/lib/python3.9/site-packages/keras/utils/traceback_utils.py", line 67, in error_handler
raise e.with_traceback(filtered_tb) from None
File "/home/user/miniconda3/envs/model/lib/python3.9/site-packages/tensorflow/python/eager/execute.py", line 54, in quick_execute
tensors = pywrap_tfe.TFE_Py_Execute(ctx._handle, device_name, op_name,
tensorflow.python.framework.errors_impl.InvalidArgumentError: Graph execution error:
Detected at node 'model/concatenate/RaggedConcat/assert_equal_1/Assert/AssertGuard/Assert' defined at (most recent call last):
File "/home/user/.config/JetBrains/PyCharmCE2021.2/scratches/scratch_19.py", line 33, in <module>
model.fit(x=col_1_data,y=col_2_data,epochs=10)
File "/home/user/miniconda3/envs/model/lib/python3.9/site-packages/keras/utils/traceback_utils.py", line 64, in error_handler
return fn(*args, **kwargs)
File "/home/user/miniconda3/envs/model/lib/python3.9/site-packages/keras/engine/training.py", line 1384, in fit
tmp_logs = self.train_function(iterator)
File "/home/user/miniconda3/envs/model/lib/python3.9/site-packages/keras/engine/training.py", line 1021, in train_function
return step_function(self, iterator)
File "/home/user/miniconda3/envs/model/lib/python3.9/site-packages/keras/engine/training.py", line 1010, in step_function
outputs = model.distribute_strategy.run(run_step, args=(data,))
File "/home/user/miniconda3/envs/model/lib/python3.9/site-packages/keras/engine/training.py", line 1000, in run_step
outputs = model.train_step(data)
File "/home/user/miniconda3/envs/model/lib/python3.9/site-packages/keras/engine/training.py", line 859, in train_step
y_pred = self(x, training=True)
File "/home/user/miniconda3/envs/model/lib/python3.9/site-packages/keras/utils/traceback_utils.py", line 64, in error_handler
return fn(*args, **kwargs)
File "/home/user/miniconda3/envs/model/lib/python3.9/site-packages/keras/engine/base_layer.py", line 1096, in __call__
outputs = call_fn(inputs, *args, **kwargs)
File "/home/user/miniconda3/envs/model/lib/python3.9/site-packages/keras/utils/traceback_utils.py", line 92, in error_handler
return fn(*args, **kwargs)
File "/home/user/miniconda3/envs/model/lib/python3.9/site-packages/keras/engine/functional.py", line 451, in call
return self._run_internal_graph(
File "/home/user/miniconda3/envs/model/lib/python3.9/site-packages/keras/engine/functional.py", line 589, in _run_internal_graph
outputs = node.layer(*args, **kwargs)
File "/home/user/miniconda3/envs/model/lib/python3.9/site-packages/keras/utils/traceback_utils.py", line 64, in error_handler
return fn(*args, **kwargs)
File "/home/user/miniconda3/envs/model/lib/python3.9/site-packages/keras/engine/base_layer.py", line 1096, in __call__
outputs = call_fn(inputs, *args, **kwargs)
File "/home/user/miniconda3/envs/model/lib/python3.9/site-packages/keras/utils/traceback_utils.py", line 92, in error_handler
return fn(*args, **kwargs)
File "/home/user/miniconda3/envs/model/lib/python3.9/site-packages/keras/layers/merge.py", line 183, in call
return self._merge_function(inputs)
File "/home/user/miniconda3/envs/model/lib/python3.9/site-packages/keras/layers/merge.py", line 531, in _merge_function
return backend.concatenate(inputs, axis=self.axis)
File "/home/user/miniconda3/envs/model/lib/python3.9/site-packages/keras/backend.py", line 3311, in concatenate
return tf.concat(tensors, axis)
Node: 'model/concatenate/RaggedConcat/assert_equal_1/Assert/AssertGuard/Assert'
assertion failed: [Inputs must have identical ragged splits] [Condition x == y did not hold element-wise:] [x (model/lstm/RaggedFromTensor/concat:0) = ] [0 8 11...] [y (model/lstm_1/RaggedFromTensor/concat:0) = ] [0 1 7...]
[[{{node model/concatenate/RaggedConcat/assert_equal_1/Assert/AssertGuard/Assert}}]] [Op:__inference_train_function_9256]
If rows in both columns contain lists of the same length then it works fine.
Is there a way to work with lists of different length using ragged tensors?
TF2.8 is used.