Skip to content Skip to sidebar Skip to footer

Change Loss Function Dynamically During Training In Keras, Without Recompiling Other Model Properties Like Optimizer

Is it possible to set model.loss in a callback without re-compiling model.compile(...) after (since then the optimizer states are reset), and just recompiling model.loss, like for

Solution 1:

I hope you have found a solution to your problem by now but using tensorflow I think you can solve this by building a custom training loop (here is the doc). this will not override the loss attribute as you requested however you can probably achieve what you are looking for.

example

initializing variable

modifying the example from the documentation, with a model and dataset as such:

inputs = tf.keras.Input(shape=(784,), name="digits")
x1 = tf.keras.layers.Dense(64, activation="relu")(inputs)
x2 = tf.keras.layers.Dense(64, activation="relu")(x1)
outputs = tf.keras.layers.Dense(10, name="predictions")(x2)
model = tf.keras.Model(inputs=inputs, outputs=outputs)


# Prepare the training dataset.
batch_size = 64
(x_train, y_train), (x_test, y_test) = tf.keras.datasets.mnist.load_data()
x_train = np.reshape(x_train, (-1, 784))
x_test = np.reshape(x_test, (-1, 784))
train_dataset = tf.data.Dataset.from_tensor_slices((x_train, y_train))
train_dataset = train_dataset.shuffle(buffer_size=1024).batch(batch_size)

we can define our two loss functions (the two I chose make no sense from a scientific point of view but allow us to check the code works)

# Instantiate an optimizer.optimizer = tf.keras.optimizers.Adam(learning_rate=1e-3)
# Instantiate a loss function.loss_1 = tf.keras.losses.SparseCategoricalCrossentropy(from_logits=True)
loss_2 = lambda y_true, y_pred: -1 * loss_1(y_true, y_pred)

training loop

we can then execute our custom training loop:

epochs = 10for epoch inrange(epochs):
    print("\nStart of epoch %d" % (epoch,))

    # Iterate over the batches of the dataset.for step, (x_batch_train, y_batch_train) inenumerate(train_dataset):

        # Open a GradientTape to record the operations run# during the forward pass, which enables auto-differentiation.
        loss_fn = loss_1 if epoch % 2else loss_2
        with tf.GradientTape() as tape:

            # Run the forward pass of the layer.# The operations that the layer applies# to its inputs are going to be recorded# on the GradientTape.
            logits = model(x_batch_train, training=True)  # Logits for this minibatch# Compute the loss value for this minibatch.
            loss_value = loss_fn(y_batch_train, logits)

        # Use the gradient tape to automatically retrieve# the gradients of the trainable variables with respect to the loss.
        grads = tape.gradient(loss_value, model.trainable_weights)

        # Run one step of gradient descent by updating# the value of the variables to minimize the loss.
        optimizer.apply_gradients(zip(grads, model.trainable_weights))
         # Log every 200 batches.if step % 200 == 0:
            print(
                "Training loss (for one batch) at step %d: %.4f"
                % (step, float(loss_value))
            )
            print("Seen so far: %s samples" % ((step + 1) * 64))

and we check the output is what we want (alternate positive and negative losses)

Startofepoch0Trainingloss(foronebatch)at step 0:-96.1003Seen so far:64samplesTrainingloss(foronebatch)at step 200:-3383849.5000Seen so far:12864samplesTrainingloss(foronebatch)at step 400:-40419124.0000Seen so far:25664samplesTrainingloss(foronebatch)at step 600:-149133008.0000Seen so far:38464samplesTrainingloss(foronebatch)at step 800:-328322816.0000Seen so far:51264samplesStartofepoch1Trainingloss(foronebatch)at step 0:580457984.0000Seen so far:64samplesTrainingloss(foronebatch)at step 200:297710528.0000Seen so far:12864samplesTrainingloss(foronebatch)at step 400:213328544.0000Seen so far:25664samplesTrainingloss(foronebatch)at step 600:159328976.0000Seen so far:38464samplesTrainingloss(foronebatch)at step 800:105737024.0000Seen so far:51264samples

drawbacks and further improvments

the problem with writing custom loops as such is that you will loose the convenience of keras's fit method. I think you can manage this by defining a custom model and overriding the train_step as shown here in the documentation

If you really need to have the loss attribute of your model changed, you can set the compiled_loss attribute using a keras.engine.compile_utils.LossesContainer (here is the reference) and set model.train_function to model.make_train_function() (so that the new loss gets taken into account).

Post a Comment for "Change Loss Function Dynamically During Training In Keras, Without Recompiling Other Model Properties Like Optimizer"