Naive CL Strategy on Permuted MNIST #412
-
Hello! I am applying the naive CL strategy to the Permuted MNIST scenario and it seems the naive method doesn't demonstrate catastrophic forgetting. After training each new experience up to 3 experiences, the evaluation accuracies are surprisingly high for all previous experiences (~90% and above with only 1 training epoch). Are these results to be expected using the Permuted MNIST dataset? Below is my code and output. Appreciate any and all feedback :)
|
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
Hi @christam96 ! Thank you for reaching out. Yes, I think your results are correct. PMNIST is a setting which shows relatively low forgetting. Moreover, with only one epoch the model is able to retain most of the original performance. Forgetting should increase if you increase the number of training epochs (e.g. from 1 to 5-10), since you focus more and more on the current data, at the expenses of previous experiences. You can also try to use more experiences, since in the literature PMNIST is usually used with 10 or more experiences. You should see a larger forgetting on the first experience as you add more experiences. |
Beta Was this translation helpful? Give feedback.
Hi @christam96 ! Thank you for reaching out. Yes, I think your results are correct. PMNIST is a setting which shows relatively low forgetting. Moreover, with only one epoch the model is able to retain most of the original performance. Forgetting should increase if you increase the number of training epochs (e.g. from 1 to 5-10), since you focus more and more on the current data, at the expenses of previous experiences. You can also try to use more experiences, since in the literature PMNIST is usually used with 10 or more experiences. You should see a larger forgetting on the first experience as you add more experiences.