Replies: 1 comment
-
This may very well be the case, depending on the model/data that you are trying to represent. Just like any machine learning model, a GP can fall into bad local minima if your data are complex or not preprocessed in an appropriate way. Many GP kernels can also struggle with high dimensional data. I would z-score your data, and try scaling down to a small problem (i.e. set D=10), and see if the issue persists.
You need to run There could be many reasons why this is taking lots of GPU memory (GPs are notoriously memory intensive), and there are modifications that you can make to the model to reduce this. If you want any feedback on this, you'll need to give us more information (i.e. what is N, D, what is the data that you are using) as well as a runable code example. |
Beta Was this translation helpful? Give feedback.
-
Hi all,
I'm new to Gpytorch, so please forgive me if my questions seem too basic. I'm currently using the Multioutput Gaussian Process (Multitask GP) for handling inputs and outputs with shapes
(N, D)
, whereN
represents the number of samples (in this case, images) andD
represents the corresponding flattened pixels.I have two specific doubts:
In the prediction phase, I have noticed that the columns in the var_pred matrix
(shape: (N_t, D))
have the same values. In other words, the model is not learning the covariance in theN
direction. As a result, when I plot the variance, it simply appears as noise.When I use
observed_pred = likelihood(model(x_t))
, it consumes a significantly large amount of GPU memory (90+ GB). However, when I use onlylikelihood(x_t)
, it consumes much less GPU memory. Here,x_t
represents the test data with shape (N, D).Could you please help me understand these issues and provide any possible solutions?
Beta Was this translation helpful? Give feedback.
All reactions