Replies: 4 comments 4 replies
-
|
Hi! NLE and NRE do have advantages for iid-observations: During training, they need only a single simulation per parameter set. This typically makes them less simulation hungry and less prone to issues such as memory outage. The downside of these methods is that sampling can become relatively slow, especially for many iid datapoints. I would probably still give it a shot with one of those two methods. Hope this helps |
Beta Was this translation helpful? Give feedback.
-
|
adding to the GPU memory problem: at inference time you would only need the net and the batch iid samples on the GPU, therefore I am surprised about the overflow. Just an idea but you could try recreating the posterior object from scratch, e.g., given your already trained does this help? |
Beta Was this translation helpful? Give feedback.
-
|
Hi, Thank you for the answers. I initially started with NRE since I needed to use an embedding network for my observation, and I had many ID data points. So, right now, my biggest limitation is the GPU memory during inference time. The whole pipeline gives reasonable results as long as I use a single observation. When I try to use five or more, I always run out of GPU memory. I started to play around with mcmc and vi samplings, but for both, I ran into VRAM issues. Is there maybe somewhere a parameter for the batch size which could be tuned ? |
Beta Was this translation helpful? Give feedback.
-
|
Hello, Thank you for your responses! I created a minimal example that produces the same error. Currently, the simulator has two parameters, but it may be expanded to include between 15 and 30 parameters in the future. In this minimal example, the simulator generates only noise, while the actual simulator produces a time series of a similar dimension. I also simplified the embedding network, while the total number of trainable parameters is the same. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hello everybody,
I have a problem, in which I have many iid observations, which I need to combine to get a precise inference.
Probably I will have around 100 to 1000 ii samples. The observations are time series which are compressed for now with an embedding network. Right now im using NRE with MCMC sampling. But my problem here is that I quickly run out of GPU memory when I try to make inference on many iid observations.
Are there other methods which are more suitable in this case?
Beta Was this translation helpful? Give feedback.
All reactions