Conversation
|
Nice! Yes, I would love to search for |
|
Not any I could think of, except some overhead on the user's side which may be unavoidable. @arrjon if you concur, I will proceed with adding the search for attention mask among simulation outputs. |
|
Sounds good to me as it gives maximal flexibility to the user! |
|
I now allowed for The case for sampling is tricky tho, as now the user will have to provide the |
|
Can you elaborate why we cannot use |
|
I can solve it with a bunch of checks for now. |
|
I don’t fully understand what you mean by this. Solve what? Making it part
of conditions?
Stefan Radev ***@***.***> schrieb am Di. 3. Feb. 2026 um
14:56:
… *stefanradev93* left a comment (bayesflow-org/bayesflow#627)
<#627 (comment)>
I can solve it with a bunch of checks for now.
—
Reply to this email directly, view it on GitHub
<#627 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ADCW2ADNWQSQIJOVGN64ALD4KCSJVAVCNFSM6AAAAACTL6MU52VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZTQNBRGQ4DEMJXGA>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
|
@arrjon Can you please check if my latest commit enables the functionality you needed? |
This PR refactors the transformers module for internal consistency and directly exposes
attention_maskanduse_causal_maskin the relevant transformers. The following changes were made:attentionmodule.Transformerwas added to easily tell apart transformer summaries from other summary nets.TimeSeriesTransformercan now act as a many-to-many network (e.g., for modeling time-varying targets)It also prepares to address ##626.
@arrjon @paul-buerkner It remains to decide how we want the attention mask passed? Should we search for it in the simulator outputs (as we do for other special args like
sample_weights)? This would have the advantage that the mask can be constructed very flexibly.