https://huggingface.co/google/gemma-3n-E4B-it-litert-lm?utm_source=chatgpt.com
States a context length of 32k.
The ReadMe of this Repo states supported context is 4096.
Hence: Is it not possible to run inference with more than 4096 tokens here, or is this a legacy description?