Hi there,
I am facing question of using the distillation cfg.
I have trained my teacher network with the observation group:
obs_groups = {"policy": ["policy", "privileged_info"], "critic": ["policy", "privileged_info"]}
Then in the rsl_rl_distillation_cfg.py, I updated the observation group as followed:
obs_groups = {"policy": ["policy" ], "teacher": ["policy", "privileged_info"]}
However, the behavior loss, as I saw from wandb, is getting huge and can up to 100 and it doesn't go down.
What are the correct way to setting up Teacher-Student Training in IsaacLab?