A new teacher-student framework called Distill-Belief decouples Bayesian correctness from computational efficiency. It uses a particle-filter teacher to provide dense information-gain signals, preventing a student model from exploiting approximation errors. This approach stops reward hacking during inverse source localization. Practitioners gain a faster way to infer latent field parameters without sacrificing uncertainty estimation.