/soft_constraint_irl

In this project, we use the maximum entropy principle in Inverse reinforcement learning to learn soft constraints from demonstrations obtained from an agent interacting with a non-deterministic MDP. In the second part of this project, we implement various strategies (orchestrators) to mix conflicting policies (e.g. pragmatic vs ethical). In one of these orchestrators, we use a cognitive model of decision making (MDFT) to enable the agent to make human-like decisions.

Primary LanguagePythonMIT LicenseMIT

Issues