Responsible AI Adherence of Language Models
Major factors affecting trustworthiness of Language Models
We intend to explore and utilize the Performance, Fairness & Bias aspects of Language Models to quantify adherence of Responsible AI (RAI)
Motivation: Why Safe and Explainable RL ?
Safety: In real-world applications, like surgical robotics or self driving cars deploying models without safety considerations can lead to undesirable consequences. Safe RL ensures that the learned policies adhere to certain constraints, preventing harmful actions.
Explainability: Understanding the decisions made by RL models is crucial for users, stakeholders, and regulators. Explainable RL helps build trust in the system by providing clear insights into why a particular decision or action was taken.
Legal and Ethical Compliance: Many industries are subject to regulations and ethical standards. Safe and explainable RL helps in complying with these requirements, avoiding legal issues and ensuring responsible AI deployment.
Notion of Safety