7 December 2023
By: Girish kumar anshul
Regulating AI is complex and faces challenges like defining AI and addressing opaque black box algorithms, but global efforts are underway to promote responsible development and mitigate risks.
AI's transformative potential has risen rapidly, sparking global efforts to regulate its development and address potential risks.
– G7 Hiroshima AI Process: Adopted international principles and code of conduct for developing advanced AI systems. – US Executive Order: Emphasises safety, security, and trust in AI development and use.
– UK Bletchley Declaration: Focuses on responsible AI development and safety. – EU AI Act: Classifies AI technologies by risk, prohibits certain practices, and promotes transparency and accountability.
– Frontier Model Forum: Leading companies collaborate to anticipate regulations and promote responsible AI use. – White House AI Pledge: Companies voluntarily commit to managing AI risks.
– Defining AI: No single definition exists, making regulation complex. – Black box problem: Algorithmic decision-making can be opaque, hindering accountability and transparency. – Balancing innovation and risk mitigation: Regulations should not stifle innovation.
– Focus on responsible use: Human-centric design, controls, risk management throughout the AI lifecycle. – Trustworthy technology: Fairness, transparency, bias mitigation, accountability in decision-making.
EU AI Act: Prohibits high-risk practices like facial recognition. US Executive Order: Requires advanced AI developers to share safety tests. – India: No specific AI regulations yet, but recent court ruling recognizes limitations of AI.
This visual story utilises symbolic imagery and metaphors to convey the complex topic of AI regulation in a clear and engaging way.