Artwork

เนื้อหาจัดทำโดย Katie Harbath เนื้อหาพอดแคสต์ทั้งหมด รวมถึงตอน กราฟิก และคำอธิบายพอดแคสต์ได้รับการอัปโหลดและจัดหาให้โดยตรงจาก Katie Harbath หรือพันธมิตรแพลตฟอร์มพอดแคสต์ของพวกเขา หากคุณเชื่อว่ามีบุคคลอื่นใช้งานที่มีลิขสิทธิ์ของคุณโดยไม่ได้รับอนุญาต คุณสามารถปฏิบัติตามขั้นตอนที่แสดงไว้ที่นี่ https://th.player.fm/legal
Player FM - แอป Podcast
ออฟไลน์ด้วยแอป Player FM !

AI and Human Rights: A Balancing Act

46:42
 
แบ่งปัน
 

Manage episode 440708521 series 3508712
เนื้อหาจัดทำโดย Katie Harbath เนื้อหาพอดแคสต์ทั้งหมด รวมถึงตอน กราฟิก และคำอธิบายพอดแคสต์ได้รับการอัปโหลดและจัดหาให้โดยตรงจาก Katie Harbath หรือพันธมิตรแพลตฟอร์มพอดแคสต์ของพวกเขา หากคุณเชื่อว่ามีบุคคลอื่นใช้งานที่มีลิขสิทธิ์ของคุณโดยไม่ได้รับอนุญาต คุณสามารถปฏิบัติตามขั้นตอนที่แสดงไว้ที่นี่ https://th.player.fm/legal

We’re back! Welcome to season three of Impossible Tradeoffs. This season will focus mainly on the U.S. election, how the campaigns use tech, how tech companies respond, and what people outside the United States are watching.

First up is Paolo Carozza, a member of the Meta Oversight Board and law professor at Notre Dame. Earlier this week, the Oversight Board released a report on using AI in content moderation.

In this conversation, we discuss the complexities of content moderation, particularly in the context of AI and human rights. Carozza shares insights from his experience on the Oversight Board, emphasizing the importance of balancing operational feasibility with principled human rights considerations. They explore the challenges of scale in content moderation, the need for effective recommendations, and the role of transparency and accountability in the board's work. The discussion also delves into the findings from a recent report on AI's impact on content moderation, the difficulties of understanding context, and the board's approach to handling controversial cases.

Here are some of the key moments:

* Introduction to Paolo Carozza and the Oversight Board

* Challenges of Scale in Content Moderation

* Balancing Human Rights and Operational Feasibility

* Implementing Effective Recommendations

* Navigating Transparency and Accountability

* Findings from the AI Report

* Understanding Context in Content Moderation

* The Evolution of Trust and Safety Practices

* Expedited Decisions in Urgent Contexts

* Handling Controversial Cases and Public Reactions

During the conversation, I mentioned how I wrote a piece about how to write effective recommendations for tech companies based on things the Oversight Board learned. You can find that here.

Enjoy!

Please support the curation and analysis I’m doing with this newsletter. As a paid subscriber, you make it possible for me to bring you in-depth analyses of the most pressing issues in tech and politics.

Get full access to Anchor Change with Katie Harbath at anchorchange.substack.com/subscribe

  continue reading

50 ตอน

Artwork
iconแบ่งปัน
 
Manage episode 440708521 series 3508712
เนื้อหาจัดทำโดย Katie Harbath เนื้อหาพอดแคสต์ทั้งหมด รวมถึงตอน กราฟิก และคำอธิบายพอดแคสต์ได้รับการอัปโหลดและจัดหาให้โดยตรงจาก Katie Harbath หรือพันธมิตรแพลตฟอร์มพอดแคสต์ของพวกเขา หากคุณเชื่อว่ามีบุคคลอื่นใช้งานที่มีลิขสิทธิ์ของคุณโดยไม่ได้รับอนุญาต คุณสามารถปฏิบัติตามขั้นตอนที่แสดงไว้ที่นี่ https://th.player.fm/legal

We’re back! Welcome to season three of Impossible Tradeoffs. This season will focus mainly on the U.S. election, how the campaigns use tech, how tech companies respond, and what people outside the United States are watching.

First up is Paolo Carozza, a member of the Meta Oversight Board and law professor at Notre Dame. Earlier this week, the Oversight Board released a report on using AI in content moderation.

In this conversation, we discuss the complexities of content moderation, particularly in the context of AI and human rights. Carozza shares insights from his experience on the Oversight Board, emphasizing the importance of balancing operational feasibility with principled human rights considerations. They explore the challenges of scale in content moderation, the need for effective recommendations, and the role of transparency and accountability in the board's work. The discussion also delves into the findings from a recent report on AI's impact on content moderation, the difficulties of understanding context, and the board's approach to handling controversial cases.

Here are some of the key moments:

* Introduction to Paolo Carozza and the Oversight Board

* Challenges of Scale in Content Moderation

* Balancing Human Rights and Operational Feasibility

* Implementing Effective Recommendations

* Navigating Transparency and Accountability

* Findings from the AI Report

* Understanding Context in Content Moderation

* The Evolution of Trust and Safety Practices

* Expedited Decisions in Urgent Contexts

* Handling Controversial Cases and Public Reactions

During the conversation, I mentioned how I wrote a piece about how to write effective recommendations for tech companies based on things the Oversight Board learned. You can find that here.

Enjoy!

Please support the curation and analysis I’m doing with this newsletter. As a paid subscriber, you make it possible for me to bring you in-depth analyses of the most pressing issues in tech and politics.

Get full access to Anchor Change with Katie Harbath at anchorchange.substack.com/subscribe

  continue reading

50 ตอน

ทุกตอน

×
 
Loading …

ขอต้อนรับสู่ Player FM!

Player FM กำลังหาเว็บ

 

คู่มืออ้างอิงด่วน