1:"$Sreact.fragment" 2:I[22016,["/_next/static/chunks/0y90oe1t50yh1.js","/_next/static/chunks/0rq1gphgwlz40.js","/_next/static/chunks/0d3shmwh5_nmn.js"],""] 3:I[58541,["/_next/static/chunks/0y90oe1t50yh1.js","/_next/static/chunks/0rq1gphgwlz40.js","/_next/static/chunks/0d3shmwh5_nmn.js"],"default"] 6:I[97367,["/_next/static/chunks/0y90oe1t50yh1.js","/_next/static/chunks/0rq1gphgwlz40.js","/_next/static/chunks/0d3shmwh5_nmn.js"],"OutletBoundary"] 7:"$Sreact.suspense" 0:{"rsc":["$","$1","c",{"children":[["$","div",null,{"className":"py-12 sm:py-16","children":["$","div",null,{"className":"max-w-3xl mx-auto px-4 sm:px-6 lg:px-8","children":[["$","$L2",null,{"href":"/blog","className":"inline-flex items-center gap-1.5 text-sm font-medium text-muted hover:text-acm-blue transition-colors mb-8","children":[["$","$L3",null,{"iconNode":[["path",{"d":"m12 19-7-7 7-7","key":"1l729n"}],["path",{"d":"M19 12H5","key":"x3x0zl"}]],"className":"lucide-arrow-left","size":14}],"Back to Blog"]}],["$","header",null,{"className":"mb-10","children":[["$","span",null,{"className":"inline-block px-3 py-1 rounded-full text-xs font-semibold bg-acm-yellow/15 text-amber-700 mb-4","children":"Event Coverage"}],["$","h1",null,{"className":"text-3xl sm:text-4xl font-extrabold text-dark-text tracking-tight leading-tight","children":"Inside the Responsible AI Symposium: Key Moments and Takeaways"}],["$","div",null,{"className":"flex flex-wrap items-center gap-4 mt-5 text-sm text-muted","children":[["$","span",null,{"className":"font-medium text-dark-text","children":"Santiago Jerald"}],["$","span",null,{"className":"flex items-center gap-1","children":[["$","$L3",null,{"iconNode":[["path",{"d":"M8 2v4","key":"1cmpym"}],["path",{"d":"M16 2v4","key":"4m81vk"}],["rect",{"width":"18","height":"18","x":"3","y":"4","rx":"2","key":"1hopcy"}],["path",{"d":"M3 10h18","key":"8toen8"}]],"className":"lucide-calendar","size":14}],"March 25, 2026"]}],["$","span",null,{"className":"flex items-center gap-1","children":[["$","$L3",null,{"iconNode":[["circle",{"cx":"12","cy":"12","r":"10","key":"1mglay"}],["path",{"d":"M12 6v6l4 2","key":"mmk7yg"}]],"className":"lucide-clock","size":14}],"6 min read"]}]]}]]}],["$","article",null,{"className":"prose-custom","children":[["$","p","0",{"className":"text-muted leading-relaxed my-4","children":"Our Responsible AI Symposium brought together over 300 attendees from across the university and industry partners for two days of deep discussions on the future of ethical AI development."}],["$","h2","1",{"className":"text-xl sm:text-2xl font-bold text-dark-text mt-8 mb-4","children":"Day 1: Setting the Stage"}],["$","p","2",{"className":"text-muted leading-relaxed my-4","children":"The symposium kicked off with a powerful keynote on \"AI Governance in the Age of Large Language Models\" that highlighted the growing need for frameworks that ensure AI systems are fair, transparent, and accountable."}],["$","p","3",{"className":"text-muted leading-relaxed my-4","children":"The panel discussion on bias in machine learning featured perspectives from researchers, industry practitioners, and policy advocates. Key themes included the importance of diverse training data, the role of human oversight in AI decision-making, and the need for standardized auditing procedures."}],["$","h2","4",{"className":"text-xl sm:text-2xl font-bold text-dark-text mt-8 mb-4","children":"Day 2: Hands-On Learning"}],["$","p","5",{"className":"text-muted leading-relaxed my-4","children":"The second day focused on practical workshops where participants got hands-on experience with fairness-aware ML tools. Teams worked on real datasets to identify and mitigate bias, learning techniques like adversarial debiasing, calibrated equalized odds, and fairness constraints in model training."}],["$","h2","6",{"className":"text-xl sm:text-2xl font-bold text-dark-text mt-8 mb-4","children":"Student Hackathon Results"}],["$","p","7",{"className":"text-muted leading-relaxed my-4","children":"The 6-hour mini-hackathon challenged teams to build responsible AI prototypes. The winning team created an automated bias detection tool for hiring algorithms that evaluates candidate scoring models across multiple demographic dimensions."}],["$","h2","8",{"className":"text-xl sm:text-2xl font-bold text-dark-text mt-8 mb-4","children":"Looking Ahead"}],["$","p","9",{"className":"text-muted leading-relaxed my-4","children":"The success of this symposium has reinforced our commitment to responsible computing education. We're already planning follow-up workshops and a semester-long reading group on AI ethics."}]]}],"$L4"]}]}],null,"$L5"]}],"isPartial":false,"staleTime":300,"varyParams":null,"buildId":"UcxJ1GnvZqjXLrdSu2fiy"} 4:["$","div",null,{"className":"mt-16 pt-8 border-t border-border","children":["$","$L2",null,{"href":"/blog","className":"inline-flex items-center gap-1.5 text-sm font-semibold text-acm-blue hover:text-acm-blue-dark transition-colors","children":[["$","$L3",null,{"iconNode":"$0:rsc:props:children:0:props:children:props:children:0:props:children:0:props:iconNode","className":"lucide-arrow-left","size":14}],"Back to all posts"]}]}] 5:["$","$L6",null,{"children":["$","$7",null,{"name":"Next.MetadataOutlet","children":"$@8"}]}] 8:null