{"id":4899,"date":"2024-04-30T15:17:20","date_gmt":"2024-04-30T15:17:20","guid":{"rendered":"https:\/\/xaiworldconference.com\/2024\/?post_type=mp-event&#038;p=4899"},"modified":"2024-05-09T10:46:01","modified_gmt":"2024-05-09T10:46:01","slug":"s-17-a-1-2-2-2","status":"publish","type":"mp-event","link":"https:\/\/xaiworldconference.com\/2024\/timetable\/event\/s-17-a-1-2-2-2\/","title":{"rendered":"Fairness, trustworthiness and trust calibration in XAI"},"content":{"rendered":"\n<figure class=\"wp-block-table is-style-stripes has-small-font-size\"><table class=\"has-fixed-layout\"><thead><tr><th>Time<\/th><th>Authors<\/th><th>Title<\/th><\/tr><\/thead><tbody><tr><td>15:45-16:00<\/td><td>Raphael C. Engelhardt, Moritz Lange, Laurenz Wiskott and Wolfgang Konen<\/td><td>Exploring the Reliability of SHAP Values in Reinforcement Learning<\/td><\/tr><tr><td>16:00-16:15<\/td><td>Francesco Giannini, Stefano Fioravanti, Pietro Barbiero, Alberto Tonda, Pietro Lio and Elena Di Lavore<\/td><td>Categorical Foundation of Explainable AI: A Unifying Theory<\/td><\/tr><tr><td>16:15-16:30<\/td><td>Alireza Torabian and Ruth Urner<\/td><td>Investigating Calibrated Classification Scores through the Lens of Interpretability<\/td><\/tr><tr><td>16:30-16:45<\/td><td>Sarah Seifi, Tobias Sukianto, Maximilian Strobel, Cecilia Carbonelli, Lorenzo Servadei and Robert Wille<\/td><td>XentricAI: A Gesture Sensing Calibration Approach through Explainable and User-Centric AI<\/td><\/tr><tr><td>16:45-17:00<\/td><td>Niklas Koenen and Marvin N. Wright<\/td><td>Toward Understanding the Disagreement Problem in Neural Network Feature Attribution<\/td><\/tr><\/tbody><\/table><\/figure>\n","protected":false},"excerpt":{"rendered":"<p>Time Authors Title 15:45-16:00 Raphael C. Engelhardt, Moritz Lange, Laurenz Wiskott and Wolfgang Konen Exploring the Reliability of SHAP Values in Reinforcement Learning 16:00-16:15 Francesco Giannini, Stefano Fioravanti, Pietro Barbiero, Alberto Tonda, Pietro Lio and Elena Di Lavore Categorical Foundation of Explainable AI: A Unifying Theory 16:15-16:30 Alireza Torabian and Ruth Urner Investigating Calibrated Classification &hellip; <\/p>\n","protected":false},"author":2,"featured_media":0,"menu_order":0,"comment_status":"open","ping_status":"closed","template":"","mp-event_category":[],"mp-event_tag":[],"class_list":["post-4899","mp-event","type-mp-event","status-publish","hentry","mp-event-item"],"_links":{"self":[{"href":"https:\/\/xaiworldconference.com\/2024\/wp-json\/wp\/v2\/mp-event\/4899","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/xaiworldconference.com\/2024\/wp-json\/wp\/v2\/mp-event"}],"about":[{"href":"https:\/\/xaiworldconference.com\/2024\/wp-json\/wp\/v2\/types\/mp-event"}],"author":[{"embeddable":true,"href":"https:\/\/xaiworldconference.com\/2024\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/xaiworldconference.com\/2024\/wp-json\/wp\/v2\/comments?post=4899"}],"wp:attachment":[{"href":"https:\/\/xaiworldconference.com\/2024\/wp-json\/wp\/v2\/media?parent=4899"}],"wp:term":[{"taxonomy":"mp-event_category","embeddable":true,"href":"https:\/\/xaiworldconference.com\/2024\/wp-json\/wp\/v2\/mp-event_category?post=4899"},{"taxonomy":"mp-event_tag","embeddable":true,"href":"https:\/\/xaiworldconference.com\/2024\/wp-json\/wp\/v2\/mp-event_tag?post=4899"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}