Submissions for MLOps November edition
On ML workflows, tools, automation and running ML in production
This project is accepting submissions for MLOps November conference edition.
The first edition of the MLOps conference was held on 23, 24 and 27 July. Details about the conference including videos and blog posts are published at https://hasgeek.com/fifthelephant/mlops-conference/
Contact information: For inquiries, contact The Fifth Elephant on fifthelephant.editorial@hasgeek.com or call 7676332020.
Hosted by
Ramjee Ganti
Topic:
As more companies adopt AI, more people question the impact AI creates on society, especially on algorithmic fairness. However, most metrics that measure the fairness of AI algorithms today don’t capture the critical nuance of intersectionality. Instead, they hold a binary view of fairness, e.g., protected vs. unprotected groups. In this talk, we’ll discuss the latest research on intersectional group fairness using worst-case comparisons.
3 Key Takeaways:
The importance of fairness in AI
Why AI fairness is even more critical today
Why intersectional group fairness is critical to improving AI fairness
Slides: https://docs.google.com/presentation/d/1sqxhzOw3cItJc7lxrpJwgfBZqIgnHB4VSVVB9CAkb8U/edit?usp=sharing
Speaker:
I am Ramjee, work at Fiddler Labs Inc(http://fiddler.ai/). We believe that every AI solution should be responsible and ethical. With the mission to help every organization build trust into AI, we’ve developed a model performance monitoring platform, powered by explainable AI. Our platform empowers data science and engineering teams to track and troubleshoot ML issues, such as drift and integrity, faster with our unique XAI technology. The XAI not only pinpoints causes of the performance drops, it enables our users to quickly and easily validate and compare different models, and detect bias.
Hosted by
{{ gettext('Login to leave a comment') }}
{{ gettext('Post a comment…') }}{{ errorMsg }}
{{ gettext('No comments posted yet') }}