In this SEI Podcast, Dr. Eric Heim, a senior machine learning research scientist at Carnegie Mellon University's Software Engineering Institute (SEI), discusses the quantification of uncertainty in machine-learning (ML) systems. ML systems can make wrong predictions and give inaccurate estimates for the uncertainty of their predictions. It can be difficult to predict when their predictions will be wrong. Heim also discusses new techniques to quantify uncertainty, identify causes of uncertainty, and efficiently update ML models to reduce uncertainty in their predictions. The work of Heim and colleagues at the SEI Emerging Technology Center closes the gap between the scientific and mathematical advances from the ML research community and the practitioners who use the systems in real-life contexts, such as software engineers, software developers, data scientists, and system developers.
A Discussion on Automation with Watts Humphrey Award Winner Rajendra Prasad
Enabling Transition From Sustainment to Engineering Within the DoD
The Silver Thread of Cyber in the Global Supply Chain
Measuring DevSecOps: The Way Forward
Bias in AI: Impact, Challenges, and Opportunities
My Story in Computing with Rachel Dzombak
Agile Strategic Planning: Concepts and Methods for Success
Applying Scientific Methods in Cybersecurity
Zero Trust Adoption: Benefits, Applications, and Resources
11 Rules for Ensuring a Security Model with AADL and Bell–LaPadula
Benefits and Challenges of Model-Based Systems Engineering
Fostering Diversity in Software Engineering
Can DevSecOps Make Developers Happier?
Is Your Organization Ready for AI?
My Story in Computing with Marisa Midler
Managing Vulnerabilities in Machine Learning and Artificial Intelligence Systems
AI Workforce Development
Moving from DevOps to DevSecOps
My Story in Computing with David Zubrow
Create your
podcast in
minutes
It is Free