AI for Testers
Artificial Intelligence (AI) has taken the world by storm, increasing the productivity of workers in a wide range of industries, especially software. But, it’s also understandably led to uncertainty and fear about the personal and professional implications for disciplines such as software testing.
If you’re interested in cutting through the hype and understanding how AI affects the testing profession, then this course is for you. In this class, you will gain a solid understanding of AI and Machine Learning (ML), how to test systems with AI components, and how to apply AI to the process of testing itself.
Key takeaways from this class include:
- Understanding what AI is in its current state along with expected upcoming trends.
- Leveraging AI to support testing activities like planning, test analysis, test design, implementation, execution, and completion.
- Effectively testing a system that includes AI components
- Introducing AI testing tools
Who Should Attend
This course is ideal for those who test AI-based systems or use (or wish to use) AI to support their testing activities. This includes those in hands-on testing roles or test managers, as well as software developers and development managers. In addition, those who want a basic familiarity with these critical topics, such as those in project management, leadership, and consulting roles, will derive value from this course.
Laptop and RDP Required
This class involves hands-on activities using sample software to better facilitate learning. Each student should bring a laptop with a remote desktop protocol (RDP) client pre-installed. Connection specifics and credentials will be supplied during class. Please work with your IT Admin before class to verify that your RDP client can be used to access a virtual machine running in the Amazon Web Services (AWS) environment. If you or your Admin have questions about the specific applications involved, contact our Client Support team.
- Overfitting and Underfitting
- Overfitting
- Underfitting
- Hands-On Exercise: Demonstrate Overfitting and Underfitting
Session 4: Testing AI-Based Systems Overview
- Specification of AI-Based Systems
- Test Levels for AI-Based Systems
- Input Data Testing
- ML Model Testing
- Component Testing
- Component Integration Testing
- System Testing
- Acceptance Testing
- Test Data for Testing AI-based Systems
- Testing for Automation Bias in AI-Based Systems
- Documenting an AI Component
- Testing for Concept Drift
- Selecting a Test Approach for an ML System
- Hands-On Exercise: Selecting a Test Approach for an ML System
Session 5: Using AI for Testing
- AI Technologies for Testing
- Hands-On Exercise: The Use of AI in Testing
- Using AI to Analyze Reported Defects
- Using AI for Test Case Generation
- Using AI for Healing or to Create Self-Healing Test Automation
- Using AI for the Optimization of Regression Test Suites
- Using AI for Defect Prediction
- Hands-On Exercise: Build a Defect Prediction System
- Using AI for Testing User Interfaces
- Using AI to Test Through the Graphical User Interface (GUI)
- Using AI to Test the GUI
Sign-In/Registration 7:30 - 8:30 a.m.
Morning Session 8:30 a.m. - 12:00 p.m.
Lunch 12:00 - 1:00 p.m.
Afternoon Session 1:00 - 5:00 p.m.
Times represent the typical daily schedule. Please confirm your schedule at registration.
• Digital course materials
• Continental breakfasts and refreshment breaks
• Lunches
Jonathan Kauffman works as an agile software development and test consultant at Coveros, a company that helps organizations develop secure software using agile methods. In this role, Jonathan has helped both government and commercial organizations develop and test high-quality applications, and he has gained his experience by working with health care, biomedical device, and research organizations. Jonathan also presents at and attends Meetups to help maintain his connection with the software testing community and to stay abreast of recent industry developments. Before joining Coveros, he earned his B.S. in computer science from Allegheny College, where he published research on techniques for optimizing regression test suites.