Conference Presentations

STARWEST 2001: Bug Hunting: Going on a Software Safari

This presentation is about bugs: where they hide, how you find them, and how you tell other people they exist so they can be fixed. Explore the habitats of the most common types of software bugs. Learn how to make bugs more likely to appear and discover ways to present information about the bugs you find to ensure they get fixed. Drawing on real-world examples of bug reports, Elisabeth Hendrickson reveals tips and techniques for capturing the wiliest and most squirmy critters crawling around in your software.

Elisabeth Hendrickson, Quality Tree Software
Managing the Test Effort Using Requirements-Based Testing Metrics

It's difficult to quantify the true state of a test effort. Often, it's measured by quantity of work combined with deadline compliance. But if this is the case, then the true level of quality remains unknown. The Requirements-Based Testing (RBT) process offers a set of metrics that can be utilized throughout the development cycle. These metrics can provide an accurate picture of the test effort at any given time.

Gary Mogyorodi, Bit Inc.
Evolution of Automated Testing for Enterprise Systems

The key to accelerating test automation in any project is for a well-rounded, cohesive team to emerge that can marry its business knowledge with its technical expertise. This session is an in-depth case study of the evolution of automated testing at the BNSF Railroad. From record-and-playback to database-driven robust test scripts, this session will take you through each step of the $24 billion corporation's efforts to implement test automation.

Cherie Coles, BNSF Railroad
Introduction to Usability Testing

What is usability? Why is it important? If these questions wake you in the middle of the night, then this presentation is for you. Cheryl Nesta discusses the relevance of usability testing within the broad framework of quality assurance and appropriate expectations based on its uses and applicability. Explore methodology, process flow, goal identification, and definition. Real-world examples create a hands-on introductory experience.

Cheryl L. Nesta, Vanteon
How to Find the Level of Quality Your Sponsor Wants

In this paper, I'll talk about how to focus yourself to do effective testing. I'll cover where to start, what to look for, what and who are your resources. There is a level of understanding you need about the quality goals for a project. I'll give some examples of ways to figure that out without even asking questions. Or, who to ask if you could and then what to do with the information to be
effective in your testing efforts. Some of this is similar to risk based testing, which you can use in a complementary fashion. The way it is different is in its focus on the sponsor’s expectations.

Sue Bartlett, Step Technology
Concise, Standardized, Organized Testing in Complex Test Environments

There's a need for standardized, organized hardware and software infrastructure, and for a common framework, in a complex test environment. Gerhard Strobel focuses on the experience of testing diverse products on many different platforms (UNIX, Windows, OS2, z/OS, OS400)-how they differ and how much they have in common. He explains how to configure and profile test machines, then highlights the technical areas where test efficiency can be increased. He also covers methods of execution control.

Gerhard Strobel, IBM Germany
Revealing the Mysteries of Test Measurement

Contrary to popular belief, test measurement is not a mysterious art. Rather, it's a vital part of test management. This presentation will assist those new to test measurement-and those that have been intimidated by measurement and its complexities in the past-discover that test measurement can and should be a part of your test plan. Learn metrics and techniques that are easy to understand as well as implement.

David Hutcheson, Glen Abbot Ltd.
Test Result Checking Patterns

Determining how a test case detects a product failure requires several test case design trade-offs. These trade-offs include the characteristics of test data used and when comparisons are done. This document addresses how result checking impacts test design.

Keith Stobie, Microsoft
Data in Functional Testing-You Can't Live Without It

This paper sets out to illustrate some of the ways that data can influence the test process, and will show that testing can be improved by a careful choice of input data. In doing this, the paper will concentrate most on data-heavy applications; those which use databases or are heavily influenced by the data they hold. The paper will focus on input data, rather than output data or the transitional states the data passes through during processing, as input data has the greatest influence on functional testing and is the simplest to manipulate. The paper will not consider areas where data is important to non-functional testing, such as operational profiles, massive datasets and environmental tuning.

James Lyndsay, Workroom Productions
A Framework for Testing Real-Time and Embedded Systems

What do we mean when we say local, remote, simultaneous, and distributed testing? Alan Haffenden of The Open Group explores the differences, and explains why the architecture of a distributed test execution system must be different from that of non-distributed systems. An overview of POSIX 1003.13 profiles and units of functionality helps advanced users build a good foundation for testing both their real-time and embedded systems.

Alan Haffenden, The Open Group

Pages

AgileConnection is a TechWell community.

Through conferences, training, consulting, and online resources, TechWell helps you develop and deliver great software every day.