The Quality Barometer A QA manager is often faced with measuring the impossible. Here is a simple, post-ship metric to help judge the test effort's effectiveness. The Quality Barometer method uses the bug counts found during testing, calculates a percentage, and then uses that percentage as the defect target number that can be tolerated after shipment. |
George Hamblen
June 26, 2002 |
|
Risk-Based Testing: How to Conduct Heuristic Risk Analysis Software testing is often motivated by risk. If you accept this premise, you might well wonder how the term "risk-based testing" is not merely redundant. However, conducting a heuristic risk analysis by employing a checklist of open-ended questions, suggestions, or guidewords is a proven approach to help you find the most important risks for developing your testing plans. |
||
Lessons in Test Automation Elfriede Dustin has worked on many projects at various companies where automated testing tools were introduced to a test program lifecycle for the first time. In reviewing these projects, she has accumulated a list of "Automated Testing Lessons Learned," taken from actual experiences and test engineer feedback. In this article, she will share examples of this feedback, hoping that this information will help you avoid some typical false starts and roadblocks. |
||
Testing Web-based Applications To be most effective in analyzing and reproducing errors in a Web environment, you need to have a command over the operating environment. You also need to understand how environment-specific variables may affect your ability to replicate errors. With the application of some of the skills covered in this article, your Web testing experience should be less frustrating and more enjoyable. |
Hung Nguyen
June 26, 2002 |
|
Designing Useful Metrics: Using Observation, Modeling, and Measurement to Make Decisions First-order measurement can help you understand what's going on, make decisions, and improve results. Observation, modeling, and simple data gathering are things that you can implement in your work group without a big measurement program or big funding. Start by modeling your system and working out on paper how different measures will affect your system. Then involve your team, expand your model, and try some simple data gathering. This approach to measurement is one more tool in your toolkit, and it will move your organization toward better quality. |
||
The Influential Test Manager: How to Develop and Use Influence to Help Your Test Group - and Project - Succeed Test managers often feel that while somebody might be in control of schedules or resources, they certainly are not. An experienced test manager shares ways to develop and use professional influence to help the test group. |
||
Retiring Lifecycle Dinosaurs: Adaptive Software Development Adaptive Software Development (ASD) is one of a growing number of alternatives to traditional, process-centric software management methods. Extreme Programming (XP), Lean Development, SCRUM, and Crystal Light methods—although different in many respects—are tied together by a focus on people, results, minimal methods, and maximum collaboration. They are geared to the high speed and high change of today's e-business projects. |
Jim Highsmith
June 26, 2002 |
|
Software Installation Testing: How to Automate Tests for Smooth System Installation Installation testing—especially manual testing—can sometimes be grueling. Here are several aspects of installation testing that are best suited to automated methods. |
Chris Agruss
June 26, 2002 |
|
Cem Kaner on Rethinking Software Metrics The theory underlying a measurement must take into account at least nine factors. This article defines these nine factors (e.g., the scope of the measurement, the scale of the instrument, and the variation of measurements made with the instrument) and applies them to a few examples. |
Cem Kaner
June 26, 2002 |
|
Avoiding Scalability Shock Web application scalability tops the list of challenges for those designing and developing e-commerce sites. Here are five steps to managing the performance of e-business applications: architecture validation, performance benchmarking, performance regression testing, performance tuning and acceptance, and continuous performance monitoring. |
Billie Shea
June 26, 2002 |
Pages
Upcoming Events
Apr 27 |
STAREAST Software Testing Conference in Orlando & Online |
Jun 08 |
AI Con USA An Intelligence-Driven Future |
Sep 21 |
STARWEST Software Testing Conference in Anaheim & Online |