Documente Academic
Documente Profesional
Documente Cultură
Q: What is validation?
Q: What is a walk-through?
A: A walk-through is an informal meeting for evaluation or informational
purposes.
Q: What is an inspection?
Q: What is quality?
A: A good code is code that works, is free of bugs and is readable and
maintainable. Organizations usually have coding standards all developers should
adhere to, but every programmer and software engineer has different ideas about
what is best and what are too many or too few rules. We need to keep in mind
that excessive use of rules can stifle both productivity and creativity. Peer reviews
and code analysis tools can be used to check for problems and enforce
standards.
A: Design could mean to many things, but often refers to functional design or
internal design. Good functional design is indicated by software functionality can
be traced back to customer and end-user requirements. Good internal design is
indicated by software code whose overall structure is clear, understandable,
easily modifiable and maintainable; is robust with sufficient error handling and
status logging capability; and works correctly when implemented.
A: Software life cycle begins when a software product is first conceived and ends
when it is no longer in use. It includes phases like initial concept, requirements
analysis, functional design, internal design, documentation planning, test
planning, coding, document preparation, integration, testing, maintenance,
updates, re-testing and phase-out.
A: It depends on the size of the organization and the risks involved. For large
organizations with high-risk projects, a serious management buy-in is required
and a formalized QA process is necessary. For medium size organizations with
lower risk projects, management and organizational buy-in and a slower, step-by-
step process is required. Generally speaking, QA processes should be balanced
with productivity, in order to keep any bureaucracy from getting out of hand. For
smaller groups or projects, an ad-hoc process is more appropriate. A lot depends
on team leads and managers, feedback to developers and good communication
is essential among customers, managers, developers, test engineers and testers.
Regardless the size of the company, the greatest value for effort is in managing
requirement processes, where the goal is requirements that are clear, complete
and
testable.
A: Yes and no. For larger projects, or ongoing long-term projects, they can be
valuable. But for small projects, the time needed to learn and implement them is
usually not worthwhile. A common type of automated tool is the record/playback
type. For example, a test engineer clicks through all combinations of menu
choices, dialog box choices, buttons, etc. in a GUI and has an automated testing
tool record and log the results. The recording is typically in the form of text, based
on a scripting language that the testing tool can interpret. If a change is made
(e.g. new buttons are added, or some underlying code in the application is
changed), the application is then re-tested by just playing back the recorded
actions and compared to the logged results in order to check effects of the
change. One problem with such tools is that if there are continual changes to the
product being tested, the recordings have to be changed so often that it becomes
a very time-consuming task to continuously update the scripts. Another problem
with such tools is the interpretation of the results (screens, data, logs, etc.) that
can be a time-consuming task.
Good test engineers have a "test to break" attitude, they take the point of view of the
customer, have a strong desire for quality and an attention to detail. Tact and diplomacy
are useful in maintaining a cooperative relationship with developers and an ability to
communicate with both technical and non-technical people. Previous software
development experience is also helpful as it provides a deeper understanding of the
software development process, gives the test engineer an appreciation for the
developers' point of view and reduces the learning curve in automated test tool
programming.
A: The same qualities a good test engineer has are useful for a QA engineer. Additionally,
Rob Davis understands the entire software development process and how it fits into the
business approach and the goals of the organization. Rob Davis' communication skills
and the ability to understand various sides of issues are important.
Good QA engineers understand the entire software development process and how it fits
into the business approach and the goals of the organization. Communication skills and
the ability to understand various sides of issues are important.
A: QA/Test Managers are familiar with the software development process; able to
maintain enthusiasm of their team and promote a positive atmosphere; able to
promote teamwork to increase productivity; able to promote cooperation between
Software and Test/QA Engineers, have the people skills needed to promote
improvements in QA processes, have the ability to withstand pressures and say
*no* to other managers when quality is insufficient or QA processes are not being
adhered to; able to communicate with technical and non-technical people; as well
as able to run meetings and keep them focused.
A: A test case is a document that describes an input, action, or event and its
expected result, in order to determine if a feature of an application is working
correctly. A test case should contain particulars such as a...
Please note, the process of developing test cases can help find problems in the
requirements or design of an application, since it requires you to completely think
through the operation of the application. For this reason, it is useful to prepare
test cases early in the development cycle, if possible.
A: In this situation the best bet is to have test engineers go through the process
of reporting whatever bugs or problems initially show up, with the focus being on
critical bugs. Since this type of problem can severely affect schedules and
indicates deeper problems in the software development process, such as
insufficient unit testing, insufficient integration testing, poor design, improper build
or release procedures, managers should be notified and provided with some
documentation as evidence of the problem.
A: Since it's rarely possible to test every possible aspect of an application, every
possible combination of events, every dependency, or everything that could go
wrong, risk analysis is appropriate to most software development projects. Use
risk analysis to determine where testing should be focused. This requires
judgment skills, common sense and experience. The checklist should include
answers to the following questions:
A: Consider the impact of project errors, not the size of the project. However, if
extensive testing is still not justified, risk analysis is again needed and the
considerations listed under "What if there isn't enough time for thorough testing?"
do apply. The test engineer then should do "ad hoc" testing, or write up a limited
test plan based on the risk analysis.
• Ensure the code is well commented and well documented; this makes
changes easier for the developers.
• Use rapid prototyping whenever possible; this will help customers feel
sure of their requirements and minimize changes.
• In the project's initial schedule, allow for some extra time to
commensurate with probable changes.
• Move new requirements to a 'Phase 2' version of an application and use
the original requirements for the 'Phase 1' version.
• Negotiate to allow only easily implemented new requirements into the
project; move more difficult, new requirements into future versions of the
application.
• Ensure customers and management understand scheduling impacts,
inherent risks and costs of significant requirements changes. Then let
management or the customers decide if the changes are warranted; after
all, that's their job.
• Balance the effort put into setting up automated testing with the expected
effort required to redo them to deal with changes.
• Design some flexibility into automated test scripts;
• Focus initial automated testing on application aspects that are most likely
to remain unchanged;
• Devote appropriate effort to risk analysis of changes, in order to minimize
regression-testing needs;
• Design some flexibility into test cases; this is not easily done; the best
bet is to minimize the detail in the test cases, or set up only higher-level
generic-type test plans;
• Focus less on detailed test plans and test cases and more on ad-hoc
testing with an understanding of the added risk this entails.
A: Because testing during the design phase can prevent defects later on. I
recommend we verify three things...
A: Quality Assurance ensures all parties concerned with the project adhere to the
process and procedures, standards and templates and test readiness reviews.
Rob Davis' QA service depends on the customers and projects. A lot will depend on team
leads or managers, feedback to developers and communications among customers,
managers, developers' test engineers and testers.
A: Detailed and well-written processes and procedures ensure the correct steps are
being executed to facilitate a successful completion of a task. They also ensure a process
is repeatable. Once Rob Davis has learned and reviewed customer's business processes
and procedures, he will follow them. He will also recommend improvements and/or
additions.
A: All documents should be written to a certain standard and template. Standards and
templates maintain document uniformity. It also helps in learning where information is
located, making it easier for a user to find what they want. Lastly, with standards and
templates, information will not be accidentally omitted from a document. Once Rob Davis
has learned and reviewed your standards and templates, he will use them. He will also
recommend improvements and/or additions.
A: Black box testing is functional testing, not based on any knowledge of internal
software design or code. Black box testing is based on requirements and functionality.
A: White box testing is based on knowledge of the internal logic of an application's code.
Tests are based on coverage of code statements, branches, paths and conditions.
A: Unit testing is the first level of dynamic testing and is first the responsibility of
developers and then that of the test engineers. Unit testing is performed after the
expected test results are met or differences are explainable/acceptable.
A: Parallel/audit testing is testing where the user reconciles the output of the new system
to the output of the current system to verify the new system performs the operations
correctly.
A: Usability testing is testing for 'user-friendliness'. Clearly this is subjective and depends
on the targeted end-user or customer. User interviews, surveys, video recording of user
sessions and other techniques can be used. Test engineers are needed, because
programmers and developers are usually not appropriate as usability testers.
A: Upon completion of unit testing, integration testing begins. Integration testing is black
box testing. The purpose of integration testing is to ensure distinct components of the
application still work in accordance to customer requirements. Test cases are developed
with the express purpose of exercising the interfaces between the components. This
activity is carried out by the test team. Integration testing is considered complete, when
actual results and expected results are either in line or differences are
explainable/acceptable based on client input.
A: System testing is black box testing, performed by the Test Team, and at the start of the
system testing the complete system is configured in a controlled environment. The
purpose of system testing is to validate an application's accuracy and completeness in
performing the functions as designed. System testing simulates real life scenarios that
occur in a "simulated real life" test environment and test all functions of the system that
are required in real life. System testing is deemed complete when actual results and
expected results are either in line or differences are explainable or acceptable, based on
client input.
Upon completion of integration testing, system testing is started. Before system testing,
all unit and integration test results are reviewed by SWQA to ensure all problems have
been resolved. For a higher level of testing it is important to understand unresolved
problems that originate at unit and integration test levels.
A: End-to-end testing is similar to system testing, the *macro* end of the test
scale; it is the testing a complete application in a situation that mimics real life
use, such as interacting with a database, using network communication, or
interacting with other hardware, application, or system.
A: Load testing is testing an application under heavy loads, such as the testing of
a web site under a range of loads to determine at what point the system
response time will degrade or fail.
A: Depending on the organization, the following roles are more or less standard
on most testing projects: Testers, Test Engineers, Test/QA Team Lead, Test/QA
Manager, System Administrator, Database Administrator, Technical Analyst, Test
Build Manager and Test Configuration Manager. Depending on the project, one
person may wear more than one hat. For instance, Test Engineers may also wear
the hat of Technical Analyst, Test Build Manager and Test Configuration Manager.
A: Test Build Managers deliver current software versions to the test environment,
install the application's software and apply software patches, to both the
application and the operating system, set-up, maintain and back up test
environment hardware. Depending on the project, one person may wear more
than one hat. For instance, a Test Engineer may also wear the hat of a Test Build
Manager.
A: The test schedule is a schedule that identifies all tasks required for a
successful testing effort, a schedule of all test activities and resource
requirements.
This methodology can be used and molded to your organization's needs. Rob
Davis believes that using this methodology is important in the development and
ongoing maintenance of his customers' applications.
• An approved and signed off test strategy document, test plan, including
test cases.
• Testing issues requiring resolution. Usually this requires additional
negotiation at the project management level.
• Test cases and scenarios are designed to represent both typical and
unusual situations that may occur in the application.
• Test engineers define unit test requirements and unit test cases. Test
engineers also execute unit test cases.
• It is the test team who, with assistance of developers and clients,
develops test cases and scenarios for integration and system testing.
• Test scenarios are executed through the use of test procedures or
scripts.
• Test procedures or scripts define a series of steps necessary to perform
one or more test scenarios.
• Test procedures or scripts include the specific data that will be used for
testing the process or transaction.
• Test procedures or scripts may cover multiple test scenarios.
• Test scripts are mapped back to the requirements and traceability
matrices are used to ensure each test is within scope.
• Test data is captured and base lined, prior to testing. This data serves as
the foundation for unit and system testing and used to exercise system
functionality in a controlled environment.
• Some output data is also base-lined for future comparison. Base-lined
data is used to support future application maintenance via regression
testing.
• A pre-test meeting is held to assess the readiness of the application and
the environment and data to be tested. A test readiness document is
created to indicate the status of the entrance criteria of the release.
• Approved test documents, e.g. Test Plan, Test Cases, Test Procedures.
• Test tools, including automated test tools, if applicable.
• Developed scripts.
• Changes to the design, i.e. Change Request Documents.
• Test data.
• Availability of the test team and project team.
• General and Detailed Design Documents, i.e. Requirements Document,
Software Design Document.
• A software that has been migrated to the test environment, i.e. unit tested
code, via the Configuration/Build Manager.
• Test Readiness Document.
• Document Updates.
• Log and summary of the test results. Usually this is part of the Test
Report. This needs to be approved and signed-off with revised testing
deliverables.
• Changes to the code, also known as test fixes.
• Test document problems uncovered as a result of testing. Examples are
Requirements document and Design Document problems.
• Reports on software design issues, given to software developers for
correction. Examples are bug reports on code issues.
• Formal record of test incidents, usually part of problem tracking.
• Base-lined package, also known as tested source and object code, ready
for migration to the next level.
TABLE OF CONTENTS
1 INTRODUCTION 23
2 TESTING METHODOLOGIES 24
3 TEMPLATES 28
1.1 Abstract
The testing approach for a website will be entirely different when compared to the testing
approach for any other applications. There are various types of testing and they are
confined with different solutions for different platforms. This document aims at having a
Unique Strategy for testing a website and aims at describing the various Testing
Methodologies and Guidelines that are going to be used for Testing Solutions.
2 Testing Methodologies
GUI Verification
Compatibility Testing
Based upon this, the Test Cases have to be developed. Checks have to be done
to ensure that the page that is displayed to the user is authorized to be viewed by
him.
Both are considered as positive tests. Benchmarking needs the result of Stress
and Load Testing.
3.1.2 Verification
This sheet carries the following attributes.
• UI Expected Title
• UI Actual Title
• UI Control
• Present (Yes/ No)
• Expected Style
• Actual Style
• Expected Label
• Actual Label
The Page Expected Title identifies the page. The UI Control Name is the content
of the UI Control Attribute. The Expected Style is captured from the Style Sheet.
The Actual Title, Actual Style and the Actual Label are captured by WYS.
3.1.3 Global Validation
This sheet carries the following attributes.
• UI
• Control Name
• Control Value
• Operation
• Expected Result
• Actual Result
• Test Result
• Severity
• Status
• Remarks/Suggestion
The set of Control Names and Control Values for an operation is listed down. The
Expected and Actual Results for that operation is written down. If Test Result
fails, the Severity and Status have to be tracked.
3.1.4 Scenario
This sheet carries the following attributes.
• Scenario
• Expected Result
• Actual Result
• Test Result
• Severity
• Status
• Remarks/Suggestion
The Scenarios of operations are captured and its Expected and Actual Results
are written down. If the Test Result fails, the Severity and Status have to be
tracked.
3.1.5 Negative
This sheet carries the following attributes.
• User Action
• Expected Result
• Actual Result
• Test Result
• Severity
• Status
• Remarks/Suggestion
The User Action should be very descriptive of about which page the user uses to
do the explained action. Expected and Actual Results are written down. If the
Test Result fails, the Severity and Status have to be tracked.
• UI
• Item Name
• Input Values 1 (or) Input User Action 1
• Input Values 2 (or) Input User Action 2
• :
• :
• Input Values N (or) Input User Action N
If the Input Value is a desired one for that control, letter ‘D’ should be written
down in the cell that appears at the intersection of the Input Values (N) Attribute
and the Item Name. This serves as the Test Case. If this Test Case get Pass, the
background of the cell will be turned to green else red.