SE Unit4
SE Unit4
UNIT 4
Contents:
Coding and Testing: Coding, Code Review, Software Documentation, Testing, Unit Testing,
Black-Box Testing, White-Box Testing, Debugging, Integration Testing, System Testing.
Introduction:
● Coding is undertaken once the design phase is complete and the design documents
have been successfully reviewed.
● In the coding phase, every module specified in the design document is coded and unit
tested. During unit testing, each module is tested in isolation from other modules.
● After all the modules of a system have been coded and unit tested, the integration and
system testing phase is undertaken
● Integration and testing of modules is carried out according to an integration plan.
● The full product takes shape only after all the modules have been integrated together.
System testing is conducted on the full product. During system testing, the product is
tested against its requirements as recorded in the SRS document.
● Testing is an important phase in software development, requires the maximum effort
and requires the maximum effort.
Coding:
● The input to the coding phase is the design document produced at the end of the design
phase.
● The design document contains not only the high-level design of the system in the form
of a module structure (e.g., a structure chart), but also the detailed design.
● The detailed design is usually documented in the form of module specifications where
the data structures and algorithms for each module are specified.
● The objective of the coding phase is to transform the design of a system into code in a
high-level language, and then to unit test this code.
● good software development organisations require their programmers to adhere to some
well-defined and standard style of coding which is called their coding standard.
● organisations formulate their own coding standards and require their developers to
follow the standards rigorously.
● The main advantages of adhering to a standard:
○ A coding standard gives a uniform appearance to the codes written by different
engineers.
○ It facilitates code understanding and code reuse.
○ It promotes good programming practices.
● It is mandatory for the programmers to follow the coding standards. Compliance of their
code to coding standards is verified during code inspection. Any code that does not
Unit 4 : Software Engineering [20IT3302] PVP20
conform to the coding standards is rejected during code review and the code is
reworked by the concerned programmer.
● In contrast, coding guidelines provide some general suggestions regarding the coding
style to be followed but leave the actual implementation of these guidelines to the
discretion of the individual developers.
Usually code review is carried out to ensure that the coding standards are followed and
also to detect as many errors as possible before testing. Reviews are an efficient way of
removing errors from code.
Good software development organisations usually develop their own coding standards and
guidelines.
● Rules for limiting the use of globals: These rules list what types of data can be declared
global and what cannot, with a view to limit the data that needs to be defined with global
scope.
● Standard headers for different modules: The header of different modules should have
standard format and information for ease of understanding and maintenance.
● Naming conventions for global variables, local variables, and constant identifiers: A
popular naming convention is that variables are named using mixed case lettering. Example
:GlobalData, localData, CONSTDATA
● Conventions regarding error return values and exception handling mechanisms: The way
error conditions are reported by different functions in a program should be standard within
an organisation.
● Do not use a coding style that is too clever or too difficult to understand: Code should be
easy to understand. Many inexperienced engineers actually take pride in writing cryptic and
incomprehensible code.
● Avoid obscure side effects: The side effects of a function call include modifications to the
parameters passed by reference, modification of global variables, and I/O operations. An
obscure side effect is one that is not obvious from a casual examination of the code.
Obscure side effects make it difficult to understand a piece of code.
● Do not use an identifier for multiple purposes: Programmers often use the same identifier
to denote several temporary entities. There are several things wrong with this approach and
hence should be avoided.
● Code should be well-documented: As a rule of thumb, there should be at least one
comment line on the average for every three source lines of code.
Unit 4 : Software Engineering [20IT3302] PVP20
● Length of any function should not exceed 10 source lines: A lengthy function is usually
very difficult to understand as it probably has a large number of variables and carries out
many different types of computations.
● Do not use GOTO statements: Use of GOTO statements makes a program unstructured.
This makes the program very difficult to understand, debug, and maintain.
Code Review:
Code inspection.
● During code inspection, the code is examined for the presence of some common
programming errors.
● The principal aim of code inspection is to check for the presence of some common types
of errors that usually creep into code due to programmer mistakes and oversights and to
check whether coding standards have been adhered to.
● The inspection process has several beneficial side effects, other than finding errors. The
programmer usually receives feedback on programming style, choice of algorithm, and
programming techniques. The other participants gain by being exposed to another
programmer’s errors.
Unit 4 : Software Engineering [20IT3302] PVP20
Code walkthrough.
Self Study:
Clean Room Testing
Unit 4 : Software Engineering [20IT3302] PVP20
Software Documentation:
When a software is developed, in addition to the executable files and the source code, several
kinds of documents such as users’ manual, software requirements specification (SRS)
document, design document, test document, installation manual, etc., are developed as part of
the software engineering process.
All these documents are considered a vital part of any good software development practice.
Good documents are helpful in the following ways:
● Good documents help enhance understandability of code.
● Documents help the users to understand and effectively use the system.
● Good documents help to effectively tackle the manpower turnover problem
● Production of good documents helps the manager to effectively track the progress of
the project
Different types of software documents can broadly be classified into the following:
Internal documentation:
● These are provided in the source code itself. Internal documentation can be provided in the
code in several forms. The important types of internal documentation are the following:
○ Comments embedded in the source code.
○ Use of meaningful variable names.
○ Module and function headers.
○ Code indentation.
○ Code structuring (i.e., code decomposed into modules and functions).
○ Use of enumerated types.
○ Use of constant identifiers.
○ Use of user-defined data types.
● Even when a piece of code is carefully commented, meaningful variable names have been
found to be the most helpful in understanding the code.
External documentation:
● These are the supporting documents such as SRS document, installation document, user
manual, design document, and test document.
● A systematic software development style ensures that all these documents are of good
quality and are produced in an orderly fashion.
● An important feature that is required of any good external documentation is consistency
with the code.
● If the different documents are not consistent, a lot of confusion is created for somebody
trying to understand the software.
● Every change made to the code should be reflected in the relevant external documents.
● Another important feature required for external documents is proper understandability by
the category of users for whom the document is designed.
● Gunning’s Fog Index:
○ Gunning’s fog index (developed by Robert Gunning in 1952) is a metric that has been
designed to measure the readability of a document.
Unit 4 : Software Engineering [20IT3302] PVP20
○ The computed metric value (fog index) of a document indicates the number of years of
formal education that a person should have, in order to be able to comfortably
understand that document.
○ The Gunning’s fog index of a document D can be computed as follows:
Testing
● The aim of program testing is to help realise/identify all defects in a program.
● However, in practice, even after satisfactory completion of the testing phase, it is not
possible to guarantee that a program is error free.
● This is because the input data domain of most programs is very large, and it is not practical
to test the program exhaustively with respect to each value that the input can assume.
● We must remember that careful testing can expose a large percentage of the defects
existing in a program
Testing terminology:
● Mistake: A mistake is essentially any programmer action that later shows up as an
incorrect result during program execution. A programmer may commit a mistake in almost
any development activity.
● Error: An error is the result of a mistake committed by a developer in any of the
development activities. Among the extremely large variety of errors that can exist in a
program. The terms error, fault, bug, and defect are considered to be synonyms.
● Failure: A failure of a program essentially denotes an incorrect behaviour exhibited by the
program during its execution. An incorrect behaviour is observed either as an incorrect
result produced or as an inappropriate activity carried out by the program.
● Test-case: A test case is a triplet [I , S, R], where I is the data input to the program under
test, S is the state of the program at which the data is to be input, and R is the result
expected to be produced by the program. The state of a program is also called its execution
mode.
○ A positive test case is designed to test whether the software correctly performs a
required functionality
○ A negative test case is designed to test whether the software carries out something
that is not required of the system.
● Test scenario: A test scenario is an abstract test case in the sense that it only identifies the
aspects of the program that are to be tested without identifying the input, state, or output. A
test case can be said to be an implementation of a test scenario.
● Test script: A test script is an encoding of a test case as a short program. Test scripts are
developed for automated execution of the test cases.
● Test suite: A test suite is the set of all tests that have been designed by a tester to test a
given program.
Unit 4 : Software Engineering [20IT3302] PVP20
● Error correction: After the error is located during debugging, the code is appropriately
changed to correct the error
Testing Process
Self Study:
Why design test cases?
Testing in small vs testing in Large?
Unit Testing
● Unit testing is undertaken after a module has been coded and reviewed.
● This activity is typically undertaken by the coder of the module himself in the coding phase.
● Before carrying out unit testing, the unit test cases have to be designed and the test
environment for the unit under test has to be developed.
● In order to test a single module, we need a complete environment to provide all relevant
code that is necessary for execution of the module.
● That is, besides the module under test, the following are needed to test the module:
○ The procedures belonging to other modules that the module under test calls.
○ Non-local data structures that the module accesses.
○ A procedure to call the functions of the module under test with appropriate
parameters.
● Modules required to provide the necessary environment (which either call or are called by
the module under test) are usually not available until they too have been unit tested.
● In this context, stubs and drivers are designed to provide the complete environment for a
module so that testing can be carried out.
Stub: A stub procedure is a dummy procedure that has the same I/O parameters as the function
called by the unit under test but has a highly simplified behaviour.
Unit 4 : Software Engineering [20IT3302] PVP20
Driver: A driver module should contain the non-local data structures accessed by the module
under test. Additionally, it should also have the code to call the different functions of the unit
under test with appropriate parameter values for testing.
● Unit testing is referred to as testing in the small, whereas integration and system testing are
referred to as testing in the large.
Black-Box testing:
● In black-box testing, test cases are designed from an examination of the input/output
values only and no knowledge of design or code is required.
● The following are the two main approaches available to design black box test cases:
○ Equivalence class partitioning
○ Boundary value analysis
classes are {A,B,C}, then the invalid equivalence class is U-{A,B,C} where U is the
universe of possible input values.
White-Box Testing:
● White-box testing is an important type of unit testing. A large number of white-box testing
strategies exist.
● Each testing strategy essentially designs test cases based on analysis of some aspect of
source code and is based on some heuristic.
Basic Concepts:
A white-box testing strategy can either be coverage-based or fault based.
Fault-based testing: A fault-based testing strategy targets to detect certain types of faults.
These faults that a test strategy focuses on constitutes the fault model of the strategy..
Coverage-based testing: A coverage-based testing strategy attempts to execute (or cover)
certain elements of a program.
Self study
Testing criterion
Stronger versus weaker testing & Complementary testing strategies.
Unit 4 : Software Engineering [20IT3302] PVP20
○ A CFG is a directed graph consisting of a set of nodes and edges (N, E), such that
each node n ∈ N corresponds to a unique program statement and an edge exists
between two nodes if control can transfer from one node to the other.
○ We can easily draw the CFG for any program, if we know how to represent the
sequence, selection, and iteration types of statements in the CFG.
○ If a set of paths is linearly independent of each other, then no path in the set can be
obtained through any linear operations (i.e., additions or subtractions) on the other
paths in the set.
McCabe’s Cyclomatic Complexity Metric:
● It is straightforward to identify the linearly independent paths for simple programs, for
more complex programs it is not easy to determine the number of independent paths.
● McCabe’s cyclomatic complexity metric is an important result that lets us compute
the number of linearly independent paths for any arbitrary program.
● McCabe’s cyclomatic complexity defines an upper bound for the number of linearly
independent paths through a program.
1. Method 1:
Given a control flow graph G of a program, the cyclomatic complexity V(G) can be
computed as:
V(G) = E – N + 2
where, N is the number of nodes of the control flow graph and E is the number of
edges in the control flow graph.
For the CFG of example shown in above figure (GCD function),, E = 7 and N = 6.
Therefore, the value of the Cyclomatic complexity = 7 – 6 + 2 = 3.
2. Method 2:
An alternate way of computing the cyclomatic complexity of a program is based
on a visual inspection of the control flow graph.
In this method, the cyclomatic complexity V (G) for a graph G is given by the
following expression:
V(G) = Total number of non-overlapping bounded areas + 1
Consider the CFG example shown in above figure (GCD function). From a visual
examination of the CFG the number of bounded areas is 2. Therefore the
cyclomatic complexity, computed with this method is also 2+1=3.
3. Method 3:
The cyclomatic complexity of a program can also be easily computed by
computing the number of decision and loop statements of the program. If N is
the number of decision and loop statements of a program, then the McCabe’s
metric is equal to N + 1.
Steps to carry out path coverage-based testing:
The following is the sequence of steps that need to be undertaken for deriving the path
coverage-based test cases for a program:
1. Draw control flow graph for the program.
2. Determine the McCabe’s metric V(G).
3. Determine the cyclomatic complexity. This gives the minimum number of test cases
required to achieve path coverage.
4. Repeat
a. Test using a randomly designed set of test cases.
b. Perform dynamic analysis to check the path coverage achieved.
Unit 4 : Software Engineering [20IT3302] PVP20
Debugging:
After a failure has been detected, it is necessary to first identify the program statement(s) that
are in error and are responsible for the failure, the error can then be fixed.
Debugging Approaches:
The following are some of the approaches that are popularly adopted by the programmers for
debugging:
1. Brute force method:
● This is the most common method of debugging but is the least efficient method.
● In this approach, print statements are inserted throughout the program to print the
intermediate values with the hope that some of the printed values will help to identify
the statement in error.
● This approach becomes more systematic with the use of a symbolic debugger,
because values of different variables can be easily checked and breakpoints and
watchpoints can be easily set to test the values of variables effortlessly.
2. Backtracking:
● This is also a fairly common approach. In this approach, starting from the statement
at which an error symptom has been observed, the source code is traced backwards
until the error is discovered.
3. Cause elimination method:
● In this approach, once a failure is observed, the symptoms of the failure are noted.
● Based on the failure symptoms, the causes which could possibly have contributed to
the symptom is developed and tests are conducted to eliminate each.
4. Program slicing:
● This technique is similar to back tracking. In the backtracking approach, one often has
to examine a large number of statements.
● However, the search space is reduced by defining slices.
● A slice of a program for a particular variable and at a particular statement is the set of
source lines preceding this statement that can influence the value of that variable.
Debugging guidelines:
Debugging is often carried out by programmers based on their ingenuity and experience. The
following are some general guidelines for effective debugging:
● Many times debugging requires a thorough understanding of the program design. Trying to
debug based on a partial understanding of the program design may require an inordinate
amount of effort to be put into debugging even for simple problems.
● Debugging may sometimes even require full redesign of the system. In such cases, a
common mistake that novice programmers often make is attempting not to fix the error
but its symptoms.
● One must beware of the possibility that an error correction may introduce new errors.
Therefore after every round of error-fixing, regression testing must be carried out.
Unit 4 : Software Engineering [20IT3302] PVP20
Integration Testing:
● Integration testing is carried out after all (or at least some of) the modules have been unit
tested.
● Successful completion of unit testing, to a large extent, ensures that the unit (or module) as
a whole works satisfactorily.
● In this context, the objective of integration testing is to detect the errors at the module
interfaces (call parameters).
● The objective of integration testing is to check whether the different modules of a program
interface with each other properly.
● During integration testing, different modules of a system are integrated in a planned
manner using an integration plan.
● The integration plan specifies the steps and the order in which modules are combined to
realise the full system.
● After each integration step, the partially integrated system is tested.
● By examining the structure chart, the integration plan can be developed.
● Any one (or a mixture) of the following approaches can be used to develop the test plan:
1. Big-bang approach to integration testing:
● Big-bang testing is the most obvious approach to integration testing. In this
approach, all the modules making up a system are integrated in a single step.
● In simple words, all the unit tested modules of the system are simply linked
together and tested.
● However, this technique can meaningfully be used only for very small systems.
● The main problem with this approach is that once a failure has been detected
during integration testing, it is very difficult to localise the error as the error may
potentially lie in any of the modules.
2. Bottom-up approach to integration testing:
● Large software products are often made up of several subsystems.
● A subsystem might consist of many modules which communicate among each
other through well-defined interfaces.
● In bottom-up integration testing, first the modules for each subsystem are
integrated.
● Thus, the subsystems can be integrated separately and independently.
● The primary purpose of carrying out the integration testing of a subsystem is to
test whether the interfaces among various modules making up the subsystem
work satisfactorily.
● In a pure bottom-up testing no stubs are required, and only test-drivers are
required.
3. Top-down approach to integration testing:
● Top-down integration testing starts with the root module in the structure chart and
one or two subordinate modules of the root module.
● After the top-level ‘skeleton’ has been tested, the modules that are at the
immediately lower layer of the ‘skeleton’ are combined with it and tested.
Unit 4 : Software Engineering [20IT3302] PVP20
System Testing
● After all the units of a program have been integrated together and tested, system testing is
taken up.
● System tests are designed to validate a fully developed system to assure that it meets its
requirements.
● The test cases are therefore designed solely based on the SRS document.
● There are essentially three main kinds of system testing depending on who carries out
testing:
1. Alpha Testing: Alpha testing refers to the system testing carried out by the test team
within the developing organisation.
2. Beta Testing: Beta testing is the system testing performed by a select group of
friendly customers.
3. Acceptance Testing: Acceptance testing is the system testing performed by the
customer to determine whether to accept the delivery of the system.
● In each of the above types of system tests, the test cases can be the same, but the
difference is with respect to who designs test cases and carries out testing.
● The system test cases can be classified into functionality and performance test cases.
● Before a fully integrated system is accepted for system testing, smoke testing is
performed.
Smoke Testing
● Smoke testing is carried out before initiating system testing to ensure that system
testing would be meaningful, or whether many parts of the software would fail.
Unit 4 : Software Engineering [20IT3302] PVP20
● The idea behind smoke testing is that if the integrated program cannot pass even the
basic tests, it is not ready for vigorous testing.
● For smoke testing, a few test cases are designed to check whether the basic
functionalities are working.
● The system test cases can be classified into functionality and performance test cases.
Performance Testing:
● Performance testing is an important type of system testing.
● Performance testing is carried out to check whether the system meets the nonfunctional
requirements identified in the SRS document.
● There are several types of performance testing corresponding to various types of
non-functional requirements.
● All performance tests can be considered as black-box tests.
1. Stress testing:
● Stress testing is also known as endurance testing.
● Stress testing evaluates system performance when it is stressed for short periods of
time.
● Stress tests are black-box tests which are designed to impose a range of abnormal
and even illegal input conditions so as to stress the capabilities of the software.
● Input data volume, input data rate, processing time, utilisation of memory, etc., are
tested beyond the designed capacity.
● Stress testing is especially important for systems that under normal circumstances
operate below their maximum capacity but may be severely stressed at some peak
demand hours.
2. Volume testing:
● Volume testing checks whether the data structures (buffers, arrays, queues, stacks,
etc.) have been designed to successfully handle extraordinary situations.
3. Configuration testing:
● Configuration testing is used to test system behaviour in various hardware and
software configurations specified in the requirements.
● Sometimes systems are built to work in different configurations for different users.
4. Compatibility testing ;
● This type of testing is required when the system interfaces with external systems
(e.g., databases, servers, etc.).
● Compatibility aims to check whether the interfaces with the external systems are
performing as required.
5. Regression testing:
● This type of testing is required when a software is maintained to fix some bugs or
enhance functionality, performance
6. Recovery testing:
● Recovery testing tests the response of the system to the presence of faults, or loss of
power, devices, services, data, etc.
Unit 4 : Software Engineering [20IT3302] PVP20
● The system is subjected to the loss of the mentioned resources (as discussed in the
SRS document) and it is checked if the system recovers satisfactorily.
7. Maintenance testing:
● This addresses testing the diagnostic programs, and other procedures that are
required to help maintenance of the system.
● It is verified that the artifacts exist and they perform properly.
8. Security testing:
● Security testing is essential for software that handle or process confidential data that
is to be guarded against pilfering.
● It needs to be tested whether the system is fool-proof from security attacks such as
intrusion by hackers.