ExamGecko
Home Home / ISTQB / CTFL4

ISTQB CTFL4 Practice Test - Questions Answers, Page 8

Question list
Search
Search

Related questions











How can testing contribute to higher quality?

A.
Testing help to measure the quality of software.
A.
Testing help to measure the quality of software.
Answers
B.
Testing ensures that remaining defects are documented.
B.
Testing ensures that remaining defects are documented.
Answers
C.
Testing removes errors in the software.
C.
Testing removes errors in the software.
Answers
D.
Testing eliminates the risk with software.
D.
Testing eliminates the risk with software.
Answers
Suggested answer: A

Explanation:

Testing can contribute to higher quality by helping to measure the quality of software. Quality is defined as the degree to which a component or system satisfies specified requirements and customer or user needs and expectations. Testing is a process of evaluating a component or system by applying inputs and observing outputs, and comparing them with expected results. Testing can help to measure the quality of software by providing information on its functionality, performance, usability, security, reliability, etc. Testing can also help to identify and report defects in software, which can lead to improvement actions and quality assurance activities. The other options are not accurate descriptions of how testing can contribute to higher quality. Testing does not ensure that remaining defects are documented, but rather that detected defects are reported. Testing does not remove errors in software, but rather finds defects in software behavior or quality. Testing does not eliminate the risk with software, but rather assesses and manages the risk with software. Verified

Reference:A Study Guide to the ISTQB Foundation Level 2018 Syllabus - Springer, page 3.

A program got 100% decision coverage in a test. Which of the following statements is then guaranteed to be true?

A.
Every executable statement Is covered.
A.
Every executable statement Is covered.
Answers
B.
Every output equivalence class has been tested.
B.
Every output equivalence class has been tested.
Answers
C.
Every input equivalence class has been tested.
C.
Every input equivalence class has been tested.
Answers
D.
The 'dead' code has not been covered.
D.
The 'dead' code has not been covered.
Answers
Suggested answer: A

Explanation:

If a program got 100% decision coverage in a test, then it is guaranteed that every executable statement is covered. Decision coverage (also known as branch coverage) is a type of structural coverage (also known as white-box coverage) that measures how many decision outcomes have been exercised by a test suite. A decision outcome is a possible result of a decision point (such as an if-then-else statement) in a program's code. Decision coverage requires that each decision point has both true and false outcomes executed at least once by a test suite. Decision coverage implies statement coverage, which is another type of structural coverage that measures how many executable statements have been executed by a test suite. Statement coverage requires that each executable statement is executed at least once by a test suite. Therefore, if a program got 100% decision coverage in a test, then it also got 100% statement coverage in a test, which means that every executable statement is covered. The other options are not guaranteed to be true if a program got 100% decision coverage in a test. Every output equivalence class has been tested and every input equivalence class has been tested are not guaranteed to be true if a program got 100% decision coverage in a test, because equivalence classes are based on functional requirements or specifications, not on code structure or logic. Equivalence classes are used in specification-based testing (also known as black-box testing), which is a type of testing that does not consider the internal structure or implementation of the system under test. Decision coverage is used in structure-based testing (also known as white-box testing), which is a type of testing that considers the internal structure or implementation of the system under test. Therefore, achieving 100% decision coverage does not imply achieving 100% equivalence class coverage. The ''dead'' code has not been covered is not guaranteed to be true if a program got 100% decision coverage in a test, because dead code (also known as unreachable code) is code that can never be executed due to logical errors or design flaws. Dead code can reduce readability and maintainability of the code, as well as increase complexity and size. Decision coverage does not account for dead code, as it only considers the decision outcomes that are possible to execute. Therefore, achieving 100% decision coverage does not imply that the dead code has not been covered. Verified

Reference:A Study Guide to the ISTQB Foundation Level 2018 Syllabus - Springer, page 36.

Which of the following should be included in a test status report?

I . Estimation details

II . Total number of open and closed defects

III Actual effort spent

IV . Defect reports

V . Number of executed, failed, blocked tests

A.
III .V
A.
III .V
Answers
B.
II, III
B.
II, III
Answers
C.
I . II . IV
C.
I . II . IV
Answers
D.
II, III .V
D.
II, III .V
Answers
Suggested answer: D

Explanation:

The following should be included in a test status report: total number of open and closed defects, actual effort spent, and number of executed, failed, and blocked tests. A test status report is a document that provides information on the results and status of testing activities for a given period or phase. A test status report should include information that is relevant, accurate, and timely for the intended audience and purpose. Some of the information that should be included in a test status report are: total number of open and closed defects, which can indicate the defect trend and defect density of the software product; actual effort spent, which can indicate the productivity and efficiency of the testing process; number of executed, failed, and blocked tests, which can indicate the test progress and test coverage of the software product. The following should not be included in a test status report: estimation details, defect reports, and impact analysis. Estimation details are not part of a test status report, but rather part of a test plan or a test estimation document. Estimation details provide information on the expected time, resources, and costs for testing activities, not on the actual results or status of testing activities. Defect reports are not part of a test status report, but rather separate documents that provide detailed information on individual defects found during testing. Defect reports include information such as defect description, defect severity, defect priority, defect status, defect resolution, etc. Defect reports can be referenced or summarized in a test status report, but not included in full. Impact analysis is not part of a test status report, but rather part of a risk assessment or prioritization process. Impact analysis provides information on the potential effects or consequences of a change or a defect on the software product or project. Impact analysis can be used to evaluate the amount or scope of testing to be performed, but not to report the results or status of testing activities. Verified

Reference:A Study Guide to the ISTQB Foundation Level 2018 Syllabus - Springer, page 141.

In maintenance testing, what is the relationship between impact analysis and regression testing?

A.
Impact analysis requires a regression testing for only the tests that have detected faults in previous SW release
A.
Impact analysis requires a regression testing for only the tests that have detected faults in previous SW release
Answers
B.
There is no relationship between impact analysis and regression testing.
B.
There is no relationship between impact analysis and regression testing.
Answers
C.
Impact analysis requires a regression testing for all program elements which were newly integrated (new functionalities).
C.
Impact analysis requires a regression testing for all program elements which were newly integrated (new functionalities).
Answers
D.
The impact analysis is used to evaluate the amount of regression testing to be performed.
D.
The impact analysis is used to evaluate the amount of regression testing to be performed.
Answers
Suggested answer: D

Explanation:

In maintenance testing, the relationship between impact analysis and regression testing is that the impact analysis is used to evaluate the amount of regression testing to be performed. Maintenance testing is a type of testing that is performed on an existing software product after it has been delivered or deployed, in order to ensure that it still meets its requirements and functions correctly after a change or a modification. Maintenance testing can be triggered by various reasons, such as corrective maintenance (fixing defects), adaptive maintenance (adapting to new environments), perfective maintenance (improving performance), preventive maintenance (avoiding future problems), etc. Impact analysis is a technique that is used to assess the extent and nature of changes introduced by maintenance activities on the software product or project. Impact analysis helps to identify which parts of the software product are affected by the changes, which parts need to be modified or updated accordingly, which parts need to be retested or verified for correctness or compatibility, etc. Regression testing is a type of testing that verifies that previously tested software still performs correctly after a change or a modification. Regression testing helps to detect any side effects or unintended consequences of maintenance activities on the software product's functionality or quality. Regression testing can be performed at various levels and scopes depending on the impact analysis results. Therefore, in maintenance testing, impact analysis is used to evaluate the amount of regression testing to be performed. Verified

Reference:A Study Guide to the ISTQB Foundation Level 2018 Syllabus - Springer, page 20.

A software system checks age in order to determine which welcome screen to display. Age groups are:

Group I: 0-12

Group II; 13-18

Group III: over 18

Which of the below represent boundary values?

A.
(-1.0.12.13.18,19)
A.
(-1.0.12.13.18,19)
Answers
B.
(-1.0,11.12.13,14,18.19)
B.
(-1.0,11.12.13,14,18.19)
Answers
C.
(0.12.13.18.19)
C.
(0.12.13.18.19)
Answers
D.
(4.5.15.20)
D.
(4.5.15.20)
Answers
Suggested answer: A

Explanation:

A correct list of boundary values for the age input should include the minimum and maximum values of each age group (0, 12, 13, 18), as well as the values just below and above each boundary (-1, 19). Boundary value analysis is a test design technique that involves testing the values at or near the boundaries of an input domain or output range, as these values are more likely to cause errors than values in the middle. Option A satisfies this condition, as it has all six boundary values (-1, 0, 12, 13, 18, 19). Option B has two values from the same equivalence class (12 and 13), option C has only four boundary values (0, 12, 18, 19), and option D has no boundary values at all. Verified

Reference:A Study Guide to the ISTQB Foundation Level 2018 Syllabus - Springer, page 34.

Which of the following is an example of black-box dynamic testing?

A.
Functional Testing
A.
Functional Testing
Answers
B.
Code inspection
B.
Code inspection
Answers
C.
Checking memory leaks for a program by executing it
C.
Checking memory leaks for a program by executing it
Answers
D.
Coverage analysis
D.
Coverage analysis
Answers
Suggested answer: A

Explanation:

Functional testing is an example of black-box dynamic testing. Black-box testing (also known as specification-based testing) is a type of testing that does not consider the internal structure or implementation of the system under test, but rather its external behavior or functionality. Dynamic testing is a type of testing that involves executing the system under test with various inputs and observing its outputs. Functional testing is a type of black-box dynamic testing that verifies that the system under test performs its intended functions according to its requirements or specifications. Functional testing can be performed at various levels and scopes depending on the objectives and criteria of testing. The other options are not examples of black-box dynamic testing. Code inspection is an example of white-box static testing. White-box testing (also known as structure-based testing) is a type of testing that considers the internal structure or implementation of the system under test. Static testing is a type of testing that does not involve executing the system under test, but rather analyzing it for defects, errors, or violations of standards. Code inspection is a type of white-box static testing that involves examining the source code of the system under test for quality, readability, maintainability, etc. Checking memory leaks for a program by executing it is an example of white-box dynamic testing. Memory leaks are defects that occur when a program fails to release memory that it has allocated but no longer needs. Checking memory leaks for a program by executing it requires knowledge and access to the internal structure or implementation of the program, such as memory allocation and deallocation mechanisms, pointers, references, etc. Coverage analysis is an example of white-box static testing. Coverage analysis is a technique that measures how much of the code or structure of the system under test has been exercised by a test suite. Coverage analysis requires knowledge and access to the internal structure or implementation of the system under test, such as statements, branches, paths, conditions, etc. Verified

Reference:A Study Guide to the ISTQB Foundation Level 2018 Syllabus - Springer, page 7.

What is 'Component Testing'?

A.
Integration Testing
A.
Integration Testing
Answers
B.
Functional testing
B.
Functional testing
Answers
C.
Experience-based testing
C.
Experience-based testing
Answers
D.
A test level
D.
A test level
Answers
Suggested answer: D

Explanation:

Component testing is a test level. A test level is a group of test activities that are organized and managed together based on some common characteristics or objectives. A test level can be defined based on various factors, such as the scope and target of testing, the phase and model of development, the stakeholders and roles involved in testing, etc. Component testing (also known as unit testing or module testing) is a test level that focuses on verifying the functionality and quality of individual software components (such as modules, classes, functions, methods, etc.). Component testing can be performed by developers or testers using various techniques and tools depending on the type and complexity of the components. The other options are not test levels. Integration testing is another test level that focuses on verifying the functionality and quality of groups of software components that interact with each other or with external systems. Functional testing is a type of black-box dynamic testing that verifies that the system under test performs its intended functions according to its requirements or specifications. Experience-based testing is a category of test design techniques that rely on the tester's knowledge and intuition to derive and select test cases based on their experience with similar systems, technologies, domains, risks, etc. Verified

Reference:A Study Guide to the ISTQB Foundation Level 2018 Syllabus - Springer, page 19.

Which of the following is NOT an objective of testing?

A.
Finding defects
A.
Finding defects
Answers
B.
Providing information for decision-making
B.
Providing information for decision-making
Answers
C.
Gaining confidence about the level of quality of the software
C.
Gaining confidence about the level of quality of the software
Answers
D.
Analyzing and removing the cause of failures
D.
Analyzing and removing the cause of failures
Answers
Suggested answer: D

Explanation:

Analyzing and removing the cause of failures is not an objective of testing, but rather a task of development or maintenance. A failure is an event or behavior that deviates from the expected or specified result of a system under test. A failure is caused by an error (also known as a mistake or a fault) in the software code, design, or specification. Analyzing and removing the cause of failures is a process of locating and fixing errors in the software code, design, or specification, which is also known as debugging or defect resolution. Analyzing and removing the cause of failures does not aim to find or report defects, but rather to correct or prevent them. The other options are objectives of testing. Finding defects is one of the main objectives of testing, as it helps to improve the quality and reliability of the software product. Providing information for decision-making is another objective of testing, as it helps to support decision making and risk management. Gaining confidence about the level of quality of the software is another objective of testing, as it helps to assure that the software product meets its requirements and customer or user needs and expectations. Verified

Reference:A Study Guide to the ISTQB Foundation Level 2018 Syllabus - Springer, page 3.

An Incident Management tool implements the following defect states; Open, Assigned, Solved,

Closed Consider the following defect report:

Id T000561

Test Object 'Warehouse Management' application

Tester name; John Bishop

Date: 10th. April 2010

Test Case MRT558I

Status OPEN

Severity Serious

Priority

Problem- After inputting the Total Quantity item = 450 in the SV034 screen, the system shows an unexpected Error message=47

Correction:

Developer name:

Closing date:

Which of the following is a valid criticism of this report?

A.
The Priority, the Correction description and the Developer name are missing
A.
The Priority, the Correction description and the Developer name are missing
Answers
B.
The version of the application is missing
B.
The version of the application is missing
Answers
C.
There is no link to the applicable requirement (traceability)
C.
There is no link to the applicable requirement (traceability)
Answers
D.
The description is not highlighting the source of the problem
D.
The description is not highlighting the source of the problem
Answers
Suggested answer: B

Explanation:

A valid criticism of this report is that the version of the application is missing. The version of the application is an important piece of information that should be included in a defect report, as it helps to identify which release or build of the software product contains the defect. The version of the application can also help to reproduce and debug the defect, as different versions may have different behaviors or features. The other options are not valid criticisms of this report. The priority, the correction description and the developer name are not missing, but rather not applicable for this report. The priority is a measure of how urgently a defect needs to be fixed, which can be assigned by the project manager or the defect tracking system, not by the tester who reports the defect. The correction description and the developer name are information that are added after the defect has been resolved, not when it has been reported. There is no link to the applicable requirement (traceability) is not a valid criticism of this report, because traceability is not a mandatory attribute of a defect report, but rather an optional one. Traceability is a relationship between two or more entities (such as requirements, test cases, defects, etc.) that shows how they are related or dependent on each other. Traceability can help to verify that the requirements are met by the test cases and defects, but it is not essential for reporting a defect. The description is not highlighting the source of the problem is not a valid criticism of this report, because highlighting the source of the problem is not a responsibility of the tester who reports the defect, but rather of the developer who fixes the defect. The description should provide enough information to describe what happened when the defect occurred, such as input values, expected results, actual results, error messages, screenshots, etc., but it does not need to explain why or how it happened. Verified

Reference:A Study Guide to the ISTQB Foundation Level 2018 Syllabus - Springer, page 140.

Which of the following is a key characteristic of informal reviews?

A.
Kick-off meeting
A.
Kick-off meeting
Answers
B.
Low cost
B.
Low cost
Answers
C.
Individual preparation
C.
Individual preparation
Answers
D.
Metrics analysis
D.
Metrics analysis
Answers
Suggested answer: B

Explanation:

A key characteristic of informal reviews is low cost. Informal reviews are a type of review that does not follow a formal process or have any formal documentation. Informal reviews are usually performed by individuals or small groups of peers or colleagues who have some knowledge or interest in the product under review. Informal reviews can be done at any time and for any purpose, such as checking for errors, clarifying doubts, sharing ideas, etc. Informal reviews have low cost, as they do not require much time, effort, or resources to conduct. The other options are not key characteristics of informal reviews. Kick-off meeting is a characteristic of formal reviews, such as inspections or walkthroughs. Kick-off meeting is a meeting that is held before the review process starts, where the roles and responsibilities of the participants are defined, the objectives and scope of the review are agreed, and the logistics and schedule of the review are planned. Individual preparation is a characteristic of formal reviews, such as inspections or walkthroughs. Individual preparation is an activity that is performed by the reviewers before the review meeting, where they examine the product under review and identify any issues or questions that need to be discussed or resolved during the review meeting. Metrics analysis is a characteristic of formal reviews, such as inspections or walkthroughs. Metrics analysis is an activity that is performed after the review process is completed, where the data and results of the review are collected and analyzed to measure the effectiveness and efficiency of the review, as well as to identify any improvement actions or lessons learned for future reviews. Verified

Reference:A Study Guide to the ISTQB Foundation Level 2018 Syllabus - Springer, page 9.

Total 288 questions
Go to page: of 29