Testing Terminology

  • November 2019
  • PDF

This document was uploaded by user and they confirmed that they have the permission to share it. If you are author or own the copyright of this book, please report to us by using this DMCA report form. Report DMCA


Overview

Download & View Testing Terminology as PDF for free.

More details

  • Words: 2,901
  • Pages: 5
WELCOME

QA/QC SOLUTIONS

GLOSSARY OF TESTING TERMINOLOGY

PRACTICE OBJECTIVE This glossary of testing terminology has two objectives: first to define the test terms that will be used throughout this manual; and second to provide a basis for establishing a glossary of testing terminology for your organization. Information services (I/S) organizations that use a common testing vocabulary are better able to accelerate the maturity of their testing process. QAI believes that the testing terminology as defined in this glossary is the most commonly held definition for these terms. Therefore, QAI recommends that this glossary be adopted as a set of core testing term definitions. The glossary can then be supplemented with testing terms that are specific to your organization. These additional terms might include: ° Names of acquired testing tools ° Testing tools developed in your organization ° Names of testing libraries ° Names of testing reports ° Terms which cause specific action to occur

Audit: An inspection/assessment activity that verifies compliance with plans, policies, and procedures, and ensures that resources are conserved. Audit is a staff function; it serves as the "eyes and ears" of management. Automated Testing: That part of software testing that is assisted with software tool(s) that does not require operator input, analysis, or evaluation. Beta Testing: Testing conducted at one or more end user sites by the end user of a delivered software product or system. Black-box Testing: Functional testing based on requirements with no knowledge of the internal program structure or data. Also known as closed-box testing. Bottom-up Testing: An integration testing technique that tests the low-level components first using test drivers for those components that have not yet been developed to call the low-level components for test. Boundary Value Analysis: A test data selection technique in which values are chosen to lie along data extremes. Boundary values include maximum, minimum, just inside/outside boundaries, typical values, and error values. Brainstorming: A group process for generating creative and diverse ideas.

GLOSSARY OF TERMS Acceptance Testing: Formal testing conducted to determine whether or not a system satisfies its acceptance criteria—enables an end user to determine whether or not to accept the system. Affinity Diagram: A group process that takes large amounts of language data, such as a list developed by brainstorming, and divides it into categories. Alpha Testing: Testing of a software product or system conducted at the developer's site by the end user.

Branch Coverage Testing: A test method satisfying coverage criteria that requires each decision point at each possible branch to be executed at least once. Bug: A design flaw that will result in symptoms exhibited by some object (the object under test or some other object) when an object is subjected to an appropriate test. Cause-and-Effect (Fishbone) Diagram: A tool used to identify possible causes of a problem by representing the relationship between some effect and its possible cause.

Copyright © 1997 · Quality Assurance Institute® · Orlando, Florida GLOSSARY OF TESTING TERMINOLOGY April 1997

QC. 1.1.1

QA/QC SOLUTIONS Cause-effect Graphing: A testing technique that aids in selecting, in a systematic way, a high-yield set of test cases that logically relates causes to effects to produce test cases. It has a beneficial side effect in pointing out incompleteness and ambiguities in specifications. Checksheet: A form used to record data as it is gathered. Clear-box Testing: Another term for white-box testing. Structural testing is sometimes referred to as clear-box testing, since "white boxes" are considered opaque and do not really permit visibility into the code. This is also known as glass-box or open-box testing. Client: The end user that pays for the product received, and receives the benefit from the use of the product. Control Chart: A statistical method for distinguishing between common and special cause variation exhibited by processes. Customer (end user): The individual or organization, internal or external to the producing organization, that receives the product. Cyclomatic Complexity: A measure of the number of linearly independent paths through a program module. Data Flow Analysis: Consists of the graphical analysis of collections of (sequential) data definitions and reference patterns to determine constraints that can be placed on data values at various points of executing the source program. Debugging: The act of attempting to determine the cause of the symptoms of malfunctions detected by testing or by frenzied user complaints. Defect: NOTE: Operationally, it is useful to work with two definitions of a defect: 1) From the producer's viewpoint: a product requirement that has not been met or a product attribute possessed by a product or a function performed by a product that is not in the statement of requirements that define the product. 2) From the end user's viewpoint: anything that causes end user dissatisfaction, whether in the statement of requirements or not. Defect Analysis: Using defects as data for continuous quality improvement. Defect analysis generally seeks to classify defects into categories and identify possible causes in order to direct process improvement efforts.

QC. 1.1.2

Defect Density: Ratio of the number of defects to program length (a relative number). Desk Checking: A form of manual static analysis usually performed by the originator. Source code documentation, etc., is visually checked against requirements and standards. Dynamic Analysis: The process of evaluating a program based on execution of that program. Dynamic analysis approaches rely on executing a piece of software with selected test data. Dynamic Testing: Verification or validation performed which executes the system's code. Error: 1) A discrepancy between a computed, observed, or measured value or condition and the true, specified, or theoretically correct value or condition; and 2) a mental mistake made by a programmer that may result in a program fault. Error-based Testing: Testing where information about programming style, error-prone language constructs, and other programming knowledge is applied to select test data capable of detecting faults, either a specified class of faults or all possible faults. Evaluation: The process of examining a system or system component to determine the extent to which specified properties are present. Execution: The process of a computer carrying out an instruction or instructions of a computer. Exhaustive Testing: Executing the program with all possible combinations of values for program variables. Failure: The inability of a system or system component to perform a required function within specified limits. A failure may be produced when a fault is encountered. Failure-directed Testing: Testing based on the knowledge of the types of errors made in the past that are likely for the system under test. Fault: A manifestation of an error in software. A fault, if encountered, may cause a failure. Fault-based Testing: Testing that employs a test data selection strategy designed to generate test data capable of demonstrating the absence of a set of prespecified faults, typically, frequently occurring faults.

April 1997 GLOSSARY OF TESTING TERMINOLOGY

QA/QC SOLUTIONS Fault Tree Analysis: A form of safety analysis that written material that consists of two dominant assesses hardware safety to provide failure statistics and components: product (document) improvement and sensitivity analyses that indicate the possible effect of process improvement (document production and critical failures. inspection). Flowchart: A diagram showing the sequential steps of a process or of a workflow around a product or service. Formal Review: A technical review conducted with the end user, including the types of reviews called for in the standards. Functional Testing: Application of test data derived from the specified functional requirements without regard to the final program structure. Also known as black-box testing. Function Points: A consistent measure of software size based on user requirements. Data components include inputs, outputs, etc. Environment characteristics include data communications, performance, reusability, operational ease, etc. Weight scale: 0 = not present; 1 = minor influence, 5 = strong influence. Heuristics Testing: Another term for failure-directed testing. Histogram: A graphical description of individual measured values in a data set that is organized according to the frequency or relative frequency of occurrence. A histogram illustrates the shape of the distribution of individual values in a data set along with information regarding the average and variation. Hybrid Testing: A combination of top-down testing combined with bottom-up testing of prioritized or available components. Incremental Analysis: Incremental analysis occurs when (partial) analysis may be performed on an incomplete product to allow early feedback on the development of that product. Infeasible Path: Program statements sequence that can never be executed. Inputs: Products, services, or information needed from suppliers to make a process work. Inspection: 1) A formal evaluation technique in which software requirements, design, or code are examined in detail by a person or group other than the author to detect faults, violations of development standards, and other problems. 2) A quality improvement process for

GLOSSARY OF TESTING TERMINOLOGY April 1997

Instrument: To install or insert devices or instructions into hardware or software to monitor the operation of a system or component. Integration: The process of combining software components or hardware components, or both, into an overall system. Integration Testing: An orderly progression of testing in which software components or hardware components, or both, are combined and tested until the entire system has been integrated. Interface: A shared boundary. An interface might be a hardware component to link two devices, or it might be a portion of storage or registers accessed by two or more computer programs. Interface Analysis: Checks the interfaces between program elements for consistency and adherence to predefined rules or axioms. Intrusive Testing: Testing that collects timing and processing information during program execution that may change the behavior of the software from its behavior in a real environment. Usually involves additional code embedded in the software being tested or additional processes running concurrently with software being tested on the same platform. IV&V: Independent verification and validation is the verification and validation of a software product by an organization that is both technically and managerially separate from the organization responsible for developing the product. Life Cycle: The period that starts when a software product is conceived and ends when the product is no longer available for use. The software life cycle typically includes a requirements phase, design phase, implementation (code) phase, test phase, installation and checkout phase, operation and maintenance phase, and a retirement phase. Manual Testing: That part of software testing that requires operator input, analysis, or evaluation. Mean: A value derived by adding several qualities and dividing the sum by the number of these quantities.

QC. 1.1.3

QA/QC SOLUTIONS

QA/QC SOLUTIONS

Quality: A product is a quality product if it is defect Measure: Toproducer ascertaina or appraise comparing to aif it free. To the product is aby quality product standard; to apply to a metric. meets or conforms the statement of requirements that defines the product. This statement is usually shortened Measurement: 1) meets The act or process of NOTE: measuring. 2) to: quality means requirements. A figure, extent,the orwork amount obtained Operationally, quality refersbytomeasuring. products.

that includes desk checking, walkthroughs, technical Problem: Any deviation from defined standards. Same reviews, peer reviews, formal reviews, and inspections. as defect. Run Chart: A graph of data points in chronological Procedure: step-by-step followed to ensure order used toThe illustrate trendsmethod or cycles of the characthat standards are met. for the purpose of suggesting an teristic being measured assignable cause rather than random variation. Metric: Assurance Quality A measure(QA): of the extent The setorofdegree support to activities which a Process: The work effort that produces a product. This Scatter (correlation A graph designed product possesses (including facilitation, and exhibits training,ameasurement, certain quality,and includesPlot efforts of peoplediagram): and equipment guided by to show whether there a relationship between two property,needed analysis) or attribute. to provide adequate confidence that policies, standards, andisprocedures. processes are established and continuously improved in changing factors. order to produce that specifications Mutation Testing:products A method to meet determine test set and Process Improvement: To change a process to make 1) The relationship of characters or a group are fit for use.by measuring the extent to which a test set Semantics: thoroughness the process produce a given product faster, more characters to meanings, independent of the can discriminate the program from slight variants of the of economically, ortheir of higher quality. Such changes may Quality Control (QC): The process by which product manner of their interpretation and use. 2) The rate must program. require the product to be changed. The defect quality is compared with applicable standards; and the relationships symbols and their meanings. be maintainedbetween or reduced. action taken when nonconformance detected. Its Nonintrusive Testing: Testing that isistransparent to the focus is defect removal. This is change a line the Software software underdetection test; i.e., and testing that does not Product: Characteristic: The output of a An process; inherent, the work possibly product. accifunction, is, the performance of of these is the timing or that processing characteristics the tasks software There are dental, trait, three quality, usefulorclasses property of products: of software manu(for responsibility of its thebehavior people working the process. example, under test from in a realwithin environment. factured products functionality, (standard performance, and custom), attributes, administradesign Usually involves additional hardware that collects tive/information constraints, number products of states, (invoices, lines ofletters, branches). etc.), and Quality To changeand a production timing orImprovement: processing information processes process that service products (physical, intellectual, physiological, so that the rate at whichplatform. defective products (defects) are Software Feature: AProducts softwareare characteristic information on another and psychological). defined by specified a statement produced is reduced. Some process changes may require or implied by requirements documentation (for example, of requirements; they are produced by one or more Operational Requirements: the product to be changed. Qualitative and quantitafunctionality, performance, people working in a process.attributes, or design tive parameters that specify the desired operational constraints). Random Testing: An essentially Product Improvement: To change the statement of capabilities of a system and serveblack-box as a basistesting for deterapproach which a program is tested bysuitability randomlyof a Software requirements Tool:that A computer defines a product programtoused make to the helpproduct mining theinoperational effectiveness and choosing a subset of all possible input values. The more satisfying develop, test, analyze, and attractive or maintain to the another end user computer (more system prior to deployment. distribution may be arbitrary or may attempt to competitive). program or itsSuch documentation; changes maye.g., addautomated to or delete design from Operationalreflect Testing: Testing performed by in thethe end accurately the distribution of inputs the listcompilers, tools, of attributes testand/or tools, and the list maintenance of functions tools. defining user on software in its normal operating environment. application environment. a product. Such changes frequently require the process Standards: TheNOTE: measureThis usedprocess to evaluate to be changed. couldproducts result in and a Regression Testing: services, Selectiveor retesting to detect faultsto identify nonconformance. The basis upon which Outputs: Products, information supplied totally new product. introduced during modification of a system or system adherence to policies is measured. meet end user needs. Productivity: Procedures The ratio ofare theimplemented output of a process to the component, to verify that modifications have not caused Standardize: to ensure input, usually in the same units.atItaisdesired unintended adverse effects, or to verify that atomodified Path Analysis: Program analysis performed identify that the outputmeasured of a process is maintained frequently useful to compare the value added to a system or system still meetstoits specified all possible paths component through a program, detect incomlevel. requirements. plete paths, or to discover portions of the program that product by a process to the value of the input resources required (using fair market both input and Statement Coverage Testing:values A testfor method satisfying are not on any path. Reliability: The probability of failure-free operation for coverage output). criteria that requires each statement be aPath specified period. Coverage Testing: A test method satisfying executed at least once. Proof Checker: A program that checks formal proofs coverage criteria that each logical path through the Requirement: A formal of: program 1) an attribute to Statement Requirements: The correctness. exhaustive list of of programofproperties for logical program is tested. Pathsstatement through the often are be possessed the product or a function to from be each requirements that define a product. NOTE: The grouped into by a finite set of classes; one path Prototyping: Evaluating requirements or designs at the performed by the product; 2) the performance standard statement of requirements should document requireclass is tested. conceptualization phase, the requirements for the attribute or function; or 3) the measuring process ments proposed and rejected (including theanalysis reason for Peer A methodical examination of software phase, or design phasethe byrequirements quickly building scaled-down to be Reviews: used in verifying that the standard has been met. the rejection) during determination work products by the producer's peers to identify defects process. components of the intended system to obtain rapid Review: way tochanges use theare diversity and power of a and areasAwhere needed. feedback of analysis and design decisions. group of people to point out needed improvements in a Static Testing: Verification performed without Policy: Managerial desires andof intents concerning Qualification Formal testing, conproduct or confirm those parts a product in which executing the Testing: system's code. Also calledusually static analysis. either process (intended objectives) or products (desired ducted by the developer for the end user, to demonstrate improvement is either not desired or not needed. A Statistical Processmeets Control: The use of statistical attributes). that the software its specified requirements. review is a general work product evaluation technique

QC. 1.1.4

GLOSSARY OF TESTING TERMINOLOGY April 1997

April 1997 GLOSSARY OF TESTING TERMINOLOGY

QC. 1.1.5

Related Documents

Testing Terminology
November 2019 18
Terminology
July 2020 18
Terminology
July 2020 19
Terminology
November 2019 42
Terminology
May 2020 24