Validity of an instrument/tool refers to the
extent to which a data gathering instrument/tool measures what it
is supposed to measure, i.e., the extent to which the
instrument/tool obtain data that are relevant to what is being
measured. The greater the validity of an instrument, the more
confidence researcher can have that the instrument will obtain data
that will answer the research question or test the research
hypothesis.
There are following main approaches for establishing the
validity of a tool.
- Content validity The content validity of a
tool is the extent to which the tool represents the phenomenon
under study. It refers to the appropriateness of the content of an
instrument.
- Construct Validity It is an evaluation of the
extent to which an instrument measures the theoretical construct
that the investigator wishes to measure. A basic approach to
establishing construct validity is the known group technique, in
which the instrument is administered to several groups known to
differ on a certain construct.
- Criterion Validity Criterion validity of an
instrument is determined by comparing the results of an instrument
under testing with results of some “Gold
Standard”, it refers to the measuring instrument that is
being universally accepted or beyond doubt. Validity based on a
gold standard (GS) is called as Criterion
Validity.
- Formative Validity Formative validity is used
to assess how well a tool is able to provide information to help
improve the program under study.
Reliability is also known as precision,
reproducability, consistency and repeatability. Reliability is
usually expressed as a number, called a coefficient. Reliability
coefficient ranges between -1.00 to +1.00, with higher coefficients
indicating higher levels of reliability. However, most of the
instrument report reliability between 0.50 and 0.80.The three key
aspects of reliability are stability, internal consistency, and
equivalence.
- Stability It is a measure of the
instrument/tool to obtain similar results by administering the same
test twice over a period of time to a group of individuals. The
procedure used to assess the stability is the test-retest
reliability.
- Internal consistency Internal consistency
reliability or scale homogeneity is the extent to which all items
of an instrument measure the same variable. This type of
reliability is appropriate only when the instrument is examining
one concept or construct at a time. It is the most widely used
approach to assess the reliability particularly in psychological
instruments. The scales and tests that involve summing up of items
are usually evaluated for their internal consistency, for example
the scales measuring the attributes, attitudes etc.
Split-half technique also called odd-even
reliability yields the coefficient of internal
consistency.
- Equivalence The equivalence approach is
primarily used in observational measures. The commonest method used
to establish the reliability in equivalence approach is
Inter-rater / inter-observer reliability.
Inter-rater reliability is a measure of reliability used to assess
the degree to which different observers agree in their assessment
decisions. A researcher uses an observation technique.