FEATURE
Information Quality, Liability, and Corrections
By Stephen Adams
All of us have suffered the consequences of
poor-quality information. For most of us, most of the
time, the impact has minor significance and is of short
duration. Perhaps we missed a bus or flight connection
as a result of using an out-of-date timetable, or we
lost an insurance claim because we failed to note the
change in exemptions to the policy when we last renewed.
As frustrating or painful as these examples may be,
they are rarely fatal. However, in a small percentage
of cases, poor quality information has direct, devastating
consequences. For example, many of the arguments concerning
personal privacy are based on the knowledge that an
adverse comment on a person's reputation perpetuates
itself, even after a formal retraction is published
or a libel case is won. Some sorts of information are
more "sticky" than others. Just as the garden weeds
are more robust than the desired plants, bad information
rears its ugly head more virulently than good information.
Nonetheless, information on its own is neither inherently
good nor bad. It is often a sequence of events
that leads to the consequences of what we simplistically
refer to as bad information. One striking example is
the Johns Hopkins clinical trials case, in which an
insufficient search in PubMed resulted in a death [1].
It is relatively rare to be able to track a chain of
events so directly to poor quality information capture,
dissemination, or retrieval. However, there are a few
examples when blame has been contested in the courts,
with varying results [2]. In one case, Dun & Bradstreet
was found liable for having inaccurately reported that
Greenmoss Builders had filed for voluntary bankruptcy
[3], while another case in Germany against a medical
publisher concerned a missing decimal point (advice
to "inject a 2.5% NaCl infusion" appeared as a "25%
NaCl infusion") [4]. If these cases prove anything,
it is that the "big, bad Internet" is not behind every
instance of poor quality information: The problem predates
that platform and goes much deeper.
Before considering any solutions to the problems
presented to information professionals by imperfect
information, we need to understand the nature of the
problemor, more accurately, multiple problemsat
the heart of "when information is wrong." Only then
can we make progress towards modifying our processes
in order to cope better in future.
DEFINING "WRONG" IN WRONG INFORMATION
It is vital to realize that each point in
the information dissemination chain is equally prone
to breakage. The fault, if it can be isolated at all,
can lie anywhere along the complex processes of publication,
collection, storage, dissemination, retrieval, or utilization.
Blaming the quality of the source data may be temptingly
easy, but simplistic. The Johns Hopkins case was due
to searcher shortcomings, and, to some extent, the
dissemination platform. Decisions were taken not because
the relevant information did not exist, or was wrong,
but because it was not found. Hence, the more we can
strengthen each individual link in the information
chain, the greater the chance of ensuring successful
transfer of the right information.
I can perceive at least five distinct ways in which
information can be wrong.
Inappropriate Quality
During the 1980s and '90s, when Total Quality Management
(TQM) was all the rage, many and various definitions
of "quality" were put forward. One of the most succinct
is simply that quality is "fitness for purpose," in
other words, the output matches the specification or
requirements of the putative user. It thus becomes
possible to speak of a "quality" Trabant car as well
as a "quality" Rolls Royce. The issue is not where
the vehicle fits on some hypothetical scale of reliability
or comfort, but whether it meets the reasonable expectations
of the user and is built in accordance with a set of
criteria known to both producer and consumer.
In information terms, this means that information
items geared towards one set of consumers may
be perceived as poor quality when located by a
different set. For example, most teenage mathematics
students know that the quantity pi has
been defined to a million decimal places, but
the home handyman will still be happy to use the
approximation of 22/7 to calculate most circular
perimeters. Likewise, my theoretical chemistry
tutor spent much of his professional life working
on the problem of whether benzene has six equal
ring bonds or three short and three long onesyet
in organic tutorials, we were not criticized for
drawing a regular hexagon to represent the compound.
Problems start when it becomes difficult to discern
the intended user of a piece of information, or when
users expecting one quality level encounter information
built to a different quality level. In past years,
a textbook became recognized over several editions
as an authoritative source by means of a process of
(more or less rigorous) testing in the marketplace.
Works tended to be focused on a defined user community.
A classic example is the CRC Press Handbook of Chemistry
and Physics, a 6-inch-thick tome of data tables
that has held open many a chemistry undergrad's room
door. The result of this process of targeted marketing
and testing was to establish a tacit hierarchy of information
qualitythe experienced user came to know what was
reliable and what was not.
The Internet publishing phenomenon has demolished
much of this information hierarchyanyone can now
produce a "world-leading" reference Web site or an
apparently authoritative Weblog. As the classic New
Yorker cartoon puts it, "On the Internet, no one
knows you're a dog." The information market has widened
substantially; the same information is now being offered
by multiple sources, increasing the risk that information
that is of perfectly suitable quality for one purpose
is being offered to, accessed, and used by users with
a totally different intended purpose. This leaves the
door open to both innocent mistakes and deliberate
abuse.
Ambiguous or Deliberately Fraudulent
Writing as a chemist, it is tempting to believe that
some types of information (such as physical properties)
can be measured by an absolute standard, and that this
standard becomes more precise as time goes on. Yet
recent discussions on the impact of the United States'
Data Quality Act [5] have highlighted the fallacy of
this belief. As a consequence of the Act, the Environmental
Protection Agency (EPA) has drafted a set of "Assessment
Factors for Evaluating the Quality of Information from
External Sources" [6], which attempt to establish standards
for data quality. However, to illustrate the scale
of the problem, a recent U.S. Geological Survey report
evaluated the variation in a single data item (the
octanol-water partition coefficient for certain pesticides)
reported in the chemical literature. More than 700
papers published over a 55-year period showed "up to
4 orders of magnitude variation with no convergence
with time" for the same data [7]. Possibly an extreme
case, but hardly an inspiring starting point.
In addition to "innocent" variation in information
quality as an artifact of the method of presentation,
there are known examples of deliberate fraud. One of
the most recent is that of J. H. Schön of Bell
Laboratories who "engaged in scientific misconduct
by falsifying and fabricating experimental data between
1998 and 2001" [8]. His publication record is extensive
and, although we now know his data is made up, there
remains an electronic trail in scientific/technical
databases.
Biased or Non-Objective
Prior to the advent of the Internet, professional
users assumed that information retrieved from an electronic
service was a faithful reflection of their search strategy.
The implementation of paid-for links on Internet search
engines has demolished that assumption. We can no longer
assume that the results presented to us as "the most
relevant" are the best fit to our strategy.
A failure to be objective in reporting is one of
the most subtle corruptions of information quality,
since it is one of the most difficult to detect by
anyone other than a specialist peer group. It is also
one of the most difficult to correct, if it manages
to enter the information chain. Consider the editorial
written by Trevor Laird, editor of Organic Process
Research & Development [9]. He notes that some
authors in his journal appear to deliberately skew
their citation listing, avoiding work by competitors
in the apparent hope that reviewers will not be aware
of competing or contradicting articles. Such a failure
of objective reporting can perpetuate itself through
the bibliographic system and result in the undermining
of the basic assumptions surrounding the citation process.
This can have particularly damaging effects when citation
analysis is used in many research evaluation techniques
and grant-awarding considerations.
Incomplete
On a visit to a film studio in California during
the 1980s, Pope John-Paul II made a speech to a group
of film directors and producers. This was at a time
when the portrayal of violence and sex in movies was
being debated as an issue touching upon the exercise
of the right to freedom of speech. The Pope made the
comment that "the proper exercise of the freedom of
speech demands that what is communicated is complete
and true." In the context of quality information,
I take this to mean that the distribution of a piece
of information, whether it's visual, verbal, or printed,
can become just as misleading through what it fails to
communicate as by what it actually attempts to show.
Consider the following examples. In the house journal
of the Royal Society of Chemistry, a researcher reported
that an accident had occurred in which the lid of a
steel drum, used for disposal of laboratory solvents,
had been blown off, apparently due to an increase in
internal pressure [10]. The contents of the drum were
known, as was the composition of the most recent additions.
The most likely cause was considered to be a reaction
between acetonitrile and ethanolamine. However, one
of the points of the letter was to highlight that "the
MSDS [Materials Safety Data Sheets] for ethanolamine
and acetonitrile...contain some confusing information." Clearly,
the author of the letter was a technically qualified
person, yet the very information source that should
have helped him to assess risk was not able to convey
the information in a concise and technically relevant
manner. One possibility is that the process of distilling
information into a usable note resulted in information
loss, which led to the accident.
A similar failure to ensure the communication of
the whole picture can be cited from my own field of
patent information. The U.S. Patent and Trademark Office
(USPTO) has a well-regarded Web site covering many
years' worth of U.S. patent documents. However, an
unskilled user can come away from this site with a
misleading picture of the patent situation. Under USPTO
practice, it is possible for an applicant to apply
for a Re-issue certificate (Re) if new information
affecting the proper scope of their patent becomes
available after the grant. If the Re-issue certificate
is granted, the predecessor document is treated as
withdrawn for the purpose of determining the owner's
enforceable rights. However, although the later (Re)
document cites its "parent," there is no link in the
other directionfrom the withdrawn document to its
successor. Only a specific search designed to locate
Re documents will reveal the true picture. By contrast,
value-added commercial databases, such as IFI Claims
and Derwent's WPI, assist the process by linking the
two documents. For the searcher who knows what can
happen, either source is acceptablebut for the nonskilled
searcher, the USPTO database's structure, allied with
insufficient searching knowledge, can lead to the wrong
answer.
Out-of-Date
The final possibility for defining "wrong" information
is that although a search strategy may be adequate,
the source unbiased, and the results technically accurate,
the answer may be out-of-date. This is clearly a live
issue when dealing with dynamic information such as
financial or business data. Knowledge of database update
policy is important to assess the usefulness of that
database to your particular type of search. In some
instances, no harm is done by a failure to identify
the most current data, as the situation will be rectified
later on in the supply chain. For example, a book search
may identify the second edition of a title. If the
requestor places an order, the book supplier should
alert the buyer that the third edition is now available.
However, as more and more end users are searching directly
for information, there may be no safety net of third-party
intervention, and out-of-date information may be recalled
and used as if it was the most current.
The best control over this eventuality is the application
of searcher skill and database knowledge, although
sometimes information suppliers can assist the process.
During the 1980s the British Standards Institution
(BSI), which publishes technical standards for manufacturing
in the U.K., experimented with printing its standards
in black text on red paper. This was an attempt to
control photocopying and hence prevent the perpetuation
of out-of-date technical standards; anyone other than
the original purchaser of the standard had to refer
to the BSI to obtain a new master copy. That reference
process ensured that the most up-to-date edition was
supplied.
FAILURE TO LEARN
Even in the best-regulated system, there's a failing
that is extremely hard to make provision forthe failure
to learn from experience. There are myriad examples
of information users reinventing the wheelor more
commonly, making the same mistakeas a result of a
failure to retrieve and utilize a comprehensive set
of teaching from previous recorded literature. In one
sense, the Johns Hopkins case illustrates this. The
search strategy may have been right, but the fault
lay in the failure to search comprehensively, back
to the earliest relevant item.
The phenomenon of"literature half-life" has been
recognized for many years. The frequency of citation
and/or retrieval decreases in an exponential fashion
over time. It becomes very easy to assume that the
long "tail" of older items retrieved in a search will
become less and less relevant, particularly within
a fast-moving technical field, where received wisdom
would state that "the relevant material can only have
been published in recent years." Today, this natural
tendency is exacerbated by a generation of searchers
who seem to assume that all information activities
started on the same day as the Internet, and that old
non-electronic information is passé. This
is anathema, particularly to patent searching, when
the law requires a presumption that literature of any age
has novelty-destroying potential. Consider two examples
of"bibliographic amnesia."
In September 2001, there was a fatal explosion
at an ammonium nitrate plant owned by Atofina.
In subsequent months, various letters appeared [11] speculating that
the explosion was caused by the same or similar
mechanism to that which occurred in the harbor of
Texas City
in April 1947, involving the same chemical. One
of the letters noted that a possible mechanistic
explanation
of the Texas City explosion had been in the literature
since 1960.
There has been much interest recently in so-called "economy-class
syndrome," a form of deep-vein thrombosis (DVT) apparently
linked to sitting for long periods in cramped conditions,
such as on long-haul aircraft flights. A letter [12] noted the existence of a "travel advisory" strikingly
similar in content to the current airline "well-being
programs," such as taking periodic breaks to move
around and exercising limbs in a sitting position.
The author
concerned had been writing about travel by stagecoach
in 1789.
THE PROCESS OF DISSEMINATION
In addition to the above two factors"wrong" information
being retrieved and a failure to learn from what could
be retrievedthere is a further risk: the perpetuation
of incorrect data. This is an acute problem now that
electronic communication has vastly increased the speed
at which new information can be added to the corpus
of knowledge, disseminated widely to both skilled and
unskilled users, reproduced, and transmitted again.
The speed with which an "urban myth" becomes established
is astonishing.
The damage is bad enough when something that started
out as correct information goes through a process of "electronic
Chinese whispers," but the wary user at least has the
chance to cross-check against earlier sources, as a
guard against corruption. But if the error is very
early in the dissemination chain, no amount of faithful
onward communication can alter it. Sometimes the error
may remain in the information system for decades or
even centuries. One example is the story of the clarification
of the chemical structure of Perkin's mauve, the first
synthetic dyestuff. It was originally synthesized in
1856 and had been assigned an incorrect chemical structure
since at least 1924. It was not until 1994 that modern
analysis of an original sample proved the correct structure
[13].
A similar example in more recent years comes from
the field of biochemistry. Authors McDonagh and Lightner
wrote to Chemical & Engineering News [14] to note that the chemical structure of bilirubin had
been incorrectly illustrated in an earlier article
in the same periodical. The author of the original
paper responded to their letter, conceding that they
were correct, but observing moreover that all three
major graduate-level biochemistry textbooks in the
U.S. [15, 16, 17] also reported the incorrect structure
in the current editions! Interestingly, in the course
of verifying the latter references, I discovered a
Web page entitled Howlers in General Biochemistry Textbooks
[http://bip.cnrs-mrs.fr/bip10/howler.htm] that also
laments the short-term influence of publishing information
about errors.
Peer review in the primary literature should guard
against the perpetuation of wrong data, backed up by
a periodic consolidation into secondary and tertiary
sources. A single paper in a reputable chemical journal
would be reviewed by one or more independent and qualified
peers before publication. At intervals, review articles
(secondary literature) collected all the related primary
papers into a more-or-less comprehensive bibliography,
with or without informed comment upon their content.
Over time, the teaching of individual papers would
be progressively distilled into the tertiary literature
of handbooks, encyclopedias and databooks, and student
textbooks.
Bottle [19] presents data suggesting that the length
of time between initial publication and first appearance
of the same information in a high school textbook is
decreasing. However, the speed of consolidation of
informationwith concomitant possibilities for erroris
not the only concern. There appears to be a parallel
reduction in the use of the entire primary-secondary-tertiary
information chain. Access to primary literature is
now so easy and so powerful that users are tempted
to merely re-run searches against a primary source
at regular intervals instead of utilizing the slower
process of independent data consolidation. For example,
the Beilstein Handbook of Organic Chemistry was
for decades acknowledged as a slow-publishing but fanatically
accurate data compendium. In the last 20 years, it
has moved away from its position of strength in an
effort to re-invent itself as a more rapid service,
abstracting the primary literature.
The dangers of this time pressure, leading to a breakdown
in the data distillation process, are twofold:
The primary literature now contains a larger
proportion of material that has not been
peer-reviewed at all.
The quality audit provided by
the secondary and tertiary services, which attempted to
place the primary literature in its proper context, has largely been swept
away.
Therefore, it becomes more difficult to see primary
literature in its proper context, geared towards a
specific user community, generated to a specific quality
standard. Open access can merely mean that it is open
to misuse. The second consequence is the failure of
the mechanism that has provided a useful filter at
the source (identifying "citation classics" and de-emphasizing
ephemeral items) and a form of bibliographic "version
control," collating any corrections and revisions into
a common format that ensures their subsequent retrieval.
IMPROVING THE SYSTEM
How can information science deal efficiently with
correcting information in this imperfect world? What
mechanisms do we have on hand that can help us cope
with errors and control or prevent their perpetuation?
The simple answer, at least in relation to much of
the scientific literature, seems to be, "Not many." Consider
the instances of known fraud. Although individual journals
have noted the "withdrawal" of the Schön papers,
the bibliographic databases rarely exert any form of
quality control, and references to these misleading
papers remain to be retrieved in future years, alongside
all the legitimate ones. There is no "health warning" attached
to the records. There is a fair likelihood that their
very notoriety will ensure that the Schön papers
continue to be cited, and probably more heavily than
would otherwise have happened. There is no control
in the scientific citation system to distinguish between
a critical citation and an approving one. Hence the
presence of these papers in the scientific literature
system is perpetuated. [Editor's note: It's distressing
that not much has changed since I wrote about problems
with identifying corrected articles. See Ojala, Marydee, "Oops!
Retractions, Corrections, and Amplifications in Online
Environments." Searcher, vol. 4, no. 1 (January
1996): pp. 30-41.]
There are a few examples of more formal correction
and quality control mechanisms, and it is worth considering
them in turn, for the extent to which they are currently
used and the possibility of expanding their application.
Publishing Retractions and Errata
Many established journals publish retractions or
errata notices. For example, the publishers of Online
Information Review in volume 26 number 2 published
a retraction of certain remarks made by one of their
columnists in volume 25 number 4. The later notice
appeared several issues after the original article.
This might have been some help to a regular reader
of the paper journal, but is next to useless for electronic
retrieval. Searchers are most unlikely to be alerted
to the existence of a retraction notice when they locate
the original article, since no link is made to the
later correction.
The hazard for the searcher remains; unless they
know, or have reason to suspect, in advance that a
correction or retraction has been issued, it does not
normally fall out in the standard search process. Surely
it cannot be right that modern retrieval mechanisms
for a later item which corrects an acknowledged mistake
lag so far behind those which serve to locate the erroneous
item? This would seem to be exactly what the hyperlink
was invented forbut journal publishers and secondary
service providers do not so far seem to have been applying
it for this purpose.
Capturing Ephemera
Many of the same remarks on the issue of published
retractions also apply to a situation where a first
author writes a full article, which is subsequently
commented upon by a second author through the letter
pages of the same journal. Not all bibliographic databases
cover the letters pages, even in respected journals,
so even if the later author cites the earlier article
(which is not guaranteed), there is a risk that the
later comments will be lost to the searcher. Yet these
pages can contain not only valuable comment upon papers,
but also original items. The letter pages of Chemistry
in Britain, the house journal of the U.K. Royal
Society of Chemistry, often contain items along the
lines of"I tried out the synthesis method of Smith
et al. and it blew up." It is notable that one of the
frequent respondents to these letters is the editor
of Bretherick's Handbook of Reactive Chemical Hazards,
which suggests that the publishers have a regular policy
of monitoring these informal reports.
Managing Expectations
In light of my comments on data quality, it is clear
that one hazard is the same piece of information being
used by many different communities. What is "quality" to
one is not to another. It is worth noting the efforts
that database producers have put into trying to tackle
this problem. The INSPEC bibliographic database of
engineering literature, produced by the Institution
of Electrical Engineers (IEE) in the U.K., commendably
devised a system of Treatment Codes, applied to each
entry in the database. The intent is to expand the
normal Document Type code to include an explicit indication
of the "expertise level" of each record. In a similar
fashion, the huge esp@cenet Web server of patent information
is intended as a "first-cut" source for small businesses
and provides only limited search possibilities and
limited help in interpreting the output. At the urging
of the professional user community, the service now
comes with a "health warning" that the output from
searches on this site cannot be considered authoritative
on questions of patentability.
Publishing Explicit Correction Documents
Returning to my own specialty of patents, there is
an interesting example of a policy for dealing with
corrections. After many years of negotiation, the World
Intellectual Property Organisation (WIPO) promulgated
its bibliographic standard ST.50 in 1998. The full
text is available at the WIPO Web site [19]. The purpose
of ST.50 is to establish a recommended mechanism whereby
patent offices ensure that corrections are published.
Corrections have to be notified in a way that ensures
they are retrieved by any search that would have retrieved
the original erroneous document. The principal mechanism
for this ensures that the document identifiers (specifically
the publication number) of both the original and correction
document are substantially identical, with merely a
modified suffix to denote the point in the information
chain to which the document belongs. For example, published
patent application EP 1266661 A1 is superseded by a
reprinted version EP 1266661 A9. All other bibliographic
data (applicant name, inventors' names, classification)
are identical between the two versions (unless of course
it is an error in one of these fields that the correction
document is replacing).
The details of ST.50 allow for the possibility of
different errors entering the information chain at
different points in the dissemination processfor
example, the paper document may be correct but its
corresponding facsimile edition on CD-ROM may be corrupted.
Specific document codes enable the user to identify
the nature of the correction made, and hence to determine
which is the "authority" version appropriate to the
search.
It cannot be claimed that the system is perfect,
especially as (with all WIPO standards) it is advisory
rather than mandatory and thus far has been adopted
by relatively few patent offices. However, it is being
applied to European Patent Office documents, and correction
documents are now entering the bibliographic system
through Web databases such as esp@cenet and conventional
online files such as Derwent's World Patent Index (WPI).
I suggest that it would be instructive for major database
producers to examine the principles and operation of
the standard, with a view to seeing whether it could
be applied to other forms of literature.
Utilizing the Tertiary Literature
The example of Perkin's mauve is striking not only
in the longevity of the error, but also in the argument
that it makes for intelligent use of a literature distillation
process. In all the years when the primary literature
was perpetuating the error, the Beilstein Handbook remained
uncommitted as to the structure of mauveine. At the
present time, there seems to be a consistent view that
the "compendium" approach to information is slow and
expensive; it is cheaper and easier simply to re-search
and re-collate the primary literature. I believe that
this attitude overlooks the very valuable contribution
made by compilationsthat of applying a second set
of eyes to the data, without time pressure, to produce
a considered opinion on the veracity of each piece
of primary information. It is true that the process
is labor-intensive and slow, but the question is ultimately
one of information qualityhow important to you is
the answer?
DO YOU WANT IT NOW OR DO YOU WANT IT TO BE CORRECT?
Today every searcher, be they expert or part-time,
is expected to produce results at high speed with a
high degree of accuracy. I believe that part of the
role of the information professional is to ask this
pertinent question: "Do you want an incorrect answer
quickly or a correct one more slowly?" This is not
a Luddite argument against technologyit is
an argument for quality, in the full meaning
of the word. There will be times when we need an answer,
any answer, and serve our user communities best by
providing an adequate response in a timely fashion.
But equally, there will be times when important commercial
consequences hang upon information provision, when
we may serve our users better by maintaining professional
standards and a rigorous approach to information retrieval.
If this means that we have to ask them to wait while
we check before delivering an answer, then so be it.
How Thomson ISI Handles
Corrected and Retracted Articles
Thomson ISI has, for all practical purposes,
cornered the market on citation reference searching.
How does the company handle corrections to articles
in its databases?
If either the publisher of the article or Thomson
ISI makes an error in any of the bibliographic
elements of an article, Thomson ISI will correct
the data. No general notice is given for this
type of correction. If an individual researcher
notifies Thomson ISI of an error in the bibliographic
elements of an article or citation referring
to their work, Thomson ISI will correct the data
and notify the researcher that the correction
has been made. If the publisher corrects the
content of a previously published article and
publishes this subsequent correction, Thomson
ISI indexes the new, corrected item, including
the original volume, page, and year of publication,
and lists the document type as a "correction" in
the Web of Science. When a publisher retracts
a previously published article and publishes
the retraction, Thomson ISI indexes it in the
same manner, assigns a document type of "correction," but
also includes the word "retraction" in the article
title. Thomson ISI also modifies the title of
the original article record to include the statement: "(Retracted
article. See vol X, pg XXX, YYYY)."
As Thomson ISI evaluates journals, the overall
quality of the science included is determined
by a number of factors, including the application
of peer review by the publisher, the prestige
of the sponsoring organization if present, and,
to some extent, the reputation of the publisher.
Obviously, it would be impossible to replicate
research from millions of articles as part of
this process.
Citations to published articles may occur for
either positive or negative reasons. Thomson
ISI records all citations without further qualification. |
REFERENCES
[1] Perkins, Eva, "Johns Hopkins Tragedy: Could
Librarians Have Prevented a Death?" [www.infotoday.com/newsbreaks/nb010806-1.htm] [2] Denis, S. and Y. Poullet, "Questions of
Liability in the Provision of Information Services." EUSIDIC
Newsletter (Newsidic) No. 100 (Apr. 1990): pp.
7-19.
[3] Dun & Bradstreet versus Greenmoss Builders.
472 US 749 (1985).
[4] Bundesgerichthof, Neue Juristische Wochenschrift
(1970), 1963.
[5] U.S. Public Law No. 106-554
[6] Hogue, C., "Assessing Data for Quality." Chemical & Engineering
News (February 10, 2003): pp. 21-22.
[7] Pontolillo, J. and R. P. Eganhouse, "The
Search for Reliable Aqueous Solubility (Sw) and Octanol-Water
Partition Coefficient (Kow) Data for Hydrophobic
Organic Compounds: DDT and DDE as a Case Study." Water-Resources
Investigations Report No. 01-4201, U.S. Geological
Survey (2001).
[8] Available online at http://www.lucent.com/news_events/researchreview.html [9] Laird, T., "The Importance of Adequate and
Accurate References." Organic Process Research & Development [http://dx.doi.org/10.1021/op030004z].
[10] Letters to the Editor. Chemistry in
Britain, vol. 38, no. 7 (July 2002): p. 20
[11] Letters to the Editor. Chemistry in
Britain, vol. 38, no. 2 (February 2002): p. 20
and vol. 38 no. 4 (April 2002): p. 22.
[12] Letters to the Editor. Chemistry in
Britain, vol. 38, no. 4 (April 2002): p. 24.
[13] Meth-Cohn, O. and A. S. Travis, "The Mauveine
Mystery." Chemistry in Britain, vol. 31, no.
7 (July1995): pp. 547-549.
[14] McDonagh, A. F. and D. A. Lightner, "Attention
to Stereochemistry." Chemical & Engineering
News (February 3, 2003): p. 2.
[15] Voet, D. and J. Voet, Biochemistry.
New York: Wiley, 1995 (2nd edition).
[16] Berg, J., J. Tymoczko, and L. Stryer, Biochemistry.
New York: Freeman, 2002 (5th edition).
[17] Lehninger, A. L., D. L. Nelson, and M.
M. Cox, Principles of Biochemistry. New York:
Worth, 1993 (2nd edition).
[18] Bottle, R. T., "Changes in the Communication
of Chemical Information. II: An Updated Model." Journal
of Information Science, vol. 6 (1983): pp. 109-113.
[19] Corrections, Alterations and Supplements
Relating to Patent Information. WIPO Standard ST.50
(1998) [www.wipo.int/scit/en/standards/pdf/st_50.pdf].
Stephen R. Adams, M.Sc., MCLIP [stevea@magister.co.uk] is managing director of Magister Ltd.
Comments? E-mail letters to the editor to marydee@xmission.com.
|