HOW TO ASSESS A MATHEMATICIAN
Dawn, 16th August 2009
Professor Philip Hall (1904 - 1982),
FRS, an eminent mathematician was Sadlerian professor of Pure Mathematics at
Cambridge from 1953 to 1967. In daily Telegraph it has been stated that he was
the world’s leading group theorist of the 20th century. Recipient of
prestigious the Sylvester Medal of the Royal Society, the de Morgan Medal and
the Larimor prize of the London Mathematical Society, he was elected Honorary
Secretary and the President of the London Mathematical Society. Hall exercised
a profound influence on English mathematics, and an influence which was felt
throughout the mathematical world. Hall has published 40 papers in his career
of 47 years.
Hall’s
cumulative Impact Factor, based on
HEC’s criterion, is only 19.844. According to the HEC criterion for civil award
of Pakistan (Impact Factor for
Tamgha-i-Imtiaz 34 – 49, Pride of Performance 50 – 99, Sitara-i-Imtiaz 100 – 198, Hilal-i-Imtiaz 200), Hall will not qualify even for
Tamgha-i-Imtiaz. There are many more such outstanding mathematicians who would
not qualify the HEC criterion for awards, research projects, distinguish
professorship, etc., whereas, ironically, there are some Pakistani
mathematicians who are no match to such outstanding mathematicians such as
Professor Philip Hall and yet have received awards and research projects worth
millions of rupees.
The impact factor is a measure of the relative number of
citations in years 2 and 3. It is calculated by dividing the number of current
citations a journal receives to articles published in the two previous years by
the number of articles published in those same years. So, for example, the 1999
impact factor is the citations in 1999 to articles published in 1997 and 1998
divided by the number articles published in 1997 and 1998. The number that
results can be thought of as the average number of citations the average
article receives per annum in the two years after the publication year.
In June 2008, the
International Mathematical Union (IMU) issued a report “Citation Statistics”,
which addresses the use of quantitative measures in assessment of research. It
states that while numbers, that is, impact factors, appear to be “objective:
their objectivity can be illusory. The meaning of a citation can be even more
subjective than peer review. Because this subjectivity is less obvious for
citations, those who use citation data are less likely to understand their
limitations. It states that the sole reliance on citation data provides at best
an incomplete and often shallow understanding of research. Numbers are not
inherently superior to sound judgments. But citation data provide only a
limited and incomplete view of research quality, and the statistics derived
from citation data are sometimes poorly understood and misused. Research is too
important to measure its value with only a single coarse tool. It is important
that those involved in assessment will understand that if we set high standards
for the conduct of science, they should set equally high standards for
assessing its quality. (For details see the September 2008 issue of the NOTICES
of the American Mathematical Society.)
The value of the
impact factor is affected by sociological and statistical factors. Sociological
factors include the subject area of the journal, the type of journal (letters,
full papers, reviews), and the average number of authors per paper (which is
related to subject area). Statistical factors include the size of the journal
and the size of the citation measurement window.
In general,
fundamental and pure subject areas have lower impact factors than specialized
or applied ones. The variation is so significant that the top journal in one
field may have an impact factor lower than the bottom journal in another area.
Closely connected to subject area variation is the phenomenon of multiple
authorship. The average number of collaborators on a paper varies according to
subject area, from pure mathematics (with about two authors per paper) to
applied mathematics (where there are four or more). Not unsurprisingly, given
the tendency of authors to refer to their own work, there is a strong and
significant correlation between the average number of authors per paper and the
average impact factor for a subject area. So comparisons of impact factors
should only be made for journals in the same subject area.
The value of the
impact factor is affected by the subject area, type and size of a journal, and
the period of measurement used. Journals of Physics and Engineering for
instance, have much greater impact factors than the mathematical journals not
because they are qualitatively better but because they have a wider readership
and the time spent from acceptance of a paper to its publication is much
shorter. The ISI has listed 321 journals under
the subject of mathematics, and only 15.58 % mathematics journals
have impact factors greater than 1. Only 4 journals have impact
factors greater than 2, the highest being 2.75.
Use of impact
factors, outside of the context of other journals within the same subject area,
is virtually meaningless; journals ranked top in one field may be bottom in
another. Extending the use of the journal impact factor from the journal to the
authors of papers in the journal is highly suspect; the error margins can
become so high as to make any value meaningless. The use of journal impact
factors for evaluating individual scientists is even more dubious, given the
statistical and sociological variability in journal impact factors. Impact
factors, as one citation measure, are useful in establishing the influence
journals have within the literature of a discipline. Nevertheless, they are not
a direct measure of quality and must not be used.
In Pakistan, the
rationale for the use of the Impact Factor is to “help the administrators of
science to evaluate the quality and output of scientists who seek key
positions”. The National Commission on Science has made it a part of its policy
to rate scientists and their work on the basis of impact factors of their
research papers in accordance with the list of Impact Factors published by the
ISI. However, as Professor Milman in his article has argued, ISI’s impact
factor puts mathematicians in a disadvantageous position because the index is
not suitable for research in mathematics.
It is
bizarre that one's status could be determined by the arbitrary assignments of
numbers to the journals in which one happened to publish. Most intelligent
scientists and administrators are well aware of two facts: firstly, that we do
not yet have reliable bibliographic measures for comparing or making absolute
ratings of the value of the work done by research workers; secondly, that in
any event, bibliographic measures appropriate in one field are inappropriate in
others. An impact value based on the
simple measurement of how many times a journal is cited makes no sense as a
measure of the quality of the papers published in it, let alone the quality of
the mathematicians publishing there. Such use of management-type figures which
claim to enable comparisons to be made, can be utterly misleading and
damaging. Figures are only as good as
the premises on which the figures are based and often, the premises of many
widely-touted management figures are seriously flawed. The only real criterion of an individual’s
scholarship is quality of work and that does not admit of simple numerical
assessment.
The misuse of
impact factors and the citation index has already borne adverse effects in
Pakistan. Applied mathematicians for instance are at a more advantageous
position. They have the choice of publishing papers in journals of physics,
computational mathematics and engineering. Their choice of publishing a paper
also increases. These journals by and large have much higher impact factors
than the journals of mathematics. Consequently the cumulative impact factors of
mathematicians who have the choice of publishing papers in such journals
increase phenomenally.
Young Pakistani
mathematicians are now reluctant to do research in pure mathematics as they
feel that publishing papers in top mathematical journals is not only difficult
but receives no recognition or appreciation due to low impact factors and
citations. Fake multiple authorship has increased. Production of short papers
is on the rise. Worthwhile and exhaustive research is compromised. Meaningless
mathematical modeling is gaining popularity. Rate race of producing shallow
mathematical results with no scientific value has damaged the very purpose of
research in basic sciences.
Allocation
of funds, grants, awards, appointments, and promotions on the basis of Impact
Factor will be disastrous for science in Pakistan. The particular nature and
special characteristics of each branch of science have been completely ignored
in devising the Rs15.7 billion National Scientific and Technological Research
and Development Fund for encouraging scientific research in the country.
It does not
require much effort to understand from the information and analysis provided
above that the use of Impact Factor needs some modifications to suit the very
nature of mathematics. We can adopt a more pragmatic approach to this problem
rather than a protectionist approach.
Comments
Post a Comment