Thursday, May 17, 2012

Rational citation based incentives to maintain Panjab University H-Index rank(and rewards).

A message to colleagues that I think is applicable more generally.

Rational citation based incentives  to  maintain  Panjab University H-Index rank(and rewards).

Recently Hyderabad University, Panjab University,  Delhi University and Banaras Hindu University,  which rank as the top 4 Universities in the country with (SCOPUS)H-indices of 87, 85, 84, 78 respectively, were recognized  by  Department of Science and Technology of the Govt. of India and rewarded for this high H-index performance by the release of multi-crore PURSE grants.  In view of these developments it seems that  the obscurity shrouding  general understanding of the meaning of Citations, Impact Factors and other arcana of the Internet Publication Era is now liable to rapidly clear : since there is nothing quite like economic  self interest  to clear the mind ! This message is a contribution towards turning that inevitable processtowards rational policy and thus  a virtuous cycle .

   Three years ago I had strongly advocated that the use of Impact factor based incentives to reward faculty was absurd  (On the Absurdity of the use of Journal Impact factors as a measure of Individual Academic Excellence: http://chanaulakh.blogspot.in/2009/12/on-absurdity-of-use-of-journal-impact.html).  I argued that citation based incentives were much more likely to select faculty whose research would bring laurels and largesse to the University. Unfortunately no attention was paid  to my fulminations and the PURSE awards were used to hand out considerable amounts of discretionary spending money to reward individuals on the basis of  the utterly   meaningless concept of “Aggregate Impact Factor’’.   The absurdity of those awards is underlined by the fact that many  of the awardees do not figure at all   in the list of 85 papers which actually contributed to the University H-index of 85(see attached SCOPUS files), and conversely the authors actually on the H-index list received little if any recognition !  The persons rewarded were not even  selected  on the basis of their number of contributions   with more than, say, 50  citations: who could therefore rationally be expected to contribute to raising the University’s H-index and visibility. Instead,  based upon a complete misunderstanding of the meaning and significance of the very concepts of Journal Impact factors  and Citations,    PURSE monies were used to encourage not those with really outstanding research but instead those with other qualifications, connections or pretensions.  Now that all know that National authorities will base their recognition and  funding distribution  on the Citation achievements,  all stakeholders have a vital interest that the University sets up a reward system tuned to encourage real Citation achievements rather than irrelevant parameters such as the Impact Factors of journals published in. As  I had already argued in the cited blog and e-mail,  recognition on the basis of  Impact Factor of Journals  may actually   reward    individuals who lowered the IF of the journal they published in  if the citation count for the individual  publications is not   checked !  It is like rewarding someone for managing to drink tea at an exclusive club(measured by JIF) even if the result is that the club has to work twice as hard to clean up the mess he left behind !

 Since the bulk of the PURSE monies are shared among all the Science Departments and Science Faculty it is clear that such perverse policies are not only  unfair exploitation of the work of dedicated researchers but  also  damage the  entire University and Faculty in the long run. As I concluded earlier :

The UGC and university authorities should wake up to the developing absurdity that will soon entrench itself and become established wisdom that will entail another 50 years of academic mediocrity . They should  adopt numerical measures-- if at all they have the ability and self confidence to use numerical measures objectively and not a la Disraeli i.e as the evil third in ``lies , damn lies and statistics"-- that are in line with the best metrics available globally for measuring individual performance, and not further nourish the absurd  laurels for mediocrity  that are the bane of the quest for excellence in the Indian academic system.

I therefore propose that to encourage continued high rank of Panjab University(why not first ?!)  on the national H-index list the University should base its distribution of  special recognition funding (which is only a miniscule fraction of the total) from the PURSE grants released strictly on the basis  of  

1)   Contribution to the University H-index of 85 : those faculty and scholars whose papers figure on the list of 85 papers counted for the University H-Index should be recognized with special Research and Travel funding. When counting shared contributions of several faculty to the same paper   fractions may be used so that resources are not used up unfairly when say a paper on the list has, say,  10 co-authors from the University !

2)    Presence on the University’s top 200 by (SCOPUS)citation list(attached) /i60 list (i.e authors of those papers from the university which have more than 60 citations -this can be deduced from the attached list- but are not yet on the H-index=85  list) should  be considered for recognition, since it is a rational expectation that it is these authors/papers  who are likely to actually push the Panjab University H-index even higher next year or the next time the University’s H-index is evaluated by funding agencies.

Use of these criteria and a mathematical formula for proportionally distributing the earmarked funding on the basis of numerical scores associated with these criteria would go a long way in removing the  corrosive and counterproductive resentments felt by researchers when administrators select individuals for rewards in no sense commensurate with their achievements (not to speak of the later use of those very rewards as counters to garner yet more recognition : a perfect vicious cycle!). If these simple and rational measures are adopted  we may succeed in setting up a virtuous cycle by which each round of citation recognition funding (e.g PURSE) at the institutional level not only improves the facilities and funding for the University and Faculty as a whole but also  is sensibly and directly used to recognize and motivate those who will carry the University’s publication performance to new heights and thus to win it even more funding and recognition at the national level and so on.  Any other arbitrary system of rewards would  be more of  the  “kill  the golden goose” policies that bedevil our polity and academe. I appeal to the faculty to support the institution of a  fair and transparent  system  of research achievement recognition at the earliest.

  Charanjit S. Aulakh

On the usefulness of Citations in Academic Evaluation


 This article was actually written in March 2011. I just got around to "sounding off" about it now ! 
There is  increasing recognition by  governments world over  that  the vitality and excellence of   a  nation's  academic    and research  system  plays a central and   determining role in all aspects of  National  well being. It particularly   affects the ability of   nations to compete effectively in the globalized economic arena of  an ever  more knowledge driven world economy.   Increasing resources are being allocated to fund higher  education and research  and  there is a concomitant   increase in  anxiety - both among the governors and the governed - regarding the emplacement of  mechanisms of evaluation that will ensure that the resources invested are not feed for white elephants or polish for ivory towers  but rather select real talent,  nurture real  excellence , reward real hard work  and inspire real striving  : which will lead to  original contributions to the growth of  basic knowledge,   the development of  technology and the promotion of  human welfare  and culture in society at  large.  In this context the increasing use of Citation based measures of individual and institutional performance in all developed scientific systems  challenges  the Indian academic system’s  continued reliance on largely subjective or formal  indices of academic and scientific worth  coupled with an often  decried mediocrity in motivation,   achievement and innovation.  This challenge is particularly relevant  at this  conjuncture when the system needs to be reformed to evade the stultifying effects of entrenched networks of  patronage and mutual reward  of academic  and scientific elites . The recent vogue of publication of Global University rankings  (by  a multitude of agencies ranging from the Times of London to the University of Western Australia  or Shanghai ) which use Citation performance of University faculty and scholars as an important component of the  ranking method has also brought the issue to the front page of public perception.
Starting from the chronic underdevelopment of a colonial society,  India has made much progress  in  providing cohorts of  scientific and technical   staff who can contribute well when transplanted  to the   scientific ecosystems of  the developed world.  However the achievement of scientific excellence within the country, specially as regards  advances in fundamental science – as opposed to   localization of established scientific techniques and procedures - has been much more modest.  A considerable component  of the obstacles to   emergence  from scientific underdevelopment, backwardness,   and mediocrity  may well be the  fact that  hitherto the evaluation of  scientific merit of individuals and groups  has largely been left to the subjective evaluation   of the very elites whose mediocrity is  lamented . It has long been the lament of émigré  achievers that it was the certainty of being frustrated by brazenly unfair and opaque  under valuation of their talents that drove them to quit the country.     By the very nature of  science,  competence to evaluate scientific  work is generally  restricted  to   co-workers in the subfield ( for example   Particle  Physics is a sub field of Physics , which is itself divided into  practically autonomous  subdivisions, each with it’s own  standard working `lore’ ,  distinct sociology and  accepted  norms of performance and networking.  There are  divisions such as  Experimental High Energy Physics,  Phenomenology , Lattice Gauge Theory , and String Theory, and it is a rare researcher in one area who is competent to judge work in another area .  Since the density of workers in any sub-field  is rather low  (typically of the order of  1-3  individuals for any division actually present  in an academic Department) ,  the total  pool  of competent   evaluators in the country as a whole is typically less than a 100 individuals.   Naturally this very limited pool of  individuals are linked together by various professional ties   and transactions  as well as mutual interest in fostering professional development including placement in  jobs, placement of  students,  evaluation of theses,  promotions,  recognition  for awards and  so on .    The  restriction of   evaluation  to the subjective opinion of that  very set  of  national senior scientists  and academics, whose stagnation,  imitative repetitiveness and  mediocrity of achievement  the  Indian scientific system needs to bootstrap itself  away from,  functions as a  Catch-22  that    retards   progress.     It is an axiom of scientific administration that   mediocrity perpetuates  itself   by recognizing only the mediocre  while notionally genuflecting to distant  and  irrelevant eminences (the  ``scientists in America have shown syndrome’’ !).  
 There is thus a need   to by pass this obstacle  by introducing   objective   factors of evaluation that cannot be easily faked by a consensus of  the  mediocre,   already in positions of control or eminence,   whose deficient evaluation is difficult to  successfully challenge without recourse to indices that transcend  or bypass the local stagnation and underperformance.   If introduced , such mechanisms  will bias the  entire system positively toward the recognition of talent and performance  and lead to the  evolution of a pool  of actual (as opposed to notional)  talents.  In short there is an urgent and inescapable need for  objective and quantifiable inputs to the performance evaluation process  that can escape  the dangers of nepotism and patronage  influenced evaluation.  
  The need for  indices of achievement that are harder to fake by collusion amongst local/national  evaluators can  now be partially met by taking recourse to   statistical measures of scientific achievement  such as citations  of research workers in research publications  by co-workers in the same   field of research  and various derived measures based on such citations.  Of course the mere use of such indicators cannot  replace,  but only supplement, the  traditional indicators of talent and performance .  Application of mind –in  good faith !-  to the entire available range of information about performance   is - and will continue to be-  necessary to succeed  in attributing   merit    and rewards  in a way that will in the long run lead to the  development of excellence by global standards . Thus   one should  always keep in mind the caveats that: (i)    Evaluation    of  performance and potential can never become purely objective.(ii)  Nor can it escape the need for a the presumption of goodwill  and faith on the part of the evaluators. (iii)  Nor can the same yardstick and benchmarks  be applied uncritically across different subfields  and specializations.  If these reservations – and   additional  modulations  required in the light of subsequent experience when such metrics are used in practice -  are kept in mind then consistent  input  of citation information can     usefully supplement and improve  evaluation .
   For reasons very similar to those adduced above in the National context ,  and in spite  of well functioning and innovative scientific establishments , there is also Internationally a  drive towards greater reliance on   innovative  statistical   measures  designed  to identify  and rank  academics and   researchers according to the  depth  and range of their impact on society and  peers.   The increasing ability to amass and sort vast amounts of information using the Internet  has led to an qualitative change  in actual and perceived role of   bibliographic metrics for the evaluation of   academic performance.  Of course  citation databases  allow  one to rapidly trace the evolution of a scientific idea.  In addition,    researchers in most branches of science now routinely use  Citation Counts, H-index,  Author Rank(based on the Google  web page ranking algorithm !),  Impact Factors  and  so on to rapidly    evaluate  the scientific output of candidates and peers. We shall briefly review some of these terms below and comment on their significance.
  The citation trail that an individual 's scientific  publications generate is      accepted as a means of ascertaining the importance and relevance of the work done.   While scientists as individuals  may be prey to all the human frailties, nevertheless as  a group  they are committed to an “inter-subjective” conception of     truth : the quest for which motivates and drives their work.   The “truth'' in any subfield'' is arrived at by a dialectical process of  discussion and criticism in the format of the scientific journals which  generally  require    that  the prior history of study and research be adequately summarized and referred to as a precondition for publication.    Unjustified claims of scientific seminality or preeminence are always vigorously contested not only by those directly affected but also by  other knowledgeable individuals in the sub-field. Thus  there is an an internal policing mechanism within each  scientific sub-field   to prevent  pretensions and plagiarism   and thus award citation levels justified by actual merit. Moreover no geographical area is now  so dominant as to get away with     recognizing  only its own contributions.  Any sensible and genuine scientific worker knows he has little possibility to obtain   recognition for innovation or study unless he clearly identifies and distinguishes the work done prior to him .    Thus even  if they are reluctant to cite the works of others, specially  competitors (!) ,  there is really no alternative to doing so if they have scientific priority and precedence.   In this way tracking the citation counts and connections of an individual or group is generally an  effective way of tracking their actual impact on science.  
The  inter-subjectivity of   the scientific  process has long been consciously appreciated by scientists.  However   it is only over the   last half century  that rigorous bibliographic research  has established and conceptualized   how science progresses through (and  therefore  can be tracked by tracing)  the  linked percolation process  connecting  waves of scientific   papers that inspire succeeding papers that cite the papers that inspired them.  Such  research   has clearly established the “  inherent topological and graphical nature of the worldwide citation network”  as   a  crucial property of    scientific  discourse :  as proposed by R. Garner in 1965.   
“The use of citation counts to rank journals was a technique used in the early part of the nineteenth century but the systematic ongoing measurement of these counts for scientific journals was initiated by Eugene Garfield at the Institute for Scientific Information who also pioneered the use of these counts to rank authors and papers. In a landmark paper of 1965 he and Irving Sher showed the correlation between citation frequency and eminence in demonstrating that Nobel Prize winners published five times the average number of papers while their work was cited 30 to 50 times the average.”[Wikipedia]
 Since then multiple studies have re-confirmed the strong correlation between citation frequencies and scientific originality, relevance and influence at all levels of the achievement pyramid.  Before the advent of the digital revolution and the internet,  tracking bibliographic information  and making it available for reference by  scientists or scholars was a complex and almost prohibitive enterprise.  However the Information  Revolution has brought with it easy access to large bibliographic data bases that are now essential to almost every scientist in the internet era.   Some of the well known data bases are the  general purpose Science Citation Index  from Thomson Reuters available online through the Web of Science  database   and the SCOPUS data base available from  Elsevier as a part of the ScienceDirect  database .  Various specialized data bases such as the  Neuroscience Citation Index from  Thomson Reuters,   and the path breaking  and user charge free Stanford Public  Information Retrieval  System  for  High Energy Physics (HEP-SPIRES) are also commonly now  used via the internet thousands of times daily.
  I   illustrate my comments by reference  to HEP-SPIRES since it is available freely on the internet (http://www.slac.stanford.edu/spires/  and at many other ``mirror’’ sites maintained at various high energy physics  research centers across the globe ), is complete for it’s subfield (over a million records updated daily ) and simple to use and understand.  Similar databases are now available for Astrophysics (NASA), Chaos (Maryland), Computing(Computing reviews) ,  Mathematics(MathScinet,  Zentralblatt  MATH), Nuclear Physics (Nuclear Science References W3 Retrieval System )  and so on.  Commercially available data bases  such as the Web of Science  and Science Direct   offer many more `bells and whistles’’ on the basic   functionality of SPIRES .  A successor system (INSPIRE ) has begun to offer many of these  more complex bibliographical metrics.
As an   example  we give in Table 1  the results of a search of the HEP-SPIRES/INSPIRE    database  for   renowned Indian Theoretical High   Physicist Asoke Sen’s bibliographic data.  The search instantly   yields a summary page that shows just how high the  performance bar can be objectively  set even for scientists working within the country.  The summary reveals an extraordinary level of scientific productivity, excellence  and information. Over 212  scientific papers of which 202  are  papers published in recognized (by HEP-SPIRES criteria) scientific  media  and have been cited over 18000 times by various  scientific articles.     With an average number of citations per paper of 86  and an extraordinary  number of   over 100  ``TOPCITE’’  papers ( two with more than 500 but less than 700 citations , 19 with 250-499   citations ,  30 with 100-249 citations and so on)  the impressive citation counts are supplemented by his extraordinarily  high  H(for Hirsch)-index of  75.   The H-index is defined as the maximal number  H  of papers with H or more citations each  for the author in question. Thus  this author has  at 75 , but not 76, papers with 75 or more citations each.  It is known that the total number of citations  $N$ scales with the square of the h-index (with a coefficient of   3.5-4) .

Citation summary results
All papers
Published only
Total number of citable papers analyzed:
212
202
Total number of citations:
18,501
17,534
Average citations per paper:
87.3
86.8
Breakdown of papers by citations:


Renowned papers (500+)
2
2
Famous papers (250-499)
19
18
Very well-known papers (100-249)
32
31
Well-known papers (50-99)
48
45
Known papers (10-49)
86
82
Less known papers (1-9)
22
21
Unknown papers (0)
3
3



Additional Citation Metrics 


h-index
75
73
Table 1 :  Results of a search for citation data of renowned Indian String Theorist Asoke Sen on March 13, 2011.
It is interesting to contrast this record of  a largely Mathematically oriented physicist with that of the three giants who defined the Standard Model of Particle physics : Steven Weinberg (184 papers, 49012 citations, H-index 94, 114 topcites, 1 paper cited more than 7000 times one more than 2000 times and 7 more than 1000 times),  Abdus Salam (232 papers, 15679 citations, H-index 53, 58 topcites, 1 paper cited more than 3000 times  and one   more than 1000 times) and Sheldon Glashow  (150 papers, 26921 citations, H-index 59, 65 topcites, 2 papers cited more than 4000 times  and one each more than 3000, 2000 and 1000 times respectively).    At least to those knowledgeable  about these subjects the  citation levels and details   contain a wealth of information on the contributions   and stature of these scientists and the sociology of their fields of research. 
A use of the  freely available SPIRES database to study the citation levels of various other more or less eminent Indian  High Energy Physicists  is also most instructive. One finds   instances   of eminences,  members of various National Academies   and Directors of National institutes who   have accumulated only a few hundred citations in a long career spanning forty years yet habitually sit in exacting judgment on scientists 30 years their junior who already have better recognized work  as far as citation levels go!  Like the dynamic effects of the recently introduced  Right to Information  laws   on entrenched bureaucracies and patronage networks ,  citation  information   has the potential  to upset many comfortable academic apple carts  of the  powerful but often comatose  guardians of entrenched mediocrity in our academic and research system .
The HEP-SPIRES data base allows various methods for further refining the  metrics presented to remove doubts whether the citation record  presents a true picture of the esteem in which the  work done by the author is actually held by his scientific peers.    For instance  one may subtract  self citations  to  check whether the citation picture changes drastically thereby.  Another   check  that can prove  very revealing is the number of cites divided by the number of authors.  In fields such as Experimental  High Energy Physics where teams of thousands of individuals labour on an experiment and publish collectively  it often happens that the papers of the collaboration are cited thousands of times. Similarly when a working group of tens or hundreds of theorists publishes a ``white paper’’ summarizing a consensus in an evolving field it often receives a large number of citations. In such situations a re-normalization  is called for in order.  For example  a division by the number of authors  can however yield a figure of only a few citations per author ! Clearly the true picture regarding the scientific contribution of the individual scientists  in the huge collaborations must be sought by other means than a simple citation count.  In fact there exist methods in such fields for obtaining a true picture of an individuals talent and contribution by studying the  Collaboration Notes contributed by small sub-groups and their citation by the final published paper,  opinions of group leaders  etc.  This example underlines the dangers of uninformed  use of citation numbers,  specially when making comparisons between different subjects.  Thus the use of citation data requires an informed  formulation of appropriate norms for citation performance in each field separately.  The reader who repeats  the above  exercise (performed  using the INSPIRE database  at http://inspirebeta.net/search) will also soon  realize that the automated databases are bedeviled by familiar digital bugbears : the need to discriminate between individuals carrying closely similar names or initials.  There is thus also a need to provide  unique identity labels to authors which has not been addressed satisfactorily so far. 
Another source of confusion is the  use of  Journal Impact Factors  to judge the excellence of  individuals .  Some  administrators have begun to reward and recognize faculty on the basis of Cumulative  (Journal) Impact Factors.  To appreciate the absurdity   and negative impact of such a policy one should first be clear about what   Impact Factors  actually refer to.  Unfortunately many Indian  academics have only a hazy   idea of the meaning of this terms, or  perhaps  actually support the   introduction  of this policy for  even less admirable reasons.  The Wikipedia defines the Impact Factor(IF)  as follows :
"In a given year, the impact factor of a journal is the average number of citations to those papers that were published during the two preceding year. For example, the 2003 impact factor of a journal would be calculated as follows: A = the number of times articles published in 2001 and 2002 were cited by indexed journals during 2003. B = the total number of "citable items" published in 2001 and 2002. ("Citable items" are usually articles, reviews, proceedings, or notes; not editorials or Letters-to-the-Editor.) .2003 impact factor = A/B”
It also succinctly describes the reasons for the absurdity of using IF to judge individuals and how this fact is also widely appreciated by administrators in advanced countries :

"The impact factor is often misused to evaluate the importance of an individual publication or evaluate an individual researcher. This does not work well since a small number of publications are cited much more than the majority - for example, about 90% of Nature's 2004 impact factor was based on only a quarter of its publications, and thus the importance of any one publication will be different from, and in most cases less than, the overall number. The impact factor, however, averages over all articles and thus underestimates the citations of the most cited articles while exaggerating the number of citations of the majority of articles. Consequently, the Higher Education Funding Council for England was urged by the House of Commons Science and Technology Select Committee   to remind Research Assessment Exercise panels that they are obliged to assess the quality of the content of individual articles, not the reputation of the journal in which they are published."
 Note that  a policy that spuriously uses an index appropriate for journals to judge individuals is not only unjust but also  is   clearly  anti-National inasmuch it seeks to introduce a bias that will  tend to prevent Indian researchers from publishing in national journals such as Pramana     since these do  not have  anywhere near the same IF as many other journals.  Even a person who  performs work  that is  well cited (even though published in a low IF journal)  will not be recognized under such policies.   Indeed over time all sub-dominant If journals would be driven towards a fixed point of low or zero Impact Factor !  The logic behind such thinking seems to be  that of an outsider  who recounts the instances when he manages to penetrate an exclusive club  as a proof of his social merit : even if the club itself may rue the lowering of it’s tone(i.e of  it’s IF !)  by his entry (publication in that journal) ! Like all reasoning  based on snobbery, whether direct or inverted,  it cannot lead to healthy development or behaviour  and specially will not encourage individuals to  mount the challenges to orthodoxy and dominance that are the heart of  good science. Rather it is a genuflection to the established  scientific dominance  order which    countries like India need to challenge if they are ever to come into their own.
In spite of these facts being   well known and  critiques being easily accessible,  administrators in the Indian context continue to use IF and CIF as measures of individual performance.   It also bears mention that the detailed evaluation   criteria published by the UGC as a part of the revision of Pay scales under the 6th Pay Commission  make  no mention of  Citation levels as a basis for judgment of  research performance .   Little interest is evinced by teachers and researchers in Universities and National Institutions towards a reform  of the evaluation system to include citation data to provide a more accurate and objective picture  of individual performance.  Nor has the manifest perversity of using IF and CIF as opposed to more appropriate measures like  cumulative citation counts and  H-index, average citations per paper etc elicited much protest .  Indeed one gathers the impression that  many Indian academics are  motivated to cast doubts  on the citation based evaluation methodologies  using   arguments based on real or imagined anomalies that may occur  and thus  argue against even the  introduction of the use of citation data. It is amusing, however, that often the very same critics of citation based  evaluation are capable of waxing eloquent upon the influence of their own papers when they gain even a modicum of citations.  We have already mentioned that various caveats are in order   when using citation data  but it is our firm belief that  the use of citation data is essential and inevitable at least for those sectors of the  system that have performance levels that are not so low as to escape evaluation on the basis of citations completely !   India can ill afford an ostrich like attitude towards  such a positive  and enabling technological development which has the potential to debunk and dis-empower entrenched but dysfunctional elites and  turbo-charge our evolution as  a Global  scientific power.  Introduction and publication of citation data relevant to individuals selected for any scientific or academic position(and their unsuccessful competitors !)  would go a long way in reforming the academic evaluation system . Such an information  and   driven  reform  would be consonant with the spirit of the age as already seen in the radical effects of the Right to Information and Asset Disclosure  laws which have already opened the working of the Bureaucracy , Legislature and Judiciary to public oversight and reform and induced a welcome diffidence  among those in positions of  authority and power.  There is no reason why Academia and the Scientific Establishment   should continue to   shelter  from these refreshing winds of change and transparency behind the opacity of purely subjective metrics of evaluation.

Charanjit S. Aulakh

Saturday, July 10, 2010

IISER Admissions 2010 : A cruel Baboo designed Theater of the Absurd.

                                           

I have been troubled by the IISER admission issue for some months now, with my daughter's needs serving to focus my mind on it. Now that the issue of her personal benefit is almost settled (Although she topped the Chandigarh Tricity ICSE Medical and Non-medical results she has been denied a chance to compete for IISER admission even though just 0.3% short of the 93.3% Class XII cutoff for ICSE (based on 2009 results) and 94.2% in Class X relative to 95.6% cutoff based on 2007 Class X results).

As a parent I am quite upset,  as an academic I am puzzled about what should be my honest and upright attitude to the very fixed responses of the IISER bureaucracy on the issue of the distorted norms adopted by them for determining eligibility for IISER admission. I am trying to get some closure by thinking it through. However the more I think about it the more bizarre, defective and unexamined the whole rigged up scheme of IISER admission channels seems to me. The first two channels of admission to IISERs i.e via KVPY and IIT-JEE, though also defective inasmuch as they discriminate against Biology and perhaps Chemistry majors, are of minor importance in the overall scheme, since the bulk of admissions will be done through the so called third or Direct Channel based on school leaving marks. It seems to me that the debate on this channel so far has uncovered a glaring defect arising from the non-currency of the cutoffs used, but several fundamental issues concerning the very nature of the third channel have somehow remained completely unquestioned.

The IISERs were set up with much fanfare to extend quality scientific education and research opportunities to a much wider section of our young people than previously available in most parts of the country. To my knowledge the admission process was to be entirely merit based with no reservations other than a relaxation in cutoff marks available to SC-ST/OBC students to the extent of 10% / 5% concessions in the cutoff marks . However the third channel has imported a completely novel scheme of offering the top 1% cohort of each of the 29 School boards in India a chance to apply for admission to the 600 or so integrated 5 year MSc seats on offer at the various IISERs this year (of which perhaps 20% at most may be filled via KVPY or IIT-JEE scholars). Note that this is just a chance to apply and not admission itself. For normalization I believe about 900 applications have been accepted in this channel in 2010 though this remains to be extracted by RTI from the JAC-IISER. It also remains to be seen who these candidates are, how many of them will be offered admission and how many will finally join.

What is indisputable is that –apparently to circumvent the difficulty of normalizing the marks of different boards and thus to avoid injustice – a new principle has been used to determine merit according to the school Board origin of the student. Due to this novel new principle the entry level qualification varies from 68.1 %(Nagaland Board) to 95.6%(ICSE Board ) i.e by 27.5 % at the CLASS X level and from 67.2%(Jharkhand) to 95.8%(Viswa Bharati Board) i.e a 28.6% variation at the Class XII level ! These cutoffs are also known to exhibit yearly fluctuations to the extent of 5% as thoroughly discussed by Prof. A. Nangia in a scholarly article in Current Science (CURRENT SCIENCE, VOL. 98, NO. 11, 10 JUNE 2010) . Another absurd aspect of these cutoffs that has emerged from an RTI enquiry addressed to DST is that they are based upon figures provided by the different School boards irrespective of the stream taken by students ! Thus the cutoffs used to determine eligibility for science students may well be based on the performance of Commerce or other students besides being based on a previous year’s performance !!

At the very least one must conclude that the School Board results are considered almost completely unreliable to evaluate relative merit ! However the Baboos of the DST and IISER have ventured to cut the Gordian knot of determining the relative worth of different Board results by imposing at one fell swoop an equivalence ( for determining scientific merit for scholarshiops and IISER admission) between 67.2 % in the Jharkhand board to a score of 95.8% in the ViswaBharati Board or 93.3% in the ICSE Board ! This presumption is so absurd that it leaves one speechless. Any person with the slightest knowledge of Indian social reality knows that such an equivalence is completely untenable. All attempts to persuade these baboos that these manifest spatio-temporal fluctuations demand that the candidates falling within a reasonable band of variation must be permitted to compete for the IISER admission test have fallen on deaf ears. The JAC-IISER Chairman and the other science bureaucrats have stonewalled very attempt to prevent a whole year’s crop of rare aspiring scientists from suffering a grievous humiliation and disappointment due to the out of hand rejection by the IISER admission system. Short of a pronouncement from a High Court of Law there seems no way that some balance may be brought to this theatre of the absurd in which Baboo Red Queens run riot and lay down absurd equivalences as life and mind shattering Laws.

An absurd and cruel injustice will be perceived by the thousands of students denied a chance to compete when they are near the top of the best All India School Boards such as the CBSE/ICSE (to which it is well known that the best students gravitate ) while students who have as much as 27 % lower than them (and that too from State Boards often renowned for their defective examination systems and poor policing of examinations, not to speak of poor teaching and training) are entitled to apply for IISER admission. This injustice and immersion in a cruel Baboo Theater of the Absurd will embitter and make cynical some of our most talented students. How can they accept such manifestly distorted equivalences -when they are in the very throes of forming the self images that are to sustain and drive them in the long process of their scientific formation – without becoming cynical and embittered as so many Indian generations before them ? Is this not typical of the very means and ways in which the Indian Science system and society reinforces and reinstates it’s own mediocrity and injustice?

The implications for our society of this novel method of social engineering are also not trivial. Science is supposed to be based on objectivity and merit . The bureaucrats of the DST dreamt up inclusion in the top 1% of any of the 29 Board results (which is calculated irrespective of subject choice so that it may not include even a single science candidate !) as a criterion for award of DST-INSPIRE scholarship. By some far fetched social engineering principle this may seem justified. However the IISERs have adopted these same cutoffs with their enormous variations as a basis for excluding some the best and most talented students in the country from even applying for admission to IISERs while welcoming laughably inferior levels of preparation as candidates. Can this ever lead to the IISERs training crops of talented and productive scientists drawn from our best talent ?

In an earlier era 10% preferences and quotas applied in the professional educational system at large led to large scale riots and the souring of the minds of whole generations . In the present instance it is “only pure science” education that is at stake . Most of the brightest students do not even consider such a choice since it will lead to a frustrated life in our mediocrity and influence peddling dominated scientific system. The hope was that the IISERs would somehow bring quality Science education with impartial evaluation of merit and attract talented students to science to power India’s development into a Knowledge Society. Many were cynical and said it would be the same mediocrity generalizing and rewarding itself . It could well be that they will be proved right. Perversion of reason and truth and manifest injustice are not good classmates for brilliant scientists in search of natural truth.

Table 1 : http://www.iiser-admissions.in/marks/ gives the following cutoffs.
Board Class 10 % Class 12 % Andhra Pradesh 91.10 82.30 Assam 77.60 73.40 Bihar 73.20 73.40 CBSE 93.60 92.40 Chhattisgarh 82.50 85.00 Goa 85.80 81.80 Gujarat 86.40 77.80 Haryana 87.00 82.60 Himachal Pradesh 76.30 79.40 ICSE 95.60 93.30 Jammu & Kashmir 84.40 79.20 Jharkhand 78.40 67.20 Karnataka 89.60 88.00 Kerala 93.90 91.80 Madhya Pradesh 85.80 82.60 Maharashtra 80.10 83.80 Manipur 77.00 75.80 Meghalaya 70.50 70.60 Mizoram 86.60 68.20 Nagaland 68.10 73.00 Orissa 81.30 76.10 Punjab 81.30 76.70 Rajasthan 79.10 82.40 Tamil Nadu 90.40 94.50 Tripura 68.80 68.80 Uttar Pradesh 72.50 72.60 Uttarakhand 70.40 68.20 Viswa-Bharathi 93.40 95.80 West Bengal 86.60 82.20 This list is taken from the website
http://www.inspire-dst.gov.in/Inspire-Advertisement.pdf
(2007 class X and 2009 class XII) cut off Table 1. For SC/ST Candidates, the cut-off's for
X and XII will be 10% lower for all of the above entries,
e.g. CBSE Cut-off will be (83.60%) for X and 82.40% for XII.
Similarly, cut-off for OBC Candidate will be 5% lower i.e.
for CBSE (88.60%) for X and (87.40%) for