MEMO FOR MEMBERS OF THE HUMAN RESOURCES COMMITTEE (Sanitized)

Document Type: 
Collection: 
Document Number (FOIA) /ESDN (CREST): 
CIA-RDP83M00171R001500160011-3
Release Decision: 
RIPPUB
Original Classification: 
C
Document Page Count: 
7
Document Creation Date: 
December 16, 2016
Document Release Date: 
December 2, 2004
Sequence Number: 
11
Case Number: 
Publication Date: 
August 6, 1981
Content Type: 
MF
File: 
AttachmentSize
PDF icon CIA-RDP83M00171R001500160011-3.pdf342.87 KB
Body: 
Approved For lease 2004/12/22 : CIA-RDP83M0017 01500160011-3 CONFIDENTIAL DIRECTOR OF CENTRAL INTELLIGENCE Human Resources Committee OFFICE OF THE CHAIR11AN HRC-81-72 6 August 1981 21EIIORANDU11 FOR: Members of the Human Resources Committee FROM: ' I Chairman, Human Resources Conm.ittee 1. We now have sufficient guidance from the DCI and have had several exchanges with the DDCI which together enable us to formulate a coherent approach to the Community's human source collection activities and problems. The principal elements of this approach, which will in the future govern my own activities as Chairman of the Human Resources Committee and that of, my office as both the HRC staff and the HUMINT? component of the Intelligence Community Staff, are the following: a. We will continue to provide support to the varied human source collection activities of the Community"members. In this connection, I again urge all members of the HRC to bring to my attention problems and recommendations for solutions which are not susceptible to in-house solutions, which cut across agency lines or which require cooperation and support of non-intelligence departments or agencies and in which action by the Committee or by myself personally may be helpful. As necessary, we will form subgroups to address specific HU2IINT collection problems. b. As the primary means for providing national level guidance, and as the basis for monitoring human source collection, we will complete about 65 National HUiINT Collection Plans (heretofore known as National HUMINT Tasking Plans). The plans--as now--will consist of statements of prioritized collection objectives developed in coordination with policynakers and analysts and will provide for a rational. assignment of primary and supporting collection responsibilities worked out jointly with the Intelligence Community and other USG agencies with collection potential. The Collection Capabilities Group will be redesignated the Collection Coordinating Group to more accurately reflect the reality. DERIVATIVE CL BY SIGNS 5X1 REVIEW AUGUST 1987 Approved For Release 2004/12/22: CIA-RDP83M001 CONFIDENTIAL Approved For ease 2004/12/22: CIA-RDP83MOO17 01500160011-3 CONFIDENTIAL c. The principal means for monitoring human source collection performance will be assessment (or evaluation) of our performance in satisfying the defined collection objectives. These assessments will be jointly conducted by my staff together with the evaluation or assessment components of the Community elements. n this connection, I want to express my appreciation for the cooperation already demonstrated by the principal Community members in getting this process started. At the same time, I again urge agencies which do not yet have their own evaluation or assessment capabilities to establish these soonest. As the process takes hold, we will progressively meld the FOCUS reviews of individual diplomatic mission reporting into the jointly conducted overall assessment of collection performance against a given topical or geographic target. Ile will also simplify the FOCUS process. From time-to-time and on a very selective basis we will--again jointly with the appropriate Community members--conduct special evaluations of human source collection activities or programs. Here too, I solicit your recommenda- tions on areas where such studies would be helpful. d. Finally, we will continue to make a strong effort in the open-source area for the purpose of: (1) making certain that openly available sources are fully exploited, thus obviating the need for more complex and costly collection activities; (2) providing for maximum knowledge throughout the Community of open sources available to analysts and collectors; (3) eliminating undesirable redundancy; and (4) orienting open-source collection activities toward important national intelligence collection objectives. 2. The DDCI has reviewed this memorandum and approves the approach outlined above. Also, he has asked me to convey to you once again his strong interest in improving our human source collection capabilities. He is convinced that human source collection, with its variety and location both within and outside the Community, has suffered badly from personnel and funding degradation. He very much favors a concerted effort to rebuild our human source collection capabilities and has asked me to be particularly supportive of all soundly-based actions or proposals 25X1 Approved For Release 200#/a R/~J D gPrJZW83M001 R001500160011-3 Approved Foolease 200 P83 M 001 001500160011-3 DCI/RM-80-1951 15 September 1980 25X1 MEMORANDUM FOR: Director, HIMINT, Tasking Office I Program Assessment Office VIA: Director, Program Assessment Office SLBJECT: DoD's Intelligence Report Evaluation Program--A Statistical Review REFERENCES: A. NFIP and Resource Guidance FY 82-86 (9 May 80, DC I/Il4/3275-80 ) B. Army Clandestine HIMINT--A Review (5 Mar 80, DCI/RM 80-2001, Attachment II) C. DIA Response to CTS/HTO Questions (31 July 80) The DoD Intelligence Report (IR) evaluation program was developed to reflect the degree to which DoD Human Source reporting meets the requirements levied upon it. The program calls for roughly 20% of all 'IRs to be eval uated. Some IRs are automatically eval uated due. to the collection requirements that drive them; others are evaluated at the collector's initiative, while still others are evaluated at the initiative of DoD analysts. DoD analysts provide an IR evaluation by subjectively categorizing the value of an IR as with "high", "moderate", "low", "none" or "cannot judge". This statistical review evaluates the soundness of DoD's IR evaluation program. (S) Background: Statistically, samples are selected from a larger population according to some rule or plan. Generally, samples are-obtained, by one of two methods; those selected by some form of subjective judgmentand those selected according to some chance mechanism (such as random sampling). 25X1 25X1 WARNING NOTICE INTELLIGENCE SOURCES AND METHODS INVOL arA ved For Release 2004/ - 83M 25X1 SECRET Approved Foolease 2004/12/ X10017001500160011-3 A good sample is one from which generalizations to the population can be accurately and precisely made. To generalize from a sample to a population, the laws of mathematical probability must apply--random sampling assures that these laws do apply. For thisreason random samples are preferred over judgment samples. To generalize accurately and precisely from a sample to a population, the uncertainties in the sample must be understood. There are two components of sample uncertainty: reliability (numerical precision) and validity (accuracy or realism). Reliability is controlled for the most part by sample size, and can be calculated from the data at hand. Validity, however, cannot be judged from the data and can be controlled only before sampling through sound experimental design. (U). Discussion: DOD's sample size of roughly 20% provides for sufficiently precise estimates, IF THE SAMPLE IS VALIDLY CHOSEN. The percentage of IRs rated as having h gh value, for example, are precise to better than ?3% (95% Confidence Internal) based on the 20% sampl ing (See Appendix). In fact, a sample as small as 500 evaluations, if chosen properly, will provide precision to better than ?5% (95% Confidence Internal). It must be noted parenthetically that the precision of sample estimates is proportional to the number of IRs sampled and not to the percentage of IRs sampled. Reference CC states that samples were taken from each of some 120 individual collection entities. Care must be taken when examining separately each of these collection entities since their sample sizes may be quite small. On the average, one would expect the precision of estimates within a collection entity to be on the order of ?10-20% (95% Confidence Interval). However, it is not insufficent reliability but insufficient validity that undermines DoD's evaluation program. There are three primary causes of invalidity: (1) Systematic errors. According to Reference C, there is a tendency toinitiate evaluations of high-or-low-value reports at the expense of reports rated moderate in value. This practice results in the systematic elimination of a portion of the population and a consequent bias to inferences made from the sample. Reference B surfaces another source of systematic error: the inordinate number of high evaluations that upon closer examination appear to have been unwarranted. The effects. of such systematic overrating cannot be removed through statistical analysis and thus further undermine the validity of the inferences drawn from the sample. . Approved For Release 2004/12/SEGRP83M00171R001500160011-3. 25X1 25X1 25X1 Approved FcOelease SECRET 3M001 001500160011-3 (2) Mismatch between sample and o ulation. Reference B also isolates a serious mismatch between the sample and the population it purports to represent--the sample was taken primarily from the population of mid-level DoD analysts while inferences are drawn about the population of consumers (policymakers and senior analysts both inside and outside DoD). Since the value of a report to mid level analyst appears to be different from the value of the same report to other consumers (Reference B), one must seriously question the use to which DoD's summaries can be put. Furthermore, DoD's evaluation sample does not appear to match the total IR population in several other respects. The sample was not randomly chosen (i.e., each report did not have an equal chance of being evaluated), thus invalidating the mathematical basis for making inferences. As noted before, judgment sampling is not randan, and according to Reference C, "analyst initiative" evaluations are often intentionally biased to "reduce the ... IRs which ... are evaluated as being of low or no value." Likewise, it is not clear that special and initiative evaluations are representative of the total IR population, since they represent reports of some special, not random, interest. Failure to attend to the representativeness of the sample can lead to serious underestimates of uncertainty and consequent overoptimism about the stability and realism of population inferences. And estimates for which the accuracy is unknown can be quite misleading. (3) Correlated evaluations. If one analyst evaluates a inappropriate share of reports and has a tendency to rate reports higher or lower than other analysts, his evaluation may speciously inflate (or deflate) the estimated worth of IRs. His evaluations are said to be correlated, and correlated evaluations lower the validity of an analysis. Likewise, if several evaluations are performed on a single requirement (or similar requirements), there is again the tendency for such correlated evaluations to artificially alter population estimates. There is potential for such correlated evaluations in "analyst initiative" reporting. (S) Approved For Release 2004/ R-DP83M00171R001500160011-3 25X1 Approved For Wase 2004/1 001701500160011-3 Conclusions: o If the intent is to understand the value to consumers of IRs as a whole, mandatory evaluation must be randomly assigned to 10% or so (depending upon the accuracy desired) of all reporting to match sample to population and to provide for suit icient reliability. Furthermore, since mid-level analysts provided evaluations from their own perspective, results will be valid only for these analysts. Inferences about other consumers are invalid unless it can be shown that the attitudes and perspective of mid-level analysts are like those of the other consumers. o "Initiative" and specially requested evaluations, while they may be useful for other purposes, should not be included in the data analysis due to their systematic biases and potential for correlated evaluations. o The assertion in Reference B that the Intelligence Community "cannot rely upon such evaluations for an objective view of the worth of the reporting" appears to be based on. an invalidating mismatch between sample and population. o The violation of such fundamental laws of validity renders the DoD Evaluation of questionable value for estimating the worth of intelligence reporting to consumers. (S) 25X1 25X1 Approved For Release 2004/1 25X1 Approved Forease 2004/1 L? 1 0 3M0017 01500160011-3 25X1 APPENDIX. Statitical Foundation for Estimates of Precision. DoD defines the value of an IR as either "high", "moderate", "low", "none" or "cannot judge". These categories form a well-defined statistical population known as a multinomial population. When samples are randomly placed into mul ti nomial categories, the percentage of the total sample falling in each category can easily be calculated. The variance (a measure of precision) of each percentage, P, is defined as: Variance = [P(100-P )] + N, where N is the total sample size. For example, if 70% of 2,000 eval uations are rated as "moderate" in val ue, the precision of this 70% is given by: [70(30)] + 2000 = 1.05. A 95% Confidence Internal is approximated by twice the square root of this number, or about 2. Therefore, the 70% is precise to within ?2% (at a 95% level of confidence). In other words, if this evaluation were repeated 100 times, one would expect the proportion of "moderate" ratings to be between 68% and 72% 95 times, and outside that range only 5 times. (S) Approved For Release 20 71?~-RDP83MOO171ROO1500160011-3