Intelligence analysis

Intelligence analysis is the application of individual and collective cognitive methods to weigh data and test hypotheses within a secret socio-cultural context.[1] The descriptions are drawn from what may only be available in the form of deliberately deceptive information; the analyst must correlate the similarities among deceptions and extract a common truth. Although its practice is found in its purest form inside national intelligence agencies, its methods are also applicable in fields such as business intelligence or competitive intelligence.

Overview

Analysis is part of the Intelligence Process or Cycle

Intelligence analysis is a way of reducing the ambiguity of highly ambiguous situations. Many analysts prefer the middle-of-the-road explanation, rejecting high or low probability explanations. Analysts may use their own standard of proportionality as to the risk acceptance of the opponent, rejecting that the opponent may take an extreme risk to achieve what the analyst regards as a minor gain. The analyst must avoid the special cognitive traps for intelligence analysis projecting what she or he wants the opponent to think, and using available information to justify that conclusion. Being aware that one's enemies may try to confuse is a relevant factor, especially in the areas of intelligence cycle security and its subdiscipline counterintelligence. During World War II, the German word for counterintelligence art was Funkspiel, or radio game—not a game in the sense of playing fields, but something that draws from game theory and seeks to confuse one's opponents.[2]

A set of problem-solving talents are essential for analysts. Since the other side may be hiding their intention, the analyst must be tolerant of ambiguity, of false leads, and of partial information far more fragmentary than faces the experimental scientist. According to Dick Heuer,[3] in an experiment in which analyst behavior was studied, the process is one of incremental refinement: "with test subjects in the experiment demonstrating that initial exposure to blurred stimuli interferes with accurate perception even after more and better information becomes available...the experiment suggests that an analyst who starts observing a potential problem situation at an early and unclear stage is at a disadvantage as compared with others, such as policymakers, whose first exposure may come at a later stage when more and better information is available."

The receipt of information in small increments over time also facilitates assimilation of this information into the analyst's existing views. One item of information may not be sufficient to prompt the analyst to change a previous view. The cumulative message inherent in many pieces of information may be significant but is attenuated when this information is not examined as a whole. The Intelligence Community's review of its performance before the 1973 Yom Kippur War noted [in the only declassified paragraph].[3]

The problem of incremental analysis—especially as it applies to the current intelligence process—was also at work in the period preceding hostilities. Analysts, according to their own accounts, were often proceeding on the basis of the day's take, hastily comparing it with material received the previous day. They then produced in 'assembly line fashion' items which may have reflected perceptive intuition but which [did not] accrue from a systematic consideration of an accumulated body of integrated evidence.

Writers on analysis have suggested reasons why analysts come to incorrect conclusions, by falling into cognitive traps for intelligence analysis.[4][5][6] Without falling into the trap of avoiding decisions by wanting more information, analysts also need to recognize that they always can learn more about the opponent.

Analytic tradecraft

Intelligence reflects a progressive refinement of data and information

The body of specific methods for intelligence analysis is generally referred to as analytic tradecraft.[7] The academic disciplines examining the art and science of intelligence analysis are most routinely referred to as "Intelligence Studies", and exemplified by institutions such as the Joint Military Intelligence College, University of Pittsburgh Graduate School of Public and International Affairs (Security and Intelligence Studies major), and Mercyhurst College Institute for Intelligence Studies. The goal of the Analytic Tradecraft Notes of the Central Intelligence Agency's Directorate of Intelligence (DI) include the

Pursuit of expertise in analytic tradecraft is a central element of this plan. Our tradecraft enables analysts to provide "value-added" to consumers of intelligence by ensuring:

  • Dedication to objectivity – tough-minded evaluation of information and explicit defense of judgments – which enhances our credibility with consumers dealing with complex and sensitive policy issues.
  • Delivery of our products to the right people in time to be useful in their decision-making, and using feedback and tasking from them to drive the collection of the basic intelligence that we need to produce our analysis.

Analytic tradecraft skills also serve as "force multipliers", helping us provide top-quality analysis:

  • The feedback our customers give us on our customized analysis clarifies for the analyst what questions most need answering.
  • Employing rules for evaluating information and making judgments helps analysts manage the deluge of information, discern trends, and identify attempts at deception.
  • Tradecraft standards can be used to iron out differences among experts who have complementary substantive specialties. Their interaction enhances teamwork, which allows the [Directorate of Intelligence] to be more productive.

On January 2, 2015, the Office of the Director of National Intelligence (ODNI) issued Intelligence Community Directive (ICD) 203, which "establishe[d] Intelligence Community (IC) Analytic Standards that govern the production and evaluation of analytic products; articulates the responsibility of intelligence analysts to strive for excellence, integrity, and rigor in their analytic thinking and work practices..."[8]

Setting goals for an intelligence analysis

Stating the objective from the consumer's standpoint is an excellent starting point for goal-setting:

Ambassador Robert D. Blackwill ... seized the attention of the class of some 30 [intelligence community managers] by asserting that as a policy official he never read ... analytic papers. Why? "Because they were nonadhesive." As Blackwill explained, they were written by people who did not know what he was trying to do and, so, could not help him get it done: "When I was working at State on European affairs, for example, on certain issues I was the Secretary of State. DI analysts did not know that—that I was one of a handful of key decision makers on some very important matters."

More charitably, he now characterizes his early periods of service at the NSC Staff and in State Department bureaus as ones of "mutual ignorance"

"DI analysts did not have the foggiest notion of what I did; and I did not have a clue as to what they could or should do."[9]

Blackwill explained how he used his time efficiently, which rarely involved reading general CIA reports. "I read a lot. Much of it was press. You have to know how issues are coming across politically to get your job done. Also, cables from overseas for preparing agendas for meetings and sending and receiving messages from my counterparts in foreign governments. Countless versions of policy drafts from those competing for the President's blessing. And dozens of phone calls. Many are a waste of time but have to be answered, again, for policy and political reasons.

"One more minute, please, on what I did not find useful. This is important. My job description called for me to help prepare the President for making policy decisions, including at meetings with foreign counterparts and other officials...Do you think that after I have spent long weeks shaping the agenda, I have to be told a day or two before the German foreign minister visits Washington why he is coming?"

Be bold and honest

Weasel-wording is problematic in intelligence analysis; still, some things truly are uncertain. Arguably, when uncertainties are given with probabilities or at least some quantification of likelihood, they become less a case of weasel wording and more a case of reflecting reality as it is best understood.

While a good analyst must be able to consider, thoughtfully, alternative viewpoints, an analyst must be willing to stand by his or her position. This is especially important in specialized areas, when the analyst may be the only one that reads every field report, every technical observation on a subject.

"Believe in your own professional judgments. Always be willing to listen to alternative conclusions or other points of view, but stand your ground if you really believe the intelligence supports a certain conclusion. Just because someone is your boss, is a higher grade, or has been around longer than you does not mean he or she knows more about your account than you do. You are the one who reads the traffic every day and who studies the issue". At the same time, Watanabe[10] observes, "It is better to be mistaken than wrong". Not willing to be wrong is also a disease of the highest policymaker levels, and why there needs to be a delicately balanced relationship, built of trust, between a policymaker and his closest intelligence advisors.

"Being an intelligence analyst is not a popularity contest...But your job is to pursue the truth. I recall a colleague who forwarded an analysis that called into question the wisdom behind several new US weapon systems. This analysis caused criticism of the CIA, of his office, and of himself. He stood his ground, however; the Agency supported him, and eventually he was proven right. He did not make a lot of friends, but he did his job.[10]

Intelligence analysts are expected to give policymakers' opinions both support and reality checks.[7] The most effective products have several common features:

  • Opportunities and dangers for interests of the analyst's country, especially unexpected developments that may require a reaction.
  • Motives, objectives, strengths, and vulnerabilities of adversaries, allies, and other actors.
  • Direct and indirect sources of friendly parties' leverage on foreign players and issues.
  • Tactical alternatives for advancing stated national policy goals.

Reality checking is not to be underestimated. In World War II, the Allies launched an air offensive against a target system that they really did not understand: the V-1 cruise missile. Their rationale to attack ("if the enemy apparently valued it, then it must be worth attacking") may have been rational when there were large numbers of aircraft and pilots, but it might not be applicable to current situations, at least not until analysts rule out the possibility of the target system being a decoy.[11] If the threat is real, then it might be warranted to defer attack until a massive one can be delivered.

Agreement on content

The analytic process must be interactive with the customer to succeed. For example, the first IMINT of Soviet missiles during the Cuban Missile Crisis was verified and quickly taken to the President and Secretary of Defense. The highest level of authority immediately requested more detail, but also wanted a perspective on the Soviet strategy, which was not available from photography.

Photographs convey information, not necessarily intentions

As the White House requested more CIA and Navy support for photography, it simultaneously searched for HUMINT and SIGINT from Cuba, as well as diplomatic HUMINT. Until John F. Kennedy was briefed by excellent briefers, such as Dino Brugioni, he probably did not understand the capabilities of IMINT.[12]

Frequently, the intelligence service will organize the production process and its output to mirror the customer organization. Government production by the single-source intelligence agencies is largely organized geographically or topically, to meet the needs of all-source country, region, or topic analysts in the finished-intelligence producing agencies.

In terms of intended use by the customer, both business and government producers may generate intelligence to be applied in the current, estimative, operational, research, science and technology, or warning context. Serendipity plays a role here, because the collected and analyzed information may meet any or all of these criteria.

A good example is warning intelligence.[13] Military and political analysts are always watching for predefined indication that an emergency, such as outbreak of war, or a political coup, is imminent. When an indicator is approved, policymakers are alerted and a crisis team is often convened, with the mission of providing time-sensitive intelligence on the situation to all relevant customers.

Orienting oneself to the consumers

Experienced analysts recommend seeing oneself as a specialist on a team, with 5–10 key players. Learn something about each of them, both in terms of how they express themselves, and how you can reinforce their strengths and support their weaknesses. The analyst must constantly ask himself, "what do they want/need to know? How do they prefer to have it presented? Are they still trying to select the best course of action, or have they committed and now need to know the obstacles and vulnerabilities on their chosen path?"

Others on the team may know, how to handle the likely challenges. The analyst's contribution is in recognizing the unlikely, or providing connections that are not obvious. Consumers must get information in a timely manner, not after they commit to a decision they might not have made having rougher information available sooner.

Sometimes, when the producer is struggling with how to meet the needs of both internal and external customers, the solution is to create two different types of products, one for each type of customer. An internal product might contain detail of sources, collection methods, and analytic techniques, while an external product is more like journalism. Remember that journalists always address:

  1. Who
  2. What
  3. When
  4. Where
  5. Why

"How" is often relevant to journalists, but, in intelligence, may wander into that delicate area of sources and methods, appropriate only for internal audiences. The external consumer needs to know more of potential actions. Actions exist in three phases:

  1. The decision to act
  2. The action
  3. Disengagement from the action[14]

Internal products contain details about the sources and methods used to generate the intelligence, while external products emphasize actionable target information. Similarly, the producer adjusts the product content and tone to the customer's level of expertise.

Orienting oneself to peers

Even in professional sports, where there are strict anti-fraternization rules on the playing field, players often have deep friendships with counterparts on opposing teams. They might have been on a college team together, or are simply aware that the team they oppose today might be the team to which they might be traded tomorrow. If a technique is personal, rather than a proprietary idea of a coach, one professional might be quite willing to show a nominal opponent how he does some maneuver. Watanabe observed

If you are examining a problem and there is no intelligence available, or the available intelligence is insufficient, be aggressive in pursuing collection and in energizing collectors. ... As an analyst, you have the advantage of knowing both what the consumer needs to know (sometimes better than the consumer knows himself) and which collectors can obtain the needed intelligence.

Aggressively pursue collection of information you need. In the Intelligence Community, we have the unique ability to bring substantial collection resources to bear in order to collect information on important issues. An analyst needs to understand the general capabilities and limitations of collection systems...If the analyst is in a technical discipline, the analyst might have an insight about a collection system that the operators have not considered ... If you are not frequently tasking collectors and giving them feedback on their reporting, you are failing to do an important part of your job.[10]

Peers, both consumer and analyst, also have a psychological context. Johnston[15] suggests the three major components of that context are:

  1. socialization
  2. degree of risk taking or risk aversion
  3. organizational-historical context

Devlin[16] observes that while traditional logical work does not consider socialization, work on extending logic into the real world of intelligence requires it. "The first thing to note, and this is crucial, is that the process by which an agent attaches meaning to a symbol always takes place in a context, indeed generally several contexts, and is always dependent on those contexts. An analytic study of the way that people interpret symbols comes down to an investigation of the mechanism captured by the diagram:

[agent] + [symbol] + [context] +. . . + [context] → [interpretation]

Things that are true about contexts include:

  1. Contexts are pervasive
  2. Contexts are primary
  3. Contexts perpetuate
  4. Contexts proliferate
  5. Contexts are potentially pernicious

The discipline of critical discourse analysis will help organize the context. Michael Crichton,[17] in giving examples of physicians communicating with other physicians, points out that laymen have trouble following such discourses not only because there is specialized vocabulary in use, but the discourse takes place in an extremely high context. One physician may ask a question about some diagnostic test, and the other will respond with a result from an apparently unrelated test. The shared context was that the first test looked for evidence of a specific disease, while the answer cited a test result that ruled out the disease. The disease itself was never named, but, in the trained context, perfectly obvious to the participants in the discourse.

Intelligence analysis is also extremely high context. Whether the subject is political behavior or weapons capabilities, the analysts and consumers share a great deal of context. Intelligence consumers express great frustration with generic papers that waste their time by giving them context they already have internalized.

Organizing what you have

Collection processes provide analysts with assorted kinds of information, some important and some irrelevant, some true and some false (with many shades in between), and some requiring further preprocessing before they can be used in analysis. Raw information reports use a standard code for the presumed reliability of the source and of the information. The U.S. Intelligence Community uses some formal definition of the kinds of information.[7]

Term Definition Example
Fact Verified information; something known to exist or to have happened. A confirmed inventory of a resource of one's own service
Direct Information The content of reports, research, and analytic reflection on an intelligence issue that helps analysts and their consumers evaluate the likelihood that something is factual and thereby reduces uncertainty, Information relating to an intelligence issue under scrutiny the details of which can, as a rule, be considered factual, because of the nature of the source, the source's direct access to the information, and the concrete and readily verifiable character of the contents COMINT or OSINT quoting what a foreign official said; IMINT providing a count of the number of ships at a pier. HUMINT from a US diplomatic officer who directly observed an event.
Indirect Information Information relating to an intelligence issue the details of which may or may not be factual, the doubt reflecting some combination of the source's questionable reliability, the source's lack of direct access, and the complex character of the contents HUMINT from a reliable agent, citing secondhand what an informant said that a government official said. OSINT providing a foreign government document that gives the number of ships at a pier. Indirect OSINT from a US embassy officer. COMINT that contains a report by a foreign official to his government, about what something he cannot confirm, but states with a probability.
Direct Data Organized information that provides context for evaluating the likelihood that a matter under scrutiny is factual. A chronology of events based on observations by US officers
Indirect Data Organized information that provides context for evaluating the likelihood that a matter under scrutiny is factual. A chronology based on reports from a liaison intelligence service

Collation describes the process of organizing raw data, interpolating known data, evaluating the value of data, putting in working hypotheses. The simplest approaches often are an excellent start. With due regard for protecting documents and information, a great deal can be done with pieces of paper, a whiteboard, a table, and perhaps a corkboard. Maps often are vital adjuncts, maps that can be written upon.

There are automated equivalents of all of these functions, and each analyst will have a personal balance between manual and machine-assisted methods. Unquestionably, when quantitative methods such as modeling and simulation are appropriate, the analyst will want computer assistance, and possibly consultation from experts in methodology. When combining maps and imagery, especially different kinds of imagery, a geographic information system is usually needed to normalize coordinate systems, scale and magnification, and the ability to suppress certain details and add others.

Outlining, possibly in a word processing program, or using visualization tools such as mind maps can give structure, as can file folders and index cards. Data bases, with statistical techniques such as correlation, factor analysis, and time series analysis can give insight.

Mind-map showing a wide range of nonhierarchical relationships

Some analysts speak of a Zen-like state in which they allow the data to "speak" to them. Others may meditate, or even seek insight in dreams, hoping for an insight such as that given to August Kekulé in a daydream that resolved one of the fundamental structural problems of organic chemistry.

Krizan[5] took criteria from.[18] Regardless of its form or setting, an effective collation method will have the following attributes:

  1. Be impersonal. It should not depend on the memory of one analyst; another person knowledgeable in the subject should be able to carry out the operation.
  2. Not become the "master" of the analyst or an end in itself.
  3. Be free of bias in integrating the information.
  4. Be receptive to new data without extensive alteration of the collating criterion.

Semantic maps are related to mind maps, but are more amenable to computer discovery of relationships.

Semantic network; compare formalism to mind map

The more interactive that the relationship between producer and consumer becomes, the more important will be tools:[19]

  • Collaboration tools. These include all media: voice, video, instant messaging, electronic whiteboards, and shared document markup
  • Databases. Not only will these need to be interoperable, they need to reflect different models, when appropriate, such as the semantic web. There may no longer be a clear line between databases and web applications.
  • Analytic tools. These will cover a wide range of pattern recognition and knowledge organization.

The nature of analysis

An analysis should have a summary of the key characteristics of the topic, followed by the key variables and choices. Increasingly deep analysis can explain the internal dynamics of the matter being studied, and eventually to prediction, known as estimation.

The purpose of intelligence analysis is to reveal to a specific decision maker the underlying significance of selected target information. Analysts should begin with confirmed facts, apply expert knowledge to produce plausible but less certain findings, and even forecast, when the forecast is appropriately qualified. Analysts should not, however, engage in fortunetelling that has no basis in fact.

Food chain in intelligence analysis: the bigger the "fish", the more unlikely it is

The mnemonic "Four Fs Minus One" may serve as a reminder of how to apply this criterion. Whenever the intelligence information allows, and the customer's validated needs demand it, the intelligence analyst will extend the thought process as far along the Food Chain as possible, to the third "F" but not beyond to the fourth.

Types of reasoning

Objectivity is the intelligence analyst's primary asset in creating intelligence that meets the Four Fs Minus One criterion. To produce intelligence objectively, the analyst must employ a process tailored to the nature of the problem. Four basic types of reasoning apply to intelligence analysis: induction, deduction, abduction and the scientific method.[5]

Induction: seeking causality

The induction process is one of discovering relationships among the phenomena under study. It may come from human pattern recognition ability, looking at a seemingly random set of events, perhaps writing them on cards and shuffling them until a pattern emerges.[5]

An analyst might notice that when Country X's command post with call sign ABC sent out a message on frequency 1 between Thursday and Saturday, an air unit will move to a training range within one week. The acknowledgement will take one day, so the analyst should recommend intensified COMINT monitoring of the appropriate frequencies between Friday and Sunday. Another kind of causality could come from interviews, in which soldiers might describe the things that warn them of an impending attack, or how the ground might look when an improvised explosive device has been emplaced.

While induction, for human beings, is usually not at a fully rational level, do not discount the potential role of software that uses statistical or logical techniques for finding patterns. Induction is subtly different from intuition: there usually is a pattern that induction recognizes, and this pattern may be applicable to other situations.

Deduction: applying the general

Deduction, is the classic process of reasoning from the general to the specific, a process made memorable by Sherlock Holmes: "How often have I said to you that when you have eliminated the impossible, whatever remains, however improbable, must be the truth?" Deduction can be used to validate a hypothesis by working from premises to conclusion.[5]

The pattern of air maneuvers described above may be a general pattern, or it may be purely General X's personal command style. Analysts need to look at variables, such as personalities, to learn whether a pattern is truly general doctrine, or simply idiosyncratic.

Not all intelligence officers regard this as a desirable approach. At his confirmation hearing for CIA Director, Gen. Michael V. Hayden said he believes that intelligence analysis should be done by "induction", under which "all the data" are gathered and general conclusions determined, rather than by "deduction", under which you have a conclusion and seek out the data that support it.[20]

Trained intuition

Analysts need to harness trained intuition: the recognition that one has come to a spontaneous insight. The steps leading there may not be apparent, although it is well to validate the intuition with the facts and tools that are available.[5]

Polish cryptanalysts first were reading German Enigma ciphers in 1932, although the commercial version may have been broken by the British cryptanalyst, Dilwyn Knox, in the 1920s. Poland gave critical information to the French and British in 1939, and production British cryptanalysis was well underway in 1940. The Enigma, with German military enhancements, was quite powerful for a mechanical encryption device, and it might not have been broken as easily had the Germans been more careful about operating procedures. Throughout the war, Germany introduced enhancements, but never realized the British were reading the traffic almost as fast as the Germans.

Ultimately, no code is unbreakable, including Enigma's, if security is compromised

US cryptanalysts had broken several Japanese diplomatic ciphers, but, without ever seeing the PURPLE machine until after the war, they deduced the logic. Purple was actually mechanically simpler than Enigma, but the U.S. Army team struggled with a mechanical reproduction until Leo Rosen had the unexplained insight that the critical building block in the Purple machine was a telephone-type stepping switch rather than the rotor used in Enigma and in more advanced U.S. and UK machines. Rosen, Frank Rowlett, and others of the team recognized Rosen's insight as based on nothing but a communication engineer's intuition.

Experienced analysts, and sometimes less experienced ones, will have an intuition about some improbable event in a target country, and will collect more data, and perhaps send out collection requests within his or her authority. These intuitions are useful just often enough that wise managers of analysts, unless the situation is absolutely critical, allow them a certain amount of freedom to explore.

Scientific method

Astronomers and nuclear physicists, at different ends of the continuum from macroscopic to microscopic, share the method of having to infer behavior, consistent with hypothesis, not by measuring phenomena to which they have no direct access, but by measuring phenomena that can be measured and that hypothesis suggests will be affected by the mechanism of interest. Other scientists may be able to set up direct experiments, as in chemistry or biology. If the experimental results match the expected outcome, then the hypothesis is validated; if not, then the analyst must develop a new hypothesis and appropriate experimental methods.[5]

In intelligence analysis, the analyst rarely has direct access to the observable subject, but gathers information indirectly. Even when the intelligence subject at hand is a technical one, analysts must remain aware that the other side may be presenting deliberately deceptive information.

From these gathered data, the analyst may proceed with the scientific method by generating tentative explanations for a subject event or phenomenon. Next, each hypothesis is examined for plausibility and compared against newly acquired information, in a continual process toward reaching a conclusion. Often the intelligence analyst tests several hypotheses at the same time, whereas the scientist usually focuses on one at a time. Furthermore, intelligence analysts cannot usually experiment directly upon the subject matter as in science, but must generate fictional scenarios and rigorously test them through methods of analysis suggested below.

Methods of analysis

As opposed to types of reasoning, which are ways the analyst drafts the product, the following methods are ways of validating the analyst's results of reasoning. Structured analytic techniques are used to help challenge judgments, identify mental mindsets, overcome biases, stimulate creativity, and manage uncertainty. Examples include the key assumptions check, analysis of competing hypotheses, Devil's advocacy, Red Team Analysis, and Alternative Futures/Scenarios analysis, among others.[21]

Opportunity analysis

Opportunity analysis identifies for policy officials opportunities or vulnerabilities that the customer's organization can exploit to advance a policy, as well as dangers that could undermine a policy. Lawyers apply the test cui bono (who benefits?) in a rather similar way.

To make the best use of opportunity analysis, there needs to be a set of objectives for one's own country, preferably with some flexibility to them. The next step is to examine personalities and groups in that target country to see if there are any with a commonality of interest. Even though the different sides might want the same thing, it is entirely possible that one or the other might have deal-breaking conditions. If that is the case, then ways to smooth that conflict need to be identified, or no more work should be spent on that alternative.

Conversely, if there are elements that would be utterly opposed to the objectives of one's side, ways of neutralizing those elements need to be explored. They may have vulnerabilities that could render them impotent, or there may be a reward, not a shared opportunity, that would make them cooperate.

Linchpin analysis

Linchpin analysis proceeds from information that is certain, or with a high probability of being certain. In mathematics and physics, a similar problem formation, which constrains the solution by certain known or impossible conditions, is the boundary value condition.

By starting from knowns (and impossibilities), the analyst has a powerful technique for showing consumers, peers, and managers that a problem has both been thoroughly studied and constrained to reality.[22] Linchpin analysis was introduced to CIA by deputy director for Intelligence (1993–1996) Doug MacEachin, as one of the "muscular" terms he pressed as an alternative to academic language, which was unpopular with many analysts. He substituted linchpin analysis for the hypotheses driving key variables. MacEachin required the hypotheses—or linchpins—needed to be explicit, so policymakers could be aware of coverage, and also aware of changes in assumptions.

This method is an "anchoring tool" that seeks to reduce the hazard of self-inflicted intelligence error as well as policymaker misinterpretation. It forces use of the checkpoints listed below, to be used when drafting reports:

  1. Identify the main uncertain factors or key variables judged likely to drive the outcome of the issue, forcing systematic attention to the range of and relationships among factors at play.
  2. Determine the linchpin premises or working assumptions about the drivers. This encourages testing of the key subordinate judgments that hold the estimative conclusion together.
  3. Marshal findings and reasoning in defense of the linchpins, as the premises that warrant the conclusion are subject to debate as well as error.
  4. Address the circumstances under which unexpected developments could occur. What indicators or patterns of development could emerge to signal that the linchpins were unreliable? And what triggers or dramatic internal and external events could reverse the expected momentum?

Analysis of competing hypotheses

Dick Heuer spent years in the CIA Directorate of Operations (DO) as well as the DI, and worked on methodology of analysis both in his later years and after retirement.[3] Some of his key conclusions, coming from both experience and an academic background in philosophy, include:

  1. The mind is poorly "wired" to deal effectively with both inherent uncertainty (the natural fog surrounding complex, indeterminate intelligence issues) and induced uncertainty (the man-made fog fabricated by denial and deception operations).
  2. Even increased awareness of cognitive and other "unmotivated" biases, such as the tendency to see information confirming an already-held judgment more vividly than one sees "disconfirming" information, does little by itself to help analysts deal effectively with uncertainty.
  3. Tools and techniques that gear the analyst's mind to apply higher levels of critical thinking can substantially improve analysis on complex issues on which information is incomplete, ambiguous, and often deliberately distorted. Key examples of such intellectual devices include techniques for structuring information, challenging assumptions, and exploring alternative interpretations.

In 1980, he wrote an article, "Perception: Why Can't We See What Is There to be Seen?" which suggests to Davis[22] that Heuer's ideas were compatible with linchpin analysis. Given the difficulties inherent in the human processing of complex information, a prudent management system should

  1. Encourage products that (a) clearly delineate their assumptions and chains of inference and (b) specify the degree and source of the uncertainty involved in the conclusions.
  2. Emphasize procedures that expose and elaborate alternative points of view—analytic debates, devil's advocates, interdisciplinary brainstorming, competitive analysis, intra-office peer review of production, and elicitation of outside expertise.

According to Heuer, analysts construct a reality based on objective information, filtered through complex mental processes that determine which information is attended to, how it is organized, and the meaning attributed to it. What people perceive, how readily they perceive it, and how they process this information after receiving it are all strongly influenced by past experience, education, cultural values, role requirements, and organizational norms, as well as by the specifics of the information received. To understand how the analysis results, one must use good mental models to create the work, and understand the models when evaluating it. Analysts need to be comfortable with challenge, refinement, and challenge. To go back to linchpin analysis, the boundary conditions give places to challenge and test, reducing ambiguity.

More challenge, according to Heuer, is more important than more information. He wanted better analysis to be applied to less information, rather than the reverse. Given the immense volumes of information that modern collection systems produce, the mind is the limiting factor. Mirror-imaging is one of Heuer's favorite example of a cognitive trap, in which the analyst substitutes his own mindset for that of the target. "To see the options faced by foreign leaders as these leaders see them", according to Heuer, " one must understand [the foreign leaders'] values and assumptions and even their misperceptions and misunderstandings. ... Too frequently, foreign behavior appears "irrational" or "not in their own best interest." Projecting American values created models that were inappropriate for the foreign leader.

A significant problem during the Vietnam War is that Secretary of Defense Robert S. McNamara, an expert on statistical decision-making, assumed that Ho Chi Minh, Võ Nguyên Giáp, and other North Vietnamese officials would approach decision-making as he did. For example, in McNamara's thinking, if the United States did not attack SA-2 anti-aircraft missiles, the enemy would interpret that as "restraint" and not use them against U.S. aircraft [23] The North Vietnamese leadership, not privy to McNamara's thinking, were unaware of the "signaling" and did their best to shoot down U.S. aircraft with those missiles.

Heuer's answer was making the challenge of Analysis of Competing Hypotheses (ACH) the core of analysis. In ACH, there is competition among competing hypotheses of the foreign leader's assumptions, which will reduce mirror-imaging even if they do not produce the precise answer. The best use of information, in this context, is to challenge the assumption the analyst likes best.

One of the key motivations for ACH, according to Heuer, is to avoid rejecting deception out of hand, because the situation looks straightforward. Heuer observed that good deception looks real. "Rejecting a plausible but unproven hypothesis too early tends to bias the subsequent analysis, because one does not then look for the evidence that might support it. The possibility of deception should not be rejected until it is disproved or, at least, until a systematic search for evidence has been made and none has been found."

The steps in ACH are:[24]

  1. Identify the possible hypotheses to be considered. Use a group of analysts with different perspectives to brainstorm the possibilities.
  2. Make a list of significant evidence and arguments for and against each hypothesis.
  3. Prepare a matrix with hypotheses across the top and evidence down the side. Analyze the "diagnosticity" of the evidence and arguments—that is, identify which items are most helpful in judging the relative likelihood of the hypotheses.
  4. Refine the matrix. Reconsider the hypotheses and delete evidence and arguments that have no diagnostic value.
  5. Draw tentative conclusions about the relative likelihood of each hypothesis. Proceed by trying to disprove the hypotheses rather than prove them.
  6. Analyze how sensitive your conclusion is to a few critical items of evidence. Consider the consequences for your analysis if that evidence were wrong, misleading, or subject to a different interpretation.
  7. Report conclusions. Discuss the relative likelihood of all the hypotheses, not just the most likely one.
  8. Identify milestones for future observation that may indicate events are taking a different course than expected.

Keith Devlin has been researching the use of mathematics and formal logic in implementing Heuer's ACH paradigm.[16]

Analogy

Analogy is common in technical analysis, but engineering characteristics seeming alike do not necessarily mean that the other side has the same employment doctrine for an otherwise similar thing. Sometimes, the analogy was valid for a time, such as the MiG-25 aircraft being designed as a Soviet counter to the perceived threat of the high-altitude, supersonic B-70 bomber. The Soviets could have canceled the MiG-25 program when the US changed doctrines to low altitude penetration and canceled the B-70 program, but they continued building the MiG-25.

One of the Soviet variants was a high-speed, high-altitude reconnaissance aircraft (MiG-25RB), which, for a time, was thought comparable to the US SR-71 aircraft. Several additional points of data, however, showed that an analogy between the SR-71 and MiG-25RB was not complete. HUMINT revealed that a single Mach 3.2 flight of the MiG wrecked the engines beyond hope of repair, and the cost of replacement was prohibitive. The SR-71, however, could make repeated flights with the same engines. The dissimilarity of engine life was not only expensive, but meant that the MiG-25RB could operate only from bases with the capability to change engines.

Top speed MiG-25RB reconnaissance flights damaged its engines beyond repair

The United States had applied "reverse engineering" to the MiG, essentially saying "if we had an aircraft with such capabilities, what would we do with it?" In the fighter-interceptor role, however, the US gives the pilot considerable flexibility in tactics, where the Soviets had a doctrine of tight ground control. For the U.S. doctrine, the aircraft was too inflexible for American fighter tactics, but made sense for the Soviets as an interceptor that could make one pass at a penetrating bomber, using an extremely powerful radar to burn through jamming for final targeting.

Many of these assumptions fell apart after Viktor Belenko flew his MiG-25 to the West, where TECHINT analysts could examine the aircraft, and doctrinal specialists could interview Belenko.[25]

The analytic process

Analysts should follow a series of sequential steps:

Define the problem

Policy makers will have questions based on their intelligence requirements. Sometimes questions are clear and can easily be addressed by the analyst. Sometimes however, clarification is required due to vagueness, multiple layers of bureaucracy between customer and analyst, or due to time constraints. Just as analysts need to try to understand the thinking of the adversary, analysts need to know the thinking of their customers and allies.

Generate hypotheses

Once the problem is defined, the analyst is able to generate reasonable hypotheses based on the question. For example, a business may want to know whether a competitor will lower their prices in the next quarter. From this problem, two obvious hypotheses are:

  1. The competitor will lower prices or
  2. The competitor will not lower prices.

However, with a little brainstorming, additional hypotheses may become apparent. Perhaps the competitor will offer discounts to long term customers, or perhaps they may even raise prices. At this point, no hypothesis should be discarded.

Determine information needs and gather information

In intelligence, collection usually refers to the step in the formal intelligence cycle process. In many cases, the information needed by the analyst is either already available or is already being sought by collection assets (such as spies, imagery satellites). If not, the analyst may request collection on the subject, or if this is not possible identify this information gap in their final product. The analyst will generally also research other sources of info, such as open source (public record, press reporting), historical records, and various databases.

Evaluate sources

Information used for military, commercial, state, and other forms of intelligence analysis has often been obtained from individuals or organizations that are actively seeking to keep it secret, or may provide misleading information. Adversaries do not want to be analyzed correctly by competitors. This withholding of information is known as counterintelligence, and is very different from similar fields of research, such as science and history where information may be misleading, incomplete or wrong, but rarely does the subject of investigation actively deny the researcher access. So, the analyst must evaluate incoming information for reliability (has the source reported accurate information in the past?), credibility (does the source reasonably have access to the information claimed? Has the source lied in the past?), and for possible denial and deception (even if the source is credible and reliable, they may have been fooled).

Evaluate (test) hypotheses

All hypotheses must be rigorously tested. Methods such as Analysis of Competing Hypotheses or link charts are key.[26] It is essential to triage which may be valid, which fail readily, and which require more information to assess.

Be especially alert to cognitive and cultural biases in and out of the organization. Recent scholarship on theories of the sociology of knowledge raise important caveats.

As Jones and Silberzahn documented in the 2013 volume Constructing Cassandra: Reframing Intelligence Failure at the CIA, 1947–2001, while hypotheses are essential to sorting "signals" from "noise" in raw intelligence data, the variety, types and boundaries of the types of hypotheses an intelligence organization entertains are a function of the collective culture and identity of the intelligence producer. Often, these hypotheses are shaped not merely by the cognitive biases of individual analysts, but by complex social mechanism both inside and outside that analytic unit. After many strategic surprises, "Cassandras" – analysts or outsiders who offered warnings, but whose hypotheses were ignored or sidelined – are discovered. Therefore, careful analysts should recognize the key role that their own and their organization's identity and culture play in accepting or rejecting hypotheses at each step in their analysis.[27]

Production and packaging

Once hypotheses have been evaluated, the intelligence product must be created for the consumer. Three key features of the intelligence product are:

  • Timeliness. Timeliness includes not only the amount of time required to deliver the product, but also the usefulness of the product to the customer at a given moment.
  • Scope. Scope involves the level of detail or comprehensiveness of the material contained in the product.
  • Periodicity. Periodicity describes the schedule of product initiation and generation.

Government intelligence products are typically packaged as highly structured written and oral presentations, including electrical messages, hardcopy reports, and briefings. Many organizations also generate video intelligence products, especially in the form of live daily "newscasts", or canned documentary presentations.

Analysts should understand the relationship between the analyst's and the consumer's organization. There may be times that while the ultimate consumer and originating analyst simply want to pass information, a manager in either chain of command may insist on a polished format.

Peer review

Peer review is essential to assess and confirm accuracy. "Coordination with peers is necessary...If you think you are right, and the coordinator disagrees, let the assessment reflect that difference of opinion and use a footnote, called a reclama,[28] inside the U.S. intelligence community if necessary. But never water down your assessment to a lowest common denominator just to obtain coordination. When everyone agrees on an issue, something probably is wrong. "As an example, following the collapse of the Soviet Union, there was an almost unanimous belief that large numbers of Russian ballistic missile specialists would flood into the Third World and aid missile programs in other states (the so-called brain drain)...As it turned out, there was no [expected] mass departure of Russian missile specialists, but Russian expertise was supplied to other states in ways that had been ignored due to the overemphasis on the brain drain.

In large intelligence establishments, analysts have peers at other agencies. The practical amount of coordination, indeed inside one's own agency, will depend on the secure collaboration tools available (wikis, analyst webpages, email), the schedule and availability of the other analysts, any restrictions on dissemination of the material, and the analyst's ability to play nicely with others. Extremely specialized issues might have very few people who could meaningfully look at it.

An intelligence community document, as opposed to a spot report from a single agency, is expected to be coordinated and reviewed. For example, in reports on the Iraqi WMD program, given a field report that aluminum tubes were on order, which might have been received both at the geographic desk and the Counterproliferation Center, someone might have thought they were for use in uranium separation centrifuges. It has been reported that some analysts thought they might be used for rocket casings, which apparently was the correct interpretation. The question needs to be asked "did the original analyst contact a technical specialist in separation centrifuges, perhaps at Department of Energy intelligence?"

Such an analyst might have mentioned that while aluminum has been used, maraging steel is the material of choice for Zippe-type centrifuges. The alternative, the Helikon vortex separation process, has no moving parts and thus less demand on the tubes, but takes much more energy. If the Helikon had been under consideration, the consultation could have gone farther, perhaps to IMINT analysts familiar with power generation in the area or infrared MASINT specialists who could look for the thermal signature of power generation or the cascade itself. Both Zippe and Helikon techniques take a great deal of energy, and often have been placed near hydroelectric dam power plants so power will be nearby.

Customer feedback and production evaluation

The production phase of the intelligence process does not end with delivering the product to the customer. Rather, it continues in the same manner in which it began: with interaction between producer and customer. For the product to be useful, the analyst and policymaker need to hear feedback from one another, and they refine both analysis and requirements.

Feedback procedures between producers and customers includes key questions, such as: Is the product usable? Is it timely? Was it in fact used? Did the product meet expectations? If not, why not? What next? The answers to these questions lead to refined production, greater use of intelligence by decision makers, and further feedback sessions. Thus, production of intelligence generates more requirements in this iterative process.

Never forget the end user

Effective intelligence analysis is ultimately tailored to the end user. William Donovan, the head of the World War II OSS, began to get FDR's ear because he gave vividly illustrated, well-organized briefings that would be common today, but were unprecedented in World War II. Today, there is danger of becoming too entranced with the presentation and less with its subject. This is also a delicate dance of overemphasizing the subjects that interest high officials, and what they want to hear declared true about them, rather than hearing what the analysts believe is essential.

Most consumers do not care how attractive a report looks or whether the format is correct. I have lost count of the number of times consumers have told me they do not care if an assessment has a CIA seal on it, if it is in the proper format, or even if it has draft stamped all over it; they just want the assessment in their hands as soon as possible, at least in time to help make a decision. Unfortunately, a number of mid-level managers get overly worried about form, and wise top-level intelligence officials make sure that does not happen.[10]

At the same time, analysts must always be wary of mirroring the desires, attitudes, and views of intelligence consumers. They must raise awkward facts and ask probing questions, even if this makes the decision-maker's job harder.[29]

See also

References

  1. ^ Hayes, Joseph (2007), "Analytic Culture in the U.S. Intelligence Community. Chapter One. Working Definitions.", History Staff, Center for the Study of Intelligence, Central Intelligence Agency, archived from the original on June 13, 2007, retrieved 2016-05-24
  2. ^ Lee, Bartholomew (2006), Radio Spies – Episodes in the Ether Wars (PDF), archived from the original (PDF) on 2018-10-06, retrieved 2007-11-29
  3. ^ a b c Heuer, Richards J. Jr. (1999), "Psychology of Intelligence Analysis. Chapter 2. Perception: Why Can't We See What Is There To Be Seen?", History Staff, Center for the Study of Intelligence, Central Intelligence Agency, archived from the original on July 15, 2001, retrieved 2007-10-29
  4. ^ North Atlantic Treaty Organization (November 2001), NATO Open Source Intelligence Handbook (PDF), archived from the original (PDF) on 2020-11-07, retrieved 2007-10-23
  5. ^ a b c d e f g Krizan, Lisa (June 1999), Intelligence Essentials for Everyone, Joint Military Intelligence College, archived from the original on 2009-05-24, retrieved 2009-05-24
  6. ^ Alfred Rolington. Strategic Intelligence for the 21st Century: The Mosaic Method, Oxford University Press, 2013.
  7. ^ a b c Central Intelligence Agency, Directorate of Intelligence (February 1997), A Compendium of Analytic Tradecraft Notes, archived from the original on April 4, 2004, retrieved 2007-12-03
  8. ^ "Objectivity". www.dni.gov. Retrieved 2022-12-15.
  9. ^ Davis, Jack (1995), "A Policymaker's Perspective On Intelligence Analysis", Studies in Intelligence, 38 (5), archived from the original on June 13, 2007, retrieved 2007-10-28
  10. ^ a b c d Watanabe, Frank (1997), "Fifteen Axioms for Intelligence Analysts: How To Succeed in the DI [Directorate of Intelligence]", Studies in Intelligence, retrieved 2007-10-23
  11. ^ Kalisch, Robert B. (July–August 1971), "Air Force Technical Intelligence", Air University Review, retrieved 2007-10-27
  12. ^ May, Ernest R.; Zelikow, Philip D. (1996), The Kennedy Tapes: Inside the White House During the Cuban Missile Crisis, retrieved 2007-10-23
  13. ^ Morgan, Brent A. (September 1995), Employment of Indications and Warning Intelligence Methods to Forecast a Potentially Hostile Revolution in Military Affairs, US Naval Postgraduate School, archived from the original on April 8, 2013
  14. ^ Ikle, Fred (2005), Every War Must End, Columbia University Press
  15. ^ Johnston, Rob (2005). "Analytic Culture in the US Intelligence Community: An Ethnographic Study". Center for the Study of Intelligence, Central Intelligence Agency. Archived from the original on June 13, 2007. Retrieved 2007-10-29.
  16. ^ a b Devlin, Keith (July 15, 2005). "Confronting context effects in intelligence analysis: How can mathematics help?" (PDF). Center for the Study of Language and Information, Stanford University. Retrieved 2007-10-29.
  17. ^ Crichton, Michael (1970). Five Patients. Ballantine Books. ISBN 0-345-35464-8.
  18. ^ Mathams, Robert (1995), "The Intelligence Analyst's Notebook", in Douglas Dearth; R. Thomas Goodden (eds.), Strategic Intelligence: Theory and Application, Joint Military Intelligence Training Center
  19. ^ National Intelligence Production Board (2001), Strategic Investment Plan for Intelligence Community Analysis, NIPB-2001, archived from the original on June 13, 2007, retrieved 2007-10-28
  20. ^ Priest, Dana; Pincus, Walter (May 19, 2006), "Nominee Has Ability To Bear Bad News: Some Senators Unsure He Will Use It With Bush", Washington Post, retrieved 2007-10-28
  21. ^ CIA-A Tradecraft Primer:Structured Analytic Techniques for Improving Intelligence Analysis-March 2009
  22. ^ a b Davis, Jack (1999), "Improving Intelligence Analysis at CIA: Dick Heuer's Contribution to Intelligence Analysis", Psychology of Intelligence Analysis, Center for the Study of Analysis, Central Intelligence Agency, Davis 1999, archived from the original on July 15, 2001, retrieved 2007-10-27
  23. ^ McMaster, H. R. (1998), Dereliction of Duty: Johnson, McNamara, the Joint Chiefs of Staff, and the Lies That Led to Vietnam, Harper Perennial, ISBN 0-06-092908-1
  24. ^ Heuer, Richards J. Jr. (1999), "Psychology of Intelligence Analysis. Chapter 8: Analysis of Competing Hypotheses", History Staff, Center for the Study of Intelligence, Central Intelligence Agency, archived from the original on July 15, 2001, retrieved 2007-10-28
  25. ^ Barron, John (1983), Mig Pilot: The Final Escape of Lt. Belenko, Avon Books, ISBN 0-380-53868-7
  26. ^ "3.21.00 Link Chart". Archived from the original on 2007-11-21. Retrieved 2007-11-12.
  27. ^ Jones, Milo L. and; Silberzahn, Philippe (2013). Constructing Cassandra, Reframing Intelligence Failure at the CIA, 1947-2001. Stanford University Press. ISBN 978-0804793360.
  28. ^ US Department of Defense (12 July 2007), Joint Publication 1-02 Department of Defense Dictionary of Military and Associated Terms (PDF), archived from the original (PDF) on 23 November 2008, retrieved 2007-10-01
  29. ^ Pages 234-252 Jones, Milo L. and; Silberzahn, Philippe (2013). Constructing Cassandra, Reframing Intelligence Failure at the CIA, 1947-2001. Stanford University Press. ISBN 978-0804793360.

Further reading