In Proceeding of the Third Information Survivability Workshop, Boston, Massachusetts, October 24-26, 2000. Sponsored by the IEEE Computer Society and the US State Department. Organized by the CERT Coordination Center, Software Engineering Institute.
© 2000 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.
This material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders. All persons copying this information are expected to adhere to the terms and constraints invoked by each author's copyright. In most cases, these works may not be reposted without the explicit permission of the copyright holder.

Trapping and Tracking Hackers:
Collective security for survival in the Internet age

Douglas B. Moran
Vice President, Research & Development
Recourse Technologies, Inc
2450 El Camino Real, Suite 100, Palo Alto, CA 94306
http://www.recourse.com/

Introduction

Why are Internet attacks so difficult to deal with? A significant factor is that the coverage of the current set of computer security tools is a poor match to the threat, analogous to a fence with multiple gaps. Most current tools fall into two broad categories:
  1. Prevention: configuration checking, identification-and-authentication, other encryption-based approaches, ...
  2. Limited response to individual instances of known vulnerabilities and attacks: firewalls, intrusion detection systems, ...
Problems arise because these tools are too often expected to provide broader coverage than their designs permit. The limitations of "prevention" tools are that (1) there is typically a large gap in time between the identification of the vulnerability and the availability of the solution, and (2) partial and faulty deployment resulting in gaps in coverage. Tools for response to individual attacks typically have significant false negative rates (e.g., tunneling through firewalls) or high false positive rates (intrusion detection systems). With these tools, there is also a gap between the onset of new types of attack and when signatures/rules become available to handle them.

"I don't need to outrun the bear, I just need to outrun you." This punchline reflects the philosophy inherent in many sites' security measures. Increasing the difficulty of breaking into your site can be effective in persuading some attackers to shift their efforts to some other site. Notice that this is most effective against attackers who would constitute a mere annoyance, for example, script kiddies who are looking for sites to break into, without much regard for which ones. This approach is least effective against the more serious threats: attackers specifically targeted on your site and insiders.

This defensive approach blocks a wide range of known attacks, but gives you little leverage for preventing future attacks. Military doctrine has long recognized the futility of a purely defensive strategy—the aggressor, given the twin advantages of initiative and time, will inevitably find an exploitable weak spot. However, the military analogy's remedy of offensive actions must be adapted for the Internet because simple offensive actions are typically impractical, immoral and/or illegal [1]. Instead, the remedy is to deprive the attacker of the initiative and the time to stage further attackers. To achieve this, computer security tools must better leverage information about attacks and attackers to better protect the larger cluster, the enterprise, and the larger Internet.

Handling new attacks is a major critical gap. Our approach to this is based on the philosophy of time-based security [2] and it has two components:

  1. Intelligence gathering: collecting information about the attacker, his intentions, and his capabilities
  2. Delaying action: This has multiple purposes:
    • Slowing down the attack can limit the damage done. Many attacks indicate an awareness by the perpetrator that risk of detection rises disproportionately with time: they are heavily automated and increasingly segmented into separate sessions. Causing the attacker to waste his time on low-value targets reduces or eliminates the time available for attacking high-value targets.
    • Delaying the onset of the attack against the high-value targets allows the defender to use that time to strengthen the defenses of those systems and to use the intelligence gathered during the early stages of the attack to prioritize the measures taken.
    • Allowing time to perform additional intelligence gathering, especially tracking the attacker back through a chain of relay/cutout hosts, enhances the breadth and depth of the response.

A secondary benefit of this approach is that it not only helps fill the gap in the metaphorical fence, but makes the gap narrower: the improved intelligence gathering allows other security tools (e.g., configuration checkers, IDSes) to be upgraded to handle new attacks earlier.

Currently, collecting information about attacks has multiple impediments:

  1. The rate of detection of attacks is low [3].
  2. The rate of reporting detected attacks is low [4].
  3. The quality of the data collected is typically poor [5]. The combination of automated attacks, delayed detection and delayed response almost always means that the attacker has plenty of opportunity to cleanup after himself. Additional evidence is lost to normal operations of the computer and its legitimate users.
  4. The quality of the reports of the data collected tends to be poor [6]. Even when expertise is available, the cost of collecting the data and assembling it into a report exceeds the available resources of the invariably overloaded system administrators.
  5. Substantial delays between the reporting of an attack and the receipt of countermeasures destroys any sense of urgency or importance.
  6. The cost and delays to sanitize the data in a report to be submitted to an outside agency is often prohibitive.
  7. Reporting that you have been successfully hacked can be not only embarrassing, but may involve penalties (e.g.,loss of investor confidence).

Deception Servers and Deception Hosts

There are some existing technologies to help fill this gap, but they are little used: Deception servers emulate a range of network servers (e.g., Telnet, FTP, SMTP), often for a range of platforms, but do not provide actual access to the underlying host. Most allow the operator to provide a dummy version of some likely targets (e.g., the password file). This is very useful for discovering details of new attacks in which the opening gambit is the key move, but because they cut off access immediately after that, they give negligible information about the attacker's intentions. And with the increasingly rapid dissemination of new exploits from elite hackers to the script kiddies, these tools often fail to provide information about the skill level of attackers (important when you want to focus your efforts on the more dangerous ones).

Deception hosts have a long history of use in computer intrusions, but most have been ad hoc. They typically have been created and deployed in the midst of a serious incident. Long-term use of a deception host presents challenges not seen during an ongoing attack. ManTrapTM is a system for creating and managing a collection of deception hosts that reside permanently on a network.

ManTrapTM Key Features

Monitoring: The typical attacker bolts at the slightest sign that he is being monitored, so the monitoring and management of the deception host must be virtually undetectable. Ad hoc deception hosts are typically monitored with packet sniffers, which work well if the attack is conducted primarily with interactive commands. However, for automated attacks, a packet sniffer does not show what is happening on the system—the operator is forced to analyze the attack scripts and binaries, or, more dangerously, run them on a comparable system and monitor the effects. For many networks, packet sniffing is an unacceptable long-term monitoring tool because data captured presents both storage and confidentiality problems.

With ManTrap, the attacker is actually running inside a cage that contains a virtual environment that he can explore and change. The monitoring processes run outside the cage, both hiding and protecting them from the attacker. Multiple monitoring processes are running, providing the operator with multiple views of the attackers activities (e.g., packet logging, process invocation).

These logs facilitate creating reports on an attack. Additional tools for summarizing the log contents are planned. ManTrap has minimal operator costs: it has a negligible false positive rate, and the operator sets a threshold for being notified.

Containment: The fundamental concept of a deception host is that you want the attacker to have apparent full access to the host so that you can study him. However, this brings with it the danger of him using the deception host as a base to attack other hosts, both inside and outside your network. During an ongoing attack, monitoring the deception host can be handled manually—it represents a small addition to the monitoring that typically is already being done. And since you are actively monitoring events, you are able to respond quickly. With ManTrap, containment is provided by a combination of the cage and its placement in the network, with multiple alternatives available. Different choices provide trade-offs between setup costs, the realism of the network environment, and how quickly the operator needs to begin active monitoring of an attack.

Convincing content: Not only must a deception host not scare off the attacker, it must induce him to stay connected and continue to perform actions that provide useful information to you. Ad hoc deception hosts typically use real (live) systems that have already been compromised, allowing some additional controlled compromises for the opportunity to prevent uncontrolled compromises.

This is not acceptable for most deployments of permanent deception hosts. ManTrap includes a Content Generation Module that probabilistically selects and instantiates templates using a combination of generic and locally-specified values. This generated data is not intended to withstand careful examination–the attacker is unlikely to expend valuable connect time on this. Instead, this data is intended to provide a convincing volume of false hits for the types of searches that attackers often perform. This fabricated content is typically supplemented with non-confidential real content, for example, the Web pages from the public Web server.

Because the contents of a ManTrap has been "pre-sanitized," this removes one of the substantial impediment to reporting (above). Also, since an attack captured by a ManTrap represents a success of your security measures, reporting an attack has lost the stigma of failure, thereby eliminating another of the impediments.

Variability of content: As ManTrap becomes more widely deployed, attackers will attempt to detect whether they have connected to one. To make detection more difficult, ManTrap probabilistically introduces variability, both in the configuration of the platform and in the text generated from the templates. Variations in the text include modifying line breaks and introducing common typos and misspelling.

Faithful representation of the platform: ManTrap does not emulate a platform, but instead implements the cage with a small set of modifications to the kernel. With this approach, the attacker sees the expected capabilities and idiosyncrasies of the target platform. The approach works well for an operating system that has unified APIs for the kernel (e.g., Solaris), but for OSes that have multiple APIs to the same kernel functionality (e.g., Linux), the work required to provide a consistent deception is greatly increased.

Resetting the trap: It is common for a serious hacker to "own" hundreds of hosts at one time, and thus it is not uncommon for significant time to elapse between the initial compromise of a host and when it is used to stage attacks on other hosts. An effective deception host needs to leave the attacker's downloads on the system for use in later visits. However, if the deception host is being used only for uninteresting activity, such as a "Warez" server (for pirated software), the operator will usually want to erase all changes (by resetting the deception host) and wait for a more interesting (threatening) attack. With ManTrap, the contents of the cage is generated from data stored outside the cage, and thus is trivially regenerated.

Deployment Schemes

Minefield: Currently, the most common scheme for deploying ManTrap systems is to place them among likely high-value targets. One site is deploying one ManTrap for every three of its Web servers. Legitimate requests are distributed among only the real Web servers, but someone attempting to gain access to the Web servers via non-httpd means has a 25% chance of selecting a ManTrap system as his first target, and since the typical attacker will attempt to "own" all the Web servers in the cluster, he will quickly wind up in a ManTrap, revealing both his presence and his actions.

Shield: Attacks against high-value targets are actively redirected into the ManTrap by a firewall or router. For example, an attempt to connect to the public Web server through an un-permitted service (e.g., SMTP) is automatically redirected (via Network Address Translation) to a ManTrap whose contents mirror the Web server but that sits in a separate DMZ (for containment).

Zoo: This is an all-deception-host network, with the deception hosts configured with different applications just as you would expect for the normal mix of servers and workstations, and with different mixes of vulnerabilities. The purpose of such networks is to encourage the attacker to exercise more of his "bag of tricks" and to reveal his skill level and intentions, thereby allowing efforts to be focused on the more serious threats. Although such networks are under consideration, we are currently not aware of any operational deployments.

ManHuntTM

A common practice of attackers is to use a series of intervening hosts (relays/cutouts) to hide their true location. Current trackback methods, especially across administrative boundaries (such as ISPs and enterprises), are too slow: the trail too often has gone cold after only one or two hops. ManHunt uses the existing infrastructure to provide rapid trackback across network segments and administrative boundaries [18].

Interlocking Fields of Fire

Current security tools are designed as independent bastions and this philosophy leads to them being deployed as multiple lines of defense. Incorporating the concept of interlocking fields of fire into such tools can provide a substantial boost in the effectiveness of the overall collection. Although this concept provides for redundant coverage, its most important attribute is that an attacker comes under fire from several different angles: an effective attack on one defender can leave him highly exposed to others.

For example, it is inevitable that attackers will discover ways to identify a ManTrap system. This potential problem can be turned into an advantage by including a suitable IDS. The anticipated tests for ManTrap all produce strong signatures, creating a dilemma for the attacker: If he tests for ManTrap but is on a normal system, he has announced his presence to any IDS aware of those signatures. Thus the mere presence of ManTrap, or even suspected presence, can provide enhanced protection for other hosts on the network. However, since ManTrap is just beginning to appear in operational environments, it has not yet become a factor in attackers' plans. It will be interesting to see how attackers react, and to see if individual sites can influence (to their advantage) the attacker's decision of whether to test for ManTrap.


Notes and References

[1] Example: Attackers rarely stage a Denial-of-Service (DoS) attack from their own site, instead launching it from another compromised site. Thus, a site that is DoS'ing your site is likely an unwitting victim of the real perpetrator.
[2] Time Based Security, Winn Schwartau. Interpact Press, 1999.
[3] In the DISA Vulnerability Analysis and Assessment Program (VAAP), the detection rate was roughly 5% [2-pg iii, 17].
[4] In the VAAP, the reporting rate during the first half was about 5% (or 0.25% of successful attacks) [2-pg iii], but later rose to 27% [17], almost certainly as a consequence of actions taken to remedy the low reporting rate. Hence, the 5% rate of reporting detected attacks is probably the better estimate for generic organizations.
[5] Personal communications from staff of various incident response teams and from law enforcement officials.
[6] The representative of a major computer security company (ISS) classified their customers thusly: 60% had expertise in neither their OS nor computer security, 20% had expertise in both, and the remaining 20% had expertise in OSs but not computer security. Comment made during a panel session at Recent Advances in Intrusion Detection (RAID) Workshop, Purdue U., 1999.
[7] "There Be Dragons," Steven Bellovin. In Proceedings of the Third USENIX UNIX Security Symposium, Baltimore, MD, September 1992. ftp://ftp.cerias.purdue.edu/pub/doc/true_stories/Steven_Bellovin_dragon.ps.Z
[8] "An Evening with Berferd In Which a Cracker is Lured, Endured, and Studied," Bill Cheswick, 1991. ftp://ftp.cerias.purdue.edu/pub/doc/true_stories/Bill_Cheswick_berferd.ps.Z
[9] Deception ToolKit (DTK) from Fred Cohen & Associates. http://all.net/dtk/dtk.html
[10] CyberCop Sting from Network Associates. http://www.pgp.com/asp_set/products/tns/ccsting_intro.asp
[11] SPECTER Intrusion Detection System. http://www.specter.ch/
[12] NetFacade from GTE/BBN. http://www.itsecure.bbn.com/NetFacade.htm
[13] "Stalking the Wily Hacker," Clifford Stoll. In Communications of the ACM, 31(5):484, May 1988.
[14] The Cuckoo's Egg: Tracking a Spy Through the Maze of Computer Espionage, Clifford Stoll. Doubleday, 1989.
[15] "To Build a Honeypot," Lance Spitzer, 2000 June 7. http://www.enteract.com/~lspitz/honeypot.html
[16] "Know Your Enemy: Motives: The Motives and Psychology of the Black-hat Community," The Honeynet Project, 2000 June 27. http://www.enteract.com/~lspitz/motives/
[17] Information Security: Computer Attacks at Department of Defense Pose Increasing Risks, General Accounting Office, U.S. Congress, Chapter Report, 05/22/1996, GAO/AIMD-96-84, section 2:1 Number of Attacks Increasing. http://www.gao.gov/AIndexFY96/abstracts/ai96084.htm
[18] ManHunt is scheduled to be available in 4Q'2000.