| 12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037 | \documentclass{llncs}\usepackage{url}\usepackage{amsmath}\usepackage{epsfig}%\setlength{\textwidth}{5.9in}%\setlength{\textheight}{8.4in}%\setlength{\topmargin}{.5cm}%\setlength{\oddsidemargin}{1cm}%\setlength{\evensidemargin}{1cm}\newenvironment{tightlist}{\begin{list}{$\bullet$}{  \setlength{\itemsep}{0mm}    \setlength{\parsep}{0mm}    %  \setlength{\labelsep}{0mm}    %  \setlength{\labelwidth}{0mm}    %  \setlength{\topsep}{0mm}    }}{\end{list}}\begin{document}\title{Design of a blocking-resistant anonymity system}%\author{Roger Dingledine\inst{1} \and Nick Mathewson\inst{1}}\author{Roger Dingledine \and Nick Mathewson}\institute{The Free Haven Project\\\email{\{arma,nickm\}@freehaven.net}}\maketitle\pagestyle{plain}\begin{abstract}Websites around the world are increasingly being blocked bygovernment-level firewalls. Many people use anonymizing networks likeTor to contact sites without letting an attacker trace their activities,and as an added benefit they are no longer affected by local censorship.But if the attacker simply denies access to the Tor network itself,blocked users can no longer benefit from the security Tor offers.Here we describe a design that builds upon the current Tor networkto provide an anonymizing network that resists blockingby government-level attackers.\end{abstract}\section{Introduction and Goals}Anonymizing networks such as Tor~\cite{tor-design} bounce traffic arounda network of relays. They aim to hide not only what is being said, butalso who is communicating with whom, which users are using which websites,and so on. These systems have a broad range of users, including ordinarycitizens who want to avoid being profiled for targeted advertisements,corporations who don't want to reveal information to their competitors,and law enforcement and government intelligence agencies who need to dooperations on the Internet without being noticed.Historically, research on anonymizing systems has assumed a passiveattacker who monitors the user (call her Alice) and tries to discover heractivities, yet lets her reach any piece of the network. In more modernthreat models such as Tor's, the adversary is allowed to perform activeattacks such as modifying communications in hopes of tricking Aliceinto revealing her destination, or intercepting some of her connectionsto run a man-in-the-middle attack. But these systems still assume thatAlice can eventually reach the anonymizing network.An increasing number of users are making use of the Tor softwarenot so much for its anonymity properties but for its censorshipresistance properties -- if they access Internet sites like Wikipediaand Blogspot via Tor, they are no longer affected by local censorshipand firewall rules. In fact, an informal user study (described inAppendix~\ref{app:geoip}) showed China as the third largest user basefor Tor clients, with perhaps ten thousand people accessing the Tornetwork from China each day.The current Tor design is easy to block if the attacker controls Alice'sconnection to the Tor network --- by blocking the directory authorities,by blocking all the server IP addresses in the directory, or by filteringbased on the signature of the Tor TLS handshake. Here we describe adesign that builds upon the current Tor network to provide an anonymizingnetwork that also resists this blocking. Specifically,Section~\ref{sec:adversary} discusses our threat model --- that is,the assumptions we make about our adversary; Section~\ref{sec:current-tor}describes the components of the current Tor design and how they can beleveraged for a new blocking-resistant design; Section~\ref{sec:related}explains the features and drawbacks of the currently deployed solutions;and ...%And adding more different classes of users and goals to the Tor network%improves the anonymity for all Tor users~\cite{econymics,usability:weis2006}.\section{Adversary assumptions}\label{sec:adversary}The history of blocking-resistance designs is littered with conflictingassumptions about what adversaries to expect and what problems arein the critical path to a solution. Here we try to enumerate our bestunderstanding of the current situation around the world.In the traditional security style, we aim to describe a strong attacker--- if we can defend against this attacker, we inherit protectionagainst weaker attackers as well. After all, we want a general designthat will work for people in China, people in Iran, people in Thailand,whistleblowers in firewalled corporate networks, and people in whateverturns out to be the next oppressive situation. In fact, by designing witha variety of adversaries in mind, we can take advantage of the fact thatadversaries will be in different stages of the arms race at each location.We assume there are three main network attacks in use by censorscurrently~\cite{clayton:pet2006}:\begin{tightlist}\item Block destination by automatically searching for certain stringsin TCP packets.\item Block destination by manually listing its IP address at thefirewall.\item Intercept DNS requests and give bogus responses for certaindestination hostnames.\end{tightlist}We assume the network firewall has very limited CPU perconnection~\cite{clayton:pet2006}. Against an adversary who spendshours looking through the contents of each packet, we would needsome stronger mechanism such as steganography, which introduces itsown problems~\cite{active-wardens,tcpstego,bar}.More broadly, we assume that the chance that the authorities try toblock a given system grows as its popularity grows. That is, a systemused by only a few users will probably never be blocked, whereas awell-publicized system with many users will receive much more scrutiny.We assume that readers of blocked content are not in as much dangeras publishers. So far in places like China, the authorities mainly goafter people who publish materials and coordinate organized movementsagainst the state~\cite{mackinnon}. If they find that a user happensto be reading a site that should be blocked, the typical response issimply to block the site. Of course, even with an encrypted connection,the adversary may be able to distinguish readers from publishers byobserving whether Alice is mostly downloading bytes or mostly uploadingthem --- we discuss this issue more in Section~\ref{subsec:upload-padding}.We assume that while various different regimes can coordinate and sharenotes, there will be a significant time lag between one attacker learninghow to overcome a facet of our design and other attackers picking it up.Similarly, we assume that in the early stages of deployment the insiderthreat isn't as high of a risk, because no attackers have put seriouseffort into breaking the system yet.We assume that government-level attackers are not always uniform acrossthe country. For example, there is no single centralized place in Chinathat coordinates its censorship decisions and steps.We assume that our users have control over their hardware andsoftware --- they don't have any spyware installed, there are nocameras watching their screen, etc. Unfortunately, in many situationsthese threats are very real~\cite{zuckerman-threatmodels}; yetsoftware-based security systems like ours are poorly equipped to handlea user who is entirely observed and controlled by the adversary. SeeSection~\ref{subsec:cafes-and-livecds} for more discussion of what littlewe can do about this issue.We assume that the user will be able to fetch a genuineversion of Tor, rather than one supplied by the adversary; seeSection~\ref{subsec:trust-chain} for discussion on helping the userconfirm that he has a genuine version and that he can connect to thereal Tor network.\section{Components of the current Tor design}\label{sec:current-tor}Tor is popular and sees a lot of use. It's the largest anonymitynetwork of its kind.Tor has attracted more than 800 routers from around the world.A few sentences about how Tor works.In this section, we examine some of the reasons why Tor has taken off,with particular emphasis to how we can take advantage of these propertiesfor a blocking-resistance design.Tor aims to provide three security properties:\begin{tightlist}\item 1. A local network attacker can't learn, or influence, yourdestination.\item 2. No single router in the Tor network can link you to yourdestination.\item 3. The destination, or somebody watching the destination,can't learn your location.\end{tightlist}For blocking-resistance, we care most clearly about the firstproperty. But as the arms race progresses, the second propertywill become important --- for example, to discourage an adversaryfrom volunteering a relay in order to learn that Alice is readingor posting to certain websites. The third property is not so clearlyimportant in this context, but we believe it will turn out to be helpful:consider websites and other Internet services that have been pressuredrecently into treating clients differently depending on their networklocation~\cite{google-geolocation}.% and cite{goodell-syverson06} once it's finalized.The Tor design provides other features as well over manual or adhoc circumvention techniques.Firstly, the Tor directory authorities automatically aggregate, test,and publish signed summaries of the available Tor routers. Tor clientscan fetch these summaries to learn which routers are available andwhich routers have desired properties. Directory information is cachedthroughout the Tor network, so once clients have bootstrapped they neverneed to interact with the authorities directly. (To tolerate a minorityof compromised directory authorities, we use a threshold trust scheme ---see Section~\ref{subsec:trust-chain} for details.)Secondly, Tor clients can be configured to use any directory authoritiesthey want. They use the default authorities if no others are specified,but it's easy to start a separate (or even overlapping) Tor network justby running a different set of authorities and convincing users to prefera modified client. For example, we could launch a distinct Tor networkinside China; some users could even use an aggregate network made up ofboth the main network and the China network. But we should not be tooquick to create other Tor networks --- part of Tor's anonymity comes fromusers behaving like other users, and there are many unsolved anonymityquestions if different users know about different pieces of the network.Thirdly, in addition to automatically learning from the chosen directorieswhich Tor routers are available and working, Tor takes care of buildingpaths through the network and rebuilding them as needed. So the usernever has to know how paths are chosen, never has to manually pickworking proxies, and so on. More generally, at its core the Tor protocolis simply a tool that can build paths given a set of routers. Tor isquite flexible about how it learns about the routers and how it choosesthe paths. Harvard's Blossom project~\cite{blossom-thesis} makes thisflexibility more concrete: Blossom makes use of Tor not for its securityproperties but for its reachability properties. It runs a separate setof directory authorities, its own set of Tor routers (called the Blossomnetwork), and uses Tor's flexible path-building to let users view Internetresources from any point in the Blossom network.Fourthly, Tor separates the role of \emph{internal relay} from therole of \emph{exit relay}. That is, some volunteers choose just to relaytraffic between Tor users and Tor routers, and others choose to also allowconnections to external Internet resources. Because we don't force allvolunteers to play both roles, we end up with more relays. This increaseddiversity in turn is what gives Tor its security: the more options theuser has for her first hop, and the more options she has for her last hop,the less likely it is that a given attacker will be watching both endsof her circuit~\cite{tor-design}. As a bonus, because our design attractsmore internal relays that want to help out but don't want to deal withbeing an exit relay, we end up with more options for the first hop ---the one most critical to being able to reach the Tor network.Fifthly, Tor is sustainable. Zero-Knowledge Systems offered the commercialbut now-defunct Freedom Network~\cite{freedom21-security}, a design withsecurity comparable to Tor's, but its funding model relied on collectingmoney from users to pay relays. Modern commercial proxy systems similarlyneed to keep collecting money to support their infrastructure. On theother hand, Tor has built a self-sustaining community of volunteers whodonate their time and resources. This community trust is rooted in Tor'sopen design: we tell the world exactly how Tor works, and we provide allthe source code. Users can decide for themselves, or pay any securityexpert to decide, whether it is safe to use. Further, Tor's modularityas described above, along with its open license, mean that its impactwill continue to grow.Sixthly, Tor has an established user base of hundreds ofthousands of people from around the world. This diversity ofusers contributes to sustainability as above: Tor is used byordinary citizens, activists, corporations, law enforcement, andeven governments and militaries~\cite{tor-use-cases}, and they canonly achieve their security goals by blending together in the samenetwork~\cite{econymics,usability:weis2006}. This user base also providessomething else: hundreds of thousands of different and often-changingaddresses that we can leverage for our blocking-resistance design.We discuss and adapt these components further inSection~\ref{sec:components}. But first we examine the strengths andweaknesses of other blocking-resistance approaches, so we can expandour repertoire of building blocks and ideas.\section{Current proxy solutions}\label{sec:related}Relay-based blocking-resistance schemes generally have two maincomponents: a relay component and a discovery component. The relay partencompasses the process of establishing a connection, sending trafficback and forth, and so on --- everything that's done once the user knowswhere he's going to connect. Discovery is the step before that: theprocess of finding one or more usable relays.For example, we described several pieces of Tor in the previous section,but we can divide them into the process of building paths and sendingtraffic over them (relay) and the process of learning from the directoryservers about what routers are available (discovery). With this distinctionin mind, we now examine several categories of relay-based schemes.\subsection{Centrally-controlled shared proxies}Existing commercial anonymity solutions (like Anonymizer.com) are basedon a set of single-hop proxies. In these systems, each user connects toa single proxy, which then relays the user's traffic. These public proxysystems are typically characterized by two features: they control andoperator the proxies centrally, and many different users get assignedto each proxy.In terms of the relay component, single proxies provide weak securitycompared to systems that distribute trust over multiple relays, since acompromised proxy can trivially observe all of its users' actions, andan eavesdropper only needs to watch a single proxy to perform timingcorrelation attacks against all its users' traffic. Worse, all usersneed to trust the proxy company to have good security itself as well asto not reveal user activities.On the other hand, single-hop proxies are easier to deploy, and theycan provide better performance than distributed-trust designs like Tor,since traffic only goes through one relay. They're also more convenientfrom the user's perspective --- since users entirely trust the proxy,they can just use their web browser directly.Whether public proxy schemes are more or less scalable than Tor isstill up for debate: commercial anonymity systems can use some of theirrevenue to provision more bandwidth as they grow, whereas volunteer-basedanonymity systems can attract thousands of fast relays to spread the load.The discovery piece can take several forms. Most commercial anonymousproxies have one or a handful of commonly known websites, and their userslog in to those websites and relay their traffic through them. Whenthese websites get blocked (generally soon after the company becomespopular), if the company cares about users in the blocked areas, theystart renting lots of disparate IP addresses and rotating through themas they get blocked. They notify their users of new addresses by email,for example. It's an arms race, since attackers can sign up to receive theemail too, but they have one nice trick available to them: because theyhave a list of paying subscribers, they can notify certain subscribersabout updates earlier than others.Access control systems on the proxy let them provide service only tousers with certain characteristics, such as paying customers or peoplefrom certain IP address ranges.Discovery despite a government-level firewall is a complex and unsolvedtopic, and we're stuck in this same arms race ourselves; we explore itin more detail in Section~\ref{sec:discovery}. But first we examine theother end of the spectrum --- getting volunteers to run the proxies,and telling only a few people about each proxy.\subsection{Independent personal proxies}Personal proxies such as Circumventor~\cite{circumventor} andCGIProxy~\cite{cgiproxy} use the same technology as the public ones asfar as the relay component goes, but they use a different strategy fordiscovery. Rather than managing a few centralized proxies and constantlygetting new addresses for them as the old addresses are blocked, theyaim to have a large number of entirely independent proxies, each managingits own (much smaller) set of users.As the Circumventor site~\cite{circumventor} explains, ``You don'tactually install the Circumventor \emph{on} the computer that is blockedfrom accessing Web sites. You, or a friend of yours, has to install theCircumventor on some \emph{other} machine which is not censored.''This tactic has great advantages in terms of blocking-resistance ---recall our assumption in Section~\ref{sec:adversary} that the attentiona system attracts from the attacker is proportional to its number ofusers and level of publicity. If each proxy only has a few users, andthere is no central list of proxies, most of them will never get noticed.On the other hand, there's a huge scalability question that so far hasprevented these schemes from being widely useful: how does the fellowin China find a person in Ohio who will run a Circumventor for him? Insome cases he may know and trust some people on the outside, but in manycases he's just out of luck. Just as hard, how does a new volunteer inOhio find a person in China who needs it?%discovery is also hard because the hosts keep vanishing if they're%on dynamic ip. But not so bad, since they can use dyndns addresses.This challenge leads to a hybrid design --- centrally-distributedpersonal proxies --- which we will investigate in more detail inSection~\ref{sec:discovery}.\subsection{Open proxies}Yet another currently used approach to bypassing firewalls is to locateopen and misconfigured proxies on the Internet. A quick Google searchfor ``open proxy list'' yields a wide variety of freely available listsof HTTP, HTTPS, and SOCKS proxies. Many small companies have sprung upproviding more refined lists to paying customers.There are some downsides to using these oen proxies though. Firstly,the proxies are of widely varying quality in terms of bandwidth andstability, and many of them are entirely unreachable. Secondly, unlikenetworks of volunteers like Tor, the legality of routing traffic throughthese proxies is questionable: it's widely believed that most of themdon't realize what they're offering, and probably wouldn't allow it ifthey realized. Thirdly, in many cases the connection to the proxy isunencrypted, so firewalls that filter based on keywords in IP packetswill not be hindered. And lastly, many users are suspicious that someopen proxies are a little \emph{too} convenient: are they run by theadversary, in which case they get to monitor all the user's requestsjust as single-hop proxies can?A distributed-trust design like Tor resolves each of these issues forthe relay component, but a constantly changing set of thousands of openrelays is clearly a useful idea for a discovery component. For example,users might be able to make use of these proxies to bootstrap theirfirst introduction into the Tor network.\subsection{JAP}Stefan's WPES paper is probably the closest related work, and isthe starting point for the design in this paper.\subsection{steganography}infranet\subsection{break your sensitive strings into multiple tcp packets;ignore RSTs}\subsection{Internal caching networks}Freenet is deployed inside China and caches outside content.\subsection{Skype}port-hopping. encryption. voice communications not so susceptible tokeystroke loggers (even graphical ones).\subsection{Tor itself}And lastly, we include Tor itself in the list of current solutionsto firewalls. Tens of thousands of people use Tor from countries thatroutinely filter their Internet. Tor's website has been blocked in mostof them. But why hasn't the Tor network been blocked yet?We have several theories. The first is the most straightforward: tens ofthousands of people are simply too few to matter. It may help that Tor isperceived to be for experts only, and thus not worth attention yet. Themore subtle variant on this theory is that we've positioned Tor in thepublic eye as a tool for retaining civil liberties in more free countries,so perhaps blocking authorities don't view it as a threat. (We revisitthis idea when we consider whether and how to publicize a a Tor variantthat improves blocking-resistance --- see Section~\ref{subsec:publicity}for more discussion.)The broader explanation is that most government-level filters are notcreated by people setting out to block all possible ways to bypassthem. They're created by people who want to do a good enough job thatthey can still appear in control. They realize that there will alwaysbe ways for a few people to get around the firewall, and as long as Torhas not publically threatened their control, they see no urgent need toblock it yet.We should recognize that we're \emph{already} in the arms race. Theseconstraints can give us insight into the priorities and capabilities ofour various attackers.\section{Components of a blocking-resistant design}\label{sec:components}We need to address three problems:- adapting the relay component of Tor so it resists blocking better.- Discovery.- Tor's network signature.Here we describe the new pieces we need to add to the current Tor design.\subsection{Bridge relays}Some Tor users on the free side of the network will opt to become\emph{bridge relays}. They will relay a small amount of bandwidth intothe main Tor network, and they won't need to allow exits.They sign up on the bridge directory authorities (described below),and they use Tor to publish their descriptor so an attacker observingthe bridge directory authority's network can't enumerate bridges....need to outline instructions for a Tor config that will publishto an alternate directory authority, and for controller commandsthat will do this cleanly.\subsection{The bridge directory authority (BDA)}They aggregate server descriptors just like the main authorities, andanswer all queries as usual, except they don't publish full directoriesor network statuses.So once you know a bridge relay's key, you can get the most recentserver descriptor for it.Since bridge authorities don't answer full network statuses, weneed to add a new way for users to learn the current status for asingle relay or a small set of relays --- to answer such questions as``is it running?''  or ``is it behaving correctly?'' We describe inSection~\ref{subsec:enclave-dirs} a way for the bridge authority topublish this information without resorting to signing each answerindividually.\subsection{Putting them together}If a blocked user has address information for one working bridge relay,then he can use it to make secure connections to the BDA to update hisknowledge about other bridgerelays, and he can make secure connections to the main Tor networkand directory servers to build circuits and connect to the rest ofthe Internet.So now we've reduced the problem from how to circumvent the firewallfor all transactions (and how to know that the pages you get have notbeen modified by the local attacker) to how to learn about a workingbridge relay.The following section describes ways to bootstrap knowledge of your firstbridge relay, and ways to maintain connectivity once you know a fewbridge relays. (See Section~\ref{subsec:first-bridge} for a discussionof exactly what information is sufficient to characterize a bridge relay.)\section{Discovering and maintaining working bridge relays}\label{sec:discovery}In the first subsection we describe how to find a first bridge.Thus they can reach the BDA. From here we either assume a socialnetwork or other mechanism for learning IP:dirport or key fingerprintsas above, or we assume an account server that allows us to limit thenumber of new bridge relays an external attacker can discover.Going to be an arms race. Need a bag of tricks. Hard to saywhich ones will work. Don't spend them all at once.\subsection{Bootstrapping: finding your first bridge}\label{subsec:first-bridge}Most government firewalls are not perfect. They allow connections toGoogle cache or some open proxy servers, or they let file-sharing orSkype or World-of-Warcraft connections through.For users who can't use any of these techniques, hopefully they knowa friend who can --- for example, perhaps the friend already knows somebridge relay addresses.(If they can't get around it at all, then we can't help them --- theyshould go meet more people.)Some techniques are sufficient to get us an IP address and a port,and others can get us IP:port:key. Lay out some plausible optionsfor how users can bootstrap into learning their first bridge.Round one:- the bridge authority server will hand some out.- get one from your friend.- send us mail with a unique account, and get an automated answer.- Round two:- social network thingattack: adversary can reconstruct your social network by learning whoknows which bridges.\subsection{Centrally-distributed personal proxies}Circumventor, realizing that its adoption will remain limited if would-beusers can't connect with volunteers, has started a mailing list todistribute new proxy addresses every few days. From experimentationit seems they have concluded that sending updates every 3 or 4 days issufficient to stay ahead of the current attackers.If there are many volunteer proxies and many interested users, a centralwatering hole to connect them is a natural solution. On the other hand,at first glance it appears that we've inherited the \emph{bad} parts ofeach of the above designs: not only do we have to attract many volunteerproxies, but the users also need to get to a single site that is sureto be blocked.There are two reasons why we're in better shape. Firstly, the users don'tactually need to reach the watering hole directly: it can respond toemail, for example. Secondly, % In fact, the JAP%project~\cite{web-mix,koepsell:wpes2004} suggested an alternative approach%to a mailing list: new users email a central address and get an automated%response listing a proxy for them.% While the exact details of the%proposal are still to be worked out, the idea of giving out\subsection{Discovery based on social networks}A token that can be exchanged at the BDA (assuming youcan reach it) for a new IP:dirport or server descriptor.The account serverruns as a Tor controller for the bridge authorityUsers can establish reputations, perhaps based on social networkconnectivity, perhaps based on not getting their bridge relays blocked,Probably the most critical lesson learned in past work on reputationsystems in privacy-oriented environments~\cite{p2p-econ} is the need forverifiable transactions. That is, the entity computing and advertisingreputations for participants needs to actually learn in a convincingway that a given transaction was successful or unsuccessful.(Lesson from designing reputation systems~\cite{p2p-econ}: easy toreward good behavior, hard to punish bad behavior.\subsection{How to allocate bridge addresses to users}Hold a fraction in reserve, in case our currently deployed tricksall fail at once --- so we can move to new approaches quickly.(Bridges that sign up and don't get used yet will be sad; but thisis a transient problem --- if bridges are on by default, nobody willmind not being used.)Perhaps each bridge should be known by a single bridge directoryauthority. This makes it easier to trace which users have learned aboutit, so easier to blame or reward. It also makes things more brittle,since loss of that authority means its bridges aren't advertised untilthey switch, and means its bridge users are sad too.(Need a slick hash algorithm that will map our identity key to abridge authority, in a way that's sticky even when we add bridgedirectory authorities, but isn't sticky when our authority goesaway. Does this exist?)Divide bridges into buckets based on their identity key.[Design question: need an algorithm to deterministically map a bridge'sidentity key into a category that isn't too gameable. Take a keyedhash of the identity key plus a secret the bridge authority keeps?An adversary signing up bridges won't easily be able to learn whatcategory he's been put in, so it's slow to attack.]One portion of the bridges is the public bucket. If you ask thebridge account server for a public bridge, it will give you a randomone of these. We expect they'll be the first to be blocked, but they'llhelp the system bootstrap until it *does* get blocked, and remember thatwe're dealing with different blocking regimes around the world that willprogress at different rates.The generalization of the public bucket is a bucket based on the bridgeuser's IP address: you can learn a random entry only from the subbucketyour IP address (actually, your /24) maps to.Another portion of the bridges can be sectioned off to be given out ina time-release basis. The bucket is partitioned into pieces which aredeterministically available only in certain time windows.And of course another portion is made available for the social networkdesign above.Captchas.Is it useful to load balance which bridges are handed out? The abovebucket concept makes some bridges wildly popular and others less so.But I guess that's the point.\subsection{How do we know if a bridge relay has been blocked?}We need some mechanism for testing reachability from inside theblocked area.The easiest answer is for certain users inside the area to sign up astesting relays, and then we can route through them and see if it works.First problem is that different network areas block different net masks,and it will likely be hard to know which users are in which areas. Soif a bridge relay isn't reachable, is that because of a network blocksomewhere, because of a problem at the bridge relay, or just a temporaryoutage?Second problem is that if we pick random users to test random relays, theadversary should sign up users on the inside, and enumerate the relayswe test. But it seems dangerous to just let people come forward anddeclare that things are blocked for them, since they could be trickingus. (This matters even moreso if our reputation system above relies onwhether things get blocked to punish or reward.)Another answer is not to measure directly, but rather let the bridgesreport whether they're being used. If they periodically report to theirbridge directory authority how much use they're seeing, the authoritycan make smart decisions from there.If they install a geoip database, they can periodically report to theirbridge directory authority which countries they're seeing use from. Thismight help us to track which countries are making use of Ramp, and canalso let us learn about new steps the adversary has taken in the armsrace. (If the bridges don't want to install a whole geoip subsystem, theycan report samples of the /24 network for their users, and the authoritiescan do the geoip work. This tradeoff has clear downsides though.)Worry: adversary signs up a bunch of already-blocked bridges. If we'restingy giving out bridges, users in that country won't get useful ones.(Worse, we'll blame the users when the bridges report they're notbeing used?)Worry: the adversary could choose not to block bridges but just recordconnections to them. So be it, I guess.\subsection{How to learn how well the whole idea is working}We need some feedback mechanism to learn how much use the bridge networkas a whole is actually seeing. Part of the reason for this is so we canrespond and adapt the design; part is because the funders expect to seeprogress reports.The above geoip-based approach to detecting blocked bridges gives us asolution though.\section{Security considerations}\label{sec:security}\subsection{Hiding Tor's network signatures}\label{subsec:enclave-dirs}A short paragraph about Tor's current network appearance.The simplest format for communicating information about a bridge relayis as an IP address and port for its directory cache. From there, theuser can ask the directory cache for an up-to-date copy of that bridgerelay's server descriptor, to learn its current circuit keys, the portit uses for Tor connections, and so on.However, connecting directly to the directory cache involves a plaintextHTTP request. A censor could create a network signature for therequest and/or its response, thus preventing these connections. Thereforewe've modified the Tor protocol so that users can connect to the directorycache via the main Tor port --- they establish a TLS connection withthe bridge as normal, and then send a Tor "begindir" relay cell toestablish a connection to its directory cache.Predictable SSL ports:We should encourage most servers to listen on port 443, which iswhere SSL normally listens.Is that all it will take, or should we set things up so some fractionof them pick random ports? I can see that both helping and hurting.Predictable TLS handshakes:Right now Tor has some predictable strings in its TLS handshakes.These can be removed; but should they be replaced with nothing, orshould we try to emulate some popular browser? In any case ourprotocol demands a pair of certs on both sides --- how much will thismake Tor handshakes stand out?\subsection{Minimum info required to describe a bridge}In the previous subsection, we described a way for the bridge userto bootstrap into the network just by knowing the IP address andTor port of a bridge. What about local spoofing attacks? That is,since we never learned an identity key fingerprint for the bridge,a local attacker could intercept our connection and pretend to bethe bridge we had in mind. It turns out that giving false informationisn't that bad --- since the Tor client ships with trusted keys for thebridge directory authority and the Tor network directory authorities,the user can learn whether he's being given a real connection to thebridge authorities or not. (If the adversary intercepts every connectionthe user makes and gives him a bad connection each time, there's nothingwe can do.)What about anonymity-breaking attacks from observing traffic? Not so badeither, since the adversary could do the same attacks just by monitoringthe network traffic.Once the Tor client has fetched the bridge's server descriptor at leastonce, he should remember the identity key fingerprint for that bridgerelay. Thus if the bridge relay moves to a new IP address, the clientcan then query the bridge directory authority to look up a fresh serverdescriptor using this fingerprint.So we've shown that it's \emph{possible} to bootstrap into the networkjust by learning the IP address and port of a bridge, but are theresituations where it's more convenient or more secure to learn itsidentity fingerprint at the beginning too? We discuss that questionmore in Section~\ref{sec:bootstrapping}, but first we introduce moresecurity topics.\subsection{Observers can tell who is publishing and who is reading}\label{subsec:upload-padding}Should bridge users sometimes send bursts of long-range drop cells?\subsection{Anonymity effects from becoming a bridge relay}Against some attacks, becoming a bridge relay can improve anonymity. Thesimplest example is an attacker who owns a small number of Tor servers. Hewill see a connection from the bridge, but he won't be able to knowwhether the connection originated there or was relayed from somebody else.There are some cases where it doesn't seem to help: if an attacker canwatch all of the bridge's incoming and outgoing traffic, then it's easyto learn which connections were relayed and which started there. (In thiscase he still doesn't know the final destinations unless he is watchingthem too, but in this case bridges are no better off than if they werean ordinary client.)There are also some potential downsides to running a bridge. First, whilewe try to make it hard to enumerate all bridges, it's still possible tolearn about some of them, and for some people just the fact that they'rerunning one might signal to an attacker that they place a high valueon their anonymity. Second, there are some more esoteric attacks on Torrelays that are not as well-understood or well-tested --- for example, anattacker may be able to ``observe'' whether the bridge is sending trafficeven if he can't actually watch its network, by relaying traffic throughit and noticing changes in traffic timing~\cite{attack-tor-oak05}. Onthe other hand, it may be that limiting the bandwidth the bridge iswilling to relay will allow this sort of attacker to determine if it'sbeing used as a bridge but not whether it is adding traffic of its own.It is an open research question whether the benefits outweigh the risks. Alot of the decision rests on which the attacks users are most worriedabout. For most users, we don't think running a bridge relay will bethat damaging.\subsection{Trusting local hardware: Internet cafes and LiveCDs}\label{subsec:cafes-and-livecds}Assuming that users have their own trusted hardware is notalways reasonable.For Internet cafe Windows computers that let you attach your own USB key,a USB-based Tor image would be smart. There's Torpark, and hopefullythere will be more options down the road. Worries about hardware orsoftware keyloggers and other spyware --- and physical surveillance.If the system lets you boot from a CD or from a USB key, you can gaina bit more security by bringing a privacy LiveCD with you. Hardwarekeyloggers and physical surveillance still a worry. LiveCDs also usefulif it's your own hardware, since it's easier to avoid leaving breadcrumbseverywhere.\subsection{Forward compatibility and retiring bridge authorities}Eventually we'll want to change the identity key and/or locationof a bridge authority. How do we do this mostly cleanly?\subsection{The trust chain}\label{subsec:trust-chain}Tor's ``public key infrastructure'' provides a chain of trust tolet users verify that they're actually talking to the right servers.There are four pieces to this trust chain.Firstly, when Tor clients are establishing circuits, at each stepthey demand that the next Tor server in the path prove knowledge ofits private key~\cite{tor-design}. This step prevents the first nodein the path from just spoofing the rest of the path. Secondly, theTor directory authorities provide a signed list of servers along withtheir public keys --- so unless the adversary can control a thresholdof directory authorities, he can't trick the Tor client into using otherTor servers. Thirdly, the location and keys of the directory authorities,in turn, is hard-coded in the Tor source code --- so as long as the usergot a genuine version of Tor, he can know that he is using the genuineTor network. And lastly, the source code and other packages are signedwith the GPG keys of the Tor developers, so users can confirm that theydid in fact download a genuine version of Tor.But how can a user in an oppressed country know that he has the correctkey fingerprints for the developers? As with other security systems, itultimately comes down to human interaction. The keys are signed by dozensof people around the world, and we have to hope that our users have metenough people in the PGP web of trust~\cite{pgp-wot} that they can learnthe correct keys. For users that aren't connected to the global securitycommunity, though, this question remains a critical weakness.% XXX make clearer the trust chain step for bridge directory authorities\section{Performance improvements}\label{sec:performance}\subsection{Fetch server descriptors just-in-time}I guess we should encourage most places to do this, so blockedusers don't stand out.network-status and directory optimizations. caching better. partitioningissues?\section{Maintaining reachability}\subsection{How many bridge relays should you know about?}If they're ordinary Tor users on cable modem or DSL, many of them willdisappear and/or move periodically. How many bridge relays should ablockee knowabout before he's likely to have at least one reachable at any given point?How do we factor in a parameter for "speed that his bridges get discoveredand blocked"?The related question is: if the bridge relays change IP addressesperiodically, how often does the bridge user need to "check in" in orderto keep from being cut out of the loop?\subsection{Cablemodem users don't provide important websites}\label{subsec:block-cable}...so our adversary could just block all DSL and cablemodem networks,and for the most part only our bridge relays would be affected.The first answer is to aim to get volunteers both from traditionally``consumer'' networks and also from traditionally ``producer'' networks.The second answer (not so good) would be to encourage more use of consumernetworks for popular and useful websites.Other attack: China pressures Verizon to discourage its users fromrunning bridges.\subsection{Scanning-resistance}If it's trivial to verify that we're a bridge, and we run on a predictableport, then it's conceivable our attacker would scan the whole Internetlooking for bridges. (In fact, he can just scan likely networks likecablemodem and DSL services --- see Section~\ref{block-cable} for a relatedattack.) It would be nice to slow down this attack. It wouldbe even nicer to make it hard to learn whether we're a bridge withoutfirst knowing some secret.Password protecting the bridges.Could provide a password to the bridge user. He provides a nonced hash ofit or something when he connects. We'd need to give him an ID key for thebridge too, and wait to present the password until we've TLSed, else theadversary can pretend to be the bridge and MITM him to learn the password.\subsection{How to motivate people to run bridge relays}One of the traditional ways to get people to run software that benefitsothers is to give them motivation to install it themselves.  An oftensuggested approach is to install it as a stunning screensaver so everybodywill be pleased to run it. We take a similar approach here, by leveragingthe fact that these users are already interested in protecting theirown Internet traffic, so they will install and run the software.Make all Tor users become bridges if they're reachable -- needs more workon usability first, but we're making progress.Also, we can make a snazzy network graph with Vidalia that emphasizesthe connections the bridge user is currently relaying. (Minor anonymityimplications, but hey.) (In many cases there won't be much activity,so this may backfire. Or it may be better suited to full-fledged Torservers.)\subsection{What if the clients can't install software?}Bridge users without Tor clientsBridge relays could always open their socks proxy. This is bad though,firstlybecause they learn the bridge users' destinations, and secondly becausewe've learned that open socks proxies tend to attract abusive users whohave no idea they're using Tor.Bridges could require passwords in the socks handshake (not supportedby most software including Firefox). Or they could run web proxiesthat require authentication and then pass the requests into Tor. Thisapproach is probably a good way to help bootstrap the Psiphon network,if one of its barriers to deployment is a lack of volunteers willingto exit directly to websites. But it clearly drops some of the niceanonymity features Tor provides.\subsection{Publicity attracts attention}\label{subsec:publicity}both good and bad.\subsection{The Tor website: how to get the software}\section{Related work}\section{Future designs}\subsection{Bridges inside the blocked network too}Assuming actually crossing the firewall is the risky part of theoperation, can we have some bridge relays inside the blocked area too,and more established users can use them as relays so they don't need tocommunicate over the firewall directly at all? A simple example here isto make new blocked users into internal bridges also -- so they sign upon the BDA as part of doing their query, and we give out their addressesrather than (or along with) the external bridge addresses. This designis a lot trickier because it brings in the complexity of whether theinternal bridges will remain available, can maintain reachability withthe outside world, etc.Hidden services as bridges. Hidden services as bridge directory authorities.\bibliographystyle{plain} \bibliography{tor-design}\appendix\section{Counting Tor users by country}\label{app:geoip}\end{document}ship geoip db to bridges. they look up users who tls to them in the db,and upload a signed list of countries and number-of-users each day. thebridge authority aggregates them and publishes stats.bridge relays have buddiesthey ask a user to test the reachability of their buddy.leaks O(1) bridges, but not O(n).we should not be blockable by ordinary cisco censorship features.that is, if they want to block our new design, they will need toadd a feature to block exactly this.strategically speaking, this may come in handy.hash identity key + secret that bridge authority knows. startout dividing into 2^n buckets, where n starts at 0, and we choosewhich bucket you're in based on the first n bits of the hash.Bridges come in clumps of 4 or 8 or whatever. If you know one bridgein a clump, the authority will tell you the rest. Now bridges canask users to test reachability of their buddies.Giving out clumps helps with dynamic IP addresses too. Whether itshould be 4 or 8 depends on our churn.the account server. let's call it a database, it doesn't have tobe a thing that human interacts with.rate limiting mechanisms:energy spent. captchas. relaying traffic for others?send us $10, we'll give you an accountso how do we reward people for being good?
 |