blocking.tex 70 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445
  1. \documentclass{llncs}
  2. \usepackage{url}
  3. \usepackage{amsmath}
  4. \usepackage{epsfig}
  5. \setlength{\textwidth}{5.9in}
  6. \setlength{\textheight}{8.4in}
  7. \setlength{\topmargin}{.5cm}
  8. \setlength{\oddsidemargin}{1cm}
  9. \setlength{\evensidemargin}{1cm}
  10. \newenvironment{tightlist}{\begin{list}{$\bullet$}{
  11. \setlength{\itemsep}{0mm}
  12. \setlength{\parsep}{0mm}
  13. % \setlength{\labelsep}{0mm}
  14. % \setlength{\labelwidth}{0mm}
  15. % \setlength{\topsep}{0mm}
  16. }}{\end{list}}
  17. \begin{document}
  18. \title{Design of a blocking-resistant anonymity system\\DRAFT}
  19. %\author{Roger Dingledine\inst{1} \and Nick Mathewson\inst{1}}
  20. \author{Roger Dingledine \and Nick Mathewson}
  21. \institute{The Free Haven Project\\
  22. \email{\{arma,nickm\}@freehaven.net}}
  23. \maketitle
  24. \pagestyle{plain}
  25. \begin{abstract}
  26. Websites around the world are increasingly being blocked by
  27. government-level firewalls. Many people use anonymizing networks like
  28. Tor to contact sites without letting an attacker trace their activities,
  29. and as an added benefit they are no longer affected by local censorship.
  30. But if the attacker simply denies access to the Tor network itself,
  31. blocked users can no longer benefit from the security Tor offers.
  32. Here we describe a design that builds upon the current Tor network
  33. to provide an anonymizing network that resists blocking
  34. by government-level attackers.
  35. \end{abstract}
  36. \section{Introduction and Goals}
  37. Anonymizing networks such as Tor~\cite{tor-design} bounce traffic around
  38. a network of relays. They aim to hide not only what is being said, but
  39. also who is communicating with whom, which users are using which websites,
  40. and so on. These systems have a broad range of users, including ordinary
  41. citizens who want to avoid being profiled for targeted advertisements,
  42. corporations who don't want to reveal information to their competitors,
  43. and law enforcement and government intelligence agencies who need to do
  44. operations on the Internet without being noticed.
  45. Historically, research on anonymizing systems has focused on a passive
  46. attacker who monitors the user (call her Alice) and tries to discover her
  47. activities, yet lets her reach any piece of the network. In more modern
  48. threat models such as Tor's, the adversary is allowed to perform active
  49. attacks such as modifying communications to trick Alice
  50. into revealing her destination, or intercepting some connections
  51. to run a man-in-the-middle attack. But these systems still assume that
  52. Alice can eventually reach the anonymizing network.
  53. An increasing number of users are using the Tor software
  54. less for its anonymity properties than for its censorship
  55. resistance properties---if they use Tor to access Internet sites like
  56. Wikipedia
  57. and Blogspot, they are no longer affected by local censorship
  58. and firewall rules. In fact, an informal user study (described in
  59. Appendix~\ref{app:geoip}) showed China as the third largest user base
  60. for Tor clients, with perhaps ten thousand people accessing the Tor
  61. network from China each day.
  62. The current Tor design is easy to block if the attacker controls Alice's
  63. connection to the Tor network---by blocking the directory authorities,
  64. by blocking all the server IP addresses in the directory, or by filtering
  65. based on the signature of the Tor TLS handshake. Here we describe a
  66. design that builds upon the current Tor network to provide an anonymizing
  67. network that also resists this blocking. Specifically,
  68. Section~\ref{sec:adversary} discusses our threat model---that is,
  69. the assumptions we make about our adversary; Section~\ref{sec:current-tor}
  70. describes the components of the current Tor design and how they can be
  71. leveraged for a new blocking-resistant design; Section~\ref{sec:related}
  72. explains the features and drawbacks of the currently deployed solutions;
  73. and ...
  74. % The other motivation is for places where we're concerned they will
  75. % try to enumerate a list of Tor users. So even if they're not blocking
  76. % the Tor network, it may be smart to not be visible as connecting to it.
  77. %And adding more different classes of users and goals to the Tor network
  78. %improves the anonymity for all Tor users~\cite{econymics,usability:weis2006}.
  79. \section{Adversary assumptions}
  80. \label{sec:adversary}
  81. The history of blocking-resistance designs is littered with conflicting
  82. assumptions about what adversaries to expect and what problems are
  83. in the critical path to a solution. Here we try to enumerate our best
  84. understanding of the current situation around the world.
  85. In the traditional security style, we aim to describe a strong
  86. attacker---if we can defend against this attacker, we inherit protection
  87. against weaker attackers as well. After all, we want a general design
  88. that will work for citizens of China, Iran, Thailand, and other censored
  89. countries; for
  90. whistleblowers in firewalled corporate network; and for people in
  91. unanticipated oppressive situations. In fact, by designing with
  92. a variety of adversaries in mind, we can take advantage of the fact that
  93. adversaries will be in different stages of the arms race at each location,
  94. so a server blocked in one locale can still be useful in others.
  95. We assume there are three main network attacks in use by censors
  96. currently~\cite{clayton:pet2006}:
  97. \begin{tightlist}
  98. \item Block a destination or type of traffic by automatically searching for
  99. certain strings or patterns in TCP packets.
  100. \item Block a destination by manually listing its IP address at the
  101. firewall.
  102. \item Intercept DNS requests and give bogus responses for certain
  103. destination hostnames.
  104. \end{tightlist}
  105. We assume the network firewall has limited CPU and memory per
  106. connection~\cite{clayton:pet2006}. Against an adversary who carefully
  107. examines the contents of every packet, we would need
  108. some stronger mechanism such as steganography, which introduces its
  109. own problems~\cite{active-wardens,tcpstego,bar}.
  110. More broadly, we assume that the authorities are more likely to
  111. block a given system as its popularity grows. That is, a system
  112. used by only a few users will probably never be blocked, whereas a
  113. well-publicized system with many users will receive much more scrutiny.
  114. We assume that readers of blocked content are not in as much danger
  115. as publishers. So far in places like China, the authorities mainly go
  116. after people who publish materials and coordinate organized
  117. movements~\cite{mackinnon}.
  118. If they find that a user happens
  119. to be reading a site that should be blocked, the typical response is
  120. simply to block the site. Of course, even with an encrypted connection,
  121. the adversary may be able to distinguish readers from publishers by
  122. observing whether Alice is mostly downloading bytes or mostly uploading
  123. them---we discuss this issue more in Section~\ref{subsec:upload-padding}.
  124. We assume that while various different regimes can coordinate and share
  125. notes, there will be a time lag between one attacker learning
  126. how to overcome a facet of our design and other attackers picking it up.
  127. Similarly, we assume that in the early stages of deployment the insider
  128. threat isn't as high of a risk, because no attackers have put serious
  129. effort into breaking the system yet.
  130. We do not assume that government-level attackers are always uniform across
  131. the country. For example, there is no single centralized place in China
  132. that coordinates its censorship decisions and steps.
  133. We assume that our users have control over their hardware and
  134. software---they don't have any spyware installed, there are no
  135. cameras watching their screen, etc. Unfortunately, in many situations
  136. these threats are real~\cite{zuckerman-threatmodels}; yet
  137. software-based security systems like ours are poorly equipped to handle
  138. a user who is entirely observed and controlled by the adversary. See
  139. Section~\ref{subsec:cafes-and-livecds} for more discussion of what little
  140. we can do about this issue.
  141. We assume that widespread access to the Internet is economically,
  142. politically, and/or
  143. socially valuable to the policymakers of each deployment country. After
  144. all, if censorship
  145. is more important than Internet access, the firewall administrators have
  146. an easy job: they should simply block everything. The corollary to this
  147. assumption is that we should design so that increased blocking of our
  148. system results in increased economic damage or public outcry.
  149. We assume that the user will be able to fetch a genuine
  150. version of Tor, rather than one supplied by the adversary; see
  151. Section~\ref{subsec:trust-chain} for discussion on helping the user
  152. confirm that he has a genuine version and that he can connect to the
  153. real Tor network.
  154. \section{Components of the current Tor design}
  155. \label{sec:current-tor}
  156. Tor is popular and sees a lot of use. It's the largest anonymity
  157. network of its kind.
  158. Tor has attracted more than 800 volunteer-operated routers from around the
  159. world. Tor protects users by routing their traffic through a multiply
  160. encrypted ``circuit'' built of a few randomly selected servers, each of which
  161. can remove only a single layer of encryption. Each server sees only the step
  162. before it and the step after it in the circuit, and so no single server can
  163. learn the connection between a user and her chosen communication partners.
  164. In this section, we examine some of the reasons why Tor has become popular,
  165. with particular emphasis to how we can take advantage of these properties
  166. for a blocking-resistance design.
  167. Tor aims to provide three security properties:
  168. \begin{tightlist}
  169. \item 1. A local network attacker can't learn, or influence, your
  170. destination.
  171. \item 2. No single router in the Tor network can link you to your
  172. destination.
  173. \item 3. The destination, or somebody watching the destination,
  174. can't learn your location.
  175. \end{tightlist}
  176. For blocking-resistance, we care most clearly about the first
  177. property. But as the arms race progresses, the second property
  178. will become important---for example, to discourage an adversary
  179. from volunteering a relay in order to learn that Alice is reading
  180. or posting to certain websites. The third property helps keep users safe from
  181. collaborating websites: consider websites and other Internet services
  182. that have been pressured
  183. recently into revealing the identity of bloggers~\cite{arrested-bloggers}
  184. or treating clients differently depending on their network
  185. location~\cite{google-geolocation}.
  186. % and cite{goodell-syverson06} once it's finalized.
  187. The Tor design provides other features as well over manual or ad
  188. hoc circumvention techniques.
  189. First, the Tor directory authorities automatically aggregate, test,
  190. and publish signed summaries of the available Tor routers. Tor clients
  191. can fetch these summaries to learn which routers are available and
  192. which routers are suitable for their needs. Directory information is cached
  193. throughout the Tor network, so once clients have bootstrapped they never
  194. need to interact with the authorities directly. (To tolerate a minority
  195. of compromised directory authorities, we use a threshold trust scheme---
  196. see Section~\ref{subsec:trust-chain} for details.)
  197. Second, Tor clients can be configured to use any directory authorities
  198. they want. They use the default authorities if no others are specified,
  199. but it's easy to start a separate (or even overlapping) Tor network just
  200. by running a different set of authorities and convincing users to prefer
  201. a modified client. For example, we could launch a distinct Tor network
  202. inside China; some users could even use an aggregate network made up of
  203. both the main network and the China network. (But we should not be too
  204. quick to create other Tor networks---part of Tor's anonymity comes from
  205. users behaving like other users, and there are many unsolved anonymity
  206. questions if different users know about different pieces of the network.)
  207. Third, in addition to automatically learning from the chosen directories
  208. which Tor routers are available and working, Tor takes care of building
  209. paths through the network and rebuilding them as needed. So the user
  210. never has to know how paths are chosen, never has to manually pick
  211. working proxies, and so on. More generally, at its core the Tor protocol
  212. is simply a tool that can build paths given a set of routers. Tor is
  213. quite flexible about how it learns about the routers and how it chooses
  214. the paths. Harvard's Blossom project~\cite{blossom-thesis} makes this
  215. flexibility more concrete: Blossom makes use of Tor not for its security
  216. properties but for its reachability properties. It runs a separate set
  217. of directory authorities, its own set of Tor routers (called the Blossom
  218. network), and uses Tor's flexible path-building to let users view Internet
  219. resources from any point in the Blossom network.
  220. Fourth, Tor separates the role of \emph{internal relay} from the
  221. role of \emph{exit relay}. That is, some volunteers choose just to relay
  222. traffic between Tor users and Tor routers, and others choose to also allow
  223. connections to external Internet resources. Because we don't force all
  224. volunteers to play both roles, we end up with more relays. This increased
  225. diversity in turn is what gives Tor its security: the more options the
  226. user has for her first hop, and the more options she has for her last hop,
  227. the less likely it is that a given attacker will be watching both ends
  228. of her circuit~\cite{tor-design}. As a bonus, because our design attracts
  229. more internal relays that want to help out but don't want to deal with
  230. being an exit relay, we end up with more options for the first hop---the
  231. one most critical to being able to reach the Tor network.
  232. Fifth, Tor is sustainable. Zero-Knowledge Systems offered the commercial
  233. but now defunct Freedom Network~\cite{freedom21-security}, a design with
  234. security comparable to Tor's, but its funding model relied on collecting
  235. money from users to pay relay operators. Modern commercial proxy systems
  236. similarly
  237. need to keep collecting money to support their infrastructure. On the
  238. other hand, Tor has built a self-sustaining community of volunteers who
  239. donate their time and resources. This community trust is rooted in Tor's
  240. open design: we tell the world exactly how Tor works, and we provide all
  241. the source code. Users can decide for themselves, or pay any security
  242. expert to decide, whether it is safe to use. Further, Tor's modularity
  243. as described above, along with its open license, mean that its impact
  244. will continue to grow.
  245. Sixth, Tor has an established user base of hundreds of
  246. thousands of people from around the world. This diversity of
  247. users contributes to sustainability as above: Tor is used by
  248. ordinary citizens, activists, corporations, law enforcement, and
  249. even government and military users~\cite{tor-use-cases}, and they can
  250. only achieve their security goals by blending together in the same
  251. network~\cite{econymics,usability:weis2006}. This user base also provides
  252. something else: hundreds of thousands of different and often-changing
  253. addresses that we can leverage for our blocking-resistance design.
  254. We discuss and adapt these components further in
  255. Section~\ref{sec:bridges}. But first we examine the strengths and
  256. weaknesses of other blocking-resistance approaches, so we can expand
  257. our repertoire of building blocks and ideas.
  258. \section{Current proxy solutions}
  259. \label{sec:related}
  260. Relay-based blocking-resistance schemes generally have two main
  261. components: a relay component and a discovery component. The relay part
  262. encompasses the process of establishing a connection, sending traffic
  263. back and forth, and so on---everything that's done once the user knows
  264. where she's going to connect. Discovery is the step before that: the
  265. process of finding one or more usable relays.
  266. For example, we can divide the pieces of Tor in the previous section
  267. into the process of building paths and sending
  268. traffic over them (relay) and the process of learning from the directory
  269. servers about what routers are available (discovery). With this distinction
  270. in mind, we now examine several categories of relay-based schemes.
  271. \subsection{Centrally-controlled shared proxies}
  272. Existing commercial anonymity solutions (like Anonymizer.com) are based
  273. on a set of single-hop proxies. In these systems, each user connects to
  274. a single proxy, which then relays traffic between the user and her
  275. destination. These public proxy
  276. systems are typically characterized by two features: they control and
  277. operate the proxies centrally, and many different users get assigned
  278. to each proxy.
  279. In terms of the relay component, single proxies provide weak security
  280. compared to systems that distribute trust over multiple relays, since a
  281. compromised proxy can trivially observe all of its users' actions, and
  282. an eavesdropper only needs to watch a single proxy to perform timing
  283. correlation attacks against all its users' traffic and thus learn where
  284. everyone is connecting. Worse, all users
  285. need to trust the proxy company to have good security itself as well as
  286. to not reveal user activities.
  287. On the other hand, single-hop proxies are easier to deploy, and they
  288. can provide better performance than distributed-trust designs like Tor,
  289. since traffic only goes through one relay. They're also more convenient
  290. from the user's perspective---since users entirely trust the proxy,
  291. they can just use their web browser directly.
  292. Whether public proxy schemes are more or less scalable than Tor is
  293. still up for debate: commercial anonymity systems can use some of their
  294. revenue to provision more bandwidth as they grow, whereas volunteer-based
  295. anonymity systems can attract thousands of fast relays to spread the load.
  296. The discovery piece can take several forms. Most commercial anonymous
  297. proxies have one or a handful of commonly known websites, and their users
  298. log in to those websites and relay their traffic through them. When
  299. these websites get blocked (generally soon after the company becomes
  300. popular), if the company cares about users in the blocked areas, they
  301. start renting lots of disparate IP addresses and rotating through them
  302. as they get blocked. They notify their users of new addresses (by email,
  303. for example). It's an arms race, since attackers can sign up to receive the
  304. email too, but operators have one nice trick available to them: because they
  305. have a list of paying subscribers, they can notify certain subscribers
  306. about updates earlier than others.
  307. Access control systems on the proxy let them provide service only to
  308. users with certain characteristics, such as paying customers or people
  309. from certain IP address ranges.
  310. Discovery in the face of a government-level firewall is a complex and
  311. unsolved
  312. topic, and we're stuck in this same arms race ourselves; we explore it
  313. in more detail in Section~\ref{sec:discovery}. But first we examine the
  314. other end of the spectrum---getting volunteers to run the proxies,
  315. and telling only a few people about each proxy.
  316. \subsection{Independent personal proxies}
  317. Personal proxies such as Circumventor~\cite{circumventor} and
  318. CGIProxy~\cite{cgiproxy} use the same technology as the public ones as
  319. far as the relay component goes, but they use a different strategy for
  320. discovery. Rather than managing a few centralized proxies and constantly
  321. getting new addresses for them as the old addresses are blocked, they
  322. aim to have a large number of entirely independent proxies, each managing
  323. its own (much smaller) set of users.
  324. As the Circumventor site~\cite{circumventor} explains, ``You don't
  325. actually install the Circumventor \emph{on} the computer that is blocked
  326. from accessing Web sites. You, or a friend of yours, has to install the
  327. Circumventor on some \emph{other} machine which is not censored.''
  328. This tactic has great advantages in terms of blocking-resistance---recall
  329. our assumption in Section~\ref{sec:adversary} that the attention
  330. a system attracts from the attacker is proportional to its number of
  331. users and level of publicity. If each proxy only has a few users, and
  332. there is no central list of proxies, most of them will never get noticed by
  333. the censors.
  334. On the other hand, there's a huge scalability question that so far has
  335. prevented these schemes from being widely useful: how does the fellow
  336. in China find a person in Ohio who will run a Circumventor for him? In
  337. some cases he may know and trust some people on the outside, but in many
  338. cases he's just out of luck. Just as hard, how does a new volunteer in
  339. Ohio find a person in China who needs it?
  340. % another key feature of a proxy run by your uncle is that you
  341. % self-censor, so you're unlikely to bring abuse complaints onto
  342. % your uncle. self-censoring clearly has a downside too, though.
  343. This challenge leads to a hybrid design---centrally-distributed
  344. personal proxies---which we will investigate in more detail in
  345. Section~\ref{sec:discovery}.
  346. \subsection{Open proxies}
  347. Yet another currently used approach to bypassing firewalls is to locate
  348. open and misconfigured proxies on the Internet. A quick Google search
  349. for ``open proxy list'' yields a wide variety of freely available lists
  350. of HTTP, HTTPS, and SOCKS proxies. Many small companies have sprung up
  351. providing more refined lists to paying customers.
  352. There are some downsides to using these open proxies though. First,
  353. the proxies are of widely varying quality in terms of bandwidth and
  354. stability, and many of them are entirely unreachable. Second, unlike
  355. networks of volunteers like Tor, the legality of routing traffic through
  356. these proxies is questionable: it's widely believed that most of them
  357. don't realize what they're offering, and probably wouldn't allow it if
  358. they realized. Third, in many cases the connection to the proxy is
  359. unencrypted, so firewalls that filter based on keywords in IP packets
  360. will not be hindered. And last, many users are suspicious that some
  361. open proxies are a little \emph{too} convenient: are they run by the
  362. adversary, in which case they get to monitor all the user's requests
  363. just as single-hop proxies can?
  364. A distributed-trust design like Tor resolves each of these issues for
  365. the relay component, but a constantly changing set of thousands of open
  366. relays is clearly a useful idea for a discovery component. For example,
  367. users might be able to make use of these proxies to bootstrap their
  368. first introduction into the Tor network.
  369. \subsection{JAP}
  370. Stefan's WPES paper~\cite{koepsell:wpes2004} is probably the closest
  371. related work, and is
  372. the starting point for the design in this paper.
  373. \subsection{steganography}
  374. infranet
  375. \subsection{break your sensitive strings into multiple tcp packets;
  376. ignore RSTs}
  377. \subsection{Internal caching networks}
  378. Freenet is deployed inside China and caches outside content.
  379. \subsection{Skype}
  380. port-hopping. encryption. voice communications not so susceptible to
  381. keystroke loggers (even graphical ones).
  382. \subsection{Tor itself}
  383. And last, we include Tor itself in the list of current solutions
  384. to firewalls. Tens of thousands of people use Tor from countries that
  385. routinely filter their Internet. Tor's website has been blocked in most
  386. of them. But why hasn't the Tor network been blocked yet?
  387. We have several theories. The first is the most straightforward: tens of
  388. thousands of people are simply too few to matter. It may help that Tor is
  389. perceived to be for experts only, and thus not worth attention yet. The
  390. more subtle variant on this theory is that we've positioned Tor in the
  391. public eye as a tool for retaining civil liberties in more free countries,
  392. so perhaps blocking authorities don't view it as a threat. (We revisit
  393. this idea when we consider whether and how to publicize a Tor variant
  394. that improves blocking-resistance---see Section~\ref{subsec:publicity}
  395. for more discussion.)
  396. The broader explanation is that the maintainance of most government-level
  397. filters is aimed at stopping widespread information flow and appearing to be
  398. in control, not by the impossible goal of blocking all possible ways to bypass
  399. censorship. Censors realize that there will always
  400. be ways for a few people to get around the firewall, and as long as Tor
  401. has not publically threatened their control, they see no urgent need to
  402. block it yet.
  403. We should recognize that we're \emph{already} in the arms race. These
  404. constraints can give us insight into the priorities and capabilities of
  405. our various attackers.
  406. \section{The relay component of our blocking-resistant design}
  407. \label{sec:bridges}
  408. Section~\ref{sec:current-tor} describes many reasons why Tor is
  409. well-suited as a building block in our context, but several changes will
  410. allow the design to resist blocking better. The most critical changes are
  411. to get more relay addresses, and to distribute them to users differently.
  412. %We need to address three problems:
  413. %- adapting the relay component of Tor so it resists blocking better.
  414. %- Discovery.
  415. %- Tor's network signature.
  416. %Here we describe the new pieces we need to add to the current Tor design.
  417. \subsection{Bridge relays}
  418. Today, Tor servers operate on less than a thousand distinct IP addresses;
  419. an adversary
  420. could enumerate and block them all with little trouble. To provide a
  421. means of ingress to the network, we need a larger set of entry points, most
  422. of which an adversary won't be able to enumerate easily. Fortunately, we
  423. have such a set: the Tor users.
  424. Hundreds of thousands of people around the world use Tor. We can leverage
  425. our already self-selected user base to produce a list of thousands of
  426. often-changing IP addresses. Specifically, we can give them a little
  427. button in the GUI that says ``Tor for Freedom'', and users who click
  428. the button will turn into \emph{bridge relays} (or just \emph{bridges}
  429. for short). They can rate limit relayed connections to 10 KB/s (almost
  430. nothing for a broadband user in a free country, but plenty for a user
  431. who otherwise has no access at all), and since they are just relaying
  432. bytes back and forth between blocked users and the main Tor network, they
  433. won't need to make any external connections to Internet sites. Because
  434. of this separation of roles, and because we're making use of software
  435. that the volunteers have already installed for their own use, we expect
  436. our scheme to attract and maintain more volunteers than previous schemes.
  437. As usual, there are new anonymity and security implications from running a
  438. bridge relay, particularly from letting people relay traffic through your
  439. Tor client; but we leave this discussion for Section~\ref{sec:security}.
  440. %...need to outline instructions for a Tor config that will publish
  441. %to an alternate directory authority, and for controller commands
  442. %that will do this cleanly.
  443. \subsection{The bridge directory authority}
  444. How do the bridge relays advertise their existence to the world? We
  445. introduce a second new component of the design: a specialized directory
  446. authority that aggregates and tracks bridges. Bridge relays periodically
  447. publish server descriptors (summaries of their keys, locations, etc,
  448. signed by their long-term identity key), just like the relays in the
  449. ``main'' Tor network, but in this case they publish them only to the
  450. bridge directory authorities.
  451. The main difference between bridge authorities and the directory
  452. authorities for the main Tor network is that the main authorities provide
  453. a list of every known relay, but the bridge authorities only give
  454. out a server descriptor if you already know its identity key. That is,
  455. you can keep up-to-date on a bridge's location and other information
  456. once you know about it, but you can't just grab a list of all the bridges.
  457. The identity key, IP address, and directory port for each bridge
  458. authority ship by default with the Tor software, so the bridge relays
  459. can be confident they're publishing to the right location, and the
  460. blocked users can establish an encrypted authenticated channel. See
  461. Section~\ref{subsec:trust-chain} for more discussion of the public key
  462. infrastructure and trust chain.
  463. Bridges use Tor to publish their descriptors privately and securely,
  464. so even an attacker monitoring the bridge directory authority's network
  465. can't make a list of all the addresses contacting the authority.
  466. Bridges may publish to only a subset of the
  467. authorities, to limit the potential impact of an authority compromise.
  468. %\subsection{A simple matter of engineering}
  469. %
  470. %Although we've described bridges and bridge authorities in simple terms
  471. %above, some design modifications and features are needed in the Tor
  472. %codebase to add them. We describe the four main changes here.
  473. %
  474. %Firstly, we need to get smarter about rate limiting:
  475. %Bandwidth classes
  476. %
  477. %Secondly, while users can in fact configure which directory authorities
  478. %they use, we need to add a new type of directory authority and teach
  479. %bridges to fetch directory information from the main authorities while
  480. %publishing server descriptors to the bridge authorities. We're most of
  481. %the way there, since we can already specify attributes for directory
  482. %authorities:
  483. %add a separate flag named ``blocking''.
  484. %
  485. %Thirdly, need to build paths using bridges as the first
  486. %hop. One more hole in the non-clique assumption.
  487. %
  488. %Lastly, since bridge authorities don't answer full network statuses,
  489. %we need to add a new way for users to learn the current status for a
  490. %single relay or a small set of relays---to answer such questions as
  491. %``is it running?'' or ``is it behaving correctly?'' We describe in
  492. %Section~\ref{subsec:enclave-dirs} a way for the bridge authority to
  493. %publish this information without resorting to signing each answer
  494. %individually.
  495. \subsection{Putting them together}
  496. \label{subsec:relay-together}
  497. If a blocked user knows the identity keys of a set of bridge relays, and
  498. he has correct address information for at least one of them, he can use
  499. that one to make a secure connection to the bridge authority and update
  500. his knowledge about the other bridge relays. He can also use it to make
  501. secure connections to the main Tor network and directory servers, so he
  502. can build circuits and connect to the rest of the Internet. All of these
  503. updates happen in the background: from the blocked user's perspective,
  504. he just accesses the Internet via his Tor client like always.
  505. So now we've reduced the problem from how to circumvent the firewall
  506. for all transactions (and how to know that the pages you get have not
  507. been modified by the local attacker) to how to learn about a working
  508. bridge relay.
  509. There's another catch though. We need to make sure that the network
  510. traffic we generate by simply connecting to a bridge relay doesn't stand
  511. out too much.
  512. %The following section describes ways to bootstrap knowledge of your first
  513. %bridge relay, and ways to maintain connectivity once you know a few
  514. %bridge relays.
  515. % (See Section~\ref{subsec:first-bridge} for a discussion
  516. %of exactly what information is sufficient to characterize a bridge relay.)
  517. \subsubsection{Multiple questions about directory authorities}
  518. % This dumps many of the notes I had in one place, because I wanted
  519. % them to get into the tex document, rather than constantly living in
  520. % a separate notes document. They need to be changed and moved, but
  521. % now they're in the right document. -PFS
  522. 9. Bridge directories must not simply be a handful of nodes that
  523. provide the list of bridges. They must flood or otherwise distribute
  524. information out to other Tor nodes as mirrors. That way it becomes
  525. difficult for censors to flood the bridge directory servers with
  526. requests, effectively denying access for others. But, there's lots of
  527. churn and a much larger size than Tor directories. We are forced to
  528. handle the directory scaling problem here much sooner than for the
  529. network in general.
  530. I think some kind of DHT like scheme would work here. A Tor node is
  531. assigned a chunk of the directory. Lookups in the directory should be
  532. via hashes of keys (fingerprints) and that should determine the Tor
  533. nodes responsible. Ordinary directories can publish lists of Tor nodes
  534. responsible for fingerprint ranges. Clients looking to update info on
  535. some bridge will make a Tor connection to one of the nodes responsible
  536. for that address. Instead of shutting down a circuit after getting
  537. info on one address, extend it to another that is responsible for that
  538. address (the node from which you are extending knows you are doing so
  539. anyway). Keep going. This way you can amortize the Tor connection.
  540. 10. We need some way to give new identity keys out to those who need
  541. them without letting those get immediately blocked by authorities. One
  542. way is to give a fingerprint that gets you more fingerprints, as
  543. already described. These are meted out/updated periodically but allow
  544. us to keep track of which sources are compromised: if a distribution
  545. fingerprint repeatedly leads to quickly blocked bridges, it should be
  546. suspect, dropped, etc. Since we're using hashes, there shouldn't be a
  547. correlation with bridge directory mirrors, bridges, portions of the
  548. network observed, etc. It should just be that the authorities know
  549. about that key that leads to new addresses.
  550. This last point is very much like the issues in the valet nodes paper,
  551. which is essentially about blocking resistance wrt exiting the Tor network,
  552. while this paper is concerned with blocking the entering to the Tor network.
  553. In fact the tickets used to connect to the IPo (Introduction Point),
  554. could serve as an example, except that instead of authorizing
  555. a connection to the Hidden Service, it's authorizing the downloading
  556. of more fingerprints.
  557. Also, the fingerprints can follow the hash(q + '1' + cookie) scheme of
  558. that paper (where q = hash(PK + salt) gave the q.onion address). This
  559. allows us to control and track which fingerprint was causing problems.
  560. Note that, unlike many settings, the reputation problem should not be
  561. hard here. If a bridge says it is blocked, then it might as well be.
  562. If an adversary can say that the bridge is blocked wrt
  563. $\mathcal{censor}_i$, then it might as well be, since
  564. $\mathcal{censor}_i$ can presumably then block that bridge if it so
  565. chooses.
  566. 11. How much damage can the adversary do by running nodes in the Tor
  567. network and watching for bridge nodes connecting to it? (This is
  568. analogous to an Introduction Point watching for Valet Nodes connecting
  569. to it.) What percentage of the network do you need to own to do how
  570. much damage. Here the entry-guard design comes in helpfully. So we
  571. need to have bridges use entry-guards, but (cf. 3 above) not use
  572. bridges as entry-guards. Here's a serious tradeoff (again akin to the
  573. ratio of valets to IPos) the more bridges/client the worse the
  574. anonymity of that client. The fewer bridges/client the worse the
  575. blocking resistance of that client.
  576. \section{Hiding Tor's network signatures}
  577. \label{sec:network-signature}
  578. \label{subsec:enclave-dirs}
  579. Currently, Tor uses two protocols for its network communications. The
  580. main protocol uses TLS for encrypted and authenticated communication
  581. between Tor instances. The second protocol is standard HTTP, used for
  582. fetching directory information. All Tor servers listen on their ``ORPort''
  583. for TLS connections, and some of them opt to listen on their ``DirPort''
  584. as well, to serve directory information. Tor servers choose whatever port
  585. numbers they like; the server descriptor they publish to the directory
  586. tells users where to connect.
  587. One format for communicating address information about a bridge relay is
  588. its IP address and DirPort. From there, the user can ask the bridge's
  589. directory cache for an up-to-date copy of its server descriptor, and
  590. learn its current circuit keys, its ORPort, and so on.
  591. However, connecting directly to the directory cache involves a plaintext
  592. HTTP request. A censor could create a network signature for the request
  593. and/or its response, thus preventing these connections. To resolve this
  594. vulnerability, we've modified the Tor protocol so that users can connect
  595. to the directory cache via the main Tor port---they establish a TLS
  596. connection with the bridge as normal, and then send a special ``begindir''
  597. relay command to establish an internal connection to its directory cache.
  598. Therefore a better way to summarize a bridge's address is by its IP
  599. address and ORPort, so all communications between the client and the
  600. bridge will use ordinary TLS. But there are other details that need
  601. more investigation.
  602. What port should bridges pick for their ORPort? We currently recommend
  603. that they listen on port 443 (the default HTTPS port) if they want to
  604. be most useful, because clients behind standard firewalls will have
  605. the best chance to reach them. Is this the best choice in all cases,
  606. or should we encourage some fraction of them pick random ports, or other
  607. ports commonly permitted through firewalls like 53 (DNS) or 110
  608. (POP)? Or perhaps we should use other ports where TLS traffic is
  609. expected, like 993 (IMAPS) or 995 (POP3S). We need more research on our
  610. potential users, and their current and anticipated firewall restrictions.
  611. Furthermore, we need to look at the specifics of Tor's TLS handshake.
  612. Right now Tor uses some predictable strings in its TLS handshakes. For
  613. example, it sets the X.509 organizationName field to ``Tor'', and it puts
  614. the Tor server's nickname in the certificate's commonName field. We
  615. should tweak the handshake protocol so it doesn't rely on any unusual details
  616. in the certificate, yet it remains secure; the certificate itself
  617. should be made to resemble an ordinary HTTPS certificate. We should also try
  618. to make our advertised cipher-suites closer to what an ordinary web server
  619. would support.
  620. Tor's TLS handshake uses two-certificate chains: one certificate
  621. contains the self-signed identity key for
  622. the router, and the second contains a current TLS key, signed by the
  623. identity key. We use these to authenticate that we're talking to the right
  624. router, and to limit the impact of TLS-key exposure. Most (though far from
  625. all) consumer-oriented HTTPS services provide only a single certificate.
  626. These extra certificates may help identify Tor's TLS handshake; instead,
  627. bridges should consider using only a single TLS key certificate signed by
  628. their identity key, and providing the full value of the identity key in an
  629. early handshake cell. More significantly, Tor currently has all clients
  630. present certificates, so that clients are harder to distinguish from servers.
  631. But in a blocking-resistance environment, clients should not present
  632. certificates at all.
  633. Last, what if the adversary starts observing the network traffic even
  634. more closely? Even if our TLS handshake looks innocent, our traffic timing
  635. and volume still look different than a user making a secure web connection
  636. to his bank. The same techniques used in the growing trend to build tools
  637. to recognize encrypted Bittorrent traffic~\cite{bt-traffic-shaping}
  638. could be used to identify Tor communication and recognize bridge
  639. relays. Rather than trying to look like encrypted web traffic, we may be
  640. better off trying to blend with some other encrypted network protocol. The
  641. first step is to compare typical network behavior for a Tor client to
  642. typical network behavior for various other protocols. This statistical
  643. cat-and-mouse game is made more complex by the fact that Tor transports a
  644. variety of protocols, and we'll want to automatically handle web browsing
  645. differently from, say, instant messaging.
  646. % Tor cells are 512 bytes each. So TLS records will be roughly
  647. % multiples of this size? How bad is this? -RD
  648. % Look at ``Inferring the Source of Encrypted HTTP Connections''
  649. % by Marc Liberatore and Brian Neil Levine (CCS 2006)
  650. % They substantially flesh out the numbers for the web fingerprinting
  651. % attack. -PS
  652. % Yes, but I meant detecting the signature of Tor traffic itself, not
  653. % learning what websites we're going to. I wouldn't be surprised to
  654. % learn that these are related problems, but it's not obvious to me. -RD
  655. \subsection{Identity keys as part of addressing information}
  656. We have described a way for the blocked user to bootstrap into the
  657. network once he knows the IP address and ORPort of a bridge. What about
  658. local spoofing attacks? That is, since we never learned an identity
  659. key fingerprint for the bridge, a local attacker could intercept our
  660. connection and pretend to be the bridge we had in mind. It turns out
  661. that giving false information isn't that bad---since the Tor client
  662. ships with trusted keys for the bridge directory authority and the Tor
  663. network directory authorities, the user can learn whether he's being
  664. given a real connection to the bridge authorities or not. (After all,
  665. if the adversary intercepts every connection the user makes and gives
  666. him a bad connection each time, there's nothing we can do.)
  667. What about anonymity-breaking attacks from observing traffic, if the
  668. blocked user doesn't start out knowing the identity key of his intended
  669. bridge? The vulnerabilities aren't so bad in this case either---the
  670. adversary could do similar attacks just by monitoring the network
  671. traffic.
  672. % cue paper by steven and george
  673. Once the Tor client has fetched the bridge's server descriptor, it should
  674. remember the identity key fingerprint for that bridge relay. Thus if
  675. the bridge relay moves to a new IP address, the client can query the
  676. bridge directory authority to look up a fresh server descriptor using
  677. this fingerprint.
  678. So we've shown that it's \emph{possible} to bootstrap into the network
  679. just by learning the IP address and ORPort of a bridge, but are there
  680. situations where it's more convenient or more secure to learn the bridge's
  681. identity fingerprint as well as instead, while bootstrapping? We keep
  682. that question in mind as we next investigate bootstrapping and discovery.
  683. \section{Discovering and maintaining working bridge relays}
  684. \label{sec:discovery}
  685. Tor's modular design means that we can develop a better relay component
  686. independently of developing the discovery component. This modularity's
  687. great promise is that we can pick any discovery approach we like; but the
  688. unfortunate fact is that we have no magic bullet for discovery. We're
  689. in the same arms race as all the other designs we described in
  690. Section~\ref{sec:related}.
  691. In this section we describe three approaches to adding discovery
  692. components for our design. Note that we should deploy all the schemes
  693. at once---bridges and blocked users can then use the discovery approach
  694. that is most appropriate for their situation.
  695. \subsection{Independent bridges, no central discovery}
  696. The first design is simply to have no centralized discovery component at
  697. all. Volunteers run bridges, and we assume they have some blocked users
  698. in mind and communicate their address information to them out-of-band
  699. (for example, through Gmail). This design allows for small personal
  700. bridges that have only one or a handful of users in mind, but it can
  701. also support an entire community of users. For example, Citizen Lab's
  702. upcoming Psiphon single-hop proxy tool~\cite{psiphon} plans to use this
  703. \emph{social network} approach as its discovery component.
  704. There are several ways to do bootstrapping in this design. In the simple
  705. case, the operator of the bridge informs each chosen user about his
  706. bridge's address information and/or keys. A different approach involves
  707. blocked users introducing new blocked users to the bridges they know.
  708. That is, somebody in the blocked area can pass along a bridge's address to
  709. somebody else they trust. This scheme brings in appealing but complex game
  710. theoretic properties: the blocked user making the decision has an incentive
  711. only to delegate to trustworthy people, since an adversary who learns
  712. the bridge's address and filters it makes it unavailable for both of them.
  713. Note that a central set of bridge directory authorities can still be
  714. compatible with a decentralized discovery process. That is, how users
  715. first learn about bridges is entirely up to the bridges, but the process
  716. of fetching up-to-date descriptors for them can still proceed as described
  717. in Section~\ref{sec:bridges}. Of course, creating a central place that
  718. knows about all the bridges may not be smart, especially if every other
  719. piece of the system is decentralized. Further, if a user only knows
  720. about one bridge and he loses track of it, it may be quite a hassle to
  721. reach the bridge authority. We address these concerns next.
  722. \subsection{Families of bridges, no central discovery}
  723. Because the blocked users are running our software too, we have many
  724. opportunities to improve usability or robustness. Our second design builds
  725. on the first by encouraging volunteers to run several bridges at once
  726. (or coordinate with other bridge volunteers), such that some fraction
  727. of the bridges are likely to be available at any given time.
  728. The blocked user's Tor client would periodically fetch an updated set of
  729. recommended bridges from any of the working bridges. Now the client can
  730. learn new additions to the bridge pool, and can expire abandoned bridges
  731. or bridges that the adversary has blocked, without the user ever needing
  732. to care. To simplify maintenance of the community's bridge pool, each
  733. community could run its own bridge directory authority---reachable via
  734. the available bridges, and also mirrored at each bridge.
  735. \subsection{Public bridges with central discovery}
  736. What about people who want to volunteer as bridges but don't know any
  737. suitable blocked users? What about people who are blocked but don't
  738. know anybody on the outside? Here we describe a way to make use of these
  739. \emph{public bridges} in a way that still makes it hard for the attacker
  740. to learn all of them.
  741. The basic idea is to divide public bridges into a set of buckets based on
  742. identity key, where each bucket has a different policy for distributing
  743. its bridge addresses to users. Each of these \emph{distribution policies}
  744. is designed to exercise a different scarce resource or property of
  745. the user.
  746. How do we divide bridges into buckets such that they're evenly distributed
  747. and the allocation is hard to influence or predict, but also in a way
  748. that's amenable to creating more buckets later on without reshuffling
  749. all the bridges? We compute the bucket for a given bridge by hashing the
  750. bridge's identity key along with a secret that only the bridge authority
  751. knows: the first $n$ bits of this hash dictate the bucket number,
  752. where $n$ is a parameter that describes how many buckets we want at this
  753. point. We choose $n=3$ to start, so we have 8 buckets available; but as
  754. we later invent new distribution policies, we can increment $n$ to split
  755. the 8 into 16 buckets. Since a bridge can't predict the next bit in its
  756. hash, it can't anticipate which identity key will correspond to a certain
  757. bucket when the buckets are split. Further, since the bridge authority
  758. doesn't provide any feedback to the bridge about which bucket it's in,
  759. an adversary signing up bridges to fill a certain bucket will be slowed.
  760. % This algorithm is not ideal. When we split buckets, each existing
  761. % bucket is cut in half, where half the bridges remain with the
  762. % old distribution policy, and half will be under what the new one
  763. % is. So the new distribution policy inherits a bunch of blocked
  764. % bridges if the old policy was too loose, or a bunch of unblocked
  765. % bridges if its policy was still secure. -RD
  766. The first distribution policy (used for the first bucket) publishes bridge
  767. addresses in a time-release fashion. The bridge authority divides the
  768. available bridges into partitions which are deterministically available
  769. only in certain time windows. That is, over the course of a given time
  770. slot (say, an hour), each requestor is given a random bridge from within
  771. that partition. When the next time slot arrives, a new set of bridges
  772. are available for discovery. Thus a bridge is always available when a new
  773. user arrives, but to learn about all bridges the attacker needs to fetch
  774. the new addresses at every new time slot. By varying the length of the
  775. time slots, we can make it harder for the attacker to guess when to check
  776. back. We expect these bridges will be the first to be blocked, but they'll
  777. help the system bootstrap until they \emph{do} get blocked. Further,
  778. remember that we're dealing with different blocking regimes around the
  779. world that will progress at different rates---so this bucket will still
  780. be useful to some users even as the arms race progresses.
  781. The second distribution policy publishes bridge addresses based on the IP
  782. address of the requesting user. Specifically, the bridge authority will
  783. divide the available bridges in the bucket into a bunch of partitions
  784. (as in the first distribution scheme), hash the requestor's IP address
  785. with a secret of its own (as in the above allocation scheme for creating
  786. buckets), and give the requestor a random bridge from the appropriate
  787. partition. To raise the bar, we should discard the last octet of the
  788. IP address before inputting it to the hash function, so an attacker
  789. who only controls a ``/24'' address only counts as one user. A large
  790. attacker like China will still be able to control many addresses, but
  791. the hassle of needing to establish connections from each network (or
  792. spoof TCP connections) may still slow them down. (We could also imagine
  793. a policy that combines the time-based and location-based policies to
  794. further constrain and rate-limit the available bridge addresses.)
  795. The third policy is based on Circumventor's discovery strategy. Realizing
  796. that its adoption will remain limited without some central coordination
  797. mechanism, the Circumventor project has started a mailing list to
  798. distribute new proxy addresses every few days. From experimentation it
  799. seems they have concluded that sending updates every three or four days
  800. is sufficient to stay ahead of the current attackers. We could give out
  801. bridge addresses from the third bucket in a similar fashion
  802. The fourth policy provides an alternative approach to a mailing list:
  803. users provide an email address, and receive an automated response
  804. listing an available bridge address. We could limit one response per
  805. email address. To further rate limit queries, we could require a CAPTCHA
  806. solution~\cite{captcha} in each case too. In fact, we wouldn't need to
  807. implement the CAPTCHA on our side: if we only deliver bridge addresses
  808. to Yahoo or GMail addresses, we can leverage the rate-limiting schemes
  809. that other parties already impose for account creation.
  810. The fifth policy ties in
  811. ...
  812. reputation system
  813. Pick some seeds---trusted people in the blocked area---and give
  814. them each a few hundred bridge addresses. Run a website next to the
  815. bridge authority, where they can log in (they only need persistent
  816. pseudonyms). Give them tokens slowly over time. They can use these
  817. tokens to delegate trust to other people they know. The tokens can
  818. be exchanged for new accounts on the website.
  819. Accounts in ``good standing'' accrue new bridge addresses and new
  820. tokens.
  821. This is great, except how do we decide that an account is in good
  822. standing? One answer is to measure based on whether the bridge addresses
  823. we give it end up blocked. But how do we decide if they get blocked?
  824. Other questions below too.
  825. \ref{sec:accounts}
  826. Buckets six through eight are held in reserve, in case our currently
  827. deployed tricks all fail at once---so we can adapt and move to
  828. new approaches quickly, and have some bridges available for the new
  829. schemes. (Bridges that sign up and don't get used yet may be unhappy that
  830. they're not being used; but this is a transient problem: if bridges are
  831. on by default, nobody will mind not being used yet.)
  832. \subsubsection{Bootstrapping: finding your first bridge.}
  833. \label{subsec:first-bridge}
  834. How do users find their first public bridge, so they can reach the
  835. bridge authority to learn more?
  836. Most government firewalls are not perfect. That is, they allow connections to
  837. Google cache or some open proxy servers, or they let file-sharing traffic or
  838. Skype or World-of-Warcraft connections through. We assume that the
  839. users have some mechanism for bypassing the firewall initially.
  840. For users who can't use any of these techniques, hopefully they know
  841. a friend who can---for example, perhaps the friend already knows some
  842. bridge relay addresses.
  843. (If they can't get around it at all, then we can't help them---they
  844. should go meet more people.)
  845. Is it useful to load balance which bridges are handed out? The above
  846. bucket concept makes some bridges wildly popular and others less so.
  847. But I guess that's the point.
  848. Families of bridges: give out 4 or 8 at once, bound together.
  849. \subsection{Advantages of deploying all solutions at once}
  850. For once we're not in the position of the defender: we don't have to
  851. defend against every possible filtering scheme, we just have to defend
  852. against at least one.
  853. \subsection{Remaining unsorted notes}
  854. In the first subsection we describe how to find a first bridge.
  855. Thus they can reach the BDA. From here we either assume a social
  856. network or other mechanism for learning IP:dirport or key fingerprints
  857. as above, or we assume an account server that allows us to limit the
  858. number of new bridge relays an external attacker can discover.
  859. Going to be an arms race. Need a bag of tricks. Hard to say
  860. which ones will work. Don't spend them all at once.
  861. Some techniques are sufficient to get us an IP address and a port,
  862. and others can get us IP:port:key. Lay out some plausible options
  863. for how users can bootstrap into learning their first bridge.
  864. attack: adversary can reconstruct your social network by learning who
  865. knows which bridges.
  866. %\section{The account / reputation system}
  867. \section{Social networks with directory-side support}
  868. \label{sec:accounts}
  869. Perhaps each bridge should be known by a single bridge directory
  870. authority. This makes it easier to trace which users have learned about
  871. it, so easier to blame or reward. It also makes things more brittle,
  872. since loss of that authority means its bridges aren't advertised until
  873. they switch, and means its bridge users are sad too.
  874. (Need a slick hash algorithm that will map our identity key to a
  875. bridge authority, in a way that's sticky even when we add bridge
  876. directory authorities, but isn't sticky when our authority goes
  877. away. Does this exist?)
  878. \subsection{Discovery based on social networks}
  879. A token that can be exchanged at the bridge authority (assuming you
  880. can reach it) for a new bridge address.
  881. The account server runs as a Tor controller for the bridge authority.
  882. Users can establish reputations, perhaps based on social network
  883. connectivity, perhaps based on not getting their bridge relays blocked,
  884. Probably the most critical lesson learned in past work on reputation
  885. systems in privacy-oriented environments~\cite{rep-anon} is the need for
  886. verifiable transactions. That is, the entity computing and advertising
  887. reputations for participants needs to actually learn in a convincing
  888. way that a given transaction was successful or unsuccessful.
  889. (Lesson from designing reputation systems~\cite{rep-anon}: easy to
  890. reward good behavior, hard to punish bad behavior.
  891. \subsection{How do we know if a bridge relay has been blocked?}
  892. We need some mechanism for testing reachability from inside the
  893. blocked area.
  894. The easiest answer is for certain users inside the area to sign up as
  895. testing relays, and then we can route through them and see if it works.
  896. First problem is that different network areas block different net masks,
  897. and it will likely be hard to know which users are in which areas. So
  898. if a bridge relay isn't reachable, is that because of a network block
  899. somewhere, because of a problem at the bridge relay, or just a temporary
  900. outage?
  901. Second problem is that if we pick random users to test random relays, the
  902. adversary should sign up users on the inside, and enumerate the relays
  903. we test. But it seems dangerous to just let people come forward and
  904. declare that things are blocked for them, since they could be tricking
  905. us. (This matters even moreso if our reputation system above relies on
  906. whether things get blocked to punish or reward.)
  907. Another answer is not to measure directly, but rather let the bridges
  908. report whether they're being used. If they periodically report to their
  909. bridge directory authority how much use they're seeing, the authority
  910. can make smart decisions from there.
  911. If they install a geoip database, they can periodically report to their
  912. bridge directory authority which countries they're seeing use from. This
  913. might help us to track which countries are making use of Ramp, and can
  914. also let us learn about new steps the adversary has taken in the arms
  915. race. (If the bridges don't want to install a whole geoip subsystem, they
  916. can report samples of the /24 network for their users, and the authorities
  917. can do the geoip work. This tradeoff has clear downsides though.)
  918. Worry: adversary signs up a bunch of already-blocked bridges. If we're
  919. stingy giving out bridges, users in that country won't get useful ones.
  920. (Worse, we'll blame the users when the bridges report they're not
  921. being used?)
  922. Worry: the adversary could choose not to block bridges but just record
  923. connections to them. So be it, I guess.
  924. \subsection{How to learn how well the whole idea is working}
  925. We need some feedback mechanism to learn how much use the bridge network
  926. as a whole is actually seeing. Part of the reason for this is so we can
  927. respond and adapt the design; part is because the funders expect to see
  928. progress reports.
  929. The above geoip-based approach to detecting blocked bridges gives us a
  930. solution though.
  931. \section{Security considerations}
  932. \label{sec:security}
  933. \subsection{Possession of Tor in oppressed areas}
  934. Many people speculate that installing and using a Tor client in areas with
  935. particularly extreme firewalls is a high risk---and the risk increases
  936. as the firewall gets more restrictive. This is probably true, but there's
  937. a counter pressure as well: as the firewall gets more restrictive, more
  938. ordinary people use Tor for more mainstream activities, such as learning
  939. about Wall Street prices or looking at pictures of women's ankles. So
  940. if the restrictive firewall pushes up the number of Tor users, then the
  941. ``typical'' Tor user becomes more mainstream.
  942. Hard to say which of these pressures will ultimately win out.
  943. ...
  944. % Nick can rewrite/elaborate on this section?
  945. \subsection{Observers can tell who is publishing and who is reading}
  946. \label{subsec:upload-padding}
  947. Should bridge users sometimes send bursts of long-range drop cells?
  948. \subsection{Anonymity effects from acting as a bridge relay}
  949. Against some attacks, relaying traffic for others can improve anonymity. The
  950. simplest example is an attacker who owns a small number of Tor servers. He
  951. will see a connection from the bridge, but he won't be able to know
  952. whether the connection originated there or was relayed from somebody else.
  953. There are some cases where it doesn't seem to help: if an attacker can
  954. watch all of the bridge's incoming and outgoing traffic, then it's easy
  955. to learn which connections were relayed and which started there. (In this
  956. case he still doesn't know the final destinations unless he is watching
  957. them too, but in this case bridges are no better off than if they were
  958. an ordinary client.)
  959. There are also some potential downsides to running a bridge. First, while
  960. we try to make it hard to enumerate all bridges, it's still possible to
  961. learn about some of them, and for some people just the fact that they're
  962. running one might signal to an attacker that they place a high value
  963. on their anonymity. Second, there are some more esoteric attacks on Tor
  964. relays that are not as well-understood or well-tested---for example, an
  965. attacker may be able to ``observe'' whether the bridge is sending traffic
  966. even if he can't actually watch its network, by relaying traffic through
  967. it and noticing changes in traffic timing~\cite{attack-tor-oak05}. On
  968. the other hand, it may be that limiting the bandwidth the bridge is
  969. willing to relay will allow this sort of attacker to determine if it's
  970. being used as a bridge but not whether it is adding traffic of its own.
  971. It is an open research question whether the benefits outweigh the risks. A
  972. lot of the decision rests on which attacks the users are most worried
  973. about. For most users, we don't think running a bridge relay will be
  974. that damaging.
  975. Need to examine how entry guards fit in. If the blocked user doesn't use
  976. the bridge's entry guards, then the bridge doesn't gain as much cover
  977. benefit. If he does, first how does that actually work, and second is
  978. it turtles all the way down (need to use the guard's guards, ...)?
  979. \subsection{Trusting local hardware: Internet cafes and LiveCDs}
  980. \label{subsec:cafes-and-livecds}
  981. Assuming that users have their own trusted hardware is not
  982. always reasonable.
  983. For Internet cafe Windows computers that let you attach your own USB key,
  984. a USB-based Tor image would be smart. There's Torpark, and hopefully
  985. there will be more thoroughly analyzed options down the road. Worries
  986. about hardware or
  987. software keyloggers and other spyware---and physical surveillance.
  988. If the system lets you boot from a CD or from a USB key, you can gain
  989. a bit more security by bringing a privacy LiveCD with you. Hardware
  990. keyloggers and physical surveillance still a worry. LiveCDs also useful
  991. if it's your own hardware, since it's easier to avoid leaving breadcrumbs
  992. everywhere.
  993. \subsection{Forward compatibility and retiring bridge authorities}
  994. Eventually we'll want to change the identity key and/or location
  995. of a bridge authority. How do we do this mostly cleanly?
  996. \subsection{The trust chain}
  997. \label{subsec:trust-chain}
  998. Tor's ``public key infrastructure'' provides a chain of trust to
  999. let users verify that they're actually talking to the right servers.
  1000. There are four pieces to this trust chain.
  1001. First, when Tor clients are establishing circuits, at each step
  1002. they demand that the next Tor server in the path prove knowledge of
  1003. its private key~\cite{tor-design}. This step prevents the first node
  1004. in the path from just spoofing the rest of the path. Second, the
  1005. Tor directory authorities provide a signed list of servers along with
  1006. their public keys---so unless the adversary can control a threshold
  1007. of directory authorities, he can't trick the Tor client into using other
  1008. Tor servers. Third, the location and keys of the directory authorities,
  1009. in turn, is hard-coded in the Tor source code---so as long as the user
  1010. got a genuine version of Tor, he can know that he is using the genuine
  1011. Tor network. And last, the source code and other packages are signed
  1012. with the GPG keys of the Tor developers, so users can confirm that they
  1013. did in fact download a genuine version of Tor.
  1014. But how can a user in an oppressed country know that he has the correct
  1015. key fingerprints for the developers? As with other security systems, it
  1016. ultimately comes down to human interaction. The keys are signed by dozens
  1017. of people around the world, and we have to hope that our users have met
  1018. enough people in the PGP web of trust~\cite{pgp-wot} that they can learn
  1019. the correct keys. For users that aren't connected to the global security
  1020. community, though, this question remains a critical weakness.
  1021. % XXX make clearer the trust chain step for bridge directory authorities
  1022. \subsection{Security through obscurity: publishing our design}
  1023. Many other schemes like dynaweb use the typical arms race strategy of
  1024. not publishing their plans. Our goal here is to produce a design---a
  1025. framework---that can be public and still secure. Where's the tradeoff?
  1026. \section{Performance improvements}
  1027. \label{sec:performance}
  1028. \subsection{Fetch server descriptors just-in-time}
  1029. I guess we should encourage most places to do this, so blocked
  1030. users don't stand out.
  1031. network-status and directory optimizations. caching better. partitioning
  1032. issues?
  1033. \section{Maintaining reachability}
  1034. \subsection{How many bridge relays should you know about?}
  1035. If they're ordinary Tor users on cable modem or DSL, many of them will
  1036. disappear and/or move periodically. How many bridge relays should a
  1037. blockee know
  1038. about before he's likely to have at least one reachable at any given point?
  1039. How do we factor in a parameter for "speed that his bridges get discovered
  1040. and blocked"?
  1041. The related question is: if the bridge relays change IP addresses
  1042. periodically, how often does the bridge user need to "check in" in order
  1043. to keep from being cut out of the loop?
  1044. \subsection{Cablemodem users don't provide important websites}
  1045. \label{subsec:block-cable}
  1046. ...so our adversary could just block all DSL and cablemodem networks,
  1047. and for the most part only our bridge relays would be affected.
  1048. The first answer is to aim to get volunteers both from traditionally
  1049. ``consumer'' networks and also from traditionally ``producer'' networks.
  1050. The second answer (not so good) would be to encourage more use of consumer
  1051. networks for popular and useful websites. (But P2P exists; minor websites
  1052. exist; gaming exists; IM exists; ...)
  1053. Other attack: China pressures Verizon to discourage its users from
  1054. running bridges.
  1055. \subsection{Scanning-resistance}
  1056. If it's trivial to verify that we're a bridge, and we run on a predictable
  1057. port, then it's conceivable our attacker would scan the whole Internet
  1058. looking for bridges. (In fact, he can just scan likely networks like
  1059. cablemodem and DSL services---see Section~\ref{block-cable} for a related
  1060. attack.) It would be nice to slow down this attack. It would
  1061. be even nicer to make it hard to learn whether we're a bridge without
  1062. first knowing some secret.
  1063. Password protecting the bridges.
  1064. Could provide a password to the bridge user. He provides a nonced hash of
  1065. it or something when he connects. We'd need to give him an ID key for the
  1066. bridge too, and wait to present the password until we've TLSed, else the
  1067. adversary can pretend to be the bridge and MITM him to learn the password.
  1068. We could some kind of ID-based knocking protocol, or we could act like an
  1069. unconfigured HTTPS server if treated like one.
  1070. We can assume that the attacker can easily recognize https connections
  1071. to unknown servers. It can then attempt to connect to them and block
  1072. connections to servers that seem suspicious. It may be that password
  1073. protected web sites will not be suspicious in general, in which case
  1074. that may be the easiest way to give controlled access to the bridge.
  1075. If such sites that have no other overt features are automatically
  1076. blocked when detected, then we may need to be more subtle.
  1077. Possibilities include serving an innocuous web page if a TLS encrypted
  1078. request is received without the authorization needed to access the Tor
  1079. network and only responding to a requested access to the Tor network
  1080. of proper authentication is given. If an unauthenticated request to
  1081. access the Tor network is sent, the bridge should respond as if
  1082. it has received a message it does not understand (as would be the
  1083. case were it not a bridge).
  1084. \subsection{How to motivate people to run bridge relays}
  1085. One of the traditional ways to get people to run software that benefits
  1086. others is to give them motivation to install it themselves. An often
  1087. suggested approach is to install it as a stunning screensaver so everybody
  1088. will be pleased to run it. We take a similar approach here, by leveraging
  1089. the fact that these users are already interested in protecting their
  1090. own Internet traffic, so they will install and run the software.
  1091. Make all Tor users become bridges if they're reachable---needs more work
  1092. on usability first, but we're making progress.
  1093. Also, we can make a snazzy network graph with Vidalia that emphasizes
  1094. the connections the bridge user is currently relaying. (Minor anonymity
  1095. implications, but hey.) (In many cases there won't be much activity,
  1096. so this may backfire. Or it may be better suited to full-fledged Tor
  1097. servers.)
  1098. \subsection{What if the clients can't install software?}
  1099. [this section should probably move to the related work section,
  1100. or just disappear entirely.]
  1101. Bridge users without Tor software
  1102. Bridge relays could always open their socks proxy. This is bad though,
  1103. first
  1104. because bridges learn the bridge users' destinations, and second because
  1105. we've learned that open socks proxies tend to attract abusive users who
  1106. have no idea they're using Tor.
  1107. Bridges could require passwords in the socks handshake (not supported
  1108. by most software including Firefox). Or they could run web proxies
  1109. that require authentication and then pass the requests into Tor. This
  1110. approach is probably a good way to help bootstrap the Psiphon network,
  1111. if one of its barriers to deployment is a lack of volunteers willing
  1112. to exit directly to websites. But it clearly drops some of the nice
  1113. anonymity and security features Tor provides.
  1114. A hybrid approach where the user gets his anonymity from Tor but his
  1115. software-less use from a web proxy running on a trusted machine on the
  1116. free side.
  1117. \subsection{Publicity attracts attention}
  1118. \label{subsec:publicity}
  1119. Many people working on this field want to publicize the existence
  1120. and extent of censorship concurrently with the deployment of their
  1121. circumvention software. The easy reason for this two-pronged push is
  1122. to attract volunteers for running proxies in their systems; but in many
  1123. cases their main goal is not to build the software, but rather to educate
  1124. the world about the censorship. The media also tries to do its part by
  1125. broadcasting the existence of each new circumvention system.
  1126. But at the same time, this publicity attracts the attention of the
  1127. censors. We can slow down the arms race by not attracting as much
  1128. attention, and just spreading by word of mouth. If our goal is to
  1129. establish a solid social network of bridges and bridge users before
  1130. the adversary gets involved, does this attention tradeoff work to our
  1131. advantage?
  1132. \subsection{The Tor website: how to get the software}
  1133. \section{Future designs}
  1134. \subsection{Bridges inside the blocked network too}
  1135. Assuming actually crossing the firewall is the risky part of the
  1136. operation, can we have some bridge relays inside the blocked area too,
  1137. and more established users can use them as relays so they don't need to
  1138. communicate over the firewall directly at all? A simple example here is
  1139. to make new blocked users into internal bridges also---so they sign up
  1140. on the BDA as part of doing their query, and we give out their addresses
  1141. rather than (or along with) the external bridge addresses. This design
  1142. is a lot trickier because it brings in the complexity of whether the
  1143. internal bridges will remain available, can maintain reachability with
  1144. the outside world, etc.
  1145. Hidden services as bridges. Hidden services as bridge directory authorities.
  1146. \section{Conclusion}
  1147. a technical solution won't solve the whole problem. after all, china's
  1148. firewall is *socially* very successful, even if technologies exist to
  1149. get around it.
  1150. but having a strong technical solution is still useful as a piece of the
  1151. puzzle.
  1152. \bibliographystyle{plain} \bibliography{tor-design}
  1153. \appendix
  1154. \section{Counting Tor users by country}
  1155. \label{app:geoip}
  1156. \end{document}
  1157. ship geoip db to bridges. they look up users who tls to them in the db,
  1158. and upload a signed list of countries and number-of-users each day. the
  1159. bridge authority aggregates them and publishes stats.
  1160. bridge relays have buddies
  1161. they ask a user to test the reachability of their buddy.
  1162. leaks O(1) bridges, but not O(n).
  1163. we should not be blockable by ordinary cisco censorship features.
  1164. that is, if they want to block our new design, they will need to
  1165. add a feature to block exactly this.
  1166. strategically speaking, this may come in handy.
  1167. Bridges come in clumps of 4 or 8 or whatever. If you know one bridge
  1168. in a clump, the authority will tell you the rest. Now bridges can
  1169. ask users to test reachability of their buddies.
  1170. Giving out clumps helps with dynamic IP addresses too. Whether it
  1171. should be 4 or 8 depends on our churn.
  1172. the account server. let's call it a database, it doesn't have to
  1173. be a thing that human interacts with.
  1174. rate limiting mechanisms:
  1175. energy spent. captchas. relaying traffic for others?
  1176. send us \$10, we'll give you an account
  1177. so how do we reward people for being good?