490

This question is (inspired by)/(shamefully stolen from) a similar question at MathOverflow, but I expect the answers here will be quite different.

We all have favorite papers in our own respective areas of theory. Every once in a while, one finds a paper so astounding (e.g., important, compelling, deceptively simple, etc.) that one wants to share it with everyone. So list these papers here! They don't have to be from theoretical computer science -- anything that you think might appeal to the community is a fine answer.

You can give as many answers as you want; please put one paper per answer! Also, notice this is community wiki, so vote on everything you like!

(Note there has been a previous question about papers in recursion-theoretic complexity but that is quite specialized.)

Ryan Williams
  • 27,465
  • 7
  • 115
  • 164
  • 71
    In the answers, I'd like to see more emphasis on whether it really is a good idea to read the original paper nowadays (or if it makes much more sense to read a modern textbook exposition of it). I have too often seen TCS papers that are truly seminal, but I'd rather save my colleagues from the pain of trying to decipher the original write-up – which is far too often a hastily-written 10-page conference abstract, with references to a "full version" that never appeared... – Jukka Suomela Sep 12 '10 at 09:46
  • 7
    Yes, I hope it is clear that papers of this type are not good for the list (if you want to share it with everyone, then it shouldn't be a pain to read) – Ryan Williams Sep 12 '10 at 16:22
  • 32
    Too many people are just posting one-liners. Any one can post 100s of unique papers without putting any thought into it. Please post why you think everyone should read those papers. This means justifying why they should read that paper instead of someone else's writeup of that result, and what is so awesome about the paper that everyone should read it. – Robin Kothari Sep 16 '10 at 19:18
  • Good question. My opinion is that if you want to understand the minds of the inventors, and possibly understand how to invent things, you have to read their own words. The more you labor, the closer you get to their actual thought process. – ixtmixilix Sep 26 '10 at 22:07
  • 1
  • There is a book along similar lines: Ideas That Created the Future: Classic Papers of Computer Science by Harry R Lewis. – Nishant Jun 10 '23 at 14:00

72 Answers72

176

"A mathematical theory of communication" by Claude Shannon, classics of information theory. Very readable.

(Mirror)

Ekene E.
  • 121
  • 4
Grigory Yaroslavtsev
  • 1,468
  • 1
  • 14
  • 25
157

The 1936 paper that arguably started computer science itself:

  • Alan Turing, "On Computable Numbers, with an Application to the Entscheidungsproblem", Proceedings of the London Mathematical Society s2-42, 230–265, 1937. doi: 10.1112/plms/s2-42.1.230

In just 36 pages, Turing formulates (but does not name) the Turing Machine, recasts Gödel's famous First Incompleteness Theorem in terms of computation, describes the concept of universality, and in the appendix shows that computability by Turing machines is equivalent to computability by $\lambda$-definable functions (as studied by Church and Kleene).

vorushin
  • 101
  • 3
Henry Yuen
  • 3,798
  • 1
  • 21
  • 33
131

Ken Thompson's "Reflections on Trusting Trust". Short, sweet, and mind-blowing.

Jeffε
  • 23,129
  • 10
  • 96
  • 163
  • 6
    Also, very approachable. I read it quite some time ago, when I had basically no CS background, no programming experience and didn't even know what a compiler was. – Jörg W Mittag Sep 13 '10 at 19:05
  • just read it -- awesome indeed! – Lev Reyzin Sep 15 '10 at 01:15
  • 1
    "Last week, Googler Ken Thompson was awarded the Japan Prize in Information and Communications for his early work on the UNIX operating system." (src: Buzz post from Life at Google) – Sebastián Grignoli May 26 '11 at 05:23
  • 5
    I would think this paper would be pretty difficult to digest without at least knowing what a compiler is. – Fixee Sep 02 '11 at 04:26
  • 3
    In the paper, I think figures 2.1 and 2.2 are swapped. – Dennis Sep 10 '13 at 06:50
  • 3
    Disagree - nothing awesome or mindblowing in this paper. TL;DR 6 pages from mid-80s about "need to change criminal code to start punishing hackers [just like thieves or burglars]". O yeah, mentions a quine, without calling it by name. – c69 Oct 29 '17 at 22:59
  • 1
    I agree with c69. This is currently the third top rated answer, but there's an enormous interestingness gap between it and the top two papers (by Shannon and Turing). Those two created new branches of mathematics. This one is... cute, I guess. – benrg Aug 13 '20 at 19:11
  • This paper is really nothing more than a high level introduction to the concept of a quine that goes absolutely off the rails at the end. Thompson's conclusion is that the existence of quines make trojan horses too hard to detect, so he demands the criminal code be updated and pop culture be engineered such that "[t]he act of breaking into a computer system has to have the same social stigma as breaking into a neighbor's house". Ridiculous. – rw-nandemo Jul 09 '21 at 20:56
98

What Every Computer Scientist Should Know About Floating-Point Arithmetic

This paper explains and reinforces the notion that floating point isn't magic. It explains overflow, underflow, what denormalized numbers are, what NaNs are, what inf is, and all the things these imply. After reading this paper, you'll know why a == a + 1.0 can be true, why a==a can be false, why running your code on two different machines can give you two different answers, why summing numbers in a different order can give you an order of magnitude difference and all the wacky stuff that happens in the world of mapping an uncountably infinite set of numbers onto a countably finite set.

An edited version is also available on the web.

Bacon
  • 101
  • 3
93

Keshav's How to Read a Paper. You can also download the paper from here.

philosodad
  • 101
  • 1
  • 5
71

Paths, Trees and Flowers by J. Edmonds. This paper about classic combinatorial optimization problem is not only well written, but also states that the notion of "polynomial-time algorithms" is essentially a synonym for efficiency.

Joshua Grochow
  • 37,260
  • 4
  • 129
  • 228
ilyaraz
  • 1,569
  • 18
  • 33
62

Reducibility Among Combinatorial Problems by Richard Karp. The paper contains what's often referred to as Karp's "original 21 NP-complete problems." In many ways, this paper truly motivated the study of NP-completeness by demonstrating its applicability to a wider domain. Very readable.

klingt.net
  • 101
  • 4
Daniel Apon
  • 6,001
  • 1
  • 37
  • 53
54

Hartmanis and Stearns, "On the computational complexity of algorithms", Transactions of the American Mathematical Society 117: 285–306 (1965)

This was the first paper that took the study of time complexity seriously, and surely was the primary impetus for Hartmanis and Stearns' joint Turing award. While their initial definitions are not quite what we use today, the paper remains extremely readable. You really get the feeling of how things were in the old "Wild West" frontier of the 60's.

Ryan Williams
  • 27,465
  • 7
  • 115
  • 164
52

Quantum Mechanical Computers (PDF) by Richard Feynman.

He introduces the idea of quantum computation, describes quantum circuits, explains how classical circuits can be simulated by quantum circuits, and shows how quantum circuits can compute functions without lots of garbage qubits (using uncomputation).

He then shows how any classical circuit can be encoded into a time-independent Hamiltonian! His proof goes through for quantum circuits too, therefore showing that time evolving Hamiltonians is BQP-hard! His Hamiltonian construction is also used in the proof of the quantum version of the Cook-Levin theorem, proved by Kitaev, which shows that k-local Hamiltonian is QMA-complete.

Jarwain
  • 3
  • 3
Robin Kothari
  • 13,617
  • 2
  • 60
  • 116
  • The link isn't valid. Do you have another source? edit> Searched on google : http://www.wjzeng.net/Ref/Feynman_QuantumMechanicalComputers.pdf Is it this one? – Klaim Oct 01 '10 at 14:21
  • That's the one. I added a new link and a link to it's page on the publisher's website. – Robin Kothari Oct 01 '10 at 23:27
  • Did the notions of BQP and QMA exist when Feynman wrote this paper? Or is there a recent paper which draws this connection? Any reference/exposition of this fact that k-local Hamiltonian is QMA complete? – Student Jan 03 '16 at 10:54
50

Expander graphs and their applications, S. Hoory, N. Linial, and A. Wigderson is an extremely nice survey on expander graphs. No surprise that it won the 2008 AMS Conant Prize.

I want to recall that expander graphs are the key ingredient in recent breakthroughs in TCS, eg.

and not so recent:

VS.
  • 539
  • 3
  • 15
Dai Le
  • 3,664
  • 1
  • 24
  • 37
45

I'm surprised that no one has come up with Hastad's "Some Optimal Inapproximability Results" (JACM 2001; originally STOC 1997). This landmark paper has been written so well, you can come to it with little other than mathematical maturity and it will make you want to learn several things well, such as its Fourier techniques, parallel repetition, gadgets, and whatnot.

44

Hundreds of Impossibility Results for Distributed Computing by Fich and Ruppert. A readable, pictorial survey that really does present hundreds of impossibility results, including the core questions of the field. A remarkable piece of expository writing.

Aaron Sterling
  • 6,994
  • 6
  • 42
  • 74
44

Les Valiant's Theory of the Learnable (1984) set the agenda for learning theory for decades, and it's a nice and readable paper!

There's also quite a bit of intuitive explanation in the paper that makes it fun and compelling. Various parts of this paper are still routinely quoted in COLT/ALT talks.

Lev Reyzin
  • 11,968
  • 13
  • 63
  • 103
43

Polynomial-Time Algorithms for Prime Factorization and Discrete Logarithms on a Quantum Computer - Peter W. Shor This paper showed that the discrete logarithm problem can be solved in $O((log N)^3)$ time when the corresponding classical algorithm takes considerably longer specifically $O\left(\exp\left(\left(\begin{matrix}\frac{64}{9}\end{matrix} b\right)^{1\over3} (\log b)^{2\over3}\right)\right)$ which is the runtime of GNFS.

Joshua Herman
  • 1,661
  • 17
  • 38
Pratik Deoghare
  • 1,924
  • 18
  • 26
43

Perhaps too basic, but I'm shocked that nobody has mentioned the original Lambda papers by Steele and Sussman: SCHEME: An Interpreter for Extended Lambda Calculus, Lambda: The Ultimate Imperative, Lambda: The Ultimate Declarative.

Klaus Draeger
  • 2,520
  • 1
  • 23
  • 19
sclv
  • 1,379
  • 1
  • 14
  • 17
42

John McCarthy's Recursive functions of symbolic expressions and their computation by machine, part I.

This is the foundational paper on Lisp. Here we find the first metacircular evaluator, fitting on a single page. Its impact cannot be overstated, and it is still eminently readable.

Per Vognsen
  • 2,161
  • 16
  • 25
37

The complexity of theorem-proving procedures by Stephen A. Cook. This paper proves that all the languages decided by polytime nondeterministic Turing machines can be (Cook-)reduced to the set of propositional tautologies.

The importance of this result is (at least) twofold: first, it shows that there exist problems in NP which are at least as hard as the whole class, the NP-complete problems; furthermore, it provides a concrete example of such a problem, which can then be reduced to others in order to prove them complete.

Nowadays Karp reductions are more commonly used than Cook reductions, but the main proof of this paper can be easily adapted to show that SAT is NP-complete with respect to Karp reductions.

  • 7
    This is one of those conference papers for which no journal version ever appeared, but this one is definitely worth going back to: well written and full of great side comments. – András Salamon Sep 12 '10 at 16:42
37

C.A.R. Hoare, An Axiomatic Basis for Computer Programming.

From the abstract: In this paper an attempt is made to explore the logical foundations of computer programming by use of techniques which were first applied in the study of geometry and have later been extended to other branches of mathematics.

It has six pages that are quite easy to follow.

Radu GRIGore
  • 4,796
  • 30
  • 69
37

Call-by-value is dual to call-by-name by Philip Wadler is a good read.

lyonanderson
  • 101
  • 1
  • 2
34

Alon, Matias and Szegedy, The space complexity of approximating the frequency moments, JCSS 58(1):137-147, 1999.

This rather magical paper was the first one to formalize streaming algorithms and prove rigorous upper and lower bounds for foundational tasks in the streaming model. Its techniques are simple, its proofs are beautiful, and its impact has been profound. The work won Alon, Matias and Szegedy the Gödel Prize in 2005.

arnab
  • 7,000
  • 1
  • 38
  • 55
30

Immerman's paper proving the theorem now known as the Immerman–Szelepcsényi theorem, is a great example of easy-to-read, clever and short paper. I love the story told in the intro.

N. Immerman, Nondeterministic space is closed under complementation, SIAM Journal on Computing 17, 1988, pp. 935–938.

Michaël Cadilhac
  • 3,946
  • 22
  • 39
  • 1
    To be fair, Szelepcsényi's paper, "The method of forced enumeration for nondeterministic automata," is just as nice. – Lev Reyzin Jun 27 '12 at 14:00
30

I recommend reading Savitch's paper. It basically states that, for any function $f(n) \ge \log(n)$,

$\text{NSPACE}\left(f\left(n\right)\right) \subseteq \text{DSPACE}\left(\left(f\left(n\right)\right)^2\right).$

The result establishes, for example, that $\text{NPSPACE} = \text{PSPACE}$; a surprising result which its "time" counterpart ($\text{P}$ vs. $\text{NP}$) is a long-standing open problem.

Savitch, Walter J. (1970), "Relationships between nondeterministic and deterministic tape complexities", Journal of Computer and System Sciences 4 (2): 177–192.

Sadeq Dousti
  • 16,479
  • 9
  • 69
  • 152
28

Russell Impagliazzo's A Personal View of Average-Case Complexity. This is a great paper because it is cleverly written, and it summarizes the state of affairs in five "worlds" where our conjectures about complexity are resolved in various ways, giving real-world consequences in each case.

Steve Flammia
  • 813
  • 8
  • 12
26

Improved approximation algorithms for maximum cut and satisfiability problems using semidefinite programming by Goemans and Williamson.

A fine example of introducing a new technique to obtain results that are much better than those known before.

25

How to Write a Proof, by Leslie Lamport.

Anthony Labarre
  • 3,264
  • 1
  • 25
  • 33
  • 5
    I read this and I read A Mathematician's Lament by Lockhart (http://www.maa.org/devlin/LockhartsLament.pdf). IMHO I believe that the strategy that Lamport suggest goes against what Lockhart's argues on the beauty of mathematics. – Marcos Villagra Apr 25 '11 at 05:02
  • 6
    Very interesting read. I understand your opinion, but if I'm not mistaken, Lamport aims his message towards people who are more "mathematically educated" than those targeted by Lockhart, who aims at helping students develop a taste for mathematics. I'll also admit that following a strict format makes proofs quite dull to read, but I agree with Lamport on the idea of proofs by levels: you do not always want/need/have time to read everything in detail, and even when you do, having a summary of what's to come can be quite helpful. Quite a lot more than those "easy to see/clearly/wlog/..." ;-) – Anthony Labarre May 17 '11 at 11:27
24

Extractors and Pseudorandom Generators by Luca Trevisan. In this paper good randomness extractor is built by the means of error-correcting codes and combinatorial designs. Construction is quite easy to understand but it is completely stunning, because it is not obvious at all what is the connection between extractors, codes and designs.

After all, it is a good example of a result in TCS that requires some fancy combinatorics.

ilyaraz
  • 1,569
  • 18
  • 33
22

The influence of variables on boolean functions, J. Kahn, G. Kalai and N. Linial

This paper introduced Fourier techniques for TCS community and solved very neat open problem.

I find this paper very readable.

ilyaraz
  • 1,569
  • 18
  • 33
21

If I may quote Sarah Palin on this issue: "All of them".

More seriously, I think most papers should not be read in the original. As time passes people figure out better way of understanding and presenting the original problem/solution. Except for the Turing original paper, which is of historical importance, I would not recommend reading most original papers if there is followup work that cleaned it up. In particular, of a lot of stuff is presented much better in books than in the original.

Sariel Har-Peled
  • 9,626
  • 31
  • 53
  • 16
    This comment is true in general, but Ryan explicitly asks for examples for which this is not true. There are many classic papers that contain conjectures not yet proved, techniques that have been overlooked, or results that tend to be forgotten but could be dusted off and put to new uses. – András Salamon Sep 17 '10 at 11:35
  • Fair enough, but IMHO for most of the papers suggested books are a better place to read this stuff. – Sariel Har-Peled Sep 23 '10 at 02:39
  • I see your point. I think that university scholarship is like one giant paper, in a way, with its own way of thinking. Some branches of philosophy refer to this process as lexification. But when it comes to someone who really invented/wrote or even lexified something huge and unique (like Turing), a mere mortal cannot hope to lexify it completely, and there will always be new secrets waiting there. – ixtmixilix Sep 26 '10 at 22:12
  • 12
    I disagree. It is true that original papers sometime are unreadable and secondary works give better exposition of the results, but sometimes the original papers contain ideas which are omitted in later works. Also reading original papers can teach us how the author came up with the idea. Take a look at this post of Timothy Chow on MO: http://mathoverflow.net/questions/28268/do-you-read-the-masters – Kaveh Sep 29 '10 at 18:35
  • 4
    Its great when this happens. I just claim that it is somewhat rare. – Sariel Har-Peled Sep 30 '10 at 04:14
  • 6
    You say "All of them", but don't you then argue for "None of them"? – Peter Taylor Jan 19 '11 at 11:28
  • 2
    @Peter Taylor, I think that's why Sarah is mentioned. :) – Radu GRIGore Feb 09 '11 at 14:32
  • 1
    I once read the original paper of Kleene "Representation of Events in Nerve Nets and Finite Automata" in which regular expression are defined for the first time. This paper is far away of the presentation of this material in todays courses, and maybe a little to much complicated, but nevertheless there I learned about nerve nets. But to mention a counter-example from mathematics, I once read the paper "Algebraische Theorie der Körper" by E. Steinitz from 1910(!) when I took a course on Galois Theory and was very impressed by the fact that the theory on fields was almost identically in my cours – StefanH Jan 08 '14 at 23:22
  • I mean the paper presented the material almost identically to the way it was introduced in the course and terminology was very close. – StefanH Jan 08 '14 at 23:23
18

Chomsky analyzes how mathematical models can be used to describe natural language, from a linguistic point of view.

András Salamon
  • 19,000
  • 3
  • 64
  • 150
mgalle
  • 356
  • 1
  • 3
  • 3
    By the way, I am not advocating this paper -- just edited to fix typos and add a link. I prefer Gold's paper if one wants a classic paper about language. – András Salamon Sep 17 '10 at 13:10
18

Kurt Gödel's On formally undecidable propositions of Principia Mathematica and related systems.

Konrad Rudolph
  • 351
  • 1
  • 9
Giorgio Camerani
  • 6,842
  • 1
  • 34
  • 64
  • 6
    This is important, though I do think that later treatments on the subject are easier to read than the original. – Rob Apr 26 '11 at 20:02
18

"Can Programming Be Liberated from the von Neumann Style? A Functional Style and Its Algebra of Programs" by John Backus. This is the 1977 ACM Turing Award Lecture in which Backus introduces functional programming to the world. ACM honored Backus with this award for his seminal work on FORTRAN and for being the B in BNF notation used for describing programming language syntax. I found this work to be really inspiring. It caused me to look at computers and programming languages in a whole new way.

It also represents the kind of paper I wish there were more of. It exposes the inspiration and thought processing behind a nest of ideas without the rigorous but limiting tone of a research paper. It is a shame that researchers have to wait for an opportunity like the ACM Turing Award to be able to express themselves in this mode. Of course, few researchers can write like John Backus. This papers clarity of vision amazes me.

Ekene E.
  • 121
  • 4
Paul Topping
  • 1
  • 1
  • 2
18

You and Your Research by Richard Hamming.

Not a peer reviewed paper, but the transcript of a seminar given in 1986. It's a great write up about lessons learned for becoming a great researcher.

Update: here is the later version of this talk.

Igor Pak
  • 812
  • 5
  • 15
Daniel
  • 1
  • 1
  • 2
16

Natural Proofs by Razborov and Rudich.

The original paper is clearly written, easy to read and requires very little background knowledge. Yet it proves what is (in my opinion) one of the most important known results in computational complexity, by showing that most naive or "natural" approaches one could think of for proving P != NP are doomed to failure. Worth reading at least as much for the mind-expanding nature of the proof as for the result itself.

Ashley Montanaro
  • 2,013
  • 18
  • 21
15

PRIMES in P by Agrawal, Kayal, Saxena

Also known as the AKS primality test, was the first deterministic, (unconditional,) polynomial time primality testing algorithm, that was proposed in the paper, in 2002.

The authors received the Gödel Prize and the Fulkerson Prize for this work.

Subhayan
  • 831
  • 9
  • 19
15

Alan Turing's Computing Machinery and Intelligence.

Giorgio Camerani
  • 6,842
  • 1
  • 34
  • 64
15

R. Moser and G. Tardos: A constructive proof of the general Lovasz Local Lemma

In general the lovasz local lemma is used to (noncronstructively) prove the existence of some (combinatorial) object. Moser and Tardos showed that you can efficiently find this object with a very simple algorithm (in most applications of the LLL).

Great result and nice paper!

Marc Bury
  • 1,338
  • 9
  • 17
14

This paper formalized quantum Turing machines and quantum complexity theory, introducing the class of efficient quantum algorithms BQP and showed the first example of a problem (Fourier sampling) in BQP but not known to be in BPP. Although there is also a previous conference paper from 1993 and Bernstein's PhD thesis, this paper in particular is very well written, easy to understand, and fun to read.

Dai Le
  • 3,664
  • 1
  • 24
  • 37
Marcos Villagra
  • 3,290
  • 27
  • 45
14

There are two essays (but not paper) which are very handy after reading all the suggested papers:
1.How to write papers
2.How not to write papers
Both by Oded Goldreich

Yasser Sobhdel
  • 654
  • 6
  • 16
14

Ronald L. Rivest, Adi Shamir, Leonard M. Adleman: A Method for Obtaining Digital Signatures and Public-Key Cryptosystems. Commun. ACM (CACM) 21(2):120-126 (1978)

If your are interested in Crypto/Security, this is a very nice reading.

VS.
  • 539
  • 3
  • 15
giuper
  • 1
  • 2
  • 2
14

The mechanical evaluation of expressions by Peter J. Landin

Introduced:

  1. the lambda calculus as a basis for defining a programming language,

  2. abstract syntax,

  3. the idea of meta-language to explain other languages,

  4. imperative constructs to the lambda calculus.

gadmm
  • 307
  • 1
  • 8
kunjan kshetri
  • 101
  • 1
  • 4
13

Factoring Polynomials with Rational Coefficients by Lenstra, Lenstra and Lovasz. They present the LLL lattice reduction algorithm for finding short vectors on integer lattices and show an application for factoring polynomials with rational coefficients in polynomial time.

While the algorithm has since been optimized and the polynomial factoring algorithm has been simplified (see Yap's book, chap. 9, for a good reference), the original paper has a good description of the lattice reduction algorithm.

arnab
  • 7,000
  • 1
  • 38
  • 55
user834
  • 2,806
  • 21
  • 27
13

Emil Post's "Recursively enumerable sets of positive integers and their decision problems." Bull. Amer. Math. Soc. 50 (1944), 284-316.

Not only is the paper readable, but it was (I believe) the first paper to introduce each of the following notions, many of which were later adapted to polynomial-time either as central ideas or for interesting results:

  • Many-one reduction (and one-one reduction)
  • Truth-table reduction
  • Simple sets (=complements of immune sets), hypersimple sets
  • Creative sets (later used in papers regarding the Berman-Hartmanis isomorphism conjecture)

Especially recommended for anyone interested in history and/or computability theory. As far as I can tell, it's also a good survey of all of computability theory up to 1944, and was really the starting point for the blossoming of the field.

Joshua Grochow
  • 37,260
  • 4
  • 129
  • 228
12

Gold proves (for a simple model) that it is not possible to learn even very simple languages (like those generated by regular grammars) if only strings that occur in the language are presented. In contrast, if one can query whether arbitrary strings are in the language, then languages generated by quite complex grammars can be learned. This set the scene for countless papers (well, at least 2660 according to Google Scholar), many of them dealing with, starting from, or criticizing the hypothesis that natural language cannot be learned without negative examples. Whether you agree or disagree with this foundation of Chomskian universal grammar, Gold's paper is well-written and clearly argued, and makes no claims about whether natural language can be learned or not. The model is simple, the results elegant, the consequences misunderstood -- read it and make up your own mind.

András Salamon
  • 19,000
  • 3
  • 64
  • 150
11

PCP Theorem by Gap Amplification by Irit Dinur

This paper should be of interest to anyone who uses the PCP theorem for approximation algorithms. It gives an alternate proof which arises much more naturally from approximation algorithms than the original proof. This one is a so-called "combinatorial proof".

Bill Fahle
  • 1
  • 1
  • 2
11

Another very nice paper in proof complexity is

Ben-Sasson, Wigderson - Short proofs are narrow, Resolution made simple

Notice that even this paper gives a new technique which simplifies previous results.

MassimoLauria
  • 1,841
  • 16
  • 21
11

Time bounds for Selection by Blum, Floyd, Pratt, Rivest and Tarjan.

Very elegant algorithm. Plus it has four Turing award winners as authors.

yzll
  • 428
  • 1
  • 3
  • 9
10

This is a classic!

Probabilistic Computations: Toward a Unified Measure of Complexity. Andrew Yao. FOCS'77.

Here Yao gives his famous Minmax principle. I read it and is so easy to read and fun. And the proof is just beautiful and the result amazing.

Marcos Villagra
  • 3,290
  • 27
  • 45
10

Smoothed analysis of algorithms: why the simplex algorithm usually takes polynomial time by Daniel Spielman and Shang-Hua Teng for introducing smoothed analysis and showing it's success in explaining the behavior of the simplex algorithm.

Opt
  • 1,311
  • 14
  • 20
9

The paper Impossibility of distributed consensus with one faulty process by Fisher, Lynch and Paterson shows that it is impossible to reach agreement in an asynchronous distributed system if 1 process might crash, no matter how many other (correct) processes are in the system! While this impossibility result can be circumvented using randomization, the methods of this paper, i.e. using indistinguishability to show that processes must behave in a certain way, have turned out to be highly useful for showing subsequent lower bound/impossibility for problems unrelated to fault-tolerant consensus (e.g. graph problems). As a plus, the paper is short but self-contained and a nice introductory read for someone new to distributed computing.

Peter
  • 1,251
  • 7
  • 11
8

I'll point a couple of papers in Proof Complexity, they are clear and explain most of the relevant techniques in the field (at least for Resolution system).

The first one is

Beame, Pitassi - Simplified and improved Resolution lower bounds

the paper really nails down the core of previous pigeon hole principle lower bounds. And that's why I think it is a pleasure to be read.

MassimoLauria
  • 1,841
  • 16
  • 21
8

Shafi Goldwasser, Silvio Micali: Probabilistic Encryption. J. Comput. Syst. Sci. 28(2): 270-299 (1984)

This paper is the theoretical foundations of modern Cryptography.

VS.
  • 539
  • 3
  • 15
giuper
  • 1
  • 2
  • 2
7

The theory of interstellar trade by Paul Krugman

Martin Schwarz
  • 5,496
  • 25
  • 41
Pratik Deoghare
  • 1,924
  • 18
  • 26
  • 2
    How does interstellar arbitrage (while undoubtedly interesting) relate to theoretical computer science? – András Salamon Sep 17 '10 at 11:32
  • 5
    In the question's explanation there is a sentence : They don't have to be from theoretical computer science -- anything that you think might appeal to the community is a fine answer. – Pratik Deoghare Sep 24 '10 at 08:30
  • 4
    The critique in the paper applies to TCS as well – Noam Oct 14 '10 at 19:39
6

Bill Gasarch's P v NP poll

Not a paper like the others mentioned but certainly interesting and still very relevant today since no significant progress has been made in proving lower bounds.

Mugizi Rwebangira
  • 1,278
  • 10
  • 18
6

I guess this question will be searched by some amateurs like me. While everybody proficient in the field will most likely know it, I found Donald Knuth's paper Dancing Links a very interesting read.

It doesn't require too much theoretical background and still gives an interesting impression on how we can find new ways to solve well known problems with some creative thinking. And it directly leads to the exact cover problem which again provided some interesting insights. This especially for everybody who may try his skills in areas like solving Sudokus or the N queens problem.

thorsten muller
  • 101
  • 1
  • 2
6

Guy E. Blelloch: Programming Parallel Algorithms

Very clear introduction to parallel algorithms.

Raphael
  • 4,565
  • 28
  • 48
5

Allen Newell and Herbert A. Simon’s “Computer Science as Empirical Inquiry: Symbols and Search” (direct PDF link).

This quote summarizes it:

“We come now to the evidence for the hypotesis that physical symbol systems are capable of intelligent action, and that general intelligent action calls for a physical symbol system”

It's also very readable!

Agos
  • 243
  • 3
  • 9
5

A Simple Proof that Toffoli and Hadamard are Quantum Universal, D. Aharonov summarising a result found by Yaoyun Shi.

Because it is a simple, well written, 4 pages paper which makes you think and realise that Quantum Computation can be done with just two elementary blocks, from which one is a universal classical gate.

Juan Bermejo Vega
  • 1,405
  • 1
  • 14
  • 30
5

Some great suggestions are included in the reading list of the Reading the Classics course given by Christos Papadimitriou at Berkeley. Some of them have been mentioned in previous answers already. Notable exception: Euler's Königsberg Bridge Problem.

Dimitris
  • 101
  • 1
  • 2
4

On Universal Learning Algorithms, Oded Goldreich and Dana Ron (1997), Information Processing Letters, Volume 63, Issue 3, pp. 131-136. (see also the updated version)

Adapting Levin's argument for the existence of an optimal algorithm for NP, the authors show that there exists a universal learning algorithm (in several learning settings, including PAC): "if a concept class is learnable, this algorithm will learn it, optimally."

Beyond the result itself, and perhaps more strikingly, this is also (as pointed out and discussed in the paper) a great illustration of the dangers of abusing $O(\cdot)$ notations and asymptotics.

Clement C.
  • 4,461
  • 29
  • 51
4

Fischer Black and Myron Scholes, The Pricing of Options and Corporate Liabilities https://web.archive.org/web/20231114103316/https://www.cs.princeton.edu/courses/archive/fall09/cos323/papers/black_scholes73.pdf

This paper is still one of the clearest descriptions of the options pricing problem and its closed-form solution. More fundamentally, it addresses how to price risk by convolving a probability distribution with a non-smooth payoff curve.

Ekene E.
  • 121
  • 4
4

The RSA cryptosystem was described in a short elegant paper; it's very readable and created quite a stir even among non-scientists.

Fixee
  • 1,003
  • 8
  • 15
3

My favorite piece of scientific writing is Charles Bennett's 1979 On Random and Hard-to-Describe Numbers which describes Chaitin's number. You should read it not because its scientific content will be useful to you, although it may be, but just for the quality of the writing. If you already know about Chaitin's number, just skip to the last paragraph on page 6 and read from there on.

I don't want to quote too much from the article, but here is the last sentence of the abstract.

Other, Cabalistic properties of $\Omega$ [Chaitin's number] are pointed out for the first time.

Aram Harrow
  • 1,060
  • 10
  • 10
3

Although not published in a scientific journal, Vannevar Bush's As We May Think has influenced much work in the computer science field, including hypertext, the personal computer, digital libraries, and the Internet. It even has it's own Wikipedia article, for god's sake.

2

Time/space tradeoffs for reversible computation (1989) by Charles Bennett: In this paper, Bennett introduces a pebble game to show that reversible computation can emulate any conventional computation with very reasonable space/time overheads. This method of emulating conventional computation with reversible computation will become increasingly practical in the future when energy efficient reversible computers and quantum computers become prominent.

Ekene E.
  • 121
  • 4
Joseph Van Name
  • 596
  • 5
  • 12
2

Ray Solomonoff:

  • A formal theory of inductive inference. Part I and Part II, 1964.

  • Complexity-based induction systems: Comparisons and convergence theorems. 1978

Kolmogorov:

  • Three Approaches to the Quantitative Definition of Information. 1965

Martin-Löf:

  • The definition of random sequences. 1966

Marvin Minsky said: The most important discovery since G"odel was the discovery by Chaitin, Solomonoff and Kolmogorov of the concept called Algorithmic Probability which is a fundamental new theory of how to make predictions given a collection of experiences and this is a beautiful theory, everybody should learn it, but it's got one problem, that is, that you cannot actually calculate what this theory predicts because it is too hard, it requires an infinite amount of work. However, it should be possible to make practical approximations to the Chaitin, Kolmogorov, Solomonoff theory that would make better predictions than anything we have today. Everybody should learn all about that and spend the rest of their lives working on it.

Xi Li
  • 21
  • 1
2

Not specifically a topic in theoretical computer science, but I think an area that is fundamental to much of the work in data analysis and machine learning is a foundational paper in what are currently known as Graphical Models:

Lauritzen, Steffen L. and Spiegelhalter, David J., ( 1988) "Local Computations with Probabilities on Graphical Structures and their Application to Expert Systems", Journal of the Royal Statistical Society, Series B (Methodological)", 50(2) pp 157--224.

The way they solve the probability updating problem harkens back to earlier work on optimizing elimination orders for linear systems, and, subsequently is expanded on by Lauritzen to a larger class of dynamic programming style problems.

VS.
  • 539
  • 3
  • 15
1

Rumelhart, David E.; Hinton, Geoffrey E., Williams, Ronald J. (8 October 1986). "Learning representations by back-propagating errors". Nature 323 (6088): 533–536. DOI:10.1038/323533a0

The paper that introduced backpropagation and resuscitated neural networks.

mrig
  • 151
  • 2
1

Expanders are special graph which are sparse yet highly connected. The challenge is to give an explicit construction of expanders. See the survey by Hoory-Linial-Wigderson-Expander Graphs and their Application for several applications of expanders. The most useful expanders are the ones with constant degree.If we have constant-sized object then we can freely use it without having to find a nice description for it. Also we can always find constant size in constant time by brute force approach.

An Elementary Construction of Constant-Degree Expanders by Alon-Schwartz-Shapira appeared in SODA 2007 and gave a wonderful construction of expanders using Replacement Product. In my opinion, the construction is so nice that this ought to be a part of lectures on Expanders. They apply the Replacement product only twice on Cayley expanders to obtain a constant-degree expanders. The construction is so combinatorial that one can even visualise the edges coming out of a cut.

akr_
  • 101
  • 3
-1

The paper Multidimensional Divide-and-conquer by Jon L. Bentley discussed multidimensional divide-and-conquer, an algorithmic paradigm to solve problems in multidimensional divide-and-conquer.

It is really easy to read and helpful to solve lots of high-dimensional computation problem.

VS.
  • 539
  • 3
  • 15
xiaom
  • 41
  • 4
-1

The Base Rate Fallacy and its implications for the Difficulty of Intrusion Detection

Stefan Axelsson, "The Base-Rate Fallacy and its Implications for the Difficulty Of Intrusion", 1999

For those poor saps stuck waiting for IDS alerts, this shows that even if you're signature is 99% accurate, if it's a packet based IDS then you will still be overrun with false positives.

Comment - More generally an understanding of how Bayes rule applies to probabilistic inference, and how probabilistic graphical models are used illuminates many fallacies that arise in working with statistical data.

Kaveh
  • 21,577
  • 8
  • 82
  • 183
user5283
  • 1
  • 1
  • 4
    I think, in the context of this site, "everyone" should mean "every theoretical computer scientist". So why should we all be persuaded to read something that isn't even theoretical computer science? – David Eppstein Dec 30 '12 at 18:36