fbpx
Wikipedia

Average-case complexity

In computational complexity theory, the average-case complexity of an algorithm is the amount of some computational resource (typically time) used by the algorithm, averaged over all possible inputs. It is frequently contrasted with worst-case complexity which considers the maximal complexity of the algorithm over all possible inputs.

There are three primary motivations for studying average-case complexity.[1] First, although some problems may be intractable in the worst-case, the inputs which elicit this behavior may rarely occur in practice, so the average-case complexity may be a more accurate measure of an algorithm's performance. Second, average-case complexity analysis provides tools and techniques to generate hard instances of problems which can be utilized in areas such as cryptography and derandomization. Third, average-case complexity allows discriminating the most efficient algorithm in practice among algorithms of equivalent best case complexity (for instance Quicksort).

Average-case analysis requires a notion of an "average" input to an algorithm, which leads to the problem of devising a probability distribution over inputs. Alternatively, a randomized algorithm can be used. The analysis of such algorithms leads to the related notion of an expected complexity.[2]: 28 

History and background edit

The average-case performance of algorithms has been studied since modern notions of computational efficiency were developed in the 1950s. Much of this initial work focused on problems for which worst-case polynomial time algorithms were already known.[3] In 1973, Donald Knuth[4] published Volume 3 of the Art of Computer Programming which extensively surveys average-case performance of algorithms for problems solvable in worst-case polynomial time, such as sorting and median-finding.

An efficient algorithm for NP-complete problems is generally characterized as one which runs in polynomial time for all inputs; this is equivalent to requiring efficient worst-case complexity. However, an algorithm which is inefficient on a "small" number of inputs may still be efficient for "most" inputs that occur in practice. Thus, it is desirable to study the properties of these algorithms where the average-case complexity may differ from the worst-case complexity and find methods to relate the two.

The fundamental notions of average-case complexity were developed by Leonid Levin in 1986 when he published a one-page paper[5] defining average-case complexity and completeness while giving an example of a complete problem for distNP, the average-case analogue of NP.

Definitions edit

Efficient average-case complexity edit

The first task is to precisely define what is meant by an algorithm which is efficient "on average". An initial attempt might define an efficient average-case algorithm as one which runs in expected polynomial time over all possible inputs. Such a definition has various shortcomings; in particular, it is not robust to changes in the computational model. For example, suppose algorithm A runs in time tA(x) on input x and algorithm B runs in time tA(x)2 on input x; that is, B is quadratically slower than A. Intuitively, any definition of average-case efficiency should capture the idea that A is efficient-on-average if and only if B is efficient on-average. Suppose, however, that the inputs are drawn randomly from the uniform distribution of strings with length n, and that A runs in time n2 on all inputs except the string 1n for which A takes time 2n. Then it can be easily checked that the expected running time of A is polynomial but the expected running time of B is exponential.[3]

To create a more robust definition of average-case efficiency, it makes sense to allow an algorithm A to run longer than polynomial time on some inputs but the fraction of inputs on which A requires larger and larger running time becomes smaller and smaller. This intuition is captured in the following formula for average polynomial running time, which balances the polynomial trade-off between running time and fraction of inputs:

 

for every n, t > 0 and polynomial p, where tA(x) denotes the running time of algorithm A on input x, and ε is a positive constant value.[6] Alternatively, this can be written as

 

for some constants C and ε, where n = |x|.[7] In other words, an algorithm A has good average-case complexity if, after running for tA(n) steps, A can solve all but a nc/(tA(n))ε fraction of inputs of length n, for some ε, c > 0.[3]

Distributional problem edit

The next step is to define the "average" input to a particular problem. This is achieved by associating the inputs of each problem with a particular probability distribution. That is, an "average-case" problem consists of a language L and an associated probability distribution D which forms the pair (L, D).[7] The two most common classes of distributions which are allowed are:

  1. Polynomial-time computable distributions (P-computable): these are distributions for which it is possible to compute the cumulative density of any given input x. More formally, given a probability distribution μ and a string x ∈ {0, 1}n it is possible to compute the value   in polynomial time. This implies that Pr[x] is also computable in polynomial time.
  2. Polynomial-time samplable distributions (P-samplable): these are distributions from which it is possible to draw random samples in polynomial time.

These two formulations, while similar, are not equivalent. If a distribution is P-computable it is also P-samplable, but the converse is not true if PP#P.[7]

AvgP and distNP edit

A distributional problem (L, D) is in the complexity class AvgP if there is an efficient average-case algorithm for L, as defined above. The class AvgP is occasionally called distP in the literature.[7]

A distributional problem (L, D) is in the complexity class distNP if L is in NP and D is P-computable. When L is in NP and D is P-samplable, (L, D) belongs to sampNP.[7]

Together, AvgP and distNP define the average-case analogues of P and NP, respectively.[7]

Reductions between distributional problems edit

Let (L,D) and (L′, D′) be two distributional problems. (L, D) average-case reduces to (L′, D′) (written (L, D) ≤AvgP (L′, D′)) if there is a function f that for every n, on input x can be computed in time polynomial in n and

  1. (Correctness) xL if and only if f(x) ∈ L′
  2. (Domination) There are polynomials p and m such that, for every n and y,  

The domination condition enforces the notion that if problem (L, D) is hard on average, then (L′, D′) is also hard on average. Intuitively, a reduction should provide a way to solve an instance x of problem L by computing f(x) and feeding the output to the algorithm which solves L'. Without the domination condition, this may not be possible since the algorithm which solves L in polynomial time on average may take super-polynomial time on a small number of inputs but f may map these inputs into a much larger set of D' so that algorithm A' no longer runs in polynomial time on average. The domination condition only allows such strings to occur polynomially as often in D'.[6]

DistNP-complete problems edit

The average-case analogue to NP-completeness is distNP-completeness. A distributional problem (L′, D′) is distNP-complete if (L′, D′) is in distNP and for every (L, D) in distNP, (L, D) is average-case reducible to (L′, D′).[7]

An example of a distNP-complete problem is the Bounded Halting Problem, BH, defined as follows:

 [7]

In his original paper, Levin showed an example of a distributional tiling problem that is average-case NP-complete.[5] A survey of known distNP-complete problems is available online.[6]

One area of active research involves finding new distNP-complete problems. However, finding such problems can be complicated due to a result of Gurevich which shows that any distributional problem with a flat distribution cannot be distNP-complete unless EXP = NEXP.[8] (A flat distribution μ is one for which there exists an ε > 0 such that for any x, μ(x) ≤ 2−|x|ε.) A result by Livne shows that all natural NP-complete problems have DistNP-complete versions.[9] However, the goal of finding a natural distributional problem that is DistNP-complete has not yet been achieved.[10]

Applications edit

Sorting algorithms edit

As mentioned above, much early work relating to average-case complexity focused on problems for which polynomial-time algorithms already existed, such as sorting. For example, many sorting algorithms which utilize randomness, such as Quicksort, have a worst-case running time of O(n2), but an average-case running time of O(n log(n)), where n is the length of the input to be sorted.[2]

Cryptography edit

For most problems, average-case complexity analysis is undertaken to find efficient algorithms for a problem that is considered difficult in the worst-case. In cryptographic applications, however, the opposite is true: the worst-case complexity is irrelevant; we instead want a guarantee that the average-case complexity of every algorithm which "breaks" the cryptographic scheme is inefficient.[11]

Thus, all secure cryptographic schemes rely on the existence of one-way functions.[3] Although the existence of one-way functions is still an open problem, many candidate one-way functions are based on hard problems such as integer factorization or computing the discrete log. Note that it is not desirable for the candidate function to be NP-complete since this would only guarantee that there is likely no efficient algorithm for solving the problem in the worst case; what we actually want is a guarantee that no efficient algorithm can solve the problem over random inputs (i.e. the average case). In fact, both the integer factorization and discrete log problems are in NPcoNP, and are therefore not believed to be NP-complete.[7] The fact that all of cryptography is predicated on the existence of average-case intractable problems in NP is one of the primary motivations for studying average-case complexity.

Other results edit

In 1990, Impagliazzo and Levin showed that if there is an efficient average-case algorithm for a distNP-complete problem under the uniform distribution, then there is an average-case algorithm for every problem in NP under any polynomial-time samplable distribution.[12] Applying this theory to natural distributional problems remains an outstanding open question.[3]

In 1992, Ben-David et al. showed that if all languages in distNP have good-on-average decision algorithms, they also have good-on-average search algorithms. Further, they show that this conclusion holds under a weaker assumption: if every language in NP is easy on average for decision algorithms with respect to the uniform distribution, then it is also easy on average for search algorithms with respect to the uniform distribution.[13] Thus, cryptographic one-way functions can exist only if there are distNP problems over the uniform distribution that are hard on average for decision algorithms.

In 1993, Feigenbaum and Fortnow showed that it is not possible to prove, under non-adaptive random reductions, that the existence of a good-on-average algorithm for a distNP-complete problem under the uniform distribution implies the existence of worst-case efficient algorithms for all problems in NP.[14] In 2003, Bogdanov and Trevisan generalized this result to arbitrary non-adaptive reductions.[15] These results show that it is unlikely that any association can be made between average-case complexity and worst-case complexity via reductions.[3]

See also edit

References edit

  1. ^ O. Goldreich and S. Vadhan, Special issue on worst-case versus average-case complexity, Comput. Complex. 16, 325–330, 2007.
  2. ^ a b Cormen, Thomas H.; Leiserson, Charles E., Rivest, Ronald L., Stein, Clifford (2009) [1990]. Introduction to Algorithms (3rd ed.). MIT Press and McGraw-Hill. ISBN 0-262-03384-4.
  3. ^ a b c d e f A. Bogdanov and L. Trevisan, "Average-Case Complexity," Foundations and Trends in Theoretical Computer Science, Vol. 2, No 1 (2006) 1–106.
  4. ^ D. Knuth, The Art of Computer Programming. Vol. 3, Addison-Wesley, 1973.
  5. ^ a b L. Levin, "Average case complete problems," SIAM Journal on Computing, vol. 15, no. 1, pp. 285–286, 1986.
  6. ^ a b c J. Wang, "Average-case computational complexity theory," Complexity Theory Retrospective II, pp. 295–328, 1997.
  7. ^ a b c d e f g h i S. Arora and B. Barak, Computational Complexity: A Modern Approach, Cambridge University Press, New York, NY, 2009.
  8. ^ Y. Gurevich, "Complete and incomplete randomized NP problems", Proc. 28th Annual Symp. on Found. of Computer Science, IEEE (1987), pp. 111–117.
  9. ^ N. Livne, "All Natural NP-Complete Problems Have Average-Case Complete Versions," Computational Complexity (2010) 19:477. https://doi.org/10.1007/s00037-010-0298-9
  10. ^ O. Goldreich, "Notes on Levin's theory of average-case complexity," Technical Report TR97-058, Electronic Colloquium on Computational Complexity, 1997.
  11. ^ J. Katz and Y. Lindell, Introduction to Modern Cryptography (Chapman and Hall/Crc Cryptography and Network Security Series), Chapman and Hall/CRC, 2007.
  12. ^ R. Impagliazzo and L. Levin, "No Better Ways to Generate Hard NP Instances than Picking Uniformly at Random," in Proceedings of the 31st IEEE Sympo- sium on Foundations of Computer Science, pp. 812–821, 1990.
  13. ^ S. Ben-David, B. Chor, O. Goldreich, and M. Luby, "On the theory of average case complexity," Journal of Computer and System Sciences, vol. 44, no. 2, pp. 193–219, 1992.
  14. ^ J. Feigenbaum and L. Fortnow, "Random-self-reducibility of complete sets," SIAM Journal on Computing, vol. 22, pp. 994–1005, 1993.
  15. ^ A. Bogdanov and L. Trevisan, "On worst-case to average-case reductions for NP problems," in Proceedings of the 44th IEEE Symposium on Foundations of Computer Science, pp. 308–317, 2003.

Further reading edit

The literature of average case complexity includes the following work:

  • Franco, John (1986), "On the probabilistic performance of algorithms for the satisfiability problem", Information Processing Letters, 23 (2): 103–106, doi:10.1016/0020-0190(86)90051-7.
  • Levin, Leonid (1986), "Average case complete problems", SIAM Journal on Computing, 15 (1): 285–286, doi:10.1137/0215020.
  • Flajolet, Philippe; Vitter, J. S. (August 1987), Average-case analysis of algorithms and data structures, Tech. Report, Institut National de Recherche en Informatique et en Automatique, B.P. 105-78153 Le Chesnay Cedex France.
  • Gurevich, Yuri; Shelah, Saharon (1987), "Expected computation time for Hamiltonian path problem", SIAM Journal on Computing, 16 (3): 486–502, CiteSeerX 10.1.1.359.8982, doi:10.1137/0216034.
  • Ben-David, Shai; Chor, Benny; Goldreich, Oded; Luby, Michael (1989), "On the theory of average case complexity", Proc. 21st Annual Symposium on Theory of Computing, Association for Computing Machinery, pp. 204–216.
  • Gurevich, Yuri (1991), "Average case completeness", Journal of Computer and System Sciences, 42 (3): 346–398, doi:10.1016/0022-0000(91)90007-R, hdl:2027.42/29307. See also 1989 draft.
  • Selman, B.; Mitchell, D.; Levesque, H. (1992), "Hard and easy distributions of SAT problems", Proc. 10th National Conference on Artificial Intelligence, pp. 459–465.
  • Schuler, Rainer; Yamakami, Tomoyuki (1992), "Structural average case complexity", Proc. Foundations of Software Technology and Theoretical Computer Science, Lecture Notes in Computer Science, vol. 652, Springer-Verlag, pp. 128–139.
  • Reischuk, Rüdiger; Schindelhauer, Christian (1993), "Precise average case complexity", Proc. 10th Annual Symposium on Theoretical Aspects of Computer Science, pp. 650–661.
  • Venkatesan, R.; Rajagopalan, S. (1992), "Average case intractability of matrix and Diophantine problems", Proc. 24th Annual Symposium on Theory of Computing, Association for Computing Machinery, pp. 632–642.
  • Cox, Jim; Ericson, Lars; Mishra, Bud (1995), The average case complexity of multilevel syllogistic (PDF), Technical Report TR1995-711, New York University Computer Science Department.
  • Impagliazzo, Russell (April 17, 1995), A personal view of average-case complexity, University of California, San Diego.
  • Paul E. Black, , in Dictionary of Algorithms and Data Structures[online]Paul E. Black, ed., U.S. National Institute of Standards and Technology. 17 December 2004.Retrieved Feb. 20/09.
  • Christos Papadimitriou (1994). Computational Complexity. Addison-Wesley.

average, case, complexity, avgp, redirects, here, other, uses, avgp, disambiguation, computational, complexity, theory, average, case, complexity, algorithm, amount, some, computational, resource, typically, time, used, algorithm, averaged, over, possible, inp. AvgP redirects here For other uses see avgp disambiguation In computational complexity theory the average case complexity of an algorithm is the amount of some computational resource typically time used by the algorithm averaged over all possible inputs It is frequently contrasted with worst case complexity which considers the maximal complexity of the algorithm over all possible inputs There are three primary motivations for studying average case complexity 1 First although some problems may be intractable in the worst case the inputs which elicit this behavior may rarely occur in practice so the average case complexity may be a more accurate measure of an algorithm s performance Second average case complexity analysis provides tools and techniques to generate hard instances of problems which can be utilized in areas such as cryptography and derandomization Third average case complexity allows discriminating the most efficient algorithm in practice among algorithms of equivalent best case complexity for instance Quicksort Average case analysis requires a notion of an average input to an algorithm which leads to the problem of devising a probability distribution over inputs Alternatively a randomized algorithm can be used The analysis of such algorithms leads to the related notion of an expected complexity 2 28 Contents 1 History and background 2 Definitions 2 1 Efficient average case complexity 2 2 Distributional problem 2 3 AvgP and distNP 3 Reductions between distributional problems 3 1 DistNP complete problems 4 Applications 4 1 Sorting algorithms 4 2 Cryptography 5 Other results 6 See also 7 References 8 Further readingHistory and background editThe average case performance of algorithms has been studied since modern notions of computational efficiency were developed in the 1950s Much of this initial work focused on problems for which worst case polynomial time algorithms were already known 3 In 1973 Donald Knuth 4 published Volume 3 of the Art of Computer Programming which extensively surveys average case performance of algorithms for problems solvable in worst case polynomial time such as sorting and median finding An efficient algorithm for NP complete problems is generally characterized as one which runs in polynomial time for all inputs this is equivalent to requiring efficient worst case complexity However an algorithm which is inefficient on a small number of inputs may still be efficient for most inputs that occur in practice Thus it is desirable to study the properties of these algorithms where the average case complexity may differ from the worst case complexity and find methods to relate the two The fundamental notions of average case complexity were developed by Leonid Levin in 1986 when he published a one page paper 5 defining average case complexity and completeness while giving an example of a complete problem for distNP the average case analogue of NP Definitions editEfficient average case complexity edit The first task is to precisely define what is meant by an algorithm which is efficient on average An initial attempt might define an efficient average case algorithm as one which runs in expected polynomial time over all possible inputs Such a definition has various shortcomings in particular it is not robust to changes in the computational model For example suppose algorithm A runs in time tA x on input x and algorithm B runs in time tA x 2 on input x that is B is quadratically slower than A Intuitively any definition of average case efficiency should capture the idea that A is efficient on average if and only if B is efficient on average Suppose however that the inputs are drawn randomly from the uniform distribution of strings with length n and that A runs in time n2 on all inputs except the string 1n for which A takes time 2n Then it can be easily checked that the expected running time of A is polynomial but the expected running time of B is exponential 3 To create a more robust definition of average case efficiency it makes sense to allow an algorithm A to run longer than polynomial time on some inputs but the fraction of inputs on which A requires larger and larger running time becomes smaller and smaller This intuition is captured in the following formula for average polynomial running time which balances the polynomial trade off between running time and fraction of inputs Pr x R D n t A x t p n t ϵ displaystyle Pr x in R D n left t A x geq t right leq frac p n t epsilon nbsp for every n t gt 0 and polynomial p where tA x denotes the running time of algorithm A on input x and e is a positive constant value 6 Alternatively this can be written as E x R D n t A x ϵ n C displaystyle E x in R D n left frac t A x epsilon n right leq C nbsp for some constants C and e where n x 7 In other words an algorithm A has good average case complexity if after running for tA n steps A can solve all but a nc tA n e fraction of inputs of length n for some e c gt 0 3 Distributional problem edit The next step is to define the average input to a particular problem This is achieved by associating the inputs of each problem with a particular probability distribution That is an average case problem consists of a language L and an associated probability distribution D which forms the pair L D 7 The two most common classes of distributions which are allowed are Polynomial time computable distributions P computable these are distributions for which it is possible to compute the cumulative density of any given input x More formally given a probability distribution m and a string x 0 1 n it is possible to compute the value m x y 0 1 n y x Pr y displaystyle mu x sum limits y in 0 1 n y leq x Pr y nbsp in polynomial time This implies that Pr x is also computable in polynomial time Polynomial time samplable distributions P samplable these are distributions from which it is possible to draw random samples in polynomial time These two formulations while similar are not equivalent If a distribution is P computable it is also P samplable but the converse is not true if P P P 7 AvgP and distNP edit A distributional problem L D is in the complexity class AvgP if there is an efficient average case algorithm for L as defined above The class AvgP is occasionally called distP in the literature 7 A distributional problem L D is in the complexity class distNP if L is in NP and D is P computable When L is in NP and D is P samplable L D belongs to sampNP 7 Together AvgP and distNP define the average case analogues of P and NP respectively 7 Reductions between distributional problems editLet L D and L D be two distributional problems L D average case reduces to L D written L D AvgP L D if there is a function f that for every n on input x can be computed in time polynomial in n and Correctness x L if and only if f x L Domination There are polynomials p and m such that for every n and y x f x y D n x p n D m n y displaystyle sum limits x f x y D n x leq p n D m n y nbsp The domination condition enforces the notion that if problem L D is hard on average then L D is also hard on average Intuitively a reduction should provide a way to solve an instance x of problem L by computing f x and feeding the output to the algorithm which solves L Without the domination condition this may not be possible since the algorithm which solves L in polynomial time on average may take super polynomial time on a small number of inputs but f may map these inputs into a much larger set of D so that algorithm A no longer runs in polynomial time on average The domination condition only allows such strings to occur polynomially as often in D 6 DistNP complete problems edit The average case analogue to NP completeness is distNP completeness A distributional problem L D is distNP complete if L D is in distNP and for every L D in distNP L D is average case reducible to L D 7 An example of a distNP complete problem is the Bounded Halting Problem BH defined as follows B H M x 1 t M is a non deterministic Turing machine that accepts x in t steps displaystyle BH M x 1 t M text is a non deterministic Turing machine that accepts x text in leq t text steps nbsp 7 In his original paper Levin showed an example of a distributional tiling problem that is average case NP complete 5 A survey of known distNP complete problems is available online 6 One area of active research involves finding new distNP complete problems However finding such problems can be complicated due to a result of Gurevich which shows that any distributional problem with a flat distribution cannot be distNP complete unless EXP NEXP 8 A flat distribution m is one for which there exists an e gt 0 such that for any x m x 2 x e A result by Livne shows that all natural NP complete problems have DistNP complete versions 9 However the goal of finding a natural distributional problem that is DistNP complete has not yet been achieved 10 Applications editSorting algorithms edit As mentioned above much early work relating to average case complexity focused on problems for which polynomial time algorithms already existed such as sorting For example many sorting algorithms which utilize randomness such as Quicksort have a worst case running time of O n2 but an average case running time of O n log n where n is the length of the input to be sorted 2 Cryptography edit For most problems average case complexity analysis is undertaken to find efficient algorithms for a problem that is considered difficult in the worst case In cryptographic applications however the opposite is true the worst case complexity is irrelevant we instead want a guarantee that the average case complexity of every algorithm which breaks the cryptographic scheme is inefficient 11 Thus all secure cryptographic schemes rely on the existence of one way functions 3 Although the existence of one way functions is still an open problem many candidate one way functions are based on hard problems such as integer factorization or computing the discrete log Note that it is not desirable for the candidate function to be NP complete since this would only guarantee that there is likely no efficient algorithm for solving the problem in the worst case what we actually want is a guarantee that no efficient algorithm can solve the problem over random inputs i e the average case In fact both the integer factorization and discrete log problems are in NP coNP and are therefore not believed to be NP complete 7 The fact that all of cryptography is predicated on the existence of average case intractable problems in NP is one of the primary motivations for studying average case complexity Other results editIn 1990 Impagliazzo and Levin showed that if there is an efficient average case algorithm for a distNP complete problem under the uniform distribution then there is an average case algorithm for every problem in NP under any polynomial time samplable distribution 12 Applying this theory to natural distributional problems remains an outstanding open question 3 In 1992 Ben David et al showed that if all languages in distNP have good on average decision algorithms they also have good on average search algorithms Further they show that this conclusion holds under a weaker assumption if every language in NP is easy on average for decision algorithms with respect to the uniform distribution then it is also easy on average for search algorithms with respect to the uniform distribution 13 Thus cryptographic one way functions can exist only if there are distNP problems over the uniform distribution that are hard on average for decision algorithms In 1993 Feigenbaum and Fortnow showed that it is not possible to prove under non adaptive random reductions that the existence of a good on average algorithm for a distNP complete problem under the uniform distribution implies the existence of worst case efficient algorithms for all problems in NP 14 In 2003 Bogdanov and Trevisan generalized this result to arbitrary non adaptive reductions 15 These results show that it is unlikely that any association can be made between average case complexity and worst case complexity via reductions 3 See also editProbabilistic analysis of algorithms NP complete problems Worst case complexity Amortized analysis Best worst and average caseReferences edit O Goldreich and S Vadhan Special issue on worst case versus average case complexity Comput Complex 16 325 330 2007 a b Cormen Thomas H Leiserson Charles E Rivest Ronald L Stein Clifford 2009 1990 Introduction to Algorithms 3rd ed MIT Press and McGraw Hill ISBN 0 262 03384 4 a b c d e f A Bogdanov and L Trevisan Average Case Complexity Foundations and Trends in Theoretical Computer Science Vol 2 No 1 2006 1 106 D Knuth The Art of Computer Programming Vol 3 Addison Wesley 1973 a b L Levin Average case complete problems SIAM Journal on Computing vol 15 no 1 pp 285 286 1986 a b c J Wang Average case computational complexity theory Complexity Theory Retrospective II pp 295 328 1997 a b c d e f g h i S Arora and B Barak Computational Complexity A Modern Approach Cambridge University Press New York NY 2009 Y Gurevich Complete and incomplete randomized NP problems Proc 28th Annual Symp on Found of Computer Science IEEE 1987 pp 111 117 N Livne All Natural NP Complete Problems Have Average Case Complete Versions Computational Complexity 2010 19 477 https doi org 10 1007 s00037 010 0298 9 O Goldreich Notes on Levin s theory of average case complexity Technical Report TR97 058 Electronic Colloquium on Computational Complexity 1997 J Katz and Y Lindell Introduction to Modern Cryptography Chapman and Hall Crc Cryptography and Network Security Series Chapman and Hall CRC 2007 R Impagliazzo and L Levin No Better Ways to Generate Hard NP Instances than Picking Uniformly at Random in Proceedings of the 31st IEEE Sympo sium on Foundations of Computer Science pp 812 821 1990 S Ben David B Chor O Goldreich and M Luby On the theory of average case complexity Journal of Computer and System Sciences vol 44 no 2 pp 193 219 1992 J Feigenbaum and L Fortnow Random self reducibility of complete sets SIAM Journal on Computing vol 22 pp 994 1005 1993 A Bogdanov and L Trevisan On worst case to average case reductions for NP problems in Proceedings of the 44th IEEE Symposium on Foundations of Computer Science pp 308 317 2003 Further reading editThe literature of average case complexity includes the following work Franco John 1986 On the probabilistic performance of algorithms for the satisfiability problem Information Processing Letters 23 2 103 106 doi 10 1016 0020 0190 86 90051 7 Levin Leonid 1986 Average case complete problems SIAM Journal on Computing 15 1 285 286 doi 10 1137 0215020 Flajolet Philippe Vitter J S August 1987 Average case analysis of algorithms and data structures Tech Report Institut National de Recherche en Informatique et en Automatique B P 105 78153 Le Chesnay Cedex France Gurevich Yuri Shelah Saharon 1987 Expected computation time for Hamiltonian path problem SIAM Journal on Computing 16 3 486 502 CiteSeerX 10 1 1 359 8982 doi 10 1137 0216034 Ben David Shai Chor Benny Goldreich Oded Luby Michael 1989 On the theory of average case complexity Proc 21st Annual Symposium on Theory of Computing Association for Computing Machinery pp 204 216 Gurevich Yuri 1991 Average case completeness Journal of Computer and System Sciences 42 3 346 398 doi 10 1016 0022 0000 91 90007 R hdl 2027 42 29307 See also 1989 draft Selman B Mitchell D Levesque H 1992 Hard and easy distributions of SAT problems Proc 10th National Conference on Artificial Intelligence pp 459 465 Schuler Rainer Yamakami Tomoyuki 1992 Structural average case complexity Proc Foundations of Software Technology and Theoretical Computer Science Lecture Notes in Computer Science vol 652 Springer Verlag pp 128 139 Reischuk Rudiger Schindelhauer Christian 1993 Precise average case complexity Proc 10th Annual Symposium on Theoretical Aspects of Computer Science pp 650 661 Venkatesan R Rajagopalan S 1992 Average case intractability of matrix and Diophantine problems Proc 24th Annual Symposium on Theory of Computing Association for Computing Machinery pp 632 642 Cox Jim Ericson Lars Mishra Bud 1995 The average case complexity of multilevel syllogistic PDF Technical Report TR1995 711 New York University Computer Science Department Impagliazzo Russell April 17 1995 A personal view of average case complexity University of California San Diego Paul E Black 8 in Dictionary of Algorithms and Data Structures online Paul E Black ed U S National Institute of Standards and Technology 17 December 2004 Retrieved Feb 20 09 Christos Papadimitriou 1994 Computational Complexity Addison Wesley Retrieved from https en wikipedia org w index php title Average case complexity amp oldid 1151098788, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.