13
ACP RESEARCH PAPER V.E.S. COLLEGE OF ARTS, SCIENCE AND COMMERCE QUANTUM COMPUTER IN CRYPTOGRAPHY YEAR 2014-15 ACP CO-ORDINATOR MENTOR NAME Prof. Aarohi Khar Prof. Shrikant Ghodke SUBMITTED BY Akshay M Shelake FROM T.Y.B.Sc (Computer Science)

Research paper of quantum computer in cryptography

Embed Size (px)

Citation preview

ACP

RESEARCH PAPER

V.E.S. COLLEGE OF ARTS, SCIENCE AND

COMMERCE

QUANTUM COMPUTER IN CRYPTOGRAPHY

YEAR 2014-15

ACP CO-ORDINATOR MENTOR NAME

Prof. Aarohi Khar Prof. Shrikant Ghodke

SUBMITTED BY

Akshay M Shelake

FROM

T.Y.B.Sc (Computer Science)

Abstract

With the introduction of quantum computing on the horizon computer security

organizations are stepping up research and development to defend against a new kind of

computer power. Quantum computers pose a very real threat to the global information

technology infrastructure of today. Many security implementations in use are based on

the difficulty for modern-day computers to perform large integer factorization. Utilizing

a specialized algorithm such as mathematician Peter Shor’s, a quantum computer can

compute large integer factoring in polynomial time versus classical computing’s sub-

exponential time. This theoretical exponential increase in computing speed has prompted

computer security experts around the world to begin preparing by devising new and

improved cryptography methods. If the proper measures are not in place by the time full-

scale quantum computers are being produced, the world’s governments and major

enterprises could suffer from security breaches and the loss of massive amounts of

encrypted data. Cryptographers are discussing alternatives to today’s methods and have

agreed that there are four major candidates that would provide immunity from a quantum

computer attack. The four possible replacement methods include: error-correcting codes,

hash-functions, lattice cryptography systems, and multivariate public-key cryptography

system

IntroductionComputer security and protecting valuable information has long been a delicate

subject in the world of information technology. With the development of the Internet,

companies had to ensure that customer data as well as their internal private data was

protected from outside intrusions; the secure socket layer (SSL) protocol was the first

step toward allowing for the secure transmission of information from client to server

and vice versa. Data encryption became a requirement for day-to-day operations of any

organization connected to the Internet and thus the world of big-business cryptography

exploded

Cryptography, while not a new practice, grew exponentially in popularity as

more computers came “online” and companies began to realize competing in the global

market of the Internet was becoming a necessity. Encryption of data for many IT

systems today relies on public-key cryptography. The concept of public-key

cryptography was introduced by Whitfield Diffie and Martin Hellman in 1976 (Diffie

& Hellman, 1976). This new method of encryption had two main purposes, encryption

and digital signatures. It entails that each person (or communicating system) gets a pair

of keys, one was dubbed the public key and the other was named the private key. The

public key is shared between the two parties and is used for identifying the end-user

while the private key remains a secret and is never transmitted. Encrypted information

is sent using the public key to identify the source but only a receiver that possesses the

private key is able to decode the message. Unfortunately, the private key, while kept a

secret from prying eyes, is linked to the public key through a mathematical algorithm

therefore presenting a weakness to this system. This “weakness” in the system became

less and less of an issue due to the complexity of being able to solve the algorithm for

common computer systems. Even when utilizing a “brute force” technique (a.k.a.

systematically trying every combination of letters, numbers, and symbols), a strongly

ciphered public-key encryption system remains untouchable by even today’s most

powerful computer systems. The complexity of arriving at a solution arises due to the

inability for today’s processors to factor increasingly larger and larger integers, the basis

for the strongest breeds of encryption to date. To put this into a clearer context,

researchers were able to successfully factor a 768-bit RSA modulus cipher using the

number field sieve factoring method which would have taken a single-core 2.2 GHz

AMD Opteron processor with 2 GB of RAM over 1500 years to process (Klein Jung, et

al., 2010). It is this length of time that makes breaking today’s most heavily encrypted

data such a daunting and near impossible endeavour.

Quantum computing presents the first serious risk associated with actually

providing a means to break the most sophisticated of encryption systems in use today.

Utilizing atoms as pseudo bits of digital information, binary processing can be achieved

in a much more sophisticated fashion than is currently possible. Quantum bits, or

“qubits”, can take on the value of 1 or 0 as the traditional digital bit does; the complexity

arises as these qubits can also take on the value for everything between 0 and 1 at the

same time. This ability allows for a single qubit to represent every possible value

between 0 and 1 simultaneously, thus permitting computations to be calculated in

parallel on every one of these values as well.

In this paper, we will explore the history of quantum computing theory and the

development of varying techniques researchers are using to accomplish building a fully

functional system today. We’ll also explore the practical applications for quantum

computers and how they could affect current cryptographic systems as well as how they

are shaping the development of new systems. Cryptography will remain a central focal

point throughout delving into varying theories and methods cryptography organizations

are testing and implementing to defend against massive computing power at an atomic

scale. Current methods of cryptography are in danger due to this unique type of threat

and security organizations are realizing that they can no longer sit idle waiting for the

first quantum computer to be produced before acting. Preparation must begin now if the

world is to have any hope of converting over from the encryption techniques that we

rely on every day to new and improved quantum computer proof methods of tomorrow.

The History of Quantum Computing

The idea of a quantum computer began in the early 1980s and was conceived by

Paul Benioff, Charles Bennett, David Deutsch, Richard Feynman, and Yuri Manin

(Bacon & Leung, 2007). The original idea of such a system was purely theoretical and

scientists were basing these ideas on years of research into quantum theory and

information science. Scientists speculated that if technology was to continue following

Moore’s Law (the observation that steady technological improvements in

miniaturization leads to a doubling of the density of transistors on new integrated

circuits every 18 months (Moore's law)) that eventually the size of circuitry on chips

would be reduced to the size of no more than a few atoms. At this size, the workings of

an integrated circuit would be governed by the laws of quantum mechanics and thus the

researchers began to question if a new kind of computer could be created based around

the study of quantum physics.

In 1982, Richard Feynman constructed an abstract model showing how a

quantum system would function and be used for doing computational work. Basically,

a classical bit in computing is used to represent two different states of an information

processing machine. One could refer to a bit as a type of light switch; rather, a bit can

be either “on” or “off”. When a bit is “off” it is said to have the value of 0 whereas when

it is “on” it is said to have the value of 1. Quantum computer bits, or “qubits”, can be a

0 or a 1 at the same time. A particle found to be in this quantum state is said to be in

“superposition”; however, the particle will take on a single location once someone or

something observes the particle. The speed advantages of quantum computers over

classical computers were not realized until the early 1990s when David Deutsch and

Richard Jozsa demonstrated that given a quantum computer utilizing the function.

f : {0,1}n →{0,1}, we can be assured that the function is either constant (0 or 1), or

balanced (returns 1 for half of the results and 0 for the other half) (Deutsch & Jozsa,

1992). Deutsch and Jozsa’s algorithm provided a man by the name of Peter Shor the

basis for constructing one of the most well-known and important quantum algorithms

the computing world had ever seen.

In 1994, mathematician Peter Shor formulated a quantum algorithm designed to

be used on a quantum computer to process integer factorization computations. Integer

factorization at its basis can be taken as “Given an integer N find its prime factors”.

Shor realized that utilizing his theory a quantum computer could “efficiently factor and

compute” the solution to large integer factoring problems (Bacon & Leung, 2007).

Shor’s quantum theory works in polynomial-time unlike the classical factoring

algorithm, the general number field sieve, which factors an L-bit number N in time

O(exp(cL1/3 log2/3 L)). The algorithm works by determining the period of the function

f(x) = ax mod N where a is a random chosen number by the quantum computer with no

factors in common with N. After obtaining this period, using number-theoretic

techniques we can now factor N with a high probability of success (Shor, 1997). Using

this method for factoring numbers results in a significant exponential difference in time

versus the general number field sieve method with time being O((log N)3). These

extreme differences in time between classical computers and Shor’s algorithm on a

quantum computer can be seen in Figure 1. The NFS curve on the left is data gathered

from a previous world record, factoring a 530-bit number in one month on 104 PCs and

workstations in 2003 (Van Meter, Itoh, & Ladd, 2005). The right curve is speculative

based on 1,000 times as much computing power of these classical computers which

works out to be around 100 PCs in 2018 based on Moore’s law. We can see how Shor’s

algorithm is much more efficient in factoring than anything that’s currently possible or

will be possible in the next decade.

Figure 1. Scaling of number field sieve (NFS) on classical computers and Shor’s algorithm for

factoring on a quantum computer, using Beckman-Chari-Devabhaktuni-Preskill modular

exponentiation with various clock rates. Both horizontal and vertical axes are log scale. The

horizontal axis is the size of the number being factored (Van Meter, Itoh, & Ladd, 2005).

In 2001, Shor’s algorithm was put to the test by IBM researchers using room

temperature liquid-state nuclear magnetic resonance techniques to manipulate nuclei in

a molecule as quantum bits (Vandersypen, Steffen, Breyta, Yannoni, Sherwood, &

Chuang, 2001). As insignificant as it sounds, the researchers, utilizing a very primitive

quantum computer, were able to apply Shor’s algorithm to successfully factor 15 giving

the results of 3 and 5. They noted that their experiment could be scaled to a much larger

system with more than the 7 qubits they utilized but that it was intended to simply

demonstrate the techniques for the control and modeling of quantum computers for the

future

Quantum cryptography has been at the forefront of purposes for developing

quantum computers since the early 1980s. Due to the way the qubits behave when

observed, it opened up the possibility of creating a new form of quantum

communication between two parties. Where before, transmission of messages relied on

the receiver having an encryption key to decode an encoded message, researchers were

able to utilize photons to send a message and detect whether the message had been

viewed along the way. While this method does not prevent an eavesdropper from

reading the message, it created a way for both the sender and receiver to know if the

message had been intercepted.

A cryptographic application of a quantum system was one of the earliest ideas

involved with quantum computation and can be accredited to Stephen Wiesner in the

1960s. Wiesner developed a theory that was meant to prevent counterfeiting of money

using the laws of physics as a basis for protection (Bacon & Leung, 2007). His method

relied on information that is encoded in quantum states thereby being able to prevent

any outside party from accessing said information without disturbing the state. This

property of quantum information has given birth to a new method of information

exchange and other companies are investing in it to develop new products giving users

the utmost security of knowing if their critical data has been intercepted or viewed by

an unintended audience outside the exchange.

In 2002 and 2003, a Swiss company called id Quantique and an American

company called MagiQ Technologies, both developed commercial communication

products leveraging this technology for message transmission and receipt (Bacon &

Leung, 2007). These two companies are noted as marketing the very first quantum key

distribution systems. This could be the preferred method for secure communication of

the future instead of relying on a receiver held private key utilized in systems based on

the famous RSA crypto architecture for example. Larger organizations are also starting

to invest in quantum technologies, such as Hewlett-Packard, Microsoft, IBM, Lucent,

Toshiba and NEC; each have active research programs exploring how quantum

cryptography can be leveraged into their future business models (Bacon & Leung,

2007).

Regarding the aforementioned past research on quantum systems, “the short-

range business concerns of these developments remains unclear at the moment, but

experience has shown that the industry needs many years to replace legacy systems –

you cannot easily change ATMs and mainframe applications,” says Andrea Simmons

of Computer Weekly (Simmons, 2009). At this point, it’s not really an “if quantum

computers come to be” it’s “when they come to be”, will we be prepared? Research is

happening right now and scientists are getting closer to understanding the hardest

questions plaguing quantum computers; in the next section, we’ll take a look at this

research and some of the concerns and issues scientists are dealing with today.

Quantum Computing Today

Many advances in research being conducted on the creation of quantum

computers have lead industry and educational sector experts to believe that the

technology could be just around the corner. “Quantum computational devices with

calculating power greater than any of today’s conventional computers could be just a

decade away, says Bristol University physicist and electrical engineer Mark

Thompson,” (Docksai, 2011). Thompson aided in the development of two quantum

photonic computer chips which process photons to provide what IBM accomplished in

2001 with factoring the integer 15. This shows how technology advancements over the

past ten years has allowed past research findings to be minimized into a form factor that

can be utilized inside a much more acceptable size for consumers.

Organizations delving into quantum computer development are faced with a

number of different methods for developing quantum systems. These varying methods

each have their advantages and disadvantages regarding factors such as scalability,

longevity, and accuracy. Current forerunners in the area include ion-trap quantum

computers and NMR (nuclear magnetic resonance) quantum computers. First, we’ll

look at the ion-trap method and the current research being done to overcome the

complexities of scaling this method.

Ion-trap quantum computation is currently being researched at the National

Institute of Standards and Technology. An ion-trap can be described as using a line of

N trapped ions where each ion has two stable or metastable states; there are also N laser

beam pairs each interacting with one of the ions (a qubit) (Steane, 1996). It is these laser

beams that essentially program the qubits by providing a pulsing form of a quantum

logic gate. This is very similar to how a transistor works by switching a classical bit

from 1 to 0 and vice versa. Scientists are looking into a “quantum charge-coupled

device” (QCCD) architecture which is essentially a large collection of these ion-traps.

The QCCD method was proposed as a possible solution to the limitations researchers

faced on scaling a single ion-trap to be able to confine a large number of ions. Incredible

complexities arise technologically and “scaling arguments suggest that this scheme is

limited to computations on tens of ions (Kielpinski, Monroe, & Wineland, 2002). The

advantage of the QCCD method is that by altering voltages of the ion-traps researchers

can confine a set number of ions in each trap or even transport ions to another trap

(Kielpinski, Monroe, & Wineland, 2002). This allows for the scaling of a multiple ion-

trap quantum computer to be achieved much easier. QCCD ion-trap quantum computers

still have a long way to go to be a feasible and scalable method for developing these

machines. If we look at the size of the very first computers taking up multiple rooms

and compare this with where these quantum computers are, we can see that in a few

decades these machines will follow Moore’s law and decrease substantially in size

while increasing in power. “Build the first one and in 25 years, they will be 25% of the

size. I bet that, after the first quantum computer, the cost of one 10 years later will be

significantly reduced,” (Docksai, 2011).

The United States Government is funding the research at NIST on QCCD ion-

trap quantum computers because they realize the implications that could arise should a

country gain quantum computational power before them. The country that first develops

a full-scale quantum computer will have the power to crumble current encryption

methods and expose an unbelievable amount of data in a very short amount of time.

Achieving the technology to build a fully functional quantum computer “is among the

great technological races of the 21st century, a race whose results may profoundly alter

the manner in which we compute (Bacon & Leung, 2007). In the book, The Quest for

the Quantum Computer by Julian Brown and David Deutsch, the authors note that “if

anyone could build a full-scale quantum computer, it’s possible that he or she would be

able to access everything from your bank account to the Pentagon’s most secret files.

It’s no surprise, then, that significant funds backing this line of research have come from

such organizations as the U.S. Department of Defense, the National Security Agency,

NATO, and the European Union,” (Brown & Deutsch, 2000).

Another method being explored is based on nuclear magnetic resonance (NMR)

spectroscopy. The aforementioned research done at IBM factoring 15 was done using a

NMR based quantum computer. NMR computers rely on the two spin states of a spin-

1/2 atomic nucleus in a magnetic field; different atoms in a molecule can be singled out

and thus a molecule can actually be used as a quantum computer (Jones, 1998). Each

spin-1/2 nucleus provides a single qubit that is manipulated by multiple logic gates

provided by radio frequency fields. To increase the capabilities of this quantum

computer the scientists need only make these logic gates more complex. As their

complexity increases, the qubits can affect other qubits in the computer and thus allows

them to work together as a single computational system. The advantages of this method

arise when looking at the coherence of the qubits in the system. Coherence describes

the ability for the qubits to retain their states despite “noise” from the external

environment. “Slow decoherence is one of the primary attractions of an NMR quantum

computer,” (Gershenfeld & Chuang, 1997).

There are a plethora of different methods, other than the ones mentioned here,

that scientists are working on to bring quantum computing to the next level. As research

continues in these areas, we inch closer to the day that a fully functional quantum

processing machine is in our grasp. The latest research has yielded better error

correction with the qubit states, faster quantum algorithms, and led to the evolution of

decoherence and entanglement in quantum computing

Computer Security Organizations and Quantum Computing

Cryptography specialists are beginning to take notice that quantum computing

may not be too far away. Many of the IT industry’s top cryptography experts have

predicted that a full-scale quantum computer could manifest in as little as 10 years

(Heger, 2009). The PQCrypto (Post-Quantum Cryptography) conference was created in

2006 to address and discuss the dangers the world of computer security may face with

the successful creation of quantum computers. Researchers from all over the world

flocked to meet in Leuven, Belgium to begin discussions on possible alternatives to the

most widely used encryption systems. The two most popular encryption methods in use

today are RSA and elliptic-curve cryptography (ECC) (Heger, 2009). These two

methods both rely on public-key architecture and digital signatures to provide a means

to communicate securely between two parties in digital communication. As we’ve

covered earlier in this paper, the entire public-key architecture is at risk of being

obliterated by the computing power of a single, albeit low-powered, quantum computer.

To gain a better understanding of where experts see cryptography technology heading,

let’s take a look at a few of the quantum-proof methods that have been discussed at past

PQCrypto conferences.

The idea of Lamport signatures was first mentioned in a paper by Leslie Lamport

from SRI International on October 18, 1979 (Lamport, 1979) where he discusses the

use of one-way hash functions to generate signatures. The idea behind this type of

encryption is that a function can be created to be irreversible in nature. The Lamport

crypto-scheme is a one-time signature scheme, therefore each time a signed message is

generated a new signature must also be created. Utilizing a hash function to implement

this type of crypto system, the text of a message is efficiently reduced to a much shorter

string of bits which is the message signature (Heger, 2009). Researchers Ray Perlner

and David Cooper (2009) from the National Institute of Standards and Technology

explain the basics to the Lamport signature in their paper on quantum resistant public-

key cryptography:

In the simplest variant of Lamport signatures, the signer generates two high-

entropy secrets, S0,k and S1,k, for each bit location, k, in the message digest that will be

used for signatures. These secrets (2n secrets are required if the digest is n bits long)

comprise the private key. The public key consists of the images of the secrets under f,

i.e., f(S0,k) and f(S1,k), concatenated together in a prescribed order (lexicographically by

subscript for example). In order to sign a message, the signer reveals half of the secrets,

chosen as follows: if bit k is a zero, the secret S0,k is revealed, and if it is one, S1,k is

revealed. The revealed secrets, concatenated together, comprise the signature. (Perlner

& Cooper, 2009)

The reason the Lamport signature is a one-time used architecture is due to the

signing method actually revealing a small amount of information about the private key.

This leak of information is not, however, enough for an attacker to build and sign a

forged message but subsequent messages must be accompanied by a newly generated

key to remain secure. The performance of a system like the Lamport signature is purely

dependent on the one-way function the signer chooses to implement. The original

implementation was greatly improved by others as the original requirement of running

numerous hash functions to generate signatures grew exponentially (Merkle, 1988).

Another possible candidate to replace public-key systems comes in the form of

multivariate public-key cryptosystems (MPKCs) and bases its strength on multivariable

nonlinear equations (Heger, 2009). Quantum computers share a weakness with their

classical brethren when trying to solve problems said to be “NP-complete”. Jintai Ding

(2008), a top researcher in MPKCs, notes that “It’s difficult to explain what NP-

complete means, but it just means very, very difficult. It is exponential, meaning that as

the size of a problem increases, the time to solve it increases exponentially. And

quantum computers have not yet been able to defeat NP-complete types of problems,”

(IEEE Spectrum, 2008). A computing machine is only as powerful as the mathematics

behind said machine and when the mathematics has not yet been developed to solve a

problem with efficiency, the problem is said to be NP-hard. Whereas traditional RSA

type cryptosystems rely on mathematics developed in the 17

th and 18th centuries (number theory), MPKCs use 20th century algebraic

geometry for their basis (Ding & Schmidt, 2006). Ding and Schmidt (2006) state on

MKPCs that, “the method relies on the proven theorem that solving a set of

multivariable polynomial equations over a finite field is in general an NP-hard

problem,” (Ding & Schmidt, 2006).

PQCrypto attendees also discussed lattice based cryptography systems which

researchers believe can be implemented in a way that makes for solving the algorithm

an NP-complete problem. “An n-dimensional lattice is the set of vectors that can be

expressed as the sum of integer multiples of a specific set of n vectors, collectively

called the basis of the lattice,” (Perlner & Cooper, 2009). The NP-complete issue for

cracking this type of cryptography arises when increasing the dimensions of the lattices

and trying to solve the shortest vector problem (Ajtai, 1998) as well as the closest vector

problem (van Emde Boas, 1981). Both problems revolve around the difficulty of solving

for the shortest vector to a random non-lattice vector.

The fourth and final candidate being researched is encryption schemes based on

the use of error-correcting codes. Basically, the idea behind this type of encryption is

that the sender of the message encrypts the message with noise, or random additional

information, therefore obfuscating the original message. Only the receiver has the

ability to “sift” through the information to deduce the true content of the message. The

first error-correcting encryption scheme was devised by Robert J. McEliece about one

year after the RSA encryption technique was proposed (Joye, 2009). Many refer to

error-correcting encryption schemes as “code-based cryptography” and researchers

Bernstein, Lange, and Peters (2011) state that “code-based cryptography has lately

received a lot of attention because it is a good candidate for public-key cryptography

that remains secure against attacks by a quantum computer,” (Bernstein, Lange, &

Peters, 2011). A major drawback in utilizing this scheme is that the public key is too

large due to the excess data for efficient communication; however, research continues

in an attempt to minimize the required information for public key creation to make this

technique a viable solution.

Of these four completely different techniques, is there a single encryption

scheme that will provide security where the aging encryption schemes fail? Jintai Ding

(2008) states on the matter, “no, I cannot really specify one area. These four systems

are all very different and each has its own advantages and disadvantages,” (IEEE

Spectrum, 2008). Researchers from all over the world continue to meet at the PQCrypto

conference, held every few years, to discuss the latest findings on each of the

aforementioned methods as well as new potential candidates. While they may disagree

on when quantum computing is coming, they all seem to agree that there is a sense of

urgency to develop encryption schemes that will be ready for when it does:

Quantum cryptography may be five years ahead; quantum computing may be 15

years away. Progress in building quantum computers tends to develop in steps as

researchers find new methods, so it is slow, but the theory is solid. We should be starting

now to evaluate the impact. (Simmons, 2009)

CONCLUSION

Quantum cryptography ensure secure communication by providing security

based on the fundamental law of physics, intead of the current state of mathematical

algorithms or computing technology unlike classical encryption algorithm quantum

cryptography does not depend factoring large integers into primes but on the

fundamental principles of quantum physics. Quantum cryptography is more secure,

because an intruder is not able to replicate the photon to recreate the key.

Integrating QKD in TLS protocol will ensure financial transaction. Instead of

using RSA, in TLS protocol .We can use Quantum Cryptography securely exchange the

secret data and avoid an attack of intruder

ACKNOWLEDGMENT

I am using this opportunity to express my gratitude to everyone who supported

me throughout the course of this seminar report. I am thankful for their aspiring

guidance, invaluably constructive criticism and friendy advice during the seminar work.

I deeply express my sincere thanks to my guide Mr. Shrikant Ghodke for his support

and guidance.

References

Ajtai, M. (1998). The shortest vector problem in L2 is NP-hard randomized

reductions. 30th ACM Symposium on Theory of Computing (pp. 10-19). New York:

ACM.

Bacon, D., & Leung, D. (2007, September). Toward a World with Quantum

Computers. Communications of the ACM, 50(9), pp. 55-59.

Bernstein, D. J., Lange, T., & Peters, C. (2011). Wild McEliece. Lecture Notes in

Computer Science, 6544/2011, 143-158.

Brown, J. R., & Deutsch, D. (2000). The quest for the quantum computer. New York,

NY: Touchstone (Simon & Schuster, Inc.).

Deutsch, D., & Jozsa, R. (1992). Rapid Solution of Problems by Quantum

Computation. Proceedings of the Royal Society of London Series A - Mathematical

Physical and Engineering Sciences (pp. 553-558). London: Royal Society of London.

Diffie, W., & Hellman, M. E. (1976). New directions in cryptography. IEEE

Transactions on Information Theory, 22(6), 644-654.

Ding, J., & Schmidt, D. (2006). Multivariable public key cryptosystems.

Contemporary Mathematics(419), 79-94.

Docksai, R. (2011). Computers making the Quantum Leap. Futurist, 45(3), pp. 10-11.

Gershenfeld, N. A., & Chuang, I. L. (1997, January 17). Bulk Spin-Resonance

Quantum Computation. Science, 275(5298), 350-356.

Heger, M. (2009, January ). Cryptographers Take On Quantum Computers. Retrieved

July 24, 2011, from IEEE Spectrum:

http://spectrum.ieee.org/computing/software/cryptographers-take-on-quantum-

computers

IEEE Spectrum. (2008, November). Q&A with post-quantum computing cryptography

researcher Jintai Ding. Retrieved August 8, 2011, from IEEE Spectrum:

http://spectrum.ieee.org/computing/networks/qa-with-postquantum-computing-

cryptography-researcher-jintai-ding/0

Jones, J. (1998). Fast searches with nuclear magnetic resonance computers. Science,

280(5361), 229.

Joye, M. (2009). Identity-based cryptography. Amsterdam: IOS Press

Kielpinski, D., Monroe, C., & Wineland, D. J. (2002, June 13). Architecture for a

large-scale ion-trap quantum computer. Nature, 417(6890), 709-711.

Kleinjung, T., Aoki, K., Franke, J., Lenstra, A. K., Thomé, E., Bos, J. W., et al. (2010,

August). Factorization of a 768-bit RSA modulus. CRYPTO'10 Proceedings of the

30th annual conference on Advances in cryptology (pp. 333-350). Berlin, Heidelberg:

Springer-Verlag.

Lamport, L. (1979, October 18). Constructing digital signatures from a one-way

function. In Technical Report CSL-98. Menlo Park, CA: SRI International.

Merkle, R. C. (1988). A digital signature based on a convential encryption function.

CRYPTO '87 A Conference on the Theory and Applications of Cryptographic

Techniques on Advances in Cryptology (pp. 369-378). London, UK: Springer-Verlag.

Moore's law. (n.d.). The American Heritage® Science Dictionary. Retrieved August 5,

2011, from Dictionary.com:

http://dictionary.reference.com/browse/moore%27s%20law

Perlner, R. A., & Cooper, D. A. (2009). Quantum resistant public key cryptography: a

survey. IDtrust '09 Proceedings of the 8th Symposium on Identity and Trust on the

Internet (pp. 85-93). New York, NY: Association for Computing Machinery.

Shor, P. (1997). Polynomial-time algorithms for prime factorization and discrete

logarithms on a quantum. SIAM Journal on Computing, 26, 1484-1509.

Simmons, A. (2009, May 19). Quantum implications for IT security. Computer

Weekly, pp. 14-15.

Steane, A. (1996). The ion trap quantum information processor. Applied Physics B:

Lasers and Optics, 64(6), 623-643.

van Emde Boas, P. (1981). Another NP-complete problem and the complexity of

computing short vectors in a lattice. Netherlands: University of Amsterdam,

Department of Mathematics.

Van Meter, R., Itoh, K. M., & Ladd, T. D. (2005). Architecture-Dependent Execution

Time of Shor's Algorithm. Retrieved from EBSCOhost.

Vandersypen, L., Steffen, M., Breyta, G., Yannoni, C., Sherwood, M., & Chuang, I.

(2001, December). Experimental realization of Shor's quantum factoring algorithm

using nuclear magnetic resonance. NATURE, 414(6866), 883-887.

Wood, L. (2010, December 17). The clock is ticking for encryption. Retrieved August

8, 2011, from Computerworld:

http://www.computerworld.com/s/article/9201281/The_clock_is_ticking_on_encrypti

on