[go: nahoru, domu]

US20170099134A1 - Differential power analysis - resistant cryptographic processing - Google Patents

Differential power analysis - resistant cryptographic processing Download PDF

Info

Publication number
US20170099134A1
US20170099134A1 US15/236,739 US201615236739A US2017099134A1 US 20170099134 A1 US20170099134 A1 US 20170099134A1 US 201615236739 A US201615236739 A US 201615236739A US 2017099134 A1 US2017099134 A1 US 2017099134A1
Authority
US
United States
Prior art keywords
key
analog
measurements
digital converter
clock
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US15/236,739
Inventor
Paul C. Kocher
Joshua M. Jaffe
Benjamin C. Jun
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Cryptography Research Inc
Original Assignee
Cryptography Research Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US09/224,682 external-priority patent/US6304658B1/en
Priority claimed from US09/326,222 external-priority patent/US6327661B1/en
Priority claimed from US09/324,798 external-priority patent/US6278783B1/en
Application filed by Cryptography Research Inc filed Critical Cryptography Research Inc
Priority to US15/236,739 priority Critical patent/US20170099134A1/en
Assigned to CRYPTOGRAPHY RESEARCH, INC. reassignment CRYPTOGRAPHY RESEARCH, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: JAFFE, JOSHUA M., JUN, BENJAMIN C., KOCHER, PAUL C.
Publication of US20170099134A1 publication Critical patent/US20170099134A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/70Protecting specific internal or peripheral components, in which the protection of a component leads to protection of the entire computer
    • G06F21/71Protecting specific internal or peripheral components, in which the protection of a component leads to protection of the entire computer to assure secure computing or processing of information
    • G06F21/77Protecting specific internal or peripheral components, in which the protection of a component leads to protection of the entire computer to assure secure computing or processing of information in smart cards
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L9/00Cryptographic mechanisms or cryptographic arrangements for secret or secure communications; Network security protocols
    • H04L9/06Cryptographic mechanisms or cryptographic arrangements for secret or secure communications; Network security protocols the encryption apparatus using shift registers or memories for block-wise or stream coding, e.g. DES systems or RC4; Hash functions; Pseudorandom sequence generators
    • H04L9/0618Block ciphers, i.e. encrypting groups of characters of a plain text message using fixed encryption transformation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • G06F1/04Generating or distributing clock signals or signals derived directly therefrom
    • G06F1/06Clock generators producing several clock signals
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • G06F1/26Power supply means, e.g. regulation thereof
    • G06F1/266Arrangements to supply power to external peripherals either directly from the computer or under computer control, e.g. supply of power through the communication port, computer controlled power-strips
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • G06F1/26Power supply means, e.g. regulation thereof
    • G06F1/32Means for saving power
    • G06F1/3203Power management, i.e. event-based initiation of a power-saving mode
    • G06F1/3206Monitoring of events, devices or parameters that trigger a change in power modality
    • G06F1/3215Monitoring of peripheral devices
    • G06F1/3225Monitoring of peripheral devices of memory devices
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/70Protecting specific internal or peripheral components, in which the protection of a component leads to protection of the entire computer
    • G06F21/71Protecting specific internal or peripheral components, in which the protection of a component leads to protection of the entire computer to assure secure computing or processing of information
    • G06F21/75Protecting specific internal or peripheral components, in which the protection of a component leads to protection of the entire computer to assure secure computing or processing of information by inhibiting the analysis of circuitry or operation
    • G06F21/755Protecting specific internal or peripheral components, in which the protection of a component leads to protection of the entire computer to assure secure computing or processing of information by inhibiting the analysis of circuitry or operation with measures against power attack
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F7/00Methods or arrangements for processing data by operating upon the order or content of the data handled
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K19/00Record carriers for use with machines and with at least a part designed to carry digital markings
    • G06K19/06Record carriers for use with machines and with at least a part designed to carry digital markings characterised by the kind of the digital marking, e.g. shape, nature, code
    • G06K19/067Record carriers with conductive marks, printed circuits or semiconductor circuit elements, e.g. credit or identity cards also with resonating or responding marks without active components
    • G06K19/07Record carriers with conductive marks, printed circuits or semiconductor circuit elements, e.g. credit or identity cards also with resonating or responding marks without active components with integrated circuit chips
    • G06K19/073Special arrangements for circuits, e.g. for protecting identification code in memory
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K19/00Record carriers for use with machines and with at least a part designed to carry digital markings
    • G06K19/06Record carriers for use with machines and with at least a part designed to carry digital markings characterised by the kind of the digital marking, e.g. shape, nature, code
    • G06K19/067Record carriers with conductive marks, printed circuits or semiconductor circuit elements, e.g. credit or identity cards also with resonating or responding marks without active components
    • G06K19/07Record carriers with conductive marks, printed circuits or semiconductor circuit elements, e.g. credit or identity cards also with resonating or responding marks without active components with integrated circuit chips
    • G06K19/073Special arrangements for circuits, e.g. for protecting identification code in memory
    • G06K19/07309Means for preventing undesired reading or writing from or onto record carriers
    • G06K19/07363Means for preventing undesired reading or writing from or onto record carriers by preventing analysis of the circuit, e.g. dynamic or static power analysis or current analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q20/00Payment architectures, schemes or protocols
    • G06Q20/30Payment architectures, schemes or protocols characterised by the use of specific devices or networks
    • G06Q20/34Payment architectures, schemes or protocols characterised by the use of specific devices or networks using cards, e.g. integrated circuit [IC] cards or magnetic cards
    • G06Q20/341Active cards, i.e. cards including their own processing means, e.g. including an IC or chip
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q20/00Payment architectures, schemes or protocols
    • G06Q20/38Payment protocols; Details thereof
    • G06Q20/382Payment protocols; Details thereof insuring higher security of transaction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q20/00Payment architectures, schemes or protocols
    • G06Q20/38Payment protocols; Details thereof
    • G06Q20/40Authorisation, e.g. identification of payer or payee, verification of customer or shop credentials; Review and approval of payers, e.g. check credit lines or negative lists
    • G06Q20/409Device specific authentication in transaction processing
    • GPHYSICS
    • G07CHECKING-DEVICES
    • G07FCOIN-FREED OR LIKE APPARATUS
    • G07F7/00Mechanisms actuated by objects other than coins to free or to actuate vending, hiring, coin or paper currency dispensing or refunding apparatus
    • G07F7/08Mechanisms actuated by objects other than coins to free or to actuate vending, hiring, coin or paper currency dispensing or refunding apparatus by coded identity card or credit card or other personal identification means
    • GPHYSICS
    • G07CHECKING-DEVICES
    • G07FCOIN-FREED OR LIKE APPARATUS
    • G07F7/00Mechanisms actuated by objects other than coins to free or to actuate vending, hiring, coin or paper currency dispensing or refunding apparatus
    • G07F7/08Mechanisms actuated by objects other than coins to free or to actuate vending, hiring, coin or paper currency dispensing or refunding apparatus by coded identity card or credit card or other personal identification means
    • G07F7/0806Details of the card
    • G07F7/0813Specific details related to card security
    • G07F7/082Features insuring the integrity of the data on or in the card
    • GPHYSICS
    • G07CHECKING-DEVICES
    • G07FCOIN-FREED OR LIKE APPARATUS
    • G07F7/00Mechanisms actuated by objects other than coins to free or to actuate vending, hiring, coin or paper currency dispensing or refunding apparatus
    • G07F7/08Mechanisms actuated by objects other than coins to free or to actuate vending, hiring, coin or paper currency dispensing or refunding apparatus by coded identity card or credit card or other personal identification means
    • G07F7/10Mechanisms actuated by objects other than coins to free or to actuate vending, hiring, coin or paper currency dispensing or refunding apparatus by coded identity card or credit card or other personal identification means together with a coded signal, e.g. in the form of personal identification information, like personal identification number [PIN] or biometric data
    • G07F7/1008Active credit-cards provided with means to personalise their use, e.g. with PIN-introduction/comparison system
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09CCIPHERING OR DECIPHERING APPARATUS FOR CRYPTOGRAPHIC OR OTHER PURPOSES INVOLVING THE NEED FOR SECRECY
    • G09C1/00Apparatus or methods whereby a given sequence of signs, e.g. an intelligible text, is transformed into an unintelligible sequence of signs by transposing the signs or groups of signs or by replacing them by others according to a predetermined system
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L9/00Cryptographic mechanisms or cryptographic arrangements for secret or secure communications; Network security protocols
    • H04L9/002Countermeasures against attacks on cryptographic mechanisms
    • H04L9/003Countermeasures against attacks on cryptographic mechanisms for power analysis, e.g. differential power analysis [DPA] or simple power analysis [SPA]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2207/00Indexing scheme relating to methods or arrangements for processing data by operating upon the order or content of the data handled
    • G06F2207/72Indexing scheme relating to groups G06F7/72 - G06F7/729
    • G06F2207/7219Countermeasures against side channel or fault attacks
    • G06F2207/7223Randomisation as countermeasure against side channel attacks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2207/00Indexing scheme relating to methods or arrangements for processing data by operating upon the order or content of the data handled
    • G06F2207/72Indexing scheme relating to groups G06F7/72 - G06F7/729
    • G06F2207/7219Countermeasures against side channel or fault attacks
    • G06F2207/7266Hardware adaptation, e.g. dual rail logic; calculate add and double simultaneously
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L2209/00Additional information or applications relating to cryptographic mechanisms or cryptographic arrangements for secret or secure communication H04L9/00
    • H04L2209/04Masking or blinding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L2209/00Additional information or applications relating to cryptographic mechanisms or cryptographic arrangements for secret or secure communication H04L9/00
    • H04L2209/08Randomization, e.g. dummy operations or using noise
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L2209/00Additional information or applications relating to cryptographic mechanisms or cryptographic arrangements for secret or secure communication H04L9/00
    • H04L2209/56Financial cryptography, e.g. electronic payment or e-cash
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L2209/00Additional information or applications relating to cryptographic mechanisms or cryptographic arrangements for secret or secure communication H04L9/00
    • H04L2209/80Wireless
    • H04L2209/805Lightweight hardware, e.g. radio-frequency identification [RFID] or sensor

Definitions

  • This application relates generally to cryptographic systems and, more specifically, to determining useful information about a cryptographic system by external monitoring of its cryptographic operations.
  • cryptographic devices can be attacked using information gathered by observing the timing of comparison operations performed by such devices during their operation. For example, if a MAC (Message Authentication Code) algorithm is strong and the key is secure, forging a MAC should require O(2 ⁇ n) attempts (where n is the MAC length in bits), but a device using a vulnerable MAC validation process is vulnerable to an O(n) timing attack.
  • MAC Message Authentication Code
  • Timing equalization system itself can be vulnerable to non-timing attacks, for example by analyzing power consumption to detect the start of processing delays. It would therefore be advantageous to protect the devices' internal operations themselves instead of (or in addition to) simply externally masking the devices' timing (or other) fluctuations.
  • a data collection and analysis system is configured with an analog-to-digital converter connected to measure the device's consumption of electrical power, or some other property of the target device, that varies during the device's processing. As the target device performs cryptographic operations, data from the A/D converter are recorded for each cryptographic operation. The stored data are then processed using statistical analysis, yielding the entire key, or partial information about the key that can be used to accelerate a brute force search or other attack.
  • Particularly serious sources of leakage include the device's power consumption and electromagnetic radiation. Observation of the microprocessor's power consumption can reveal whether the jumps are taken. Observation of the power consumption and/or timing can reveal whether the carried bits in the key rotates of each DES round equal zero or one. Operations that change the device state can reveal information about the initial and final states of the operations. Signals radiated from the address and data bus lines connecting a device to memory can be detected and analyzed to gain information which in turn can compromise the keys. Variations between individual transistors in an integrated circuit, variations in the electrical properties of wires within a chip, variations in the amount of electromagnetic radiation emitted by different wires, etc. can all provide variations that can be analyzed statistically to determine secret keys.
  • the attack collects a small amount of data related to the key each time the token is observed performing a cryptographic operation involving the key.
  • the amount of information known about the key is increased by collecting and statistically correlating (or combining) data from multiple observations of the token as it performs operations involving the key (or related key).
  • such observations may contain signal (i.e., information correlated usefully to the key). However, such observations also contain noise (i.e., information and error that hinder or are irrelevant to determination of the key).
  • the quality of the information gained from these observations is characterized by a “signal to noise” (or S/N) ratio, which is a measure of the magnitude of the signal compared to the amount of noise.
  • the number of operations needed to recover the key depends on the measurement and analysis techniques, but is generally inversely proportional to the square of the S/N ratio.
  • the constant of proportionality also depends upon the amount of confidence required. For example, data of a relatively low confidence level may be acceptable if used to guide an optimized brute force search using statistical information about key bit values. If a countermeasure is used that decreases the signal or increases the amount of measurement noise by, for example, a factor of 300, the statistical attack can generally still succeed but would be expected to require roughly 90,000 times as many observations to extract the same amount of information about the key. An attack requiring 1,000 observations to recover a key before the S/N reduction would now yield the same level of confidence in the recovered key by using on the order of 90 million observations.
  • the signal-to-noise ratio may be improved by an attacker or evaluator using methods such as aligning these measurements so that the data points corresponding to a single point of interest can be compared and analyzed across a large number of observations.
  • Averaging data collected from many operations can be an effective means of amplifying signals and filtering out noise.
  • the evaluator guesses the value of some of the bits of the key, computes an average, checks whether expected biases appear (or uses another statistical technique and check for other expected effects) in collected data, and iterates this process with multiple guesses to recover the entire key. In some cases, disconnecting the power or resetting a device during an operation, may be helpful in compromising secrets by allowing a single leaky operation to be performed repeatedly.
  • FIG. 1 illustrates an exemplary apparatus for introducing noise into a cryptosystem.
  • FIG. 2 illustrates an exemplary apparatus for implementing clock skipping.
  • This application discloses both external monitoring techniques (e.g., attacks against cryptosystems), as well as techniques for using unpredictable information to protect cryptosystems against such external monitoring techniques.
  • the external monitoring techniques operate on the principle that information leaked from smart cards and other tamper resistant cryptographic devices can be statistically analyzed to determine keys or other secret data.
  • a data collection and analysis system is configured with an analog-to-digital converter connected to measure the device's consumption of electrical power, or some other property of the target device, that varies during the device's processing. As the target device performs cryptographic operations, data from the A/D converter are recorded for each cryptographic operation. The stored data are then processed using statistical analysis, yielding the entire key, or partial information about the key that can be used to accelerate a brute force search or other attack.
  • Particularly serious sources of leakage include the device's power consumption and electromagnetic radiation. Observation of the microprocessor's power consumption can reveal whether the jumps are taken. Observation of the power consumption and/or timing can reveal whether the carried bits in the key rotates of each DES round equal zero or one. Operations that change the device state can reveal information about the initial and final states of the operations. Signals radiated from the address and data bus lines connecting a device to memory can be detected and analyzed to gain information which in turn can compromise the keys. Variations between individual transistors in an integrated circuit, variations in the electrical properties of wires within a chip, variations in the amount of electromagnetic radiation emitted by different wires, etc. can all provide variations that can be analyzed statistically to determine secret keys.
  • the attack collects a small amount of data related to the key each time the token is observed performing a cryptographic operation involving the key.
  • the amount of information known about the key is increased by collecting and statistically correlating (or combining) data from multiple observations of the token as it performs operations involving the key (or related key).
  • such observations may contain signal (i.e., information correlated usefully to the key). However, such observations also contain noise (i.e., information and error that hinder or are irrelevant to determination of the key).
  • the quality of the information gained from these observations is characterized by a “signal to noise” (or S/N) ratio, which is a measure of the magnitude of the signal compared to the amount of noise.
  • the number of operations needed to recover the key depends on the measurement and analysis techniques, but is generally inversely proportional to the square of the S/N ratio.
  • the constant of proportionality also depends upon the amount of confidence required. For example, data of a relatively low confidence level may be acceptable if used to guide an optimized brute force search using statistical information about key bit values. If a countermeasure is used that decreases the signal or increases the amount of measurement noise by, for example, a factor of 300, the statistical attack can generally still succeed but would be expected to require roughly 90,000 times as many observations to extract the same amount of information about the key. An attack requiring 1,000 observations to recover a key before the S/N reduction would now yield the same level of confidence in the recovered key by using on the order of 90 million observations.
  • the signal-to-noise ratio may be improved by an attacker or evaluator using methods such as aligning these measurements so that the data points corresponding to a single point of interest can be compared and analyzed across a large number of observations.
  • Averaging data collected from many operations can be an effective means of amplifying signals and filtering out noise.
  • the evaluator guesses the value of some of the bits of the key, computes an average, checks whether expected biases appear (or uses another statistical technique and check for other expected effects) in collected data, and iterates this process with multiple guesses to recover the entire key. In some cases, disconnecting the power or resetting a device during an operation, may be helpful in compromising secrets by allowing a single leaky operation to be performed repeatedly.
  • leakage refers to the leakage (or the reducing, masking, or minimizing thereof) of any information that is potentially useful to an attacker trying determine secret information.
  • the leaked information includes the secret information itself, but also other information pertaining to that secret information.
  • the attacked device may also leak information, such as information correlated to its internal processing operations, that is not useful to attackers.
  • leakage of non-useful information is not relevant to this description of countermeasures.
  • an attacker can gather data by observing a series of operations, perform statistical analysis on the observations, and use the results to determine the key.
  • an attacker monitors a physical property, such as power consumption, of a secure token as it performs a cryptographic operation.
  • the attacker collects a small amount of data related to the key each time the token is observed performing a cryptographic operation involving the key.
  • the attacker increases the amount of information known about the key by collecting and statistically correlating (or combining) data from multiple observations of the token as it performs operations involving the key (or a related key).
  • such observations may contain signal (i.e., information correlated usefully to the key). However, such observations also contain noise (i.e., information and error that hinder or are irrelevant to determination of the key).
  • the quality of the information gained from these observations is characterized by a “signal to noise” (or S/N) ratio, which is a measure of the magnitude of the signal compared to the amount of noise.
  • the number of operations that the attacker must analyze to recover the key depends on the measurement and analysis techniques, but is generally inversely proportional to the square of the S/N ratio.
  • the constant of proportionality also depends upon the amount of confidence the attacker requires. For example, a relatively low confidence level may be acceptable to an attacker willing to do an optimized brute force search using statistical information about key bit values. Decreasing the signal by a factor of 15 and increasing the amount of measurement noise by a factor of 20 will reduce the signal-to-noise ratio by a factor of 300. This will generally mean that an attacker will require roughly 90,000 times as many observations to extract the same amount of information about the key. An attack requiring 1,000 observations to recover a key before the S/N reduction would now require on the order of 90 million observations to gain the same level of confidence in the recovered key.
  • one approach according to the general technique of using unpredictable information to protect cryptosystems against external monitoring attacks is to implement cryptographic protocols so as to produce unpredictable state information, thereby increasing the number of observations required by an attacker to compromise a key.
  • a system designer can make the so-called work function (effort required) to break a system larger.
  • the system can be made even more secure by making the number of samples required to gain any significant amount of useful key information exceed the maximum number of transactions that can be performed using the key, exceed the number of transactions that can be performed by the device (e.g., before the key expires), or else be so large that monitoring attacks are comparable to (or of greater difficulty than) brute force and other known attacks. For example, consider a system programmed to self-destruct after one million operations—well beyond the expected operational life of most smartcards.
  • noise production system 100 includes randomness source 101 , noise processing module 102 (such as, without limitation, a linear feedback shift register or a hash function-based compression function), activation controller 103 , digital/analog converter 104 , and noise production module ( 105 ).
  • noise processing module 102 such as, without limitation, a linear feedback shift register or a hash function-based compression function
  • activation controller 103 activation controller 103
  • digital/analog converter 104 digital/analog converter
  • noise production module 105
  • Other noise production systems including none, any, or all of the components of FIG. 1 can also be used within the scope of the countermeasures.
  • Randomness source 101 creates the initial noise used to generate unpredictable information.
  • Randomness source 101 can be implemented in hardware or software. It is preferable that the random number generator be implemented in hardware because hardware implementations typically maintain less state information that can be subject to attack. If random numbers are generated via software, care should be taken to ensure that attackers cannot compromise the random number generator state and predict future random number generator outputs. For example, to help make a software random number generator resist external monitoring attacks, an implementation may incorporate extra state information and update its state frequently. Of course, as will be appreciated by those skilled in the art, truly random numbers are not always necessary or available. Therefore, as used herein, any term described as “random” will be understood to include truly random, and also pseudorandom or otherwise unpredictable, information suitable to, and depending on, the nature of the particular application at hand.
  • randomness source 101 is an analog source
  • its output is first converted to digital form, for example using digital/analog converter 104 .
  • the digital output produced by randomness source 101 or digital/analog converter 104 is then provided as an input to noise processing module 102 .
  • Noise processing module 102 converts the initial noise, (which may be biased or have other nonrandom characteristics) into either statistically random noise or noise with desired characteristics (for example, random but with a nonlinear statistical distribution).
  • the activation controller 103 can be configured so that the noise production process is activated during operations in which security is important (such as, without limitation, encryption, decryption, digital signing, data comparison, MAC verification, code verification, audit log updating, EEPROM update, and key changing), but is deactivated during non-security critical operations.
  • security such as, without limitation, encryption, decryption, digital signing, data comparison, MAC verification, code verification, audit log updating, EEPROM update, and key changing
  • a noise production activation control can thus greatly reduce many of the potential disadvantages of such a noise system (such as increased power consumption, reduced performance, increased electromagnetic radiation, decreased reliability, increased heat production, etc.).
  • Activation controller 103 can be implemented in any of a variety of ways, including without limitation in a microprocessor cryptographic accelerator, or other well-known controller device that disables power to one or more elements of noise production system 100 , forces the output of randomness source 101 (or mixer) to a particular value, forces the input or output of digital/analog converter 104 to a particular value, or disables noise production module 105 .
  • noise production module 105 When activation controller 103 enables noise production system 100 , random output from noise processing module 102 is provided to digital/analog (D/A) converter 104 .
  • the D/A output is provided to noise production module 105 , which is configured to sink power, produce electromagnetic radiation, or otherwise introduce noise into attackers' measurements, where the noise produced is a function of the D/A input.
  • the noise production module thus introduces noise into attackers' measurements, increasing the difficulty of external monitoring attacks.
  • Digital/analog conversion methods are known in the background art, and need not be described in detail here. For example, an array of current sources (e.g., transistors) and/or current sinks (e.g., resistors), as well as many other well known techniques can be used.
  • noise production module 105 can operate using the output of randomness source 101 as a direct input. Activation controller 103 can then operate by regulating the output of randomness source 101 or enabling and disabling noise production module 105 .
  • multiple noise production modules may be deployed and driven simultaneously from the same or different random sources.
  • the noise processing module can be used to combine outputs from multiple noise sources and/or provide inputs to multiple noise production modules.
  • microprocessor current usage profiles and other externally measurable characteristics such as E/M radiation
  • noise production modules may include delay lines that temporally isolate their outputs from those of the others, or they may be clocked independently, or they may be free-running.
  • the noise source can be integrated within the cryptosystem microprocessor itself.
  • the noise source and noise control circuitry can be integrated into the same chip that contains the microprocessor, secure memory, I/O interface, etc.
  • the signal-to-noise reduction techniques described herein may be implemented for use in various environments, including without limitation key management and storage systems, cryptographic accelerators (e.g., hardware DES implementations, multipliers, fast modular exponentiators, hash functions, etc.), nonvolatile memory (e.g., EEPROM, flash, etc.), data communication interfaces, buses, and (as will be evident to one of ordinary skill in the art) other computational devices and methods used in cryptographic operations.
  • cryptographic accelerators e.g., hardware DES implementations, multipliers, fast modular exponentiators, hash functions, etc.
  • nonvolatile memory e.g., EEPROM, flash, etc.
  • data communication interfaces e.g., buses, and (as will be evident to one of ordinary skill in the art) other computational devices and methods used in cryptographic operations.
  • clock skipping or clock decorrelation
  • attackers typically compare measurements from several different operations against each other. For example, an attacker might make a sequence of observations by sampling the target device's power consumption at 200 MHz during a 5 ms portion of each of 1,000 cryptographic operations done by the target device. For this exemplary attack, 1,000 observations each containing 1,000,000 data points are thus collected. The attacker would then align these measurements so that the data points corresponding to a single point of interest can be compared and analyzed across a large number of observations.
  • a “1” bit in a secret or private cryptographic key might statistically result in a power feature consisting of a 1 ⁇ A increase above average for 2 ⁇ s followed immediately by a decrease to 2 ⁇ A below average for 1 ⁇ s, while a “0” key bit might result in a power feature consisting of a 1 ⁇ A decrease below average for 2 ⁇ s followed by a 2 ⁇ A increase above average for 1 ⁇ s.
  • Differentiating such signals is easy with sub-microsecond resolution, but can be extremely difficult or impossible with only millisecond resolution unless an extraordinarily large number of samples is taken.
  • small temporal alignment variations may not be able to conceal signal characteristics that are of large amplitude or of long duration (e.g., comparable to or larger than the size of the alignment variations).
  • poor temporal alignment will reduce an attacker's ability to identify fine variations within operations and significantly increase the number of measurements required for a successful attack.
  • clock skipping or clock decorrelation
  • Clock skipping involves decorrelating cryptographic operations from the normal (external) clock cycles by creating a separate, internal clock signal that is used to control processor timing during cryptographic operations. While externally-measurable characteristics (particularly power consumption and electromagnetic radiation) can reveal when some internal clock cycles occur, clock skipping will make them much more difficult for an attacker to accurately locate points of interest in measurements, particularly if noise is introduced into the signal using the techniques disclosed herein. This will be described in more detail below with respect to an exemplary embodiment illustrated in FIG. 2 .
  • random number generator 200 (which can be, but need not be, implemented in hardware) is used to determine which clock cycles (or clock state transitions) are to be used by microprocessor core 225 .
  • Random number generator 200 produces a stream of random (or pseudorandom) digital output bits or analog noise as random output 205 .
  • Clock skipping module 240 then combines (as will be described below) random output 205 with clock signal 220 received from external smartcard interface 210 .
  • clock signal 220 can also originate from another source (for example, if the technique is implemented in environments other than smartcards).
  • random number generator 200 itself uses an external clock signal (e.g., where a random bit is output on each clock state transition)
  • random number generator 200 can, but need not, use clock signal 220 .
  • random output 205 is used to select cycles of clock signal 220 to skip in order to produce clock signal 260 .
  • random output 205 can be used to select the closest corresponding cycles of clock signal 220 to be used as clock signal 260 , or random output 205 can even be used as clock signal 260 itself
  • clock signal 260 be (partially or wholly) decorrelated from external clock signal 220 via random output 205 .
  • clock skipping module 240 can optionally apply a filter to clock signal 260 to ensure desired characteristics. For example, to ensure a minimum clock rate (as opposed to a statistical average), a transition of clock signal 260 may be forced after more than a threshold number of cycles of clock signal 260 have been skipped, either recently or consecutively (e.g., a transition of clock signal 260 can be forced if clock signal 260 has not changed during more than three transitions of clock signal 220 .)
  • clock skipping module 240 can optionally monitor the clock rate (of either clock signal 220 or 260 ) to prevent attackers from stopping the clock and analyzing the device in a halted state or from operating the device too quickly.
  • module 240 can reset microprocessor core 225 , clear memory 290 (which can be nonvolatile RAM, such as battery-backed CMOS, EEPROM, flash memory, a hard disk, or other such storage used to store the key and/or other information), clear the state of cryptographic accelerator 280 , and log the fault in memory 290 .
  • clear memory 290 which can be nonvolatile RAM, such as battery-backed CMOS, EEPROM, flash memory, a hard disk, or other such storage used to store the key and/or other information
  • clear memory 290 which can be nonvolatile RAM, such as battery-backed CMOS, EEPROM, flash memory, a hard disk, or other such storage used to store the key and/or other information
  • clear memory 290 which can be nonvolatile RAM,
  • clock skipping module 240 and microprocessor 225 are combined, such that random output 205 can force microprocessor 225 to skip clock cycles. For example, when microprocessor 225 is directed to skip a clock cycle (such as when three output bits equal to zero are received in random output 205 ), the result of the current or next instruction (or clock cycle) executed by the microprocessor is discarded and repeated.
  • the fraction of skipped clock cycles does not need to be very large; for example and without limitation, even skipping as few as one clock cycle in 20 (on average) will introduce significant measurement drift.
  • clock skipping may sometimes adversely affect operations requiring regular clock cycles.
  • one bit is sent or received on a serial I/O (input/output) line every 372 cycles of the external clock.
  • a 3.579545 MHz external clock is compatible with a serial communication rate of 9600 bits per second.
  • microprocessor 225 will operate at a different clock rate governed by signal 260 .
  • a mismatch between the data communications clock rate and the microprocessor clock rate may result, causing I/O errors to occur. Consequently, in devices implementing clock skipping, it is often advantageous for the microprocessor to be controlled by external clock 220 during I/O operations.
  • clock skipping activation signal 230 is used to select between external clock signal 220 and the (modified) internal clock that would otherwise be produced by clock skipping module 140 .
  • clock skipping activation signal 220 can be produced by a microprocessor or any other control device that is capable of knowing when to apply (or not apply) the clock skipping. Selection of whether or not to clock skip at any particular time can be performed by many well-known techniques that need not be described in detail here.
  • microprocessor 225 is well suited for such a task because it is necessarily aware of I/O operations associated with the receipt of data signals 270 .
  • microprocessor core 225 can assert control signal 230 to cause clock skipping module 240 to ignore random output 205 and provide external clock signal 220 directly as clock signal 260 .
  • Control signal 230 and the noise production activation control signal described previously can, but need not be the same signal.
  • a conventional external clock signal is used for I/O and other processing, where clock skipping is not needed to protect secret information.
  • an internal clock signal preferably but not necessarily generated in the device (for example, produced using a ring oscillator, which is well known in the background art), is used for internal (e.g., cryptographic) processing.
  • internal operations need not proceed at a speed related to or derived from the external clock rate.
  • the internal clock may be distorted or skipped, for example, as described above.
  • Clock signal selection can be performed by microprocessor 225 as mentioned previously.
  • Another technique which is especially suitable for, but not limited to smartcards, uses a UART (universal asynchronous receiver/transmitter) or other buffer between the internally clocked region and the external I/O interface to ensure that communications over the external serial I/O interface are clocked at a rate corresponding to the externally-supplied clock but may be accessed reliably by internally-clocked circuits.
  • UART universal asynchronous receiver/transmitter
  • the internally-generated clock signal can be derived from the external clock signal. This can be performed via an analog phase-locked loop, which is well known in the background art and need not be described in detail here. Such an internal clock will be rather closely correlated with the external clock, and therefore not as effective against attacks as the randomized clock signal or two separate clocks described previously. (Of course, its effectiveness can be improved by optionally using clock skipping or analog (or other) noise sources to adjust the frequency, drift, and jitter of its signal.) Also, when synthesizing the internal clock from the external clock, the clock-derivation circuitry can be configured to restrict the rate of the internal clock frequency, for example, to enforce a minimum internal clock frequency so that attackers cannot stop the clock and attack the device in the stopped state. The derived internal clock signal exhibits a number of useful properties that will be described in the following paragraph.
  • microprocessor 225 can continue operating so that it can detect and even respond to attacks that involve halting, removing, or altering the external clock while power is connected.
  • the use of an internally-generated clock additionally provides protection against attacks involving the introduction of errors into computations. For example, jitter or high frequencies supplied to the external clock would cause harmless communication errors, but would be prevented from causing erroneous computations.
  • phase locked loop can produce an internal clock signal that is a multiple of the external clock signal, it is possible to clock cryptographic operations at a rate faster than the external clock, providing improved performance.
  • this is an added benefit.
  • the internal clock can also be used to monitor the external clock to detect abnormalities introduced by attackers.
  • the microprocessor can have the option of selecting between multiple clock modes. For example, a slower mode might be used if it has a lower probability of computation error, a faster mode might be used when improved performance is needed, and clock skipping or other clock distortion might be activated when resistance to external monitoring attacks is desired.
  • clock decorrelation e.g., second clocks or phase-locked loops
  • clock decorrelation can also be effected by software as will be described below. This is useful, for example, where the environment does not allow for hardware-based clock skipping.
  • hardware clock decorrelation could be supplemented with software-based clock decorrelation for even greater protection in security critical code segments.
  • One efficient software-implementable technique for clock decorrelation takes advantage of the fact that the amount of time used for a process with a conditional jump can vary depending on whether or not the jump is actually performed. In such cases, inserting branch delays can be used as a form of clock decorrelation.
  • the assembly language clock randomizer below uses a random generator to introduce clock variations that can help prevent accurate alignment by an attacker:
  • the random information may be used to select between parallel code processes, such that the same cryptographic result will be produced regardless of which code process is selected but where the parallel processes perform different operations toward producing the result.
  • Another approach to the general technique of using unpredictable information to protect cryptosystems against external monitoring attacks involves the introduction of entropy into the order of processing operations or into the execution path while maintaining desired functionality (such as compatibility with standard cryptographic algorithm definitions). More specifically, a device can use a random number generator to cause unpredictability in the order of performing a sequence of suboperations. If attackers cannot accurately determine the order in which operations were performed, cross-correlation between samples becomes more difficult or impossible. Consequently the data collected by an attacker effectively has a significantly lower signal-to-noise ratio.
  • This example is input-ordered, meaning that processing steps are performed in the order (or inverse order) in which the input bits are supplied.
  • input bit 0 is permuted first
  • input bit 63 is permuted last.
  • output-ordered permutations are also commonly used in the background art.
  • table1 is a permutation (i.e., where one element equals each of the values 0 . . .
  • both output-ordered and input-ordered permutations can leak information about the data they process.
  • attackers' measurements of loop iteration i will be correlated to dataIn[i].
  • the attackers' measurements of loop iteration i will be correlated to dataOut[i].
  • An improved permutation method would thus be advantageous.
  • One exemplary implementation of such a method is shown in the table below. This high-entropy permutation combines several previously-described aspects of the countermeasures, including without limitation order randomization (thus being neither input-ordered nor output-ordered) and blinding techniques (to conceal further the data being permuted).
  • the magnitude of signals leaked due to variations in data values is usually smaller (often by a factor of several orders of magnitude) than signals leaked due to branches and variations in the execution path. Therefore, the high-entropy permutation operation, above, uses a constant execution path to inhibit leakage via variations in the execution path.
  • the exemplary blinded randomized-order permutation operation includes four steps, which can be performed separately or simultaneously: initialization, blinding, permutation, and unblinding. Implementations using partial blinding, which operate on already-blinded values, or those with reduced security requirements will not require all steps.
  • Initialization of the blinded randomized-order permutation operation involves constructing and randomizing a permutation table (“perm”) for determining the bit order for operations.
  • (Bit order permutation table “perm” randomizes the time at which any particular data bit is manipulated.)
  • the bit order table is created in two passes, where the first assures that the table has the correct form (i.e., contains the numbers zero through 63), and the second introduces random order into the table. Because the process of constructing the bit order table does not involve any secret inputs, the only security requirement for the process is that the final result be unknown to attackers.
  • the first permutation table initialization loop can also place random values into dataOut and temp to help whiten any leaked signals when data values are first stored in these arrays. Finally, although it is not required, more than 64 iterations of the randomization loop are used to ensure that any statistical biases remaining after the randomization loop are insignificantly small.
  • the next section of the code performs the blinding operation.
  • a random number generator produces a random blinding bit.
  • the temporary buffer (temp) is initialized with the XOR of the random bit and an input data bit, where the input data bit is selected according to the table (perm) constructed previously.
  • the output buffer (dataOut) is initialized with the blinding bit, where the dataOut bit is the result of using the input permutation table to operate on the index to temp.
  • the second part of the blinding process re-randomizes the bit order permutation table (perm).
  • Input bits are loaded in the order specified by the table (perm), permuted according to the (non-secret) externally-specified permutation table (table), and XORed onto the destination table (dataOut).
  • the leak-minimized permutation operation described dramatically reduces the amount of information leaked from a permutation operation, but is not necessarily expected to reduce such leakage to zero.
  • the input data to the function arrives in fixed order and unblinded form, and the output is similarly supplied unblinded in fixed order. Consequently, two or more measurements from the same transaction might (for example) be correlated to each other such that the strength or sign of the correlation is a function of one or more input or output data bits.
  • operands can be maintained in a blinded state during processing, to be (partially or completely) reconstituted only when nonlinear operations must be performed or at the end of the computation.
  • the number of bits manipulated does not need to equal 64, the order of steps may be changed, steps can be removed for simplified implementations (such as those that are not subject to some attacks), steps can be modified, different permutation generation and update processes can be used, and additional steps can be added.
  • Cryptographic operations should normally be checked to ensure that incorrect computations do not compromise keys or enable other attacks.
  • Cryptographic implementations of the countermeasures can be, and in a preferred embodiment are, combined with error-detection and/or error-correction logic to ensure that cryptographic operations are performed correctly.
  • a simple and effective technique is to perform cryptographic operations twice, ideally using two independent hardware processors and/or software implementations, with a comparison operation performed at the end to verify that both produce identical results. If the results produced by the two units do not match, the failed comparison will prevent the defective processing result from being used.
  • the compare operation ever fails (or fails too many times) the device may self-destruct (such as by deleting internal keys) or disable itself.
  • a device might erase its key storage memory if either two defective DES operations occur sequentially or five defective DES results occur during the lifetime of the device.
  • full redundancy is not necessary.
  • methods are known in the background art for self-checking functions that can be incorporated into the cryptosystem implementation (e.g., RSA signatures can be verified after digital signing operations).
  • Detection of conditions likely to cause incorrect results may also be used.
  • active or passive sensors to detect unusually high or low voltages, high-frequency noise on voltage or signal inputs, exposure to electromagnetic fields and radiation, and physical tampering may be employed. Inappropriate operating conditions can (for example) trigger the device to reset, delete secrets, or self-destruct.
  • Self-diagnostic functions such as a POST (power-on-self-test) should also be incorporated to verify that cryptographic functions have not been damaged.
  • the self-test can be deferred until after completion of the first transaction or until a sufficient idle period is encountered. For example, a flag indicating successful POST completion can be cleared upon initialization. While the card is waiting for a command from the host system, it can attempt the POST. Any I/O received during the POST will cause an interrupt, which will cancel the POST (leaving the POST-completed flag at zero). If any cryptographic function is called, the device will check the POST flag and (if it is not set) perform the POST before doing any cryptographic operations.
  • the countermeasures are extremely useful for improving security, particularly in environments and applications with difficult engineering requirements, by enabling the construction of devices that are significantly more resistant to attack than devices of similar cost and complexity that do not use such countermeasures.
  • multiple security techniques may be required to make a system secure. For example, leak minimization and obfuscation may be used in conjunction with other security methods or countermeasures.
  • cryptographic smartcards of all kinds including without limitation smartcards substantially compliant with ISO 7816-1, ISO 7816-2, and ISO 7816-3 (“ISO 7816-compliant smartcards”); contactless and proximity-based smartcards and cryptographic tokens; stored value cards and systems; cryptographically secured credit and debit cards; customer loyalty cards and systems; cryptographically authenticated credit cards; cryptographic accelerators; gambling and wagering systems; secure cryptographic chips; tamper-resistant microprocessors; software programs (including without limitation programs for use on personal computers, servers, etc.
  • cryptographic devices key management devices; banking key management systems; secure web servers; electronic payment systems; micropayment systems and meters; prepaid telephone cards; cryptographic identification cards and other identity verification systems; systems for electronic funds transfer; automatic teller machines; point of sale terminals; certificate issuance systems; electronic badges; door entry systems; physical locks of all kinds using cryptographic keys; systems for decrypting television signals (including without limitation, broadcast television, satellite television, and cable television); systems for decrypting enciphered music and other audio content (including music distributed over computer networks); systems for protecting video signals of all kinds; intellectual property protection and copy protection systems (such as those used to prevent unauthorized copying or use of movies, audio content, computer programs, video games, images, text, databases, etc.); cellular telephone scrambling and authentication systems (including telephone authentication smartcards); secure telephones (including key storage devices for such telephones); cryptographic PCMCIA cards; portable cryptographic tokens; and cryptographic data auditing systems.
  • key management devices banking key management systems; secure web servers; electronic payment systems; micropayment systems and meters; prepaid telephone

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Security & Cryptography (AREA)
  • Computer Hardware Design (AREA)
  • General Engineering & Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Accounting & Taxation (AREA)
  • Microelectronics & Electronic Packaging (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • General Business, Economics & Management (AREA)
  • Strategic Management (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Signal Processing (AREA)
  • Finance (AREA)
  • Storage Device Security (AREA)

Abstract

Information leaked from smart cards and other tamper resistant cryptographic devices can be statistically analyzed to determine keys or other secret data. A data collection and analysis system is configured with an analog-to-digital converter connected to measure the device's consumption of electrical power, or some other property of the target device, that varies during the device's processing. As the target device performs cryptographic operations, data from the A/D converter are recorded for each cryptographic operation. The stored data are then processed using statistical analysis, yielding the entire key, or partial information about the key that can be used to accelerate a brute force search or other attack.

Description

    RELATED APPLICATIONS
  • This patent application is a continuation of U.S. non-provisional patent application Ser. No. 11/643,349, filed Dec. 21, 2006; said U.S. patent application Ser. No. 11/643,349 is a continuation of U.S. non-provisional patent application Ser. No. 10/005,105, filed Dec. 3, 2001; said U.S. patent application Ser. No. 10/005,105 is a continuation-in-part of U.S. non-provisional patent application Ser. No. 09/326,222, filed Jun. 3, 1999 (which claims the benefit of U.S. provisional patent application Ser. No. 60/087,880, filed Jun. 3, 1998); said U.S. patent application Ser. No. 10/005,105 is also a continuation-in-part of U.S. non-provisional patent application Ser. No. 09/930,836, filed Aug. 15, 2001 (which is a continuation of U.S. non-provisional patent application Ser. No. 09/324,798, filed Jun. 3, 1999, now U.S. Pat. No. 6,278,783, which claims the benefit of U.S. provisional patent application 60/087,826, filed Jun. 3, 1998); said U.S. patent application Ser. No. 10/005,105 is also a continuation-in-part of U.S. non-provisional patent application Ser. No. 09/737,182, filed Dec. 13, 2000 (which is a divisional of U.S. non-provisional patent application Ser. No. 09/224,682, filed Dec. 31, 1998, now U.S. Pat. No. 6,304,658, which claims the benefit of U.S. provisional patent applications 60/089,529, filed Jun. 15, 1998, and 60/070,344, filed Jan. 2, 1998). All of the prior patent applications mentioned in this paragraph are hereby incorporated by reference in their entireties into the present patent application.
  • FIELD OF THE INVENTION
  • This application relates generally to cryptographic systems and, more specifically, to determining useful information about a cryptographic system by external monitoring of its cryptographic operations.
  • BACKGROUND
  • As described in U.S. Pat. No. 4,908,038 to Matsumura et al., cryptographic devices can be attacked using information gathered by observing the timing of comparison operations performed by such devices during their operation. For example, if a MAC (Message Authentication Code) algorithm is strong and the key is secure, forging a MAC should require O(2̂n) attempts (where n is the MAC length in bits), but a device using a vulnerable MAC validation process is vulnerable to an O(n) timing attack.
  • If timing is the only source of leaked information, securing the device is often relatively straightforward. Previously known countermeasures to attacks involving information leaking from cryptosystems employ large and often expensive physical shielding and/or careful filtering of inputs and outputs (e.g., U.S. government Tempest specifications). Unfortunately, these techniques are difficult to apply in constrained engineering environments. For example, physical constraints (such as size and weight), cost, and the need to conserve power can often prevent the use of such techniques. It is also known to use certain computational techniques (e.g., see Matsumura, above, or P. Kocher, “Timing Attacks on Implementations of Diffie-Hellman, RSA, DSS, and Other Systems,” Advances in Cryptology—CRYPTO '96, Springer-Verlag, 1996, pages 104-113) to equalize timing. However, sources of information leakage other than timing (e.g., a device's power consumption) provide other avenues of attack. Indeed, Matsumara's timing equalization system itself can be vulnerable to non-timing attacks, for example by analyzing power consumption to detect the start of processing delays. It would therefore be advantageous to protect the devices' internal operations themselves instead of (or in addition to) simply externally masking the devices' timing (or other) fluctuations.
  • SUMMARY
  • Information leaked from smart cards and other tamper resistant cryptographic devices can be statistically analyzed to determine keys or other secret data. A data collection and analysis system is configured with an analog-to-digital converter connected to measure the device's consumption of electrical power, or some other property of the target device, that varies during the device's processing. As the target device performs cryptographic operations, data from the A/D converter are recorded for each cryptographic operation. The stored data are then processed using statistical analysis, yielding the entire key, or partial information about the key that can be used to accelerate a brute force search or other attack.
  • Particularly serious sources of leakage include the device's power consumption and electromagnetic radiation. Observation of the microprocessor's power consumption can reveal whether the jumps are taken. Observation of the power consumption and/or timing can reveal whether the carried bits in the key rotates of each DES round equal zero or one. Operations that change the device state can reveal information about the initial and final states of the operations. Signals radiated from the address and data bus lines connecting a device to memory can be detected and analyzed to gain information which in turn can compromise the keys. Variations between individual transistors in an integrated circuit, variations in the electrical properties of wires within a chip, variations in the amount of electromagnetic radiation emitted by different wires, etc. can all provide variations that can be analyzed statistically to determine secret keys.
  • In an exemplary embodiment, the attack collects a small amount of data related to the key each time the token is observed performing a cryptographic operation involving the key. The amount of information known about the key is increased by collecting and statistically correlating (or combining) data from multiple observations of the token as it performs operations involving the key (or related key).
  • In the case of a cryptosystem which is leaking information, such observations may contain signal (i.e., information correlated usefully to the key). However, such observations also contain noise (i.e., information and error that hinder or are irrelevant to determination of the key). The quality of the information gained from these observations is characterized by a “signal to noise” (or S/N) ratio, which is a measure of the magnitude of the signal compared to the amount of noise.
  • The number of operations needed to recover the key depends on the measurement and analysis techniques, but is generally inversely proportional to the square of the S/N ratio. The constant of proportionality also depends upon the amount of confidence required. For example, data of a relatively low confidence level may be acceptable if used to guide an optimized brute force search using statistical information about key bit values. If a countermeasure is used that decreases the signal or increases the amount of measurement noise by, for example, a factor of 300, the statistical attack can generally still succeed but would be expected to require roughly 90,000 times as many observations to extract the same amount of information about the key. An attack requiring 1,000 observations to recover a key before the S/N reduction would now yield the same level of confidence in the recovered key by using on the order of 90 million observations.
  • After making a large number of measurements, the signal-to-noise ratio may be improved by an attacker or evaluator using methods such as aligning these measurements so that the data points corresponding to a single point of interest can be compared and analyzed across a large number of observations. Averaging data collected from many operations can be an effective means of amplifying signals and filtering out noise.
  • In one embodiment, the evaluator guesses the value of some of the bits of the key, computes an average, checks whether expected biases appear (or uses another statistical technique and check for other expected effects) in collected data, and iterates this process with multiple guesses to recover the entire key. In some cases, disconnecting the power or resetting a device during an operation, may be helpful in compromising secrets by allowing a single leaky operation to be performed repeatedly.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 illustrates an exemplary apparatus for introducing noise into a cryptosystem.
  • FIG. 2 illustrates an exemplary apparatus for implementing clock skipping.
  • DETAILED DESCRIPTION
  • This application discloses both external monitoring techniques (e.g., attacks against cryptosystems), as well as techniques for using unpredictable information to protect cryptosystems against such external monitoring techniques.
  • External Monitoring Techniques
  • The external monitoring techniques operate on the principle that information leaked from smart cards and other tamper resistant cryptographic devices can be statistically analyzed to determine keys or other secret data. A data collection and analysis system is configured with an analog-to-digital converter connected to measure the device's consumption of electrical power, or some other property of the target device, that varies during the device's processing. As the target device performs cryptographic operations, data from the A/D converter are recorded for each cryptographic operation. The stored data are then processed using statistical analysis, yielding the entire key, or partial information about the key that can be used to accelerate a brute force search or other attack.
  • Particularly serious sources of leakage include the device's power consumption and electromagnetic radiation. Observation of the microprocessor's power consumption can reveal whether the jumps are taken. Observation of the power consumption and/or timing can reveal whether the carried bits in the key rotates of each DES round equal zero or one. Operations that change the device state can reveal information about the initial and final states of the operations. Signals radiated from the address and data bus lines connecting a device to memory can be detected and analyzed to gain information which in turn can compromise the keys. Variations between individual transistors in an integrated circuit, variations in the electrical properties of wires within a chip, variations in the amount of electromagnetic radiation emitted by different wires, etc. can all provide variations that can be analyzed statistically to determine secret keys.
  • In an exemplary embodiment, the attack collects a small amount of data related to the key each time the token is observed performing a cryptographic operation involving the key. The amount of information known about the key is increased by collecting and statistically correlating (or combining) data from multiple observations of the token as it performs operations involving the key (or related key).
  • In the case of a cryptosystem which is leaking information, such observations may contain signal (i.e., information correlated usefully to the key). However, such observations also contain noise (i.e., information and error that hinder or are irrelevant to determination of the key). The quality of the information gained from these observations is characterized by a “signal to noise” (or S/N) ratio, which is a measure of the magnitude of the signal compared to the amount of noise.
  • The number of operations needed to recover the key depends on the measurement and analysis techniques, but is generally inversely proportional to the square of the S/N ratio. The constant of proportionality also depends upon the amount of confidence required. For example, data of a relatively low confidence level may be acceptable if used to guide an optimized brute force search using statistical information about key bit values. If a countermeasure is used that decreases the signal or increases the amount of measurement noise by, for example, a factor of 300, the statistical attack can generally still succeed but would be expected to require roughly 90,000 times as many observations to extract the same amount of information about the key. An attack requiring 1,000 observations to recover a key before the S/N reduction would now yield the same level of confidence in the recovered key by using on the order of 90 million observations.
  • After making a large number of measurements, the signal-to-noise ratio may be improved by an attacker or evaluator using methods such as aligning these measurements so that the data points corresponding to a single point of interest can be compared and analyzed across a large number of observations. Averaging data collected from many operations can be an effective means of amplifying signals and filtering out noise.
  • In one embodiment, the evaluator guesses the value of some of the bits of the key, computes an average, checks whether expected biases appear (or uses another statistical technique and check for other expected effects) in collected data, and iterates this process with multiple guesses to recover the entire key. In some cases, disconnecting the power or resetting a device during an operation, may be helpful in compromising secrets by allowing a single leaky operation to be performed repeatedly.
  • The foregoing external monitoring techniques are described in more detail in the following U.S. patents and patent applications, all of which are hereby incorporated by reference herein in their entirety: U.S. non-provisional patent application Ser. No. 09/326,222, filed on Jun. 3, 1999 (which claims the benefit of U.S. provisional patent application No. 60/087,880, filed on Jun. 3, 1998); U.S. non-provisional patent application Ser. No. 09/930,836, filed on Aug. 15, 2001 (which is a continuation of U.S. non-provisional patent application Ser. No. 09/324,798, filed on Jun. 3, 1999, now U.S. Pat. No. 6,278,783, which claims the benefit of U.S. provisional patent application No. 60/087,826, filed on Jun. 3, 1998); and U.S. non-provisional patent application Ser. No. 09/737,182, filed on Dec. 13, 2000 (which is a divisional of U.S. non-provisional patent application Ser. No. 09/224,682, filed on Dec. 31, 1998, now U.S. Pat. No. 6,304,658, which claims the benefit of U.S. provisional patent applications No. 60/089,529, filed on Jun. 15, 1998, and 60/070,344, filed on Jan. 2, 1998). This application is formally a continuation-in-part application of each of the foregoing patents and patent applications; however, no new matter has been added over the combination thereof.
  • The techniques for protecting cryptosystems (i.e., countermeasures) are described in various embodiments as set forth below. Although these embodiments differ in the details of their implementations, those skilled in the art will appreciate the fundamental commonality in their essential operation—using randomness or other sources of unpredictability to decorrelate secret information from externally monitorable signals in such a way that deters external monitoring attacks (including those involving statistical accumulation and analysis of collected data) upon cryptographic systems.
  • Reduction of Signal-to-Noise Ratios
  • Unless noted otherwise, it shall be assumed herein that leakage (or the reducing, masking, or minimizing thereof) refers to the leakage (or the reducing, masking, or minimizing thereof) of any information that is potentially useful to an attacker trying determine secret information. Thus, the leaked information includes the secret information itself, but also other information pertaining to that secret information. Of course, the attacked device may also leak information, such as information correlated to its internal processing operations, that is not useful to attackers. However, such leakage of non-useful information is not relevant to this description of countermeasures.
  • To obtain a secret key from a cryptosystem that leaks information, an attacker can gather data by observing a series of operations, perform statistical analysis on the observations, and use the results to determine the key. In a common situation, an attacker monitors a physical property, such as power consumption, of a secure token as it performs a cryptographic operation. The attacker collects a small amount of data related to the key each time the token is observed performing a cryptographic operation involving the key. The attacker increases the amount of information known about the key by collecting and statistically correlating (or combining) data from multiple observations of the token as it performs operations involving the key (or a related key).
  • In the case of a cryptosystem which is leaking information, such observations may contain signal (i.e., information correlated usefully to the key). However, such observations also contain noise (i.e., information and error that hinder or are irrelevant to determination of the key). The quality of the information gained from these observations is characterized by a “signal to noise” (or S/N) ratio, which is a measure of the magnitude of the signal compared to the amount of noise.
  • The number of operations that the attacker must analyze to recover the key depends on the measurement and analysis techniques, but is generally inversely proportional to the square of the S/N ratio. The constant of proportionality also depends upon the amount of confidence the attacker requires. For example, a relatively low confidence level may be acceptable to an attacker willing to do an optimized brute force search using statistical information about key bit values. Decreasing the signal by a factor of 15 and increasing the amount of measurement noise by a factor of 20 will reduce the signal-to-noise ratio by a factor of 300. This will generally mean that an attacker will require roughly 90,000 times as many observations to extract the same amount of information about the key. An attack requiring 1,000 observations to recover a key before the S/N reduction would now require on the order of 90 million observations to gain the same level of confidence in the recovered key.
  • Thus, one approach according to the general technique of using unpredictable information to protect cryptosystems against external monitoring attacks is to implement cryptographic protocols so as to produce unpredictable state information, thereby increasing the number of observations required by an attacker to compromise a key. By reducing the available signal size and/or increasing the amount of error, noise, and uncertainty in attackers' measurements, a system designer can make the so-called work function (effort required) to break a system larger.
  • The system can be made even more secure by making the number of samples required to gain any significant amount of useful key information exceed the maximum number of transactions that can be performed using the key, exceed the number of transactions that can be performed by the device (e.g., before the key expires), or else be so large that monitoring attacks are comparable to (or of greater difficulty than) brute force and other known attacks. For example, consider a system programmed to self-destruct after one million operations—well beyond the expected operational life of most smartcards. If a design not using the countermeasures requires five operations to break, and the countermeasures reduce the signal-to-noise ratio by a factor of 1000, the number of operations required to break the system (i.e., isolate the signal or key from the noise) might increase by a factor of roughly one million (i.e., to approximately 5 million) exceeding the lifetime of the secret or the device. Thus, attackers will be unable to collect enough measurements to compromise the secret.
  • Random Noise Generation
  • An exemplary apparatus for introducing noise into a cryptosystem is illustrated in FIG. 1. In FIG. 1, noise production system 100 includes randomness source 101, noise processing module 102 (such as, without limitation, a linear feedback shift register or a hash function-based compression function), activation controller 103, digital/analog converter 104, and noise production module (105). Other noise production systems including none, any, or all of the components of FIG. 1 can also be used within the scope of the countermeasures.
  • Randomness source 101 creates the initial noise used to generate unpredictable information. Randomness source 101 can be implemented in hardware or software. It is preferable that the random number generator be implemented in hardware because hardware implementations typically maintain less state information that can be subject to attack. If random numbers are generated via software, care should be taken to ensure that attackers cannot compromise the random number generator state and predict future random number generator outputs. For example, to help make a software random number generator resist external monitoring attacks, an implementation may incorporate extra state information and update its state frequently. Of course, as will be appreciated by those skilled in the art, truly random numbers are not always necessary or available. Therefore, as used herein, any term described as “random” will be understood to include truly random, and also pseudorandom or otherwise unpredictable, information suitable to, and depending on, the nature of the particular application at hand.
  • Where randomness source 101 is an analog source, its output is first converted to digital form, for example using digital/analog converter 104. The digital output produced by randomness source 101 or digital/analog converter 104 is then provided as an input to noise processing module 102. Noise processing module 102 converts the initial noise, (which may be biased or have other nonrandom characteristics) into either statistically random noise or noise with desired characteristics (for example, random but with a nonlinear statistical distribution).
  • Many cryptosystems spend a relatively small fraction of total processing time performing security-critical operations. Therefore, the activation controller 103 can be configured so that the noise production process is activated during operations in which security is important (such as, without limitation, encryption, decryption, digital signing, data comparison, MAC verification, code verification, audit log updating, EEPROM update, and key changing), but is deactivated during non-security critical operations. A noise production activation control can thus greatly reduce many of the potential disadvantages of such a noise system (such as increased power consumption, reduced performance, increased electromagnetic radiation, decreased reliability, increased heat production, etc.). Activation controller 103 can be implemented in any of a variety of ways, including without limitation in a microprocessor cryptographic accelerator, or other well-known controller device that disables power to one or more elements of noise production system 100, forces the output of randomness source 101 (or mixer) to a particular value, forces the input or output of digital/analog converter 104 to a particular value, or disables noise production module 105.
  • When activation controller 103 enables noise production system 100, random output from noise processing module 102 is provided to digital/analog (D/A) converter 104. The D/A output is provided to noise production module 105, which is configured to sink power, produce electromagnetic radiation, or otherwise introduce noise into attackers' measurements, where the noise produced is a function of the D/A input. The noise production module thus introduces noise into attackers' measurements, increasing the difficulty of external monitoring attacks. Digital/analog conversion methods are known in the background art, and need not be described in detail here. For example, an array of current sources (e.g., transistors) and/or current sinks (e.g., resistors), as well as many other well known techniques can be used.
  • In an embodiment where randomness source 101 is an analog noise source, noise production module 105 can operate using the output of randomness source 101 as a direct input. Activation controller 103 can then operate by regulating the output of randomness source 101 or enabling and disabling noise production module 105.
  • To prevent noise from being observably correlated to clock transitions or other externally-measurable events, multiple noise production modules may be deployed and driven simultaneously from the same or different random sources. Alternatively, the noise processing module can be used to combine outputs from multiple noise sources and/or provide inputs to multiple noise production modules. Also, because microprocessor current usage profiles (and other externally measurable characteristics such as E/M radiation) are instruction-dependent and carry significant detail within each clock period, it may be advantageous to drive noise production modules faster than (or independently from) the clock rate applied to cryptosystem microprocessor. For example, noise production modules may include delay lines that temporally isolate their outputs from those of the others, or they may be clocked independently, or they may be free-running.
  • All of the foregoing components may be implemented separately or in various combinations, using analog or digital techniques as appropriate. Those skilled in the art will also appreciate that various of the components can be implemented in hardware, or even software, although hardware implementations will generally provide greater security. For example, the noise source can be integrated within the cryptosystem microprocessor itself. In single-chip environments (such as smartcards and secure microprocessors), the noise source and noise control circuitry can be integrated into the same chip that contains the microprocessor, secure memory, I/O interface, etc.
  • The signal-to-noise reduction techniques described herein may be implemented for use in various environments, including without limitation key management and storage systems, cryptographic accelerators (e.g., hardware DES implementations, multipliers, fast modular exponentiators, hash functions, etc.), nonvolatile memory (e.g., EEPROM, flash, etc.), data communication interfaces, buses, and (as will be evident to one of ordinary skill in the art) other computational devices and methods used in cryptographic operations.
  • Clock Skipping
  • Another approach to the general technique of using unpredictable information to protect cryptosystems against external monitoring attacks involves what will be referred to herein as clock skipping (or clock decorrelation).
  • During statistical attacks using power consumption or electromagnetic radiation, attackers typically compare measurements from several different operations against each other. For example, an attacker might make a sequence of observations by sampling the target device's power consumption at 200 MHz during a 5 ms portion of each of 1,000 cryptographic operations done by the target device. For this exemplary attack, 1,000 observations each containing 1,000,000 data points are thus collected. The attacker would then align these measurements so that the data points corresponding to a single point of interest can be compared and analyzed across a large number of observations.
  • Therefore, security can be improved by preventing attackers from locating points of interest within collected data sets and from identifying corresponding regions between observations. Indeed, causing an attacker to include incorrectly-aligned data is one way to decrease the effective signal-to-noise ratio of the attacker's data (see previous section), since the noise increases significantly (due to the inclusion of uncorrelated samples) and the useful signal decreases (due to the presence of fewer good samples).
  • Without accurate temporal alignment, the temporal resolution of the attacker's observations decreases greatly, making it much more difficult for the attacker to identify a signal containing fine structure. For example, a “1” bit in a secret or private cryptographic key might statistically result in a power feature consisting of a 1 μA increase above average for 2 μs followed immediately by a decrease to 2 μA below average for 1 μs, while a “0” key bit might result in a power feature consisting of a 1 μA decrease below average for 2 μs followed by a 2 μA increase above average for 1 μs. Differentiating such signals is easy with sub-microsecond resolution, but can be extremely difficult or impossible with only millisecond resolution unless an extraordinarily large number of samples is taken. Of course, small temporal alignment variations may not be able to conceal signal characteristics that are of large amplitude or of long duration (e.g., comparable to or larger than the size of the alignment variations). In general, then, poor temporal alignment will reduce an attacker's ability to identify fine variations within operations and significantly increase the number of measurements required for a successful attack.
  • Many conventional systems, including commonly available smartcards, simply use external clocks for their cryptographic operations—even though attackers can freely observe and manipulate the external clock. This greatly facilitates the ability of attackers to make the measurements necessary to attack the system. One embodiment of the countermeasures uses clock skipping (or clock decorrelation) to inhibit such attacks by reducing attackers' ability to predict the system state. Clock skipping involves decorrelating cryptographic operations from the normal (external) clock cycles by creating a separate, internal clock signal that is used to control processor timing during cryptographic operations. While externally-measurable characteristics (particularly power consumption and electromagnetic radiation) can reveal when some internal clock cycles occur, clock skipping will make them much more difficult for an attacker to accurately locate points of interest in measurements, particularly if noise is introduced into the signal using the techniques disclosed herein. This will be described in more detail below with respect to an exemplary embodiment illustrated in FIG. 2.
  • Referring now to FIG. 2, random number generator 200 (which can be, but need not be, implemented in hardware) is used to determine which clock cycles (or clock state transitions) are to be used by microprocessor core 225. Random number generator 200 produces a stream of random (or pseudorandom) digital output bits or analog noise as random output 205. Clock skipping module 240 then combines (as will be described below) random output 205 with clock signal 220 received from external smartcard interface 210. Of course, clock signal 220 can also originate from another source (for example, if the technique is implemented in environments other than smartcards). In embodiments where random number generator 200 itself uses an external clock signal (e.g., where a random bit is output on each clock state transition), random number generator 200 can, but need not, use clock signal 220.
  • Within clock skipping module 240, random output 205 is used to select cycles of clock signal 220 to skip in order to produce clock signal 260. Alternatively, random output 205 can be used to select the closest corresponding cycles of clock signal 220 to be used as clock signal 260, or random output 205 can even be used as clock signal 260 itself Still other approaches are possible, as will be appreciated by those skilled in the art; the basic point being that clock signal 260 be (partially or wholly) decorrelated from external clock signal 220 via random output 205.
  • If desired, clock skipping module 240 can optionally apply a filter to clock signal 260 to ensure desired characteristics. For example, to ensure a minimum clock rate (as opposed to a statistical average), a transition of clock signal 260 may be forced after more than a threshold number of cycles of clock signal 260 have been skipped, either recently or consecutively (e.g., a transition of clock signal 260 can be forced if clock signal 260 has not changed during more than three transitions of clock signal 220.)
  • Additionally, clock skipping module 240 can optionally monitor the clock rate (of either clock signal 220 or 260) to prevent attackers from stopping the clock and analyzing the device in a halted state or from operating the device too quickly. When module 240 detects such a clock fault, it can reset microprocessor core 225, clear memory 290 (which can be nonvolatile RAM, such as battery-backed CMOS, EEPROM, flash memory, a hard disk, or other such storage used to store the key and/or other information), clear the state of cryptographic accelerator 280, and log the fault in memory 290. Methods and apparatuses for detecting such clock faults are well known in the background art and need not be described in detail here.
  • In an alternative embodiment, clock skipping module 240 and microprocessor 225 are combined, such that random output 205 can force microprocessor 225 to skip clock cycles. For example, when microprocessor 225 is directed to skip a clock cycle (such as when three output bits equal to zero are received in random output 205), the result of the current or next instruction (or clock cycle) executed by the microprocessor is discarded and repeated.
  • In all of the foregoing, it should be noted that the fraction of skipped clock cycles does not need to be very large; for example and without limitation, even skipping as few as one clock cycle in 20 (on average) will introduce significant measurement drift.
  • One consideration introduced by clock skipping is the effect on other functions of the system besides the cryptographic operations. In particular, clock skipping may sometimes adversely affect operations requiring regular clock cycles. For example, in many smartcards, one bit is sent or received on a serial I/O (input/output) line every 372 cycles of the external clock. (Thus, a 3.579545 MHz external clock is compatible with a serial communication rate of 9600 bits per second.) However, with clock decorrelation, microprocessor 225 will operate at a different clock rate governed by signal 260. A mismatch between the data communications clock rate and the microprocessor clock rate may result, causing I/O errors to occur. Consequently, in devices implementing clock skipping, it is often advantageous for the microprocessor to be controlled by external clock 220 during I/O operations.
  • This can be implemented via clock skipping activation signal 230, which is used to select between external clock signal 220 and the (modified) internal clock that would otherwise be produced by clock skipping module 140. As with the noise generator activation signal of FIG. 1, clock skipping activation signal 220 can be produced by a microprocessor or any other control device that is capable of knowing when to apply (or not apply) the clock skipping. Selection of whether or not to clock skip at any particular time can be performed by many well-known techniques that need not be described in detail here. For example, in the exemplary embodiment of FIG. 2, microprocessor 225 is well suited for such a task because it is necessarily aware of I/O operations associated with the receipt of data signals 270. In general, when I/O is performed or when other non-security-critical operations are in progress, microprocessor core 225 can assert control signal 230 to cause clock skipping module 240 to ignore random output 205 and provide external clock signal 220 directly as clock signal 260. Control signal 230 and the noise production activation control signal described previously can, but need not be the same signal.
  • In an alternative solution to the synchronization failure problem, two separate clocks are used. A conventional external clock signal is used for I/O and other processing, where clock skipping is not needed to protect secret information. However, an internal clock signal, preferably but not necessarily generated in the device (for example, produced using a ring oscillator, which is well known in the background art), is used for internal (e.g., cryptographic) processing. Thus, internal operations need not proceed at a speed related to or derived from the external clock rate. The internal clock may be distorted or skipped, for example, as described above. Alternatively, or in addition, where an analog process is used to generate the internal clock, significant sources of randomness can also be incorporated to adjust the frequency, drift, and jitter of the clock signal to prevent accurate prediction of clock state transitions. Clock signal selection can be performed by microprocessor 225 as mentioned previously. Another technique, which is especially suitable for, but not limited to smartcards, uses a UART (universal asynchronous receiver/transmitter) or other buffer between the internally clocked region and the external I/O interface to ensure that communications over the external serial I/O interface are clocked at a rate corresponding to the externally-supplied clock but may be accessed reliably by internally-clocked circuits.
  • In yet another approach, the internally-generated clock signal can be derived from the external clock signal. This can be performed via an analog phase-locked loop, which is well known in the background art and need not be described in detail here. Such an internal clock will be rather closely correlated with the external clock, and therefore not as effective against attacks as the randomized clock signal or two separate clocks described previously. (Of course, its effectiveness can be improved by optionally using clock skipping or analog (or other) noise sources to adjust the frequency, drift, and jitter of its signal.) Also, when synthesizing the internal clock from the external clock, the clock-derivation circuitry can be configured to restrict the rate of the internal clock frequency, for example, to enforce a minimum internal clock frequency so that attackers cannot stop the clock and attack the device in the stopped state. The derived internal clock signal exhibits a number of useful properties that will be described in the following paragraph.
  • One useful property of such a slightly decorrelated internal clock is that it may be sufficiently close to the external clock that it may be used to control I/O rates reliably. In addition, because a phase-locked loop can continue to produce a valid clock signal even if the external clock changes or is removed, microprocessor 225 can continue operating so that it can detect and even respond to attacks that involve halting, removing, or altering the external clock while power is connected. The use of an internally-generated clock additionally provides protection against attacks involving the introduction of errors into computations. For example, jitter or high frequencies supplied to the external clock would cause harmless communication errors, but would be prevented from causing erroneous computations. Because a phase locked loop can produce an internal clock signal that is a multiple of the external clock signal, it is possible to clock cryptographic operations at a rate faster than the external clock, providing improved performance. In smartcards with challenging performance requirements (such as those that run interpreted codes such as Java), this is an added benefit.
  • All of the foregoing paragraphs describe various ways to generate a second, internal clock signal: via randomization, via a separate clock, or via derivation from the external clock. In all of these cases, the internal clock can also be used to monitor the external clock to detect abnormalities introduced by attackers. Regardless of whether the clock is produced internally or derived from the external clock, the microprocessor can have the option of selecting between multiple clock modes. For example, a slower mode might be used if it has a lower probability of computation error, a faster mode might be used when improved performance is needed, and clock skipping or other clock distortion might be activated when resistance to external monitoring attacks is desired.
  • Much of the foregoing has been described with respect to hardware techniques for clock decorrelation (e.g., second clocks or phase-locked loops), but clock decorrelation can also be effected by software as will be described below. This is useful, for example, where the environment does not allow for hardware-based clock skipping. Alternatively, hardware clock decorrelation could be supplemented with software-based clock decorrelation for even greater protection in security critical code segments.
  • One efficient software-implementable technique for clock decorrelation takes advantage of the fact that the amount of time used for a process with a conditional jump can vary depending on whether or not the jump is actually performed. In such cases, inserting branch delays can be used as a form of clock decorrelation. For example, the assembly language clock randomizer below uses a random generator to introduce clock variations that can help prevent accurate alignment by an attacker:
  • Assembly Language Clock Randomizer:
     [...]
     inp reg5,RANDOM_GENERATOR # get a random byte
     add reg5,reg5 # shift reg5 left once
     brc delay1 # branch if carry
     nop # extra delay if bit is 0
    delay1: # continue execution
     [...] # ...more code...
     add reg5,reg5 # shift reg5 left again
     brc delay_2 # branch if carry
      # put any code here # obfuscating code/delay
    delay2: # continue execution
     [...] # ...more code...
  • In an alternative embodiment, instead of using random information to determine whether to take a delay branch, the random information may be used to select between parallel code processes, such that the same cryptographic result will be produced regardless of which code process is selected but where the parallel processes perform different operations toward producing the result.
  • This section has described temporal obfuscation techniques that are useful in preventing reliable identification and alignment of specific features in measurements of cryptographic device characteristics such as power consumption and electromagnetic radiation. However, such techniques may not always be sufficient for preventing attacks based on timing, since introduced timing delays will have a predictable statistical distribution for which attackers can compensate. Another embodiment of the general technique of implementing cryptographic protocols using unpredictable information, described below, is useful in (but is not limited to) such contexts.
  • Execution Path and Operation Order Entropy
  • Another approach to the general technique of using unpredictable information to protect cryptosystems against external monitoring attacks involves the introduction of entropy into the order of processing operations or into the execution path while maintaining desired functionality (such as compatibility with standard cryptographic algorithm definitions). More specifically, a device can use a random number generator to cause unpredictability in the order of performing a sequence of suboperations. If attackers cannot accurately determine the order in which operations were performed, cross-correlation between samples becomes more difficult or impossible. Consequently the data collected by an attacker effectively has a significantly lower signal-to-noise ratio.
  • As an illustrative example of operation order entropy, consider a bit permutation. Permutations are widely used in cryptography, for example in the Data Encryption Standard and other cryptographic algorithms. The following C language pseudocode illustrates a traditional method of implementing a permutation.
  • Input-Ordered Permutation (Background Art):
      void perm2(bool dataIn[64], bool dataOut[64], int table1[64]) {
       int i;
       for (i = 0; i < 64; i++) {
        dataOut[table1[i]] = dataIn[i];
       }
      }
  • This example is input-ordered, meaning that processing steps are performed in the order (or inverse order) in which the input bits are supplied. In the example, input bit 0 is permuted first, and input bit 63 is permuted last. Output-ordered permutations are also commonly used in the background art. Provided that table1 is a permutation (i.e., where one element equals each of the values 0 . . . 63), the pseudocode below can be made output-ordered by changing the statement inside the loop to read: “dataOut[i]=dataIn[table2[i]];”, where table2 is output-ordered (i.e., table2 is the inverse of table1 above such that table1[table2[i]]=i).
  • However, both output-ordered and input-ordered permutations can leak information about the data they process. For example, in the input-ordered permutation, attackers' measurements of loop iteration i will be correlated to dataIn[i]. In the output-ordered permutation, the attackers' measurements of loop iteration i will be correlated to dataOut[i]. An improved permutation method would thus be advantageous. One exemplary implementation of such a method is shown in the table below. This high-entropy permutation combines several previously-described aspects of the countermeasures, including without limitation order randomization (thus being neither input-ordered nor output-ordered) and blinding techniques (to conceal further the data being permuted).
  • Blinded High-Entropy Permutation:
    #define SWAP(a,b) { register int t = a; a = b; b = t; }
    #define LOOPCOUNT 128
    void perm3(bool dataIn[64], bool dataOut[64], int table[64]) {
     unsigned char trueRandom(void);   /* gives random byte */
     int i,p;
     int perm[64];
     bool b, temp[64];
     /* Initialize random permutation */
     for (i = 0; i < 64; i++) {
      perm[i] = i;
      temp[i] = trueRandom( ) & 1;
      dataOut[i] = trueRandom( ) & 1;
     }
     for (i = 0; i < LOOPCOUNT; i++) {
      p = trueRanddm( ) & 63;     /* random number mod 64 */
      SWAP(perm[p], perm[i&63]);
     }
     /* Blind: temp=blinded input, dataOut=unblinding factor */
     for (i = 0; i < 64; i++) {
      p = perm[i];
      b = (bool)(trueRandom( ) & 1);
      temp[p] = dataIn[p] {circumflex over ( )} b;
      dataOut[table[p]] = b;
     }
     for (i = 0; i < LOOPCOUNT; i++) {
      p = trueRandom( ) & 63;      /* random number mod 64 */
      SWAP(perm[p], perm[i&63]);
     }
     /* Perform the permutation on temp & unblind */
     for (i = 0; i < 64; i++) {
      p = perm[i];
      dataOut[table[p]] {circumflex over ( )}= temp[p];
      temp[p] = 0;
     }
    }
  • The magnitude of signals leaked due to variations in data values (e.g., registers and memory contents) is usually smaller (often by a factor of several orders of magnitude) than signals leaked due to branches and variations in the execution path. Therefore, the high-entropy permutation operation, above, uses a constant execution path to inhibit leakage via variations in the execution path.
  • The exemplary blinded randomized-order permutation operation includes four steps, which can be performed separately or simultaneously: initialization, blinding, permutation, and unblinding. Implementations using partial blinding, which operate on already-blinded values, or those with reduced security requirements will not require all steps.
  • Initialization of the blinded randomized-order permutation operation involves constructing and randomizing a permutation table (“perm”) for determining the bit order for operations. (Bit order permutation table “perm” randomizes the time at which any particular data bit is manipulated.) The bit order table is created in two passes, where the first assures that the table has the correct form (i.e., contains the numbers zero through 63), and the second introduces random order into the table. Because the process of constructing the bit order table does not involve any secret inputs, the only security requirement for the process is that the final result be unknown to attackers. As illustrated, the first permutation table initialization loop can also place random values into dataOut and temp to help whiten any leaked signals when data values are first stored in these arrays. Finally, although it is not required, more than 64 iterations of the randomization loop are used to ensure that any statistical biases remaining after the randomization loop are insignificantly small.
  • The next section of the code performs the blinding operation. First, for each loop iteration, a random number generator produces a random blinding bit. The temporary buffer (temp) is initialized with the XOR of the random bit and an input data bit, where the input data bit is selected according to the table (perm) constructed previously. Additionally, the output buffer (dataOut) is initialized with the blinding bit, where the dataOut bit is the result of using the input permutation table to operate on the index to temp. The second part of the blinding process re-randomizes the bit order permutation table (perm).
  • The last section performs the final bit permutation and unblinding steps. Input bits are loaded in the order specified by the table (perm), permuted according to the (non-secret) externally-specified permutation table (table), and XORed onto the destination table (dataOut).
  • Note that the leak-minimized permutation operation described dramatically reduces the amount of information leaked from a permutation operation, but is not necessarily expected to reduce such leakage to zero. The input data to the function arrives in fixed order and unblinded form, and the output is similarly supplied unblinded in fixed order. Consequently, two or more measurements from the same transaction might (for example) be correlated to each other such that the strength or sign of the correlation is a function of one or more input or output data bits. If inputs and/or outputs must be kept secret or if multiple permutations are to be performed on the same secret data (for example, through a multi-step operation such as encryption), operands can be maintained in a blinded state during processing, to be (partially or completely) reconstituted only when nonlinear operations must be performed or at the end of the computation.
  • Note that many variations on the process described are possible, as will be understood to those skilled in the art. For example and without limitation, the number of bits manipulated does not need to equal 64, the order of steps may be changed, steps can be removed for simplified implementations (such as those that are not subject to some attacks), steps can be modified, different permutation generation and update processes can be used, and additional steps can be added.
  • Other Considerations
  • Cryptographic operations should normally be checked to ensure that incorrect computations do not compromise keys or enable other attacks. Cryptographic implementations of the countermeasures can be, and in a preferred embodiment are, combined with error-detection and/or error-correction logic to ensure that cryptographic operations are performed correctly. For example, a simple and effective technique is to perform cryptographic operations twice, ideally using two independent hardware processors and/or software implementations, with a comparison operation performed at the end to verify that both produce identical results. If the results produced by the two units do not match, the failed comparison will prevent the defective processing result from being used. In situations where security is more important than reliability, if the compare operation ever fails (or fails too many times) the device may self-destruct (such as by deleting internal keys) or disable itself. For example, a device might erase its key storage memory if either two defective DES operations occur sequentially or five defective DES results occur during the lifetime of the device. In some cryptosystems, full redundancy is not necessary. For example, with RSA, methods are known in the background art for self-checking functions that can be incorporated into the cryptosystem implementation (e.g., RSA signatures can be verified after digital signing operations).
  • Detection of conditions likely to cause incorrect results may also be used. In particular, active or passive sensors to detect unusually high or low voltages, high-frequency noise on voltage or signal inputs, exposure to electromagnetic fields and radiation, and physical tampering may be employed. Inappropriate operating conditions can (for example) trigger the device to reset, delete secrets, or self-destruct.
  • Self-diagnostic functions such as a POST (power-on-self-test) should also be incorporated to verify that cryptographic functions have not been damaged. In cases where an ATR (answer-to-reset) must be provided before a comprehensive self-test can be completed, the self-test can be deferred until after completion of the first transaction or until a sufficient idle period is encountered. For example, a flag indicating successful POST completion can be cleared upon initialization. While the card is waiting for a command from the host system, it can attempt the POST. Any I/O received during the POST will cause an interrupt, which will cancel the POST (leaving the POST-completed flag at zero). If any cryptographic function is called, the device will check the POST flag and (if it is not set) perform the POST before doing any cryptographic operations.
  • CONCLUSIONS
  • The countermeasures are extremely useful for improving security, particularly in environments and applications with difficult engineering requirements, by enabling the construction of devices that are significantly more resistant to attack than devices of similar cost and complexity that do not use such countermeasures. Also, multiple security techniques may be required to make a system secure. For example, leak minimization and obfuscation may be used in conjunction with other security methods or countermeasures.
  • As those skilled in the art will appreciate, the techniques described above are not limited to particular host environments or form factors. Rather, they may be used in a wide variety of applications, including without limitation: cryptographic smartcards of all kinds including without limitation smartcards substantially compliant with ISO 7816-1, ISO 7816-2, and ISO 7816-3 (“ISO 7816-compliant smartcards”); contactless and proximity-based smartcards and cryptographic tokens; stored value cards and systems; cryptographically secured credit and debit cards; customer loyalty cards and systems; cryptographically authenticated credit cards; cryptographic accelerators; gambling and wagering systems; secure cryptographic chips; tamper-resistant microprocessors; software programs (including without limitation programs for use on personal computers, servers, etc. and programs that can be loaded onto or embedded within cryptographic devices); key management devices; banking key management systems; secure web servers; electronic payment systems; micropayment systems and meters; prepaid telephone cards; cryptographic identification cards and other identity verification systems; systems for electronic funds transfer; automatic teller machines; point of sale terminals; certificate issuance systems; electronic badges; door entry systems; physical locks of all kinds using cryptographic keys; systems for decrypting television signals (including without limitation, broadcast television, satellite television, and cable television); systems for decrypting enciphered music and other audio content (including music distributed over computer networks); systems for protecting video signals of all kinds; intellectual property protection and copy protection systems (such as those used to prevent unauthorized copying or use of movies, audio content, computer programs, video games, images, text, databases, etc.); cellular telephone scrambling and authentication systems (including telephone authentication smartcards); secure telephones (including key storage devices for such telephones); cryptographic PCMCIA cards; portable cryptographic tokens; and cryptographic data auditing systems. All of the foregoing illustrates exemplary embodiments and applications of the countermeasures, from which related variations, enhancements and modifications will be apparent without departing from the spirit and scope of the family of countermeasures addressed hereby. Therefore, the scope of any claims directed to countermeasures should not be limited to the foregoing disclosure, but rather construed by any such claims appended hereto. Similarly, the foregoing also illustrates exemplary embodiments and applications of the external monitoring (e.g., attacks) against which countermeasures are disclosed, from which related variations, enhancements and modifications will be apparent without departing from the spirit and scope of the of the family of external monitoring techniques addressed hereby. Therefore, the scope of any claims directed to external monitoring techniques should not be limited to the foregoing disclosure, but rather construed by any such claims appended hereto.

Claims (21)

1-7. (canceled)
8. A method for evaluating the security of a device, the method comprising:
connecting the device to an analog-to-digital converter;
sending a plurality of commands to the device, wherein each command causes the device to process data using a key;
recording a plurality of measurements of an attribute of the device by using the analog-to-digital converter; and
determining whether information about the key is leaking from the device.
9. The method of claim 8, wherein sending a plurality of commands to the device comprises sending a plurality of command sequences, further wherein each command sequence causes the device to process data using the key.
10. The method of claim 8, wherein each command sent to the device causes the device to perform a cryptographic operation using the key.
11. The method of claim 10, wherein the cryptographic operation includes transforming with a block cipher.
12. The method of claim 10, wherein recording a plurality of measurements of an attribute is performed during processing of each cryptographic operation.
13. The method of claim 8, wherein determining whether information about the key is leaking from the device comprises statistically combining the recorded plurality of measurements.
14. The method of claim 8, wherein determining whether information about the key is leaking from the device comprises temporally aligning data points corresponding to a point of interest within the recorded plurality of measurements.
15. The method of claim 8, further comprising determining information about the key.
16. The method of claim 15, wherein the information about the key comprises values of a plurality of key bits.
17. The method of claim 8, wherein:
the analog-to-digital converter is configured to measure electromagnetic radiation; and
recording a plurality of measurements of an attribute of the device by using the analog-to-digital converter comprises recording a plurality of measurements of electromagnetic radiation by using the analog-to-digital converter.
18. The method of claim 8, wherein:
the analog-to-digital converter is configured to measure variations in an amount of power consumed on an external power input to the device; and
recording a plurality of measurements of an attribute of the device by using the analog-to-digital converter comprises recording a plurality of measurements of an amount of power consumed on the external power input to the device taken by the analog-to-digital converter.
19. A system for evaluating the security of cryptographic hardware, the system comprising:
an analog-to-digital converter configured to connect to a device and measure an attribute related to operation of the device, the device comprising a key and a circuit configured to perform operations using the key;
a data storage system configured to record a plurality of measurements of the attribute taken by the analog-to-digital converter; and
a processor configured to determine whether information about the key is leaking from the device.
20. The system of claim 19, wherein the circuit is configured to perform cryptographic operations using the key.
21. The system of claim 20, wherein the cryptographic operations include transforming with a block cipher.
22. The system of claim 20, wherein the data storage system is configured to record a plurality of measurements of the attribute during processing of each cryptographic operation.
23. The system of claim 19, wherein the processor is configured to determine whether information about the key is leaking from the device by statistically combining the recorded plurality of measurements.
24. The system of claim 19, further comprising a data filtering system configured to temporally align data points corresponding to a point of interest within the recorded plurality of measurements.
25. The system of claim 19, wherein the processor is further configured to determine information about the key.
26. The system of claim 19, wherein:
the analog-to-digital converter is configured to measure electromagnetic radiation; and
the data storage system is configured to record a plurality of measurements of electromagnetic radiation taken by the analog-to-digital converter.
27. The system of claim 19, wherein:
the analog-to-digital converter is configured to measure variations in an amount of power consumed on an external power input to the device; and
the data storage system is configured to record a plurality of measurements of an amount of power consumed on the external power input to the device taken by the analog-to-digital converter.
US15/236,739 1998-01-02 2016-08-15 Differential power analysis - resistant cryptographic processing Abandoned US20170099134A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US15/236,739 US20170099134A1 (en) 1998-01-02 2016-08-15 Differential power analysis - resistant cryptographic processing

Applications Claiming Priority (14)

Application Number Priority Date Filing Date Title
US7034498P 1998-01-02 1998-01-02
US8782698P 1998-06-03 1998-06-03
US8788098P 1998-06-03 1998-06-03
US8952998P 1998-06-15 1998-06-15
US09/224,682 US6304658B1 (en) 1998-01-02 1998-12-31 Leak-resistant cryptographic method and apparatus
US09/326,222 US6327661B1 (en) 1998-06-03 1999-06-03 Using unpredictable information to minimize leakage from smartcards and other cryptosystems
US09/324,798 US6278783B1 (en) 1998-06-03 1999-06-03 Des and other cryptographic, processes with leak minimization for smartcards and other cryptosystems
US09/737,182 US6381699B2 (en) 1998-01-02 2000-12-13 Leak-resistant cryptographic method and apparatus
US09/930,836 US7668310B2 (en) 1998-06-03 2001-08-15 Cryptographic computation using masking to prevent differential power analysis and other attacks
US10/005,105 US7587044B2 (en) 1998-01-02 2001-12-03 Differential power analysis method and apparatus
US11/643,349 US7634083B2 (en) 1998-01-02 2006-12-21 Differential power analysis
US12/637,565 US8879724B2 (en) 1998-01-02 2009-12-14 Differential power analysis—resistant cryptographic processing
US14/530,905 US9419790B2 (en) 1998-01-02 2014-11-03 Differential power analysis—resistant cryptographic processing
US15/236,739 US20170099134A1 (en) 1998-01-02 2016-08-15 Differential power analysis - resistant cryptographic processing

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US14/530,905 Continuation US9419790B2 (en) 1998-01-02 2014-11-03 Differential power analysis—resistant cryptographic processing

Publications (1)

Publication Number Publication Date
US20170099134A1 true US20170099134A1 (en) 2017-04-06

Family

ID=27578324

Family Applications (6)

Application Number Title Priority Date Filing Date
US10/005,105 Expired - Fee Related US7587044B2 (en) 1998-01-02 2001-12-03 Differential power analysis method and apparatus
US11/643,349 Expired - Fee Related US7634083B2 (en) 1998-01-02 2006-12-21 Differential power analysis
US11/978,364 Expired - Fee Related US7599488B2 (en) 1998-01-02 2007-10-29 Differential power analysis
US12/637,565 Expired - Fee Related US8879724B2 (en) 1998-01-02 2009-12-14 Differential power analysis—resistant cryptographic processing
US14/530,905 Expired - Fee Related US9419790B2 (en) 1998-01-02 2014-11-03 Differential power analysis—resistant cryptographic processing
US15/236,739 Abandoned US20170099134A1 (en) 1998-01-02 2016-08-15 Differential power analysis - resistant cryptographic processing

Family Applications Before (5)

Application Number Title Priority Date Filing Date
US10/005,105 Expired - Fee Related US7587044B2 (en) 1998-01-02 2001-12-03 Differential power analysis method and apparatus
US11/643,349 Expired - Fee Related US7634083B2 (en) 1998-01-02 2006-12-21 Differential power analysis
US11/978,364 Expired - Fee Related US7599488B2 (en) 1998-01-02 2007-10-29 Differential power analysis
US12/637,565 Expired - Fee Related US8879724B2 (en) 1998-01-02 2009-12-14 Differential power analysis—resistant cryptographic processing
US14/530,905 Expired - Fee Related US9419790B2 (en) 1998-01-02 2014-11-03 Differential power analysis—resistant cryptographic processing

Country Status (1)

Country Link
US (6) US7587044B2 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022029443A1 (en) 2020-08-07 2022-02-10 Pugged Code Limited Method and apparatus for reducing the risk of successful side channel and fault injection attacks

Families Citing this family (111)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7587044B2 (en) * 1998-01-02 2009-09-08 Cryptography Research, Inc. Differential power analysis method and apparatus
US6304658B1 (en) * 1998-01-02 2001-10-16 Cryptography Research, Inc. Leak-resistant cryptographic method and apparatus
ES2660057T3 (en) * 1998-05-18 2018-03-20 Giesecke + Devrient Mobile Security Gmbh Protected access data storage support
EP1090480B1 (en) 1998-06-03 2019-01-09 Cryptography Research, Inc. Improved des and other cryptographic processes with leak minimization for smartcards and other cryptosystems
US6539092B1 (en) 1998-07-02 2003-03-25 Cryptography Research, Inc. Leak-resistant cryptographic indexed key update
JP4317607B2 (en) * 1998-12-14 2009-08-19 株式会社日立製作所 Information processing equipment, tamper resistant processing equipment
FR2818847A1 (en) * 2000-12-26 2002-06-28 St Microelectronics Sa VARIABLE POLARITY LOGIC CIRCUIT
DE10162309A1 (en) * 2001-12-19 2003-07-03 Philips Intellectual Property Method and arrangement for increasing the security of circuits against unauthorized access
JP3672874B2 (en) * 2002-01-25 2005-07-20 Necマイクロシステム株式会社 Data processing device
US8074081B2 (en) * 2002-04-15 2011-12-06 Infineon Technologies Ag Method for replacing contents of a data storage unit
US7376235B2 (en) * 2002-04-30 2008-05-20 Microsoft Corporation Methods and systems for frustrating statistical attacks by injecting pseudo data into a data system
FR2840083A1 (en) * 2002-05-24 2003-11-28 St Microelectronics Sa TEST OF AN ALGORITHM EXECUTED BY AN INTEGRATED CIRCUIT
FR2844891A1 (en) 2002-09-20 2004-03-26 St Microelectronics Sa Microprocessor handled digital data masking method involves factorizing digital data by residue number system based on finite base of numbers or polynomials prime to one another
FR2849232B1 (en) * 2002-12-24 2005-02-25 Trusted Logic METHOD FOR SECURING COMPUTER SYSTEMS INCORPORATING A CODE INTERPRETATION MODULE
FR2849233B1 (en) * 2002-12-24 2005-05-20 Trusted Logic METHOD FOR SECURING COMPUTER SYSTEMS BY SOFTWARE CONFINEMENT
US20040128528A1 (en) * 2002-12-31 2004-07-01 Poisner David I. Trusted real time clock
SE526551C2 (en) * 2003-04-29 2005-10-04 Haldex Brake Prod Ab Location of brake caliper
EP1678568A1 (en) * 2003-10-23 2006-07-12 Koninklijke Philips Electronics N.V. Method for protecting an information carrier comprising an integrated circuit
WO2005043299A2 (en) 2003-10-24 2005-05-12 Atmel Corporation Method and apparatus for a variable processing period in an integrated circuit
FR2861474B1 (en) * 2003-10-24 2007-04-27 Atmel Corp METHOD AND APPARATUS FOR A VARIABLE PROCESSING PERIOD IN AN INTEGRATED CIRCUIT
JP4674440B2 (en) * 2004-03-04 2011-04-20 ソニー株式会社 Data processing circuit
DE102004016342B4 (en) * 2004-04-02 2006-07-06 Infineon Technologies Ag Detection device for detecting manipulative attacks on electrical circuits, especially for chip card use, wherein the characteristic current over a specified time period is recorded and compared with a stored characteristic curve
US20050286719A1 (en) * 2004-06-29 2005-12-29 Canon Kabushiki Kaisha Generating entropy through image capture
WO2006028558A1 (en) * 2004-09-03 2006-03-16 Virgina Tech Intellectual Properties, Inc. Detecting software attacks by monitoring electric power consumption patterns
US8621597B1 (en) * 2004-10-22 2013-12-31 Xilinx, Inc. Apparatus and method for automatic self-erasing of programmable logic devices
US7478902B2 (en) * 2004-11-04 2009-01-20 Hewlett-Packard Development Company, L.P. Inkjet compositions
US7764785B2 (en) 2004-11-08 2010-07-27 King Fahd University Of Petroleum And Minerals Method for communicating securely over an insecure communication channel
US8511558B2 (en) * 2005-04-12 2013-08-20 Sandisk Il Ltd. Smartcard power management
US7509250B2 (en) * 2005-04-20 2009-03-24 Honeywell International Inc. Hardware key control of debug interface
US8874477B2 (en) 2005-10-04 2014-10-28 Steven Mark Hoffberg Multifactorial optimization system and method
DE602006020010D1 (en) * 2005-12-19 2011-03-24 St Microelectronics Sa Protection of the execution of a DES algorithm
DE102006005053B4 (en) * 2006-02-03 2012-10-25 Infineon Technologies Ag Apparatus and method for detecting an attack on an electrical circuit
US20070220369A1 (en) * 2006-02-21 2007-09-20 International Business Machines Corporation Fault isolation and availability mechanism for multi-processor system
US20070288738A1 (en) 2006-06-09 2007-12-13 Dale Jason N System and method for selecting a random processor to boot on a multiprocessor system
US7594104B2 (en) * 2006-06-09 2009-09-22 International Business Machines Corporation System and method for masking a hardware boot sequence
US20070288740A1 (en) * 2006-06-09 2007-12-13 Dale Jason N System and method for secure boot across a plurality of processors
US20070288761A1 (en) * 2006-06-09 2007-12-13 Dale Jason N System and method for booting a multiprocessor device based on selection of encryption keys to be provided to processors
US7774616B2 (en) * 2006-06-09 2010-08-10 International Business Machines Corporation Masking a boot sequence by providing a dummy processor
US8301890B2 (en) * 2006-08-10 2012-10-30 Inside Secure Software execution randomization
US7613907B2 (en) * 2006-08-11 2009-11-03 Atmel Corporation Embedded software camouflage against code reverse engineering
US7984301B2 (en) * 2006-08-17 2011-07-19 Inside Contactless S.A. Bi-processor architecture for secure systems
US7554865B2 (en) * 2006-09-21 2009-06-30 Atmel Corporation Randomizing current consumption in memory devices
JP4680876B2 (en) * 2006-12-11 2011-05-11 ルネサスエレクトロニクス株式会社 Information processing apparatus and instruction fetch control method
US7761484B2 (en) * 2007-02-09 2010-07-20 Microsoft Corporation Complete mapping between the XML infoset and dynamic language data expressions
EP2000936A1 (en) * 2007-05-29 2008-12-10 Gemplus Electronic token comprising several microprocessors and method of managing command execution on several microprocessors
FR2925968B1 (en) * 2007-12-26 2011-06-03 Ingenico Sa MICROPROCESSOR SECURING METHOD, COMPUTER PROGRAM AND CORRESPONDING DEVICE
DE102008003531A1 (en) * 2008-01-08 2009-07-09 Giesecke & Devrient Gmbh software identification
FR2928060B1 (en) * 2008-02-25 2010-07-30 Groupe Des Ecoles De Telecommunications Get Ecole Nat Superieure Des Telecommunications Enst METHOD FOR TESTING CRYPTOGRAPHIC CIRCUITS, SECURED CRYPTOGRAPHIC CIRCUIT FOR TESTING, AND METHOD FOR WIRING SUCH CIRCUIT.
FR2935059B1 (en) * 2008-08-12 2012-05-11 Groupe Des Ecoles De Telecommunications Get Ecole Nationale Superieure Des Telecommunications Enst METHOD FOR DETECTING ANOMALIES IN A DIFFERENTIAL LOGIC-PROTECTED CRYPTOGRAPHIC CIRCUIT AND CIRCUIT USING SUCH A METHOD
FR2935817B1 (en) * 2008-09-11 2011-12-09 Oberthur Technologies DATA PROCESSING METHOD AND ASSOCIATED DEVICE.
US9263085B2 (en) 2009-05-20 2016-02-16 Sony Dadc Austria Ag Method for copy protection
EP2282279A1 (en) 2009-07-20 2011-02-09 Nxp B.V. Method and device for detection of light attacks
KR101141600B1 (en) 2009-08-19 2012-05-17 한국전자통신연구원 Apparatus for measuring power consumption and generating a trigger for side channel analysis and method thereof
EP2290575A1 (en) * 2009-08-31 2011-03-02 Incard SA IC Card comprising an improved processor
US8255702B1 (en) 2009-12-03 2012-08-28 Altera Corporation Programmable logic device with improved security
EP2507708B1 (en) 2009-12-04 2019-03-27 Cryptography Research, Inc. Verifiable, leak-resistant encryption and decryption
KR101418962B1 (en) * 2009-12-11 2014-07-15 한국전자통신연구원 Secure device and method for preventing side chnannel attack
FR2957440B1 (en) 2010-03-09 2012-08-17 Proton World Int Nv PROTECTION OF A SECURITY MODULE IN A TELECOMMUNICATION DEVICE COUPLED TO AN NFC CIRCUIT
FR2957438B1 (en) 2010-03-09 2012-03-30 Proton World Int Nv DETECTION OF A DEROUTEMENT OF A COMMUNICATION CHANNEL OF A TELECOMMUNICATION DEVICE COUPLED TO AN NFC CIRCUIT
FR2957439B1 (en) 2010-03-09 2012-03-30 Proton World Int Nv PROTECTION OF A COMMUNICATION CHANNEL BETWEEN A SECURITY MODULE AND AN NFC CIRCUIT
DE102010010851A1 (en) 2010-03-10 2011-09-15 Giesecke & Devrient Gmbh Spying protection when executing an operation sequence in a portable data carrier
GB2479871A (en) * 2010-04-26 2011-11-02 David Coyne System for preventing side channel attacks on a synchronous logic device.
US8966253B1 (en) 2010-06-01 2015-02-24 Xilinx, Inc. Method and apparatus for authenticating a programmable device bitstream
US8539254B1 (en) 2010-06-01 2013-09-17 Xilinx, Inc. Method and integrated circuit for protecting against differential power analysis attacks
US8583944B1 (en) 2010-08-04 2013-11-12 Xilinx, Inc. Method and integrated circuit for secure encryption and decryption
US8832462B2 (en) 2010-09-08 2014-09-09 Xilinx, Inc. Protecting against differential power analysis attacks on sensitive data
US8650408B2 (en) 2010-09-08 2014-02-11 Xilinx, Inc. Protecting against differential power analysis attacks on decryption keys
FR2969341B1 (en) 2010-12-20 2013-01-18 Proton World Int Nv MANAGING COMMUNICATION CHANNELS IN A TELECOMMUNICATION DEVICE COUPLED TO AN NFC CIRCUIT
GB2487901B (en) * 2011-02-03 2019-12-04 Advanced Risc Mach Ltd Power signature obfuscation
US8909941B1 (en) 2011-03-31 2014-12-09 Xilinx, Inc. Programmable integrated circuit and a method of enabling the detection of tampering with data provided to a programmable integrated circuit
US8958550B2 (en) * 2011-09-13 2015-02-17 Combined Conditional Access Development & Support. LLC (CCAD) Encryption operation with real data rounds, dummy data rounds, and delay periods
KR101869064B1 (en) 2012-05-31 2018-07-19 한국전자통신연구원 Side-channel analyzer and method based on profile
DE102012209404A1 (en) 2012-06-04 2013-12-05 Robert Bosch Gmbh Apparatus for executing a cryptographic method and method of operation therefor
EP2674608B1 (en) 2012-06-13 2015-08-12 Delphi International Operations Luxembourg S.à r.l. Fuel injector
WO2014028005A1 (en) * 2012-08-14 2014-02-20 Empire Technology Development Llc Software-based side-channel attack prevention
US10103876B2 (en) 2012-10-17 2018-10-16 Synopsys, Inc. System and method for multichannel cryptographic processing
DE102012219205A1 (en) 2012-10-22 2014-05-08 Robert Bosch Gmbh Apparatus and method for carrying out a cryptographic method
CN104854531B (en) * 2012-12-13 2018-05-18 相干逻辑公司 Clock generating circuit reconfigures
TWI517655B (en) * 2013-05-23 2016-01-11 晨星半導體股份有限公司 Cryptographic device and secret key protection method
WO2014205398A1 (en) * 2013-06-21 2014-12-24 Cryptography Research, Inc. Energy analysis for differential power analysis resistance
US9009495B2 (en) 2013-06-28 2015-04-14 Envieta, LLC High speed cryptographic combining system, and method for programmable logic devices
US9172380B2 (en) * 2013-07-04 2015-10-27 Samsung Electronics Co., Ltd. Method and apparatus for supporting self-destruction function in baseband modem
US9423820B2 (en) * 2013-07-31 2016-08-23 Fairchild Semiconductor Corporation Complimentary bit slicing side channel attack defense
FR3010210B1 (en) * 2013-08-29 2017-01-13 Stmicroelectronics Rousset PROTECTION OF CALCULATION AGAINST HIDDEN CHANNEL ATTACKS
US9172718B2 (en) * 2013-09-25 2015-10-27 International Business Machines Corporation Endpoint load rebalancing controller
EP3211824B1 (en) * 2014-09-25 2020-12-09 Nec Corporation Analysis system, analysis method, and analysis program
WO2016083864A1 (en) 2014-11-25 2016-06-02 Institut Mines-Telecom Methods for recovering secret data of a cryptographic device and for evaluating the security of such a device
US9735953B2 (en) * 2015-03-06 2017-08-15 Qualcomm Incorporated Side channel analysis resistant architecture
US10530566B2 (en) * 2015-04-23 2020-01-07 Cryptography Research, Inc. Configuring a device based on a DPA countermeasure
US10642962B2 (en) 2015-07-28 2020-05-05 Western Digital Technologies, Inc. Licensable function for securing stored data
US10489611B2 (en) * 2015-08-26 2019-11-26 Rambus Inc. Low overhead random pre-charge countermeasure for side-channel attacks
WO2017049111A1 (en) * 2015-09-18 2017-03-23 Jung-Min Park Group signatures with probabilistic revocation
EP3217307B1 (en) * 2016-02-22 2018-11-07 Eshard Method of testing the resistance of a circuit to a side channel analysis of second order or more
CN107465645A (en) * 2016-06-03 2017-12-12 华邦电子股份有限公司 Breaking device and method
US10255462B2 (en) 2016-06-17 2019-04-09 Arm Limited Apparatus and method for obfuscating power consumption of a processor
US10200192B2 (en) 2017-04-19 2019-02-05 Seagate Technology Llc Secure execution environment clock frequency hopping
US10459477B2 (en) 2017-04-19 2019-10-29 Seagate Technology Llc Computing system with power variation attack countermeasures
US10270586B2 (en) 2017-04-25 2019-04-23 Seagate Technology Llc Random time generated interrupts in a cryptographic hardware pipeline circuit
US10511433B2 (en) 2017-05-03 2019-12-17 Seagate Technology Llc Timing attack protection in a cryptographic processing system
US10771236B2 (en) 2017-05-03 2020-09-08 Seagate Technology Llc Defending against a side-channel information attack in a data storage device
US10997322B2 (en) 2017-05-22 2021-05-04 Arm Limited Efficient power distribution
US10924261B2 (en) 2017-05-22 2021-02-16 Arm Limited Efficient power distribution
EP3506049A1 (en) 2017-12-26 2019-07-03 Ikerlan, S. Coop. Computing device
US11308239B2 (en) * 2018-03-30 2022-04-19 Seagate Technology Llc Jitter attack protection circuit
CN109086130B (en) * 2018-06-06 2022-06-10 北京嘉楠捷思信息技术有限公司 Chip frequency modulation method and device of computing equipment, computing force board, computing equipment and storage medium
US10785016B2 (en) 2018-07-25 2020-09-22 Silicon Laboratories, Inc. Countermeasure for power injection security attack
US11232195B2 (en) 2019-07-29 2022-01-25 Intertrust Technologies Corporation Systems and methods for managing state
US11087030B2 (en) 2019-11-19 2021-08-10 Silicon Laboratories Inc. Side-channel attack mitigation for secure devices with embedded sensors
US11704443B2 (en) 2019-11-19 2023-07-18 Silicon Laboratories Inc. Block cipher side-channel attack mitigation for secure devices
FR3117232B1 (en) 2020-12-08 2024-02-23 St Microelectronics Sa Scrambling the current signature of an integrated circuit
US11394308B1 (en) 2021-05-05 2022-07-19 Arm Limited Apparatuses and methods for power isolation

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5533123A (en) * 1994-06-28 1996-07-02 National Semiconductor Corporation Programmable distributed personal security
US5917909A (en) * 1992-12-23 1999-06-29 Gao Gesellschaft Fur Automation Und Organisation Mbh System for testing the authenticity of a data carrier

Family Cites Families (247)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US2733432A (en) 1956-01-31 Breckman
FR561910A (en) 1922-02-11 1923-10-30
US4214126A (en) 1945-04-30 1980-07-22 Rca Corporation Cadence suppression system
US2632058A (en) 1946-03-22 1953-03-17 Bell Telephone Labor Inc Pulse code communication
US3816762A (en) 1973-01-02 1974-06-11 Fairchild Camera Instr Co Noise suppression circuit
US4243890A (en) 1976-08-23 1981-01-06 Miller Bruce J Isolator/switching assembly for data processing terminal
US4107458A (en) 1976-08-23 1978-08-15 Constant James N Cipher computer and cryptographic system
US4139839A (en) 1977-03-18 1979-02-13 Nasa Digital data reformatter/deserializer
US4077061A (en) * 1977-03-25 1978-02-28 Westinghouse Electric Corporation Digital processing and calculating AC electric energy metering system
FR2401459A1 (en) 1977-08-26 1979-03-23 Cii Honeywell Bull PORTABLE INFORMATION MEDIA EQUIPPED WITH A MICROPROCESSOR AND A PROGRAMMABLE DEAD MEMORY
US4295041A (en) 1977-08-26 1981-10-13 Compagnie Internationale Pour L'informatique Cii-Honeywell Bull (Societe Anonyme) Device for the protection of access to a permanent memory of a portable data carrier
US4200770A (en) 1977-09-06 1980-04-29 Stanford University Cryptographic apparatus and method
US4202051A (en) 1977-10-03 1980-05-06 Wisconsin Alumni Research Foundation Digital data enciphering and deciphering circuit and method
CH623271A5 (en) 1977-11-15 1981-05-29 Hasler Ag
US4203166A (en) 1977-12-05 1980-05-13 International Business Machines Corporation Cryptographic file security for multiple domain networks
US4405829A (en) 1977-12-14 1983-09-20 Massachusetts Institute Of Technology Cryptographic communications system and method
US4309569A (en) 1979-09-05 1982-01-05 The Board Of Trustees Of The Leland Stanford Junior University Method of providing digital signatures
US4369332A (en) 1979-09-26 1983-01-18 Burroughs Corporation Key variable generator for an encryption/decryption device
US4268898A (en) 1980-03-20 1981-05-19 Lorain Products Corporation Semiconductor switching circuit with clamping and energy recovery features
DE3127843A1 (en) 1981-07-15 1983-05-26 AEG-Telefunken Nachrichtentechnik GmbH, 7150 Backnang METHOD FOR PREVENTING "COMPROMISING RADIATION" IN PROCESSING AND TRANSMITTING SECRET DATA INFORMATION
US4965825A (en) * 1981-11-03 1990-10-23 The Personalized Mass Media Corporation Signal processing apparatus and methods
JPS58187015A (en) 1982-04-26 1983-11-01 Nippon Telegr & Teleph Corp <Ntt> Switched capacitor circuit
US4605921A (en) 1983-06-20 1986-08-12 Riddle Herbert S Digital word-framing technique and system
US4569052A (en) 1983-07-14 1986-02-04 Sperry Corporation Coset code generator for computer memory protection
US4759063A (en) 1983-08-22 1988-07-19 Chaum David L Blind signature systems
US4776011A (en) 1983-10-24 1988-10-04 Sony Corporation Recursive key schedule cryptographic system
US4605820A (en) 1983-11-10 1986-08-12 Visa U.S.A. Inc. Key management system for on-line communication
US4570084A (en) 1983-11-21 1986-02-11 International Business Machines Corporation Clocked differential cascode voltage switch logic systems
US4799258A (en) 1984-02-13 1989-01-17 National Research Development Corporation Apparatus and methods for granting access to computers
JPS60146361U (en) 1984-03-08 1985-09-28 三洋電機株式会社 Light emitting diode mounting device
NL8401989A (en) 1984-06-22 1986-01-16 Nederlanden Staat VIDEO ENTRY STATION WITH IMAGE LINE SCRAPE.
JPS61102167A (en) 1984-10-23 1986-05-20 Yokogawa Hokushin Electric Corp Dc/dc converter
US4661658A (en) 1985-02-12 1987-04-28 International Business Machines Corporation Offline PIN validation with DES
US4694491A (en) 1985-03-11 1987-09-15 General Instrument Corp. Cryptographic system using interchangeable key blocks and selectable key fragments
US4972472A (en) 1985-03-15 1990-11-20 Tandem Computers Incorporated Method and apparatus for changing the master key in a cryptographic system
US4686392A (en) 1985-10-30 1987-08-11 International Business Machines Corporation Multi-functional differential cascode voltage switch logic
JPH0532965Y2 (en) 1985-11-14 1993-08-23
GB8608172D0 (en) 1986-04-03 1986-05-08 Walker S M Computer security devices
JPS62166489U (en) 1986-04-11 1987-10-22
JPS62260406A (en) 1986-05-06 1987-11-12 Nec Corp White noise generator
FR2600183B1 (en) 1986-06-13 1990-10-12 Eurotechnique Sa INTEGRATED CIRCUIT FOR CONFIDENTIALLY STORING AND PROCESSING INFORMATION COMPRISING AN ANTI-FRAUD DEVICE
US4937866A (en) 1986-08-13 1990-06-26 U.S. Philips Corporation System for decoding transmitted scrambled signals
JPS6370550A (en) 1986-09-12 1988-03-30 Nec Corp Semiconductor integrated circuit
US5341423A (en) 1987-02-06 1994-08-23 General Electric Company Masked data transmission system
FR2617976B1 (en) 1987-07-10 1989-11-10 Thomson Semiconducteurs BINARY LOGIC LEVEL ELECTRIC DETECTOR
ES2040786T3 (en) 1987-08-14 1993-11-01 Siemens Aktiengesellschaft DATA TRANSMISSION PROCEDURE.
JPS6481087A (en) 1987-09-22 1989-03-27 Hitachi Maxell Ic card data transmission system
JPH01114995A (en) 1987-10-29 1989-05-08 Toppan Printing Co Ltd Identification card
JP2698588B2 (en) 1987-11-13 1998-01-19 株式会社東芝 Portable electronic devices
US5412379A (en) 1988-05-27 1995-05-02 Lectron Products, Inc. Rolling code for a keyless entry system
JPH022475A (en) 1988-06-15 1990-01-08 Omron Tateisi Electron Co Ic card
NO165698C (en) 1988-07-05 1991-03-20 System Sikkerhet As DIGITAL EQUIPMENT PROTECTION SYSTEM.
DE3825880C1 (en) 1988-07-29 1995-12-21 Siemens Ag Key device
GB8819767D0 (en) 1988-08-19 1989-07-05 Ncr Co Public key diversification method
US4932057A (en) 1988-10-17 1990-06-05 Grumman Aerospace Corporation Parallel transmission to mask data radiation
US4905176A (en) 1988-10-28 1990-02-27 International Business Machines Corporation Random number generator circuit
FR2638869B1 (en) 1988-11-10 1990-12-21 Sgs Thomson Microelectronics SECURITY DEVICE AGAINST UNAUTHORIZED DETECTION OF PROTECTED DATA
JPH02187888A (en) 1989-01-17 1990-07-24 Toshiba Corp Certification system
US5293029A (en) 1989-01-17 1994-03-08 Kabushiki Kaisha Toshiba System for mutually certifying an IC card and an IC card terminal
SE462935B (en) 1989-01-30 1990-09-17 Cominvest Res Ab KEEPING AND DEVICE PROVIDING EXTERNAL DETECTION OF SIGNAL INFORMATION
US5181243A (en) 1989-05-19 1993-01-19 Syntellect, Inc. System and method for communications security protection
NL8901323A (en) 1989-05-26 1990-12-17 Nederland Ptt METHOD FOR Ciphering a series of messages consisting of at least one symbol.
US5086467A (en) 1989-05-30 1992-02-04 Motorola, Inc. Dummy traffic generation
US5440749A (en) * 1989-08-03 1995-08-08 Nanotronics Corporation High performance, low cost microprocessor architecture
FR2651347A1 (en) 1989-08-22 1991-03-01 Trt Telecom Radio Electr SINGLE NUMBER GENERATION METHOD FOR MICROCIRCUIT BOARD AND APPLICATION TO COOPERATION OF THE BOARD WITH A HOST SYSTEM.
US5412730A (en) 1989-10-06 1995-05-02 Telequip Corporation Encrypted data transmission system employing means for randomly altering the encryption keys
US5136643A (en) 1989-10-13 1992-08-04 Fischer Addison M Public/key date-time notary facility
IT1238529B (en) 1989-11-10 1993-08-18 Data Protection Srl PROTECTIVE DEVICE FOR COMPUTERS AND SIMILAR, TO PREVENT THE CAPTURE, REGISTRATION AND UNDUE USE OF DATA FROM THE SAME DURING THEIR FUNCTIONING AND TO PROTECT THEM FROM TRANSITIONAL DISTURBANCES, WITH HIGH LEVEL ENERGY CONTENT, VERIFYING ON THE MAINS NETWORK POWER SUPPLY.
US5249294A (en) 1990-03-20 1993-09-28 General Instrument Corporation Determination of time of execution of predetermined data processing routing in relation to occurrence of prior externally observable event
GB2242797B (en) 1990-04-07 1993-12-08 Ferranti Int Plc Signal generation using digital-to-analogue conversion
US5177430A (en) 1990-04-19 1993-01-05 Moshe Mohel Circuit for securing a power supply
US5115133A (en) 1990-04-19 1992-05-19 Inomet, Inc. Testing of body fluid constituents through measuring light reflected from tympanic membrane
CA2044051A1 (en) 1990-06-29 1991-12-30 Paul C. Wade System and method for error detection and reducing simultaneous switching noise
US5136646A (en) 1991-03-08 1992-08-04 Bell Communications Research, Inc. Digital document time-stamping with catenate certificate
JPH0778975B2 (en) 1990-09-27 1995-08-23 インターナシヨナル・ビジネス・マシーンズ・コーポレーシヨン Optical disk drive
FR2667715A1 (en) 1990-10-09 1992-04-10 Gemplus Card Int METHOD AND DEVICE FOR INCREASING THE PROTECTION OF A MEMORY CARD.
US5144667A (en) 1990-12-20 1992-09-01 Delco Electronics Corporation Method of secure remote access
JPH04251186A (en) 1991-01-08 1992-09-07 Kobe Steel Ltd Microwave melting furnace for treating liquid
US5149992A (en) 1991-04-30 1992-09-22 The State Of Oregon Acting By And Through The State Board Of Higher Education On Behalf Of Oregon State University MOS folded source-coupled logic
US5241598A (en) 1991-05-22 1993-08-31 Ericsson Ge Mobile Communications, Inc. Rolling key resynchronization in cellular verification and validation system
SE500276C2 (en) 1991-06-24 1994-05-24 Shield Research In Sweden Ab Method and apparatus for preventing external detection of signal information
US5142578A (en) 1991-08-22 1992-08-25 International Business Machines Corporation Hybrid public key algorithm/data encryption algorithm key distribution method based on control vectors
US5159632A (en) 1991-09-17 1992-10-27 Next Computer, Inc. Method and apparatus for public key exchange in a cryptographic system
WO1993006695A1 (en) 1991-09-23 1993-04-01 Z-Microsystems Enhanced security system for computing devices
JP3083187B2 (en) 1991-09-30 2000-09-04 富士通株式会社 Key management method of electronic wallet system
EP0558133B1 (en) 1992-02-27 1997-06-18 Koninklijke Philips Electronics N.V. CMOS integrated circuit
JP2821306B2 (en) 1992-03-06 1998-11-05 三菱電機株式会社 Authentication method and system between IC card and terminal
IT1259383B (en) 1992-04-02 1996-03-12 PROTECTION DEVICE FOR COMPUTERS AND SIMILAR
US5600324A (en) 1992-05-11 1997-02-04 Rockwell International Corporation Keyless entry system using a rolling code
US5268962A (en) 1992-07-21 1993-12-07 Digital Equipment Corporation Computer network with modified host-to-host encryption keys
US5297201A (en) 1992-10-13 1994-03-22 J.D. Technologies, Inc. System for preventing remote detection of computer data from tempest signal emissions
JP3259428B2 (en) * 1993-03-24 2002-02-25 ソニー株式会社 Apparatus and method for concealing digital image signal
FR2704081B1 (en) 1993-04-16 1995-05-19 France Telecom Method for updating a memory card and memory card for implementing this method.
JPH08510365A (en) 1993-05-05 1996-10-29 リウ,ザンクアン Method and apparatus for data encryption
US5297207A (en) 1993-05-24 1994-03-22 Degele Steven T Machine generation of cryptographic keys by non-linear processes similar to processes normally associated with encryption of data
EP0701718A4 (en) * 1993-06-02 2000-03-29 Verifone Inc System and method for revaluation of stored tokens in ic cards
US5483598A (en) 1993-07-01 1996-01-09 Digital Equipment Corp., Patent Law Group Message encryption using a hash function
US5914471A (en) * 1993-07-20 1999-06-22 Koninklijke Ptt Nederland N.V. Method and apparatus for recording usage data of card operated devices
JP2750072B2 (en) 1993-07-27 1998-05-13 松下電工株式会社 Power converter
US5399996A (en) 1993-08-16 1995-03-21 At&T Global Information Solutions Company Circuit and method for minimizing electromagnetic emissions
US5510013A (en) * 1993-09-10 1996-04-23 Robert Bosch Gmbh Layer system for electrochemical probes
DE69312328T2 (en) * 1993-09-20 1998-01-08 Ibm SYSTEM AND METHOD FOR CHANGING THE KEY OR PASSWORD IN A COMMUNICATION NETWORK WITH KEY DISTRIBUTION
US5454037A (en) 1993-10-28 1995-09-26 Grayline International Limited Portable secure-telephone communications module
US5369706A (en) 1993-11-05 1994-11-29 United Technologies Automotive, Inc. Resynchronizing transmitters to receivers for secure vehicle entry using cryptography or rolling code
US5710834A (en) * 1995-05-08 1998-01-20 Digimarc Corporation Method and apparatus responsive to a code signal conveyed through a graphic image
US5515438A (en) * 1993-11-24 1996-05-07 International Business Machines Corporation Quantum key distribution using non-orthogonal macroscopic signals
FR2713419B1 (en) 1993-12-02 1996-07-05 Gemplus Card Int Method for generating DSA signatures with low cost portable devices.
EP0656708A1 (en) 1993-12-03 1995-06-07 International Business Machines Corporation System and method for the transmission and validation of an updated encryption key between two users
US5404402A (en) * 1993-12-21 1995-04-04 Gi Corporation Clock frequency modulation for secure microprocessors
US5491749A (en) 1993-12-30 1996-02-13 International Business Machines Corporation Method and apparatus for entity authentication and key distribution secure against off-line adversarial attacks
JP3029381B2 (en) 1994-01-10 2000-04-04 富士通株式会社 Data converter
US5434919A (en) 1994-01-11 1995-07-18 Chaum; David Compact endorsement signature systems
CA2176032A1 (en) 1994-01-13 1995-07-20 Bankers Trust Company Cryptographic system and method with key escrow feature
US5631492A (en) 1994-01-21 1997-05-20 Motorola Standard cell having a capacitor and a power supply capacitor for reducing noise and method of formation
US5668878A (en) 1994-02-28 1997-09-16 Brands; Stefanus Alfonsus Secure cryptographic methods for electronic transfer of information
US5412723A (en) 1994-03-01 1995-05-02 International Business Machines Corporation Mechanism for keeping a key secret from mobile eavesdroppers
US5420925A (en) 1994-03-03 1995-05-30 Lectron Products, Inc. Rolling code encryption process for remote keyless entry system
JPH07322602A (en) 1994-05-23 1995-12-08 Fujitsu Ltd Power supply device
US5551013A (en) 1994-06-03 1996-08-27 International Business Machines Corporation Multiprocessor for hardware emulation
US5414614A (en) 1994-06-06 1995-05-09 Motorola, Inc. Dynamically configurable switched capacitor power supply and method
EP0693836A1 (en) * 1994-06-10 1996-01-24 Sun Microsystems, Inc. Method and apparatus for a key-management scheme for internet protocols.
US5506905A (en) 1994-06-10 1996-04-09 Delco Electronics Corp. Authentication method for keyless entry system
IL110181A (en) * 1994-06-30 1998-02-08 Softchip Israel Ltd Microprocessor device and peripherals
US5778071A (en) 1994-07-12 1998-07-07 Information Resource Engineering, Inc. Pocket encrypting and authenticating communications device
US5546463A (en) 1994-07-12 1996-08-13 Information Resource Engineering, Inc. Pocket encrypting and authenticating communications device
US5606612A (en) * 1994-07-25 1997-02-25 General Instrument Corporation, Jerrold Communications Division Method and apparatus for television signal scrambling using a line expansion technique
US5511123A (en) 1994-08-04 1996-04-23 Northern Telecom Limited Symmetric cryptographic system for data encryption
US5557346A (en) 1994-08-11 1996-09-17 Trusted Information Systems, Inc. System and method for key escrow encryption
US5600273A (en) 1994-08-18 1997-02-04 Harris Corporation Constant delay logic circuits and methods
US5514982A (en) 1994-08-18 1996-05-07 Harris Corporation Low noise logic family
BE1008699A3 (en) * 1994-09-09 1996-07-02 Banksys Method and arrangement for selective give access to a security system.
US5663896A (en) 1994-09-22 1997-09-02 Intel Corporation Broadcast key distribution apparatus and method using Chinese Remainder
US5559887A (en) 1994-09-30 1996-09-24 Electronic Payment Service Collection of value from stored value systems
US5544086A (en) 1994-09-30 1996-08-06 Electronic Payment Services, Inc. Information consolidation within a transaction network
US5633930A (en) * 1994-09-30 1997-05-27 Electronic Payment Services, Inc. Common cryptographic key verification in a transaction network
US5636157A (en) * 1994-10-03 1997-06-03 International Business Machines Corporation Modular 64-bit integer adder
US5737419A (en) * 1994-11-09 1998-04-07 Bell Atlantic Network Services, Inc. Computer system for securing communications using split private key asymmetric cryptography
JP2825064B2 (en) * 1994-12-19 1998-11-18 株式会社日本自動車部品総合研究所 Encryption device
US5602917A (en) 1994-12-30 1997-02-11 Lucent Technologies Inc. Method for secure session key generation
CN1096186C (en) * 1994-12-30 2002-12-11 汤姆森消费电子有限公司 Modem with automatic callback provisions
US5625692A (en) 1995-01-23 1997-04-29 International Business Machines Corporation Method and system for a public key cryptosystem having proactive, robust, and recoverable distributed threshold secret sharing
US5892900A (en) * 1996-08-30 1999-04-06 Intertrust Technologies Corp. Systems and methods for secure transaction management and electronic rights protection
DE19505097C1 (en) 1995-02-15 1996-06-05 Siemens Ag Encryption device
US5483182A (en) 1995-03-06 1996-01-09 Motorola, Inc. Method and apparatus for a DC-DC converter an current limiting thereof
DE19511298B4 (en) 1995-03-28 2005-08-18 Deutsche Telekom Ag Procedure for issuing and revoking the authorization to receive broadcasts and decoders
IL113375A (en) 1995-04-13 1997-09-30 Fortress U & T Ltd Internationally regulated system for one to one cryptographic communications with national sovereignty without key escrow
US5796836A (en) 1995-04-17 1998-08-18 Secure Computing Corporation Scalable key agile cryptography
US5675645A (en) * 1995-04-18 1997-10-07 Ricoh Company, Ltd. Method and apparatus for securing executable programs against copying
US5638444A (en) * 1995-06-02 1997-06-10 Software Security, Inc. Secure computer communication method and system
US5613004A (en) * 1995-06-07 1997-03-18 The Dice Company Steganographic method and device
US5778074A (en) * 1995-06-29 1998-07-07 Teledyne Industries, Inc. Methods for generating variable S-boxes from arbitrary keys of arbitrary length including methods which allow rapid key changes
AU728942B2 (en) * 1995-06-30 2001-01-18 Canon Kabushiki Kaisha A communication apparatus and a communication system
US5727062A (en) * 1995-07-06 1998-03-10 Ritter; Terry F. Variable size block ciphers
US5812669A (en) 1995-07-19 1998-09-22 Jenkins; Lew Method and system for providing secure EDI over an open network
US5852290A (en) * 1995-08-04 1998-12-22 Thomson Consumer Electronics, Inc. Smart-card based access control system with improved security
FR2738971B1 (en) * 1995-09-19 1997-10-10 Schlumberger Ind Sa METHOD FOR DETERMINING AN ENCRYPTION KEY ASSOCIATED WITH AN INTEGRATED CIRCUIT
FR2738970B1 (en) * 1995-09-19 1997-10-10 Schlumberger Ind Sa METHOD FOR DETERMINING A DIVERSIFIED KEY ASSOCIATED WITH AN INTEGRATED CIRCUIT
JP3336826B2 (en) 1995-09-29 2002-10-21 株式会社デンソー Anti-theft device
FR2739469B1 (en) 1995-10-03 1997-12-26 Gemplus Card Int PROCESS FOR PUBLIC KEY CRYPTOGRAPHY BASED ON DISCRETE LOGARITHM
US6141652A (en) 1995-10-10 2000-10-31 British Telecommunications Public Limited Company Operating apparatus
NL1001659C2 (en) * 1995-11-15 1997-05-21 Nederland Ptt Method for writing down an electronic payment method.
US5727063A (en) * 1995-11-27 1998-03-10 Bell Communications Research, Inc. Pseudo-random generator
JPH09163469A (en) 1995-12-11 1997-06-20 Alpha Corp Device and method for remote control
US5754659A (en) 1995-12-22 1998-05-19 General Instrument Corporation Of Delaware Generation of cryptographic signatures using hash keys
US5838794A (en) 1996-01-11 1998-11-17 Teledyne Electronic Technologies Method and apparatus for inter-round mixing in iterated block substitution systems
US5637929A (en) 1996-01-16 1997-06-10 Ford Motor Company Method and apparatus for enhanced vehicle protection
JP3504050B2 (en) * 1996-01-26 2004-03-08 株式会社東芝 Power-residue calculation method and apparatus
US6453296B1 (en) 1996-01-31 2002-09-17 Canon Kabushiki Kaisha Electronic credit system and communication apparatus
FR2745135B1 (en) * 1996-02-15 1998-09-18 Cedric Colnot METHOD FOR AUTHORIZING ACCESS BY A SERVER TO A SERVICE FROM PORTABLE MEMORY CARD TYPE ELECTRONIC MICROCIRCUIT DEVICES
FR2745099B1 (en) * 1996-02-19 1998-03-27 Sgs Thomson Microelectronics METHOD FOR SEQUENCING AN INTEGRATED CIRCUIT
US5761306A (en) * 1996-02-22 1998-06-02 Visa International Service Association Key replacement in a public key cryptosystem
FR2745924B1 (en) 1996-03-07 1998-12-11 Bull Cp8 IMPROVED INTEGRATED CIRCUIT AND METHOD FOR USING SUCH AN INTEGRATED CIRCUIT
JP3525209B2 (en) 1996-04-05 2004-05-10 株式会社 沖マイクロデザイン Power-residue operation circuit, power-residue operation system, and operation method for power-residue operation
US5778069A (en) * 1996-04-10 1998-07-07 Microsoft Corporation Non-biased pseudo random number generator
US5835599A (en) 1996-04-15 1998-11-10 Vlsi Technology, Inc. Muti-cycle non-parallel data encryption engine
CA2177622A1 (en) * 1996-05-29 1997-11-30 Thierry Moreau Cryptographic data integrity apparatus and method based on pseudo-random bit generators
US5764766A (en) 1996-06-11 1998-06-09 Digital Equipment Corporation System and method for generation of one-time encryption keys for data communications and a computer program product for implementing the same
US5859548A (en) * 1996-07-24 1999-01-12 Lg Semicon Co., Ltd. Charge recycling differential logic (CRDL) circuit and devices using the same
US5745577A (en) * 1996-07-25 1998-04-28 Northern Telecom Limited Symmetric cryptographic system for data encryption
US5995917A (en) 1996-08-08 1999-11-30 Trimble Navigation Limited Optimal ring antenna determination system
JPH1084223A (en) 1996-09-10 1998-03-31 Mitsubishi Electric Corp Noise fm signal generation circuit
EP0831433A1 (en) * 1996-09-24 1998-03-25 Koninklijke KPN N.V. Method of making recoverable smart card transactions, a method of recovering such a transaction, as well as a smart card allowing recoverable transactions
WO1998018251A2 (en) 1996-10-23 1998-04-30 Philips Electronics N.V. Payment scheme for a mobile communication service
EP0840477B1 (en) 1996-10-31 2012-07-18 Panasonic Corporation Secret key transfer method which is highly secure and can restrict the damage caused when the secret key is leaked or decoded
GB9624127D0 (en) * 1996-11-20 1997-01-08 British Telecomm Transaction system
DE19649292A1 (en) * 1996-11-28 1998-06-04 Deutsche Telekom Ag Access protection method for pay television
JPH10171717A (en) 1996-12-05 1998-06-26 Matsushita Electric Ind Co Ltd Ic card and cipher communication system using the same
US5848159A (en) 1996-12-09 1998-12-08 Tandem Computers, Incorporated Public key cryptographic apparatus and method
US5821775A (en) 1996-12-27 1998-10-13 Intel Corporation Method and apparatus to interface monotonic and non-monotonic domino logic
JPH10197610A (en) 1996-12-31 1998-07-31 Sony Corp Noise generator and waveform generator employing it
US5892829A (en) * 1997-01-08 1999-04-06 Bell Communications Research, Inc. Method and apparatus for generating secure hash functions
US6587563B1 (en) 1997-02-15 2003-07-01 Apple Computer, Inc. Cryptographic system using chaotic dynamics
US6690795B1 (en) * 1997-03-04 2004-02-10 Lucent Technologies Inc. Multiple keys for decrypting data in restricted-access television system
US6049613A (en) * 1997-03-07 2000-04-11 Jakobsson; Markus Method and apparatus for encrypting, decrypting, and providing privacy for data values
US6069957A (en) 1997-03-07 2000-05-30 Lucent Technologies Inc. Method and apparatus for providing hierarchical key system in restricted-access television system
US5995624A (en) 1997-03-10 1999-11-30 The Pacid Group Bilateral authentication and information encryption token system and method
US6247129B1 (en) * 1997-03-12 2001-06-12 Visa International Service Association Secure electronic commerce employing integrated circuit cards
AUPO799197A0 (en) 1997-07-15 1997-08-07 Silverbrook Research Pty Ltd Image processing method and apparatus (ART01)
GB9707349D0 (en) * 1997-04-11 1997-05-28 Univ Waterloo A dynamic current mode logic family
US6193153B1 (en) * 1997-04-16 2001-02-27 Francis Lambert Method and apparatus for non-intrusive biometric capture
US6748410B1 (en) * 1997-05-04 2004-06-08 M-Systems Flash Disk Pioneers, Ltd. Apparatus and method for modular multiplication and exponentiation based on montgomery multiplication
ES2293677T3 (en) * 1997-05-04 2008-03-16 Sandisk Il Ltd IMPROVED APPLIANCE AND METHOD FOR MODULAR MULTIPLICATION AND EXPOSURE BASED ON MONTGOMERY MULTIPLICATION.
US5991415A (en) 1997-05-12 1999-11-23 Yeda Research And Development Co. Ltd. At The Weizmann Institute Of Science Method and apparatus for protecting public key schemes from timing and fault attacks
US5917754A (en) * 1997-05-21 1999-06-29 Atmel Corporation Semiconductor memory having a current balancing circuit
US5905399A (en) * 1997-06-30 1999-05-18 Sun Microsystems, Inc. CMOS integrated circuit regulator for reducing power supply noise
WO1999008411A2 (en) 1997-08-08 1999-02-18 Jonathan Stiebel New operation for key insertion with folding
US6003014A (en) * 1997-08-22 1999-12-14 Visa International Service Association Method and apparatus for acquiring access using a smart card
US6128391A (en) * 1997-09-22 2000-10-03 Visa International Service Association Method and apparatus for asymetric key management in a cryptographic system
US5999623A (en) 1997-11-05 1999-12-07 Globalstar L.P. Broadcast data access controller communication system
US6064740A (en) * 1997-11-12 2000-05-16 Curiger; Andreas Method and apparatus for masking modulo exponentiation calculations in an integrated circuit
US6345359B1 (en) * 1997-11-14 2002-02-05 Raytheon Company In-line decryption for protecting embedded software
US6041412A (en) * 1997-11-14 2000-03-21 Tl Technology Rerearch (M) Sdn. Bhd. Apparatus and method for providing access to secured data or area
US6090153A (en) 1997-12-05 2000-07-18 International Business Machines Corporation Multi-threshold-voltage differential cascode voltage switch (DCVS) circuits
US6046608A (en) * 1997-12-08 2000-04-04 Intel Corporation Differential precharge circuit
US6448981B1 (en) 1997-12-09 2002-09-10 International Business Machines Corporation Intermediate user-interface definition method and system
US6185685B1 (en) * 1997-12-11 2001-02-06 International Business Machines Corporation Security method and system for persistent storage and communications on computer network systems and computer network systems employing the same
US6046931A (en) * 1997-12-11 2000-04-04 Evsx, Inc. Method and apparatus for a RAM circuit having N-nary output interface
US6069497A (en) * 1997-12-11 2000-05-30 Evsx, Inc. Method and apparatus for a N-nary logic circuit using 1 of N signals
US6066965A (en) * 1997-12-11 2000-05-23 Evsx, Inc. Method and apparatus for a N-nary logic circuit using 1 of 4 signals
US6107835A (en) 1997-12-11 2000-08-22 Intrinsity, Inc. Method and apparatus for a logic circuit with constant power consumption
US6211456B1 (en) * 1997-12-11 2001-04-03 Intrinsity, Inc. Method and apparatus for routing 1 of 4 signals
JPH11191149A (en) * 1997-12-26 1999-07-13 Oki Electric Ind Co Ltd Lsi for ic card and using method therefor
US7587044B2 (en) * 1998-01-02 2009-09-08 Cryptography Research, Inc. Differential power analysis method and apparatus
US6304658B1 (en) 1998-01-02 2001-10-16 Cryptography Research, Inc. Leak-resistant cryptographic method and apparatus
US6226750B1 (en) 1998-01-20 2001-05-01 Proact Technologies Corp. Secure session tracking method and system for client-server environment
US6101477A (en) 1998-01-23 2000-08-08 American Express Travel Related Services Company, Inc. Methods and apparatus for a travel-related multi-function smartcard
US6041122A (en) * 1998-02-27 2000-03-21 Intel Corporation Method and apparatus for hiding crytographic keys utilizing autocorrelation timing encoding and computation
FR2776445A1 (en) 1998-03-17 1999-09-24 Schlumberger Ind Sa Cryptographic algorithm security technique
FR2776410B1 (en) 1998-03-20 2002-11-15 Gemplus Card Int DEVICES FOR MASKING THE OPERATIONS CARRIED OUT IN A MICROPROCESSOR CARD
US6336188B2 (en) * 1998-05-01 2002-01-01 Certicom Corp. Authenticated key agreement protocol
CN100530025C (en) 1998-05-29 2009-08-19 西门子公司 Method and device for processing data
EP1090480B1 (en) * 1998-06-03 2019-01-09 Cryptography Research, Inc. Improved des and other cryptographic processes with leak minimization for smartcards and other cryptosystems
EP1084548B1 (en) 1998-06-03 2008-12-17 Cryptography Research Inc. Secure modular exponentiation with leak minimization for smartcards and other cryptosystems
ATE548819T1 (en) * 1998-06-03 2012-03-15 Cryptography Res Inc SYMMETRIC CRYPTOGRAPHIC COMPUTING METHOD AND DEVICE FOR MINIMIZING LOSS IN CHIP CARDS AND OTHER ENCRYPTION SYSTEMS
ATE385089T1 (en) 1998-06-03 2008-02-15 Cryptography Res Inc USE OF UNPREDICTABLE INFORMATION TO MINIMIZE LEAKS OF CHIPCARDS AND OTHER CRYPTO SYSTEMS
US5998978A (en) * 1998-06-29 1999-12-07 Motorola, Inc. Apparatus and method for reducing energy fluctuations in a portable data device
US6075865A (en) 1998-07-01 2000-06-13 Tecsec Incorporated Cryptographic communication process and apparatus
US6539092B1 (en) * 1998-07-02 2003-03-25 Cryptography Research, Inc. Leak-resistant cryptographic indexed key update
US6816968B1 (en) * 1998-07-10 2004-11-09 Silverbrook Research Pty Ltd Consumable authentication protocol and system
US6343280B2 (en) * 1998-12-15 2002-01-29 Jonathan Clark Distributed execution software license server
FR2796738B1 (en) 1999-07-22 2001-09-14 Schlumberger Systems & Service SECURE MICRO-CONTROLLER AGAINST CURRENT ATTACKS
US6289455B1 (en) 1999-09-02 2001-09-11 Crypotography Research, Inc. Method and apparatus for preventing piracy of digital content
US6724894B1 (en) 1999-11-05 2004-04-20 Pitney Bowes Inc. Cryptographic device having reduced vulnerability to side-channel attack and method of operating same
CA2298990A1 (en) 2000-02-18 2001-08-18 Cloakware Corporation Method and system for resistance to power analysis
GB2371460B (en) * 2001-01-19 2004-12-22 Pixelfusion Ltd Computer graphics
FR2935078B1 (en) 2008-08-12 2012-11-16 Groupe Des Ecoles De Telecommunications Get Ecole Nationale Superieure Des Telecommunications Enst METHOD OF PROTECTING THE DECRYPTION OF CONFIGURATION FILES OF PROGRAMMABLE LOGIC CIRCUITS AND CIRCUIT USING THE METHOD
EP2507708B1 (en) 2009-12-04 2019-03-27 Cryptography Research, Inc. Verifiable, leak-resistant encryption and decryption

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5917909A (en) * 1992-12-23 1999-06-29 Gao Gesellschaft Fur Automation Und Organisation Mbh System for testing the authenticity of a data carrier
US5533123A (en) * 1994-06-28 1996-07-02 National Semiconductor Corporation Programmable distributed personal security

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022029443A1 (en) 2020-08-07 2022-02-10 Pugged Code Limited Method and apparatus for reducing the risk of successful side channel and fault injection attacks

Also Published As

Publication number Publication date
US8879724B2 (en) 2014-11-04
US20150052368A1 (en) 2015-02-19
US20100091982A1 (en) 2010-04-15
US7599488B2 (en) 2009-10-06
US7634083B2 (en) 2009-12-15
US7587044B2 (en) 2009-09-08
US20080022146A1 (en) 2008-01-24
US20080059826A1 (en) 2008-03-06
US20020124178A1 (en) 2002-09-05
US9419790B2 (en) 2016-08-16

Similar Documents

Publication Publication Date Title
US9419790B2 (en) Differential power analysis—resistant cryptographic processing
US6327661B1 (en) Using unpredictable information to minimize leakage from smartcards and other cryptosystems
US11861047B2 (en) Masked gate logic for resistance to power analysis
CA2333095C (en) Improved des and other cryptographic processes with leak minimization for smartcards and other cryptosystems
EP3220305B1 (en) Method of testing the resistance of a circuit to a side channel analysis of second order or more
Moore et al. Improving smart card security using self-timed circuits
US7941666B2 (en) Payment smart cards with hierarchical session key derivation providing security against differential power analysis and other attacks
US9571289B2 (en) Methods and systems for glitch-resistant cryptographic signing
US20070019805A1 (en) System employing systematic robust error detection coding to protect system element against errors with unknown probability distributions
Leng Smart card applications and security
EP1926241A2 (en) Using unpredictable information to minimize leakage from smartcards and other cryptosystems
EP1933496A2 (en) Improved DES and other cryptographic processes with leak minimization for smartcards and other cryptosystems

Legal Events

Date Code Title Description
AS Assignment

Owner name: CRYPTOGRAPHY RESEARCH, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KOCHER, PAUL C.;JAFFE, JOSHUA M.;JUN, BENJAMIN C.;REEL/FRAME:040731/0974

Effective date: 19990602

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION