US 7788312 B2
A method and apparatus processes signals in a set of analog circuit components of an analog circuit white enforcing a set of explicit constraints corresponding to a set of implicit constraints to reduce errors in output signals.
1. A method for processing signals in an analog circuit, comprising the steps of:
processing analog input signals using a set of analog circuit components; and
enforcing a set of explicit constraints corresponding to a set of implicit constraints to reduce errors in output signals.
2. The method of
3. The method of
4. The method of
5. The method of
6. The method of
7. The method of
8. The method of
9. The method of
10. The method of
11. The method of
12. The method of
13. The method of
14. The method of
15. The method of
16. The method of
17. The method of
18. The method of
19. The method of
20. The method of
21. The method of
22. The method of
23. The method of
24. The method of
25. The method of
26. The method of
27. The method of
28. The method of
29. The method of
30. The method of
31. The method of
32. The method of
33. The method of
34. The method of
35. The method of
36. The method of
37. The method of
38. The method of
39. The method of
40. The method of
41. The method of
42. The method of
43. The method of
44. The methods of
45. The method of
46. The method of
47. The method of
48. The method of
49. The method of
50. The method of
51. The method of
52. The method of
53. The method of
54. The method of
55. The method of
56. The method of
57. The method of
58. The method of
59. The method of
60. The method of
61. The method of
62. The method of
63. The method of
64. The method of
65. The method of
66. The method of
67. The method of
68. An apparatus configured to process signals, comprising:
a set of analog circuit components of an analog circuit configured to process analog input signals while enforcing a set of explicit constraints corresponding to a set of implicit constraints to reduce errors in output signals.
69. The apparatus of
70. The apparatus of
71. The apparatus of
72. The apparatus of
73. The apparatus of
74. The apparatus of
75. The apparatus of
76. The apparatus of
77. The apparatus of
78. The apparatus of
79. The apparatus of
80. The method of
81. The apparatus of
82. The apparatus of
83. The apparatus of
84. The apparatus of
85. The apparatus of
86. The apparatus of
87. The apparatus of
88. The apparatus of
89. The apparatus of
90. The apparatus of
91. The apparatus of
92. The apparatus of
93. The apparatus of
94. The apparatus of
95. The apparatus of
96. The apparatus of
97. The apparatus of
98. The apparatus of
99. The apparatus of
100. The apparatus of
101. The apparatus of
102. The apparatus of
103. The apparatus of
104. The apparatus of
105. The apparatus of
106. The method of
107. The apparatus of
108. The apparatus of
109. The apparatus of
110. The apparatus of
111. The apparatus of
112. The apparatus of
113. The apparatus of
114. The apparatus of
115. The apparatus of
116. The apparatus of
117. The apparatus of
118. The apparatus of
119. The apparatus of
120. The apparatus of
121. The apparatus of
122. The apparatus of
123. The apparatus of
124. The method of
125. The apparatus of
126. The apparatus of
127. The apparatus of
128. The apparatus of
129. The apparatus of
130. The apparatus of
131. The apparatus of
132. The apparatus of
133. The apparatus of
134. The apparatus of
135. The apparatus of
The present invention relates generally to a method and apparatus for processing analog signals in analog circuits, and more particularly to reducing the effect of errors while processing the analog signals.
There is an important difference between analog and digital circuits, especially when the circuits are used for complex computations and other complex processing. Analog circuits are primarily subject to the laws of physics, while digital circuits must obey the rules of logic. This has numerous implications.
At one time, all electrical devices, such telephones, radios, televisions used analog circuits. Even the first computers were analog. Generally, analog circuits are faster, less complex, use less power, and are smaller in size than equivalent digital circuits to perform similar processing tasks. But, perhaps most important, analog circuits can operate on analog values and analog states that represent, for example, real or complex numbers.
However, analog circuits have a major disadvantage when compared with digital circuits. Analog circuits are more prone to errors than, digital circuits. This is because analog circuits are relatively susceptible to noise, uncontrollable variations in fabrication processes, systematic or non-systematic faults, parasitic effects, defects, component mismatch, offsets, non-linearities, and sometimes hard to control environmental conditions. This makes it difficult to use analog circuits in the mass production of large complex systems, as demanded for modern electronic devices.
Consequently, digital circuits are now ubiquitous, filling an insatiable consumer market. Computers came first, later followed by digital telephones, television and radio, and innumerable other electronic digital devices. However, digital circuits also have disadvantages. When compared with analog circuits, digital circuits are relatively slow. But more important, digital circuits can only operate on discrete values, for example, 0 and 1. This requires that the input signals need first to sampled and quantized. Switching between permissible discrete states in digital circuits also takes more time than smoothly slewing analog states that have close together values in an analog circuit,
Herein, error correction and error reduction are distinguished. Error correcting codes and error correcting decoders work together to form a system for removing errors from data that has been corrupted by being sent across a noisy channel. The existence of error correcting codes was described by Claude Shannon in 1958 when he proved his well known channel capacity theorem. The first examples of error correcting codes were described by Hamming in 1960, Recently the field of coding was revolutionized when turbo codes and then low-density parity check (LDPC) codes were shown to both achieve very close to the Shannon channel capacity and to be decodable by relatively low-complexity error correction decoding algorithms. So called soft decoders are in fact these low-complexity error-correction decoders for turbo codes and LDPC codes. Soft decoders have been implemented with analog circuits, and have been shown to correct noise introduced by a noisy channel. For this to be possible, however, the data sent through the channel must first be encoded using an error correcting code. Those analog error correction decoders do not correct noise introduced by the circuit itself, they only correct errors due to channel noise. Error reduction is distinguished below.
Errors in Analog Circuits
Hans-Andreas Loeliger states that, “It is commonplace that analog circuits are sensitive to noise, temperature, and component variations, and are therefore hard to design and expensive to manufacture.” Hans-Andreas Loeliger, “Analog decoding and beyond,” Information Theory Workshop, 2001. Proceedings, ISBN: 0-7803-7119-4, 2001 IEEE, pp. 126-127, September 2001. He goes on to state that, “it has always been known that analog computation is sometimes much faster or less power consuming than digital computation,” and that “indeed, Carver Mead has demonstrated unconventional adaptive analog systems for a number of signal processing tasks (primarily in image processing) that share the robustness of digital systems but use several orders of magnitude less power,” and finally that “the most interesting modes of analog computation, may yet remain to be discovered.”
Analog Fast-Fourier Transform
Donald A. Gaubatz, in “FFT-Based Analog Beam-forming Processor” Ultrasonics Symposium Proceedings, pages 676-681, 1976, describes an example of processing with analog circuits. Gaubatz states that, “Analog signal processing requires stringent design constraints to assure accuracy and repeatability, but the resulting speed and relative economy are compensating factors.” Although Gaubatz describes a method for implementing a fast Fourier transform (FFT) with analog circuits, “the stringent design constraints” he employs to “assure accuracy” is to tediously selected discrete devices, each on its own die and in its own package, and to test each device to assure that the device matches the other devices before using the device in the FFT. Clearly, that technique is not amenable to mass production.
Transistor mismatch can be due to transistors being manufactured either too large or too small in either width or length, by variations in the distribution of dopant atoms from transistor to transistor, by variations in oxide thickness, or by other causes. Gaubatz proposes to use discrete transistors so that each, transistor can be tested individually. He discards individual devices that do not match one another sufficiently. However, analog circuits manually constructed from a large set of individual discrete components are not cost competitive with modern, very-large-scale digital integrated circuits.
It would be impossible to implement Gaubatz's analog FFT circuit in modern integrated circuits, because an integrated version, would not be able to employ his tedious method of only using matching devices, and discarding mismatched devices.
In successive generations of integrated circuits, as transistors have been manufactured in smaller and smaller sizes, the effect of mismatch has increased because improvements in manufacturing tolerances have not kept up with deceased device scaling. For this reason, engineers designing integrated circuits often select to use integrated devices that are much larger than a minimum feature size of the process to avoid the worst effects of mismatch. This increases cost and power, and makes the circuits slower. The end result is that analog circuits do not benefit from Moore's law scaling,
Analog designers can also use other methods to reduce mismatch errors. For example U.S. Pat. No. 4,386,155, “High-accuracy four-quadrant multiplier which also is capable of four-quadrant division” issued to Gilbert on Apr. 29, 1986, describes a four-quadrant analog multiplier. Gilbert describes a method for reducing errors introduced by temperature fluctuations in the circuit wherein “resistors are connected . . . and current which is proportional-to-absolute-temperature is caused to flow through the resistors. There, the resistors are laser-trimmed until VBE mismatch distortion is nulled.” That method requires a separate circuit to measure the amount of mismatch or error due to temperature fluctuation, and to compensate by supplying additional current. Furthermore, in that method, resistors must be matched by laser trimming. Such methods are not suited for mass production, or very large scale integrated systems.
The Gaubatz method for assuring repeatability, in the presence of noise, is again to use large discrete devices operating at relatively large voltages, so that the average noise voltage in the circuit is small with respect to the overall voltage swing of the devices. As semiconductor fabrication processes improve and the size of transistors decreases, the supply voltage VDD also decreases. This means that the available voltage swing of devices decreases. However, the average noise-power does not decrease significantly. If a very small, low-power, integrated version of Gaubatz's circuits were manufactured in the attempt to be competitive against the power and area consumption of a digital circuit, the noise would be extremely disruptive to the processing because the average noise voltage would be equal to a significant percentage of the total voltage swing in the circuits.
U.S. Pat. No. 5,495,554, “Analog Wavelet Transform Circuitry,” issued to R. Timothy Edwards and Michael D on Feb. 27, 1996, describes “an analog circuit implementing a continuous wavelet transform.” That analog circuit is estimated to be about one-hundredth ( 1/100) the size and power of a digital wavelet transform circuit. Edwards et al. state that “the analog wavelet outputs of the analog wavelet transform chip is directly determined without the loss of information due to the digital sampling.” They do not describe a method for reducing the effects of errors introduced in their analog processing.
U.S. Pat. No. 6,954,423, “Analog Implementation of Linear Transforms,” issued to Frank A. Tinker on Oct. 11, 2005, describes a system that performs “a linear transformation of a data set of discrete values . . . provided as a set of analog signals to the input nodes.” The transform “is achieved by judiciously adjusting the signal amplitude produced at the output of the phase-shift components and summing the resulting output signals as required to simulate the transformation of interest.” They do not describe a method for reducing the effects of errors introduced in their analog processing.
The processes that can be used with the embodiments of the invention can include linear transforms, linearized transforms, unitary transforms, statistical inferences, normalized belief propagations, solving linear differential equations, solving linearized differential equations, matrix inverses, minimizing functions, or other functions that obey any conservation or scaling law.
Analog Processing of Error Correction Decoding for Communications
Like the FFT, error-correction decoding is an important and computationally intensive processing task performed by communication transceivers. Analog circuits for decoding error-correcting codes are known. In contrast to analog Viterbi decoders, the decoders are based on turbo codes described by factor graphs.
U.S. Pat. No. 7,071,846 “Analog Decoding Method and Decoder” issued to Moerz on Jul. 4, 2006, describes “an analog decoder,” Moerz applies to parallelization of decoding of convolutional turbo codes. Error reduction is not described.
U.S. Pat. No. 6,282,559 “Method and electronic circuit for signal processing, in particular for the computation of probability distributions” issued to Helfenstein et al. on Aug. 28, 2001, describes a circuit module wherein “the currents of the outputs correspond to the product of the currents through the individual inputs. By combining the outputs, sum products can be calculated, especially for processing discrete probability distributions. The combination of several circuit modules allows to solve complex signal processing tasks.” They do not describe a method for reducing the effects of errors introduced in their analog processing. The error correction decoder they describe only corrects the effects of errors introduced in the communication channel before the noisy signal enters the error correction decoder system. Noise and other errors introduced by the decoder itself are not addressed.
Although the circuits according to Loeliger et al. can perform error correction decoding of data received over a noisy channel, their circuits still suffer from internal errors introduced while the processing, see Felix Lustenberger and Hans-Andrea Loeliger, “On Mismatch Errors In Analog-VLSI Error Correcting Decoders,” Proceedings of ISCAS, May 2001, They analyze their “new type of nonlinear analog transistor networks . . . proposed for ‘turbo’ decoding of error correcting codes.” They state that, “the influence of various non-idealities on the performance of such analog decoders is not yet well understood.” They describe “the performance degradation due to transistor mismatch.” They “assume . . . that each transistor in the circuit is affected by transistor mismatch,” and they “compare the accuracy of analog decoders with that of digital decoders,” Again, Loeliger et al. only analyze the effect of mismatch, in the operation of the decoder circuit. They do not describe any method or apparatus to remedy the effects of mismatch in their circuits.
Loeliger et al. only analyze how to derive output-referred errors due to transistor mismatch in analog translinear circuits. Their circuits are capable of processing two probability distributions as inputs to produce a third probability distribution as an output. All inputs and outputs are discrete probability distributions, such as are commonly found in a histogram. In a probability distribution for a discrete stochastic variable, each possible discrete state of the variable is assigned a probability such that a sum of the probabilities that the variable is in any of its possible states is 100%. They represent the “analog” probability of a given discrete state of a variable as an analog current on a wire. Each discrete state that a variable may occupy is signaled on an associated wire. For example, for a binary variable that can be either zero or one, they use two wires where one wire signals the probability that the variable is a one, and the other wire signals the probability that the variable is a zero. Because that system uses wires carrying analog values, and devices that directly process these analog values, it is an analog circuit operating on discrete variables and states (0 or 1).
They only describe circuits with exponential-components stating that, “the transistors will be modeled as exponentially behaving voltage controlled current sources.” As described above, the use of exponential-components requires that the circuit be composed of either a set of MOSFETs operating in the below-threshold mode, or BJTs.
There are problems with those restrictions because MOSFETs operating in the below-threshold mode have the disadvantage of being very slow. Generally, MOSFETS cannot operate faster than a few hundred kHz, and usually only operate in the tens of kHz range. MOSFETs certainly cannot attain the more than GHz speeds achieved by above-threshold MOSFET devices employed in conventional digital processors. To make up for the slow speed, one can sometimes use more MOSFETs to operate in parallel. However, then leakage currents become an additional cause for errors, and the circuit size increases.
BJTs require a more complicated and more expensive manufacturing process than MOSFETs. Generally, BJT require quite a large amount of power and are also bigger than MOSFETs. Thus, BJTs require more semiconductor area. Primarily because of their much greater cost and manufacturing difficulty, BJTs are used infrequently in large-scale applications.
Errors in Digital Circuits
As stated above, digital circuits operate exclusively on discrete values and discrete states. Most often, 0 and 1. This makes it relatively easy to detect and correct errors, when compared with analog circuits. There are two basic methods for correcting errors introduced in digital circuits by any cause.
The first method is to use only discrete or “digital” states to represent information in the processing. In a binary digital circuit, a state must be a either zero or one in order to be considered a valid state. Comparators or comparator-like components in digital circuits force any state that is found to be in between zero and one to be made into a zero or one.
For example, if in a digital circuit ground GND=0, and a supply voltage VDD=1V, then a state=0.6V cannot propagate through the circuit. Because the state is greater than 0.5V, the state is forced to VDD by the digital circuits, or failing this, the state may simply be considered invalid and the entire computation can fail. Errors in a digital, circuit cannot cause states to be greater than one, because a one is the same as VDD and a state cannot achieve a voltage greater than VDD. Similarly, a state cannot achieve a voltage less than GND, so states cannot be less than zero. In this way, each individual bit in a digital computing circuit is always forced to be either VDD or GND, i.e., 1 or 0 respectively.
The second method for error correction, in digital processing due to noise deviations is to use error-correcting codes (ECC).
That method has always required the use of digital states. When using digital (binary) states or bits, the system can then make a copy of these bits or add parity check bits. Then, at some later time, the system can take advantage of the extra parity bits to detect and even correct errors that have been introduced.
Again, consider the digital circuit, which uses GND=0V and VDD=1V. Strong noise, mismatch, or other kinds of defects can invert a zero state to a one state, or vice versa. For example, a state that should be at GND could end up at 0.6V because of noise. Then, the digital circuit forces the state to VDD because the state is greater than 0.5V. If bits are inverted in a computation, this can result in wrong answers or even in the digital computer failing to complete its task.
Error correcting codes have been used to successfully correct errors due to noise in channels. When using error-correcting codes, redundant bits are sent. That decreases throughput. This slow-down due to the redundant bits results in a lower channel capacity, a maximum, rate at which information can be sent across the channel.
Methods have been described for treating noise that affects bits in computing circuitry as if the noise were noise on the actual bits traversing the channel. In contrast to noisy channels, noise in digital computing circuits has somewhat worse consequences. Not only does the need to error correct in digital circuits result in a similar slowdown in the rate of useful computation, but it also requires extra circuitry to implement the error-correction functionality.
For example, as shown in
More complex techniques use recursive redundancy, block codes, or Reed-Solomon codes, or other kinds of more sophisticated codes. All of those techniques are essentially nothing more than complicated ways to structure redundant logic, and eventually “count votes.” The additional overhead for applying error-correction codes to digital computing circuits has meant that those techniques tend only to be used in mission critical circuits, where the additional expenditure of area and power is necessary.
U.S. Pat. No. 7,006,267 “Techniques for high fidelity quantum teleportation and computing” issued to Franson, et al. on Feb. 28, 2006 describes a method for using ancilla photons to assure high fidelity quantum teleportation. Ancilla photons are used to provide extra discrete states that function analogous to that of parity bits in error-correction codes. The additional states carry redundant information so that errors can be corrected,
Quantum error correction is known in the art, see for example, Seth Lloyd and Jean-Jacques Slotine, “Analog Quantum Error Correction,” Physical Review Letters 80, 4088-4091, Issue 18, May 1998, They describe an idea for error correction on analog variables, but only for quantum-mechanical analog variables, that are quantum entangled. The mathematics of quantum mechanics are quite different from that of classical physical systems. The idea for quantum analog error correction described requires quantum entanglement and quantum measurement to be available in order to be implemented. They state clearly that their idea may only be a theoretical curiosity and they do not propose a practical system for implementing the idea. Furthermore their idea requires quantum an cilia bits, which are essentially discrete parity bits for a quantum system.
U.S. Pat. No. 7,131,054, “Apparatus and method for efficient decoder normalization” issued to Greenberg et al. on Oct. 31, 2006 describes an apparatus and method “for normalizing a set of state metric values stored in a set of accumulators,” They describe a method for performing normalization comprising: if a specified normalization condition is met, subtracting a normalization amount from a branch metric value. That system uses an “accumulator in each . . . unit [which] has a fixed precision. Therefore, all accumulators are normalized periodically to prevent overflow. They describe a method for performing normalization only when a “specified normalization condition is met.” The normalization condition occurs when the accumulators are close to overflow.
Avoiding overflow is a completely different goal than reducing errors. The use of the term normalization is misleading, because they use normalization to mean reducing the magnitude of a variable.
When processing according to probability distributions, it is frequently necessary to perform normalization in order to assure that intermediate or final probability distributions are obey the rules for a properly defined probability distribution, namely that the total chance that all possible events occur cannot exceed 100%. This normalization is necessary even if the system has perfectly error-free computing hardware.
It is desired to provide an analog circuit that is a substantially error free.
A method and apparatus processes signals in a set of circuit components of an analog circuit while enforcing a set of explicit constraints corresponding to a set of implicit constraints to reduce errors in output signals.
In contrast to the prior art error correction, the invention performs error reduction. The error reduction techniques according to the embodiments of the invention are very different from convention error correcting codes because the techniques work on analog mined variables and can therefore be implemented in analog circuits, whereas all known error-correcting codes only work on discrete states. The method according to the invention can be applied to analog circuits performing a wide range of processing tasks, including but not limited to soft error-correction decoding.
The embodiments of the invention can be applied to MOSFETs operating in any mode, cut-off or below-threshold mode, triode or linear mode, and saturation or above-threshold mode, not just in sub-threshold as in the prior art, as well to any other kind of transistor such as BJTs, JFETs, and HEMTs. The invention can also be worked with analog circuits based on molectronics, spintronics, quantum dots, carbon nanostructures, biological structures, as known in the art.
It is a goal of the invention to substantially reduce errors in the processing caused by noise, uncontrollable variations in fabrication processes, systematic or non-systematic faults, parasitic effects, defects, component mismatch, offsets, current leakage, non-linearities, and sometimes hard to control environmental conditions. This makes it difficult to use analog circuits in the mass production of large complex systems, as demanded for modern electronic devices, or other sources of error from the analog circuits themselves, or from any other source, and even to enable asymptotically error-free processing using analog circuits.
Apparatus and Method Overview
As shown in
This remarkable result is achieved by enforcing a set of analog constraints 104 while processing the analog signals. A wide variety of analog processes and constraint are described herein to illustrate the scope of my invention.
As defined and used herein, the analog input signals 101 take on a continuous range of values, either according to current, voltage, or phase. Thus, the analog signals processed by my invention can truly represent analog-valued numbers. Thus, as defined herein, when I use the term “variable,” the variable can be an analog signal, or a real number or a complex number represented by the signal, a vector or matrix of real or complex numbers, or any other analog values, or a set of variables. As used herein conventionally, a set or a subset thereof can have one or more members. These analog-valued signals may be clocked discrete time signals, or smoothly varying analog-time signals.
It should be noted, that my invention can also be applied to analog charges, energies and magnetic spins, as described below.
This is in contrast with digital circuits, where the digital signals must have discrete values, e.g., ground (GND) typically represents logical 0, and some supply voltage VDD, represents logical 1. Obviously, a digital circuit can only process an approximation of a analog-valued signal.
The processes that can be used with my invention can include linear transforms, linearized transforms, unitary transforms, statistical inference, normalized belief propagation, solving linear differential equations, solving linearized differential equations, solving linearized partial differential equations, matrix inverses, minimizing functions, or other functions that obey any conservation or scaling law.
In addition, the embodiments of my invention can reduce errors in processes where an energy or magnitude is preserved, such as in a Fourier transform (FT), wavelet transforms, fast (FFT), convolution, filtering, correlation, any unitary transform, or any function, including nonlinear functions, that can be embedded in or otherwise posed as a unitary transformation or linear transform.
To the best of my knowledge constraints for reducing errors are not used in any prior art analog computations for any of the above functions. I do not consider error reduction in finite state machines, elementary cellular automata, or digital convolutional encoders and decoders. I am not aware of any technique that can use my method of reducing errors by enforcing explicit constraints on analog circuits can be applied to inherently discrete-state systems or methods.
Some processes have implicit constraints on what constitutes a valid state. For example, for unitary transforms, a magnitude of an input vector must be equal to a magnitude of an output vector.
The embodiments of my invention use these implicit constraints to define explicit constraints that are enforced while processing the analog signals. By analog signals, I mean the analog input signals, the analog output signals, or any intermediate analog signals, or combinations thereof, processed in the analog circuit.
When the processes are implemented using analog circuits according to embodiments of my invention, these constraints can be enforced very elegantly as part of the analog circuit operation, or by efficiently adding a small amount of additional circuitry.
One goal of my invention is to substantially reduce errors in the processing caused by noise, uncontrollable variations in fabrication processes, systematic or non-systematic faults, parasitic effects, defects, component mismatch, offsets, current leakage, non-linearities, and sometimes hard to control environmental conditions. This makes it difficult to use analog circuits in the mass production of large complex systems, as demanded for modern electronic devices, or other sources of error from the analog circuits themselves, or from any other source, and even to enable asymptotically error-free processing using analog circuits.
To help understand how enforcing explicit constraints can substantially reduce errors while processing analog signals in analog circuits, it is helpful to understand how error-correction codes can be used to reduce errors in digital circuits. Parity check constraints in the digital circuits typically involve a sum over a group of bits modulo 2.
For example, a discrete variable is expressed as two data bits x1 and x2, and a third parity bit x3. Then, a digital error-correcting code can specify the discrete constraint
If noise inverts one or all three of these bits, then the constraint is no longer true, so these errors can be detected. With a sufficient number of parity bits, and more complex correction apparatus, some errors can be detected or corrected.
The embodiments of my invention use analog-valued real numbers or analog-valued complex numbers, rather than discrete or binary values. For example, if there are three analog-valued variables, an example analog constraint according to an embodiment of my invention can take the form
It should be noted that ray summation constraint is distinguished from the normalization according to Green berg et al, distinguished above. My constraint is applied to a summation of variables and not a normalization by reducing the magnitude of a variable.
Parseval Constraint and FFT Processing
A unitary transform should not change a length of an input vector, the transform can only rotate the vector. In a FFT, the input vector and the output vector must obey Parseval's theorem. Parseval's theorem requires that the result of a Fourier transform is unitary. In other words, the sum (or integral) of the square of an input function is equal to twice the sum (or integral) of the square of its output transform:
As shown in
Parseval Constraint and FFT Butterfly
To provide error reduction, an output vector can be forced to have the same complex magnitude as the input vector as shown in
Similar constraints on the input and output magnitude can be applied to any unitary transform, such as unitary matrix multiplication, filtering, convolution, correlation, FFT, Fourier transform, wavelet transform, filtering, other kernel transforms or convolutions, as well as any other operator that can be embedded in a unitary transform.
By restoring a set of variables to a valid state according to the set of constraints as described herein, errors produced while processing in the analog circuit are substantially reduced. In many of the embodiments of my invention, the apparatus or method for enforcing the constraints to reduce errors requires much less overhead than conventional digital circuits with digital error correction. Furthermore, the invention exploits analog-valued resources, which of course is impossible with conventional digital circuits.
In general, my constraints can be applied to a set of variable as shown in
As shown in
By restoring a set of variables to a valid state according to the constraint as described herein, errors are reduced, just as they are in conventional digital circuit. However, as stated above, analog processing has a number of advantages over conventional digital processing.
In this embodiment of my invention, the apparatus for enforcing the constraints to reduce errors requires much less overhead than conventional digital circuits that detect and correct errors. Furthermore, this embodiment exploits analog-valued resources, unlike conventional digital circuits operating only on discrete resources.
The embodiments of my invention can apply to MOSFETs operating in any mode, cut-off or below-threshold mode, triode or linear mode, and saturation or above-threshold mode, not just in sub-threshold as in the prior art, as well to any other kind of transistor such as BJTs, JFETs, and HEMTs in any of the above modes. The invention can also be worked with analog circuits based on molectronics, spintronics, quantum dots, carbon nanostructures, biological structures, as known in the art.
For example, if voltage V1 represents variable x1, and V2 represents x2, and so forth, then by charging adjacent capacitive components with these voltages another adjacent capacitive component will assume an average of the voltages. This embodiment could apply for example to quantum dots, or quantum dot cellular automata, see
If voltages V1 through VN are connected to a single electrical node then these voltages must all be equal. In this embodiment Kirchhoff's voltage law (KVL), described in further detail below, enforces the equality constraint over the real variables.
In another embodiment, charges C1 through CN represent variables x1 through xN respectively, and the constraint C1+C2+ . . . +CN=C0 is enforced by limiting the total amount of charge C0 that is available to the circuit. This kind of embodiment applies to adiabatic computing circuits for example.
In another embodiment, spins S1 through S1 represent variables x1 through xN respectively, and E(S1) represents the energy of a given spin state relative to its magnetic environment. The constraint E(S1)+E(S2)+ . . . +E(SN)=E0 is enforced by limiting the total amount of energy that is available to the circuit. This kind of embodiment applies to computing with spintronics. This method of applying the law of conservation of energy to enforce error reducing constraints on analog variables can be applied to any application where a conserved energy or other conserved quantity is defined for every analog state.
Kirchhoff Law Constraints
Constraints based on the associative rule can be applied by applying a summation constraint using Kirchhoff s voltage law (KVL) or Kirchhoff's current law (KCL), on an ordering the subsets of analog states each time the constraint is applied, and then converting the current to voltage and using KVL to enforce equality. For example one current sum using KCL determines (I1+I2)+I3=C*VA where C is an arbitrary constant. Another constraint determines I1+(I2+I3)=C*VB. Then, the constraint VA=VB is enforced by KVL. Alternatively, the equality between two currents can be enforced by a current mirror, as known in the art. A current mirror is an adjustable current regulator that “copies” a current flowing through one device by controlling the current in another device. This constraints the output current to be constant regardless of the load. The current being “copied” can vary.
In another embodiment, ancilla variables are employed. Ancilla variables are variables that act as parity bits in an error correcting code. They do not carry actual data, but are present to act as a reservoir for entropy, e.g., noise, errors, etc. In one embodiment using ancilla variables, constraints are enforced over both some sets of variables and ancilla variables, and the ancilla variables are initialized at a known value. Subsequently, an external system continues to maintain the ancilla variables at a known value as they participate in the constraints on the set of variables. Ancilla variables as used here are not discrete quantum ancilla variables, but analog valued “parity” states.
Constraints and Analog Variables
I describe a wide variety of ways that constraints can be enforces according to the embodiments of my invention. In
As shown in
As shown in
As shown in
As shown in
As shown in
As shown in
As shown in
In one embodiment of the invention as shown in
As shown in
represents V2, and so forth to VN 1802, then by charging adjacent capacitive components 1803, 1804, 1805 with these voltages another adjacent capacitive component 1806 assumes an average of the voltages V4 1807. This embodiment of the invention could apply, for example, to quantum dots or quantum dot cellular automata.
Glossary of Terms
The following terms are used below are defined herein as below. Any of the analog circuit components described below alone, or in various combinations can be used in embodiments of the invention.
A bipolar-junction-transistor or BJT is a transistor with a transfer-function. Terminal 1 of a BJT is called the collector. Terminal 2 of a BJT is called an emitter. Terminal 3 of a BJT is called the base. The transfer-function of the BJT can be modeled most simplistically by the equation
IE is the emitter current,
A metal-oxide-semiconductor field-effect transistor or MOSFET is a transistor with terminal 1 called the drain. Terminal 2 of the MOSFET is called a source, and terminal 3 is called the gate. The MOSFET has an entirely different transfer-function depending on the settings of VDS and VGS.
A particular transfer-function of a MOSFET is called the operating-mode. The most important operating-modes of a MOSFET are called below-threshold-mode (VGS<VTH), linear-mode (VGS>VTH and VDS<VGS−VTH), and saturation-mode (VGS>VTH and VDS>VGS−VTH). The transfer-function of each, operating-mode is fundamentally different than the transfer-function for other operating-modes.
The transfer-function for a MOSFET in the below-threshold-mode (VGS<VTH) is similar to the transfer-function of the BJT. It is given by,
The transfer-function for a MOSFET in linear-mode is similar to the transfer-function of the BJT. It is given by,
The transfer-function for a MOSFET in saturation-mode is similar to the transfer-function of the BJT. This function is given by,
IDS is the current from the drain to the source,
L is the length of the MOSFET, and
COX is the gate capacitance per area set by the oxide thickness
Because the transfer-function and other aspects of operation of a circuit including MOSFETs and other circuit-components is determined by the transfer-functions, the design of a circuit assumes a given operating-mode for each of the MOSFETs. If the operating-mode of one or more MOSFET circuit-components in a circuit is changed, then the transfer-function of the circuit changes, and the circuit will almost always fail to produce the desired output for a given input. This require that the circuit is re-designed to achieve the desired transfer-function and operation using the new operating-mode or operating-modes.
Rearranging the transfer-function of the BJT yields
Rearranging the transfer-function of the MOSFET yields
Other field effect transistors are junction gate field-effect transistor (JFET), and high electron mobility transistor, or heterostructure (HFET).
A quadratic-component is a circuit or circuit-component that performs a transfer function given at least in part by a second-order polynomial and/or a square-root function.
An exponential-component is a circuit or circuit component that performs a transfer function given at least in part by an exponential or logarithmic function.
A saturation-mode-MOSFET-circuit is a circuit that includes at least one MOSFET serving as a quadratic-component. This also means that this MOSFET or MOSFETs is operating in the saturation-mode.
A below-threshold-MOSFET-circuit is a circuit that includes at least one MOSFET the serving as an exponential-component. This also means that this MOSFET or MOSFETs is operating in the below-threshold-mode.
A differential-pair includes two transistors where terminal 3 of one transistor is electrically-connected to terminal 3 of the other transistor.
A matched-transistor-set is a set of two or more transistors where terminal 3 of each transistor is electrically-connected to terminal 3 of all of the other transistors in the set.
Transistor-mismatch is a difference in the transfer function between two different transistors.
Although the invention has been described by way of examples of preferred embodiments, it is to be understood that various other adaptations and modifications may be made within the spirit and scope of the invention. Therefore, it is the object of the appended claims to cover all such variations and modifications as come within the true spirit and scope of the invention.