Search Images Maps Play YouTube News Gmail Drive More »
Sign in
Screen reader users: click this link for accessible mode. Accessible mode has the same essential features but works better with your reader.

Patents

  1. Advanced Patent Search
Publication numberUS20020035724 A1
Publication typeApplication
Application numberUS 09/912,216
Publication dateMar 21, 2002
Filing dateJul 24, 2001
Priority dateAug 23, 1994
Also published asCA2156679A1, CA2156679C, CN1099196C, CN1126412A, DE69524162D1, EP0701368A2, EP0701368A3, EP0701368B1, US6326999
Publication number09912216, 912216, US 2002/0035724 A1, US 2002/035724 A1, US 20020035724 A1, US 20020035724A1, US 2002035724 A1, US 2002035724A1, US-A1-20020035724, US-A1-2002035724, US2002/0035724A1, US2002/035724A1, US20020035724 A1, US20020035724A1, US2002035724 A1, US2002035724A1
InventorsAdrian Wise
Original AssigneeWise Adrian Philip
Export CitationBiBTeX, EndNote, RefMan
External Links: USPTO, USPTO Assignment, Espacenet
Data rate conversion
US 20020035724 A1
Abstract
A method for converting frame data at a slower rate into field data at a faster rate in a video decoder comprises determining a basic field repetition rate such that a field is repeated an integer number of times in a frame period, calculating a ratio differential of the repetition rate by subtracting from the speed-up ratio of the faster to the slower rate, the ratio of the fields per frame period to the slower rate, comparing the ratio differential with the differential of the field repetition rate and adding or subtracting extra fields when the two are substantially at variance.
Images(6)
Previous page
Next page
Claims(20)
What is claimed:
1. A method for use in converting input sets of data having an associated input rate into output sets of data, based on the input sets, for use at an associated output rate that differs from the input rate, comprising:
receiving at least one stored value indicative of the difference between the input rate and the output rate, performing at least one computation based on the at least one stored value, the result of the computation being used, in generating of the output sets of data, to control the insertion of additional sets of data or the deletion of sets of data.
2. The method of claim 1 in which the input sets and the output sets comprise video frames.
3. The method of claim 1 in which the input sets are based on compressed digital information.
4. The method of claim 3 in which the compressed digital information complies with a data compression standard for video information.
5. The method of claim 4 in which the standard comprises MPEG, JPEG, or H.261.
6. The method of claim 1 in which the output sets comprise video fields that satisfy a video broadcast standard.
7. The method of claim 6 in which the video broadcast standard comprises NTSC, SECAM, or PAL.
8. The method of claim 1 in which the input rate is one of 24, 25, or 30 frames per second.
9. The method of claim 1 in which the output rate is one of 50 or 60.
10. The method of claim 1 in which the at least one computation is based on Breshenham's algorithm.
11. The method of claim 1 in which the at least one stored value include a positive incremental value and a negative incremental value.
12. The method of claim 1 in which the at least one computation comprises a computation performed with respect to each of the input sets.
13. The method of claim 1 in which the insertion or deletion of sets of data occurs at regular intervals.
14. The method of claim 1 in which the ratio of the input rate to the output rate may be any arbitrary rate.
15. The method of claim 1 further comprising delivering the output sets of data at the output rate.
16. The method of claim 1 in which the additional sets of data that are inserted comprise copies of ones of the input sets of data.
17. A method comprising:
receiving a succession of compressed digital video input frames that comply with a digital video standard, the input frames having an associated input frame rate in accordance with the standard,
decompressing the input frames,
using the decompressed input frames to generate a succession of output video frames that comply with a broadcast video standard, the output frames having an associated output frame rate that is different from the input frame rate, and
in connection with the generation of the output video frames performing a computation that controls the insertion of additional frames into the output frames or the deletion of frames from the output frames to achieve the output frame rate.
18. A method comprising:
receiving MPEG video frames having an associated frame rate,
decompressing the MPEG frames,
based on the MPEG frames, generating a sequence of video frames for displaying at a frame rate different from the frame rate associated with the MPEG video frames, by performing a computation that controls insertion of additional frames into the display sequence.
19. A video decoder associated with a displaying device and comprising stored values indicative of a difference in data set rates between input video data sets received by the decoder and a video data set rate associated with the displaying device, and
a logic unit controlled to generate video data sets at the frame rate associated with the displaying device by inserting or deleting data sets based on computations performed using the stored values.
20. The video decoder of claim 19 in which the logic unit comprises hardware.
Description
CROSS-REFERENCE TO RELATED APPLICATIONS

[0001] This application is a continuation of application Ser. No. 08/516,038 filed Aug. 17, 1995. This application claims priority from British Application No. 9417138.6 filed Aug. 23, 1994.

TECHNICAL FIELD

[0002] This is invention relates to data rate conversion and particularly, though not exclusively, to video data frame/field rate conversion.

BACKGROUND OF THE INVENTION

[0003] There are various ways of compressing video information. In particular, there are three standards under which compression may be carried out: JPEG, MPEG AND H.261. These are discussed, for example, in U.S. Pat. No. 5,212,742.

[0004] Video information is commonly formatted as a series of fields. The original information which is to be converted into, and displayed in, a video format may not be immediately compatible with the field rate at which the information is to be displayed. For example, a celluloid film is shot at a rate of 24 frame/sec. (24 Hz) while, for example, the NTSC television system has a field rate of almost 60 Hz. The technique of increasing the frame rate of the film images to match that of the television system film rate is known as pulldown conversion.

[0005] Continuing with the above example of displaying a film in a NTSC standard format, a ‘⅔ pulldown’ conversion could be used in which each film frame is repeated for either two or three consecutive field periods at the video field repetition rate. The number of repetitions alternates so that the first frame is displayed twice in two consecutive field periods, the second frame is displayed three times in three consecutive field periods and so on. Thus, in one second twelve film frames at 24 Hz will each have been generated twice (i.e. for 24 field periods) while the other twelve film frames will each have been generated three times (i.e. for 36 field periods). The total (24+36) equals the 60 fields in one second at a 60 Hz field rate.

[0006] Pulldown instructions may be generated remotely and signalled to the video decoder associated with the displaying device or be generated locally at the video decoder. In the signalled pulldown, the encoder performs the pulldown calculations and signals specifically which frames are to be repeated, for example using the ‘repeat-first-field’ flag in the MPEG-2 video syntax. The decoder simply obeys the remotely generated instructions received.

[0007] In local pulldown, the encoder encodes the film information and transmits it to the receiving device. There is no information in the transmitted signal to tell the decoder at the receiving device how to perform the appropriate pulldown conversion (e.g. the ‘⅔ pulldown’ referred to above). The decoder must, therefore, calculate how to perform the appropriate conversion from the transmitted film frame rate to the displayed field rate.

[0008] If only pulldown conversion from the 24 Hz frame rate to a 60 Hz field rate were required, the single ⅔ pulldown conversion would be relatively easy to implement. However, other pulldown schemes are required. For example, the 24 Hz film frame rate may need to be converted to a 50 Hz field rate for the PAL television format.

[0009] Furthermore, an additional complexity in the NTSC television system is that the actual field rate is not 60 Hz but 60000/1001 Hz. Thus, the regular alternating ⅔ pulldown yields a field rate that is actually too high.

BRIEF DESCRIPTION OF DRAWINGS

[0010] The invention may be put into practice in a number of ways, one of which will now be described with reference to the accompanying drawings in which:

[0011]FIG. 1 illustrates Breshenham's line drawing algorithm;

[0012]FIG. 2 is a block diagram illustrating the data flow through a video decoder; and

[0013] FIGS. 3 to 6 and flow charts of various aspects of the invention.

DESCRIPTION OF INVENTION

[0014] According to the invention there is provided a method of converting frames of data received at a slower rate into fields of data generated at a faster rate, the method comprising:

[0015] determining a basic integer number of repetitions of fields in a frame period;

[0016] calculating a differential of the field repetition rate from the difference between the ratio of the faster to the slower rates and the ratio of the basic repetition number of fields in the frame period to the slower frame rate;

[0017] additionally repeating or deleting selected ones of the repeated fields, when the differential of the rate of repeating fields is substantially at variance with the calculated differential of the field repetition rate, to maintain the repetition of the fields at the faster rate.

[0018] The terms ‘frame’ and ‘field’ are used for convenience. Both are intended to refer to any frame, field, packet or other discreet quantity of data sent, received and/or constructed as a set. The invention allows the selected repetition rate to be modified by the inclusion or extraction of frames of repeated and, therefore, redundant data to fulfill the faster field data rates. Preferably, the selected basic integer repetition rate is less than the faster field rate. In this case, the method will add additionally repeated frames at the repetition rate. The repetition rate may be less than half the field rate.

[0019] The method does not have to select a slower basic integer repetition rate. In the alternative, it could equally well select a faster rate and then the method would be arranged to delete repeated frames where necessary.

[0020] The invention also extends to apparatus for converting frames of data received at a slower rate into fields of data generated at a faster rate, the apparatus comprising:

[0021] means for determining a basic integer number of repetitions of fields in a frame period;

[0022] means for calculating a differential of the field repetition rate from the difference between the ration of the faster to the slower rates and the ratio of the basic repetition number of fields in the frame period to the slower frame rate;

[0023] means for additionally repeating or deleting selected ones of the repeated fields, when the differential of the rate of repeating fields is substantially at variance with the calculated differentials of the field repetition rate, to maintain the repetition of the fields at the faster rate.

[0024] Preferably, the apparatus includes means for generating a repeat or delete frame signal for actuating the means for repeating or deleting selected ones of the repeated frames.

[0025] The present invention provides a generalized solution to the pulldown calculations that allow data at 23.98 Hz, 24 Hz and 25 Hz frame rates to be displayed at 50 Hz field rate and 23.98 Hz, 24 Hz, 25Hz and 29.97 Hz to be displayed at 59.94 Hz field rate.

[0026] Breshenham's line drawing algorithm is a method of drawing lines of arbitrary slope on display devices which are divided into a series of rectangular picture elements (pels). A description of Breshenham's algorithm can be found between pages 433 and 436 of the book ‘Fundamentals of Interactive Computer Graphics’ by Foley et al., published by Addison-Wesley.

[0027] In the line-drawing case illustrated in FIG. 1 (for lines that have a slope between 0 and 1) the algorithm approximates the desired line by deciding, for each co-ordinate in the X axis, which pel in the Y axis is closest to the line. This pel is illuminated or colored in as appropriate for the application.

[0028] As the algorithm moves from left to right in the diagram from say (n−1) to (n) it decides whether to select the pel in the same Y coordinate as for (n−1) or whether to increment the Y coordinate. In the diagram the Y coordinate is incremented at (n) and (n+2) but not at (n+1).

[0029] The decision of whether or not to increment the Y coordinate is used, in its application to the current invention, in the decision of whether to display the current frame for 3 field periods rather than the normal 2 field periods when deriving a faster field rate from an incoming frame rate in a video decoder.

[0030] In the simple case of conversion from 24 Hz frame rate to 60 Hz field rate, the desired speed-up ratio is {fraction (60/24)}. However, the important decision is made in determining whether or not a frame is displayed for three field periods (rather than two field periods) in a frame period. If there were no three field period frames then the 24 Hz frames rate would yield 48 fields. Thus, the ratio of the number of twice repeated fields can be subtracted from the speed-up ratio: 60 24 - 48 24 = 12 24 = 1 2 1

[0031] Plotting a line with slope will then allow us to calculate the pulldown pattern. Clearly, for a line of slope the Y coordinate is incremented once for every other step of the X coordinate. this is the expected result since we know that we display alternate film frames for 3 field times in order to perform {fraction (3/2)} pulldown.

[0032] Our U.S. patent application No. 9,415,413.5 filed on Jul. 29, 1994 entitled ‘Method and Apparatus for Video Decompression’ describes a multi-standard video decoder and is incorporated herein by reference. the present invention can be implemented in relation to this decoder receiving the MPEG-2 standard.

[0033] Referring to FIG. 2, in a preferred embodiment of the decoder described in the above patent application coded MPEG data (MPEG-1 or MPEG-2) is transferred into the device via a coded data input circuit 200. This data is then transferred via signals 202 to the Start Code Detector (SCD) 204. The SCD 204 recognizes a number of start codes which are unique patterns of bits, these are replaced by corresponding Tokens that may easily be recognized by subsequent circuitry. The remainder of the data (other than the start codes) is carried by a DATA Token. This stream of “start code” and DATA Tokens is transferred via signals 206 to formatting circuitry 208 that arranges the data into a suitable format for storage in external memory. This data is transferred via signals 210 to the Synchronous Dynamic Random Access Memory (SDRAM) interface circuitry 212.

[0034] The SDRAM interface circuitry 212 deals with a number of streams of data which are multiplexed over a single set of interface signals 230 in order that they may be written to or read from the external SDRAM device (or devices) 228. In each case data is temporarily stored in a swing buffer (214, 216, 218, 220, 222, and 224) each comprising two separate RAM arrays. Addresses for the SDRAM are generated by the address generator 330 and transferred via signals 332 to the SDRAM interface circuitry 212 where they are further processed by the DRAM interface controller 226 before being applied via the SDRAM interface 230 to the external SDRAM 228. The address generation is such that such that a coded data buffer 234 and a number of framestores 232 are maintained in the external SDRAM.

[0035] The formatted stream of “start code” tokens and DATA tokens mentioned previously is transferred to the SDRAM interface circuitry 212 via the signals 210 where it is stored temporarily in the swing buffer 214. This data is written into the area of the external SDRAM 228 that comprises a Coded Data Buffer (CDB) 234. This buffer has the function of a FIFO (First In, First Out) in that the order of the data is maintained. Data returning from the CDB 234 is stored temporarily in the swing buffer 216 before leaving the SDRAM interface circuitry via signals 236. The data on the signals 236 is the same as that on the signals 210, except that it has been delayed by a (variable) time in the CDB 234.

[0036] The data returning from the CDB is unformatted in the circuitry 238 which undoes the formatting, suitable for storage in the external SDRAM, previously performed by the formatter 208. It should however be noted that there is no restriction that the bus width of the signals 206 be the same as the signals 240. In the preferred embodiment a wider bus width is used by the signals 240 in order that a higher instantaneous data bandwidth may be supported at this point then by the signals 206.

[0037] The data (still comprising the “start code” tokens and the remainder of the data carried as DATA tokens) is passed via the signals 240 to the video parser circuitry 242. This circuitry as a whole has the task of further processing the coded video data. In particular the structure of the video data is “parsed” in order that its component parts are identified. The video parser comprises a Microprogrammed State Machine (MSM) 244 which has a stored program. Instructions are passed via signals 250 to a Huffman decoder 246. some parts of the instruction are interpreted by the Huffman decoder 246. The remainder of the instruction, together with the data produced by the Huffman decoder is transferred via signals 255 to an Arithmetic and Logic Unit (ALU) 248. Here again some parts of the instruction are interpreted by the ALU itself whilst the remainder of the instruction and the data produced by the ALU are transferred via signals 256 to a Token Formatter 258. The Huffman decoder 246 can signal error conditions to the MSM 244 via the signals 252. The ALU 248 may feedback condition-codes to the MSM 244 via signals 254. This enables the MSM to perform a “JUMP” instruction that is conditional on data being processed in the ALU 248. The ALU includes within it a register file in order that selected information may also be stored. The “start code⇄ tokens effectively announce the type of data (contained in the DATA tokens) that follow. This allows the MSM to decide which instruction sequence to follow to decode the data. In addition to this gross decision based on the Tokens derived from the start codes the finer structure of the video data is followed by the mechanism, previously described, of storing that information that defines the structure of the video data in the register file within the ALU and using this to perform conditional “JUMP” instructions depending on the value of the decoded data to choose alternative sequences of instructions to decode the precise sequence of symbols in the coded data.

[0038] The decoded data, together with the remaining instruction bits (that are not used by the Huffman Decoder) are passed via signals 256 to the Token Formatter 258. The data is formatted in response to the instruction bits, into Tokens, which can be recognized by subsequent processing stages. The resulting tokens are transferred to three separate destinations via the signals 260, 262, and 264. One stream of Tokens 262 passes to the Address Generator 330 where it is interpreted to generate suitable addresses to maintain the Coded Data Buffer and the framestores, as previously described. The second stream of Tokens 264 is interpreted by Video Timing Generation circuitry 326 in order to control certain aspects of the final display of decoded video information. A third stream of tokens 260 is passed to the Inverse Modeller 266 and on to subsequent processing circuitry. It should be understood that whilst each of the three streams of tokens (260, 262 and 264) is identical the information that is extracted is different in each case. Those Tokens that are irrelevant to the functioning of the specific circuitry are discarded. The tokens that are usefully interpreted in the cases of streams 262 and 264 are essentially control information while those usefully interpreted in the circuitry connected to the stream 260 may more usefully be considered as data.

[0039] The Inverse Modeller 266 has the task of expanding runs of zero coefficients present in the data so that the resulting data consists of blocks of data with precisely 64 coefficients, this is transferred via signals 268 to the Inverse Zig-Zag circuit 270. This circuit re-orders the stream of data according to one of two predefined patterns and results in data that might be considered two-dimensional. The Inverse Zig-Zag circuit includes a small Random Access Memory (RAM) 272 in which data is temporarily stored whilst being reordered. The resulting data is transferred via signals 274 to the Inverse Quantiser 276. Here the coefficients are unquantized and returned to their proper numerical value in preparation for an Inverse Discrete Cosine (DCT) function. The Inverse DCT is a separable transform so that it must be applied twice, once in a vertical direction and once in a horizontal direction. In this embodiment a single one dimensional Inverse DCT function is used twice to perform the full two dimensional transform. The data first enters an Inverse DCT circuit 280 via signals 278. The resulting data is transferred via signals 284 and stored in a Transpose RAM 282. The data is read out of the transpose RAM, but in a different order to that in which it was written in order that the data is transposed (i.e. rows and columns are swapped). This transposed data is transferred via signals 286 to the Inverse DCT 280 where it is processed a second time, the data resulting from this second transform being transferred via signals 288 to Field/Frame circuitry 290.

[0040] The Field/Frame circuitry again reorders data in certain cases such that the data that is transferred via signals 294 is in the same organization (Field or Frame) as that read as prediction data from the framestores in the external SDRAM. The Field/Frame circuitry 290 stores data temporarily in a RAM 292 for the purpose of this reordering.

[0041] Prediction data is read from the framestores that are maintained, as previously described, in the external SDRAM. Predictions are read via two paths (one nominally for “forward predictions” and the other nominally for “backwards predictions” although this is not strictly adhered to). One path comprises the swing buffer 222 and signals 296 whilst the other comprises the swing buffer 224 and signals 298. The data is filtered by the Prediction Filters 300 where the two predictions (“forward” and “backward”) may be averaged if required by the particular prediction mode indicated for that data. The resulting prediction is transferred via signals 302 to a prediction adder 304 where it is added to the data transferred from the Field/Frame circuitry via the signals 294. The resulting decoded picture information is written back into a third framestore via signals 306 and the swing buffer 220.

[0042] In order to produce a video signal the decoded information is read from the SDRAM via the swing buffer 218 and is then transferred via one of two signal paths. The chrominance data is transferred via signals 308 to a vertical upsampler 31 2 which up-samples the data so that there are the same number scan lines as used for the luminance signal. The vertical upsampler 312 stores one scan line of each of the two chrominance signals in the line store 314. The two resulting chrominance signals (the blue color difference signal and red color difference signal) are transferred via signals 31 6 and 318 to a Horizontal upsampler 320. The luminance signal (that did not require vertical upsampling) is also transferred via signals 310 to the horizontal upsampler. The horizontal upsampler 320 has the task of resampling the data by one of a number of preset scale factors to produce a suitable number of pels for the final scan line. The scale factor is selected via signals 324 which are provided by the Video Timing Generation (VTG) circuitry 326. This information is simply extracted from one of the Tokens supplied to the VTG via signals 264.

[0043] The data produced by the horizontal upsampler is transferred via signals 322 to an Output Multiplex 327. The multiplexes the actual video data signal arriving via the signals 322 with synchronization, blanking and border information produced internally by the output multiplex in response to timing signals 328 generated by the Video Timing Generator (VTG) circuitry 326. In order for the correct timing signals to be generated, particularly in the aspect of generating the correct amount of border information, the VTG uses information transferred in Tokens via the signals 264.

[0044] The final resulting video signal, together with a number of strobes, synchronization and blanking signals are transferred via signals 334 to a video output interface 336. The video signal may then be transferred to some suitable video display device.

[0045] A number of other interfaces are provided. A microprocessor interface 340 enables an external microprocessor to be connected to the signals 338. Signals 342 connect to many of the blocks of circuitry allowing the current status of the video decoding device to be read by the eternal microprocessor. In addition, certain features may be controlled by the external microprocessor writing to various control registers via this interface.

[0046] A JTAG (Joint Test Action Group) interface 346 allows various aspects of the device to be controlled via an external device connected to signals 344. The JTAG interface 346 is often used only for printed circuit board testing (after assembly) in which it is only necessary to control the external signals of the video decoding device. In this embodiment additional test capability is provided and for this reason the JTAG interface 346 is connected via signals 348 to all blocks of circuitry.

[0047] Circuitry 352 is provided for the generation and distribution of clock signals 354 from the external clock signals 350. This includes various Phase Locked Loops (PLLs) that enable higher speed internal clocks to be generated from external lower speed clocks.

[0048] In the context of the video decoder described in the above patent application the display rate is known because of a configuration pin (NTSC/PAL) which indicates whether a 59.94 Hz or 50 Hz display raster is being produced. The film frame rate is transmitted in the MPEG-2 video stream as the frame-rate parameter.

[0049] In the normal course of events any progressive frame will be displayed for two field times. A bit in the PICTURE-TYPE token controls the repeating of the first field to make the frame display for three field times.

TABLE 1
E 7 6 5 4 3 2 1 0
1 1 1 1 0 1 1 1 1
0 d x f p s s t t

[0050] The ‘f” bit is set to ‘1’ to repeat the first field. In the case that the sequence is interlaced, the ‘f” bit in this token directly matches the ‘repeat-first-field’ bit in the MPEG-2 sequence. (This is the signal pulldown case). However, in the case that ‘progressive-sequence’(as the term is defined in the MPEG standard) is ‘one’, indicating that the sequence is coded as a progressive sequence, local pulldown is enabled and the ‘f’ bit is calculated according to the algorithm described herein.

[0051] The algorithm is executed on the microprogrammable state machine (MSM) and is therefore specified in microcode (rather than the more familiar ‘C’ programme that illustrates the algorithm in the appendix hereto). The MSM is a 16-bit machine and this causes some minor complications because of the limited number range that can be represented in 16 bits.

[0052] This is dealt with by reducing the size of the denominator and numerator of the slope by common factors. In the example program given at the end of this document this is done by cancelling any common factor of 1001 and then dividing by 2 until either the numerator or the denominator is odd. Even this simple case yields numbers, dx and dy which will not exceed 16-bit number range as indicated by the ‘min’ and ‘max’ values shown in the results. In the said video decoder, the numbers dx and dy are precalculated and stored in tables that are indexed to determine the correct dx and dy values. As a result, the ratios can be further reduced by the smallest possible numerator and denominator as shown below:

TABLE 2
Display Field Rate (Hz)
Frame Rate 50 60000/1001
(Hz) Full Form Reduced Full Form Reduced
24000/1001 ( 50 1001 ) 24000 - 2 41 480 60000 1001 24000 1001 - 2 1 2
24 50 24 - 2 4 12 60000 1001 24 - 2 498 1001
25 50 25 - 2 0 60000 1001 25 - 2 398 1001
30000/1001 Not Supported 0
60000 1001 30000 1001 - 2

[0053] The variable “d” is a decision variable which is updated at each x coordinate (or each film frame). At each x coordinate the ideal value of y (represented by the line in FIG. 1) lies between two pels (one black and one white). d is proportional to the difference between the distance to the upper pel and the distance to the lower pel.

[0054] If d is negative than the ideal line lies closer to the lower pel.

[0055] If d is positive then the ideal line lies closer to the upper pel.

[0056] At each x coordinate the algorithm must choose either the lower or upper pel and then update the value of d in readiness for the next x coordinate (next frame).

[0057] If d is negative, the lower pel is chosen. d is updated by adding on incr1. Since incr1 is positive d will become less negative reflecting the fact that the line will now be farther from the lower pel (at the next x coordinate) in value.

[0058] If d is positive, the upper pel is chosen. d is updated by adding on incr2. Since incr2 is negative d will become less positive reflecting the fact that the line will now be farther from the upper pel (at the next x coordinate).

[0059] incr1 and incr2 therefore represent the change in d (i.e., the change in the difference between the distance from the ideal notional line to the upper pel and the distance from the line to the lower pel) for the two possible decisions that the algorithm may take. Thus, having chosen a basic integer value of repetitions of the field in a frame period the notional slope of the notional line is determined in accordance with equation 1 from which the algorithm is used to decide whether to add a field repeat in a frame period or not to maintain the running average rate of the repetiton of fields at the faster field rate.

[0060] In the examples in Table 2 the basic integer value of repetitions of the field in a frame period is conveniently chosen as 2. Because of this repetitions of fields have to be added according to the pulldown pattern to maintain the running average. However, a basic integer value resulting in an overall excess of repetitions of the fields could be chosen, such that selected field repetitions are deleted according to the pulldown pattern established.

[0061] It will also be noted that, as an alternative to storing dx and dy in the table and calculating incr1 and incr2 it will be equally valid to store precalculated values of incr1 and incr2 in the table directly.

[0062]FIG. 3 illustrates the procedure for decoding and displaying a field for an appropriate number of times.

[0063]FIG. 4 shows an example algorithm to determine dx and dy from field-rate and frame-rate. In this example, values of field-rate and frame-rate that are larger than 1000 are interpreted as representing a multiple of 1001, e.g. a frame-rate of 24000 actually represents a frame rate of 24000/1001 Hz.

[0064]FIG. 5 shows an algorithm to initialize incr1, incr2 and d. The algorithm is used before the first frame. The values of dx and dy are integers such that the fraction dy/dx represents the “slope of the line”.

[0065]FIG. 6 shows an algorithm to determine whether to display a frame for two or three field times. The algorithm is used once for each frame. The values of incr1 and incr2 are those determined by the initialization algorithm. The value of d is that produced by this algorithm for the previous frame or the initial is a time algorithm in the case of the first frame.

[0066] The principal advantages of this method are:

[0067] 1. All of the required pulldown conversions are performed using the same arithmetic.

[0068] 2. Once the parameters dx and dy are known (and these can be stored in the table) no multiplications or divisions are required.

[0069] 3. The algorithm works for arbitrarily long sequences of frames—none of the numbers grow indefinitely (which would eventually lead to number representation problems irrespective of the word width).

[0070] 4. The frames which have the repeated field are distributed evenly throughout the sequence of frames.

[0071] 5. Very little state needs to be maintained in order for the algorithm to operate. Just a current value of ‘d’ and probably INCR1 and INCR2 (although these could be recalculated or looked in a table each frame period).

[0072] The following procedure is Breshenham's line drawing algorithm.

void breshenham(x1, y1, x2, y2)
{
int dx, dy, incr1, incr2, d, x, y, xend;
dx = abs(x2-x1);
dy = abs(y2-y1);
d = 2 * (dy - dx);
incr1 = 2 * dy;
incr2 = d;
if (x1 > x2)
{
x = x2;
y = y2;
xend = x1;
}
else
{
x = x1;
y = y1;
xend = x2;
}
printf (“(%d, %d)\n”, x, y);
while (x < xend)
{
x ++;
if (d < 0)
d += incr1;
else
{
y ++;
d + = incr2;
}
printf(“(%d, %d)\n”, x, y);
}
}

[0073] The following program shows the modified algorithm (as the procedure three_fields()) to calculate which frames to display for three field-times. Each possible conversion is checked out by testing over one million film frames to ensure that the field rate does indeed approach the required value.

#include < compiler.h>
#include <pddtypes.h>
#include <stdlib.h>
#include <stdio.h>
Boolean three_fields(int dx, int dy, int *d, Boolean initalise)
{
int incr1, incr2, x, y, xend;
int r = False;
incr1 = 2 * dy;
incr2 = 2 * (dy - dx);
if (initalise)
{
*d = incr2;
}
else
{
if (*d < 0)
*d += incr1;
else
{
r = True;
*d += incr2;
}
}
return r;
}
double check_ratio(int dx, int dy, int limits[2])
{
int d;
int frame, field = 0, num_frames =10000000;
double ratio, field_rate;
int three = 0, two = 0;
(void) three_fields(dx, dy, &d, True );
limits[0] = limits[1] = 0;
if (d < limits[0]) limits[0] = d;
if (d > limits[1]) limits[1] = d;
for (frame = 0; frame < num_frames; frame ++)
{
if ( three_fields (dx, dy, &d, False ) )
{
field += 3;
three ++;
}
else
{
field += 2;
two ++;
}
if (d < limits[0]) limits[0] =d;
if (d > limits[1]) limits[1] =d;
}
ratio = ( (double)field) / ((double) frame);
return ratio;
}
static int frame_rates[ ] = /* input frame rates */
{
−1,
24000, /* numbers > 1000 express a numerator − denominator =
1001 */
24,
25,
30000,
30,
50,
60000,
60,
−1, −1, −1, −1, −1, −1, −1};
static int field_rates[ ] = /* output display rates */
{
50, 60000
};
double real_rate (int rate)
{
if (rate > 1000)
return ((double) rate) / 1001.0;
else
return ((double) rate);
}
void main (int argc, char **argv)
{
int dx, dy, field_index, frame_index;
int limits[2];
double ratio, field_rate;
for (field_index = 0; field_index < 2; field_index ++)
{
for (frame_index = 1; frame_rates[frame_index] > 0; frame_index
++)
{
if ( (real_rate(frame_rates[frame_index] ) *2.0) < =
real_rate(field_rates[field_index]) )
{
dy = field_rates[field_index];
dx = frame_rates[frame_index];
if ( (field_rates[field_index] < = 1000) ||
(frame_rates[frame_index] < = 1000) )
/* NB if both have the 1001 then don't bother! */
{
if (field_rates[field_index] > 1000)
dx * = 1001;
if (frame_rates[frame_index] > 1000)
dy * = 1001;
}
dy − = (2 * dx);
/* limit ratio by dividing by two */
while ( ( (dx & 1) = = 0) && ( (dy& 1) = = 0) )
{
dx >> = 1;
dy >> = 1:
}
ratio = check_ratio (dx, dy, limits);
printf(“output field rate = %d%s, input frame rate =
%d%s\n”,
field_rates[field_index],
( (field_rates[field_index]> 1000) ? “/1001”:“”),
frame_rates[frame_index],
( (frame_rates[frame_index]> 1000) ? “/1001”:“”) ):
printf(“dx = %d, dy = %d\n”, dx, dy);
field_rate = frame_rates[frame_index] * ratio;
if (frame_rates[frame_index]> 1000)
field_rate / = 1001.00;
printf (“ratio = %4.12g, field_rate = %4.12g\n”, ratio,
field_rate);
printf(“(field_rate = %4.12g/1001)\n”, field_rate * 1001);
printf(“min = %d, max = %d\n\n”, limits[0], limits[1]):
}
}
}
}
The program of the preceding pages yields the following output:
output field rate = 50, input frame rate = 24000/1001
dx = 12000, dy = 1025
ratio = 2.0854166, field_rate = 49.9999984016
(field_rate = 50049.9984/1001)
min = −21950, max = 2000
output field rate = 50, input frame rate = 24
dx = 12, dy = 1
ratio = 2.0833333, field_rate = 49.9999992
(field_rate = 50049.9991992/1001)
min = −22, max = 0
output field rate = 50, input frame rate = 25
dx = 25, dy = 0
ratio = 2, field_rate = 50
(field_rate = 50050/1001)
min = −50, max = 0
output field rate = 60000/1001, input frame rate = 24000/1001
dx = 750, dy = 375
ratio = 2.5, field_rate = 59.9400599401
(field_rate = 60000/1001)
min = −750, max = 0
output field rate = 60000/1001, input frame rate = 24
dx = 3003, dy = 1494
ratio = 2.4975024, field_rate = 59.9400576
(field_rate = 59999.9976576/1001)
min = −3018, max = 2982
output field rate = 60000/1001, input frame rate = 25
dx = 25025, dy = 9950
ratio = 2.3976023, field rate = 59.9400575
(field_rate = 59999.9975575/1001)
min = −30150, max = 19850
output field rate = 60000/1001, input frame rate = 30000/1001
dx = 1875, dy = 0
ratio = 2, field_rate = 59.9400599401
(field_rate = 60000/1001)
min = −3750, max = 0

Patent Citations
Cited PatentFiling datePublication dateApplicantTitle
US2151733May 4, 1936Mar 28, 1939American Box Board CoContainer
CH283612A * Title not available
FR1392029A * Title not available
FR2166276A1 * Title not available
GB533718A Title not available
Referenced by
Citing PatentFiling datePublication dateApplicantTitle
US7860321 *Sep 21, 2005Dec 28, 2010Canon Kabushiki KaishaImage pickup apparatus with function of rate conversion processing and control method therefor
US8184660 *Sep 15, 2005May 22, 2012Lsi CorporationTransparent methods for altering the video decoder frame-rate in a fixed-frame-rate audio-video multiplex structure
Classifications
U.S. Classification725/1, 375/E07.198, 348/E07.003, 375/E07.211, 375/E07.253
International ClassificationH04N5/253, H04N7/01, H04N7/46, H04N7/26, H04N7/50
Cooperative ClassificationH04N19/40, H04N7/01, H04N19/61, H04N19/587
European ClassificationH04N7/46T, H04N7/50, H04N7/26T, H04N7/01