|Publication number||US20090256855 A1|
|Application number||US 12/485,121|
|Publication date||Oct 15, 2009|
|Filing date||Jun 16, 2009|
|Priority date||May 12, 2006|
|Also published as||US7827491, US20070262995, US20110230232|
|Publication number||12485121, 485121, US 2009/0256855 A1, US 2009/256855 A1, US 20090256855 A1, US 20090256855A1, US 2009256855 A1, US 2009256855A1, US-A1-20090256855, US-A1-2009256855, US2009/0256855A1, US2009/256855A1, US20090256855 A1, US20090256855A1, US2009256855 A1, US2009256855A1|
|Inventors||Bao Q. Tran|
|Original Assignee||Tran Bao Q|
|Export Citation||BiBTeX, EndNote, RefMan|
|Non-Patent Citations (1), Referenced by (8), Classifications (8), Legal Events (1)|
|External Links: USPTO, USPTO Assignment, Espacenet|
This application is a continuation of application Ser. No. 11/433,897, filed May 12, 2006, the content of which is incorporated by reference.
The present invention relates to a high performance computer system.
In the past, the best way for computer makers to improve performance was to turn up the clock speed on the processor. As computers decrease in price and increase in processing power, sophisticated application such as CAD, financial modeling and software-based computer-implemented video editing systems are being purchased by more consumers. Such computer-implemented video editing systems manipulate video data in a digital format, such as the Motion Pictures Expert Group (MPEG) compression format MPEG-2.
A conventional user interface for controlling a video editing system permits a user to select a clip of the video segment being edited. The selected video clip may then be copied or cut and pasted into another video segment. A conventional user interface can include a timeline bar with a fixed length, and two markers or other indicia in the timeline bar to represent the beginning and end of the selected clip. In order to select the video clip from the source video, the user may need to play the source video, stop the video at the frame where the user wishes to start the clip, mark the starting frame, resume playing the video, stop the video again where the user wishes to end the clip, and mark the ending frame. Some users can find these conventional interfaces confusing, unintuitive, inefficient, time-consuming or difficult to master.
U.S. Pat. No. 6,597,375 discloses a user interface for selection of a video clip in a video editing system that includes a first control element, a first video window, a second control element, and a second video window. The first control element is positioned to select a starting frame of a video clip from a video data source. The first video window is operatively coupled to the first control element to display the starting frame of the video clip. The second control element is positioned to select an ending frame of the video clip. The second video window is operatively coupled to the second control element to display the ending frame of the video clip simultaneously with the display of the starting frame in the first video window. The first video window updates dynamically as the first control element is moved, and the second video window updates dynamically as the second control element is moved.
U.S. Pat. No. 6,725,354 discloses a microprocessor having a first processor core and a second processor core. The first core includes a first processing block. The first processing block includes an execution unit suitable for executing a first type of instruction. The second core includes a second processing block. The second processing block includes an execution unit suitable for executing an instruction if the instruction is of the first type. The processor further includes a shared execution unit. The first and second processor cores are adapted to forward an instruction to the shared execution unit for execution if the instruction is of a second type. In one embodiment, the first type of instruction includes fixed point instructions, load/store instructions, and branch instructions and the second type of instruction includes floating point instructions.
Systems and methods are disclosed method to increase processing power of a computer by generating a plurality of tasks for an application software; distributing the tasks over a plurality of general purpose processor cores and a graphics processing unit (GPU); and receiving the outputs from the cores and the GPU.
In one aspect, a process to edit video includes: receiving a source video; generating a plurality of tasks including one or more transitions, one or more effects, and one or more audio edit to be applied to the source video; assigning the tasks to a plurality of general purpose processor cores and a graphics processing unit (GPU); and receiving the outputs from the cores and the GPU; and rendering a video clip as an output.
Implementations of the above aspect may include one or more of the following. The processor can be 64-bit processor. The GPU comprises a graphics card. The system can distribute tasks among the plurality of general purpose processors and the GPU to accelerate the preview and rendering of motion, opacity, color, and image distortion effects. The video can be scaled to one of: DV, Digital Betacam, HDV, HDCAM, DVCPRO HD, and D5 HD. The system can capture and edit HDV or HDTV content. The system can play-back video in real-time. Multiple video channels can be played at full resolution and with titles, transitions, effects, and color correction. HDTV or HDV can be played in real time. Multiple video tracks from a multicam shoot can be viewed and the system can switch between tracks in real time. The system can perform color correction in real time. The system can apply primary and secondary color-correction tools to match shots; change colors; correct exposure errors; and modify highlights, midtones, and shadows across a clip, a sequence, or a project. The editing can include switching multiple nestable timeliness The system can generate 5.1 surround sound. The system can create and manipulate multichannel audio to produce surround sound. The system can generate interactive DVDs for digital dailies, test discs, or final delivery. The system can render transitions and effects using a first general purpose processor; render audio using a second general purpose processor; and render the video clip output using the GPU.
In another aspect, a system includes a plurality of general purpose processor cores; a graphics processing unit (GPU) coupled to the plurality of general purpose processor cores; and code executable by the processor cores and the GPU to receive a source video; process the source video with the plurality of general purpose processors and the GPU; and render a video output.
In implementations of the above aspect, the tasks are distributed among the plurality of general purpose processors and the GPU to accelerate the preview and rendering of motion, opacity, color, and image distortion effects. An HDV or HDTV camera can be connected to the processor cores.
In another aspect, a computer readable disk includes code for receiving and editing HDV or HDTV content; specifying tasks to be applied to the content including transitions, effects, and audio; assigning the tasks to a plurality of general purpose processor cores and a graphics processing unit (GPU) to accelerate the preview and rendering of title, transition, motion, opacity, color, and image distortion effects; and rendering a video output.
In another aspect, a telephone includes a camera and a processor coupled to the camera, the processor having code to capture a video using the camera; display frames of the video for editing; select one or more frames to be cut; select one or more transitions to be applied to the video; select one or more audio tracks to add to the video; and adjust the volume of the video. The telephone may be a cell phone, a WiFi phone, a WiMAX phone, or an FCC part-15 compatible phone such as a corded or cordless telephone.
In another aspect, a camcorder includes a camera and a processor coupled to the camera, the processor having code to capture a video using the camera; display frames of the video for editing; select one or more frames to be cut; select one or more transitions to be applied to the video; select one or more audio tracks to add to the video; and adjust the volume of the video.
In another aspect, a camera includes a processor having code to capture a video using the camera; display frames of the video for editing; select one or more frames to be cut; select one or more transitions to be applied to the video; select one or more audio tracks to add to the video; and adjust the volume of the video.
In yet another aspect, a mobile device includes a camera; a graphics processing unit (GPU); and a processor coupled to the camera and the GPU, the processor having code to capture a video using the camera and to edit the video on both the processor and the GPU.
Implementations of the mobile device may include a light projector coupled to the processor to project an image as a video output. Alternatively, a video server can be linked to the telephone to store and/or play video on demand.
In another aspect, the invention is directed to a computer program product tangibly stored on a computer-readable medium to perform the methods of the invention.
Advantages of the invention may include the following. The video editing system can be implemented on a multi-core desktop computer or workstation. Alternatively, the system can be implemented on mobile devices such as cell phones, video camcorders, cameras, and video cameras, among others. The system provides major improvements in terms of video capabilities of mobile networks. The system supports high performance mobile communications and computing and offers consumers and enterprises mobile computing and communications anytime, anywhere and enables new revenue generating/productivity enhancement opportunities. Further, in addition to enabling access to data anytime and anywhere, the equipment is easier and cheaper to deploy than wired systems. Besides improving the overall capacity, the system's broadband wireless features create new demand and usage patterns, which will in turn, drive the development and continuous evolution of services and infrastructure.
As shown in
The video editing system 10 can be connected to one or more video signal sources 26. The sources can be digital, such as digital camcorder, or analog, such as broadcast television. If the source of the video signal uses an analog format, the video editing system 10 can convert the analog video signal into a digital video signal. This can be done in a variety of ways, such as with a video interface card or as computer software running on the computer 14. Furthermore, the video editing system 10 can be connected to one or more output devices 28 that store or transmit analog or digital video data. The video editing system 10 can also include communications hardware and software by which the computer 14 is connected to other computer systems with an intranet or Internet link to send or receive streaming video data. Although illustrated as an individual computer, the video editing system 10 can be implemented on a computer network, and the data and programs used in the video editing system 10 can be stored and run on separate computers.
During the making of a video, the user can build a Storyboard, and the system divides the video into segments to make it easier to drag and drop the parts the user wants onto the storyboard where the video or movie can be edited. Once the user decides which ones he/she wants to put in the video, the user can click and drag the clips to the Storyboard in the order in which they should appear in the video or movie. Next, the user can edit clips. In one embodiment, in a timeline view, the user can click on a particular clip to be trimmed. In a Preview window, the user can drag the scroll bar slowly and watch as the video progresses. At the point where the clip is to be trimmed, the user can set a Start Trim Point (indicating the first frame to be cut), and the user can drag the progress indicator until it reaches the desired end point of your clip and the user can then click a Set End Trim Point (indicating the last frame to be cut) and the system deletes the video segment defined by the Start and End Trim Points.
Additionally, clips can be made. A clip is a smaller segment of a larger video file. A video clip can be split into a plurality of clips. This is useful if the user wants to insert either a picture or a video transition in the middle of a clip. The user can split a clip that appears on the storyboard/timeline of a current project, or you can split the clip in the Contents pane. The user can combine two or more contiguous video clips. Contiguous means the clips were captured consecutively so that the end of one clip is immediately followed by the beginning of the next. Combining clips is useful if the user has several short clips and wants to view them as one clip on the storyboard/timeline. The user can hide unwanted parts of the clip. Trimming does not remove the information from the source material and the user can restore the clip to its original length at any time. Clips can only be trimmed after they have been added to the storyboard/timeline.
Typically, abrupt transitions occur at the deleted video segment, so one or more transitions can be added to improve the changing of one clip to another in the video. The user can add a transition effect between each movie and picture. Transitions are visual tools that bridge the gap between one scene and another. The user can select transitions such as Bars, Checkerboard, Dissolve, Eye, Keyhole, Circle, among others. Adding a video transition is as easy as adding a picture. Simply drag the transition to the small square in-between each clip. The user can also add a video effect, which visually distorts the video or picture. For example, the user can age the movie or add color distortions. Video effects add video movement to animate a still image. Effects can be added to any clip by dragging the effect onto any given clip in the storyboard.
A soundtrack can really draw the audience in and add fun to the video or movie. The user can import favorite music clips and then drag them to the timeline. The user can adjust the volume of the music being added in relationship to and the volume of the audio on the video clip. For example, the user can turn the audio track on the video clip all the way down so that only the music is heard, or adjust the levels so that both are audible. A slider can be dragged left or right, depending on whether the user wants the audio from the video to be more prominent or the music to be more prominent.
In one embodiment, AMD's dual core 64-bit processor with AMD's Direct Connect technology places two 64-bit processors on a single die, that is, twice the processing power sharing the same physical footprint of a single CPU. The instructions sets delivered to the CPU via the software can be spread over two CPUs. AMD's Direct Connect architecture directly connects the CPU to the I/O, memory controllers and other CPUs, eliminating the bottlenecks and reducing latency. The integrated DDR memory controller within this architecture reduces latencies during memory access when compared to traditional front-side bus based memory controllers. The HyperTransport technology offers a high speed point-to-point that connects directly to the CPUs and the I/Os. With Direct Connect architecture, the CPUs have their own dedicated memory controller, which does away with the bottleneck and optimizes the performance of system memory at the same time.
An optional graphics processing unit (GPU) 7 is connected to the processor 1. For example, the GPU 7 may be NVIDIA's GoForce 5500 which focuses mainly on video decoding/encoding and 3D acceleration. The GPU 7 can playback H.264, WMV9 and MPEG4 (DivX/Xvid) in real time at native DVD resolutions and can also handle up to a 10-megapixel image size.
A cellular transceiver 6A is connected to the processor 1 to access cellular network including data and voice. The cellular transceiver 6A can communicate with CDMA, GPRS, EDGE or 4G cellular networks. In addition, a broadcast transceiver 6B allows the device to receive satellite transmissions or terrestrial broadcast transmissions. The transceiver 6B supports voice or video transmissions as well as Internet access. Other alternative wireless transceiver can be used. For example, the wireless transceiver can be WiFi, WiMax, 802.X, Bluetooth, infra-red, cellular transceiver all, one or more, or any combination thereof.
In one implementation, the transceiver 6B can receive XM Radio signals or Sirius signals. XM Radio broadcasts digital channels of music, news, sports and children's programming direct to cars and homes via satellite and a repeater network, which supplements the satellite signal to ensure seamless transmission. The channels originate from XM's broadcast center and uplink to satellites or high altitude planes or balloons acting as satellites. These satellites transmit the signal across the entire continental United States. Each satellite provides 18 kw of total power making them the two most powerful commercial satellites, providing coast-to-coast coverage. Sirius is similar with 3 satellites to transmit digital radio signals. Sirius's satellite audio broadcasting systems include orbital constellations for providing high elevation angle coverage of audio broadcast signals from the constellation's satellites to fixed and mobile receivers within service areas located at geographical latitudes well removed from the equator.
In one implementation, the transceiver 6B receives Internet protocol packets over the digital radio transmission and the processor enables the user to browse the Internet at high speed. The user, through the device, makes a request for Internet access and the request is sent to a satellite. The satellite sends signals to a network operations center (NOC) who retrieves the requested information and then sends the retrieved information to the device using the satellite.
In another implementation, the transceiver 6B can receive terrestrial Digital Audio Broadcasting (DAB) signal that offers high quality of broadcasting over conventional AM and FM analog signals. In-Band-On-Channel (IBOC) DAB is a digital broadcasting scheme in which analog AM or FM signals are simulcast along with the DAB signal The digital audio signal is generally compressed such that a minimum data rate is required to convey the audio information with sufficiently high fidelity. In addition to radio broadcasts, the terrestrial systems can also support internet access. In one implementation, the transceiver 6B can receive signals that are compatible with the Ibiquity protocol.
In yet another embodiment, the transceiver 6B can receive Digital Video Broadcast (DVB) which is a standard based upon MPEG-2 video and audio. DVB covers how MPEG-2 signals are transmitted via satellite, cable and terrestrial broadcast channels along with how such items as system information and the program guide are transmitted. In addition to DVB-S, the satellite format of DVB, the transceiver can also work with DVB-T which is DVB/MPEG-2 over terrestrial transmitters and DVB-H which uses a terrestrial broadcast network and an IP back channel. DVB-H operates at the UHF band and uses time slicing to reduce power consumption. The system can also work with Digital Multimedia Broadcast (DMB) as well as terrestrial DMB.
In yet another implementation, Digital Video Recorder (DVR) software can store video content for subsequent review. The DVR puts TV on the user's schedule so the user can watch the content at any time. The DVR provides the power to pause video and do own instant replays. The user can fast forward or rewind recorded programs.
In another embodiment, the device allows the user to view IPTV over the air. Wireless IPTV (Internet Protocol Television) allows a digital television service to be delivered to subscribing consumers using the Internet Protocol over a wireless broadband connection. Advantages of IPTV include two-way capability lacked by traditional TV distribution technologies, as well as point-to-point distribution allowing each viewer to view individual broadcasts. This enables stream control (pause, wind/rewind etc.) and a free selection of programming much like its narrowband cousin, the web. The wireless service is often provided in conjunction with Video on Demand and may also include Internet services such as Web access and VOIP telephony, and data access (Broadband Wireless Triple Play). A set-top box application software running on the processor 210 and through cellular or wireless broadband internet access, can receive IPTV video streamed to the handheld device.
IPTV covers both live TV (multicasting) as well as stored video (Video on Demand VOD). Video content can be MPEG protocol. In one embodiment, MPEG2TS is delivered via IP Multicast. In another IPTV embodiment, the underlying protocols used for IPTV are IGMP version 2 for channel change signaling for live TV and RTSP for Video on Demand. In yet another embodiment, video is streamed using the H.264 protocol in lieu of the MPEG-2 protocol. H.264, or MPEG-4 Part 10, is a digital video codec standard, which is noted for achieving very high data compression. It was written by the ITU-T Video Coding Experts Group (VCEG) together with the ISO/IEC Moving Picture Experts Group (MPEG) as the product of a collective partnership effort known as the Joint Video Team (JVT). The ITU-T H.264 standard and the ISO/IEC MPEG-4 Part 10 standard (formally, ISO/IEC 14496-10) are technically identical, and the technology is also known as AVC, for Advanced Video Coding. H.264 is a name related to the ITU-T line of H.26x video standards, while AVC relates to the ISO/IEC MPEG side of the partnership project that completed the work on the standard, after earlier development done in the ITU-T as a project called H.26L. It is usual to call the standard as H.264/AVC (or AVC/H.264 or H.264/MPEG-4 AVC or MPEG-4/H.264 AVC) to emphasize the common heritage. H.264/AVC/MPEG-4 Part 10 contains features that allow it to compress video much more effectively than older standards and to provide more flexibility for application to a wide variety of network environments. H.264 can often perform radically better than MPEG-2 video-typically obtaining the same quality at half of the bit rate or less. Similar to MPEG-2, H.264/AVC requires encoding and decoding technology to prepare the video signal for transmission and then on the screen 230 or substitute screens (STB and TV/monitor, or PC). H.264/AVC can use transport technologies compatible with MPEG-2, simplifying an up-grade from MPEG-2 to H.264/AVC, while enabling transport over TCP/IP and wireless. H.264/AVC does not require the expensive, often proprietary encoding and decoding hardware that MPEG-2 depends on, making it faster and easier to deploy H.264/AVC solutions using standards-based processing systems, servers, and STBs. This also allows service providers to deliver content to devices for which MPEG-2 cannot be used, such as PDA and digital cell phones.
The H.264/AVC encoder system in the main office turns the raw video signals received from content providers into H.264/AVC video streams. The streams can be captured and stored on a video server at the headend, or sent to a video server at a regional or central office (CO), for video-on-demand services. The video data can also be sent as live programming over the network. Standard networking and switching equipment routes the video stream, encapsulating the stream in standard network transport protocols, such as ATM. A special part of H.264/AVC, called the Network Abstraction Layer (NAL), enables encapsulation of the stream for transmission over a TCP/IP network. When the video data reaches the handheld device through the transceiver 6B, the application software decodes the data using a plug-in for the client's video player (Real Player and Windows Media Player, among others).
In addition to the operating system and user selected applications, another application, a VOIP phone application executes on the processing unit or processor 1 Phone calls from the Internet directed toward the mobile device are detected by the mobile radio device and sent, in the form of an incoming call notification, to the phone device (executing on the processing unit 1). The phone device processes the incoming call notification by notifying the user by an audio output such as ringing. The user can answer the incoming call by tapping on a phone icon, or pressing a hard button designated or preprogrammed for answering a call. Outgoing calls are placed by a user by entering digits of the number to be dialed and pressing a call icon, for example. The dialed digits are sent to the mobile radio device along with instructions needed to configure the mobile radio device for an outgoing call using either the cellular transceiver 6A or the wireless broadcast transceiver 6B. If the call is occurring while the user is running another application such as video viewing, the other application is suspended until the call is completed. Alternatively, the user can view the video in mute mode while answering or making the phone call.
The light projector 4 includes a light source such as a white light emitting diode (LED) or a semiconductor laser device or an incandescent lamp emitting a beam of light through a focusing lens to be projected onto a viewing screen. The beam of light can reflect or go through an image forming device such as a liquid crystal display (LCD) so that the light source beams light through the LCD to be projected onto a viewing screen.
Alternatively, the light projector 4 can be a MEMS device. In one implementation, the MEMS device can be a digital micro-mirror device (DMD) available from Texas Instruments, Inc., among others. The DMD includes a large number of micro-mirrors arranged in a matrix on a silicon substrate, each micro-mirror being substantially of square having a side of about 16 microns.
Another MEMS device is the grating light valve (GLV). The GLV device consists of tiny reflective ribbons mounted over a silicon chip. The ribbons are suspended over the chip with a small air gap in between. When voltage is applied below a ribbon, the ribbon moves toward the chip by a fraction of the wavelength of the illuminating light and the deformed ribbons form a diffraction grating, and the various orders of light can be combined to form the pixel of an image. The GLV pixels are arranged in a vertical line that can be 1,080 pixels long, for example. Light from three lasers, one red, one green and one blue, shines on the GLV and is rapidly scanned across the display screen at a number of frames per second to form the image.
In one implementation, the light projector 4 and the camera 5 face opposite surfaces so that the camera 5 faces the user to capture user finger strokes during typing while the projector 4 projects a user interface responsive to the entry of data. In another implementation, the light projector 4 and the camera 5 on positioned on the same surface. In yet another implementation, the light projector 4 can provide light as a flash for the camera 5 in low light situations.
As shown in
In another embodiment, as part of the content upload, the user captures and edits video taken with a mobile device such as a camcorder, a camera, a mobile phone, or a cell phone. The user performs simple edits to the video segment. The system allows the editing user more creative freedom at each step in the process, such as being able to preview and correct each edit decision on the fly. The video editing process becomes similar to putting together a document or graphics presentation where the user cuts and pastes the segments together adding effects and titles.
The software can provide Linear Editing where the content can only be edited sequentially similar to older mechanical techniques of cutting films to perform the edit functions. The software can alternatively provide Non-Linear Editing where editing in this environment is essentially is a visual Cut-and-Paste method and the user can edit any part of the video at will.
The system can provide In-Camera Editing: Video shots are structured in such a way that they are shot in order and of correct length. In another embodiment, the system allows the user to assemble edit: Video shots are not structured in a specific order during shooting but are rearranged and unneeded shots deleted at the time of transferring (copying). This process requires at the least, a Camcorder and VCR. the original footage remains intact, but the rearranged footage is transferred to a new tape. Each scene or cut is “assembled” on a blank tape either one-at-a-time or in a sequence. The system can provide two types of Assemble Editing: 1) A Roll—Editing from a single source, with the option of adding an effect, such as titles or transitioning from a frozen image the start of the next cut or scene and 2) A/B Roll—Editing from a minimum of two sources or Camcorders and recording to a third source. The system can also support insert editing where new material is recorded over existing footage. This technique can be used during the original shooting process or during a later editing process. The system provides Titles on Cardboard, Paper, or other Opaque Media—Painting titles on opaque media and recording the pages on videotape and inserting or assembling the title between scenes, previously shot, during the editing process.
The system supports audio or sound mixing where two or more sound sources can be connected to a sound mixer and then inputted into the video. The system also supports Audio Dubbing for adding audio to footage that is already edited together or previously shot. The audio is added to the video tape without altering the previously recorded video and, in some cases, without altering the previously recorded audio.
The above process is suitable for editing consumer produced content which tends to be short. In certain contents such as news or movies that take too long to transmit or view, the contents need to be reduced into chunks of one, five, ten or fifteen minutes, for example, to allow easy viewing while the user is traveling or otherwise don't have full attention on the device for an extended period. In one embodiment, video is micro-chunked to reduce entertainment to its simplest discrete form, be it a blog post, a music track, or a skit. Next, the system makes the content available and lets people download, view, read, or listen. The system lets consumers subscribe to content through RSS- and podcast-style feeds so they can enjoy it wherever and whenever they like. Optionally, the system can put ads and tracking systems into the digital content itself to provide revenue. In one implementation, the system provides microchunk videos entirely free, but it plays in a pop-up window alongside an ad or alternatively short commercials also play before some segments. The microchunks can be e-mailed, linked to, searched for, downloaded, remixed, and made available on-line.
The user or producer can embed meta data into the video or music. Exemplary meta data for video or musical content such as CDs includes artist information such as the name and a list of albums available by that artist. Another meta data is album information for the title, creator and Track List. Track metadata describes one audio track and each track can have a title, track number, creator, and track ID. Other exemplary meta data includes the duration of a track in milliseconds. The meta data can describe the type of a release with possible values of: TypeAlbum, TypeSingle, TypeEP, TypeCompilation, TypeSoundtrack, TypeSpokenword, TypeInterview, TypeAudiobook, TypeLive, TypeRemix, TypeOther. The meta data can contain release status information with possible values of: StatusOfficial, StatusPromotion, StatusBootleg. Other meta data can be included as well.
The meta-data can be entered by the videographer, the producer, the record company, or by a viewer or purchaser of the content. In one implementation, a content buyer (such as a video buyer of video content) can store his or her purchased or otherwise authorized content on the server in the buyer's own private directory that no one else can access. When uploading the multimedia files to the server, the buyer annotates the name of the files and other relevant information into a database on the server. Only the buyer can subsequently download or retrieve files he or she uploaded and thus content piracy is minimized. The meta data associated with the content is stored on the server and is searchable and accessible to all members of the community, thus facilitating searching of multimedia files for everyone.
In one implementation that enables every content buyer to upload his/her content into a private secured directory that cannot be shared with anyone else, the system prevents unauthorized distribution of content. In one implementation for music sharing that allows one user to access music stored by another user, the system pays royalty on behalf of its users and supports the webcasting of music according to the Digital Millennium Copyright Act, 17 U.S.C. 114. The system obtains a statutory license for the non-interactive streaming of sound recordings from Sound Exchange, the organization designated by the U.S. Copyright Office to collect and distribute statutory royalties to sound recording copyright owners and featured and non featured artists. The system is also licensed for all U.S. musical composition performance royalties through its licenses with ASCAP, BMI and SESAC. The system also ensures that any broadcast using the client software adheres to the sound recording performance complement as specified in the DMCA. Similar licensing arrangements are made to enable sharing of images and/or videos/movies.
The system is capable of indexing and summarizing images, music clips and/or videos. The system also identifies music clips or videos in a multimedia data stream and prepares a summary of each music video that includes relevant image, music or video information. The user can search the music using the verbal search system discussed above. Also, for game playing, the system can play the music or the micro-chunks of video in accordance with a search engine or a game engine instruction to provide better gaming enjoyment.
The methods described may be implemented in hardware, firmware, software, or combinations thereof, or in a computer program product tangibly embodied in a computer readable storage device. Storage devices suitable for tangibly embodying the computer program include all forms of volatile and non-volatile memory, including semiconductor memory devices, magnetic disks, magneto-optical disks, and optical disks.
|1||*||Kevin Parrish, "Nvidia GPUs Speed Up Win 7 & Leopard," April 2009|
|Citing Patent||Filing date||Publication date||Applicant||Title|
|US8321038||Jun 11, 2007||Nov 27, 2012||Porto Vinci Ltd. Limited Liability Company||Presentation of still image data on display devices using a wireless home entertainment hub|
|US8583530||Mar 17, 2011||Nov 12, 2013||Hartford Fire Insurance Company||Code generation based on spreadsheet data models|
|US8799955 *||Aug 26, 2008||Aug 5, 2014||At&T Intellectual Property I, Lp||Apparatus and method for managing media content|
|US8830245||Dec 14, 2010||Sep 9, 2014||Amazon Technologies, Inc.||Load balancing between general purpose processors and graphics processors|
|US9052959||Aug 12, 2014||Jun 9, 2015||Amazon Technologies, Inc.||Load balancing between general purpose processors and graphics processors|
|US20100058430 *||Aug 26, 2008||Mar 4, 2010||At&T Intellectual Property I, L.P.||Apparatus and method for managing media content|
|US20120154410 *||Jun 21, 2012||Baik Hyun-Ki||Apparatus and method for processing a frame in consideration of the processing capability and power consumption of each core in a multicore environment|
|WO2012082589A2 *||Dec 12, 2011||Jun 21, 2012||Amazon Technologies, Inc.||Load balancing between general purpose processors and graphics processors|
|U.S. Classification||345/589, 455/556.1, 345/502|
|International Classification||G06F15/16, H04M1/00, G09G5/02|
|Jan 11, 2012||AS||Assignment|
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TRAN, BAO;REEL/FRAME:027518/0779
Owner name: MUSE GREEN INVESTMENTS LLC, DELAWARE
Effective date: 20111209