WO2002079913A2 - Coordinating images displayed on devices with two or more displays - Google Patents

Coordinating images displayed on devices with two or more displays Download PDF

Info

Publication number
WO2002079913A2
WO2002079913A2 PCT/US2002/010458 US0210458W WO02079913A2 WO 2002079913 A2 WO2002079913 A2 WO 2002079913A2 US 0210458 W US0210458 W US 0210458W WO 02079913 A2 WO02079913 A2 WO 02079913A2
Authority
WO
WIPO (PCT)
Prior art keywords
display
user
direct
image
microdisplay
Prior art date
Application number
PCT/US2002/010458
Other languages
French (fr)
Other versions
WO2002079913A3 (en
Inventor
Matthew G. Pallakoff
Original Assignee
Move Mobile Systems, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Move Mobile Systems, Inc. filed Critical Move Mobile Systems, Inc.
Priority to EP02719425A priority Critical patent/EP1391044A2/en
Priority to AU2002250508A priority patent/AU2002250508A1/en
Priority to KR10-2003-7013029A priority patent/KR20030097820A/en
Priority to CA002441301A priority patent/CA2441301A1/en
Priority to JP2002577682A priority patent/JP2004535699A/en
Publication of WO2002079913A2 publication Critical patent/WO2002079913A2/en
Publication of WO2002079913A3 publication Critical patent/WO2002079913A3/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/14Digital output to display device ; Cooperation and interconnection of the display device with other functional units
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • G06F1/16Constructional details or arrangements
    • G06F1/1613Constructional details or arrangements for portable computers
    • G06F1/1626Constructional details or arrangements for portable computers with a single-body enclosure integrating a flat display, e.g. Personal Digital Assistants [PDAs]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • G06F1/16Constructional details or arrangements
    • G06F1/1613Constructional details or arrangements for portable computers
    • G06F1/1633Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
    • G06F1/1637Details related to the display arrangement, including those related to the mounting of the display in the housing
    • G06F1/1647Details related to the display arrangement, including those related to the mounting of the display in the housing including at least an additional display
    • G06F1/165Details related to the display arrangement, including those related to the mounting of the display in the housing including at least an additional display the additional display being small, e.g. for presenting status information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/02Constructional features of telephone sets
    • H04M1/0202Portable telephone sets, e.g. cordless phones, mobile phones or bar type handsets
    • H04M1/0206Portable telephones comprising a plurality of mechanically joined movable body parts, e.g. hinged housings
    • H04M1/0208Portable telephones comprising a plurality of mechanically joined movable body parts, e.g. hinged housings characterized by the relative motions of the body parts
    • H04M1/0214Foldable telephones, i.e. with body parts pivoting to an open position around an axis parallel to the plane they define in closed position
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/14Digital output to display device ; Cooperation and interconnection of the display device with other functional units
    • G06F3/1423Digital output to display device ; Cooperation and interconnection of the display device with other functional units controlling a plurality of local displays, e.g. CRT and flat panel display
    • G06F3/1431Digital output to display device ; Cooperation and interconnection of the display device with other functional units controlling a plurality of local displays, e.g. CRT and flat panel display using a single graphics controller
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2340/00Aspects of display data processing
    • G09G2340/14Solving problems related to the presentation of information to be displayed
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2340/00Aspects of display data processing
    • G09G2340/14Solving problems related to the presentation of information to be displayed
    • G09G2340/145Solving problems related to the presentation of information to be displayed related to small screens
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/72445User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality for supporting Internet browser applications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2250/00Details of telephonic subscriber devices
    • H04M2250/12Details of telephonic subscriber devices including a sensor for measuring a physical value, e.g. temperature or motion
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2250/00Details of telephonic subscriber devices
    • H04M2250/16Details of telephonic subscriber devices including more than one display unit

Definitions

  • the invention generally relates to hand-held communication and computing devices, such as cellular- phone handsets, smart-phones, Web-phones, and PDAs. More specifically, the invention relates to systems and methods for coordinating images displayed on hand-held devices that have two or more displays, especially those that include a direct-view display for viewing relatively low-resolution text or images (like the displays built into most cellular phones today), and a microdisplay that uses magnifying optics to display images that, when the microdisplay is held near a user's, eye can appear larger and higher resolution than the images shown on the direct-view display.
  • a direct-view display for viewing relatively low-resolution text or images (like the displays built into most cellular phones today)
  • microdisplay that uses magnifying optics to display images that, when the microdisplay is held near a user's, eye can appear larger and higher resolution than the images shown on the direct-view display.
  • microdisplay module technology Using recently developed high-resolution microdisplay module technology, device makers can embed microdisplays in devices, or attach microdisplays including wearable displays to devices, and allow users to view full-size web-pages on pocket-size devices - as well as being able to view large video and still images, blueprints, office documents, and other high-resolution content.
  • Typing text into a Web page or other document displayed by a device on a microdisplay typically involves either using a cursor control to "type" on a virtual keyboard image shown on the microdisplay (e.g. moving the cursor over each character displayed on the virtual keyboard and clicking), or using a separate physical device (such as a keypad, a keyboard, a tablet or writing pad, or a virtual keyboard).
  • Typing using a virtual keyboard displayed on the microdisplay is tedious and slow (particularly when holding a device near-to-eye) because it requires the user to carefully position the cursor over each little character on the virtual keyboard to type each character.
  • At least one display is a direct-view display, like those on most cell phones or PDAs in use in 2001 , for viewing at normal reading distance of approximately 12 to 24 inches (which this patent will refer to as "arms'-length" viewing).
  • at least one of the displays is a microdisplay, a tiny display with magnifying optical elements, for viewing larger, higher-resolution images when the microdisplay is positioned close to the eye.
  • the microdisplays can be embedded in devices, or they can be attached to the devices, or the microdisplays can be in wearable displays attached to the devices (wirelessly or through a cable).
  • microdisplays for their view-finders: When a user brings the view-finder to their eye when the camera is playing back video, the user can watch a larger image of the video-clip than could be seen on small direct-view display.
  • high resolution microdisplays have been developed that can present images large enough to display most Web pages.
  • Inviso Inc.'s 3430 Optiscape II microdisplay displays an 800x600 pixel color image that appears to users to be as large as a 19-inch monitor roughly 2.5 feet away. Since optics are used to create an image that looks large, despite the fact that the microdisplay module is small, these images are often called “virtual images”. And in this patent we will sometimes call them "apparent images”.
  • Preferred embodiments of this invention would normally be implemented on devices using microdisplays that can display color images at least 800 pixels wide and at least 600 pixels tall.
  • An important common characteristic of the preferred embodiments of the present invention is that they are designed to facilitate a seamless, natural transition between near-to-eye viewing of content on a device's microdisplay (or microdisplays) and arms'-length viewing of content on a direct-view display. Key to enabling this seamless transition is software, hardware and storage mechanisms that are used to intelligently coordinate images on the microdisplay with images on the direct-view display.
  • a preferred embodiment of my invention is a "smart-phone" supporting functions such as instant messaging, email, and contact management, in addition to phone calling, with both a microdisplay and a direct-view display. Since most instant messages, for example, are relative short text messages, this smart-phone could display instant messages on the direct-view display so that users can see the messages at arm's length, and can concurrently see the keys, buttons and controls on the phone handset so the user can easily type replies to the messages.
  • an instant message includes a reference to a Web page URL, or includes an attached photo or other relatively large high- resolution image or document
  • the smart-phone can display the Web page or attachment in the microdisplay instead of the direct-view display. This allows the user to see the Web page or attachment, which may be relatively large and high-resolution, by bringing the microdisplay to the user's eye, much like a PC display like image.
  • the device when the device is displaying an image (or is preparing to display an image) in one display but not the others, then the device can display a text or graphics indication on one or more of the displays that is not displaying that image, indicating that one of the other displays is displaying that image (or that the device is preparing to display that image).
  • the device can display a flashing eye icon on the direct-view display; and when the Web page is ready for viewing on the microdisplay, then the device can display a static icon on the direct-view display. This lets the user know that the user can now bring the device near-to-eye to view the Web page in a high resolution, magnified form.
  • the same technique can be used any time the user can operate functions using the direct-view display that result in content then appearing on the microdisplay, or vice-versa. (If the user has the device near-to-eye and operates a function that results in new content appearing on the direct-view display, the microdisplay can display some indication that there is new content to view on the direct-view display.) It is important to provide these cues, since a user typically can not see the content on a microdisplay when holding the device at arms-length, and a user typically can not see content on a direct-view display (or doesn't think to look at the direct-view display) when looking into a microdisplay that is near-to-eye.
  • Device designers can choose to use the direct-view display as the primary display for many smart- phone functions, including instant messaging, email, phone dialing, contact management, schedule management, calculator functions, and the like, and to use the microdisplay as a high resolution viewer coordinated with the direct-view display.
  • the direct-view display When the direct-view display displays this region of interest, the direct-view display acts as a window onto the larger image that the user had been viewing on the microdisplay. That is, the region of interest portion of the image seen on the microdisplay is mirrored on the direct-view display. The user should then be able to read text or interact with content in that region of interest, or use controls on the device to scroll the region of interest left, right, up, down, or other directions, to see different parts of the image other than the subset of the image currently shown in the direct-view display.
  • the device can graphically indicate where the region of interest has moved - for example, by overlaying a rectangle in the shape of the region of interest (which will generally correspond to the pixel width and height of the direct-view display) on top of the image seen in the microdisplay.
  • Device makers can choose not to explicitly show a rectangle or other explicit representation of the region of interest on the image viewed in the microdisplay, or choose to make this kind of explicit "region of ref indicator optional for the user.
  • the device can mirror the region of interest in the direct-view display, as described above.
  • the region of interest does not necessarily have to be rectangular.
  • a device designer could choose to use a circular or other shaped direct-view display (for example, on a watch or a non-conventional cell-phone or PDA), and then choose to use a correspondingly shaped region of interest.
  • the usage model facilitated by this invention mirrors what happens when a person uses a desktop computer monitor: Typically a person briefly "takes in” the entire monitor display (i.e., looks at the entire image without focusing attention on any specific region of the display at first), and then focuses attention on a specific region of interest (such as text they want to read, or a text-insertion point where they want to type, or some other item of interest).
  • This invention allows microdisplay-enhanced devices to model this behavior.
  • this invention enables better ergonomics and better power-consumption characteristics than previous designs for many types of microdisplay-enhanced devices.
  • a user can choose to briefly observe a large area of content displayed on the microdisplay, and quickly skim the large area of content to identify a region of interest where they want to focus attention - and then (if they want) they can bring the device to arms'-length to view and interact with that region of interest on the direct-view display.
  • users can quickly orient themselves and quickly skim the large area of content to find a specific region of interest - just as they do on desktop monitors.
  • the microdisplay when the user is not holding the device near-to-eye, the microdisplay can be turned off or idled, saving power- so devices that use this invention can use significantly less power than devices that leave microdisplays on continuously.
  • direct-view displays on the device when the user is holding the device near-to-eye, direct-view displays on the device can be turned off or dimmed or idled, saving power.
  • the primary motivitation for this invention has to make it feasible for people to comfortably access and interact with full size Web content (and other high-resolution or large visual content) on pocket-size devices with built-in or attachable microdisplays.
  • the present invention's embodiments related to intelligently coordinating direct-view displays and microdisplays are applicable even when the microdisplays are part of a wearable display connected to the device rather than being embedded in or attached directly to the device.
  • Wearable displays can include one or two microdisplays.
  • Inviso Inc.'s binocular wearable display product concept called "eShades" included two microdisplays, one for each eye.
  • a monocular wearable display i.e. one with just one microdisplay positioned in front of one eye
  • many binocular wearable displays like Inviso Inc.'s eShades concept product
  • Fig. 1 A is a side view of a microdisplay-enhanced handset.
  • Fig. 1 B is a front view of a microdisplay-enhanced handset.
  • Fig. 1C is a front view of a microdisplay-enhanced handset, held in a person's hand.
  • Fig. 2A is a side view of a flip-style handset with both a direct-view display and a microdisplay.
  • Fig. 2B is a front view of a flip-style handset with both a direct-view display and a microdisplay.
  • Fig. 2C is a front view of a flip-style handset with both a direct-view display and a microdisplay, held in a person's hand.
  • Fig. 3A is a side view of an alternative configuration for a handset with both a microdisplay and a direct-view display.
  • Fig. 3B is a front view of an alternative configuration for a handset with both a microdisplay and a direct-view display.
  • Fig. 3C is a front view of an alternative configuration for a handset with both a microdisplay and a direct-view display, held in a person's hand.
  • Fig. 4 is an illustration of a handset with two displays and with buttons for highlighting and selecting selectable items on a list of selectable items displayed on one of the displays.
  • Fig. 5 is an illustration of a handset according to my invention, showing the coordination of images on the direct-view display and on the microdisplay.
  • Fig. 6 is an illustration of a handset according to my invention, showing an indication on the direct- view display (shown in close up on the left of Fig. 6) that there is new content ready for viewing on the handset's microdisplay.
  • Fig. 1 A shows a microdisplay-enhanced handset 118 in side-view.
  • Fig. 1B shows the same in front-view.
  • Fig. 1C shows the same in front-view, shown held in a person's hand.
  • This handset includes a microdisplay 100 positioned above a direct-view display 101.
  • Most of the elements of this handset could be applied to a wide range of devices - including cellular phones, "Smart-phones", Web-phones, PDAs, hand-held computers, remote controls, and others.
  • most references to "handsets" can refer to any of these types of devices.
  • the direct-view display can be any kind of direct-view display, but preferably it would be a thin bit-map LCD display with less resolution than the microdisplay but enough to render an identifiable portion of a larger image.
  • a 160x120-pixel direct-view display with 3-bits of gray scale (white, black, and six intensities of gray in-between white and black) would be sufficient for displaying text and simple graphics while still keeping the cost of the direct-view display relatively low.
  • Fig. 1 A Below the side-view of Fig. 1 A and the first front-view illustrations of the handset in Fig. 1 B are side- view 116 and front-view 117 illustrations of Inviso's 3430 Optiscape microdisplay module which includes electronics and optics - just to illustrate their size relative to the overall handset. Many other elements on this handset are common on cell phone handsets today such as the speaker 114, a power on/off button 108, a microphone 115, "Start Call” and. "End Call” buttons 112.
  • Handset users frequently interact with various functions supported on the handset, such as instant messaging, email, contact management, dialing from a list of contacts, calculator, Web browsing, and so on.
  • functions supported on the handset such as instant messaging, email, contact management, dialing from a list of contacts, calculator, Web browsing, and so on.
  • users need to see (or hear) feedback as they use the functions.
  • the feedback presented by the handset is relatively simple and brief - capable of being expressed in a few lines of text or in a small low-resolution bit-mapped display.
  • relatively simpler or brief feedback should usually be presented on the direct-view display 100 of Fig. 1B, as is done on a traditional handset that has only a direct-view display.
  • a handset built according to this invention should display that image on the microdisplay 101 of Fig. 1B.
  • designers can also allow users to interact with functions while holding the device near-to-eye and using the microdisplay by designing the handset so it displays some feedback on the microdisplay (instead of, or in addition to, displaying that feedback on the direct-view display).
  • a one-piece handset such as the one illustrated in Fig. 1
  • the microdisplay near the top of the handset and placing the direct-view display in the middle of the phone (above the keypad).
  • the device can also be designed such that depressing a given button on the device (or depressing a combination of one or more buttons or controls) results in the device popping up and displaying (on one or more displays) a menu in a menu bar associated with the device or with the front window on the device (such as the "File" menu at the top left of documents on most PC documents) while concurrently highlighting the top menu item in that menu. Then depressing left or right arrow buttons on the device, such as those in control 113 on Fig. 1 , could closes the currently popped-up menu and causes popping up the previous or next menu in the menu bar and highlights the top item on the menu.
  • the device could be designed such that further depressing up or down arrows on the device (such as those in control 113 on Fig. 1) highlights the next or previous menu item in the currently popped up menu. At that point, pressing an "enter” or “select” button on the phone results in the highlighted menu item being selected.
  • Different types of items can do different things when selected in this way.
  • selecting an item should cause the same behavior as would occur on a PC if the user used a mouse to click on that type of item. For example, selecting a check box that is not currently checked should cause it to be checked; and selecting a check box that is currently checked should cause it to be unchecked.
  • the device can further be designed such that spoken commands via speech recognition software can be used to pop up and traverse menus or to select menu items; and in which the device's speech recognition features are enabled only when the user holds in a button designed for that function on the device.
  • This "push-to-talk" button could be on the side of the device where a user could depress it with a finger or thumb on the hand holding the device.
  • Fig. 2A through Fig 2C illustrate a flip-style handset with both a direct-view display 200 and a microdisplay 201.
  • Microdisplays are relatively thick (because of their magnifying optical elements) and direct-view displays are relatively thin (since they are lower resolution and do not require magnifying optical elements). So a flip-style phone would place the direct-view display on the thin part of the phone that flips away from the main body of the phone. In Fig. 2B, this is the top part of the phone.
  • a flip-style phone could also be designed in which the keyboard and the direct-view display are on a relatively thin bottom part ofthe phone, and the microdisplay is on a relatively thick top part of the phone that flips up. But that would tend to make the phone feel top-heavy.
  • Fig. 3A through Fig. 3C illustrate a compelling alternative configuration for a handset with both a microdisplay 300 and a direct-view display 301 as seen in the side-view representation of Fig. 3A.
  • the direct-view display is placed directly over the microdisplay.
  • the user simply uses the direct-view display for feedback, since the feedback consists of relatively low-resolution text or images.
  • the user looks at the front of the phone 305 in Fig. 3B, he or she see the relatively low-resolution text or images on the direct-view display 302.
  • a handset can be designed to have a button (or other control) that the user uses to make the direct-view display become transparent (and to turn the microdisplay on) so the user can see through to the microdisplay.
  • a preferred embodiment would include a distance-sensor 304 in Fig. 3B that automatically senses when the user's eye is close to the microdisplay (within 2-3 inches) and then automatically makes the direct- view display (302 in Fig. 3A and Fig. 3B) transparent and turns on the microdisplay (300 in Fig.
  • Fig. 3A and 303 in Fig. 3C so the user can see through the direct-view display 301 to the image presented by the microdisplay 300.
  • Various distance-sensing technologies including ultrasonic distance sensors - are used in current digital video cameras, still cameras, and other electronic devices.
  • the Sony Mavica MVC CD1000 digital still camera has both a direct-view display on its back and a separate mid-resolution microdisplay in its viewfmder, and it includes a distance sensor so that when a user puts their eye close to the viewfmder the microdisplay-based viewfmder automatically turns on.
  • the handset would automatically turn on the microdisplay only when the handset is displaying a high-resolution image and when the sensor senses that the user's eye is close enough to the microdisplay - within three inches or so. At all other times, the handset would automatically turn the microdisplay off, saving considerable power relative to a device that leaves the microdisplay on all the time or one that leaves it on even when the user's eye is not close to the microdisplay.
  • the phrase "turn the microdisplay off” means to deactivate the microdisplay such that it is not generating images or generating any light that would prevent the correct operation of the direct-view display. It does not necessarily mean that no power is being supplied to the microdisplay. However, the result is substantial power savings.
  • This handset of Fig. 3A through Fig. 3C can also be built less tall than handsets that separate the two displays (as in Fig. 1A through Fig. 1C), since the two displays overlap one another in this design, although it could make the phones a bit thicker when viewed from the side.
  • a handset could also be designed with a direct-view display on one side of the handset (such as the front) and a microdisplay on the other side of the handset (such as the back). But this would require users to turn the handset over when switching between viewing the direct-view display and viewing the microdisplay - which is more awkward than simply bringing the handset to the eye to view an image in the microdisplay.
  • either or both of the displays could be touch sensitive (like many PDAs today), allowing users to input data and control aspects of the device by moving their fingers, a pen, or a stylus across the touch-sensitive area of the display(s).
  • Each of the handsets illustrated in Fig. 1 A through Fig. 3C is shown with three side-buttons on the right side -"Shift", “2nd Letter", and “3rd Letter” side-buttons - which a user can press with the hand holding the phone while simultaneously using the other hand to type the handset's face-keys.
  • the user would just type the "6" key.
  • the user would type the "6" key while holding in the "Shift" side-button.
  • To type the lowercase letter "u” the user would type the "8” key while holding in the "2nd Letter” side-button.
  • the device can also include a roller control 102 of Fig. 1 B which, when rolled in one direction, changes the highlighting in a displayed list of selectable items from the currently highlighted item to the next selectable item; and, when rolled in the other direction, changes the highlighting from the currently highlighted item to the previous selectable item.
  • a roller control 102 of Fig. 1 B which, when rolled in one direction, changes the highlighting in a displayed list of selectable items from the currently highlighted item to the next selectable item; and, when rolled in the other direction, changes the highlighting from the currently highlighted item to the previous selectable item.
  • a web-page has a "list” of selectable items, as seen in the Web-page's HTML, even if those items are arranged nonlinearly in a two-dimensional design.
  • a "selectable” item can be any displayed item that can either be selected or, in the case of a text-entry box, typed into.
  • the device can also be designed to apply to items that are part of Web pages and other documents, as well as to items that are displayed by the handset itself such as the list of functions available on the phone device.
  • a roller control can also be used to traverse through lists of items and to select items. The roller control 102 can also be pressed to select the highlighted item in any displayed list of items.
  • the device includes a button 401 that, when pressed briefly, changes the highlighting in a displayed list of selectable items from one selectable item to the previous one.
  • This patent also refers to the highlighted selectable item as the "in-focus” item, the "item in focus", or the "focus” item.
  • the device can also include a button 403 that, when pressed briefly, changes the highlighted item in a displayed list of selectable items from one selectable item to the next one.
  • the device can also include a button 402 that, when pressed, selects the highlighted item in the displayed list of items. These buttons do not have to be configured as buttons 401 , 402, and 403 appearing in Fig.
  • buttons 4 They could be configured as arrow keys and other buttons elsewhere on the device - such as the arrows and enter buttons 113 on Fig. 1B.
  • Each of these buttons can be multiple-speed (which this patent will also refer to as "multi-speed") - for example, moving the in-focus item quickly from the currently highlighted selectable item to the next or previous one when pressed hard, but moving the highlighting slowly when pressed lightly.
  • These buttons can go by many names - such as Tab-forward and Tab-back, up and down, and so on. This patent will refer to them generically as up and down or up/down. If they are multiple-speed, this patent will refer to them as multi-speed up/down controls.
  • the buttons can be placed under a shared length of plastic to form a toggle switch, with one button under one side of the toggle and the other button under the other side of the toggle, as shown in Fig. 5.
  • buttons 401 (or 403) of Fig. 4 when the button 401 (or 403) of Fig. 4 is pressed and held for more than a defined period of time (for example, more than 0.5 second), then instead of the highlighting moving from one selectable item to the previous (or next) item just once, the highlighting moves to the previous (or next) selectable item, stays briefly (e. g. 0.25 sec), moves to the previous-to- the-previous (or next-to-the-next) selectable item, stays briefly, and continues moving back (or forward) through the list of selectable items for as long as the user continues holding the corresponding button 401 (or 403).
  • This is simply an alternative to continuously rolling the roller control outlined above to quickly move the highlighting hack and forth through the list of selectable items.
  • the device can also be designed such that the longer the button 401 or 403 is held, the faster the highlighting moves through the list of selectable items, up to some maximum speed (and with the speed accelerating in well defined increments from relatively slow to relatively fast).
  • My invention can be embodied in a handset or other hand-held device with two or more displays, one of which can show content (such as an image, document, Web page, text, or other content) that includes one or more references to other content, such as referring to an image, document, Web page, text or other content.
  • content such as an image, document, Web page, text, or other content
  • references to other content such as referring to an image, document, Web page, text or other content.
  • a user can use the direct-view display of the device to read an email or instant message that has an attached photo or a reference to a Web page URL and when the user clicks on or selects that attachment or Web page URL, the handset displays the attachment or Web page URL in the microdisplay.
  • the device could show a full Web page (or a large portion of a Web page) in a microdisplay, and then show a subset of that Web page in a direct-view display.
  • the microdisplay is displaying a document or image (such as a Web page or application form) with items that can be selected or typed in, then at any given time, zero or one of those items is "in focus" and that item is visually highlighted in some way, for example, by outlining that item in light gray.
  • “In focus” means that if the user were to use a "select” control, by whatever name, on the handset or were to begin typing, then that "in focus” item would be the item selected, or, if it were a text entry box, it would be the item into which the characters would be typed. This is similar to the way items can be in "focus” in Web browsers or other applications on a desktop Windows PC. The item in "focus” is usually highlighted graphically in a desktop application, for example with a light gray outline.
  • the microdisplay is displaying a document or image (such as a Web page or application form) and if one of the items on the document or image is "in focus," then a subset of the document or image surrounding the "in focus" item can be displayed on the direct-view display.
  • This can provide a close-up view of the item that would be typed in or selected if the user were to type or use the handset's selection control. The user could see this area while holding the device a foot or two away, since it is displayed in the direct-view display. And this in turn would allow the user to see the buttons and controls on the device more easily as compared to holding the handset near-to-eye.
  • microdisplay 500 displays the web page while the subset thereof is displayed on the direct-view display 501.
  • the subset shown on the direct-view display would include both the "in focus” item and a text or bit-map representation of any text or graphic labels or symbols corresponding to that item on the form. For example, if the "in focus” item is a "Last name:” text-entry box on a Web-page form, then the direct-view display would show the label "Last name:” as well as the corresponding text-entry box, into which the user will type his or her last name.
  • the item "in focus” (and hence the image subset shown on the direct-view display) can be changed by the user typing, depressing buttons, or using other controls on the device. For example, the user can use the roller control 102 in Fig.
  • Fig. 4 illustrates this concept. Fig. 5 is explained below.
  • the item in focus (and hence the image subset shown on the direct-view display) can be changed by the user speaking into the device, with the device itself, or a service accessed by the device over a network, using speech recognition software to interpret the user's spoken commands.
  • the item in focus (and hence the image subset shown on the direct-view display) can be changed by software running on the computer processor within the device. This could be used to generate animations by having a large image made up of many smaller frames, and rapidly changing which subset of the larger image is displayed in the direct-view display.
  • the item in focus can also be changed by the user typing, pressing buttons, speaking, or using other controls on the device and, as the focus changes, the image shown in the microdisplay auto-scrolls to keep the in-focus item in view.
  • This is similar to what happens if a user views a large Web-page form on a computer monitor and the form extends past the bottom of the window in which the form is being displayed. That is, if the user uses the Tab button repeatedly to move the focus toward the end of the form, then when the focus goes past the bottom of the window to an item that can be typed-in or selected that appears on the part of the form below the window, the Web browser will typically auto- scroll the form upward within the window so that the new in-focus item will be visible within the window.
  • a user changes the state of a check-box, radio-button, or other "state" control that appears in more than one display, then all displays showing that same state control should show the state control in its current state.
  • a user types into a text-entry field shown on one or more displays, or uses arrow keys 113 of Fig. 1 or other controls to change the text-entry point or selection area, then if the text-entry point (or the end of the selection area) moves out of the subset of the document image visible in any given display, that display will automatically scroll the image so that the text-entry point (or the end of the selection area) will reappear in that display.
  • this device may display different subsets of the same document in different displays, it must auto-scroll the image in different displays at different times. For example, if a relatively low-resolution direct-view display shows a small subset of a document shown in the much higher-resolution microdisplay, then the direct-view display may have to auto-scroll several times as the user types a long message into that document, while the microdisplay could auto-scroll much less often.
  • the microdisplay is displaying a relatively high-resolution large apparent image (i.e. a virtual image, as discussed earlier)
  • the direct-view display can show a subset of the larger, higher- resolution apparent image shown in the microdisplay. That is, the direct-view display acts as a window into the larger image, showing a "region of interest" from the larger apparent image.
  • controls on the hand-held device can move the region of interest window (i. e. the borders of the subset of the Web Page, application document, or other image shown in the direct-view display) down/up right/left or in other directions across the larger image.
  • the region of interest window i. e. the borders of the subset of the Web Page, application document, or other image shown in the direct-view display
  • the direct-view display might show an upper-left corner subset of the image seen in the microdisplay, which itself might be a subset of a larger scrollable document or Web page.
  • the user can use the controls to have the direct-view show a different subset of the image seen in the microdisplay.
  • the subset of the document or Web Page that includes the text-entry point can automatically be scrolled into view in both the direct-view display and the microdisplay. That is, the "region of interest" can be automatically moved to surround the text-entry point.
  • the region of interest shown in the direct-view display is an area of bits around the current "focus" shown in the larger image displayed on the microdisplay.
  • a Web page has a set of selectable items such as links, buttons, editable text fields, and the like.
  • one of the selectable items can be in "focus", and users can move the focus back and forth through the set of selectable items using the Tab Backward and Tab Forward key.
  • Tab-Backward shift Tab Forward
  • Users can select that selectable item that is currently in focus by depressing a button (usually the space bar).
  • a manufacturer can design a cell phone handset or other hand-held device on which a user could use a thumb-roller or other control to move the focus back and forth through the list of selectable items on the Web page or screen.
  • the image of the Web page or application document shown in the microdisplay should auto-scroll to always keep the focus in view in the microdisplay, just as occurs on desktop PCs when user tabs through selectable items (using the Tab key) on a large Web page displayed in a smaller Web browser window.
  • the image of the Web page or application document shown in the direct- view display should auto-scroll to always keep the focus in view in the direct-view display.
  • each display when viewed would show both the focus item (i. e. the selectable item that would get selected if the user hits the select key) and the area immediately surrounding the focus item (which will often include text or image labels corresponding to the selectable item in focus, such as "Last Name:" labels).
  • Each display thus acts as a window onto the Web Page or Document being viewed, with the direct-view display being a smaller window than the microdisplay.
  • the direct-view display can show text corresponding to the item in focus on the Web Page.
  • the item in focus is a text-entry field in a form on a Web page in which a user is supposed to type their last name
  • the direct-view display can show any text immediately to the left (and/or right) of that text entry field (e. g. "Last Name: "). This text can often be extracted from the
  • HTML that defines the Web Page (by interpreting the HTML to determine what text lies to the left of the selectable item in focus). But often it will not be possible to extract appropriate text corresponding to the Focus item (e. g. because no text corresponding to the focus item exists). For this reason, a preferred embodiment of this device would use a bit-map direct-view display, as discussed above, instead of a text-only direct-view display.
  • the device when a user is viewing a Web page 502 on a device's microdisplay 500 held near- to-eye, the device should allow the user to position a cursor (such as the arrow cursor commonly seen when moving the mouse on a desktop computer) or a rectangular outline 503 or both (or some other indication of a "region of interest") over any chosen part of the Web page 502. Then when the user moves the device out to arms'-length the user should be able to view that "region of interest" on a direct-view display 501 on the device - i.e. viewing a subset of the larger image that appeared on the microdisplay.
  • a cursor such as the arrow cursor commonly seen when moving the mouse on a desktop computer
  • a rectangular outline 503 or both or some other indication of a "region of interest
  • the lines extending from the left corners of the Web page window 502 to the phone handset's embedded microdisplay 500 are meant to show that, when the microdisplay is held near the user's eye, the user could see a relatively large, high-resolution image like the Web page window 502.
  • the longer narrower lines extending from the "region of interest" rectangle 503 to the handset's direct-view display 501 are meant to show that a small portion of the Web page could be made visible on the direct-view display.
  • the current region of interest is a small form on the Web page where the user could enter an address to look up a map.
  • the direct-view display 501 When the direct-view display 501 displays this region of interest (the area within the rectangle 503), the direct-view display acts as a window onto the larger image that would be seen on the microdisplay if the microdisplay were held near-to-eye. While viewing the direct-view display 501 at arms'-length (that is, at a normal reading distance), a user should be able to read text or interact with content in that region of interest, or use controls on the device to scroll the region of interest left, right, up, down, or other directions, to see different parts of the image other than the subset of the image currently shown in the direct-view display.
  • the device can graphically indicate where the region of interest has moved - for example, by overlaying a rectangle 503 in the shape and size of the region of interest on top of the image seen in the microdisplay.
  • the shape and size of the region of interest should generally be the same as, or close to, the pixel width and height of the direct-view display 501.
  • the direct-view display 501 will typically have substantially fewer pixels than the microdisplay.
  • today's smartphones typically have displays that are less than 200 pixels wide and less than 250 pixels tall
  • today's PDAs typically have displays that are less than 300 pixels wide and less than 400 pixels tall.
  • Inviso Inc.'s current microdisplays are capable of displaying 800x600 pixel areas.
  • the size and shape of the direct-view display can be used to define the size and shape of the region of interest.
  • the device can show the location of the region of interest by displaying a rectangle 503 of that size on the image being viewed in the microdisplay 500. As the user moves the region of interest, the rectangle moves around the image. If the user then views the region of interest in the direct-view display 501 at arms'-length, and then uses controls on the device (or other means) to scroll the region of interest to other parts of the content being viewed (as described earlier), then when the user brings the device near-to-eye to view the content in the microdisplay, the rectangle showing the region of interest should be shown on the microdisplay in its new location on the image being viewed.
  • the device could use a range of techniques to indicate the region of interest on the microdisplay. Or device makers can choose not to explicitly mark the region of interest on the microdisplay image - choosing instead just to treat the area surrounding the cursor or surrounding highlighted "in focus" areas (discussed elsewhere in this patent) as the region of interest. Alternatively, if the region of interest is marked on the microdisplay (with a rectangle, for example), then device makers can give users the option of hiding that mark or showing that mark.
  • Fig. 6 illustrates a preferred embodiment for providing this visual cue.
  • the sample phone on the right side of Fig. 6 has a microdisplay 604 near the top and a direct-view display 600 below the microdisplay.
  • the direct-view display 601 is shown in close-up on the left side of Fig. 6.
  • the user has received an instant message.
  • the user would typically read instant messages at arms'-length (i.e.
  • the instant message in this example includes a URL 603 referring to a Web page with a map (similar to the Web page shown in Fig. 5.)
  • the device when the user operates controls on the phone to open the Web page associated with the URL, the device will display a flashing eye icon 602 on the direct-view display while the device is downloading and preparing that Web page for display. And when the Web page is ready for viewing on the microdisplay 604, the device would display a steady (i.e. non-flashing) eye icon 602.
  • the form of the indication is not important - it could be a different icon, or text, or some other visual cue, and it could flash or not-flash.
  • the user is viewing content on the microdisplay, in some cases it could be useful to provide visual cues on the microdisplay when there is significant new content available for viewing on the direct-view display, since the user may not think to look at the direct-view display while looking at the microdisplay (even if the microdisplay is on a head-mounted wearable display rather than embedded in the hand-held device with the direct-view display).
  • the present invention can be used when designing devices with a wide range of form-factors and configurations.
  • a device with a direct-view display on its face and a microdisplay embedded near the top of the device works well.
  • device makers could choose to embed or attach one or more microdisplays to any part of the device (e.g. on a hinge on the bottom of the phone or on the side) and still use this invention to intelligently coordinate content on the device's multiple displays and thereby improve the ergonomics of the device.

Abstract

A hand held electronic device (Fig. 1B) having at least two displays is disclosed. At least one display is direct-view display for viewing at normal reading distance of arms'-length viewing. The other display is a microdisplay. a tiny display with magnifying optical elements, for viewing larger, higher-resolution images when the microdisplay is positioned close to the eye. The invention allows coordinating microdisplays and direct-view displays in ways that allow people to comfortably access and interact with full-page Web content on pocket-size devices. When a user is viewing a Web page on a device's microdisplay held near-to-eye, the device allows the user to position a cursor or a rectangular outline on a particular part of the Web page, and then when the user moves the device out to arms'-length viewing, the user should be able to view that region of interest on the direct-view display.

Description

COORDINATING IMAGES DISPLAYED ON DEVICES WITH TWO OR MORE DISPLAYS
RELATED APPLICATIONS
This application claims priority, under 35 USC § 120, from US Provisional Patent Application No. 60/280,791, filed on April 2, 2001.
BACKGROUND OF THE INVENTION:
The invention generally relates to hand-held communication and computing devices, such as cellular- phone handsets, smart-phones, Web-phones, and PDAs. More specifically, the invention relates to systems and methods for coordinating images displayed on hand-held devices that have two or more displays, especially those that include a direct-view display for viewing relatively low-resolution text or images (like the displays built into most cellular phones today), and a microdisplay that uses magnifying optics to display images that, when the microdisplay is held near a user's, eye can appear larger and higher resolution than the images shown on the direct-view display.
Using recently developed high-resolution microdisplay module technology, device makers can embed microdisplays in devices, or attach microdisplays including wearable displays to devices, and allow users to view full-size web-pages on pocket-size devices - as well as being able to view large video and still images, blueprints, office documents, and other high-resolution content.
Device makers have experimented with embedding microdisplays in devices and attaching microdisplays to devices (including attaching wearable near-to-eye displays). However, prior to this invention, it has been awkward and difficult to navigate and interact with full-size Web pages and other documents and applications displayed on the microdisplays of microdisplay-enhanced handheld devices. Devices with embedded microdisplays must be held near-to-eye to see the content on the microdisplays. While users do not mind briefly bringing a device near-to-eye to view content (after all, that is what we do with cameras), users do not want to hold a device near to eye for longer periods of time. If a user is spending five minutes, ten minutes, or more reading or interacting with a set of Web pages or other documents, and if they have to hold a device near-to-eye continuously during that time, they will grow uncomfortable with the device. Furthermore, typing text on a hand-held device (already a relative awkward activity) is exceedingly awkward when the device has to be held near-to- eye to see the text as you type it. Devices with attached wearable displays, such as the head- mounted monocular displays that Xybernaut Corp. uses in their wearable computer systems, allow users to comfortably keep microdisplays near-to-eye for longer periods of time (relative to hand-held devices with embedded displays); but users still find it awkward to type text into documents viewed on wearable displays. Typing text into a Web page or other document displayed by a device on a microdisplay typically involves either using a cursor control to "type" on a virtual keyboard image shown on the microdisplay (e.g. moving the cursor over each character displayed on the virtual keyboard and clicking), or using a separate physical device (such as a keypad, a keyboard, a tablet or writing pad, or a virtual keyboard). Typing using a virtual keyboard displayed on the microdisplay is tedious and slow (particularly when holding a device near-to-eye) because it requires the user to carefully position the cursor over each little character on the virtual keyboard to type each character. Further, using a mouse on a regular computer to type text on a virtual keyboard is awkward, and it is even more awkward when using a near-to-eye microdisplay. Even when a user can type text on a physical keypad while looking at a document on a wearable display, if user's haven't learned to touch- type well (i.e. type without looking at the keypad), then users can experience an uncomfortable mental disconnect as they frequently shift their eye (or eyes) between the keypad and the image displayed on the wearable display's near-to-eye microdisplays.
However, the potential benefits of using near-to-eye microdisplays with hand-held devices are tremendous: They can allow users to view full-size web pages and other high-resolution content on pocket-size devices. So solving the problems outlined above would enable the development of compelling new types of hand-held devices. This patent describes inventions that help solve these problems.
A device with which the invention can be implemented is disclosed and claimed in U.S. patent application S/N , filed on even date herewith, assigned to the common assignee, and incorporated herein by reference.
In devices using the present invention, at least one display is a direct-view display, like those on most cell phones or PDAs in use in 2001 , for viewing at normal reading distance of approximately 12 to 24 inches (which this patent will refer to as "arms'-length" viewing). And at least one of the displays is a microdisplay, a tiny display with magnifying optical elements, for viewing larger, higher-resolution images when the microdisplay is positioned close to the eye. The microdisplays can be embedded in devices, or they can be attached to the devices, or the microdisplays can be in wearable displays attached to the devices (wirelessly or through a cable). Many digital video cameras use microdisplays for their view-finders: When a user brings the view-finder to their eye when the camera is playing back video, the user can watch a larger image of the video-clip than could be seen on small direct-view display. In recent years, high resolution microdisplays have been developed that can present images large enough to display most Web pages. For example, Inviso Inc.'s 3430 Optiscape II microdisplay displays an 800x600 pixel color image that appears to users to be as large as a 19-inch monitor roughly 2.5 feet away. Since optics are used to create an image that looks large, despite the fact that the microdisplay module is small, these images are often called "virtual images". And in this patent we will sometimes call them "apparent images". Preferred embodiments of this invention would normally be implemented on devices using microdisplays that can display color images at least 800 pixels wide and at least 600 pixels tall. An important common characteristic of the preferred embodiments of the present invention is that they are designed to facilitate a seamless, natural transition between near-to-eye viewing of content on a device's microdisplay (or microdisplays) and arms'-length viewing of content on a direct-view display. Key to enabling this seamless transition is software, hardware and storage mechanisms that are used to intelligently coordinate images on the microdisplay with images on the direct-view display. Using the descriptions in this patent and the figures for this patent (which help illustrate some of the graphical user interface elements of the invention disclosed herein), ordinary hardware and firmware engineers skilled in the art of developing computing and communication devices will be able to implement the hardware, and storage inventions disclosed herein, and ordinary programmers skilled in the art of software engineering on hand-held devices will be able to write their own code to implement the software inventions disclosed herein for devices that embody the hardware and storage inventions disclosed herein.
A preferred embodiment of my invention is a "smart-phone" supporting functions such as instant messaging, email, and contact management, in addition to phone calling, with both a microdisplay and a direct-view display. Since most instant messages, for example, are relative short text messages, this smart-phone could display instant messages on the direct-view display so that users can see the messages at arm's length, and can concurrently see the keys, buttons and controls on the phone handset so the user can easily type replies to the messages. When an instant message includes a reference to a Web page URL, or includes an attached photo or other relatively large high- resolution image or document, then when the user chooses to view that Web page or attachment, the smart-phone can display the Web page or attachment in the microdisplay instead of the direct-view display. This allows the user to see the Web page or attachment, which may be relatively large and high-resolution, by bringing the microdisplay to the user's eye, much like a PC display like image.
Additionally, when the device is displaying an image (or is preparing to display an image) in one display but not the others, then the device can display a text or graphics indication on one or more of the displays that is not displaying that image, indicating that one of the other displays is displaying that image (or that the device is preparing to display that image). For example, if the user sees an instant message in the direct-view display (on a device with one direct-view display and one microdisplay), and that instant message refers to a Web page URL, if the user chooses to view that Web page, then while the device is downloading or preparing the Web page for display on the microdisplay the device can display a flashing eye icon on the direct-view display; and when the Web page is ready for viewing on the microdisplay, then the device can display a static icon on the direct-view display. This lets the user know that the user can now bring the device near-to-eye to view the Web page in a high resolution, magnified form. The same technique can be used any time the user can operate functions using the direct-view display that result in content then appearing on the microdisplay, or vice-versa. (If the user has the device near-to-eye and operates a function that results in new content appearing on the direct-view display, the microdisplay can display some indication that there is new content to view on the direct-view display.) It is important to provide these cues, since a user typically can not see the content on a microdisplay when holding the device at arms-length, and a user typically can not see content on a direct-view display (or doesn't think to look at the direct-view display) when looking into a microdisplay that is near-to-eye.
Device designers can choose to use the direct-view display as the primary display for many smart- phone functions, including instant messaging, email, phone dialing, contact management, schedule management, calculator functions, and the like, and to use the microdisplay as a high resolution viewer coordinated with the direct-view display.
In the emerging age of Internet-enabled mobile devices, some of the most important embodiments of this invention involve coordinating microdisplays and direct-view displays in ways that allow people to comfortably access and interact with full-page Web content on pocket-size devices. One important way our invention does this is as follows: When a user is viewing a Web page (or other content) on a device's microdisplay held near-to-eye, the device should allow the user to position a cursor or a rectangular outline (or some other indication of a "region of interest") on a particular part of the Web page, and then when the user moves the device out to arms'-length the user should be able to view that region of interest on the direct-view display - that is, view a subset of the larger image that appeared on the microdisplay. This patent will refer to this subset as a "region of interest subset" or a "region of interest".
When the direct-view display displays this region of interest, the direct-view display acts as a window onto the larger image that the user had been viewing on the microdisplay. That is, the region of interest portion of the image seen on the microdisplay is mirrored on the direct-view display. The user should then be able to read text or interact with content in that region of interest, or use controls on the device to scroll the region of interest left, right, up, down, or other directions, to see different parts of the image other than the subset of the image currently shown in the direct-view display. If the user moves the region of interest while viewing the image in the direct-view display, then when the user returns to viewing the image in the microdisplay, the device can graphically indicate where the region of interest has moved - for example, by overlaying a rectangle in the shape of the region of interest (which will generally correspond to the pixel width and height of the direct-view display) on top of the image seen in the microdisplay. Device makers can choose not to explicitly show a rectangle or other explicit representation of the region of interest on the image viewed in the microdisplay, or choose to make this kind of explicit "region of interesf indicator optional for the user. Whether or not the region of interest is explicitly outlined on the image viewed in the microdisplay, the device can mirror the region of interest in the direct-view display, as described above. Also note that the region of interest does not necessarily have to be rectangular. For example, a device designer could choose to use a circular or other shaped direct-view display (for example, on a watch or a non-conventional cell-phone or PDA), and then choose to use a correspondingly shaped region of interest. The usage model facilitated by this invention mirrors what happens when a person uses a desktop computer monitor: Typically a person briefly "takes in" the entire monitor display (i.e., looks at the entire image without focusing attention on any specific region of the display at first), and then focuses attention on a specific region of interest (such as text they want to read, or a text-insertion point where they want to type, or some other item of interest). This invention allows microdisplay-enhanced devices to model this behavior.
As a result, this invention enables better ergonomics and better power-consumption characteristics than previous designs for many types of microdisplay-enhanced devices. With this invention, a user can choose to briefly observe a large area of content displayed on the microdisplay, and quickly skim the large area of content to identify a region of interest where they want to focus attention - and then (if they want) they can bring the device to arms'-length to view and interact with that region of interest on the direct-view display. By allowing the user to view a large area of content on the microdisplay, users can quickly orient themselves and quickly skim the large area of content to find a specific region of interest - just as they do on desktop monitors. This is something users cannot do on the small direct-view displays available on pocket-sized phones, PDAs, and hand-held computers that do not include embedded or attached microdisplays. Then, by allowing users to move the region of interest on the direct-view display held at arms'-length, users can choose to do most of their reading, typing, and other interacting with content while holding the device at arms'-length (i.e. normal reading distance). Being able to frequently operate the device and interact with Web pages and other content while holding the device at arms'-length is more comfortable for many people than having to hold a device near-to-eye for long periods. And it is much easier to type text on a device held at arms'-length than a device held near-to-eye. In addition, when the user is not holding the device near-to-eye, the microdisplay can be turned off or idled, saving power- so devices that use this invention can use significantly less power than devices that leave microdisplays on continuously. Similarly, when the user is holding the device near-to-eye, direct-view displays on the device can be turned off or dimmed or idled, saving power.
Clearly this method of intelligently coordinating content on microdisplays with content on direct-view displays can be applied to a wide range of content in addition to Web pages - including content such as spreadsheets, presentations, word processor files, blueprints, video, database forms, and other types of images, documents and applications.
The primary motivitation for this invention has to make it feasible for people to comfortably access and interact with full size Web content (and other high-resolution or large visual content) on pocket-size devices with built-in or attachable microdisplays. But the present invention's embodiments related to intelligently coordinating direct-view displays and microdisplays are applicable even when the microdisplays are part of a wearable display connected to the device rather than being embedded in or attached directly to the device. Wearable displays can include one or two microdisplays. For example, Inviso Inc.'s binocular wearable display product concept called "eShades" included two microdisplays, one for each eye. If a user is wearing a wearable display and looking at the near-to-eye microdisplays of the wearable display, they may find it awkward to type, particularly if the user has to look at a device's keypad frequently to find the letter keys while typing. A monocular wearable display (i.e. one with just one microdisplay positioned in front of one eye) lets the user use the other eye to look at the keypad; and many binocular wearable displays (like Inviso Inc.'s eShades concept product) allow users to see below the displays and therefore see a device's keypad held below the field of view of the wearable displays. But it is awkward to have to continuously shift ones eyes between the keypad and the image presented by the microdisplay while typing. Most users would be more comfortable if the region surrounding the current text-entry point (which might be an editable text field on a Web-based form, for example) were mirrored in the direct-view display of the device the user is using to type, ideally with the direct-view display close to the keypad on the device. This allows the user to keep their eyes focused in a small area on the device while typing.
BRIEF DESCRIPTION OF THE DRAWINGS:
Fig. 1 A is a side view of a microdisplay-enhanced handset.
Fig. 1 B is a front view of a microdisplay-enhanced handset.
Fig. 1C is a front view of a microdisplay-enhanced handset, held in a person's hand.
Fig. 2A is a side view of a flip-style handset with both a direct-view display and a microdisplay.
Fig. 2B is a front view of a flip-style handset with both a direct-view display and a microdisplay.
Fig. 2C is a front view of a flip-style handset with both a direct-view display and a microdisplay, held in a person's hand.
Fig. 3A is a side view of an alternative configuration for a handset with both a microdisplay and a direct-view display.
Fig. 3B is a front view of an alternative configuration for a handset with both a microdisplay and a direct-view display.
Fig. 3C is a front view of an alternative configuration for a handset with both a microdisplay and a direct-view display, held in a person's hand.
Fig. 4 is an illustration of a handset with two displays and with buttons for highlighting and selecting selectable items on a list of selectable items displayed on one of the displays. Fig. 5 is an illustration of a handset according to my invention, showing the coordination of images on the direct-view display and on the microdisplay.
Fig. 6 is an illustration of a handset according to my invention, showing an indication on the direct- view display (shown in close up on the left of Fig. 6) that there is new content ready for viewing on the handset's microdisplay.
DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS:
The following describes a handset on which my invention can be embodied. Fig. 1 A shows a microdisplay-enhanced handset 118 in side-view. Fig. 1B shows the same in front-view. Fig. 1C shows the same in front-view, shown held in a person's hand. This handset includes a microdisplay 100 positioned above a direct-view display 101. Most of the elements of this handset could be applied to a wide range of devices - including cellular phones, "Smart-phones", Web-phones, PDAs, hand-held computers, remote controls, and others. Throughout the rest of this patent, most references to "handsets" can refer to any of these types of devices.
The direct-view display can be any kind of direct-view display, but preferably it would be a thin bit-map LCD display with less resolution than the microdisplay but enough to render an identifiable portion of a larger image. For example, a 160x120-pixel direct-view display with 3-bits of gray scale (white, black, and six intensities of gray in-between white and black) would be sufficient for displaying text and simple graphics while still keeping the cost of the direct-view display relatively low.
Below the side-view of Fig. 1 A and the first front-view illustrations of the handset in Fig. 1 B are side- view 116 and front-view 117 illustrations of Inviso's 3430 Optiscape microdisplay module which includes electronics and optics - just to illustrate their size relative to the overall handset. Many other elements on this handset are common on cell phone handsets today such as the speaker 114, a power on/off button 108, a microphone 115, "Start Call" and. "End Call" buttons 112.
Handset users frequently interact with various functions supported on the handset, such as instant messaging, email, contact management, dialing from a list of contacts, calculator, Web browsing, and so on. To interact with those functions, users need to see (or hear) feedback as they use the functions. For example, when typing an instant message, a user needs to see the text the user is typing to know that it 's being typed correctly. For most functions, including instant messaging, the feedback presented by the handset is relatively simple and brief - capable of being expressed in a few lines of text or in a small low-resolution bit-mapped display. On handsets built according to this invention, relatively simpler or brief feedback should usually be presented on the direct-view display 100 of Fig. 1B, as is done on a traditional handset that has only a direct-view display. This allows users to see the feedback while holding the handset at arms' length (as opposed to near-to-eye) so they can still easily see and operate the handset's buttons and other controls. But when a function involves viewing or interacting with a large image, such as a Web page, a spreadsheet, or a high- resolution document attached to an email, then a handset built according to this invention should display that image on the microdisplay 101 of Fig. 1B.
While different handsets can differ in what they display on the two displays (the microdisplay 101 and the direct-view display 100), in general designers should strive to allow users to interact with the handset's functions while viewing the direct-view display. That makes it easier to see and operate more of the device's buttons and controls (e. g. for typing or for navigating through the functions' features) as compared to holding the device near-to-eye. Of course, designers can also allow users to interact with functions while holding the device near-to-eye and using the microdisplay by designing the handset so it displays some feedback on the microdisplay (instead of, or in addition to, displaying that feedback on the direct-view display).
With a one-piece handset such as the one illustrated in Fig. 1 , there are two advantages to placing the microdisplay near the top of the handset and placing the direct-view display in the middle of the phone (above the keypad). First, it makes it easier for the user to bring the microdisplay close to the eye, as compared to placing the microdisplay in the middle or at the bottom of the handset. Second, it places the direct-view display (where feedback for most text being typed would be presented) near the keyboard (so the user does not have to move their eyes far from the typing keys to the display showing the feedback).
The device can also be designed such that depressing a given button on the device (or depressing a combination of one or more buttons or controls) results in the device popping up and displaying (on one or more displays) a menu in a menu bar associated with the device or with the front window on the device (such as the "File" menu at the top left of documents on most PC documents) while concurrently highlighting the top menu item in that menu. Then depressing left or right arrow buttons on the device, such as those in control 113 on Fig. 1 , could closes the currently popped-up menu and causes popping up the previous or next menu in the menu bar and highlights the top item on the menu. Further, the device could be designed such that further depressing up or down arrows on the device (such as those in control 113 on Fig. 1) highlights the next or previous menu item in the currently popped up menu. At that point, pressing an "enter" or "select" button on the phone results in the highlighted menu item being selected. Different types of items can do different things when selected in this way. Typically, selecting an item should cause the same behavior as would occur on a PC if the user used a mouse to click on that type of item. For example, selecting a check box that is not currently checked should cause it to be checked; and selecting a check box that is currently checked should cause it to be unchecked.
The device can further be designed such that spoken commands via speech recognition software can be used to pop up and traverse menus or to select menu items; and in which the device's speech recognition features are enabled only when the user holds in a button designed for that function on the device. This "push-to-talk" button could be on the side of the device where a user could depress it with a finger or thumb on the hand holding the device.
Fig. 2A through Fig 2C illustrate a flip-style handset with both a direct-view display 200 and a microdisplay 201. Microdisplays are relatively thick (because of their magnifying optical elements) and direct-view displays are relatively thin (since they are lower resolution and do not require magnifying optical elements). So a flip-style phone would place the direct-view display on the thin part of the phone that flips away from the main body of the phone. In Fig. 2B, this is the top part of the phone. A flip-style phone could also be designed in which the keyboard and the direct-view display are on a relatively thin bottom part ofthe phone, and the microdisplay is on a relatively thick top part of the phone that flips up. But that would tend to make the phone feel top-heavy.
Fig. 3A through Fig. 3C illustrate a compelling alternative configuration for a handset with both a microdisplay 300 and a direct-view display 301 as seen in the side-view representation of Fig. 3A. In this embodiment, the direct-view display is placed directly over the microdisplay. As noted above, for most functions (placing a phone call, scrolling through lists of contacts, and instant messaging, for example) the user simply uses the direct-view display for feedback, since the feedback consists of relatively low-resolution text or images. In these instances, when the user looks at the front of the phone 305 in Fig. 3B, he or she see the relatively low-resolution text or images on the direct-view display 302. But when the user decides to view an image or document that is best viewed on fullscreen, high resolution displays (Web pages, excel spread-sheets, large photos, etc.) then the handset would display that image or document on the microdisplay. A handset can be designed to have a button (or other control) that the user uses to make the direct-view display become transparent (and to turn the microdisplay on) so the user can see through to the microdisplay. But a preferred embodiment would include a distance-sensor 304 in Fig. 3B that automatically senses when the user's eye is close to the microdisplay (within 2-3 inches) and then automatically makes the direct- view display (302 in Fig. 3A and Fig. 3B) transparent and turns on the microdisplay (300 in Fig. 3A and 303 in Fig. 3C) so the user can see through the direct-view display 301 to the image presented by the microdisplay 300. Various distance-sensing technologies - including ultrasonic distance sensors - are used in current digital video cameras, still cameras, and other electronic devices. For example, the Sony Mavica MVC CD1000 digital still camera has both a direct-view display on its back and a separate mid-resolution microdisplay in its viewfmder, and it includes a distance sensor so that when a user puts their eye close to the viewfmder the microdisplay-based viewfmder automatically turns on. This handset of Fig. 3A through Fig. 3C would automatically turn on the microdisplay only when the handset is displaying a high-resolution image and when the sensor senses that the user's eye is close enough to the microdisplay - within three inches or so. At all other times, the handset would automatically turn the microdisplay off, saving considerable power relative to a device that leaves the microdisplay on all the time or one that leaves it on even when the user's eye is not close to the microdisplay. As used herein, the phrase "turn the microdisplay off" means to deactivate the microdisplay such that it is not generating images or generating any light that would prevent the correct operation of the direct-view display. It does not necessarily mean that no power is being supplied to the microdisplay. However, the result is substantial power savings.
This handset of Fig. 3A through Fig. 3C can also be built less tall than handsets that separate the two displays (as in Fig. 1A through Fig. 1C), since the two displays overlap one another in this design, although it could make the phones a bit thicker when viewed from the side.
A handset could also be designed with a direct-view display on one side of the handset (such as the front) and a microdisplay on the other side of the handset (such as the back). But this would require users to turn the handset over when switching between viewing the direct-view display and viewing the microdisplay - which is more awkward than simply bringing the handset to the eye to view an image in the microdisplay.
On any of these dual-display devices, either or both of the displays could be touch sensitive (like many PDAs today), allowing users to input data and control aspects of the device by moving their fingers, a pen, or a stylus across the touch-sensitive area of the display(s).
Each of the handsets illustrated in Fig. 1 A through Fig. 3C is shown with three side-buttons on the right side -"Shift", "2nd Letter", and "3rd Letter" side-buttons - which a user can press with the hand holding the phone while simultaneously using the other hand to type the handset's face-keys. For example, to type the lowercase letter "m", the user would just type the "6" key. To type the uppercase letter "M", the user would type the "6" key while holding in the "Shift" side-button. To type the lowercase letter "u", the user would type the "8" key while holding in the "2nd Letter" side-button. To type the upper-case letter "U", the user would type the "8" key while holding in both the "Shift" side-button and the "2nd Letter" side-button. This can allow for faster text-typing than on current cell phones and PDAs. This is important for typing-intensive applications and functions such as instant messaging, email, interacting with some Web pages, and other applications that involve entering information into forms or documents. That fast-typing-enabling invention is the subject of patent application S/N , filed , assigned to the common assignee and incorporated herein by reference.
The device can also include a roller control 102 of Fig. 1 B which, when rolled in one direction, changes the highlighting in a displayed list of selectable items from the currently highlighted item to the next selectable item; and, when rolled in the other direction, changes the highlighting from the currently highlighted item to the previous selectable item. This is similar to using the Tab-forward and Tab-back keys on a Windows PC to move the highlighted item on a Web-page, dialog or other document - often called the "focus" by programmers - from one selectable item to the next one or to the previous one. Note the "list" of selectable items does not have to appear linear: A web-page has a "list" of selectable items, as seen in the Web-page's HTML, even if those items are arranged nonlinearly in a two-dimensional design. In fact, a "selectable" item can be any displayed item that can either be selected or, in the case of a text-entry box, typed into. The device can also be designed to apply to items that are part of Web pages and other documents, as well as to items that are displayed by the handset itself such as the list of functions available on the phone device. In all these cases, a roller control can also be used to traverse through lists of items and to select items. The roller control 102 can also be pressed to select the highlighted item in any displayed list of items.
As seen in Fig. 4, the device includes a button 401 that, when pressed briefly, changes the highlighting in a displayed list of selectable items from one selectable item to the previous one. This patent also refers to the highlighted selectable item as the "in-focus" item, the "item in focus", or the "focus" item. The device can also include a button 403 that, when pressed briefly, changes the highlighted item in a displayed list of selectable items from one selectable item to the next one. The device can also include a button 402 that, when pressed, selects the highlighted item in the displayed list of items. These buttons do not have to be configured as buttons 401 , 402, and 403 appearing in Fig. 4: They could be configured as arrow keys and other buttons elsewhere on the device - such as the arrows and enter buttons 113 on Fig. 1B. Each of these buttons can be multiple-speed (which this patent will also refer to as "multi-speed") - for example, moving the in-focus item quickly from the currently highlighted selectable item to the next or previous one when pressed hard, but moving the highlighting slowly when pressed lightly. These buttons can go by many names - such as Tab-forward and Tab-back, up and down, and so on. This patent will refer to them generically as up and down or up/down. If they are multiple-speed, this patent will refer to them as multi-speed up/down controls. The buttons can be placed under a shared length of plastic to form a toggle switch, with one button under one side of the toggle and the other button under the other side of the toggle, as shown in Fig. 5.
As additional examples of operation of the above buttons, when the button 401 (or 403) of Fig. 4 is pressed and held for more than a defined period of time (for example, more than 0.5 second), then instead of the highlighting moving from one selectable item to the previous (or next) item just once, the highlighting moves to the previous (or next) selectable item, stays briefly (e. g. 0.25 sec), moves to the previous-to- the-previous (or next-to-the-next) selectable item, stays briefly, and continues moving back (or forward) through the list of selectable items for as long as the user continues holding the corresponding button 401 (or 403). This is simply an alternative to continuously rolling the roller control outlined above to quickly move the highlighting hack and forth through the list of selectable items.
The device can also be designed such that the longer the button 401 or 403 is held, the faster the highlighting moves through the list of selectable items, up to some maximum speed (and with the speed accelerating in well defined increments from relatively slow to relatively fast).
I turn now to a description of my invention embodied in a handset of the type described above. My invention can be embodied in a handset or other hand-held device with two or more displays, one of which can show content (such as an image, document, Web page, text, or other content) that includes one or more references to other content, such as referring to an image, document, Web page, text or other content. When a user indicates the desire to view referred-to content, for example, by using a button or control to interact with the first display, such as the direct-view display, and to click on, select, or open the referred-to content, then that referred-to content is displayed on one of the other displays, such as the microdisplay, for high resolution PC display-like viewing. So a user can use the direct-view display of the device to read an email or instant message that has an attached photo or a reference to a Web page URL and when the user clicks on or selects that attachment or Web page URL, the handset displays the attachment or Web page URL in the microdisplay. In addition, the device could show a full Web page (or a large portion of a Web page) in a microdisplay, and then show a subset of that Web page in a direct-view display. When the microdisplay is displaying a document or image (such as a Web page or application form) with items that can be selected or typed in, then at any given time, zero or one of those items is "in focus" and that item is visually highlighted in some way, for example, by outlining that item in light gray. "In focus" means that if the user were to use a "select" control, by whatever name, on the handset or were to begin typing, then that "in focus" item would be the item selected, or, if it were a text entry box, it would be the item into which the characters would be typed. This is similar to the way items can be in "focus" in Web browsers or other applications on a desktop Windows PC. The item in "focus" is usually highlighted graphically in a desktop application, for example with a light gray outline. So if the microdisplay is displaying a document or image (such as a Web page or application form) and if one of the items on the document or image is "in focus," then a subset of the document or image surrounding the "in focus" item can be displayed on the direct-view display. This can provide a close-up view of the item that would be typed in or selected if the user were to type or use the handset's selection control. The user could see this area while holding the device a foot or two away, since it is displayed in the direct-view display. And this in turn would allow the user to see the buttons and controls on the device more easily as compared to holding the handset near-to-eye. This would allow the user to more easily type or use the controls on the phone while viewing that subset of the overall web-page in the direct-view display. This is best seen in Fig. 5. In that Figure the microdisplay 500 displays the web page while the subset thereof is displayed on the direct-view display 501.
One way to implement this is to provide that the subset shown on the direct-view display would include both the "in focus" item and a text or bit-map representation of any text or graphic labels or symbols corresponding to that item on the form. For example, if the "in focus" item is a "Last name:" text-entry box on a Web-page form, then the direct-view display would show the label "Last name:" as well as the corresponding text-entry box, into which the user will type his or her last name. In practice, the item "in focus" (and hence the image subset shown on the direct-view display) can be changed by the user typing, depressing buttons, or using other controls on the device. For example, the user can use the roller control 102 in Fig. 1, or down/up buttons on Fig. 4, on the phone handset to move the focus back and forth through the list of items that can be selected or typed into on the current Web page or document. As the user uses the controls to change the focus, the direct-view display should change to reflect the new focus, by showing the "region of interest" subset of the Web page or document surrounding that new in-focus item (or surrounding as much of the focus fits in the region of interest). This makes it easy for the user to always see the item that would be selected or typed-in, at any given moment, if the user were to select an item or type. Fig. 5 illustrates this concept. Fig. 5 is explained below. Additionally, the item in focus (and hence the image subset shown on the direct-view display) can be changed by the user speaking into the device, with the device itself, or a service accessed by the device over a network, using speech recognition software to interpret the user's spoken commands. Further still, the item in focus (and hence the image subset shown on the direct-view display) can be changed by software running on the computer processor within the device. This could be used to generate animations by having a large image made up of many smaller frames, and rapidly changing which subset of the larger image is displayed in the direct-view display.
The item in focus can also be changed by the user typing, pressing buttons, speaking, or using other controls on the device and, as the focus changes, the image shown in the microdisplay auto-scrolls to keep the in-focus item in view. This is similar to what happens if a user views a large Web-page form on a computer monitor and the form extends past the bottom of the window in which the form is being displayed. That is, if the user uses the Tab button repeatedly to move the focus toward the end of the form, then when the focus goes past the bottom of the window to an item that can be typed-in or selected that appears on the part of the form below the window, the Web browser will typically auto- scroll the form upward within the window so that the new in-focus item will be visible within the window. In my invention, when a user types into a text-entry field shown on one display, and that text- entry field is also shown on another display (sometimes called "echoing"), then new text that appears in the first display will also appear in the other display reflecting the fact that both displays are showing the same text-entry field. For example, if the user is wearing a wearable display attached to a hand-held device with a direct-view display, then if the user types text into a field that is being shown in both the direct-view display and in the wearable display, the new text typed will appear on both displays - so the user won't be confused as they shift their eyes back and forth between the direct-view display and the wearable display's microdisplays.
For example, if a user changes the state of a check-box, radio-button, or other "state" control that appears in more than one display, then all displays showing that same state control should show the state control in its current state. If a user types into a text-entry field shown on one or more displays, or uses arrow keys 113 of Fig. 1 or other controls to change the text-entry point or selection area, then if the text-entry point (or the end of the selection area) moves out of the subset of the document image visible in any given display, that display will automatically scroll the image so that the text-entry point (or the end of the selection area) will reappear in that display. This is similar to how documents automatically scroll in windows on desktop displays, except since this device may display different subsets of the same document in different displays, it must auto-scroll the image in different displays at different times. For example, if a relatively low-resolution direct-view display shows a small subset of a document shown in the much higher-resolution microdisplay, then the direct-view display may have to auto-scroll several times as the user types a long message into that document, while the microdisplay could auto-scroll much less often.
In contexts where the microdisplay is displaying a relatively high-resolution large apparent image (i.e. a virtual image, as discussed earlier), if the device uses a relatively low-resolution bit-map direct-view display, then at any given time the direct-view display can show a subset of the larger, higher- resolution apparent image shown in the microdisplay. That is, the direct-view display acts as a window into the larger image, showing a "region of interest" from the larger apparent image.
Further, controls on the hand-held device, such as arrow buttons or a thumb roller control or similar controls, can move the region of interest window (i. e. the borders of the subset of the Web Page, application document, or other image shown in the direct-view display) down/up right/left or in other directions across the larger image. For example, at startup, the direct-view display might show an upper-left corner subset of the image seen in the microdisplay, which itself might be a subset of a larger scrollable document or Web page. Then the user can use the controls to have the direct-view show a different subset of the image seen in the microdisplay.
Further, as the user types into a document or Web Page visible in the microdisplay, the subset of the document or Web Page that includes the text-entry point can automatically be scrolled into view in both the direct-view display and the microdisplay. That is, the "region of interest" can be automatically moved to surround the text-entry point. Usually, the region of interest shown in the direct-view display is an area of bits around the current "focus" shown in the larger image displayed on the microdisplay. For example, a Web page has a set of selectable items such as links, buttons, editable text fields, and the like. As seen on desktop PCs, at any given time, one of the selectable items can be in "focus", and users can move the focus back and forth through the set of selectable items using the Tab Backward and Tab Forward key. (On desktop PC keyboards these are usually the same key: Tab-Backward = shift Tab Forward). Users can select that selectable item that is currently in focus by depressing a button (usually the space bar). As discussed in patent application S/N , a manufacturer can design a cell phone handset or other hand-held device on which a user could use a thumb-roller or other control to move the focus back and forth through the list of selectable items on the Web page or screen. As the user does this, if the Web Page or application document being viewed is larger than the image area visible in the microdisplay, then the image of the Web page or application document shown in the microdisplay should auto-scroll to always keep the focus in view in the microdisplay, just as occurs on desktop PCs when user tabs through selectable items (using the Tab key) on a large Web page displayed in a smaller Web browser window. Similarly, as the focus is moved, if the Web Page or application document being viewed is larger than the image area visible in the direct-view display, then the image of the Web page or application document shown in the direct- view display should auto-scroll to always keep the focus in view in the direct-view display. In a preferred embodiment, in accordance with the present invention, each display (when viewed) would show both the focus item (i. e. the selectable item that would get selected if the user hits the select key) and the area immediately surrounding the focus item (which will often include text or image labels corresponding to the selectable item in focus, such as "Last Name:" labels). Each display thus acts as a window onto the Web Page or Document being viewed, with the direct-view display being a smaller window than the microdisplay.
In contexts where the microdisplay is displaying a relatively high-resolution large apparent image of a Web Page, if the device uses a text-only (i. e., non-bit-map) direct-view display, then at any given time the direct-view display can show text corresponding to the item in focus on the Web Page. For example, if the item in focus is a text-entry field in a form on a Web page in which a user is supposed to type their last name, then the direct-view display can show any text immediately to the left (and/or right) of that text entry field (e. g. "Last Name: "). This text can often be extracted from the
HTML that defines the Web Page (by interpreting the HTML to determine what text lies to the left of the selectable item in focus). But often it will not be possible to extract appropriate text corresponding to the Focus item (e. g. because no text corresponding to the focus item exists). For this reason, a preferred embodiment of this device would use a bit-map direct-view display, as discussed above, instead of a text-only direct-view display.
As noted earlier, an important way this invention allows users to comfortably interact with full-size Web pages and other high-resolution or large content on pocket-size devices is by intelligently coordinating the content of direct-view displays with the content of microdisplays. Here are descriptions of a couple important embodiments of this invention.
Referring to Fig. 5, when a user is viewing a Web page 502 on a device's microdisplay 500 held near- to-eye, the device should allow the user to position a cursor (such as the arrow cursor commonly seen when moving the mouse on a desktop computer) or a rectangular outline 503 or both (or some other indication of a "region of interest") over any chosen part of the Web page 502. Then when the user moves the device out to arms'-length the user should be able to view that "region of interest" on a direct-view display 501 on the device - i.e. viewing a subset of the larger image that appeared on the microdisplay.
In Fig. 5, the lines extending from the left corners of the Web page window 502 to the phone handset's embedded microdisplay 500 are meant to show that, when the microdisplay is held near the user's eye, the user could see a relatively large, high-resolution image like the Web page window 502. And the longer narrower lines extending from the "region of interest" rectangle 503 to the handset's direct-view display 501 are meant to show that a small portion of the Web page could be made visible on the direct-view display. In this example, the current region of interest is a small form on the Web page where the user could enter an address to look up a map.
When the direct-view display 501 displays this region of interest (the area within the rectangle 503), the direct-view display acts as a window onto the larger image that would be seen on the microdisplay if the microdisplay were held near-to-eye. While viewing the direct-view display 501 at arms'-length (that is, at a normal reading distance), a user should be able to read text or interact with content in that region of interest, or use controls on the device to scroll the region of interest left, right, up, down, or other directions, to see different parts of the image other than the subset of the image currently shown in the direct-view display. If the user moves the region of interest while viewing the image in the direct-view display, then when the user returns to viewing the image in the microdisplay, the device can graphically indicate where the region of interest has moved - for example, by overlaying a rectangle 503 in the shape and size of the region of interest on top of the image seen in the microdisplay. The shape and size of the region of interest should generally be the same as, or close to, the pixel width and height of the direct-view display 501.
As illustrated in Fig. 5, the direct-view display 501 will typically have substantially fewer pixels than the microdisplay. For example, today's smartphones typically have displays that are less than 200 pixels wide and less than 250 pixels tall, and today's PDAs typically have displays that are less than 300 pixels wide and less than 400 pixels tall. By contrast, Inviso Inc.'s current microdisplays are capable of displaying 800x600 pixel areas. The size and shape of the direct-view display can be used to define the size and shape of the region of interest. As the user uses controls on the device (or speech recognition, or other means) to move the region of interest around while viewing content on the microdisplay, the device can show the location of the region of interest by displaying a rectangle 503 of that size on the image being viewed in the microdisplay 500. As the user moves the region of interest, the rectangle moves around the image. If the user then views the region of interest in the direct-view display 501 at arms'-length, and then uses controls on the device (or other means) to scroll the region of interest to other parts of the content being viewed (as described earlier), then when the user brings the device near-to-eye to view the content in the microdisplay, the rectangle showing the region of interest should be shown on the microdisplay in its new location on the image being viewed.
Note that using a rectangle to show the region of interest is just one option. The device could use a range of techniques to indicate the region of interest on the microdisplay. Or device makers can choose not to explicitly mark the region of interest on the microdisplay image - choosing instead just to treat the area surrounding the cursor or surrounding highlighted "in focus" areas (discussed elsewhere in this patent) as the region of interest. Alternatively, if the region of interest is marked on the microdisplay (with a rectangle, for example), then device makers can give users the option of hiding that mark or showing that mark.
As discussed above, it is sometimes useful to display an icon or other visual cue on a direct-view display when there is new content for the user to look at in the microdisplay, and vice-versa, since it can be hard to see content on the microdisplay when looking at the direct-view display, and vice- versa. Fig. 6 illustrates a preferred embodiment for providing this visual cue. The sample phone on the right side of Fig. 6 has a microdisplay 604 near the top and a direct-view display 600 below the microdisplay. The direct-view display 601 is shown in close-up on the left side of Fig. 6. In this example, the user (Matt) has received an instant message. The user would typically read instant messages at arms'-length (i.e. normal reading distance) on a direct-view display 600 (shown in close up on the left 601). The instant message in this example includes a URL 603 referring to a Web page with a map (similar to the Web page shown in Fig. 5.) In this preferred embodiment, when the user operates controls on the phone to open the Web page associated with the URL, the device will display a flashing eye icon 602 on the direct-view display while the device is downloading and preparing that Web page for display. And when the Web page is ready for viewing on the microdisplay 604, the device would display a steady (i.e. non-flashing) eye icon 602. The form of the indication is not important - it could be a different icon, or text, or some other visual cue, and it could flash or not-flash. The important point is that, when a user is holding a device at arms'-length, there are times when it is difficult for them to know when there have been significant changes to the content available for viewing on the near-to-eye microdisplay (since it can be difficult to see content on a microdisplay at arms'-length), so at those times, it is useful to provide visual cues on the direct-view display. Similarly, if the user is viewing content on the microdisplay, in some cases it could be useful to provide visual cues on the microdisplay when there is significant new content available for viewing on the direct-view display, since the user may not think to look at the direct-view display while looking at the microdisplay (even if the microdisplay is on a head-mounted wearable display rather than embedded in the hand-held device with the direct-view display).
In the foregoing discussion, we have used Web pages as our primary example of content that can be accessed more easily by intelligently coordinating the images displayed on a device's direct-view display and microdisplay. But it will be appreciated by those skilled in the art that the same techniques can be applied to a wide range of content, including computer applications, documents (including spreadsheets, presentations, and word-processor documents, among others), photos, video, and more.
The present invention can be used when designing devices with a wide range of form-factors and configurations. For example, a device with a direct-view display on its face and a microdisplay embedded near the top of the device works well. But device makers could choose to embed or attach one or more microdisplays to any part of the device (e.g. on a hinge on the bottom of the phone or on the side) and still use this invention to intelligently coordinate content on the device's multiple displays and thereby improve the ergonomics of the device.
While the foregoing has been with reference to particular embodiments of the invention, it will be appreciated by those skilled in the art that changes in these embodiments may be made without departing from the principles and spirit of the invention, the scope of which is defined by the appended claims.

Claims

I CLAIM:
1. A phone handset or other hand-held device which displays images, said device having a face, said face having two or more displays for displaying said images and control elements for controlling said displays, wherein a user can operate said control elements to affect visual changes on one or more of the displays.
2. The devices of claim 1 in which at least one display is a direct-view display.
3. The device of claim 1 in which at least one of the displays is a microdisplay.
4. The device of claim 1 in which the device has exactly two displays.
5. The device of claim 4 in which the device includes one direct-view display and one microdisplay.
6. The device of claims 1 or 2 or 3 or 4 or 5 that operates on functions transmitted remotely from said device.
7. The device of claim 6 wherein such functions are selected from the group consisting of instant messaging, email, contact management, schedule management and phone calling.
8. The device of claim 6 wherein such at least one of said functions is displayed in the direct- view display.
9. The device of claim 8 wherein the displayed function includes an attachment that can be displayed in the microdisplay.
10. The device of claim 9 further including a distance sensor wherein the user affects displaying said attachment in the microdisplay by bringing the microdisplay to the user's eye.
11. The device of claim 1 in which, when an image is displayed in one display but not the others, then the device can display a text or graphics indication on one or more of the displays that is not displaying that image, said indication indicating that one of the other displays is displaying that image.
12. The device of claim 11 in which at least one display is a direct-view display.
13. The device of claim 11 in which at least one of the displays is a microdisplay.
14. The device of claim 11 in which the device has exactly two displays.
15. The device of claim 14 in which the device includes one direct-view display and one microdisplay.
16. The device of claims 9 or 10 in which (1) the displayed function is taken from the group of functions consisting of instant messaging, email, contact management, schedule maintenance, calculation functions and placing telephone calls, viewing Internet content, and viewing other types of documents, (2) the attachment is an attachment taken from the group of attachments consisting of a Web page URL, a photo image, a spreadsheet, a presentation, an Adobe Acrobat file, a Word Processor document, and other types of content, and (3) the controls are capable of operation by a user to display that attachment in the microdisplay.
17. The device of claim 16 further including control elements which, when activated, cause the device to display an indication in the direct-view display that indicates to the user when all or part of said attachment is being downloaded or prepared for display on the microdisplay, or cause the device to display an indication in the direct-view display that indicates when the attachment is ready for viewing on the microdisplay.
18. A phone handset or other hand-held device with two or more displays, said device having controls operable by a user to (1) cause one of the displays to display first content that includes one or more references to other content, and (2) cause that referred-to content to be displayed in one of the other displays.
19. The device of claim 18 wherein said first content is content selected from the group consisting of an image, a document, a Web page or text and the referred content is content selected from the group consisting of an image, a document, a Web page, text or other content.
20. The device of claim 18 in which at least one display is a direct-view display.
21. The device of claim 18 in which at least one of the displays is a microdisplay.
22. The device of claim 18 in which the device has exactly two displays.
23. The device of claim 22 in which the device includes one direct-view display and one microdisplay.
24. A phone handset or other hand-held device which displays images, said device having two or more displays for displaying said images and control elements for controlling said displays, wherein a user can operate said control elements to cause a first display to display an image and cause a second display to display a region of interest subset of said image.
25. The device of claim 24 wherein the device can display substantially a full Web page in one display and display a close up view of one subset of that Web page in another of the displays.
26. The device of claim 24 in which at least one display is a direct-view display.
27. The device of claim 24 in which at least one of the displays is a microdisplay.
28. The device of claim 24 in which the device has exactly two displays.
29. The device of claim 24 in which the device includes one direct-view display and one microdisplay.
30. The device of claim 23 in which, when a first display is displaying first content having items that can be selected or typed in, one of such items is in-focus.
31. The device of claim 30 wherein said one in-focus item is highlighted by outlining that item in a visually recognizable shade.
32. The device of claim 30 having a select function whereby a user may select said one item.
33. The device of claim 32 further including one or more face-keys which, when depressed by a user, causes a character to be typed into said one item if said one item is an editable text field.
34. The device of claim 30 further including controls operable by a user that cause, if the device has prepared an image having one item in-focus for display in the microdisplay, a subset of said image surrounding the in-focus item to be displayed in the direct-view display.
35. The device of claim 34 in which the subset displayed in the direct-view display is a close up view of said one item.
36. The device of claim 34 in which the subset displayed in the direct-view display includes both the in-focus item and a text or bit-map representation of text or graphic labels or symbols corresponding to that item.
37. The device of claim 36 in which the in-focus item is a "Last Name" text entry box on a Web- page form and the subset includes the label "Last Name" and the corresponding text-entry box into which the last name will be typed.
38. The device of claim 34 including controls operable by a user to cause the subset displayed in the direct-view display to be changed by user operation.
39. The device of claim 36 in which the device includes one control element from the group consisting of a roller control or up/down buttons or multi-speed up/down controls capable of operation by the user to cause the in-focus item to change from the currently in-focus item to the previous or next item that can be selected or typed in.
40. The device of claim 39 including software that causes the image displayed in the direct-view display to concurrently change when the in-focus item changes from one selectable item to a new in-focus item.
41. The device of claim 40 in which said image is changed by displaying a subset of the image surrounding all or part of said new in-focus item.
42. The device of claim 34 including a microphone and voice recognition software that can allow a user to cause the in-focus item to change from one selectable item to another one by the user speaking into the device.
43. The device of claim 34 including a computer processor in which both said item in-focus in said microdisplay and said image subset displayed in the direct-view display can be changed by software running on said computer processor within the device.
44. The device of claim 43 wherein said change generates animations by having a large image made up of many smaller frames, and rapidly changing which subset of the larger image is displayed in the direct-view display.
45. The device of claim 43 in which the item in-focus can be changed by the user typing, depressing buttons, speaking, or using other controls on the device, and in which, as the focus changes, if the microdisplay is displaying an image, the image shown in the microdisplay auto-scrolls to keep the in-focus item in view, and, if the direct-view display is displaying an image, the image shown in the direct-view display auto-scrolls to keep the in- focus item in view.
46. The device of claim 24 having a plurality of face-keys capable of being depressed by a user to generate new text into a text-entry field displayed in a first display, which text-entry field is also shown on a second display such that the new text that appears in said first display will also appear in said second display.
47. The device of claim 24 including at least one control button capable of being depressed by a user to generate a changed state of a check box, radio-button, or other state control that appears in more than one display, and concurrently changing to said changed state all displays showing that same check box, radio-button or other state control.
48. The device of claim 24 capable of displaying a subset of said image in at least one of said displays and further having at least one of the group consisting of (1) a plurality of face-keys capable of being depressed by a user to generate text typed into a text-entry field shown on said one or more displays, and (2) arrow keys or other controls which, when actuated by a user, change said text-entry point, wherein if the text-entry point moves out of the subset of said image visible in any given display, then that display will automatically scroll the image so that the text-entry point will reappear in that display.
49. The device of claim 29 wherein the microdisplay displays a relatively high-resolution large apparent image, the direct-view display is a low-resolution bit-map direct-view display, said direct view display acting as a window into said larger apparent image.
50. The device of claim 29 wherein the first display is a microdisplay and the second display is a direct-view display, said device further including controls operable by a user to move the borders of said region of interest subset displayed in the direct-view display down or up or right or left or other directions across the image displayed in the microdisplay.
51. The device of claim 50 further including face-keys for entering text, the image is a document or Web Page and, as the user operates said face-keys to enter text into said document or Web page, the subset of the document or Web Page that includes the text-entry point is automatically scrolled into view in both the direct view display and the microdisplay.
52. The device of claim 51 in which the subset displayed in the direct-view display is an area of bits containing all or part of the current in-focus area of the image displayed in the microdisplay.
53. A computer implemented process generating codes for interpreting the operation of control elements on a phone handset or other hand-held device which displays images, said phone handset or other device having (1) a face, said face having two or more displays for displaying said images and (2) control elements operable by a user for controlling said displays, said codes effecting visual changes on one or more of the displays in response to such operation.
54. The process of claim 53 in which at least one display is a direct-view display.
55. The process of claim 54 in which at least one of the displays is a microdisplay
56. The process of claim 55 in which the device has exactly two displays.
57. The process of claim 56 in which the device includes one direct-view display and one microdisplay.
58. The process of claims 53 or 54 or 55 or 56 or 57 that operates on functions transmitted remotely from said device.
59. The process of claim 58 wherein such functions are selected from the group consisting of instant messaging, email, contact management, schedule management and phone calling.
60. The process of claim 58 wherein such at least one of said functions is displayed in the direct- view display.
61. The process of claim 60 wherein the displayed function includes an attachment that can be displayed in the microdisplay.
62. The process of claim 61 wherein said control elements include a distance sensor, said codes effecting displaying said attachment in the microdisplay in response to the user bringing said distance sensor to the user's eye.
63. The process of claim 53 in which, in response to an image being displayed in one display but not at least one of the others, the codes effect displaying a text or graphics indication on one or more of the displays that is not displaying that image, said indication indicating that one of the other displays is displaying that image.
64. The process of claim 61 or 62 in which (1) the function is taken from the group of functions consisting of instant messaging, email, contact management, schedule maintenance, calculation functions, placing telephone calls, viewing Internet content, and viewing other types of documents, (2) the attachment is an attachment taken from the group of attachments consisting of a Web page URL or a photo image, a spreadsheet, a presentation, an Adobe Acrobat file, a Word Processor document, and other types of content, and (3) the control elements include first controls capable of operation by a user to display that attachment in the microdisplay, said codes effecting such display of such attachment in response to operation of said second controls.
65. The process of claim 64 wherein said control elements includes second control elements which, when activated by the user, display an indication in the direct-view display that indicates to the user that the attachment has been opened and is being displayed in the microdisplay and said codes effecting such display of such indication in response to said activation.
66. A computer implemented process generating codes for interpreting the operation of control elements on a phone handset or other hand-held device having two or more displays, said phone handset or other device having first control elements operable by a user to (1) cause a first display to display first content that includes one or more references to other content, and (2) to cause that referred-to content to be displayed in a second display, and said codes effecting each said displays in response to operation of said controls.
67. The process of claim 66 wherein said first content is content selected from the group consisting of an image, a document, a Web page, text or other content, and the referred-to content is content selected from the group consisting of an image, a document, a Web page, text or other content.
68. The process of claim 67 in which the device includes one direct-view display and one microdisplay.
69. A computer implemented process generating codes for interpreting the operation of control elements on a phone handset or other hand-held device which displays images, said phone handset or other device having (1) two or more displays for displaying said images and (2) control elements for controlling said displays, wherein a user can operate said control elements to cause a first display to display an image and cause a second display to display a region of interest subset of said image, and said codes effect displaying of said subset in response to the operation of said control elements.
70. The process of claim 69 wherein said control elements include first controls operable by a user to (1) display substantially a full Web page in one display and (2) display a close up view of one subset of that Web page in another of the displays, said codes effecting said displaying in said one display and said displaying in said another of the displays in response to the operation of said first controls.
71. The process of claim 69 in which the device includes one direct-view display and one microdisplay.
72. The process of claim 69 wherein said control elements include second controls operable by a user which, when a first display is displaying first content having items that can be selected or typed in, cause one of such items to be in-focus, said codes effecting said one of such items.
73. The process of claim 72 wherein, in response to the user operating said second controls, said codes effect said one in-focus item being highlighted by outlining that item in a visually recognizable shade.
74. The process of claim 73 wherein the device has a select function operable by a user to select said one in-focus item, said codes effecting selecting said one item into in response to the operation of said select function.
75. The process of claim 74 wherein the device further includes one or more face-keys which, when depressed by a user, causes a character to be typed into said one item if said one item is an editable text field, said codes effecting said character being so typed in response to the user depressing said one or more face-keys.
76. The process of claim 69 wherein said control elements include third controls operable by a user to cause, if the device has prepared an image having one item in-focus for display in the microdisplay,, a subset of said image surrounding the in-focus item to be displayed in the direct-view display, said codes effecting said subset of said image being so displayed in response to the operation of said third controls.
77. The process of claim 76 in which the subset displayed in the direct-view display is a close up view of said one item.
78. The process of claim 76 in which the subset displayed in the direct-view display includes both the in-focus item and a text or bit-map representation of text or graphic labels or symbols corresponding to that item.
79. The process of claim 76 in which the in-focus item is a "Last Name" text entry box on a Web- page form and the subset includes the label "Last Name" and the corresponding text-entry box into which the last name will be typed.
80. The process of claim 76 wherein the control elements include fourth controls operable by a user to cause the subset displayed in the direct-view display to be changed by user operation, said codes effecting said change in response to said user operation.
81. The process of claim 78 in which the device includes at least one control element from the group consisting of a roller control or up/down buttons or multi-speed up/down controls capable of operation by the user to cause the in-focus item to change from the current in- focus item to the previous or next item that can be selected or typed in.
82. The process of claim 81 wherein the process includes software that causes the image displayed in the direct-view display to concurrently change when the in-focus item changes from one selectable item to a new in-focus item.
83. The process of claim 82 in which said image is changed by displaying a subset of the image surrounding all or part of said new in-focus item.
84. The process of claim 62 in which said phone handset or other hand-held device further includes a menu button and, if a user depresses said menu button while said distance sensor senses that the microdisplay on said device is close to an object, then the device will display a menu on the microdisplay, and otherwise the device will display said menu on the direct- view display.
85. The process of claim 76 wherein the device includes a microphone and voice recognition software that allows a user to cause the in-focus item to be changed from one selectable item to another one by the user speaking into the device, said codes effecting said change in response to the user so speaking.
86. The process of claim 76 wherein the device includes a computer processor in which both said item in-focus in said microdisplay and said image subset displayed in the direct-view display can be changed by software running on said computer processor within the device.
87. The process of claim 85 wherein said change generates animations by having a large image made up of many smaller frames, and rapidly changing which subset of the larger image is displayed in the direct-view display.
88. The process of claim 69 in which the item in-focus can be changed by the user typing, depressing buttons, speaking, or using other controls on the device, and in which, as the focus changes, if the microdisplay is displaying an image, the image shown in the microdisplay auto-scrolls to keep the in-focus item in view, and, if the direct-view display is displaying an image, the image shown in the direct-view display auto-scrolls to keep the in- focus item in view.
89. The process of claim 69 wherein the device has a plurality of face-keys capable of being depressed by a user to generate new text into a text-entry field displayed in a first display, which text-entry field is also shown in a second display such that the new text that appears in said first display will also appear in said second display, said codes effecting said new text appearing in said first display an in said second display in response to the user so depressing said face keys.
90. The process of claim 69 wherein the device includes at least one control button capable of being depressed by a user to generate a changed state of a check box, radio-button, or other state control that appears in more than one display, and concurrently changing to said changed state all displays showing that same check box, radio-button or other state control.
91. The process of claim 69 in which the device is capable of displaying a subset of said image in at least one of said displays and further having at least one of the group consisting of (1) a plurality of face-keys capable of being depressed by a user to generate text typed into a text- entry field shown on said one or more displays, and (2) arrow keys or other controls which, when actuated by a user, change said text-entry point, wherein if the text-entry point moves out of the subset of said image visible in any given display, then said codes effect said any given display automatically scrolling the image so that the text-entry point will reappear in said any given display.
92. The process of claim 71 wherein the microdisplay displays a relatively high-resolution large apparent image, the direct-view display is a low-resolution bit-map direct-view display, said direct-view display acting as a window into said larger apparent image.
93. The process of claim 69 wherein a first display is a microdisplay and a second display is a direct-view display, and a subset of said control elements are operable by a user to move the borders of said region of interest subset displayed in the direct-view display down or up or right or left or other directions across the image displayed in the microdisplay, said codes effecting said movement of said borders in response to the operation of said subset of said control elements.
94. The process of claim 92 wherein the device further includes face-keys operable by a user to enter text, the image is a document or Web Page having a text entry point, said codes effecting the subset of the document or Web Page that includes the text-entry point being automatically scrolled into view in both the direct-view display and the microdisplay in response to the user operating said face-keys to enter text into said document or Web page.
95. The process of claim 92 in which the subset displayed in the direct-view display is an area of bits containing all or part of the current in-focus area of the image displayed in the microdisplay.
96. The process of claim 63 wherein the device includes one direct view display and one microdisplay.
97. A storage medium having embedded therein a computer implemented process generating codes for interpreting the operation of control elements on a phone handset or other hand- held device which displays images, said phone handset or other device having (1) a face, said face having two or more displays for displaying said images and (2) control elements operable by a user for controlling said displays, said code effecting visual changes on one or more of the displays in response to such operation.
98. The storage medium of claim 97 in which at least one display is a direct-view display.
99. The storage medium of claim 98 in which at least one of the displays is a microdisplay
100. The storage medium of claim 99 in which the device has exactly two displays.
101. The storage medium of claim 100 in which the device includes one direct-view display and one microdisplay.
102. The storage medium of claims 97 or 98 or 99 or 100 or 101 that operates on functions transmitted remotely from said device.
103. The storage medium of claim 102 wherein such functions are selected from the group consisting of instant messaging, email, contact management, schedule management and phone calling.
104. The storage medium of claim 102 wherein such at least one of said functions is displayed in the direct-view display.
105. The storage medium of claim 104 wherein the displayed function includes an attachment that can be displayed in the microdisplay.
106. The storage medium of claim 105 wherein said control elements include a distance sensor, said codes effecting displaying said attachment in the microdisplay in response to the user bringing said distance sensor to the user's eye.
107. The storage medium of claim 97 in which, in response to an image being displayed in one display but not at least one of the others, the codes effect displaying a text or graphics indication on one or more of the displays that is not displaying that image, said indication indicating that one of the other displays is displaying that image.
108. The storage medium of claim 105 or 106 in which (1) the function is taken from the group of functions consisting of instant messaging, email, contact management, schedule maintenance, calculation functions, placing telephone calls, viewing Internet content, and viewing other types of documents, (2) the attachment is an attachment taken from the group of attachments consisting of a Web page URL or a photo image, a spreadsheet, a presentation, an Adobe Acrobat file, a Word Processor document, and other types of content, and (3) the control elements include first controls capable of operation by a user to display that attachment in the microdisplay, said codes effecting such display of such attachment in response to operation of said second controls.
109. The storage medium of claim 108 wherein said control elements includes second control elements which, when activated by the user, display an indication in the direct-view display that indicates to the user that the attachment has been opened and is being displayed in the microdisplay and said codes effecting such display of such indication in response to said activation.
110. A storage medium having embedded therein a computer implemented process generating codes for interpreting the operation of control elements on a phone handset or other handheld device having two or more displays, said phone handset or other device having first control elements operable by a user to (1) cause a first display to display first content that includes one or more references to other content, and (2) to cause that referred-to content to be displayed in a second display, and said codes effecting each said displays in response to operation of said controls.
111. The storage medium of claim 110 wherein said first content is content selected from the group consisting of an image, a document, a Web page or text and the referred-to content is content selected from the group consisting of an image, a document, a Web page, text or other content.
112. The storage medium of claim 111 in which the device includes one direct-view display and one microdisplay.
113. A storage medium having embedded therein a computer implemented process generating codes for interpreting the operation of control elements on a phone handset or other handheld device which displays images, said phone handset or other device having (1) two or more displays for displaying said images and (2) control elements for controlling said displays, wherein a user can operate said control elements to cause a first display to display an image and cause a second display to display a region of interest subset of said image, and said codes effect displaying of said subset in response to the operation of said control elements.
114. The storage medium of claim 113 wherein said control elements include first controls operable by a user to (1 ) display substantially a full Web page in one display and (2) display a close up view of one subset of that Web page in another of the displays, said codes effecting said displaying in said one display and said displaying in said another of the displays in response to the operation of said first controls.
115. The storage medium of claim 113 in which the device includes one direct-view display and one microdisplay.
116. The storage medium of claim 113 wherein said control elements include second controls operable by a user which, when a first display is displaying first content having items that can be selected or typed in, cause one of such items to be in-focus, said codes effecting said one of such items.
117. The storage medium of claim 116 wherein, in response to the user operating said second controls, said codes effect said one in-focus item being highlighted by outlining that item in a visually recognizable shade.
118. The storage medium of claim 117 wherein the device has a select function operable by a user to select said one in-focus item, said codes effecting selecting said one item in response to the operation of said select function.
119. The storage medium of claim 118 wherein the device further includes one or more face-keys which, when depressed by a user, causes a character to be typed into said one item if said one item is an editable text field, said codes effecting said character being so typed in response to the user depressing said one or more face-keys.
120. The storage medium of claim 113 wherein said control elements include third controls operable by a user to cause, if the device has prepared an image having one item in-focus for display in the microdisplay, a subset of said image surrounding the in-focus item to be displayed in the direct-view display, said codes effecting said subset of said image being so displayed in response to the operation of said third controls.
121. The storage medium of claim 120 in which the subset displayed in the direct-view display is a close up view of said one item.
122. The storage medium of claim 120 in which the subset displayed in the direct-view display includes both the in-focus item and a text or bit-map representation of text or graphic labels or symbols corresponding to that item.
123. The storage medium of claim 120 in which the in-focus item is a "Last Name" text entry box on a Web-page form and the subset includes the label "Last Name" and the corresponding text-entry box into which the last name will be typed.
124. The storage medium of claim 120 wherein the control elements include fourth controls operable by a user to cause the subset displayed in the direct-view display to be changed by user operation, said codes effecting said change in response to said user operation.
125. The storage medium of claim 122 in which the device includes at least one control element from the group consisting of a roller control or up/down buttons or multi-speed up/down controls capable of operation by the user to cause the in-focus item to change from the current in-focus item to the previous or next item that can be selected or typed in.
126. The storage medium of claim 125 wherein the process includes software that causes the image displayed in the direct-view display to concurrently change when the in-focus item changes from one selectable item to a new in-focus item.
127. The storage medium of claim 126 in which said image is changed by displaying a subset of the image surrounding all or part of said new in-focus item.
128. The storage medium of claim 120 wherein the device includes a microphone and voice recognition software that allows a user to cause the item in-focus to be changed from one selectable item to another one by the user speaking into the device, said codes effecting said change in response to the user so speaking.
129. The storage medium of claim 120 wherein the device includes a computer processor in which both said item in-focus in said microdisplay and said image subset displayed in the direct- view display can be changed by software running on said computer processor within the device.
130. The storage medium of claim 129 wherein said change generates animations by having a large image made up of many smaller frames, and rapidly changing which subset of the larger image is displayed in the direct-view display.
131. The storage medium of claim 113 in which the item in-focus can be changed by the user typing, depressing buttons, speaking, or using other controls on the device, and in which, as the focus changes, if the microdisplay is displaying an image, the image shown in the microdisplay auto-scrolls to keep the in-focus item in view, and, if the direct-view display is displaying an image, the image shown in the direct-view display auto-scrolls to keep the in- focus item in view.
132. The storage medium of claim 113 wherein the device has a plurality of face-keys capable of being depressed by a user to generate new text into a text-entry field displayed in a first display, which text-entry field is also shown in a second display such that the new text that appears in said first display will also appear in said second display, said codes effecting said new text appearing in said first display an in said second display in response to the user so depressing said face keys.
133. The storage medium of claim 113 wherein the device includes at least one control button capable of being depressed by a user to generate a changed state of a check box, radio- button, or other state control that appears in more than one display, and concurrently changing to said changed state all displays showing that same check box, radio-button or other state control.
134. The storage medium of claim 113 in which the device is capable of displaying a subset of said image in at least one of said displays and further having at least one of the group consisting of (1) a plurality of face-keys capable of being depressed by a user to generate text typed into a text-entry field shown on said one or more displays, and (2) arrow keys or other controls which, when actuated by a user, change said text-entry point, wherein if the text-entry point moves out of the subset of said image visible in any given display, then said codes effect said any given display automatically scrolling the image so that the text-entry point will reappear in said any given display.
135. The storage medium of claim 115 wherein the microdisplay displays a relatively high- resolution large apparent image, the direct-view display is a low-resolution bit-map direct-view display, said direct-view display acting as a window into said larger apparent image.
136. The storage medium of claim 113 wherein a first display is a microdisplay and a second display is a direct-view display, and a subset of said control elements are operable by a user to move the borders of said region of interest subset displayed in the direct-view display down or up or right or left or other directions across the image displayed in the microdisplay, said codes effecting said movement of said borders in response to the operation of said subset of said control elements.
137. The storage medium of claim 136 wherein the device further includes face-keys operable by a user to enter text, the image is a document or Web Page having a text entry point, said codes effecting the subset of the document or Web Page that includes the text-entry point being automatically scrolled into view in both the direct-view display and the microdisplay in response to the user operating said face-keys to enter text into said document or Web page.
138. The storage medium of claim 136 in which the subset displayed in the direct-view display is an area of bits containing all or part of the current in-focus area of the image displayed in the microdisplay.
139. The storage medium of claim 107 wherein the device has one direct view display and one microdisplay.
140. The device of claim 1 or claim 24, wherein said device is a phone, smartphone, PDA, handheld computer, or watch.
141. The device of claim 1 or claim 24, wherein the device can receive and transmit data wirelessly.
142. The device of claim 1 or claim 24, wherein the device includes a microdisplay embedded in or attached near the top of the face of the device, a keypad or touchpad near the bottom of the face of the device, and a direct-view display visible on the face of the device between the microdisplay and the touchpad.
143. The device of claim 1 or claim 24, wherein the device includes a first part including a direct- view display, a second part including a microdisplay and a keypad or touchpad, and a hinge connecting said first part to said second part, wherein said microdisplay is between said hinge and said keypad or touchpad.
PCT/US2002/010458 2001-04-02 2002-04-02 Coordinating images displayed on devices with two or more displays WO2002079913A2 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
EP02719425A EP1391044A2 (en) 2001-04-02 2002-04-02 Coordinating images displayed on devices with two or more displays
AU2002250508A AU2002250508A1 (en) 2001-04-02 2002-04-02 Coordinating images displayed on devices with two or more displays
KR10-2003-7013029A KR20030097820A (en) 2001-04-02 2002-04-02 Coordinating images displayed on devices with two or more displays
CA002441301A CA2441301A1 (en) 2001-04-02 2002-04-02 Coordinating images displayed on devices with two or more displays
JP2002577682A JP2004535699A (en) 2001-04-02 2002-04-02 Adjusting the image displayed on a device with two or more displays

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US28079101P 2001-04-02 2001-04-02
US60/280,791 2001-04-02
US10/113,239 2002-04-01
US10/113,239 US7016704B2 (en) 2001-04-02 2002-04-01 Coordinating images displayed on devices with two or more displays

Publications (2)

Publication Number Publication Date
WO2002079913A2 true WO2002079913A2 (en) 2002-10-10
WO2002079913A3 WO2002079913A3 (en) 2003-04-17

Family

ID=26810836

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2002/010458 WO2002079913A2 (en) 2001-04-02 2002-04-02 Coordinating images displayed on devices with two or more displays

Country Status (8)

Country Link
US (1) US7016704B2 (en)
EP (1) EP1391044A2 (en)
JP (1) JP2004535699A (en)
KR (1) KR20030097820A (en)
CN (1) CN1533637A (en)
AU (1) AU2002250508A1 (en)
CA (1) CA2441301A1 (en)
WO (1) WO2002079913A2 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100487618B1 (en) * 2002-12-11 2005-05-03 엘지전자 주식회사 Mobile phone having various display apparatuses
WO2005062591A1 (en) * 2003-12-02 2005-07-07 Sony Ericsson Mobile Communications Ab System and method for audible control of a camera associated with a mobile phone
WO2006123294A2 (en) * 2005-05-19 2006-11-23 Koninklijke Philips Electronics, N.V. Apparatus and method to enhance navigation in a user interface for mobile devices
WO2007105508A1 (en) * 2006-03-09 2007-09-20 Casio Computer Co., Ltd. Screen display control apparatus and program product
CN1627765B (en) * 2003-12-10 2010-09-01 松下电器产业株式会社 Portable information terminal device
US7904513B2 (en) 2006-08-31 2011-03-08 Casio Computer Co., Ltd. Client apparatus, server apparatus, server-based computing system, and program
US8620997B2 (en) 2009-03-24 2013-12-31 Casio Computer Co., Ltd Client apparatus, computer system, computer readable program storage medium and display method, each for detecting change of display contents in status bar area to display the change

Families Citing this family (211)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE19741453A1 (en) * 1997-09-19 1999-03-25 Packing Gmbh Agentur Fuer Desi Digital book, esp. for reproducing textual information
US7369161B2 (en) * 1999-06-08 2008-05-06 Lightsurf Technologies, Inc. Digital camera device providing improved methodology for rapidly taking successive pictures
US8212893B2 (en) 1999-06-08 2012-07-03 Verisign, Inc. Digital camera device and methodology for distributed processing and wireless transmission of digital images
US7372485B1 (en) 1999-06-08 2008-05-13 Lightsurf Technologies, Inc. Digital camera device and methodology for distributed processing and wireless transmission of digital images
US7103357B2 (en) 1999-11-05 2006-09-05 Lightsurf Technologies, Inc. Media spooler system and methodology providing efficient transmission of media content from wireless devices
US7847833B2 (en) * 2001-02-07 2010-12-07 Verisign, Inc. Digital camera device providing improved methodology for rapidly taking successive pictures
US7047033B2 (en) * 2000-02-01 2006-05-16 Infogin Ltd Methods and apparatus for analyzing, processing and formatting network information such as web-pages
US20020016818A1 (en) * 2000-05-11 2002-02-07 Shekhar Kirani System and methodology for optimizing delivery of email attachments for disparate devices
JP2001333163A (en) * 2000-05-22 2001-11-30 Nec Corp Mobile phone and its communication method
US7526440B2 (en) * 2000-06-12 2009-04-28 Walker Digital, Llc Method, computer product, and apparatus for facilitating the provision of opinions to a shopper from a panel of peers
US7636079B2 (en) * 2000-11-29 2009-12-22 Palm Inc. Application access and activation system and method
US7305354B2 (en) 2001-03-20 2007-12-04 Lightsurf,Technologies, Inc. Media asset management system
JP2002305570A (en) * 2001-04-04 2002-10-18 Nec Corp Mobile phone
FI20010817A (en) * 2001-04-20 2003-02-14 Nokia Corp A method for displaying information on an electronic device display and an electronic device
US6907447B1 (en) * 2001-04-30 2005-06-14 Microsoft Corporation Method and apparatus for providing an instant message notification
US7219309B2 (en) 2001-05-02 2007-05-15 Bitstream Inc. Innovations for the display of web pages
EP1393190A4 (en) * 2001-05-02 2007-07-18 Bitstream Inc Methods and systems for displaying media in scaled manner and/or in orientation
US20020183099A1 (en) * 2001-05-30 2002-12-05 Lee Man Wei Multiple display panels for different modes of operation for conveying personality
ATE470897T1 (en) * 2001-06-12 2010-06-15 Research In Motion Ltd PORTABLE ELECTRONIC DEVICE WITH KEYBOARD
KR100382938B1 (en) * 2001-06-21 2003-05-09 엘지전자 주식회사 Folder type mobile phone added sub-display
JP3894546B2 (en) * 2001-06-27 2007-03-22 キヤノン株式会社 Portable information terminal, information browsing method and program
US20030038928A1 (en) * 2001-08-27 2003-02-27 Alden Ray M. Remote image projector for hand held and wearable applications
US7804489B1 (en) * 2001-08-29 2010-09-28 Palmsource Inc. Method and apparatus for displaying information in a display screen region identified by permanent printing
US7120473B1 (en) * 2001-10-14 2006-10-10 Palm, Inc. Method and apparatus for controlling a mobile device by using a pivoting input switch
US20030160755A1 (en) * 2002-02-28 2003-08-28 Palm, Inc. Detachable expandable flexible display
US20030098857A1 (en) * 2001-11-28 2003-05-29 Palm, Inc. Detachable flexible and expandable display with touch sensor apparatus and method
US20070069975A1 (en) * 2001-11-28 2007-03-29 Palm, Inc. Detachable expandable flexible display
FR2833132B1 (en) * 2001-11-30 2004-02-13 Eastman Kodak Co METHOD FOR SELECTING AND SAVING A SUBJECT OF INTEREST IN A DIGITAL STILL IMAGE
US6822852B2 (en) * 2001-12-18 2004-11-23 Qualcomm, Incorporated Handheld devices
US7724281B2 (en) 2002-02-04 2010-05-25 Syniverse Icx Corporation Device facilitating efficient transfer of digital content from media capture device
US20030193418A1 (en) * 2002-04-10 2003-10-16 Xiaodong Shi Method and Apparatus To Input Text On Devices Requiring A Small Keypad
JP2004005492A (en) * 2002-04-26 2004-01-08 Casio Comput Co Ltd Data communication device, data communication system, animated document display method and animated document display program
US6879842B2 (en) * 2002-05-31 2005-04-12 Lavaflow, Llp Foldable wireless communication device functioning as a cellular telephone and a personal digital assistant
FR2842694B1 (en) * 2002-07-22 2005-02-11 France Telecom MOBILE COMMUNICATION DEVICE
JP3872052B2 (en) * 2002-10-30 2007-01-24 埼玉日本電気株式会社 Mobile phone with remote control function, remote control method and system thereof
US20040113927A1 (en) * 2002-12-11 2004-06-17 Sandie Quinn Device and method for displaying text of an electronic document of a screen in real-time
US20040203652A1 (en) * 2003-03-07 2004-10-14 Gao Yan Method and device for dtmf wireless handset based text messaging
US7538745B2 (en) * 2003-03-24 2009-05-26 Ntag Interactive Corporation Apparatus and method for enhancing face-to-face communication
US7623892B2 (en) 2003-04-02 2009-11-24 Palm, Inc. System and method for enabling a person to switch use of computing devices
US7383061B1 (en) 2003-04-02 2008-06-03 Palm, Inc. Peripheral device for a wireless communication device
KR100547741B1 (en) * 2003-05-13 2006-01-31 삼성전자주식회사 Speaker device of mobile terminal
US8090396B2 (en) * 2003-06-30 2012-01-03 Motorola Mobility, Inc. Push-to-talk features in wireless communications devices and methods
US7205959B2 (en) * 2003-09-09 2007-04-17 Sony Ericsson Mobile Communications Ab Multi-layered displays providing different focal lengths with optically shiftable viewing formats and terminals incorporating the same
US7536440B2 (en) * 2003-09-18 2009-05-19 Vulcan Portals Inc. Method and system for email synchronization for an electronic device
US7290034B2 (en) * 2003-09-18 2007-10-30 Vulcan Portals Inc. Method and system for polling a server for new emails, downloading the new emails in a background process, and caching the downloaded emails for access by an email application of an electronic device, such as a portable computer
US20050076085A1 (en) * 2003-09-18 2005-04-07 Vulcan Portals Inc. Method and system for managing email attachments for an electronic device
WO2005041029A2 (en) * 2003-10-24 2005-05-06 Matsushita Electric Industrial Co., Ltd. Remote operation system, communication apparatus remote control system and document inspection apparatus
US20050097046A1 (en) 2003-10-30 2005-05-05 Singfield Joy S. Wireless electronic check deposit scanning and cashing machine with web-based online account cash management computer application system
US7720924B2 (en) * 2003-12-12 2010-05-18 Syniverse Icx Corporation System providing methodology for the restoration of original media quality in messaging environments
AU2003288861A1 (en) * 2003-12-18 2005-07-05 Telefonaktiebolaget Lm Ericsson (Publ) Exchangeable module for additional functionality
KR100617702B1 (en) * 2004-01-13 2006-08-28 삼성전자주식회사 Portable terminal capable of editing image and image edition method using that
DE102004010710B4 (en) * 2004-03-04 2006-06-08 Siemens Ag Method for operating a mobile device for the projection of image data and mobile projector device
US20050250570A1 (en) * 2004-05-05 2005-11-10 Edwin Kuzniar Lottery tracking and display device
US7489953B2 (en) * 2004-06-02 2009-02-10 Research In Motion Limited Mobile communication device
US7669135B2 (en) * 2004-07-15 2010-02-23 At&T Mobility Ii Llc Using emoticons, such as for wireless devices
JP2006039919A (en) * 2004-07-27 2006-02-09 Pioneer Electronic Corp Image sharing display system, terminal with image sharing function, and computer program
US8127229B2 (en) * 2004-08-31 2012-02-28 Access Co., Ltd. Controlling focus, and movement of the focus, between focus items in a multi-frame document
KR100690726B1 (en) * 2004-09-14 2007-03-09 엘지전자 주식회사 Mobile phone capable of multitasking and control method thereof
US8169410B2 (en) * 2004-10-20 2012-05-01 Nintendo Co., Ltd. Gesture inputs for a portable display device
US7548748B2 (en) * 2004-12-13 2009-06-16 Sony Ericsson Mobile Communications Ab Programmable remote unit for operating a communication device
KR101122233B1 (en) 2004-12-23 2012-03-19 삼성전자주식회사 Display device including sensing element
JP2006186655A (en) * 2004-12-27 2006-07-13 Sony Corp Mobile terminal device
US7812786B2 (en) * 2005-01-18 2010-10-12 Nokia Corporation User interface for different displays
US20060183505A1 (en) * 2005-02-15 2006-08-17 Willrich Scott Consulting Group, Inc. Digital mobile planner
US20060195354A1 (en) * 2005-02-28 2006-08-31 Ntag Interactive Corporation Method of scoring the performance of attendees at a meeting
US20060206824A1 (en) * 2005-03-10 2006-09-14 Agere Systems Inc. Mobile communication device having scroll overlap adjustment capability and method of operation thereof
US8711063B2 (en) * 2005-03-11 2014-04-29 The Invention Science Fund I, Llc Self assembly of elements for displays
US8300007B2 (en) * 2005-03-11 2012-10-30 The Invention Science Fund I, Llc Self assembling display with substrate
US8334819B2 (en) * 2005-03-11 2012-12-18 The Invention Science Fund I, Llc Superimposed displays
US20060202944A1 (en) * 2005-03-11 2006-09-14 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Elements for self assembling displays
US7977130B2 (en) * 2006-08-03 2011-07-12 The Invention Science Fund I, Llc Method of assembling displays on substrates
US7990349B2 (en) * 2005-04-22 2011-08-02 The Invention Science Fund I, Llc Superimposed displays
US7662008B2 (en) * 2005-04-04 2010-02-16 Searete Llc Method of assembling displays on substrates
US8860635B2 (en) * 2005-04-04 2014-10-14 The Invention Science Fund I, Llc Self assembling display with substrate
US9153163B2 (en) * 2005-03-11 2015-10-06 The Invention Science Fund I, Llc Self assembly of elements for displays
US8390537B2 (en) * 2005-03-11 2013-03-05 The Invention Science Fund I, Llc Method of assembling displays on substrates
TWI263436B (en) * 2005-03-18 2006-10-01 Asustek Comp Inc Mobile phone with virtual keyboard
CN100419520C (en) * 2005-03-23 2008-09-17 李卫民 Display device
US20060276220A1 (en) * 2005-06-03 2006-12-07 Judith Schure Cell phone case with magnifier and method
TWI281811B (en) * 2005-07-01 2007-05-21 Wintek Corp Mobile communication device
US7831913B2 (en) * 2005-07-29 2010-11-09 Microsoft Corporation Selection-based item tagging
US7796837B2 (en) * 2005-09-22 2010-09-14 Google Inc. Processing an image map for display on computing device
US7569781B2 (en) * 2005-09-30 2009-08-04 Palm, Inc. Switch assembly having non-planar surface and activation areas
US8849821B2 (en) * 2005-11-04 2014-09-30 Nokia Corporation Scalable visual search system simplifying access to network and device functionality
CN1767561B (en) * 2005-11-09 2010-11-17 阮立山 Method for carrying out opening or amplification for font on handset screen
US7814425B1 (en) * 2005-12-30 2010-10-12 Aol Inc. Thumbnail image previews
US7843411B2 (en) * 2006-01-18 2010-11-30 Manning Ventures, Inc. Remote cholesteric display
US8082008B2 (en) * 2006-02-10 2011-12-20 Access Co., Ltd. User-interface and architecture for portable processing device
US8918450B2 (en) 2006-02-14 2014-12-23 Casio Computer Co., Ltd Server apparatuses, server control programs, and client apparatuses for a computer system in which created drawing data is transmitted to the client apparatuses
US20070198324A1 (en) * 2006-02-22 2007-08-23 Borovoy Richard D Enabling connections between and events attended by people
WO2007120399A2 (en) * 2006-02-24 2007-10-25 Verisign, Inc. System and method for managing distribution of multi-formatted content
US20080016462A1 (en) * 2006-03-01 2008-01-17 Wyler Eran S Methods and apparatus for enabling use of web content on various types of devices
US20070236334A1 (en) * 2006-03-31 2007-10-11 Borovoy Richard D Enhancing face-to-face communication
CN101079907B (en) * 2006-05-26 2011-11-30 鸿富锦精密工业(深圳)有限公司 Display device of mobile device and display method
US7782274B2 (en) 2006-06-09 2010-08-24 Cfph, Llc Folding multimedia display device
KR100793012B1 (en) 2006-06-14 2008-01-08 (주) 엘지텔레콤 Web browsing method of mobile communication terminal
BRPI0621981A2 (en) 2006-09-01 2011-12-20 Research In Motion Ltd Method and apparatus for controlling a display on an electronic device
US7855714B2 (en) 2006-09-01 2010-12-21 Research In Motion Limited Method and apparatus for controlling a display in an electronic device
US8775452B2 (en) * 2006-09-17 2014-07-08 Nokia Corporation Method, apparatus and computer program product for providing standard real world to virtual world links
KR101139455B1 (en) 2006-10-20 2012-04-30 엘지전자 주식회사 Mobile Communication System and Controlling Method for the Same
EP2080086A4 (en) * 2006-10-31 2011-06-08 Research In Motion Ltd Controlling display images on a mobile device
US8708227B1 (en) 2006-10-31 2014-04-29 United Services Automobile Association (Usaa) Systems and methods for remote deposit of checks
US8799147B1 (en) 2006-10-31 2014-08-05 United Services Automobile Association (Usaa) Systems and methods for remote deposit of negotiable instruments with non-payee institutions
US8351677B1 (en) 2006-10-31 2013-01-08 United Services Automobile Association (Usaa) Systems and methods for remote deposit of checks
US7885451B1 (en) 2006-10-31 2011-02-08 United Services Automobile Association (Usaa) Systems and methods for displaying negotiable instruments derived from various sources
US7876949B1 (en) 2006-10-31 2011-01-25 United Services Automobile Association Systems and methods for remote deposit of checks
US7873200B1 (en) 2006-10-31 2011-01-18 United Services Automobile Association (Usaa) Systems and methods for remote deposit of checks
CN101197044B (en) * 2006-12-06 2011-02-02 鸿富锦精密工业(深圳)有限公司 Image synthesis system and method
US20080148298A1 (en) * 2006-12-18 2008-06-19 Palm, Inc. System and Methods for Providing Granular Security for Locally Running Scripted Environments and Web Applications
US20080167082A1 (en) * 2007-01-10 2008-07-10 Vladimir Gurevich Incorporated indicating arrangement for a mobile device and related methods
US8107017B2 (en) * 2007-01-31 2012-01-31 Seiko Epson Corporation Combination display apparatus
US7738503B2 (en) * 2007-02-02 2010-06-15 Palm, Inc. Multi-way, peer-to-peer synchronization
US7715444B2 (en) * 2007-02-02 2010-05-11 Palm, Inc Resuming a previously interrupted peer-to-peer synchronization operation
US8959033B1 (en) 2007-03-15 2015-02-17 United Services Automobile Association (Usaa) Systems and methods for verification of remotely deposited checks
US10380559B1 (en) 2007-03-15 2019-08-13 United Services Automobile Association (Usaa) Systems and methods for check representment prevention
US20080248834A1 (en) * 2007-04-03 2008-10-09 Palm, Inc. System and methods for providing access to a desktop and applications of a mobile device
US8478299B2 (en) * 2007-04-06 2013-07-02 Hewlett-Packard Development Company, L.P. System and methods for obtaining coarse location for a mobile device
US8059097B2 (en) * 2007-04-27 2011-11-15 Virgin Mobile USA LP Shared symbol and emoticon key and methods
US8060486B2 (en) * 2007-05-07 2011-11-15 Hewlett-Packard Development Company, L.P. Automatic conversion schema for cached web requests
US8538124B1 (en) 2007-05-10 2013-09-17 United Services Auto Association (USAA) Systems and methods for real-time validation of check image quality
US8433127B1 (en) 2007-05-10 2013-04-30 United Services Automobile Association (Usaa) Systems and methods for real-time validation of check image quality
US20090021485A1 (en) * 2007-07-16 2009-01-22 Rochelle Shapera-Gill Ergonomic thumb support device for a hand-held device
US8458612B2 (en) * 2007-07-29 2013-06-04 Hewlett-Packard Development Company, L.P. Application management framework for web applications
US10091345B2 (en) * 2007-09-04 2018-10-02 Apple Inc. Media out interface
US9058512B1 (en) 2007-09-28 2015-06-16 United Services Automobile Association (Usaa) Systems and methods for digital signature detection
US9159101B1 (en) 2007-10-23 2015-10-13 United Services Automobile Association (Usaa) Image processing
US9892454B1 (en) 2007-10-23 2018-02-13 United Services Automobile Association (Usaa) Systems and methods for obtaining an image of a check to be deposited
US9898778B1 (en) 2007-10-23 2018-02-20 United Services Automobile Association (Usaa) Systems and methods for obtaining an image of a check to be deposited
US8358826B1 (en) 2007-10-23 2013-01-22 United Services Automobile Association (Usaa) Systems and methods for receiving and orienting an image of one or more checks
EP2180666A4 (en) * 2007-10-24 2012-04-04 Nec Corp Mobile terminal device and event notification method thereof
US7996314B1 (en) 2007-10-30 2011-08-09 United Services Automobile Association (Usaa) Systems and methods to modify a negotiable instrument
US7996315B1 (en) 2007-10-30 2011-08-09 United Services Automobile Association (Usaa) Systems and methods to modify a negotiable instrument
US7996316B1 (en) 2007-10-30 2011-08-09 United Services Automobile Association Systems and methods to modify a negotiable instrument
US8001051B1 (en) 2007-10-30 2011-08-16 United Services Automobile Association (Usaa) Systems and methods to modify a negotiable instrument
US8046301B1 (en) 2007-10-30 2011-10-25 United Services Automobile Association (Usaa) Systems and methods to modify a negotiable instrument
US8320657B1 (en) 2007-10-31 2012-11-27 United Services Automobile Association (Usaa) Systems and methods to use a digital camera to remotely deposit a negotiable instrument
US8290237B1 (en) 2007-10-31 2012-10-16 United Services Automobile Association (Usaa) Systems and methods to use a digital camera to remotely deposit a negotiable instrument
US7900822B1 (en) 2007-11-06 2011-03-08 United Services Automobile Association (Usaa) Systems, methods, and apparatus for receiving images of one or more checks
US7896232B1 (en) 2007-11-06 2011-03-01 United Services Automobile Association (Usaa) Systems, methods, and apparatus for receiving images of one or more checks
US20090150433A1 (en) * 2007-12-07 2009-06-11 Nokia Corporation Method, Apparatus and Computer Program Product for Using Media Content as Awareness Cues
US8108013B2 (en) * 2007-12-13 2012-01-31 Motorola Mobility, Inc. Systems and methods for managing power consumption in a flow-based user experience
TWI410824B (en) * 2007-12-18 2013-10-01 Htc Corp Electronic device with the module integrating display unit and input unit
US20090176481A1 (en) * 2008-01-04 2009-07-09 Palm, Inc. Providing Location-Based Services (LBS) Through Remote Display
KR20090077142A (en) * 2008-01-10 2009-07-15 삼성전자주식회사 Method and apparatus for displaying information used to choose input element
US10380562B1 (en) 2008-02-07 2019-08-13 United Services Automobile Association (Usaa) Systems and methods for mobile deposit of negotiable instruments
JP5028649B2 (en) * 2008-03-07 2012-09-19 富士通モバイルコミュニケーションズ株式会社 Electronics
US8351678B1 (en) 2008-06-11 2013-01-08 United Services Automobile Association (Usaa) Duplicate check detection
US20100024010A1 (en) * 2008-07-22 2010-01-28 Ernest Samuel Baugher Mobile device with customizable login screen
US8422758B1 (en) 2008-09-02 2013-04-16 United Services Automobile Association (Usaa) Systems and methods of check re-presentment deterrent
US10504185B1 (en) 2008-09-08 2019-12-10 United Services Automobile Association (Usaa) Systems and methods for live video financial deposit
DE102008048447A1 (en) * 2008-09-23 2010-04-29 Siemens Enterprise Communications Gmbh & Co. Kg Method and arrangement for the imaging of information, in particular for use in communication terminals
EP2169570A1 (en) * 2008-09-25 2010-03-31 Infogin LTD Mobile sites detection and handling
US7885880B1 (en) 2008-09-30 2011-02-08 United Services Automobile Association (Usaa) Atomic deposit transaction
US8957835B2 (en) 2008-09-30 2015-02-17 Apple Inc. Head-mounted display apparatus for retaining a portable electronic device with display
US7974899B1 (en) 2008-09-30 2011-07-05 United Services Automobile Association (Usaa) Atomic deposit transaction
US8275710B1 (en) 2008-09-30 2012-09-25 United Services Automobile Association (Usaa) Systems and methods for automatic bill pay enrollment
US7962411B1 (en) 2008-09-30 2011-06-14 United Services Automobile Association (Usaa) Atomic deposit transaction
US8391599B1 (en) 2008-10-17 2013-03-05 United Services Automobile Association (Usaa) Systems and methods for adaptive binarization of an image
US7970677B1 (en) * 2008-10-24 2011-06-28 United Services Automobile Association (Usaa) Systems and methods for financial deposits by electronic message
US7949587B1 (en) * 2008-10-24 2011-05-24 United States Automobile Association (USAA) Systems and methods for financial deposits by electronic message
JP4781440B2 (en) * 2009-02-05 2011-09-28 ソニー エリクソン モバイル コミュニケーションズ, エービー Image capturing apparatus, control method and control program for image capturing apparatus
US8452689B1 (en) 2009-02-18 2013-05-28 United Services Automobile Association (Usaa) Systems and methods of check detection
US10956728B1 (en) 2009-03-04 2021-03-23 United Services Automobile Association (Usaa) Systems and methods of check processing with background removal
US8023975B2 (en) * 2009-03-23 2011-09-20 T-Mobile Usa, Inc. Secondary status display for mobile device
US8224407B2 (en) * 2009-03-23 2012-07-17 T-Mobile Usa, Inc. Mobile device having a movable display and associated systems and methods
DE102009022222A1 (en) * 2009-05-20 2010-11-25 Giesecke & Devrient Gmbh Arrangement for the display of information, methods for displaying information and electronic terminal equipment
US8542921B1 (en) 2009-07-27 2013-09-24 United Services Automobile Association (Usaa) Systems and methods for remote deposit of negotiable instrument using brightness correction
US9779392B1 (en) 2009-08-19 2017-10-03 United Services Automobile Association (Usaa) Apparatuses, methods and systems for a publishing and subscribing platform of depositing negotiable instruments
US8977571B1 (en) 2009-08-21 2015-03-10 United Services Automobile Association (Usaa) Systems and methods for image monitoring of check during mobile deposit
US8699779B1 (en) 2009-08-28 2014-04-15 United Services Automobile Association (Usaa) Systems and methods for alignment of check during mobile deposit
JP2011060649A (en) * 2009-09-11 2011-03-24 Toyota Motor Corp Electrode active material layer, all solid battery, manufacturing method for electrode active material layer, and manufacturing method for all solid battery
US20110193681A1 (en) * 2009-10-12 2011-08-11 Rood Richard K Office Notification Systems and Methods
KR101613619B1 (en) * 2009-11-30 2016-04-29 삼성전자주식회사 A digital photographing apparatus, a method for controlling the same, and a computer-readable medium
JP5351006B2 (en) * 2009-12-24 2013-11-27 京セラ株式会社 Portable terminal and display control program
US20110231313A1 (en) * 2010-03-22 2011-09-22 Bank Of America Corporation Creating and Utilizing an Image Replacement System
US9129340B1 (en) 2010-06-08 2015-09-08 United Services Automobile Association (Usaa) Apparatuses, methods and systems for remote deposit capture with enhanced image detection
WO2011161966A1 (en) * 2010-06-25 2011-12-29 パナソニック株式会社 Control device and script conversion method
CN108681424B (en) 2010-10-01 2021-08-31 Z124 Dragging gestures on a user interface
US8842080B2 (en) 2010-10-01 2014-09-23 Z124 User interface with screen spanning icon morphing
US20120108298A1 (en) * 2010-10-29 2012-05-03 Symbol Technologies, Inc. Portable device having a virtual display
US20120256929A1 (en) * 2011-04-11 2012-10-11 Eric Koenig Expandable multi-core telecommunication and video processing apparatus
KR20130004713A (en) * 2011-07-04 2013-01-14 삼성전자주식회사 Interface apparatus and method of mobile communication terminal
US9588668B2 (en) * 2011-07-21 2017-03-07 Imerj, Llc Methods of displaying a second view
US9606723B2 (en) * 2011-07-21 2017-03-28 Z124 Second view
US9307206B2 (en) * 2011-08-16 2016-04-05 iParse, LLC Automatic image capture
US9495012B2 (en) 2011-09-27 2016-11-15 Z124 Secondary single screen mode activation through user interface activation
US8902125B1 (en) * 2011-09-29 2014-12-02 Rockwell Collins, Inc. Reconfigurable handheld device
US9766701B2 (en) * 2011-12-28 2017-09-19 Intel Corporation Display dimming in response to user
US9870752B2 (en) 2011-12-28 2018-01-16 Intel Corporation Display dimming in response to user
CN103999145B (en) * 2011-12-28 2017-05-17 英特尔公司 Display dimming in response to user
US10380565B1 (en) 2012-01-05 2019-08-13 United Services Automobile Association (Usaa) System and method for storefront bank deposits
US8760406B2 (en) * 2012-04-03 2014-06-24 Motorola Mobility Llc Edge to edge qwerty keypad for a handheld device
US9128562B2 (en) * 2012-05-25 2015-09-08 Sony Corporation Terminal apparatus, display system, display method, and recording medium for switching between pointer mode and touch-panel mode based on handheld activation
US9152618B2 (en) * 2012-08-31 2015-10-06 Microsoft Technology Licensing, Llc Cell view mode for outsized cells
US20150358452A1 (en) * 2012-10-05 2015-12-10 Vladimir Kranz Additive Equipment to Multimedial, Health and Sport Equipment
US10552810B1 (en) 2012-12-19 2020-02-04 United Services Automobile Association (Usaa) System and method for remote deposit of financial instruments
CN103064519A (en) * 2012-12-26 2013-04-24 珠海市魅族科技有限公司 Information content display control method and terminal
CN103902177B (en) * 2012-12-27 2018-10-12 联想(北京)有限公司 A kind of display control method and electronic equipment
US9190017B2 (en) 2013-01-02 2015-11-17 International Business Machines Corporation Proportional pointer transition between multiple display devices
US9619131B2 (en) * 2013-02-22 2017-04-11 Blackberry Limited Methods and devices for displaying content
US8959562B2 (en) * 2013-02-26 2015-02-17 Roku, Inc. Method and apparatus for automatic second screen engagement
US20140267074A1 (en) * 2013-03-14 2014-09-18 Qualcomm Incorporated System and method for virtual user interface controls in multi-display configurations
KR20140132917A (en) * 2013-05-09 2014-11-19 삼성전자주식회사 Method and apparatus for displaying user interface through sub-device connectable with portable device
US11138578B1 (en) 2013-09-09 2021-10-05 United Services Automobile Association (Usaa) Systems and methods for remote deposit of currency
US9286514B1 (en) 2013-10-17 2016-03-15 United Services Automobile Association (Usaa) Character count determination for a digital image
CN109062486B (en) * 2014-05-19 2022-03-25 联想(北京)有限公司 Display window adjusting method and electronic equipment
US10402790B1 (en) 2015-05-28 2019-09-03 United Services Automobile Association (Usaa) Composing a focused document image from multiple image captures or portions of multiple image captures
WO2019040081A1 (en) * 2017-08-25 2019-02-28 Intel Corporation Configuring display screen coordinates
US11030752B1 (en) 2018-04-27 2021-06-08 United Services Automobile Association (Usaa) System, computing device, and method for document detection
JP6810720B2 (en) * 2018-05-10 2021-01-06 レノボ・シンガポール・プライベート・リミテッド Electronic devices, information processing systems, information processing methods, and programs
US11900755B1 (en) 2020-11-30 2024-02-13 United Services Automobile Association (Usaa) System, computing device, and method for document detection and deposit processing

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6073034A (en) * 1996-10-31 2000-06-06 Kopin Corporation Wireless telephone display system
US6085112A (en) * 1995-05-03 2000-07-04 Siemens Aktiengesellschaft Communication device
US6211858B1 (en) * 1997-09-26 2001-04-03 Ericsson Inc. Method and apparatus for displaying a rotating meter icon on a portable intelligent communications device
US6275714B1 (en) * 1998-02-26 2001-08-14 Inviso, Inc. Phone with ergonomic virtual image display
US6418329B1 (en) * 1997-12-22 2002-07-09 Sony Corporation Portable radio information terminal, screen displaying method, recording medium and microcomputer
US6424843B1 (en) * 1997-04-22 2002-07-23 Nokia Oyj Multi-function telecommunication device

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5425077A (en) * 1993-07-08 1995-06-13 U.S. West Advanced Technologies, Inc. Mobile telephone user interface including fixed and dynamic function keys and method of using same
JPH0926769A (en) * 1995-07-10 1997-01-28 Hitachi Ltd Picture display device
US6243596B1 (en) * 1996-04-10 2001-06-05 Lextron Systems, Inc. Method and apparatus for modifying and integrating a cellular phone with the capability to access and browse the internet
US6144358A (en) * 1997-08-20 2000-11-07 Lucent Technologies Inc. Multi-display electronic devices having open and closed configurations
AU2001228223A1 (en) * 2000-01-24 2001-07-31 Sanjay Chadha Hand-held personal computing device with microdisplay
US7224324B2 (en) * 2000-03-27 2007-05-29 Donnelly Corporation Interactive automotive rearvision system
JP2002064613A (en) * 2000-08-18 2002-02-28 Hitachi Kokusai Electric Inc Portable radio terminal set, connection method thereof, and cutoff method thereof
US20020090980A1 (en) * 2000-12-05 2002-07-11 Wilcox Russell J. Displays for portable electronic apparatus

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6085112A (en) * 1995-05-03 2000-07-04 Siemens Aktiengesellschaft Communication device
US6073034A (en) * 1996-10-31 2000-06-06 Kopin Corporation Wireless telephone display system
US6424843B1 (en) * 1997-04-22 2002-07-23 Nokia Oyj Multi-function telecommunication device
US6211858B1 (en) * 1997-09-26 2001-04-03 Ericsson Inc. Method and apparatus for displaying a rotating meter icon on a portable intelligent communications device
US6418329B1 (en) * 1997-12-22 2002-07-09 Sony Corporation Portable radio information terminal, screen displaying method, recording medium and microcomputer
US6275714B1 (en) * 1998-02-26 2001-08-14 Inviso, Inc. Phone with ergonomic virtual image display

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100487618B1 (en) * 2002-12-11 2005-05-03 엘지전자 주식회사 Mobile phone having various display apparatuses
WO2005062591A1 (en) * 2003-12-02 2005-07-07 Sony Ericsson Mobile Communications Ab System and method for audible control of a camera associated with a mobile phone
CN1627765B (en) * 2003-12-10 2010-09-01 松下电器产业株式会社 Portable information terminal device
WO2006123294A2 (en) * 2005-05-19 2006-11-23 Koninklijke Philips Electronics, N.V. Apparatus and method to enhance navigation in a user interface for mobile devices
WO2006123294A3 (en) * 2005-05-19 2007-03-01 Koninkl Philips Electronics Nv Apparatus and method to enhance navigation in a user interface for mobile devices
WO2007105508A1 (en) * 2006-03-09 2007-09-20 Casio Computer Co., Ltd. Screen display control apparatus and program product
US7904513B2 (en) 2006-08-31 2011-03-08 Casio Computer Co., Ltd. Client apparatus, server apparatus, server-based computing system, and program
US8620997B2 (en) 2009-03-24 2013-12-31 Casio Computer Co., Ltd Client apparatus, computer system, computer readable program storage medium and display method, each for detecting change of display contents in status bar area to display the change

Also Published As

Publication number Publication date
WO2002079913A3 (en) 2003-04-17
JP2004535699A (en) 2004-11-25
US7016704B2 (en) 2006-03-21
AU2002250508A1 (en) 2002-10-15
US20020151283A1 (en) 2002-10-17
CN1533637A (en) 2004-09-29
KR20030097820A (en) 2003-12-31
CA2441301A1 (en) 2002-10-10
EP1391044A2 (en) 2004-02-25

Similar Documents

Publication Publication Date Title
US7016704B2 (en) Coordinating images displayed on devices with two or more displays
US11402970B2 (en) Systems, methods, and user interfaces for interacting with multiple application windows
US11010027B2 (en) Device, method, and graphical user interface for manipulating framed graphical objects
US20210389874A1 (en) Devices, Methods, and Graphical User Interfaces for Keyboard Interface Functionalities
US20200159406A1 (en) Portable Computing Input Devices and Methods
US8509854B2 (en) Mobile terminal and method of controlling operation of the same
KR101590457B1 (en) Portable touch screen device, method, and graphical user interface for using emoji characters
US8825699B2 (en) Contextual search by a mobile communications device
US7865215B2 (en) Magnification of currently selected menu item
US20120200503A1 (en) Sizeable virtual keyboard for portable computing devices
Tarasewich Wireless device for mobile commerce: User interface design and usability
CN117425874A (en) System, method, and user interface for interacting with multiple application views
US20070234231A1 (en) Method and apparatus for inputting text effect item
Tarasewich et al. Interface design for handheld mobile devices
US20100169831A1 (en) Information Product and Method for Interacting with User

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ OM PH PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG UZ VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
WWE Wipo information: entry into national phase

Ref document number: 2441301

Country of ref document: CA

WWE Wipo information: entry into national phase

Ref document number: 02807744X

Country of ref document: CN

WWE Wipo information: entry into national phase

Ref document number: 1554/CHENP/2003

Country of ref document: IN

WWE Wipo information: entry into national phase

Ref document number: 2002577682

Country of ref document: JP

Ref document number: 2002719425

Country of ref document: EP

Ref document number: 1020037013029

Country of ref document: KR

REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

WWP Wipo information: published in national office

Ref document number: 2002719425

Country of ref document: EP

WWW Wipo information: withdrawn in national office

Ref document number: 2002719425

Country of ref document: EP