Search Images Maps Play YouTube News Gmail Drive More »
Sign in
Screen reader users: click this link for accessible mode. Accessible mode has the same essential features but works better with your reader.

Patents

  1. Advanced Patent Search
Publication numberUS20090160803 A1
Publication typeApplication
Application numberUS 12/326,991
Publication dateJun 25, 2009
Filing dateDec 3, 2008
Priority dateDec 21, 2007
Also published asUS20130300698
Publication number12326991, 326991, US 2009/0160803 A1, US 2009/160803 A1, US 20090160803 A1, US 20090160803A1, US 2009160803 A1, US 2009160803A1, US-A1-20090160803, US-A1-2009160803, US2009/0160803A1, US2009/160803A1, US20090160803 A1, US20090160803A1, US2009160803 A1, US2009160803A1
InventorsHirokazu Hashimoto
Original AssigneeSony Corporation
Export CitationBiBTeX, EndNote, RefMan
External Links: USPTO, USPTO Assignment, Espacenet
Information processing device and touch operation detection method
US 20090160803 A1
Abstract
An information processing device includes: a display unit having a touch panel on its front surface; a movement amount calculation unit that calculates the movement amount of a touch operation based on a touch point at which the touch operation is performed with respect to the touch panel and a touch release point at which the touch operation is released from the touch panel; an operation determination unit that determines whether the touch operation is a depression operation or a gesture operation depending on the calculated movement amount; and a command recognition unit that recognizes whether a received command is a command corresponding to the depression operation or the gesture operation.
Images(13)
Previous page
Next page
Claims(9)
1. An information processing device comprising:
display means having a touch panel on its front surface;
movement amount calculation means for calculating a movement amount of a touch operation based on a touch point at which the touch operation is performed with respect to the touch panel and a touch release point at which the touch operation is released from the touch panel;
operation determination means for determining whether the touch operation is a depression operation or a gesture operation depending on the calculated movement amount; and
command recognition means for recognizing whether a received command is a command corresponding to the depression operation or the gesture operation.
2. The information processing device according to claim 1, wherein
the operation determination means determines that the touch operation is a gesture operation when the movement amount exceeds a predetermined threshold.
3. The information processing device according to claim 2, wherein
the command recognition means recognizes the command associated with a pattern of the gesture operation.
4. The information processing device according to claim 1, wherein
the operation determination means determines that, when the movement amount is not more than a predetermined threshold and a button is displayed at the touch release point at which the touch operation is released, the touch operation is the depression operation with respect to the button.
5. A touch operation detection method comprising:
a touch point detection step in which touch point detection means detects a touch point at which a touch operation is performed with respect to a touch panel provided on the front surface of display means;
a touch release point detection step following the touch point detection step, in which touch release point detection means detects a touch release point at which the touch operation is released from the touch panel;
a movement amount calculation step in which movement amount calculation means calculates a movement amount of the touch operation based on the touch point and the touch release point;
an operation determination step in which operation determination means determines whether the touch operation is a depression operation or a gesture operation depending on the calculated movement amount; and
a command recognition step in which command recognition means recognizes whether a received command is a command corresponding to the depression operation or the gesture operation.
6. The touch operation detection method according to claim 5, wherein
the operation determination step determines that the touch operation is the gesture operation when the movement amount exceeds a predetermined threshold.
7. The touch operation detection method according to claim 6, wherein
the command recognition step recognizes the command associated with a pattern of the gesture operation.
8. The touch operation detection method according to claim 5, wherein
the operation determination step determines that, when the movement amount is not more than a predetermined threshold and a button is displayed at the touch release point at which the touch operation is released, the touch operation is the depression operation with respect to the button.
9. An information processing device comprising:
a display unit having a touch panel on its front surface;
a movement amount calculation unit that calculates a movement amount of a touch operation based on a touch point at which the touch operation is performed with respect to the touch panel and a touch release point at which the touch operation is released from the touch panel;
an operation determination unit that determines whether the touch operation is a depression operation or a gesture operation depending on the calculated movement amount; and
a command recognition unit that recognizes whether a received command is a command corresponding to the depression operation or the gesture operation.
Description
    CROSS REFERENCES TO RELATED APPLICATIONS
  • [0001]
    The present invention contains subject matter related to Japanese Patent Application JP2007-330849 filed in the Japanese Patent Office on Dec. 21, 2007, the entire contents of which being incorporated herein by reference.
  • BACKGROUND OF THE INVENTION
  • [0002]
    1. Field of the Invention
  • [0003]
    The present invention relates to an information processing device and a touch operation detection method, which are suitably applied to, e.g., a portable navigation device.
  • [0004]
    2. Description of the Related Art
  • [0005]
    A portable navigation device (hereinafter, referred to as “PND”) is designed to be detachably attached to the dashboard of a vehicle via a cradle.
  • [0006]
    The PND of such a type serves as a vehicle navigation device when being attached to the dashboard via a cradle and serves as a personal navigation device when being detached from the dashboard.
  • [0007]
    This type of PND aims more for portability than remote-controllability. Thus, a remote controller is not provided for the PND but a user interface that receives a command from a user via a touch panel provided on the front surface of a liquid crystal display is adopted.
  • [0008]
    There is also known an electric book display control device equipped with a display integrated tablet on which a user uses his or her finger to perform a pointing operation to realize a page-turning operation (refer to, e.g., Jpn. Pat. Appln. Laid-Open Publication No. 2004-348755).
  • SUMMARY OF THE INVENTION
  • [0009]
    The PND having such a configuration displays a plurality of destination candidates so as to allow a user to select his or her destination. If the destination candidates exist over a plurality of pages, the user needs to perform a page-turning operation in order to determine the destination.
  • [0010]
    In such a case, where the PND performs a page-turning operation in response to the user's pointing operation using his or her finger in the same manner as in the case of the abovementioned electronic book display control device, the PND often erroneously detects that a depression operation with respect to the candidate destination has been made at the moment when the user falsely touched one of the destination candidates with his or her finger, not a page-turning button. Thus, there is a possibility that the PND cannot correctly reflect the user's intention to perform a pointing operation.
  • [0011]
    The present invention has been made in view of the above points and an object thereof is to propose an information processing device and a touch operation detection method capable of correctly detect a command issued in response to a user's touch operation.
  • [0012]
    To solve the above problem, according to an aspect of the present invention, there is provided an information processing device including: a display unit having a touch panel on its front surface; a movement amount calculation unit for calculating a movement amount of a touch operation based on a touch point at which the touch operation is performed with respect to the touch panel and a touch release point at which the touch operation is released from the touch panel; an operation determination unit for determining whether the touch operation is a depression operation or a gesture operation depending on the calculated movement amount; and a command recognition unit for recognizing whether a received command is a command corresponding to the depression operation or the gesture operation.
  • [0013]
    With the above configuration, the information processing device determines whether a touch operation is a button touch operation or a gesture operation depending on the movement amount of the touch operation, thereby correctly recognizing whether a received command is a command corresponding to the depression operation or the gesture operation.
  • [0014]
    According to another aspect of the present invention, there is provided a touch operation detection method including: a touch point detection step in which a touch point detection unit detects a touch point at which a touch operation is performed with respect to a touch panel provided on the front surface of a display unit; a touch release point detection step following the touch point detection step, in which a touch release point detection unit detects a touch release point at which the touch operation is released from the touch panel; a movement amount calculation step in which a movement amount calculation unit calculates a movement amount of the touch operation based on the touch point and touch release point; an operation determination step in which an operation determination unit determines whether the touch operation is a depression operation or a gesture operation depending on the calculated movement amount; and a command recognition step in which a command recognition unit recognizes whether a received command is a command corresponding to the depression operation or the gesture operation.
  • [0015]
    With the above configuration, the touch operation detection method determines whether a touch operation is a button touch operation or a gesture operation depending on the movement amount of the touch operation, thereby correctly recognizing whether a received command is a command corresponding to the depression operation or the gesture operation.
  • [0016]
    According to the present invention, whether a touch operation is a button touch operation or a gesture operation is determined depending on the movement amount of the touch operation, allowing correct determination on whether a received command is a command corresponding to the depression operation or the gesture operation. Therefore, an information processing device and a touch operation detection method capable of correctly recognizing a command corresponding to a user's touch operation can be realized.
  • [0017]
    The nature, principle and utility of the invention will become more apparent from the following detailed description when read in conjunction with the accompanying drawings in which like parts are designated by like reference numerals or characters.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • [0018]
    In the accompanying drawings:
  • [0019]
    FIG. 1 is a perspective view schematically showing the outer appearance of a PND;
  • [0020]
    FIG. 2 is a block diagram schematically showing a circuit configuration of the PND;
  • [0021]
    FIG. 3 is a view schematically showing a guide map image;
  • [0022]
    FIG. 4 is a view schematically showing an application menu screen;
  • [0023]
    FIG. 5 is a view schematically showing a book selection screen;
  • [0024]
    FIG. 6 is a view schematically showing a password requesting screen;
  • [0025]
    FIG. 7 is a view schematically showing a spot selection screen;
  • [0026]
    FIG. 8 is a view schematically showing an introduction page screen;
  • [0027]
    FIG. 9 is a view schematically showing a detailed page screen;
  • [0028]
    FIG. 10 is a flowchart showing a procedure of command recognition processing;
  • [0029]
    FIG. 11 is a view schematically showing a touch release operation; and
  • [0030]
    FIG. 12 is a schematic view for explaining a gesture operation.
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
  • [0031]
    An embodiment of the present invention will be described in detail below with reference to the accompanying drawings.
  • (1) Outer Appearance of PND
  • [0032]
    In FIG. 1, a reference numeral 1 denotes a portable navigation device (hereinafter, referred to as “PND”) according to the embodiment of the present invention. The PND 1 has a display section 2 on which a 4.8 liquid crystal display is mounted and uses the display section 2 to present information such as a map, a current position icon PST, and a driving route toward a destination.
  • [0033]
    The PND 1 is held by a cradle 3 attached to a dashboard of a vehicle by a sucker 3A and is electrically connected thereto. In this state, the PND operates by a power supplied from a battery of the vehicle via the cradle 3. Further, when removed from the cradle 3, the PND 1 can operate by a power supplied from a battery incorporated therein.
  • (2) Circuit Configuration of PND
  • [0034]
    As shown in FIG. 2, a controller 16 having a microcomputer configuration controls the entire operation of the PND 1 according a basic program and executes various navigation processing, command recognition processing to be described later, and the like according to various application programs stores in a hard disk, etc.
  • [0035]
    Concretely, the PND 1 uses a GPS module 12 to demodulate satellite signals S1 from a plurality of GPS satellites received via a GPS antenna 11 and, based on orbital data obtained as a result of the demodulation and distance data between the GPS satellites and vehicle, measures the current position of the vehicle with accuracy and transmits the obtained current position data (current position data S2) to a route search section 13.
  • [0036]
    The orbital data is detailed orbital information (parameter) representing a detailed orbit of each of the GPS satellites. In order to measure the current position of the vehicle with accuracy, the orbital data needs to be acquired from at least three GPS satellites.
  • [0037]
    The route search section 13 reads out map data S3 representing the current position of the vehicle and its surrounding area from a map data storage section 14 based on the current position data S2, searches the map data S3 for a driving route from the current position to a destination which has been set by a user, generates a route guide map S4 including the driving route, and transmits the route guide map S4 to the display section 2 constituted by a liquid crystal display.
  • [0038]
    As shown in FIG. 3, the display section 2 displays a guide map image G1 corresponding to the route guide map data S4 and thereby allows a user to visually confirm the driving route from a current position icon PST to a destination (not shown).
  • [0039]
    When the controller 16 of the PND 1 receives a command S7 which has been issued in response to depression of an application menu button MB displayed in the lower left of the guide map image G1 via a touch panel 15, the controller 16 reads out application menu screen data S5 from a data storage section 17.
  • [0040]
    The controller 16 of the PND 1 then displays an application menu screen G2 corresponding to the application menu screen data S5 on the display section 2 as shown in FIG. 4. In the application menu screen G2, a music button B1 for reproduction of music, a video button B2 for reproduction of video, and a guidebook button B3 for display of Point Of Interest (POI) which can be candidates of a destination or a stop-off point are displayed as choices to be selected through a user's touch operation.
  • [0041]
    When detecting a touch operation made to the guidebook button B3 in the application menu screen G2 via the touch panel 15 (FIG. 2), the controller 16 of the PND 1 reads out a book selection screen data S6 associated with the guidebook button B3 from the data storage section 17.
  • [0042]
    The controller 16 of the PND 1 then displays a book selection screen G3 corresponding to the book selection screen data S6 on the display section 2 as shown in FIG. 5. The book selection screen G3 displays a list of choices of guidebooks such as “100 spots” item KL1, “1,000 spots” item KL2, “illumination information −free version” item KL3, “Ekiben (train lunch) search −free version 2)” item KL4, and an UP button UB1 and a down button DB1 for scrolling the list in the up-down direction.
  • [0043]
    The book selection screen G3 further displays a “return” button BB1. When the “return” button BB1 is touched by a user, the controller 16 of the PND 1 sets back the display content of the display section 2 from the book selection screen G3 to application menu screen G2.
  • [0044]
    When detecting, via the touch panel 15, that the “1,000 spots” item KL2 on the book selection screen G3 has been touched by a user, the controller 16 of the PND 1 reads out password requesting screen data S7 from the data storage section 17 if a password has been set for the “1,000 spots” item KL2.
  • [0045]
    Then, as shown in FIG. 6, the controller 16 of the PND 1 displays a password requesting screen G4 corresponding to the password requesting screen data S7 on the display section 2. The password requesting screen G4 displays a password input field IP, alphabet keys AK, a numeric key NK, a delete key DK, an entry key EK, as well as, a left button LB1 and a right button RB2 for moving a cursor CS of the password input field IP in the left-right direction and a “return” button BB2.
  • [0046]
    Also in this case, when the “return” button BB2 is touched by a user, the controller 16 of the PND 1 sets back the display content of the display section 2 from the password requesting screen G4 to book selection screen G3.
  • [0047]
    The controller 16 of the PND 1 displays a password input in response to a touch operation with respect to the alphabet key AK or numeric key EK in the password input field IP of the password requesting screen G4 and determines the password displayed in the password input field IP in response to a touch operation with respect to the enter key EK.
  • [0048]
    When authentication based on the password input to the password input field IP of the password requesting screen G4 is successful, the controller 16 of the PND 1 reads out spot selection screen data S8 from the data storage section 17.
  • [0049]
    The controller 16 of the PND 1 then displays a spot selection screen G5 corresponding to the spot selection screen data S8 as shown in FIG. 7. The spot selection screen G5 displays, as the POIs which can be candidates of a destination or a stop-off point, a list of spot items SA1 to SA4: “G0001 Jonan Park”, “G0002 Gold Bridge”, “G0003 Port-side Tower”, and “G0004 Yumemigaoka” and a up button UB2 and a down button DB2 for scrolling the list in the up-down direction.
  • [0050]
    Further, the spot selection screen G5 displays a “return” button BB3. When the “return” button BB3 is touched by a user, the controller 16 of the PND 1 sets back the display content of the display section 2 from the spot selection screen G5 to book selection screen G3.
  • [0051]
    When the controller 16 of the PND 1 detects via the touch panel 15 that e.g., the spot item SA1: “G0001 Jonan Park” displayed on the spot selection screen G5 has been touched, the controller 16 of the PND 1 reads out introduction page screen data S8 about “G0001 Jonan Park” from the data storage section 17.
  • [0052]
    The controller 16 of the PND 1 then displays an introduction page screen G6 corresponding to the introduction page screen data S8 on the display section 2 as shown in FIG. 8. The introduction page screen G6 displays a guidebook name display field TL1, a photo display field GA, a spot item name display field NL1, a location specifying information display field AL, a “to-map” button MB1 for returning to the abovementioned guide map image G1 (FIG. 3), and a return button BB4.
  • [0053]
    The introduction page screen G6 further displays, in the lower center thereof, a next button NT1 with an arrow for displaying a next page and a back button BK1 with an arrow for displaying a previous page. When the next button NT1 is touched by a user, the controller 16 of the PND 1 reads out detailed page screen data S9 concerning the “Jonan Park” from the data storage section 17.
  • [0054]
    The controller 16 of the PND 1 then displays a detailed page screen G7 corresponding to the detailed page screen data S9 on the display section 2 as shown in FIG. 9. The detailed page screen G7 displays a guide book name display field TL2, a spot item name display field NL2, a detailed content display field DL representing detailed content of the “G0001 Jonan Park”, a “to-map” button MB2 for returning to the guide map image G1 (FIG. 3), and a return button BB5.
  • [0055]
    The controller 16 of the PND 1 allows a user to visually confirm the detailed content display field DL of the detailed page screen G7 in the manner as described above. As a result, the user can grasp the detailed content of the “G0001 Jonan Park” and determine whether to set or not the “G0001 Jonan Park” as his or her destination or stop-off point.
  • [0056]
    Also in this case, the detailed page screen G7 displays, in the lower center thereof, a next button NT2 with an arrow for displaying a next page and a back button BK2 with an arrow for displaying a previous page. When the back button BK2 is touched by a user, the controller 16 of the PND 1 sets back the content of the display section 2 to the introduction page screen G6 concerning the “G0001 Jonan Park”.
  • [0057]
    In addition, when switching from the introduction page screen G6 to detailed page screen G7 (i.e., turn the page) or when switching from the detailed page screen G7 to introduction page screen G6 (i.e., bring back the page), the controller 16 of the PND 1 can perform the switching operation not only by detecting a user's touch operation with respect to the next buttons NT1, NT2 and back buttons BK1, BK2, but also by understanding a command corresponding to a finger gesture with respect to the touch panel 15 of the display section 2. In this case, the controller 16 of the PND 1 can perform the page switching processing according to the command. The details of this command recognition processing will be described below.
  • (3) Procedure of Command Recognition Processing
  • [0058]
    As shown in FIG. 9, the controller 16 of the PND 1 enters the processing from a starting step of a routine RT1 according to a command recognition processing program which is an application program started from a hard disk and proceeds to step SP1. In step SP1, when detecting that a touch operation has been made with respect to the touch panel 15 by a user's finger, the controller 16 proceeds to step SP2.
  • [0059]
    In step SP2, when detecting a touch release operation, i.e., detecting that the user's finger has been separated from the touch panel 15, the controller 16 of the PND 1 proceeds to step SP3.
  • [0060]
    In step SP3, the controller 16 of the PND 1 calculates the movement amount of the finger on the display section 2 from the touch point detected in step SP1 to touch release point detected in step SP2 and proceeds to step SP4.
  • [0061]
    In step SP4, the controller 16 of the PND 1 determines whether the movement amount of the finger calculated in step SP3 is not more than a predetermined threshold (e.g., 5 mm). When an affirmative result has been obtained, which means that there is little movement, the controller 16 of the PND 1 determines that a gesture operation based on a drag operation of the finger has not been performed and proceeds to step SP5.
  • [0062]
    Since the finger motion does not mean a gesture operation, the controller 16 of the PND 1 determines, in step SP5, whether any button (e.g., next button NT1) exists at the touch release point as shown in FIG. 11.
  • [0063]
    When a negative result has been obtained in step SP5, which means that the finger motion means neither a gesture operation nor a button touch operation, the controller 16 of the PND 1 determines that no command has been input and proceeds to step SP8 where the controller 16 of the PND 1 ends this flow without doing anything.
  • [0064]
    On the other hand, when an affirmative result has been obtained in step SP5, which means that the finger motion means a button touch operation with respect to, e.g., the next button NT1, the controller 16 of the PND 1 proceeds to step SP6.
  • [0065]
    In step SP6, the controller 16 of the PND 1 recognizes that a command corresponding to the button touch operation with respect to the next button NT1 has been issued and switches the display content from, e.g., the introduction page screen G6 (FIG. 8) to detailed page screen G7 (FIG. 9). Then, the controller 16 of the PND 1 proceeds to step SP8 and ends this flow.
  • [0066]
    When a negative result has been obtained in step SP4, i.e., the movement amount of the finger calculated in step SP3 exceeds a predetermined threshold (e.g., 5 mm), which means that a gesture operation based on a drag operation has been performed such that the finger is dragged from the left to right on the display section 2 as shown in FIG. 12, the controller 16 of the PND 1 proceeds to step SP7.
  • [0067]
    In step SP7, the controller 16 of the PND 1 recognizes, e.g., a page-turning command corresponding to the pattern of a gesture operation and, according to the page-turning command, switches the display content from the introduction page screen G6 (FIG. 8) to detailed page screen G7 (FIG. 9) in such a manner as if the page of a book were turned over. Then, the controller 16 of the PND 1 proceeds to step SP8 and ends this flow.
  • [0068]
    When recognizing that a gesture operation of a motion pattern in which the finger is dragged from the left to right on the display section 2 has been performed, the controller 16 of the PND 1 performs the abovementioned page-turning operation, while when recognizing that a gesture operation of a motion pattern in which the finger is dragged from the right to left on the display section 2 has been performed, the controller 16 of the PND 1 performs a page turning back operation to set back the display content from the detailed page screen G7 (FIG. 9) to introduction page screen G6 (FIG. 8).
  • [0069]
    When recognizing that a gesture operation of a motion pattern in which the finger is moved so as to draw a triangle has been performed, the controller 16 of the PND 1 recognizes that the motion means a command to search a driving route from the current position to the home of a user and displays the driving route obtained as a result of the search on the display section 2.
  • [0070]
    When recognizing that a gesture operation of a motion pattern in which the finger is swirled clockwise has been performed, the controller 16 of the PND 1 enlarges a map currently displayed on the display section 2, while when recognizing that a gesture operation of a motion pattern in which the finger is swirled counterclockwise has been performed, the controller 16 of the PND 1 reduces the map size.
  • (4) Operation and Effect
  • [0071]
    With the above configuration, the controller 16 of the PND 1 determines whether a finger motion is a button touch operation or a gesture operation based on a drag operation depending on the movement amount of the finger on the display section 2 from the touch point to touch release point, recognizes a command corresponding to the button touch operation or gesture operation, and performs predetermined processing according to the command.
  • [0072]
    That is, even if any button displayed on the display section 2 is touched by the finger, the controller 16 of the PND 1 does not determine whether the finger motion is a button touch operation or a gesture operation until the finger is separated from the button.
  • [0073]
    Therefore, even if a point at which any button exists is accidentally touched by the finger although the user intended to perform a gesture operation, it is possible to prevent the controller 16 of the PND 1 from erroneously determining that the finger motion is a button touch operation.
  • [0074]
    Further, even if the finger is applied to any button and is moved by an amount not more than a predetermined threshold until it is separated from the display section 2, the controller 16 of the PND 1 does not erroneously determine that the finger motion is a gesture operation but correctly determines that the finger motion is a button touch operation with respect to a button.
  • [0075]
    According to the above configuration, the controller 16 of the PND 1 correctly determines whether a touch operation is a button touch operation instantaneously made with respect to a button or a gesture operation based on a drag operation to thereby correctly and reliably execute predetermined processing specified by a command corresponding to the button touch operation or gesture operation.
  • (5) Other Embodiments
  • [0076]
    In the above embodiment, a value of 5 mm is used as a threshold of the movement amount of the finger from its touch point to touch release point to determine whether the finger motion is a button touch operation or a gesture operation. However, the present invention is not limited to this. For example, the value of the threshold may arbitrarily set depending on various factors such as the size of the display area of the display section 2 or size of the button.
  • [0077]
    Further, in the above embodiment, the pattern of a gesture operation includes the drag-like motion of the finger from the right to left, drag-like motion of the finger from the left to right, motion of the finger such as one by which a triangle is drawn, and clockwise or counterclockwise swirled motion of the finger. However, the present invention is not limited to this. For example, the pattern of a gesture operation may include a motion of the finger such as one by which “x” or “□” is drawn, a finger tapping motion, and other various types of motions.
  • [0078]
    Further, in the above embodiment, the controller 16 of the PND 1 determines that the finger motion is a gesture operation based on the movement amount of the finger on the display section 2 from the touch point to touch release point irrespective of whether any button displayed on the display section 2 is touched by the finger. However, the present invention is not limited to this but the controller 16 of the PND 1 may determine that the finger motion is a gesture operation based on the movement amount of the finger on the display section 2 where no button is displayed.
  • [0079]
    Further, in the above embodiment, the controller 16 of the PND 1 executes the command recognition processing procedure (FIG. 10) of the routine RT1 according to the command recognition processing program which is an application program stored in a hard disk. However, the present invention is not limited to this but the command recognition processing procedure may be executed according to a command recognition processing program installed from a program storage medium such as a compact disk (CD) or a semiconductor memory, one downloaded from the Internet, or one installed through other various routes.
  • [0080]
    Further, in the above embodiment, the PND 1 as an information processing device according to the embodiment of the present invention is constituted by the display section 2 serving as a display unit, and controller 16 serving as a movement amount calculation unit, an operation determination unit, and a command recognition unit. Alternatively, however, the information processing device according to the embodiment of the present invention may be constituted by the display unit, movement amount calculation unit, operation determination unit, and command recognition unit including other various circuit configurations.
  • [0081]
    The information processing device and touch operation detection method can be applied to various electronic apparatuses having a touch panel, such as a dashboard-mounted navigation apparatus other than the PND, a personal computer, a Personal Digital Assistant (PDA), a mobile phone, and a game apparatus.
  • [0082]
    It should be understood by those skilled in the art that various modifications, combinations, sub-combinations and alterations may occur depending on design requirements and other factors insofar as they are within the scope of the appended claims or the equivalents thereof.
Patent Citations
Cited PatentFiling datePublication dateApplicantTitle
US6094197 *May 17, 1995Jul 25, 2000Xerox CorporationGraphical keyboard
US20030206199 *May 1, 2003Nov 6, 2003Nokia CorporationMethod and apparatus for interaction with a user interface
US20050248542 *Apr 29, 2005Nov 10, 2005Pentax CorporationInput device and method for controlling input device
US20080184173 *Jan 31, 2007Jul 31, 2008Microsoft CorporationControlling multiple map application operations with a single gesture
US20080201650 *Jan 6, 2008Aug 21, 2008Lemay Stephen OWeb-Clip Widgets on a Portable Multifunction Device
US20090102806 *Feb 28, 2008Apr 23, 2009Steve TomkinsSystem having user interface using object selection and gestures
Referenced by
Citing PatentFiling datePublication dateApplicantTitle
US9086792 *Jan 8, 2013Jul 21, 2015Canon Kabushiki KaishaDisplay control apparatus and control method thereof
US9262612Mar 21, 2011Feb 16, 2016Apple Inc.Device access using voice authentication
US9283829 *Feb 28, 2011Mar 15, 2016Volkswagen AgProcess and device for displaying different information for driver and passenger of a vehicle
US9318108Jan 10, 2011Apr 19, 2016Apple Inc.Intelligent automated assistant
US9330720Apr 2, 2008May 3, 2016Apple Inc.Methods and apparatus for altering audio output signals
US9338493Sep 26, 2014May 10, 2016Apple Inc.Intelligent automated assistant for TV user interactions
US20110167350 *Jan 6, 2010Jul 7, 2011Apple Inc.Assist Features For Content Display Device
US20110221686 *Jul 30, 2010Sep 15, 2011Samsung Electronics Co., Ltd.Portable device and control method thereof
US20110238290 *Mar 24, 2010Sep 29, 2011Telenav, Inc.Navigation system with image assisted navigation mechanism and method of operation thereof
US20120299854 *Nov 29, 2012Kyocera CorporationMobile electronic device and input method
US20130047112 *Feb 28, 2011Feb 21, 2013XMethod and device for operating a user interface
US20130176256 *Jan 8, 2013Jul 11, 2013Canon Kabushiki KaishaDisplay control apparatus and control method thereof
US20140240101 *Sep 10, 2012Aug 28, 2014Nec Casio Mobile Communications, Ltd.Device and method for processing write information of electronic tag
USD702140Jan 2, 2013Apr 8, 2014Garmin Switzerland GmbhNavigation device
USD742258Jan 15, 2014Nov 3, 2015Garmin Switzerland GmbhNavigation device
USD751933Dec 22, 2014Mar 22, 2016Garmin Switzerland GmbhNavigation device
CN103197881A *Jan 8, 2013Jul 10, 2013佳能株式会社Display control apparatus and control method thereof
DE102011116187A1Oct 14, 2011Apr 18, 2013Volkswagen AktiengesellschaftMethod for providing user interface for interaction with e.g. infotainment system, in vehicle, involves expanding region in which graphic object e.g. electronic log book, is displayed on touch screen by detected gesture for page turning
EP2367098A2 *Aug 2, 2010Sep 21, 2011Samsung Electronics Co., Ltd.Portable device and control method thereof
EP2367098A3 *Aug 2, 2010Jul 30, 2014Samsung Electronics Co., Ltd.Portable device and control method thereof
EP2838774A4 *Apr 15, 2013May 20, 2015Nuance Communications IncLow-attention gestural user interface
Classifications
U.S. Classification345/173
International ClassificationG06F3/041
Cooperative ClassificationG06F3/0488, G06F3/04883
European ClassificationG06F3/0488G
Legal Events
DateCodeEventDescription
Dec 3, 2008ASAssignment
Owner name: SONY CORPORATION,JAPAN
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HASHIMOTO, HIROKAZU;REEL/FRAME:021922/0030
Effective date: 20081118