A touch processing method and system using a gui image, the method including the steps of: reading a sensed data array from a touch panel; performing a conversion procedure on a gui to generate a gui image; and performing a touch processing procedure according to the sensed data array and the gui image.
|
14. A touch processing system using a graphical user interface, comprising:
a touch panel for outputting a sensed data array corresponding to user interaction therewith, wherein said sensed data array is generated in the touch panel without going through a process for calculating touch coordinates; and
a processing unit for reading said sensed data array from said touch panel; performing a conversion procedure on a gui to generate a gui image; and performing a touch processing procedure, wherein said touch processing procedure uses said sensed data array to get touched position information or uses both said sensed data array and said icons of said gui image to get touched icon information for locating or indicating at least one of said icons and activate the object function thereof.
1. A touch processing method for a graphical user interface, the method including the steps of:
using a processing unit including at least one CPU or GPU to read a sensed data array from a touch panel, wherein said sensed data array is generated in the touch panel without going through a process for calculating touch coordinates;
using said processing unit to perform a conversion procedure on a gui to generate a gui image, wherein said gui image has a background area corresponding to a touch plane of said touch panel, and said background area contains a plurality of icons with each of said icons representing an object; and
using said processing unit for performing a touch processing procedure, wherein said touch processing procedure uses said sensed data array to get touched position information or uses both said sensed data array and said icons of said gui image to get touched icon information for locating or indicating at least one of said icons and activate the object function thereof.
2. The touch processing method for a graphical user interface using a gui imago of
3. The touch processing method for a graphical user interface of
4. The touch processing method for a graphical user interface of
5. The touch processing method for a graphical user interface of
6. The touch processing method for a graphical user interface of
7. The touch processing method for a graphical user interface of
8. The touch processing method for a graphical user interface of
9. The touch processing method for a graphical user interface
10. The touch processing method for a graphical user interface of
11. The touch processing method for a graphical user interface of
12. The touch processing method for a graphical user interface of
13. The touch processing method for a graphical user interface of
15. The touch processing system using a graphical user interface of
16. The touch processing system using a graphical user interface of
17. The touch processing system using a graphical user interface of
18. The touch processing system using a graphical user interface of
19. The touch processing system using a graphical user interfaces of
|
1. Field of the Invention
The present invention relates to a touch processing method and system, especially to a touch processing method and system using a GUI (Graphical User Interface) image.
2. Description of the Related Art
Prior art touch processing methods generally use an X-Y sensor array to detect a touch coordinate, and then transmit the touch coordinate to activate a corresponding function. However, when the X-Y sensor array is of a large size, the generation of the touch coordinate can take much time, and the activation of the corresponding function can be therefore too late to keep up with a user's touch operation.
To solve the foregoing problem, the present invention proposes a novel touch processing mechanism, which uses a GUI image to improve the efficiency in processing a touch operation.
One objective of the present invention is to disclose a touch processing method using a GUI image, so as to detect touched objects in a GUI more efficiently, and provide multiple novel touch functions without the need of calculating touch coordinates.
Another objective of the present invention is to disclose a touch processing system using a GUI image, so as to detect touched objects in a GUI more efficiently, and provide multiple novel touch functions without the need of calculating touch coordinates.
To attain the foregoing objectives, a touch processing method using a GUI image is proposed, the method including the steps of:
reading a sensed data array from a touch panel;
performing a conversion procedure on a GUI to generate a GUI image; and
performing a touch processing procedure according to the sensed data array and the GUI image.
The touch panel has a touch sensor array of resistive type, capacitive type, surface acoustic wave type, or electromagnetic type.
The sensed data array is used for conveying touched information or approached information of a touch plane, and the GUI image has a background area corresponding to the touch plane, wherein the background area contains a plurality of icons, with each of the icons representing an object.
Preferably, the conversion procedure is one selected from the group consisting of a color conversion procedure and a resolution conversion procedure.
In one embodiment, the touch processing procedure has a first function, wherein the first function is used for mapping at least one touch coordinate, which is conveyed by the sensed data array, onto the icons of the GUI image, so as to activate the object function of at least one of the icons corresponding to the at least one touch coordinate.
In another embodiment, the touch processing procedure has a second function, wherein the second function is used for performing an image operation on the sensed data array according to the icons of the GUI image, so as to locate at least one of the icons and activate the object function thereof.
In still another embodiment, the touch processing procedure has a third function, wherein the third function is used for executing an image processing procedure to combine the sensed data array and the GUI image, and thereby generate an output image.
In still another embodiment, the touch processing procedure has a first function and a second function, wherein the first function is used for mapping at least one touch coordinate, which is conveyed by the sensed data array, onto the icons of the GUI image, so as to activate the object function of at least one of the icons corresponding to the at least one touch coordinate; and the second function is used for performing an image operation on the sensed data array according to the icons of the GUI image, so as to locate at least one of the icons and activate the object function thereof.
In still another embodiment, the touch processing procedure has a first function and a third function, wherein the first function is used for mapping at least one touch coordinate, which is conveyed by the sensed data array, onto the icons of the GUI image, so as to activate the object function of at least one of the icons corresponding to the at least one touch coordinate; and the third function is used for executing an image processing procedure to combine the sensed data array and the GUI image, and thereby generate an output image.
In still another embodiment, the touch processing procedure has a second function and a third function, wherein the second function is used for performing an image operation on the sensed data array according to the icons of the GUI image, so as to locate at least one of the icons and activate the object function thereof; and the third function is used for executing an image processing procedure to combine the sensed data array and the GUI image, and thereby generate an output image.
In still another embodiment, the touch processing procedure has a first function, a second function, and a third function, wherein the first function is used for mapping at least one touch coordinate, which is conveyed by the sensed data array, onto the icons of the GUI image, so as to activate the object function of at least one of the icons corresponding to the at least one touch coordinate; the second function is used for performing an image operation on the sensed data array according to the icons of the GUI image, so as to locate at least one of the icons and activate the object function thereof; and the third function is used for executing an image processing procedure to combine the sensed data array and the GUI image, and thereby generate an output image.
Preferably, the image operation includes a convolution operation and a comparison operation.
Preferably, the image processing procedure includes a filtering operation.
Preferably, the third function is used in a multi-touch drawing application, a handwriting input application, a palm-touch control application, or a multi-finger music performing application.
To attain the foregoing objectives, a touch processing system using a GUI image is proposed, the system including:
a touch panel, used for outputting a sensed data array; and
a touch processing unit, used for: reading the sensed data array from the touch panel; performing a conversion procedure on a GUI to generate a GUI image; and performing a touch processing procedure according to the sensed data array and the GUI image.
The touch panel has a touch sensor array of resistive type, capacitive type, surface acoustic wave type, or electromagnetic type.
The sensed data array is used for conveying touched information or approached information of a touch plane. The conversion procedure is one selected from the group consisting of a color conversion procedure and a resolution conversion procedure, and the GUI image has a background area corresponding to the touch plane, wherein the background area contains a plurality of icons, with each of the icons representing an object.
The touch processing procedure has a function selected from a group consisting of a first function, a second function, and a third function, wherein the first function is used for mapping at least one touch coordinate, which is conveyed by the sensed data array, onto the icons of the GUI image, so as to activate the object function of at least one of the icons corresponding to the at least one touch coordinate; the second function is used for performing an image operation on the sensed data array according to the icons of the GUI image, so as to locate at least one of the icons and activate the object function thereof; and the third function is used for executing an image processing procedure to combine the sensed data array and the GUI image, and thereby generate an output image.
Preferably, the image operation includes a convolution operation and a comparison operation.
Preferably, the image processing procedure includes a filtering operation.
Preferably, the third function is used in a multi-touch drawing application, a handwriting input application, a palm-touch control application, or a multi-finger music performing application.
Preferably, the touch processing system using a GUI image further includes an object function execution unit, used for executing an object function under the control of the touch processing unit.
Preferably, the object function includes playing a film, playing music, displaying figures, or connecting to a website.
To make it easier for our examiner to understand the objective of the invention, its structure, innovative features, and performance, we use preferred embodiments together with the accompanying drawings for the detailed description of the invention.
The present invention will be described in more detail hereinafter with reference to the accompanying drawings that show the preferred embodiments of the invention.
Please refer to
In step a, the touch panel has a touch sensor array of resistive type, capacitive type, surface acoustic wave type, or electromagnetic type; and the sensed data array is used for conveying touched information or approached information of a touch plane.
In step b, the GUI image has a background area corresponding to the touch plane, wherein the background area contains a plurality of icons, with each of the icons representing an object. Besides, the conversion procedure is one selected from the group consisting of a color conversion procedure and a resolution conversion procedure, used for making the GUI image possess a specific color or a specific resolution.
In step c, the touch processing procedure has a function selected from a group consisting of a first function, a second function, and a third function, that is, the embodiment of the touch processing procedure can be one having the first function, or one having the second function, or one having the third function, or one having the first function and the second function, or one having the first function and the third function, or one having the second function and the third function, or one having the first function, the second function, and the third function.
The first function is used for mapping at least one touch coordinate, which is conveyed by the sensed data array, onto the icons of the GUI image, so as to activate the object function of at least one of the icons corresponding to the at least one touch coordinate; the second function is used for performing an image operation—for example but not limited to including a convolution operation and a comparison operation—on the sensed data array according to the icons of the GUI image, so as to locate at least one of the icons and activate the object function thereof, wherein the convolution operation will generate a convolution integral, and the comparison operation will compare the convolution integral with a threshold; and the third function is used for executing an image processing procedure to combine the sensed data array and the GUI image, and thereby generate an output image, wherein the image processing procedure includes performing a filtering operation and/or a color setting operation on the sensed data array.
For an operation scenario of the first function, please refer to
For an operation scenario of the second function, please refer to
For an operation scenario of the third function, please refer to
For another operation scenario of the third function, please refer to
For another operation scenario of the third function, please refer to
For still another operation scenario of the third function, please refer to
According to the foregoing specification, the present invention further proposes a touch processing system using a GUI image. Please refer to
The touch panel 110 has a touch sensor array of resistive type, capacitive type, surface acoustic wave type, or electromagnetic type; and has an output port DTP for delivering a sensed data array.
The touch processing unit 120 is used for: reading the sensed data array from the touch panel 110; performing a conversion procedure on the GUI 130 to generate a GUI image; and performing a touch processing procedure according to the sensed data array and the GUI image. The sensed data array is used for conveying touched information or approached information of a touch plane. The GUI image has a background area corresponding to the touch plane, wherein the background area contains a plurality of icons, with each of the icons representing an object. Besides, the conversion procedure is one selected from the group consisting of a color conversion procedure and a resolution conversion procedure, used for making the GUI image possess a specific color or a specific resolution.
The touch processing procedure has a function selected from a group consisting of a first function, a second function, and a third function, wherein the first function is used for mapping at least one touch coordinate, which is conveyed by the sensed data array, onto the icons of the GUI image, so as to activate the object function execution unit 150 to execute the object function of at least one of the icons corresponding to the at least one touch coordinate; the second function is used for performing an image operation—for example but not limited to a convolution operation and a comparison operation—on the sensed data array according to the icons of the GUI image, so as to locate at least one of the icons and activate the object function execution unit 150 to execute the object functions thereof, wherein the convolution operation will generate a convolution integral, and the comparison operation will compare the convolution integral with a threshold; and the third function is used for executing an image processing procedure to combine the sensed data array and the GUI image, and thereby deliver an output image DIMG to the display 140, wherein the image processing procedure includes performing a filtering operation and/or a color setting operation on the sensed data array.
The GUI 130 has a background area corresponding to the touch plane of the touch panel 110, wherein the background area contains a plurality of icons, with each of the icons representing an object.
The display 140 can be a liquid crystal display, an organic/non-organic LED display, or an e-paper display.
The object function execution unit 150 is used for playing a film, playing music, displaying figures, or connecting to a website, etc.
Preferably, the touch processing unit 120 utilizes a CPU/GPU to execute the touch processing procedure, so as to relieve the workload of the touch panel 110, and simplify the architecture of the touch panel 110. By using the powerful capacity of the CPU/GPU, the present invention's touch processing system can therefore execute relative image processing procedures in a high speed, so as to provide a novel touch control mechanism and multiple novel touch functions—including a function of multi-touch drawing, a function of handwriting input, a function of palm-touch control, and a function of multi-finger music performing.
In conclusion, the touch processing method and system using a GUI image of the present invention can relieve the workload of a touch panel, simplify the architecture of the touch panel, locate touched objects on a GUI more efficiently, and provide multiple novel touch functions. Compared to prior art touch processing method and system, the present invention possesses superior advantages.
While the invention has been described by way of example and in terms of preferred embodiments, it is to be understood that the invention is not limited thereto. To the contrary, it is intended to cover various modifications and similar arrangements and procedures, and the scope of the appended claims therefore should be accorded the broadest interpretation so as to encompass all such modifications and similar arrangements and procedures.
In summation of the above description, the present invention herein enhances the performance than the conventional structure and further complies with the patent application requirements and is submitted to the Patent and Trademark Office for review and granting of the commensurate patent rights.
Chen, Han-chang, Chia, Chung-Lin, Tu, Yen-Hung, Pan, Wen-Chieh
Patent | Priority | Assignee | Title |
10289206, | Dec 18 2015 | Intel Corporation | Free-form drawing and health applications |
9812104, | Aug 12 2015 | Samsung Electronics Co., Ltd. | Sound providing method and electronic device for performing the same |
Patent | Priority | Assignee | Title |
5511148, | Apr 30 1993 | Xerox Corporation | Interactive copying system |
6567109, | Jul 16 1999 | ACTIVISION PUBLISHING, INC | Automatic target enlargement for simplified selection |
6723929, | Apr 19 1995 | ELO TOUCH SOLUTIONS, INC | Acoustic condition sensor employing a plurality of mutually non-orthogonal waves |
6882337, | Apr 18 2002 | Microsoft Technology Licensing, LLC | Virtual keyboard for touch-typing using audio feedback |
7489306, | Dec 22 2004 | Microsoft Technology Licensing, LLC | Touch screen accuracy |
7499039, | Jan 10 2005 | 3M Innovative Properties Company | Iterative method for determining touch location |
7649807, | Aug 11 2004 | ELO TOUCH SOLUTIONS, INC | Method for locating an impact on a surface and device for implementing such a method |
7663607, | May 06 2004 | Apple Inc | Multipoint touchscreen |
7683890, | Apr 28 2005 | 3M Innovative Properties Company | Touch location determination using bending mode sensors and multiple detection techniques |
8013839, | Sep 06 2006 | Apple Inc. | Methods for determining a cursor position from a finger contact with a touch screen display |
8013846, | Feb 10 2006 | TPK HOLDING CO , LTD | Touch detection |
8269511, | Sep 08 2009 | Synaptics Incorporated | Sensing and defining an input object |
8378974, | Jul 02 2007 | ELO TOUCH SOLUTIONS, INC | Method and system for detecting touch events based on magnitude ratios |
8692810, | Apr 13 2005 | ELO TOUCH SOLUTIONS, INC | Method for determining the location of impacts by acoustic imaging |
8907930, | Aug 06 2010 | Google Technology Holdings LLC | Methods and devices for determining user input location using acoustic sensing elements |
8913039, | Jul 21 2009 | Commissariat a l Energie Atomique et aux Energies Alternatives; Centre National de la Recherche Scientifique | Method and device for locating at least one touch on a touch-sensitive surface of an object |
8988356, | Dec 31 2009 | GOOGLE LLC | Touch sensor and touchscreen user input combination |
9086764, | Jun 23 2009 | Sensitive Object | Method and a touch sensing device for implementing the method |
20020015064, | |||
20030189731, | |||
20040075671, | |||
20040136564, | |||
20040160421, | |||
20050012723, | |||
20050083313, | |||
20050146512, | |||
20050146513, | |||
20050174338, | |||
20060022956, | |||
20060139339, | |||
20060197750, | |||
20060279548, | |||
20070085821, | |||
20070120833, | |||
20070296829, | |||
20080122798, | |||
20080158169, | |||
20080168403, | |||
20080309632, | |||
20090116339, | |||
20090195517, | |||
20090309851, | |||
20100053105, | |||
20100079413, | |||
20100083190, | |||
20100116563, | |||
20100227689, | |||
20100229090, | |||
20100235793, | |||
20100288108, | |||
20100289754, | |||
20100315266, | |||
20100315372, | |||
20110080430, | |||
20110096010, | |||
20110115734, | |||
20110130096, | |||
20110134061, | |||
20110134067, | |||
20110137968, | |||
20110205185, | |||
20110261655, | |||
20110320978, | |||
20120062474, | |||
20120275642, | |||
20120293447, | |||
20130141344, | |||
20130286302, | |||
20140035871, | |||
20140139431, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Aug 31 2011 | TU, YEN-HUNG | RICH IP TECHNOLOGY INC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 027169 | /0681 | |
Aug 31 2011 | CHIA, CHUNG-LIN | RICH IP TECHNOLOGY INC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 027169 | /0681 | |
Aug 31 2011 | CHEN, HAN-CHANG | RICH IP TECHNOLOGY INC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 027169 | /0681 | |
Aug 31 2011 | PAN, WEN-CHIEH | RICH IP TECHNOLOGY INC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 027169 | /0681 | |
Nov 03 2011 | RICH IP TECHNOLOGY INC. | (assignment on the face of the patent) | / |
Date | Maintenance Fee Events |
Mar 30 2020 | M2551: Payment of Maintenance Fee, 4th Yr, Small Entity. |
Jul 01 2024 | REM: Maintenance Fee Reminder Mailed. |
Dec 16 2024 | EXP: Patent Expired for Failure to Pay Maintenance Fees. |
Date | Maintenance Schedule |
Nov 08 2019 | 4 years fee payment window open |
May 08 2020 | 6 months grace period start (w surcharge) |
Nov 08 2020 | patent expiry (for year 4) |
Nov 08 2022 | 2 years to revive unintentionally abandoned end. (for year 4) |
Nov 08 2023 | 8 years fee payment window open |
May 08 2024 | 6 months grace period start (w surcharge) |
Nov 08 2024 | patent expiry (for year 8) |
Nov 08 2026 | 2 years to revive unintentionally abandoned end. (for year 8) |
Nov 08 2027 | 12 years fee payment window open |
May 08 2028 | 6 months grace period start (w surcharge) |
Nov 08 2028 | patent expiry (for year 12) |
Nov 08 2030 | 2 years to revive unintentionally abandoned end. (for year 12) |