US20100153891A1 - Method, device and program for browsing information on a display - Google Patents
Method, device and program for browsing information on a display Download PDFInfo
- Publication number
- US20100153891A1 US20100153891A1 US12/710,944 US71094410A US2010153891A1 US 20100153891 A1 US20100153891 A1 US 20100153891A1 US 71094410 A US71094410 A US 71094410A US 2010153891 A1 US2010153891 A1 US 2010153891A1
- Authority
- US
- United States
- Prior art keywords
- display
- hand
- held device
- display device
- point
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/14—Digital output to display device ; Cooperation and interconnection of the display device with other functional units
- G06F3/1407—General aspects irrespective of display type, e.g. determination of decimal point position, display with fixed or driving decimal point, suppression of non-significant zeros
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F15/00—Digital computers in general; Data processing equipment in general
- G06F15/02—Digital computers in general; Data processing equipment in general manually operated with input through keyboard and computation using a built-in program, e.g. pocket calculators
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/1626—Constructional details or arrangements for portable computers with a single-body enclosure integrating a flat display, e.g. Personal Digital Assistants [PDAs]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/1633—Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
- G06F1/1684—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675
- G06F1/1686—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675 the I/O peripheral being an integrated camera
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/1633—Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
- G06F1/1684—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675
- G06F1/1694—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675 the I/O peripheral being a single or a set of motion sensors for pointer control or gesture input obtained by sensing movements of the portable computer
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2200/00—Indexing scheme relating to G06F1/04 - G06F1/32
- G06F2200/16—Indexing scheme relating to G06F1/16 - G06F1/18
- G06F2200/161—Indexing scheme relating to constructional details of the monitor
- G06F2200/1614—Image rotation following screen orientation, e.g. switching from landscape to portrait mode
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2200/00—Indexing scheme relating to G06F1/04 - G06F1/32
- G06F2200/16—Indexing scheme relating to G06F1/16 - G06F1/18
- G06F2200/163—Indexing scheme relating to constructional details of the computer
- G06F2200/1637—Sensing arrangement for detection of housing movement or orientation, e.g. for controlling scrolling or cursor movement on the display of an handheld computer
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2320/00—Control of display operating conditions
- G09G2320/02—Improving the quality of display appearance
- G09G2320/0261—Improving the quality of display appearance in the context of movement of objects on the screen or movement of the observer relative to the screen
Definitions
- the present invention relates to display devices where information can be browsed.
- the present invention relates to a novel and improved method and system for browsing information with hand-held devices with a display device.
- Various electronic mobile devices e.g. mobile phones, computers, Personal Digital Assistants (PDA), comprise displays. The transfer of the information to be viewed on the display is executed at least partially by a processor.
- a device typically comprises also a keypad with which the user of the device enters various commands. There are also touch-sensitive displays (touch screens). There a separate keypad is not needed. A device is controlled by touching the touch screen.
- the display of a mobile device is capable of showing only limited amount of information at a time. Because of the size of the display, e.g. a large image must be viewed part by part. In order to view such an image, the user of the device controls the display, e.g. by scrolling the display with a mouse etc.
- Devices equipped with a display have different kinds of user interfaces with which the user interacts with the device.
- a graphical user interface can be controlled with various control devices including, for example, keypad, touch screen, different kinds of cursor controlling methods, etc.
- An objective of the present invention is to adjust the view on a display device in a manner as natural as possible so that the user of the hand-held device can concentrate on the information displayed on the display device and not on the adjustment of the displayed information.
- the objective is achieved by a method, hand-held device and computer program for browsing information on a display device of a hand-held device.
- the display device is coupled to a processor mapping the information content generated by the processor into the virtual data object suitable for conveying the information to the user of the hand-held device.
- the display device displays a portion of the virtual data object at a time on the display device.
- the virtual data object comprises e.g. characters, pictures, lines, links, video or pixels that can be conveniently displayed on the display device at a time.
- the idea of the present invention is to browse information on the display device of a hand-held device naturally and logically.
- Characteristic of the invention is that information is browsed on the display device essentially in a mirror-like way. In other words, the portion of the virtual data object displayed on the display device is moved at the same direction as the hand-held device is tilted. In other words, the movements of the portion of the virtual data object displayed on the display device depend on the orientation of the hand-held device.
- An important feature of the invention is also that a certain orientation of the hand-held device always displays the same portion of the virtual data object on the display device.
- the browsing method described above is extremely logical, and the movements and responses to the movements are natural.
- the core functions of the browsing can be explained by means of the following example.
- the information is browsed with the hand-held device essentially in the same way as looking at a view from a hand mirror.
- the hand mirror is typically held in hand quite close to the viewer.
- the hand mirror represents the display device and the view behind the viewer the virtual data object. When the hand mirror is tilted, the view behind the viewer moves in response to the changes in the orientation of the hand mirror.
- the present invention is most applicable with hand-held devices with a display when a large data object is displayed by parts on the display.
- a large data object can be browsed naturally and logically from the user's perspective.
- the position memory of the muscles of a human body makes it easier to return to previously browsed points and to the starting point.
- the present invention also reduces the need to use exterior mechanical switches, keypad or other known control mechanisms for browsing information on the display device. Therefore the use of a hand-held device is easier and simpler.
- the basic functionalities of the present invention can be implemented with mass production components, and with moderate processing power. Thus, the features described in the present invention can be taken in use in consumer products without notable expense increase.
- FIG. 1 illustrates how the hand-held device is operated according to the present invention
- FIGS. 2 a , 2 b and 2 c illustrate more specific examples of how the hand-held device of FIG. 1 is handled
- FIG. 3 illustrates an exemplary viewing setup of the present invention
- FIG. 4 illustrates an example of how a view on the display device can be formed and calculated according to the viewing setup of FIG. 3 ,
- FIG. 5 is a block diagram illustrating an embodiment of the hand-held device in accordance with the present invention.
- FIG. 6 is a block diagram illustrating another embodiment of the hand-held device in accordance with the present invention.
- FIGS. 7 a , 7 b , 7 c and 7 d illustrate the view change of the display of the hand-held device in response to user actions
- FIGS. 8 a , 8 b and 8 c illustrate different ways of browsing information
- FIG. 9 is a flow diagram illustrating the operation of a preferred embodiment of the present invention.
- FIGS. 10 a - 10 d illustrate another example of how a view on the display device can be formed and calculated according to the viewing set up of FIG. 4 .
- FIG. 1 illustrates a simplified portable hand-held device according to the present invention.
- the hand-held device is e.g. a mobile phone or a Personal Digital Assistant (PDA).
- the display device of the hand-held device displays information stored on a memory of the hand-held device.
- the hand-held device is explained more specifically in later examples.
- FIG. 1 represents the basic browsing functionality. Information is browsed on the display device by tilting (rotating) the hand-held device 40 towards directions 2 , 3 , 4 , and 5 around the axis 6 and 7 .
- the memory of the hand-held device 40 comprises a virtual data object comprising characters, pictures, lines, links, video or pixels that can be conveniently displayed on the display device at a time. A portion of the virtual data object displayed on the display device is moved at the same direction as the hand-held device is tilted. Moreover, a certain orientation of the hand-held device 40 always displays the same portion of the virtual data object on the display device.
- FIGS. 2 a , 2 b and 2 c represent a more specific example of tilting the hand-held device 40 .
- a typical starting situation is that the hand-held device 40 is in a 20-30 degree angle with the horizontal plane 8 .
- This plane is in one embodiment set as a default xy-plane from which the rotation angles of the hand-held device 40 are measured. It can also be said that this starting point is the most appropriate one for viewing information with the display device. So when the user tilts the hand-held device 40 , the viewing angle changes. The view on the display device changes in real time to correspond to the new viewing angle.
- a very important feature of the invention is that the view on the display device depends on the viewing angle, and the same viewing angle displays always the same view on the display device. This feature is very natural and logical.
- angle a corresponds to the aforementioned 20-30 degrees.
- FIG. 2 a is regarded as a starting position when the browsing begins.
- the hand-held device 40 has been tilted to an angle ⁇ 1 , which is smaller than angle ⁇ .
- the view on the display device changes based on the tilting movements essentially in real time, and the movement of the information on the display device is towards the same direction as the hand-held device 40 is tilted.
- the hand-held device 40 is tilted to an angle ⁇ 2 , which is bigger than angle ⁇ .
- the angle ( ⁇ ) is a predetermined angle, and it is determined by the manufacturer of the hand-held device 40 . In the determination process it is defined that the display view plane is based on axis x_VD and y_VD, which are perpendicular to each other.
- the hand-held device is then set to a certain position ( ⁇ ), and that position is set as a default xy-plane.
- the default plane is determined based on angle a.
- the default plane can be freely determined based on any x-axis, y-axis and/or z-axis.
- the hand-held device 40 is tilted in respective to this plane.
- the default xy-plane is fixed, the user of the hand-held device is always capable of returning to a certain view by tilting the device back to the original orientation when the sensors measuring the orientation of the hand-held device do not cause any restrictions to the measured position.
- the angle ⁇ can be readjusted to a desired value.
- FIGS. 3 and 4 represent an exemplary embodiment of the setup of a “mirroring system”. It includes a viewpoint VP, a virtual screen VS and a virtual display VD.
- the viewpoint VP represents the location of a viewer of a hand-held device.
- the VD represents the display device of the hand-held device.
- the virtual screen represents the actual information browsed on the display device.
- the viewpoint VP is defined to be at point [0 0 0].
- the orientation of the virtual display VD is defined by tilting angles ⁇ x , ⁇ y , ⁇ z indicating rotation angle over each coordinate axe.
- the virtual display VD is a plane and has some size.
- FIGS. 3 and 4 represent only one embodiment of the possible positions of the VS, VP and VD, and the axes used.
- PT 1 R x ⁇ R y ⁇ R z ⁇ L
- PT 2 R x ⁇ R y ⁇ R z ⁇ M
- PNT PT 1 ⁇ PT 2 ⁇ ( cross ⁇ ⁇ product )
- PN PNT ⁇ PNT ⁇
- PN is the unit normal vector of the VD-plane.
- the PN defines the applicable orientation of the VD to be used in the projection calculation.
- the “image” on the virtual display VD is calculated. Let's assume that there is a vector beginning from the VP and being reflected via the VD. The point where the reflected vector hits on the plane VS defines the projection of the point on the VS to the point on the VD-plane. Hence, if all points on VD are processed as described above, the image on the VD can be defined.
- the points P and VP define a vector A.
- the points Q and VP define a vector B.
- the point defined as sum of the VP and 2*B defines a point R.
- the direction vector that goes via P and R defines a direction vector that hits the plane VS at point S.
- P_xyz is the coordinate of the middle point of the VD
- peili_y is the y-coordinate on the VD plane-coordinate system
- peili_z is the z-coordinate on the VD plane-coordinate system
- the point R can be defined (the reason for the factor 2 is that in mirror the arriving and departing light beam have equal angles compared to the normal vector of the surface).
- direction vector C is defined as follows:
- the system of FIG. 4 has several characteristics:
- the view on the display device moves into the same direction as it is tilted.
- the movement of the portion of the virtual data object displayed on the display device is proportional to the change amount and/or rate of the rotational movement.
- the same tilting angle causes greater movements on the virtual screen VS.
- the browsing speed of the information on the display device increases as the distance between the VP and VD increases.
- this movement factor can be adjusted by the user of the hand-held device.
- the view on the display device depends on the position and orientation of the VS, VP and VD.
- a certain VS-VP-VD position/orientation combination always constitute the same view on the display device.
- Zooming can be implemented by changing the position of the VS, VP and VD.
- Zooming can be implemented by enlarging the object on the VS or altering the radius of curvature of the mirror (VD).
- the present invention does not have to implement all the aforementioned features, but the most appropriate ones can be chosen.
- the ideal mirror-like functionality means that the information on the display device changes when:
- the information on the display device is changed at least either according to a) or b). If only a) or b) is taken into consideration, the operation of the display is not so mirror-like as if both a) and b) were implemented. In one embodiment, the display device operates according to all a), b) and c).
- FIGS. 10 a - d illustrate another example of calculation which is explained with reference to FIG. 4 .
- FIG. 10 c is a side view of FIG. 10 a
- FIG. 10 d is a side view of FIG. 10 b .
- the virtual screen is referred to as the virtual surface 200 .
- the default orientation of the display 201 is determined to be parallel with the yz-plane.
- the virtual surface (VS) 200 is above the display plane and also parallel with the yz-plane.
- a page having information to be browsed lies on the virtual surface 200 , and the size of the page is larger than the size of the display 201 .
- the reference point VP is on the virtual surface 200 .
- the x-axis (not shown) runs through the reference point VP and the middle point P of the display 201 .
- the relationship between every single point in the area ( 2 a * 2 b ) of the display 201 and the corresponding area ( 2 a * 2 b ) on the virtual surface 200 can be calculated in a similar way.
- the portion of the page ( 2 a * 2 b ) that is to be displayed then has a shape similar to the shape of the display ( 2 a * 2 b ).
- on the point S on the virtual surface 200 is the middle point of the determined rectangle 2 a * 2 b and all the other points residing around point S within the rectangle relate to the corresponding the points residing around point P on the display 201 . That portion of the page surrounding point S on the virtual surface is displayed on display 201 .
- FIGS. 10 b and 10 d the display 201 has been tilted around the y-axis, wherein the portion of the page shown on the display 201 changes in the following way:
- a reference line 203 drawn between point P and point S meets the x-axis, i.e. it is parallel with the x-axis.
- the normal of the display extending from point P is parallel with the x-axis and the reference line 203 .
- the normal 204 of the display is also tilted by angle a with respect to the x-axis.
- the reference line 203 is mirrored with respect to the normal 204 of the display wherein a mirror line 205 is generated.
- a hit point S′ is the point where the mirror line 205 hits the virtual surface 200 .
- an area (shape) of the page corresponding to the area (shape) of the display is determined.
- the display 201 then shows the portion of the page around the hit point S′ and having a shape similar to the shape of the display 201 .
- FIG. 5 represents one example of a preferred hand-held device 40 .
- the hand-held device 40 is e.g. a mobile phone.
- the hand-held device comprises a processor 30 and a display device 10 coupled to the processor 30 .
- the data memory 60 and the program memory 70 are also coupled to the processor 30 .
- the program memory 70 contains e.g. the operation system. The sizes of the memories, and the processing power of the processor 30 depend on the device and application used.
- the program memory 60 can additionally contain different kinds of software applications with which various tasks can be executed.
- Application software comprise e.g. word processing, graphical and spreadsheet software. The software applications and data used by them are loaded into the data memory 60 in order to be able to use the software.
- the display adapter 90 with the processor 30 controls the display device 10 .
- the display adapter 90 comprises a data buffer in which the information to be displayed on the display device 10 is stored.
- the hand-held device 40 comprises measuring means which in a preferred embodiment of the invention refer to acceleration sensor(s) 50 .
- the acceleration sensor(s) 50 it is possible to measure tilting movements of the hand-held device 40 .
- the processor 30 receives the measurement results and interprets them.
- the acceleration sensor(s) 50 can be e.g. piezo-electric or capacitive producing an analog voltage which is proportional to the acceleration factor.
- the acceleration sensor(s) 50 it is possible to measure one, two or three-dimensional accelerations.
- the measurement of tilting movements is based on the fact that the highest acceleration is parallel to the gravity of the earth. Therefore, the orientation of the hand-held device 40 can be defined in relation to the earth. It is also possible to use gyroscopes with its various forms to measure the orientation of the hand-held device 40 .
- the quantities measured are e.g. tilting angle and accelerations.
- the relation information between the rotation degree of the hand-held device and the memory address corresponding to the displayed view is stored e.g. on the data memory 60 .
- the processor 30 defines the orientation of the hand-held device 40 in relation to the user or a reference position.
- the processor 30 may also define the distance between the user and the hand-held device 40 or the user orientation in relation to the hand-held device 40 .
- the memory space can be implemented logically, e.g. as a two-dimensional memory space.
- the processor 30 starts the definition process of the new memory address from the current memory address so that displacement in the memory space corresponds to the direction and amount of change in orientation according to the relation information.
- the hand-held device 40 comprises also a browse lock 80 with which it is signaled when the browsing is executed.
- the orientation of the hand-held device 40 must remain in the same position in order to keep the view on the display device unchanged.
- the hand-held device 40 comprises a lock feature, e.g. a push-button, with which the browsing can be locked. The user can tilt the hand-held device back to an appropriate viewing orientation in order to view the information on the display device 10 properly. The browsing may then continue when the button is released.
- the hand-held device 40 in FIG. 6 is almost the same as the hand-held device 40 in FIG. 5 .
- the hand-held device comprises also a locator 20 . It is possible to control the view on the display device 10 also by other means than acceleration sensor(s) or equivalent means.
- the hand-held device 40 can comprise e.g. a (video) camera measuring the orientation and location of the hand-held device in relation to the user of the hand-held device 40 or to another reference point in the surroundings of the user.
- the camera 20 may be set to recognize and measure distance to a certain reference point, e.g. the eyes of the user. Therefore, when the orientation and/or position of the hand-held device 40 changes, the viewing angle measured by the camera also changes. Thus, it can be concluded that the hand-held device 40 has been tilted and/or moved towards some direction.
- the orientation of the hand-held device 40 in proportion to the reference point and the distance of the hand-held device 40 to the reference point tens of times within a second.
- the browsing functionality can be implemented merely using the video camera, so that additional acceleration sensor(s) are not necessarily needed.
- the measuring of the distance can also be implemented with an ultrasonic radar connected through an analog-digital converter to the processor 30 of the hand-held device 40 .
- the information on the display device 10 is essentially browsed in the same manner as when looking in a mirror. In other words, the view on the display 10 depends on the viewing angle in relation to the display device plane as the view in a mirror depends on the viewing angle to the mirror.
- the locator 20 comprises a video camera seeking the location of the head and eyes of the user. Heuristic algorithms and neural network seeking the location of the head and eyes can be used. Acceleration sensors are more appropriate to use in hand-held devices than a video camera, because they are cheaper. The acceleration sensors may also be a more appropriate solution in devices which do not have a built-in video camera for a default feature, e.g. in the (third generation) mobile phones.
- the advantage of the use of the video camera is that the use of the hand-held device is not restricted to the position of the hand-held device, e.g. when being on one's back the hand-held device can be used without problems.
- the display device surface level is set as an xy-plane.
- a certain relation between the x-axial and/or y-axial movement of the hand-held device and the amount of the displacement of the portion of the virtual data object displayed on the display device at a time has been determined. So, when the hand-held device 40 is moved along x- and/or y-axis, the portion of the virtual data object displayed on the display device moves in the same direction as the hand-held device is moved in the xy-plane according to the relation information.
- the processor 30 comprises also means for filtering the x-axial, y-axial and/or tilting movements before displaying the movements on the display device. Therefore, minor unintentional movements can be filtered out.
- the relation between the tilting movements and the amount of the displacement of the portion of the virtual data object displayed on the display device at a time can be changed. Therefore, a user may define e.g. that from now on a 10 degree tilting causes the same effect on the display as a 15 degree tilting earlier.
- the relation is linear. In other words, the relation between the tilting movements and the amount of the displacement of the portion of the virtual data object displayed on the display device at a time does not depend on the amount of the tilting.
- the relation is linear, but e.g. exponential.
- the amount of the displacement of the portion of the virtual data object displayed on the display device at a time depends on the amount of the tilting. For example, the value of the relation factor changes (e.g. exponentially) as the tilting amount increases.
- FIGS. 7 a - 7 d represent the situation where the size of the information on the display device depends on the zoom factor in addition to the orientation of the hand-held device.
- the zoom factor can be controlled in different ways. In one embodiment, the zoom factor depends on the distance between the user and the hand-held device.
- FIG. 7 a represent the display device 10 , on which graphical reference numerals 21 , 22 and 23 are seen. The view on the display device 10 depends on the orientation of the hand-held device or the viewing angle from which the user of the hand-held views the display device.
- a graphical FIG. 21 grows as depicted in FIGS. 7 b and 7 c .
- the zoom factor has decreased, and also the viewing angle between the user and the hand-held device has changed.
- the zoom factor can be modified with several different ways.
- the zoom factor depends on the distance between the reference point (e.g. the eyes of the user) and the hand-held device. When the distance decreases, a graphical FIG. 21 grows, and vice versa.
- the display device 10 may have to be set to a zoom mode before the zoom factor changes. If the zoom factor was all the time dependent on the distance between the reference point and the hand-held device, the browsing operation would not necessarily be practical because the view on the display 10 would change whenever the aforementioned distance changes.
- the zoom factor changes when rotating the hand-held device around the axis being essentially perpendicular to a predefined xy-plane.
- the xy-plane may be the present plane of the display device 10 or some other predetermined plane.
- the zoom factor is changed by tilting the hand-held device. Before this the display device must be set into a zoom mode. When the hand-held device is tilted, e.g. to the right the zoom factor increases, and when the hand-held device is tilted to the left, the zoom factor decreases. It is not important which predefined tilting directions are used but that the two directions can be separated sufficiently from each other.
- the aforementioned zoom mode is set on and off e.g. with a predetermined button of the hand-held device.
- FIGS. 8 a - 8 c represent different ways to implement the user interface.
- the display device 10 of the hand-held device 40 contains information to be viewed by the user.
- an A letter is on the display device 10 .
- the information on the display device 10 remains in the same position with respect to the user when the hand-held device 40 is rotated around the axis being perpendicular to the display surface plane, as depicted in FIG. 8 b .
- the information on the display device 10 remains in the same position because the information is attached to the real physical coordinates.
- the information on the display device 10 remains in the same position with respect to the hand-held device 40 when the hand-held device 40 is rotated around the axis being perpendicular to the display surface plane, as depicted in FIG. 8 c .
- the orientation of the information on the display device 10 changes with respect to the user of the hand-held device 40 because the information is not attached to the real physical coordinates but to the display device.
- FIG. 9 represents a flow diagram describing the functionality of a method of the present invention.
- FIG. 9 describes a hand-held device 40 comprising means for measuring acceleration 50 and a processor 30 .
- Means for measuring acceleration refer e.g. to a multiaxial acceleration sensor suited for measuring changes in the orientation of the hand-held device 40 .
- the hand-held device is switched on, and it is ready for browsing information on the display device, as represented in phase 100 .
- the acceleration sensor 50 measures constantly acceleration readings.
- the processor 30 receives the acceleration readings and defines the orientation of the hand-held device and also the change in the orientation compared to the prior measurement(s), as represented in phases 101 and 102 .
- phase 103 it is tested whether the browsing is on or off. If the browsing is off, the processor 30 examines if a predetermined browsing startup condition is fulfilled (phase 104 ). If it is not fulfilled, the method returns back to phase 101 . It means that the orientation of the hand-held device has not changed sufficiently, which would indicate that the user wishes to browse information on the display device of the hand-held device.
- the processor 30 sets the browsing as started (phase 106 ) and determines the browsing speed based on the current acceleration value (phase 108 ).
- the processor 30 also changes the information presented on the display device according to a relation between the rotation degree and the amount of the displacement of the portion on the virtual data object stored on the data memory 60 and the determined browsing speed (phase 108 ).
- a certain orientation of the hand-held device always causes the same view (the same portion on the virtual data object stored on the memory) on the display device. If it is observed in phase 103 that the browsing is already on, and the browsing stopping condition is fulfilled (phase 105 ), the processor 30 stops the browsing and sets the browsing as stopped (phases 107 and 109 ). If it is observed that the browsing stopping condition is not fulfilled (phase 105 ), the processor 30 returns back to phase 101 .
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Computer Hardware Design (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Computing Systems (AREA)
- User Interface Of Digital Computer (AREA)
- Position Input By Displaying (AREA)
- Calculators And Similar Devices (AREA)
- Financial Or Insurance-Related Operations Such As Payment And Settlement (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
In one embodiment, a program is provided for browsing information on a hand-held device having a display. The program includes (1) displaying a portion of the page residing around the predefined point and having a shape similar to the shape of the display, (2) generating a mirror line by mirroring the reference line in relation to a line that is perpendicular to the display surface and travels via the reference point in response to tilting of the hand-held device in relation to the spatial initial state, (3) defining a hit point (xn,yn) where the mirror line hits the virtual surface and the page containing information; and (4) showing on the display at least a portion of the page around the hit point and having a shape similar to the shape of the display, the position of the hit point on the page to correspond to the position of the reference point on the display.
Description
- This application is a continuation of U.S. Pat. No. 7,607,111, issued Oct. 20, 2009, which is a continuation-in-part of abandoned U.S. patent application Ser. No. 10/071,172, filed Feb. 8, 2002, which claims the benefit of and priority to Finnish Patent Application Serial No. 2001/1039, filed May 16, 2001, the contents of all of which are incorporated by reference herein in their entirety.
- The present invention relates to display devices where information can be browsed. In particular, the present invention relates to a novel and improved method and system for browsing information with hand-held devices with a display device.
- Various electronic mobile devices, e.g. mobile phones, computers, Personal Digital Assistants (PDA), comprise displays. The transfer of the information to be viewed on the display is executed at least partially by a processor. A device typically comprises also a keypad with which the user of the device enters various commands. There are also touch-sensitive displays (touch screens). There a separate keypad is not needed. A device is controlled by touching the touch screen.
- The display of a mobile device is capable of showing only limited amount of information at a time. Because of the size of the display, e.g. a large image must be viewed part by part. In order to view such an image, the user of the device controls the display, e.g. by scrolling the display with a mouse etc.
- Devices equipped with a display have different kinds of user interfaces with which the user interacts with the device. There are graphical user interfaces and speech controlled user interfaces. A graphical user interface can be controlled with various control devices including, for example, keypad, touch screen, different kinds of cursor controlling methods, etc.
- There are, however, drawbacks in the prior-art devices in the usability of the device, especially in the browsing of information with the device. When the information to be viewed on the display must be viewed by parts, it is difficult and slow to browse the whole information part by part. It is, for example, difficult to display a wide panorama picture on the display, while at the same time quickly and easily browsing the picture.
- For the user of a mobile hand-held device it is difficult to perceive visual entireties that can not be displayed at a time on the display. Therefore the browsing of the information should be carried out as naturally and logically as possible. A user of a mobile hand-held device must be able to learn and use the device easily and efficiently.
- From prior-art solutions it is known to use location detectors for browsing information with a device. Reference publication WO 9918495 (Telefonaktiebolaget LM Ericsson) describes a method where the display device is moved essentially in the plane of the display device, whereby different parts of a complete screen image are shown on said display device. When the display device is moved essentially in a direction perpendicular to the plane of the display device, the magnification of the screen image changes. The movement in the plane is a bit problematic. In the plane movement the necessary movements may be quite remarkable/large, and it may be difficult to maintain the display device in a proper position for reading or browsing.
- Another prior-art solution is to use tilt detectors for moving, or to be more specific, for scrolling the view on the display device. One solution of this kind is described in WO 9814863 (Philips). When the screen image is moved by scrolling (tilting the display device), the result is better than in moving the display device in the plane of the display device, as described above. However, to move the screen image fluently and to return from some point to the initial point of browsing is difficult because controlling a discontinuous motion requires continuous and precise handling of the display device. The controlling of the scrolling movement can be compared to a movement of a ball on a plane surface by tilting the plane. In order to stop the rolling of the ball, the plane surface must be perpendicular against the gravity of the earth. In other words, the control of the movements and usability are not at an acceptable level so that the use of such a device would be natural and logical.
- There are also various kinds of motion and/or location controlled display devices used in, e.g. in virtual helmets. There the display device focuses like a virtual camera. The display device displays an object to which the device (camera) points in the modeled virtual environment. To use a virtual camera model in a hand-held device is not so straightforward because displaying peripheries of a large screen image results in a disadvantageous viewing angle. Therefore, the adjustment and zooming of a display image must be implemented in a most natural and logical manner. In prior-art solutions the browsing of information on the display device is slow and awkward because the solutions are based on artificial logic.
- An objective of the present invention is to adjust the view on a display device in a manner as natural as possible so that the user of the hand-held device can concentrate on the information displayed on the display device and not on the adjustment of the displayed information.
- The objective is achieved by a method, hand-held device and computer program for browsing information on a display device of a hand-held device. In the present invention, the display device is coupled to a processor mapping the information content generated by the processor into the virtual data object suitable for conveying the information to the user of the hand-held device. The display device displays a portion of the virtual data object at a time on the display device. The virtual data object comprises e.g. characters, pictures, lines, links, video or pixels that can be conveniently displayed on the display device at a time.
- The idea of the present invention is to browse information on the display device of a hand-held device naturally and logically. Characteristic of the invention is that information is browsed on the display device essentially in a mirror-like way. In other words, the portion of the virtual data object displayed on the display device is moved at the same direction as the hand-held device is tilted. In other words, the movements of the portion of the virtual data object displayed on the display device depend on the orientation of the hand-held device. An important feature of the invention is also that a certain orientation of the hand-held device always displays the same portion of the virtual data object on the display device. The browsing method described above is extremely logical, and the movements and responses to the movements are natural.
- The core functions of the browsing can be explained by means of the following example. The information is browsed with the hand-held device essentially in the same way as looking at a view from a hand mirror. The hand mirror is typically held in hand quite close to the viewer. The hand mirror represents the display device and the view behind the viewer the virtual data object. When the hand mirror is tilted, the view behind the viewer moves in response to the changes in the orientation of the hand mirror.
- When approaching the functionality of a hand mirror the browsing of information on a display device of a hand-held device is made natural and logical.
- The present invention is most applicable with hand-held devices with a display when a large data object is displayed by parts on the display. With the present invention, a large data object can be browsed naturally and logically from the user's perspective. The position memory of the muscles of a human body makes it easier to return to previously browsed points and to the starting point.
- The present invention also reduces the need to use exterior mechanical switches, keypad or other known control mechanisms for browsing information on the display device. Therefore the use of a hand-held device is easier and simpler. The basic functionalities of the present invention can be implemented with mass production components, and with moderate processing power. Thus, the features described in the present invention can be taken in use in consumer products without notable expense increase.
- Other objects and features of the present invention will become apparent from the following detailed description considered in conjunction with the accompanying drawings. It is to be understood, however, that the drawings are designed solely for purposes of illustration and not as a definition of the limits of the invention, for which reference should be made to the appended claims. It should be further understood that the drawings are not necessarily drawn to scale and that, unless otherwise indicated, they are merely intended to conceptually illustrate the structures and procedures described herein.
- The above-mentioned features and objects of the present disclosure will become more apparent with reference to the following description taken in conjunction with the accompanying drawings wherein like reference numerals denote like elements and in which:
-
FIG. 1 illustrates how the hand-held device is operated according to the present invention, -
FIGS. 2 a, 2 b and 2 c illustrate more specific examples of how the hand-held device ofFIG. 1 is handled, -
FIG. 3 illustrates an exemplary viewing setup of the present invention, -
FIG. 4 illustrates an example of how a view on the display device can be formed and calculated according to the viewing setup ofFIG. 3 , -
FIG. 5 is a block diagram illustrating an embodiment of the hand-held device in accordance with the present invention, -
FIG. 6 is a block diagram illustrating another embodiment of the hand-held device in accordance with the present invention, -
FIGS. 7 a, 7 b, 7 c and 7 d illustrate the view change of the display of the hand-held device in response to user actions, -
FIGS. 8 a, 8 b and 8 c illustrate different ways of browsing information, -
FIG. 9 is a flow diagram illustrating the operation of a preferred embodiment of the present invention, and -
FIGS. 10 a-10 d illustrate another example of how a view on the display device can be formed and calculated according to the viewing set up ofFIG. 4 . -
FIG. 1 illustrates a simplified portable hand-held device according to the present invention. The hand-held device is e.g. a mobile phone or a Personal Digital Assistant (PDA). The display device of the hand-held device displays information stored on a memory of the hand-held device. The hand-held device is explained more specifically in later examples.FIG. 1 represents the basic browsing functionality. Information is browsed on the display device by tilting (rotating) the hand-helddevice 40 towardsdirections axis 6 and 7. The memory of the hand-helddevice 40 comprises a virtual data object comprising characters, pictures, lines, links, video or pixels that can be conveniently displayed on the display device at a time. A portion of the virtual data object displayed on the display device is moved at the same direction as the hand-held device is tilted. Moreover, a certain orientation of the hand-helddevice 40 always displays the same portion of the virtual data object on the display device. -
FIGS. 2 a, 2 b and 2 c represent a more specific example of tilting the hand-helddevice 40. It can be said that a typical starting situation is that the hand-helddevice 40 is in a 20-30 degree angle with thehorizontal plane 8. This plane is in one embodiment set as a default xy-plane from which the rotation angles of the hand-helddevice 40 are measured. It can also be said that this starting point is the most appropriate one for viewing information with the display device. So when the user tilts the hand-helddevice 40, the viewing angle changes. The view on the display device changes in real time to correspond to the new viewing angle. A very important feature of the invention is that the view on the display device depends on the viewing angle, and the same viewing angle displays always the same view on the display device. This feature is very natural and logical. - In
FIG. 2 a, angle a corresponds to the aforementioned 20-30 degrees.FIG. 2 a is regarded as a starting position when the browsing begins. InFIG. 2 b, the hand-helddevice 40 has been tilted to an angle β1, which is smaller than angle α. The view on the display device changes based on the tilting movements essentially in real time, and the movement of the information on the display device is towards the same direction as the hand-helddevice 40 is tilted. InFIG. 2 c, the hand-helddevice 40 is tilted to an angle β2, which is bigger than angle α. - In one embodiment, the angle (α) is a predetermined angle, and it is determined by the manufacturer of the hand-held
device 40. In the determination process it is defined that the display view plane is based on axis x_VD and y_VD, which are perpendicular to each other. The hand-held device is then set to a certain position (α), and that position is set as a default xy-plane. InFIGS. 2 a, 2 b and 2 c, the default plane is determined based on angle a. In another embodiment, the default plane can be freely determined based on any x-axis, y-axis and/or z-axis. - From that moment on, the hand-held
device 40 is tilted in respective to this plane. When the default xy-plane is fixed, the user of the hand-held device is always capable of returning to a certain view by tilting the device back to the original orientation when the sensors measuring the orientation of the hand-held device do not cause any restrictions to the measured position. In another embodiment, the angle α can be readjusted to a desired value. -
FIGS. 3 and 4 represent an exemplary embodiment of the setup of a “mirroring system”. It includes a viewpoint VP, a virtual screen VS and a virtual display VD. The viewpoint VP represents the location of a viewer of a hand-held device. The VD represents the display device of the hand-held device. The virtual screen represents the actual information browsed on the display device. - For simplicity in the following the viewpoint VP is defined to be at point [0 0 0]. Furthermore, the middle point of the virtual display VD is defined to be at P_xyz wherein P_xyz=[P_xyz1 P_xyz2 P_xyz3]T, and the virtual screen VS to be at plane x=kuva_shift.
- The orientation of the virtual display VD is defined by tilting angles αx, αy, αz indicating rotation angle over each coordinate axe. In
FIG. 4 , the virtual display VD is a plane and has some size. Each coordinate in this VD plane is defined using notation P=[P_xyz2+peili_y P_xyz3+peili_z] when the orientation of the VD is defined to be parallel with the x-plane. - It must be noted that
FIGS. 3 and 4 represent only one embodiment of the possible positions of the VS, VP and VD, and the axes used. - In order to the determine the orientation of the VD, two orthogonal vectors (in the x-plane) are defined as follows:
-
L=[0, 1, −1]T -
M=[0, 1, 1]T - Those vectors present the orthogonal direction vectors of the VD. Next, the orientation of the virtual display VD is defined using the rotation angles:
-
- Next the unit normal vector of the VD is calculated:
-
- where PN is the unit normal vector of the VD-plane. The PN defines the applicable orientation of the VD to be used in the projection calculation.
- Next, the “image” on the virtual display VD is calculated. Let's assume that there is a vector beginning from the VP and being reflected via the VD. The point where the reflected vector hits on the plane VS defines the projection of the point on the VS to the point on the VD-plane. Hence, if all points on VD are processed as described above, the image on the VD can be defined.
- The idea of calculation is presented using vectors in
FIG. 4 . Using the vectors the algorithm works as follows: - 1. The points P and VP define a vector A.
- 2. The projection proj of the vector A on the normal vector PN is calculated.
- 3. The sum of the vector A and proj*PN defines a point Q.
- 4. The points Q and VP define a vector B.
- 5. The point defined as sum of the VP and 2*B defines a point R.
- 6. The direction vector that goes via P and R defines a direction vector that hits the plane VS at point S.
- 7. The result of this process is that the image of point P in VD is the image of point S in VS.
- By repeating phases 1-7 for all points in the VD-plane the whole image of the virtual display VD is defined. Using vector calculation the same can be presented as follows:
- First the point P is defined:
-
P=P — xyz+R x R y R z[0 peili — y peili — z] T - where P_xyz is the coordinate of the middle point of the VD, peili_y is the y-coordinate on the VD plane-coordinate system and peili_z is the z-coordinate on the VD plane-coordinate system
- Next, the projection on the normal vector is defined:
-
- Hence the point Q can be defined:
-
Q=P−proj*PN - Further, the point R can be defined (the reason for the factor 2 is that in mirror the arriving and departing light beam have equal angles compared to the normal vector of the surface).
-
B=Q−VP -
R=VP+2*B - And finally the direction vector C is defined as follows:
-
C=R−P. - Because the VS is located at plane x=kuva_shift, the vector C hits that plane at the point
-
S=k*C+P - where
-
- where P1 is the x-component of the point P and C1 is the x-component of the vector C. Note that in this calculation the VP was defined to the origin to simplify the presentation of the algorithm. However, in practice the VP can locate freely in the coordinate space. It must be noted that the image on the virtual screen VS is horizontally inversed when the virtual screen VS is viewed from the viewpoint VP direction.
- The system of
FIG. 4 has several characteristics: - 1. The view on the display device moves into the same direction as it is tilted. In one embodiment, the movement of the portion of the virtual data object displayed on the display device is proportional to the change amount and/or rate of the rotational movement.
- 2. When the distance between the VP and VD increases, the same tilting angle causes greater movements on the virtual screen VS. In other words, the browsing speed of the information on the display device increases as the distance between the VP and VD increases. In one embodiment, this movement factor can be adjusted by the user of the hand-held device.
- 3. When rotating the display device, the view on the display device remains unchanged in relative to the user.
- 4. The view on the display device depends on the position and orientation of the VS, VP and VD.
- 5. A certain VS-VP-VD position/orientation combination always constitute the same view on the display device.
- 6. When the position of the VD alters, the viewing angle between the VP and VD changes.
- 7. Zooming can be implemented by changing the position of the VS, VP and VD.
- 8. Zooming can be implemented by enlarging the object on the VS or altering the radius of curvature of the mirror (VD).
- 9. If the figure on the VS is in the right way when viewed from the VP, the view on the VD is mirrored (horizontally inversed).
- The present invention does not have to implement all the aforementioned features, but the most appropriate ones can be chosen. The ideal mirror-like functionality means that the information on the display device changes when:
- a) the location or orientation of the hand-held device in proportion to the coordinates bound to the physical environment changes,
- b) the location of the user (VP) in proportion to the coordinates bound to the hand-held device changes,
- c) the virtual location of the data (virtual screen) displayed on the display device in proportion to the coordinates bound to the physical environment changes.
- In order to simulate the operation of a mirror to the user, the information on the display device is changed at least either according to a) or b). If only a) or b) is taken into consideration, the operation of the display is not so mirror-like as if both a) and b) were implemented. In one embodiment, the display device operates according to all a), b) and c).
-
FIGS. 10 a-d illustrate another example of calculation which is explained with reference toFIG. 4 .FIG. 10 c is a side view ofFIG. 10 a andFIG. 10 d is a side view ofFIG. 10 b. In this example the virtual screen is referred to as thevirtual surface 200. - In
FIGS. 10 a and 10 c the default orientation of thedisplay 201 is determined to be parallel with the yz-plane. The virtual surface (VS) 200 is above the display plane and also parallel with the yz-plane. A page having information to be browsed lies on thevirtual surface 200, and the size of the page is larger than the size of thedisplay 201. The reference point VP is on thevirtual surface 200. The x-axis (not shown) runs through the reference point VP and the middle point P of thedisplay 201. After calculating the point S by the method presented with reference toFIG. 4 , the result is that point S is equal to point VP. Of course, the relationship between every single point in the area (2 a*2 b) of thedisplay 201 and the corresponding area (2 a*2 b) on thevirtual surface 200 can be calculated in a similar way. The portion of the page (2 a*2 b) that is to be displayed then has a shape similar to the shape of the display (2 a*2 b). In other words, on the point S on thevirtual surface 200 is the middle point of the determined rectangle 2 a*2 b and all the other points residing around point S within the rectangle relate to the corresponding the points residing around point P on thedisplay 201. That portion of the page surrounding point S on the virtual surface is displayed ondisplay 201. - In
FIGS. 10 b and 10 d thedisplay 201 has been tilted around the y-axis, wherein the portion of the page shown on thedisplay 201 changes in the following way: - Initially (i.e. when the
virtual surface 200 and thedisplay surface 201 are parallel with respect to each other as shown inFIGS. 10 a and 10 c) areference line 203 drawn between point P and point S meets the x-axis, i.e. it is parallel with the x-axis. The normal of the display extending from point P is parallel with the x-axis and thereference line 203. Whendisplay 201 is tilted by angle a with respect to thevirtual surface 200, the normal 204 of the display is also tilted by angle a with respect to the x-axis. After thedisplay 201 is tilted as shown inFIGS. 10 b and 10 d, thereference line 203 is mirrored with respect to the normal 204 of the display wherein amirror line 205 is generated. A hit point S′ is the point where themirror line 205 hits thevirtual surface 200. In the same manner as above, an area (shape) of the page corresponding to the area (shape) of the display is determined. Thedisplay 201 then shows the portion of the page around the hit point S′ and having a shape similar to the shape of thedisplay 201. -
FIG. 5 represents one example of a preferred hand-helddevice 40. The hand-helddevice 40 is e.g. a mobile phone. The hand-held device comprises aprocessor 30 and adisplay device 10 coupled to theprocessor 30. Thedata memory 60 and theprogram memory 70 are also coupled to theprocessor 30. Theprogram memory 70 contains e.g. the operation system. The sizes of the memories, and the processing power of theprocessor 30 depend on the device and application used. Theprogram memory 60 can additionally contain different kinds of software applications with which various tasks can be executed. Application software comprise e.g. word processing, graphical and spreadsheet software. The software applications and data used by them are loaded into thedata memory 60 in order to be able to use the software. - The
display adapter 90 with theprocessor 30 controls thedisplay device 10. In order to not to use thedata memory 60 for storing display-related information, thedisplay adapter 90 comprises a data buffer in which the information to be displayed on thedisplay device 10 is stored. - The hand-held
device 40 comprises measuring means which in a preferred embodiment of the invention refer to acceleration sensor(s) 50. With the acceleration sensor(s) 50 it is possible to measure tilting movements of the hand-helddevice 40. Theprocessor 30 receives the measurement results and interprets them. The acceleration sensor(s) 50 can be e.g. piezo-electric or capacitive producing an analog voltage which is proportional to the acceleration factor. - With the acceleration sensor(s) 50 it is possible to measure one, two or three-dimensional accelerations. The measurement of tilting movements is based on the fact that the highest acceleration is parallel to the gravity of the earth. Therefore, the orientation of the hand-held
device 40 can be defined in relation to the earth. It is also possible to use gyroscopes with its various forms to measure the orientation of the hand-helddevice 40. The quantities measured are e.g. tilting angle and accelerations. - The relation information between the rotation degree of the hand-held device and the memory address corresponding to the displayed view is stored e.g. on the
data memory 60. Theprocessor 30 defines the orientation of the hand-helddevice 40 in relation to the user or a reference position. Theprocessor 30 may also define the distance between the user and the hand-helddevice 40 or the user orientation in relation to the hand-helddevice 40. - The most important point is not the way of how the aforementioned definitions are made but the fact that the orientation of the hand-held
device 40 affects the information displayed on thedisplay device 10. The memory space can be implemented logically, e.g. as a two-dimensional memory space. When browsing starts, theprocessor 30 starts the definition process of the new memory address from the current memory address so that displacement in the memory space corresponds to the direction and amount of change in orientation according to the relation information. - The hand-held
device 40 comprises also abrowse lock 80 with which it is signaled when the browsing is executed. The orientation of the hand-helddevice 40 must remain in the same position in order to keep the view on the display device unchanged. In a preferred embodiment, the hand-helddevice 40 comprises a lock feature, e.g. a push-button, with which the browsing can be locked. The user can tilt the hand-held device back to an appropriate viewing orientation in order to view the information on thedisplay device 10 properly. The browsing may then continue when the button is released. - The hand-held
device 40 inFIG. 6 is almost the same as the hand-helddevice 40 inFIG. 5 . InFIG. 5 , the hand-held device comprises also alocator 20. It is possible to control the view on thedisplay device 10 also by other means than acceleration sensor(s) or equivalent means. The hand-helddevice 40 can comprise e.g. a (video) camera measuring the orientation and location of the hand-held device in relation to the user of the hand-helddevice 40 or to another reference point in the surroundings of the user. Thecamera 20 may be set to recognize and measure distance to a certain reference point, e.g. the eyes of the user. Therefore, when the orientation and/or position of the hand-helddevice 40 changes, the viewing angle measured by the camera also changes. Thus, it can be concluded that the hand-helddevice 40 has been tilted and/or moved towards some direction. - By analyzing the video image it is possible to define the orientation of the hand-held
device 40 in proportion to the reference point and the distance of the hand-helddevice 40 to the reference point tens of times within a second. The browsing functionality can be implemented merely using the video camera, so that additional acceleration sensor(s) are not necessarily needed. The measuring of the distance can also be implemented with an ultrasonic radar connected through an analog-digital converter to theprocessor 30 of the hand-helddevice 40. In one embodiment, from the user's perspective the information on thedisplay device 10 is essentially browsed in the same manner as when looking in a mirror. In other words, the view on thedisplay 10 depends on the viewing angle in relation to the display device plane as the view in a mirror depends on the viewing angle to the mirror. - In one embodiment of
FIG. 5 , thelocator 20 comprises a video camera seeking the location of the head and eyes of the user. Heuristic algorithms and neural network seeking the location of the head and eyes can be used. Acceleration sensors are more appropriate to use in hand-held devices than a video camera, because they are cheaper. The acceleration sensors may also be a more appropriate solution in devices which do not have a built-in video camera for a default feature, e.g. in the (third generation) mobile phones. The advantage of the use of the video camera is that the use of the hand-held device is not restricted to the position of the hand-held device, e.g. when being on one's back the hand-held device can be used without problems. Also the selection of starting point of browsing is more free, and choice (of the starting point) can be given to the user of the hand-held device. In one embodiment ofFIG. 5 , the display device surface level is set as an xy-plane. A certain relation between the x-axial and/or y-axial movement of the hand-held device and the amount of the displacement of the portion of the virtual data object displayed on the display device at a time has been determined. So, when the hand-helddevice 40 is moved along x- and/or y-axis, the portion of the virtual data object displayed on the display device moves in the same direction as the hand-held device is moved in the xy-plane according to the relation information. - In a preferred embodiment of
FIGS. 5 and 6 theprocessor 30 comprises also means for filtering the x-axial, y-axial and/or tilting movements before displaying the movements on the display device. Therefore, minor unintentional movements can be filtered out. - In one embodiment of
FIGS. 5 and 6 , the relation between the tilting movements and the amount of the displacement of the portion of the virtual data object displayed on the display device at a time can be changed. Therefore, a user may define e.g. that from now on a 10 degree tilting causes the same effect on the display as a 15 degree tilting earlier. In one embodiment, the relation is linear. In other words, the relation between the tilting movements and the amount of the displacement of the portion of the virtual data object displayed on the display device at a time does not depend on the amount of the tilting. In another embodiment, the relation is linear, but e.g. exponential. In other words, the amount of the displacement of the portion of the virtual data object displayed on the display device at a time depends on the amount of the tilting. For example, the value of the relation factor changes (e.g. exponentially) as the tilting amount increases. -
FIGS. 7 a-7 d represent the situation where the size of the information on the display device depends on the zoom factor in addition to the orientation of the hand-held device. The zoom factor can be controlled in different ways. In one embodiment, the zoom factor depends on the distance between the user and the hand-held device.FIG. 7 a represent thedisplay device 10, on whichgraphical reference numerals display device 10 depends on the orientation of the hand-held device or the viewing angle from which the user of the hand-held views the display device. When the user of the hand-held device sets a graphicalFIG. 21 in the middle of the display device, and the zoom factor is increased, a graphicalFIG. 21 grows as depicted inFIGS. 7 b and 7 c. InFIG. 7 d, the zoom factor has decreased, and also the viewing angle between the user and the hand-held device has changed. - The zoom factor can be modified with several different ways. In one embodiment, the zoom factor depends on the distance between the reference point (e.g. the eyes of the user) and the hand-held device. When the distance decreases, a graphical
FIG. 21 grows, and vice versa. Thedisplay device 10 may have to be set to a zoom mode before the zoom factor changes. If the zoom factor was all the time dependent on the distance between the reference point and the hand-held device, the browsing operation would not necessarily be practical because the view on thedisplay 10 would change whenever the aforementioned distance changes. - In another embodiment, the zoom factor changes when rotating the hand-held device around the axis being essentially perpendicular to a predefined xy-plane. The xy-plane may be the present plane of the
display device 10 or some other predetermined plane. Yet in another embodiment, the zoom factor is changed by tilting the hand-held device. Before this the display device must be set into a zoom mode. When the hand-held device is tilted, e.g. to the right the zoom factor increases, and when the hand-held device is tilted to the left, the zoom factor decreases. It is not important which predefined tilting directions are used but that the two directions can be separated sufficiently from each other. The aforementioned zoom mode is set on and off e.g. with a predetermined button of the hand-held device. -
FIGS. 8 a-8 c represent different ways to implement the user interface. InFIG. 8 a, thedisplay device 10 of the hand-helddevice 40 contains information to be viewed by the user. InFIG. 8 a, an A letter is on thedisplay device 10, In one embodiment, the information on thedisplay device 10 remains in the same position with respect to the user when the hand-helddevice 40 is rotated around the axis being perpendicular to the display surface plane, as depicted inFIG. 8 b. In other words, the information on thedisplay device 10 remains in the same position because the information is attached to the real physical coordinates. - In another embodiment, the information on the
display device 10 remains in the same position with respect to the hand-helddevice 40 when the hand-helddevice 40 is rotated around the axis being perpendicular to the display surface plane, as depicted inFIG. 8 c. In other words, the orientation of the information on thedisplay device 10 changes with respect to the user of the hand-helddevice 40 because the information is not attached to the real physical coordinates but to the display device. -
FIG. 9 represents a flow diagram describing the functionality of a method of the present invention.FIG. 9 describes a hand-helddevice 40 comprising means for measuringacceleration 50 and aprocessor 30. Means for measuring acceleration refer e.g. to a multiaxial acceleration sensor suited for measuring changes in the orientation of the hand-helddevice 40. - The hand-held device is switched on, and it is ready for browsing information on the display device, as represented in
phase 100. When the hand-held device is functional, theacceleration sensor 50 measures constantly acceleration readings. Theprocessor 30 receives the acceleration readings and defines the orientation of the hand-held device and also the change in the orientation compared to the prior measurement(s), as represented inphases phase 103, it is tested whether the browsing is on or off. If the browsing is off, theprocessor 30 examines if a predetermined browsing startup condition is fulfilled (phase 104). If it is not fulfilled, the method returns back tophase 101. It means that the orientation of the hand-held device has not changed sufficiently, which would indicate that the user wishes to browse information on the display device of the hand-held device. - If the predetermined browsing startup condition is fulfilled, the
processor 30 sets the browsing as started (phase 106) and determines the browsing speed based on the current acceleration value (phase 108). Theprocessor 30 also changes the information presented on the display device according to a relation between the rotation degree and the amount of the displacement of the portion on the virtual data object stored on thedata memory 60 and the determined browsing speed (phase 108). A certain orientation of the hand-held device always causes the same view (the same portion on the virtual data object stored on the memory) on the display device. If it is observed inphase 103 that the browsing is already on, and the browsing stopping condition is fulfilled (phase 105), theprocessor 30 stops the browsing and sets the browsing as stopped (phases 107 and 109). If it is observed that the browsing stopping condition is not fulfilled (phase 105), theprocessor 30 returns back tophase 101. - While the apparatus and method have been described in terms of what are presently considered to be the most practical and preferred embodiments, it is to be understood that the disclosure need not be limited to the disclosed embodiments. It is intended to cover various modifications and similar arrangements included within the spirit and scope of the claims, the scope of which should be accorded the broadest interpretation so as to encompass all such modifications and similar structures. The present disclosure includes any and all embodiments of the following claims.
Claims (2)
1. A computer program embodied on a computer-readable medium, wherein the computer program executes the program recorded in a computer-readable medium to perform a method for browsing information on a display of a hand-held device, the computer program:
displaying a portion of the page residing around the predefined point and having a shape similar to the shape of the display,
generating a mirror line by mirroring the reference line in relation to a line that is perpendicular to the display surface and travels via the reference point in response to tilting of the hand-held device in relation to the spatial initial state,
defining a hit point (xn,yn) where the mirror line hits the virtual surface and the page containing information; and
showing on the display at least a portion of the page around the hit point and having a shape similar to the shape of the display, the position of the hit point on the page to correspond to the position of the reference point on the display.
2. A computer program according to claim 1 , wherein the computer program sets the initial state in a certain predefined angle in relation to the earth surface.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/710,944 US20100153891A1 (en) | 2001-05-16 | 2010-02-23 | Method, device and program for browsing information on a display |
Applications Claiming Priority (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
FI20011039A FI117488B (en) | 2001-05-16 | 2001-05-16 | Browsing information on screen |
FI20011039 | 2001-05-16 | ||
US10/071,172 US20020175896A1 (en) | 2001-05-16 | 2002-02-08 | Method and device for browsing information on a display |
US11/159,786 US7607111B2 (en) | 2001-05-16 | 2005-06-23 | Method and device for browsing information on a display |
US12/710,944 US20100153891A1 (en) | 2001-05-16 | 2010-02-23 | Method, device and program for browsing information on a display |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/159,786 Continuation US7607111B2 (en) | 2001-05-16 | 2005-06-23 | Method and device for browsing information on a display |
Publications (1)
Publication Number | Publication Date |
---|---|
US20100153891A1 true US20100153891A1 (en) | 2010-06-17 |
Family
ID=8561217
Family Applications (6)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/071,172 Abandoned US20020175896A1 (en) | 2001-05-16 | 2002-02-08 | Method and device for browsing information on a display |
US11/159,786 Expired - Fee Related US7607111B2 (en) | 2001-05-16 | 2005-06-23 | Method and device for browsing information on a display |
US12/569,797 Abandoned US20100020102A1 (en) | 2001-05-16 | 2009-09-29 | Method and device for browsing information on a display |
US12/691,506 Active 2027-07-29 US9727095B2 (en) | 2001-05-16 | 2010-01-21 | Method, device and program for browsing information on a display |
US12/710,944 Abandoned US20100153891A1 (en) | 2001-05-16 | 2010-02-23 | Method, device and program for browsing information on a display |
US15/650,765 Expired - Lifetime US11301196B2 (en) | 2001-05-16 | 2017-07-14 | Method, device and program for browsing information on a display |
Family Applications Before (4)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/071,172 Abandoned US20020175896A1 (en) | 2001-05-16 | 2002-02-08 | Method and device for browsing information on a display |
US11/159,786 Expired - Fee Related US7607111B2 (en) | 2001-05-16 | 2005-06-23 | Method and device for browsing information on a display |
US12/569,797 Abandoned US20100020102A1 (en) | 2001-05-16 | 2009-09-29 | Method and device for browsing information on a display |
US12/691,506 Active 2027-07-29 US9727095B2 (en) | 2001-05-16 | 2010-01-21 | Method, device and program for browsing information on a display |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/650,765 Expired - Lifetime US11301196B2 (en) | 2001-05-16 | 2017-07-14 | Method, device and program for browsing information on a display |
Country Status (10)
Country | Link |
---|---|
US (6) | US20020175896A1 (en) |
EP (1) | EP1410142B1 (en) |
JP (1) | JP4175897B2 (en) |
KR (1) | KR100671585B1 (en) |
CN (1) | CN1714326B (en) |
AT (1) | ATE317138T1 (en) |
DE (1) | DE60208995D1 (en) |
FI (1) | FI117488B (en) |
RU (1) | RU2288512C2 (en) |
WO (1) | WO2002093331A1 (en) |
Cited By (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090297062A1 (en) * | 2005-03-04 | 2009-12-03 | Molne Anders L | Mobile device with wide-angle optics and a radiation sensor |
US20090305727A1 (en) * | 2005-03-04 | 2009-12-10 | Heikki Pylkko | Mobile device with wide range-angle optics and a radiation sensor |
US20090325607A1 (en) * | 2008-05-28 | 2009-12-31 | Conway David P | Motion-controlled views on mobile computing devices |
US20100020102A1 (en) * | 2001-05-16 | 2010-01-28 | Motionip, Llc | Method and device for browsing information on a display |
US20100045667A1 (en) * | 2008-08-22 | 2010-02-25 | Google Inc. | Navigation In a Three Dimensional Environment Using An Orientation Of A Mobile Device |
US20100171691A1 (en) * | 2007-01-26 | 2010-07-08 | Ralph Cook | Viewing images with tilt control on a hand-held device |
US20110010659A1 (en) * | 2009-07-13 | 2011-01-13 | Samsung Electronics Co., Ltd. | Scrolling method of mobile terminal and apparatus for performing the same |
US20140009498A1 (en) * | 2012-07-09 | 2014-01-09 | Research In Motion Limited | System and method for determining a display orientation of a mobile device |
US9032818B2 (en) | 2012-07-05 | 2015-05-19 | Nextinput, Inc. | Microelectromechanical load sensor and methods of manufacturing the same |
US9487388B2 (en) | 2012-06-21 | 2016-11-08 | Nextinput, Inc. | Ruggedized MEMS force die |
US9902611B2 (en) | 2014-01-13 | 2018-02-27 | Nextinput, Inc. | Miniaturized and ruggedized wafer level MEMs force sensors |
EP3316117A1 (en) * | 2016-10-31 | 2018-05-02 | Nokia Technologies OY | Controlling content displayed in a display |
US10466119B2 (en) | 2015-06-10 | 2019-11-05 | Nextinput, Inc. | Ruggedized wafer level MEMS force sensor with a tolerance trench |
US10671940B2 (en) | 2016-10-31 | 2020-06-02 | Nokia Technologies Oy | Controlling display of data to a person via a display apparatus |
US10962427B2 (en) | 2019-01-10 | 2021-03-30 | Nextinput, Inc. | Slotted MEMS force sensor |
US11221263B2 (en) | 2017-07-19 | 2022-01-11 | Nextinput, Inc. | Microelectromechanical force sensor having a strain transfer layer arranged on the sensor die |
US11243125B2 (en) | 2017-02-09 | 2022-02-08 | Nextinput, Inc. | Integrated piezoresistive and piezoelectric fusion force sensor |
US11243126B2 (en) | 2017-07-27 | 2022-02-08 | Nextinput, Inc. | Wafer bonded piezoresistive and piezoelectric force sensor and related methods of manufacture |
US11255737B2 (en) | 2017-02-09 | 2022-02-22 | Nextinput, Inc. | Integrated digital force sensors and related methods of manufacture |
US11385108B2 (en) | 2017-11-02 | 2022-07-12 | Nextinput, Inc. | Sealed force sensor with etch stop layer |
US11423686B2 (en) | 2017-07-25 | 2022-08-23 | Qorvo Us, Inc. | Integrated fingerprint and force sensor |
US11579028B2 (en) | 2017-10-17 | 2023-02-14 | Nextinput, Inc. | Temperature coefficient of offset compensation for force sensor and strain gauge |
US11874185B2 (en) | 2017-11-16 | 2024-01-16 | Nextinput, Inc. | Force attenuator for force sensor |
Families Citing this family (146)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
FI20012209A (en) * | 2001-11-14 | 2003-06-24 | Nokia Corp | Method for controlling display of information in an electronic device and electronic device |
US7714880B2 (en) * | 2001-11-16 | 2010-05-11 | Honeywell International Inc. | Method and apparatus for displaying images on a display |
US7002553B2 (en) * | 2001-12-27 | 2006-02-21 | Mark Shkolnikov | Active keyboard system for handheld electronic devices |
WO2003071769A1 (en) * | 2002-02-20 | 2003-08-28 | Siemens Aktiengesellschaft | Mobile telephone comprising a wraparound display |
GB2387504B (en) * | 2002-04-12 | 2005-03-16 | Motorola Inc | Method and system of managing a user interface of a communication device |
JP4117352B2 (en) * | 2002-11-12 | 2008-07-16 | 株式会社ソニー・コンピュータエンタテインメント | File processing method and apparatus capable of using this method |
US7385624B2 (en) * | 2003-01-17 | 2008-06-10 | Nippon Telegraph And Telephone Corporation | Remote image display method, image capturing device, and method and program therefor |
CN100452868C (en) * | 2003-01-17 | 2009-01-14 | 日本电信电话株式会社 | Remote video display method, video acquisition device, method thereof, and program thereof |
GB0311177D0 (en) * | 2003-05-15 | 2003-06-18 | Qinetiq Ltd | Non contact human-computer interface |
CA2531524A1 (en) * | 2003-10-31 | 2005-05-12 | Iota Wireless Llc | Concurrent data entry for a portable device |
US20080129552A1 (en) * | 2003-10-31 | 2008-06-05 | Iota Wireless Llc | Concurrent data entry for a portable device |
JP2005277452A (en) * | 2004-03-22 | 2005-10-06 | Nec Corp | Portable electronic apparatus and its display switching method |
US20050212760A1 (en) * | 2004-03-23 | 2005-09-29 | Marvit David L | Gesture based user interface supporting preexisting symbols |
JP2008502043A (en) * | 2004-06-04 | 2008-01-24 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | Portable device for user content navigation |
KR101015352B1 (en) * | 2005-05-27 | 2011-02-16 | 샤프 가부시키가이샤 | Display device |
US7649522B2 (en) * | 2005-10-11 | 2010-01-19 | Fish & Richardson P.C. | Human interface input acceleration system |
US7652660B2 (en) | 2005-10-11 | 2010-01-26 | Fish & Richardson P.C. | Mobile device customizer |
TW200729926A (en) * | 2006-01-17 | 2007-08-01 | Inventec Appliances Corp | Method for zooming image ratio for mobile electronic device and mobile electronic device thereof |
US8730156B2 (en) * | 2010-03-05 | 2014-05-20 | Sony Computer Entertainment America Llc | Maintaining multiple views on a shared stable virtual space |
US9250703B2 (en) | 2006-03-06 | 2016-02-02 | Sony Computer Entertainment Inc. | Interface with gaze detection and voice input |
KR100772909B1 (en) * | 2006-05-30 | 2007-11-05 | 삼성전자주식회사 | Image browsing method and apparatus using the same |
US8139026B2 (en) | 2006-08-02 | 2012-03-20 | Research In Motion Limited | System and method for adjusting presentation of text and images on an electronic device according to an orientation of the device |
US8493323B2 (en) | 2006-08-02 | 2013-07-23 | Research In Motion Limited | System and method for adjusting presentation of moving images on an electronic device according to an orientation of the device |
EP2256613B1 (en) * | 2006-08-02 | 2012-10-24 | Research In Motion Limited | System and method for adjusting presentation of text and images on an electronic device according to an orientation of the device |
JP5173174B2 (en) * | 2006-09-13 | 2013-03-27 | 任天堂株式会社 | GAME DEVICE, GAME PROGRAM, GAME SYSTEM, AND GAME PROCESSING METHOD |
JP4775212B2 (en) * | 2006-09-26 | 2011-09-21 | 日本電気株式会社 | Portable terminal device, display control method thereof, program, and computer-readable recording medium |
EP1914622A3 (en) * | 2006-10-16 | 2012-11-28 | Samsung Electronics Co., Ltd. | Method and apparatus for moving list on picture plane |
KR101270700B1 (en) * | 2006-11-15 | 2013-06-03 | 삼성전자주식회사 | Method for wide viewing angle and apparatus for the same |
TWI330802B (en) * | 2006-12-13 | 2010-09-21 | Ind Tech Res Inst | Inertial sensing method and system |
US20090198132A1 (en) * | 2007-08-10 | 2009-08-06 | Laurent Pelissier | Hand-held ultrasound imaging device having reconfigurable user interface |
KR101365595B1 (en) * | 2007-08-16 | 2014-02-21 | 삼성전자주식회사 | Method for inputting of device containing display unit based on GUI and apparatus thereof |
JP2009059223A (en) * | 2007-08-31 | 2009-03-19 | Toshiba Corp | Information processing apparatus, information processing method and program |
US20090079701A1 (en) * | 2007-09-25 | 2009-03-26 | Grosskopf Jr George | Device and Method for Displaying Data and Receiving User Input |
RU2447481C2 (en) * | 2007-10-04 | 2012-04-10 | Валерий Константинович Любезнов | Method for determining position of touching screen of sensor system (versions) and optoelectronic sensor system for realising said method |
US8631358B2 (en) * | 2007-10-10 | 2014-01-14 | Apple Inc. | Variable device graphical user interface |
US8418083B1 (en) | 2007-11-26 | 2013-04-09 | Sprint Communications Company L.P. | Applying a navigational mode to a device |
WO2009082377A1 (en) * | 2007-12-26 | 2009-07-02 | Hewlett-Packard Development Company, L.P. | Touch wheel zoom and pan |
US8384718B2 (en) * | 2008-01-10 | 2013-02-26 | Sony Corporation | System and method for navigating a 3D graphical user interface |
TWI378329B (en) * | 2008-02-12 | 2012-12-01 | Shenzhen China Star Optoelect | Method for controlling angle of display and self-adjusting display device |
US8217964B2 (en) * | 2008-02-14 | 2012-07-10 | Nokia Corporation | Information presentation based on display screen orientation |
US9274681B2 (en) | 2008-03-26 | 2016-03-01 | Lg Electronics Inc. | Terminal and method of controlling the same |
US8375336B2 (en) * | 2008-05-23 | 2013-02-12 | Microsoft Corporation | Panning content utilizing a drag operation |
US20090315810A1 (en) * | 2008-06-19 | 2009-12-24 | Sony Ericsson Mobile Communications Ab | Information presentation device |
WO2009152879A1 (en) * | 2008-06-19 | 2009-12-23 | Sony Ericsson Mobile Communications Ab | Information presentation device |
US10031549B2 (en) * | 2008-07-10 | 2018-07-24 | Apple Inc. | Transitioning between modes of input |
US9285970B2 (en) * | 2008-07-25 | 2016-03-15 | Google Technology Holdings LLC | Method and apparatus for displaying navigational views on a portable device |
JP5047090B2 (en) * | 2008-07-31 | 2012-10-10 | キヤノン株式会社 | system |
KR101505198B1 (en) * | 2008-08-18 | 2015-03-23 | 엘지전자 주식회사 | PORTABLE TERMINAL and DRIVING METHOD OF THE SAME |
US8520979B2 (en) * | 2008-08-19 | 2013-08-27 | Digimarc Corporation | Methods and systems for content processing |
JP4720879B2 (en) * | 2008-08-29 | 2011-07-13 | ソニー株式会社 | Information processing apparatus and information processing method |
KR101481556B1 (en) | 2008-09-10 | 2015-01-13 | 엘지전자 주식회사 | A mobile telecommunication terminal and a method of displying an object using the same |
KR101602363B1 (en) | 2008-09-11 | 2016-03-10 | 엘지전자 주식회사 | 3 Controling Method of 3 Dimension User Interface Switchover and Mobile Terminal using the same |
JP5080409B2 (en) * | 2008-09-11 | 2012-11-21 | Kddi株式会社 | Information terminal equipment |
CN102150160B (en) * | 2008-09-12 | 2015-05-20 | 皇家飞利浦电子股份有限公司 | Navigating in graphical user interface on handheld devices |
JP4962741B2 (en) * | 2008-09-29 | 2012-06-27 | 株式会社エクォス・リサーチ | Terminal device |
US8717283B1 (en) * | 2008-11-25 | 2014-05-06 | Sprint Communications Company L.P. | Utilizing motion of a device to manipulate a display screen feature |
US8441441B2 (en) | 2009-01-06 | 2013-05-14 | Qualcomm Incorporated | User interface for mobile devices |
JP5357800B2 (en) * | 2009-02-12 | 2013-12-04 | キヤノン株式会社 | Electronic device and control method thereof |
JP5158006B2 (en) * | 2009-04-23 | 2013-03-06 | ソニー株式会社 | Information processing apparatus, information processing method, and program |
US20100275122A1 (en) * | 2009-04-27 | 2010-10-28 | Microsoft Corporation | Click-through controller for mobile interaction |
JP5263049B2 (en) * | 2009-07-21 | 2013-08-14 | ソニー株式会社 | Information processing apparatus, information processing method, and program |
US8531571B1 (en) * | 2009-08-05 | 2013-09-10 | Bentley Systmes, Incorporated | System and method for browsing a large document on a portable electronic device |
US8456430B2 (en) * | 2009-08-21 | 2013-06-04 | Motorola Mobility Llc | Tactile user interface for an electronic device |
US9092115B2 (en) * | 2009-09-23 | 2015-07-28 | Microsoft Technology Licensing, Llc | Computing system with visual clipboard |
CN102043569A (en) * | 2009-10-26 | 2011-05-04 | 环达电脑(上海)有限公司 | Browsing device and method of large-size image |
KR101651430B1 (en) * | 2009-12-18 | 2016-08-26 | 삼성전자주식회사 | Apparatus and method for controlling size of display data in portable terminal |
US8814686B2 (en) | 2010-02-03 | 2014-08-26 | Nintendo Co., Ltd. | Display device, game system, and game method |
WO2011096204A1 (en) | 2010-02-03 | 2011-08-11 | 任天堂株式会社 | Display device, game system, and game processing method |
US8339364B2 (en) | 2010-02-03 | 2012-12-25 | Nintendo Co., Ltd. | Spatially-correlated multi-display human-machine interface |
US8913009B2 (en) | 2010-02-03 | 2014-12-16 | Nintendo Co., Ltd. | Spatially-correlated multi-display human-machine interface |
JP5413250B2 (en) * | 2010-03-05 | 2014-02-12 | ソニー株式会社 | Image processing apparatus, image processing method, and program |
US8977987B1 (en) * | 2010-06-14 | 2015-03-10 | Google Inc. | Motion-based interface control on computing device |
CN101887348A (en) * | 2010-06-29 | 2010-11-17 | 深圳桑菲消费通信有限公司 | Method for inputting characters on mobile terminal |
JP6220268B2 (en) * | 2010-06-30 | 2017-10-25 | コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. | How to zoom the displayed image |
JP6243586B2 (en) | 2010-08-06 | 2017-12-06 | 任天堂株式会社 | GAME SYSTEM, GAME DEVICE, GAME PROGRAM, AND GAME PROCESSING METHOD |
US10150033B2 (en) * | 2010-08-20 | 2018-12-11 | Nintendo Co., Ltd. | Position calculation system, position calculation device, storage medium storing position calculation program, and position calculation method |
JP5840386B2 (en) | 2010-08-30 | 2016-01-06 | 任天堂株式会社 | GAME SYSTEM, GAME DEVICE, GAME PROGRAM, AND GAME PROCESSING METHOD |
JP5840385B2 (en) | 2010-08-30 | 2016-01-06 | 任天堂株式会社 | GAME SYSTEM, GAME DEVICE, GAME PROGRAM, AND GAME PROCESSING METHOD |
US9164542B2 (en) * | 2010-08-31 | 2015-10-20 | Symbol Technologies, Llc | Automated controls for sensor enabled user interface |
KR101492310B1 (en) | 2010-11-01 | 2015-02-11 | 닌텐도가부시키가이샤 | Operating apparatus and information processing apparatus |
US10146426B2 (en) * | 2010-11-09 | 2018-12-04 | Nokia Technologies Oy | Apparatus and method for user input for controlling displayed information |
KR101932688B1 (en) | 2010-11-29 | 2018-12-28 | 삼성전자주식회사 | Portable Device and Method for Providing User Interface Mode thereof |
JP5645626B2 (en) * | 2010-12-06 | 2014-12-24 | キヤノン株式会社 | Display control apparatus, display control method, program, and storage medium |
US20120200573A1 (en) * | 2011-02-07 | 2012-08-09 | Hooray LLC | E-reader with locked and unlocked content and reader tracking capability |
US9035940B2 (en) * | 2011-03-08 | 2015-05-19 | Nokia Corporation | Apparatus and associated methods |
CN102141846A (en) * | 2011-03-09 | 2011-08-03 | 中兴通讯股份有限公司 | Method for intuitively browsing contents and mobile terminal |
JP5689014B2 (en) | 2011-04-07 | 2015-03-25 | 任天堂株式会社 | Input system, information processing apparatus, information processing program, and three-dimensional position calculation method |
WO2012141352A1 (en) * | 2011-04-13 | 2012-10-18 | Lg Electronics Inc. | Gesture recognition agnostic to device orientation |
US10120438B2 (en) | 2011-05-25 | 2018-11-06 | Sony Interactive Entertainment Inc. | Eye gaze to alter device behavior |
US8381102B1 (en) * | 2011-12-06 | 2013-02-19 | Google Inc. | Systems and methods for visually scrolling through a stack of items displayed on a device |
US9223138B2 (en) | 2011-12-23 | 2015-12-29 | Microsoft Technology Licensing, Llc | Pixel opacity for augmented reality |
US9606586B2 (en) | 2012-01-23 | 2017-03-28 | Microsoft Technology Licensing, Llc | Heat transfer device |
JP6019601B2 (en) * | 2012-02-10 | 2016-11-02 | ソニー株式会社 | Information processing apparatus, information processing method, and program |
US9779643B2 (en) | 2012-02-15 | 2017-10-03 | Microsoft Technology Licensing, Llc | Imaging structure emitter configurations |
US9368546B2 (en) | 2012-02-15 | 2016-06-14 | Microsoft Technology Licensing, Llc | Imaging structure with embedded light sources |
US9297996B2 (en) | 2012-02-15 | 2016-03-29 | Microsoft Technology Licensing, Llc | Laser illumination scanning |
US9726887B2 (en) | 2012-02-15 | 2017-08-08 | Microsoft Technology Licensing, Llc | Imaging structure color conversion |
KR101894567B1 (en) | 2012-02-24 | 2018-09-03 | 삼성전자 주식회사 | Operation Method of Lock Screen And Electronic Device supporting the same |
US9075451B2 (en) * | 2012-02-24 | 2015-07-07 | Blackberry Limited | Handheld device with notification message viewing |
US9578318B2 (en) | 2012-03-14 | 2017-02-21 | Microsoft Technology Licensing, Llc | Imaging structure emitter calibration |
US11068049B2 (en) * | 2012-03-23 | 2021-07-20 | Microsoft Technology Licensing, Llc | Light guide display and field of view |
US9558590B2 (en) * | 2012-03-28 | 2017-01-31 | Microsoft Technology Licensing, Llc | Augmented reality light guide display |
US10191515B2 (en) | 2012-03-28 | 2019-01-29 | Microsoft Technology Licensing, Llc | Mobile device light guide display |
US9717981B2 (en) | 2012-04-05 | 2017-08-01 | Microsoft Technology Licensing, Llc | Augmented reality and physical games |
KR101890140B1 (en) | 2012-05-07 | 2018-09-28 | 삼성전자주식회사 | A method for controlling a display apparatus using a camera device and mobile device, display apparatus, and system thereof |
US10502876B2 (en) | 2012-05-22 | 2019-12-10 | Microsoft Technology Licensing, Llc | Waveguide optics focus elements |
US8989535B2 (en) | 2012-06-04 | 2015-03-24 | Microsoft Technology Licensing, Llc | Multiple waveguide imaging structure |
US20140002443A1 (en) * | 2012-06-29 | 2014-01-02 | Blackboard Inc. | Augmented reality interface |
JP6100497B2 (en) * | 2012-10-09 | 2017-03-22 | 任天堂株式会社 | Information processing program, information processing apparatus, information processing system, and image display method |
CN103019549A (en) * | 2012-11-28 | 2013-04-03 | 深圳市中兴移动通信有限公司 | Method and device for scrolling contents based on mobile terminal |
US10192358B2 (en) | 2012-12-20 | 2019-01-29 | Microsoft Technology Licensing, Llc | Auto-stereoscopic augmented reality display |
CN103902194A (en) * | 2012-12-28 | 2014-07-02 | 鸿富锦精密工业(武汉)有限公司 | Electronic device and control method for allowing screen image of electronic device to rotate |
US8976202B2 (en) | 2013-01-28 | 2015-03-10 | Dave CAISSY | Method for controlling the display of a portable computing device |
EP2965503A1 (en) * | 2013-03-08 | 2016-01-13 | Thomson Licensing | Method and system for stabilization and reframing |
US9658688B2 (en) * | 2013-10-15 | 2017-05-23 | Microsoft Technology Licensing, Llc | Automatic view adjustment |
JP2014197400A (en) * | 2014-05-16 | 2014-10-16 | 日本電気株式会社 | Liquid crystal display apparatus, mobile communication terminal device and liquid crystal display method |
US10782787B2 (en) * | 2014-06-06 | 2020-09-22 | Adobe Inc. | Mirroring touch gestures |
US10019140B1 (en) * | 2014-06-26 | 2018-07-10 | Amazon Technologies, Inc. | One-handed zoom |
US9304235B2 (en) | 2014-07-30 | 2016-04-05 | Microsoft Technology Licensing, Llc | Microfabrication |
US10678412B2 (en) | 2014-07-31 | 2020-06-09 | Microsoft Technology Licensing, Llc | Dynamic joint dividers for application windows |
US10254942B2 (en) | 2014-07-31 | 2019-04-09 | Microsoft Technology Licensing, Llc | Adaptive sizing and positioning of application windows |
US10592080B2 (en) | 2014-07-31 | 2020-03-17 | Microsoft Technology Licensing, Llc | Assisted presentation of application windows |
CN104216634A (en) * | 2014-08-27 | 2014-12-17 | 小米科技有限责任公司 | Method and device for displaying manuscript |
JP6465625B2 (en) * | 2014-10-10 | 2019-02-06 | キヤノン株式会社 | Control device, control method therefor, and program |
CN105786360A (en) * | 2014-12-23 | 2016-07-20 | 珠海金山办公软件有限公司 | Method and device for demonstrating PPT file with mobile terminal |
CN104548596B (en) * | 2015-02-02 | 2017-05-24 | 广州周游网络科技有限公司 | Aiming method and device of shooting games |
US9513480B2 (en) | 2015-02-09 | 2016-12-06 | Microsoft Technology Licensing, Llc | Waveguide |
US9429692B1 (en) | 2015-02-09 | 2016-08-30 | Microsoft Technology Licensing, Llc | Optical components |
US10317677B2 (en) | 2015-02-09 | 2019-06-11 | Microsoft Technology Licensing, Llc | Display system |
US11086216B2 (en) | 2015-02-09 | 2021-08-10 | Microsoft Technology Licensing, Llc | Generating electronic components |
US9827209B2 (en) | 2015-02-09 | 2017-11-28 | Microsoft Technology Licensing, Llc | Display system |
US9535253B2 (en) | 2015-02-09 | 2017-01-03 | Microsoft Technology Licensing, Llc | Display system |
US9372347B1 (en) | 2015-02-09 | 2016-06-21 | Microsoft Technology Licensing, Llc | Display system |
US9423360B1 (en) | 2015-02-09 | 2016-08-23 | Microsoft Technology Licensing, Llc | Optical components |
US10018844B2 (en) | 2015-02-09 | 2018-07-10 | Microsoft Technology Licensing, Llc | Wearable image display system |
CN105353908B (en) * | 2015-10-31 | 2018-03-27 | 广东欧珀移动通信有限公司 | A kind of menu control method and user terminal |
CN106843709B (en) | 2015-12-04 | 2020-04-14 | 阿里巴巴集团控股有限公司 | Method and device for displaying display object according to real-time information |
RU2666626C1 (en) * | 2016-07-28 | 2018-09-11 | Бейджин Сяоми Мобайл Софтвэар Ко., Лтд. | Playback state controlling method and device |
CN106708183A (en) * | 2016-11-23 | 2017-05-24 | 浪潮(苏州)金融技术服务有限公司 | Terminal and use method thereof |
CN108268056B (en) * | 2016-12-30 | 2020-12-15 | 昊翔电能运动科技(昆山)有限公司 | Handheld holder calibration method, device and system |
CN111052082A (en) * | 2017-02-22 | 2020-04-21 | 思开睿公司 | Computer network modeling |
US11439839B2 (en) * | 2017-08-09 | 2022-09-13 | Acuity Innovation And Design, Llc | Hand-held treatment device using LED light sources with interchangeable emitters |
CN108664145A (en) * | 2018-05-14 | 2018-10-16 | 珠海格力电器股份有限公司 | Interface display control method and device, storage medium and mobile terminal |
US11036048B2 (en) * | 2018-10-03 | 2021-06-15 | Project Whitecard Digital Inc. | Virtual reality system and method for displaying on a real-world display a viewable portion of a source file projected on an inverse spherical virtual screen |
WO2020096096A1 (en) * | 2018-11-09 | 2020-05-14 | Samsung Electronics Co., Ltd. | Display method and display device in portable terminal |
KR20200091522A (en) | 2019-01-22 | 2020-07-31 | 삼성전자주식회사 | Method for controlling display orientation of content and electronic device thereof |
Citations (41)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6115025A (en) * | 1997-09-30 | 2000-09-05 | Silicon Graphics, Inc. | System for maintaining orientation of a user interface as a display changes orientation |
US6151208A (en) * | 1998-06-24 | 2000-11-21 | Digital Equipment Corporation | Wearable computing device mounted on superior dorsal aspect of a hand |
US6201554B1 (en) * | 1999-01-12 | 2001-03-13 | Ericsson Inc. | Device control apparatus for hand-held data processing device |
US6288704B1 (en) * | 1999-06-08 | 2001-09-11 | Vega, Vista, Inc. | Motion detection and tracking system to control navigation and display of object viewers |
US6369794B1 (en) * | 1998-09-09 | 2002-04-09 | Matsushita Electric Industrial Co., Ltd. | Operation indication outputting device for giving operation indication according to type of user's action |
US6375572B1 (en) * | 1999-10-04 | 2002-04-23 | Nintendo Co., Ltd. | Portable game apparatus with acceleration sensor and information storage medium storing a game progam |
US20020052209A1 (en) * | 2000-10-27 | 2002-05-02 | Stig Frohlund | Portable radio communications device |
US6400376B1 (en) * | 1998-12-21 | 2002-06-04 | Ericsson Inc. | Display control for hand-held data processing device |
US20020143489A1 (en) * | 2001-03-29 | 2002-10-03 | Orchard John T. | Method and apparatus for controlling a computing system |
US6466198B1 (en) * | 1999-11-05 | 2002-10-15 | Innoventions, Inc. | View navigation and magnification of a hand-held device with a display |
US20020167699A1 (en) * | 2000-05-17 | 2002-11-14 | Christopher Verplaetse | Motion-based input system for handheld devices |
US6489945B1 (en) * | 1998-02-11 | 2002-12-03 | Agilent Technologies, Inc. | Method and system for tracking attitude |
US6552713B1 (en) * | 1999-12-16 | 2003-04-22 | Hewlett-Packard Company | Optical pointing device |
US6567101B1 (en) * | 1999-10-13 | 2003-05-20 | Gateway, Inc. | System and method utilizing motion input for manipulating a display of data |
US20030126100A1 (en) * | 2001-12-26 | 2003-07-03 | Autodesk, Inc. | Fuzzy logic reasoning for inferring user location preferences |
US6624824B1 (en) * | 1996-04-30 | 2003-09-23 | Sun Microsystems, Inc. | Tilt-scrolling on the sunpad |
US6675553B2 (en) * | 2001-02-09 | 2004-01-13 | Teepack Spezialmaschinen Gmbh & Co. Kg | Method and device for stacking and packing infusion bags |
US20040012566A1 (en) * | 2001-03-29 | 2004-01-22 | Bradski Gary R. | Intuitive mobile device interface to virtual spaces |
US6690358B2 (en) * | 2000-11-30 | 2004-02-10 | Alan Edward Kaplan | Display control for hand-held devices |
US6765443B2 (en) * | 2002-02-21 | 2004-07-20 | Ericsson Inc. | Dynamic bias controller for power amplifier circuits |
US6797937B2 (en) * | 2001-07-24 | 2004-09-28 | Agilent Technologies, Inc. | System and method for reducing power consumption in an optical screen pointing device |
US20040204067A1 (en) * | 2002-03-28 | 2004-10-14 | Nec Corporation | Portable apparatus including improved pointing device |
US6872931B2 (en) * | 2000-11-06 | 2005-03-29 | Koninklijke Philips Electronics N.V. | Optical input device for measuring finger movement |
US6939231B2 (en) * | 2000-12-22 | 2005-09-06 | Nokia Corporation | Method for controlling a terminal display and a terminal |
US20050208978A1 (en) * | 2004-03-16 | 2005-09-22 | Myorigo, L.L.C. | Mobile device with wide-angle optics and a radiation sensor |
US6977675B2 (en) * | 2002-12-30 | 2005-12-20 | Motorola, Inc. | Method and apparatus for virtually expanding a display |
US7058432B2 (en) * | 2001-04-20 | 2006-06-06 | Mitsubishi Denki Kabushiki Kaisha | Pointing device and mobile telephone |
US20060146009A1 (en) * | 2003-01-22 | 2006-07-06 | Hanno Syrbe | Image control |
US7138979B2 (en) * | 2004-08-27 | 2006-11-21 | Motorola, Inc. | Device orientation based input signal generation |
US7164411B2 (en) * | 2002-12-30 | 2007-01-16 | Nokia Corporatoin | Optical user interface for controlling portable electric device |
US7184025B2 (en) * | 2002-05-31 | 2007-02-27 | Microsoft Corporation | Altering a display on a viewing device based upon a user controlled orientation of the viewing device |
US7194816B2 (en) * | 2004-07-15 | 2007-03-27 | C&N Inc. | Mobile terminal apparatus |
US7242391B2 (en) * | 2003-12-29 | 2007-07-10 | Pixart Imaging Inc. | Optical navigation chip |
US7301528B2 (en) * | 2004-03-23 | 2007-11-27 | Fujitsu Limited | Distinguishing tilt and translation motion components in handheld devices |
US20090297062A1 (en) * | 2005-03-04 | 2009-12-03 | Molne Anders L | Mobile device with wide-angle optics and a radiation sensor |
US20090303204A1 (en) * | 2007-01-05 | 2009-12-10 | Invensense Inc. | Controlling and accessing content using motion processing on mobile devices |
US20090305727A1 (en) * | 2005-03-04 | 2009-12-10 | Heikki Pylkko | Mobile device with wide range-angle optics and a radiation sensor |
US7688306B2 (en) * | 2000-10-02 | 2010-03-30 | Apple Inc. | Methods and apparatuses for operating a portable device based on an accelerometer |
US7721968B2 (en) * | 2003-10-31 | 2010-05-25 | Iota Wireless, Llc | Concurrent data entry for a portable device |
US7859553B2 (en) * | 2004-12-30 | 2010-12-28 | Lg Electronics Inc. | Image navigation in a mobile station |
US8099124B2 (en) * | 2007-04-12 | 2012-01-17 | Symbol Technologies, Inc. | Method and system for correlating user/device activity with spatial orientation sensors |
Family Cites Families (71)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4720703A (en) | 1984-08-02 | 1988-01-19 | Tektronix, Inc. | Display method and apparatus employing cursor panning |
JPS6232527A (en) | 1985-08-06 | 1987-02-12 | Hitachi Ltd | Display picture control system |
US4787051A (en) | 1986-05-16 | 1988-11-22 | Tektronix, Inc. | Inertial mouse system |
US5100823A (en) * | 1988-02-29 | 1992-03-31 | Motorola, Inc. | Method of making buried stacked transistor-capacitor |
IL92220A (en) | 1989-11-06 | 1993-02-21 | Ibm Israel | Three-dimensional computer input device |
US5367614A (en) | 1992-04-01 | 1994-11-22 | Grumman Aerospace Corporation | Three-dimensional computer image variable perspective display system |
US5463725A (en) | 1992-12-31 | 1995-10-31 | International Business Machines Corp. | Data processing system graphical user interface which emulates printed material |
US5526481A (en) | 1993-07-26 | 1996-06-11 | Dell Usa L.P. | Display scrolling system for personal digital assistant |
US6937140B1 (en) | 1993-07-30 | 2005-08-30 | Ge Interlogix, Inc. | Personal digital assistant key for an electronic lock |
JPH0764754A (en) * | 1993-08-24 | 1995-03-10 | Hitachi Ltd | Compact information processor |
SE504846C2 (en) | 1994-09-28 | 1997-05-12 | Jan G Faeger | Control equipment with a movable control means |
KR0143459B1 (en) * | 1995-05-22 | 1998-07-01 | 한민구 | Morse-gate type power transistor |
US5602556A (en) * | 1995-06-07 | 1997-02-11 | Check Point Systems, Inc. | Transmit and receive loop antenna |
US5648670A (en) * | 1995-06-07 | 1997-07-15 | Sgs-Thomson Microelectronics, Inc. | Trench MOS-gated device with a minimum number of masks |
JPH1011249A (en) | 1996-06-20 | 1998-01-16 | Matsushita Electric Ind Co Ltd | Image forming display |
GB9620464D0 (en) * | 1996-10-01 | 1996-11-20 | Philips Electronics Nv | Hand held image display device |
WO1998018495A2 (en) | 1996-10-28 | 1998-05-07 | Marsden, John, Christopher | Improvements in or relating to diagnostic/therapeutic agents |
AU3120197A (en) | 1997-05-09 | 1998-04-02 | Remec Inc. | Computer control device |
SE516552C2 (en) | 1997-10-02 | 2002-01-29 | Ericsson Telefon Ab L M | Handheld display unit and method for displaying screens |
GB9722766D0 (en) | 1997-10-28 | 1997-12-24 | British Telecomm | Portable computers |
US6316807B1 (en) * | 1997-12-05 | 2001-11-13 | Naoto Fujishima | Low on-resistance trench lateral MISFET with better switching characteristics and method for manufacturing same |
SE511516C2 (en) | 1997-12-23 | 1999-10-11 | Ericsson Telefon Ab L M | Handheld display unit and method for displaying screens |
JP3338777B2 (en) | 1998-04-22 | 2002-10-28 | 日本電気株式会社 | Mobile terminal and screen display method thereof |
JP2000259856A (en) | 1999-03-09 | 2000-09-22 | Nippon Telegr & Teleph Corp <Ntt> | Method and device for displaying three-dimensional computer graphics |
US6433385B1 (en) * | 1999-05-19 | 2002-08-13 | Fairchild Semiconductor Corporation | MOS-gated power device having segmented trench and extended doping zone and process for forming same |
EP1063607A1 (en) | 1999-06-25 | 2000-12-27 | Siemens Aktiengesellschaft | Apparatus and method for inputting control information in a computer system |
US7194702B1 (en) | 1999-06-29 | 2007-03-20 | Gateway Inc. | System method apparatus and software for minimizing unintended cursor movement |
US6252277B1 (en) * | 1999-09-09 | 2001-06-26 | Chartered Semiconductor Manufacturing Ltd. | Embedded polysilicon gate MOSFET |
JP3847058B2 (en) * | 1999-10-04 | 2006-11-15 | 任天堂株式会社 | GAME SYSTEM AND GAME INFORMATION STORAGE MEDIUM USED FOR THE SAME |
CN1119051C (en) | 1999-11-03 | 2003-08-20 | 摩托罗拉公司 | Device and method for selecting user interface option on portable electronic equipment |
KR100399583B1 (en) * | 1999-11-29 | 2003-09-26 | 한국전자통신연구원 | Fabrication method for Trench gate DMOSFET using self-align technique |
US6864912B1 (en) | 1999-12-16 | 2005-03-08 | International Business Machines Corp. | Computer system providing hands free user input via optical means for navigation or zooming |
US7124374B1 (en) | 2000-03-06 | 2006-10-17 | Carl Herman Haken | Graphical interface control system |
AU2001256576A1 (en) | 2000-05-12 | 2001-11-20 | Zvi Lapidot | Apparatus and method for the kinematic control of hand-held devices |
JP4696335B2 (en) * | 2000-05-30 | 2011-06-08 | 株式会社デンソー | Semiconductor device and manufacturing method thereof |
US7289102B2 (en) * | 2000-07-17 | 2007-10-30 | Microsoft Corporation | Method and apparatus using multiple sensors in a device with a display |
US7302280B2 (en) * | 2000-07-17 | 2007-11-27 | Microsoft Corporation | Mobile phone operation based upon context sensing |
US7034803B1 (en) | 2000-08-18 | 2006-04-25 | Leonard Reiffel | Cursor display privacy product |
FR2814254B1 (en) | 2000-09-15 | 2006-07-21 | Thomson Csf | METHOD OF PROCESSING THE DISPLACEMENT OF A CURSOR ON A SCREEN |
US7159172B1 (en) | 2000-11-08 | 2007-01-02 | Xerox Corporation | Display for rapid text reading |
US6577296B2 (en) * | 2000-11-14 | 2003-06-10 | Vega Vista, Inc. | Fixed cursor |
US6731330B2 (en) | 2001-01-30 | 2004-05-04 | Hewlett-Packard Development Company, L.P. | Method for robust determination of visible points of a controllable display within a camera view |
US6977645B2 (en) | 2001-03-16 | 2005-12-20 | Agilent Technologies, Inc. | Portable electronic device with mouse-like capabilities |
US6603485B2 (en) | 2001-04-24 | 2003-08-05 | Hewlett-Packard Development Company, L.P. | Computer cursor spotlight |
US20020180733A1 (en) * | 2001-05-15 | 2002-12-05 | Koninklijke Philips Electronics N.V. | Method and apparatus for adjusting an image to compensate for an offset position of a user |
FI117488B (en) | 2001-05-16 | 2006-10-31 | Myorigo Sarl | Browsing information on screen |
US20030001863A1 (en) * | 2001-06-29 | 2003-01-02 | Brian Davidson | Portable digital devices |
US6847351B2 (en) | 2001-08-13 | 2005-01-25 | Siemens Information And Communication Mobile, Llc | Tilt-based pointing for hand-held devices |
JP2003140815A (en) * | 2001-11-06 | 2003-05-16 | Nec Corp | Portable terminal |
US7100123B1 (en) | 2002-01-25 | 2006-08-29 | Microsoft Corporation | Electronic content search and delivery based on cursor location |
US20030231189A1 (en) | 2002-05-31 | 2003-12-18 | Microsoft Corporation | Altering a display on a viewing device based upon a user controlled orientation of the viewing device |
US7456823B2 (en) | 2002-06-14 | 2008-11-25 | Sony Corporation | User interface apparatus and portable information apparatus |
US7042441B2 (en) | 2002-06-28 | 2006-05-09 | Microsoft Corporation | Input device including a scroll wheel assembly for manipulating an image in multiple directions |
KR100955954B1 (en) | 2002-12-16 | 2010-05-04 | 마이크로소프트 코포레이션 | Systems and methods for interfacing with computer devices |
EP1491855A1 (en) * | 2003-06-23 | 2004-12-29 | Leica Geosystems AG | Optical inclination sensor |
EP1524588A1 (en) | 2003-10-15 | 2005-04-20 | Sony Ericsson Mobile Communications AB | User input device for a portable electronic device |
CN1943229A (en) | 2004-04-08 | 2007-04-04 | 皇家飞利浦电子股份有限公司 | Mobile projectable gui |
JP5080273B2 (en) | 2005-01-07 | 2012-11-21 | クアルコム,インコーポレイテッド | Tilt sensor based on optical flow |
US7519468B2 (en) | 2005-02-28 | 2009-04-14 | Research In Motion Limited | System and method for navigating a mobile device user interface with a directional sensing device |
CN103257684B (en) | 2005-05-17 | 2017-06-09 | 高通股份有限公司 | The signal output method and device of orientation-sensitive |
NO20052656D0 (en) * | 2005-06-02 | 2005-06-02 | Lumex As | Geometric image transformation based on text line searching |
DE602005006405T2 (en) * | 2005-06-24 | 2009-07-02 | Maxpat Trading & Marketing (Far East) Ltd., Tsimshatsui | Maisabstreifer |
CN101213550B (en) | 2005-06-30 | 2011-09-28 | 诺基亚公司 | Camera control means to allow operating of a destined location of the information surface of a presentation and information system |
US20070107068A1 (en) | 2005-10-14 | 2007-05-10 | Oqo, Inc. | Hybrid hardware/firmware multi-axis accelerometers for drop detect and tumble detect |
US7667686B2 (en) | 2006-02-01 | 2010-02-23 | Memsic, Inc. | Air-writing and motion sensing input for portable devices |
US8139026B2 (en) * | 2006-08-02 | 2012-03-20 | Research In Motion Limited | System and method for adjusting presentation of text and images on an electronic device according to an orientation of the device |
WO2008094458A1 (en) | 2007-01-26 | 2008-08-07 | F-Origin, Inc. | Viewing images with tilt control on a hand-held device |
US7615751B2 (en) * | 2007-12-27 | 2009-11-10 | Honeywell International Inc. | Adjustable motion detection sensor with cam |
US9092053B2 (en) * | 2008-06-17 | 2015-07-28 | Apple Inc. | Systems and methods for adjusting a display based on the user's position |
US8355031B2 (en) | 2009-03-17 | 2013-01-15 | Harris Corporation | Portable electronic devices with adjustable display orientation |
US9582166B2 (en) * | 2010-05-16 | 2017-02-28 | Nokia Technologies Oy | Method and apparatus for rendering user interface for location-based service having main view portion and preview portion |
-
2001
- 2001-05-16 FI FI20011039A patent/FI117488B/en not_active IP Right Cessation
-
2002
- 2002-02-08 CN CN028141067A patent/CN1714326B/en not_active Expired - Lifetime
- 2002-02-08 RU RU2003136273/09A patent/RU2288512C2/en not_active IP Right Cessation
- 2002-02-08 JP JP2002589943A patent/JP4175897B2/en not_active Expired - Lifetime
- 2002-02-08 AT AT02711899T patent/ATE317138T1/en not_active IP Right Cessation
- 2002-02-08 KR KR1020037014906A patent/KR100671585B1/en not_active IP Right Cessation
- 2002-02-08 WO PCT/FI2002/000095 patent/WO2002093331A1/en active IP Right Grant
- 2002-02-08 DE DE60208995T patent/DE60208995D1/en not_active Expired - Lifetime
- 2002-02-08 EP EP02711899A patent/EP1410142B1/en not_active Expired - Lifetime
- 2002-02-08 US US10/071,172 patent/US20020175896A1/en not_active Abandoned
-
2005
- 2005-06-23 US US11/159,786 patent/US7607111B2/en not_active Expired - Fee Related
-
2009
- 2009-09-29 US US12/569,797 patent/US20100020102A1/en not_active Abandoned
-
2010
- 2010-01-21 US US12/691,506 patent/US9727095B2/en active Active
- 2010-02-23 US US12/710,944 patent/US20100153891A1/en not_active Abandoned
-
2017
- 2017-07-14 US US15/650,765 patent/US11301196B2/en not_active Expired - Lifetime
Patent Citations (45)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6624824B1 (en) * | 1996-04-30 | 2003-09-23 | Sun Microsystems, Inc. | Tilt-scrolling on the sunpad |
US6115025A (en) * | 1997-09-30 | 2000-09-05 | Silicon Graphics, Inc. | System for maintaining orientation of a user interface as a display changes orientation |
US6489945B1 (en) * | 1998-02-11 | 2002-12-03 | Agilent Technologies, Inc. | Method and system for tracking attitude |
US6151208A (en) * | 1998-06-24 | 2000-11-21 | Digital Equipment Corporation | Wearable computing device mounted on superior dorsal aspect of a hand |
US6369794B1 (en) * | 1998-09-09 | 2002-04-09 | Matsushita Electric Industrial Co., Ltd. | Operation indication outputting device for giving operation indication according to type of user's action |
US6400376B1 (en) * | 1998-12-21 | 2002-06-04 | Ericsson Inc. | Display control for hand-held data processing device |
US6201554B1 (en) * | 1999-01-12 | 2001-03-13 | Ericsson Inc. | Device control apparatus for hand-held data processing device |
US6288704B1 (en) * | 1999-06-08 | 2001-09-11 | Vega, Vista, Inc. | Motion detection and tracking system to control navigation and display of object viewers |
US6375572B1 (en) * | 1999-10-04 | 2002-04-23 | Nintendo Co., Ltd. | Portable game apparatus with acceleration sensor and information storage medium storing a game progam |
US6567101B1 (en) * | 1999-10-13 | 2003-05-20 | Gateway, Inc. | System and method utilizing motion input for manipulating a display of data |
US6466198B1 (en) * | 1999-11-05 | 2002-10-15 | Innoventions, Inc. | View navigation and magnification of a hand-held device with a display |
US6552713B1 (en) * | 1999-12-16 | 2003-04-22 | Hewlett-Packard Company | Optical pointing device |
US6933923B2 (en) * | 2000-04-05 | 2005-08-23 | David Y. Feinstein | View navigation and magnification of a hand-held device with a display |
US20020167699A1 (en) * | 2000-05-17 | 2002-11-14 | Christopher Verplaetse | Motion-based input system for handheld devices |
US7688306B2 (en) * | 2000-10-02 | 2010-03-30 | Apple Inc. | Methods and apparatuses for operating a portable device based on an accelerometer |
US20020052209A1 (en) * | 2000-10-27 | 2002-05-02 | Stig Frohlund | Portable radio communications device |
US6872931B2 (en) * | 2000-11-06 | 2005-03-29 | Koninklijke Philips Electronics N.V. | Optical input device for measuring finger movement |
US6690358B2 (en) * | 2000-11-30 | 2004-02-10 | Alan Edward Kaplan | Display control for hand-held devices |
US6939231B2 (en) * | 2000-12-22 | 2005-09-06 | Nokia Corporation | Method for controlling a terminal display and a terminal |
US6675553B2 (en) * | 2001-02-09 | 2004-01-13 | Teepack Spezialmaschinen Gmbh & Co. Kg | Method and device for stacking and packing infusion bags |
US20040012566A1 (en) * | 2001-03-29 | 2004-01-22 | Bradski Gary R. | Intuitive mobile device interface to virtual spaces |
US20040196259A1 (en) * | 2001-03-29 | 2004-10-07 | Bradski Gary R. | Intuitive mobile device interface to virtual spaces |
US20020143489A1 (en) * | 2001-03-29 | 2002-10-03 | Orchard John T. | Method and apparatus for controlling a computing system |
US7058432B2 (en) * | 2001-04-20 | 2006-06-06 | Mitsubishi Denki Kabushiki Kaisha | Pointing device and mobile telephone |
US6797937B2 (en) * | 2001-07-24 | 2004-09-28 | Agilent Technologies, Inc. | System and method for reducing power consumption in an optical screen pointing device |
US20030126100A1 (en) * | 2001-12-26 | 2003-07-03 | Autodesk, Inc. | Fuzzy logic reasoning for inferring user location preferences |
US6765443B2 (en) * | 2002-02-21 | 2004-07-20 | Ericsson Inc. | Dynamic bias controller for power amplifier circuits |
US20040204067A1 (en) * | 2002-03-28 | 2004-10-14 | Nec Corporation | Portable apparatus including improved pointing device |
US7315751B2 (en) * | 2002-03-28 | 2008-01-01 | Nec Corporation | Portable apparatus including improved pointing device |
US7184025B2 (en) * | 2002-05-31 | 2007-02-27 | Microsoft Corporation | Altering a display on a viewing device based upon a user controlled orientation of the viewing device |
US7164411B2 (en) * | 2002-12-30 | 2007-01-16 | Nokia Corporatoin | Optical user interface for controlling portable electric device |
US6977675B2 (en) * | 2002-12-30 | 2005-12-20 | Motorola, Inc. | Method and apparatus for virtually expanding a display |
US20060146009A1 (en) * | 2003-01-22 | 2006-07-06 | Hanno Syrbe | Image control |
US7721968B2 (en) * | 2003-10-31 | 2010-05-25 | Iota Wireless, Llc | Concurrent data entry for a portable device |
US7242391B2 (en) * | 2003-12-29 | 2007-07-10 | Pixart Imaging Inc. | Optical navigation chip |
US20050208978A1 (en) * | 2004-03-16 | 2005-09-22 | Myorigo, L.L.C. | Mobile device with wide-angle optics and a radiation sensor |
US7567818B2 (en) * | 2004-03-16 | 2009-07-28 | Motionip L.L.C. | Mobile device with wide-angle optics and a radiation sensor |
US7301528B2 (en) * | 2004-03-23 | 2007-11-27 | Fujitsu Limited | Distinguishing tilt and translation motion components in handheld devices |
US7194816B2 (en) * | 2004-07-15 | 2007-03-27 | C&N Inc. | Mobile terminal apparatus |
US7138979B2 (en) * | 2004-08-27 | 2006-11-21 | Motorola, Inc. | Device orientation based input signal generation |
US7859553B2 (en) * | 2004-12-30 | 2010-12-28 | Lg Electronics Inc. | Image navigation in a mobile station |
US20090297062A1 (en) * | 2005-03-04 | 2009-12-03 | Molne Anders L | Mobile device with wide-angle optics and a radiation sensor |
US20090305727A1 (en) * | 2005-03-04 | 2009-12-10 | Heikki Pylkko | Mobile device with wide range-angle optics and a radiation sensor |
US20090303204A1 (en) * | 2007-01-05 | 2009-12-10 | Invensense Inc. | Controlling and accessing content using motion processing on mobile devices |
US8099124B2 (en) * | 2007-04-12 | 2012-01-17 | Symbol Technologies, Inc. | Method and system for correlating user/device activity with spatial orientation sensors |
Cited By (45)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11301196B2 (en) | 2001-05-16 | 2022-04-12 | Apple Inc. | Method, device and program for browsing information on a display |
US9727095B2 (en) | 2001-05-16 | 2017-08-08 | Apple Inc. | Method, device and program for browsing information on a display |
US20100020102A1 (en) * | 2001-05-16 | 2010-01-28 | Motionip, Llc | Method and device for browsing information on a display |
US20100125818A1 (en) * | 2001-05-16 | 2010-05-20 | Motionip, Llc | Method, device and program for browsing information on a display |
US20090305727A1 (en) * | 2005-03-04 | 2009-12-10 | Heikki Pylkko | Mobile device with wide range-angle optics and a radiation sensor |
US20090297062A1 (en) * | 2005-03-04 | 2009-12-03 | Molne Anders L | Mobile device with wide-angle optics and a radiation sensor |
US20100171691A1 (en) * | 2007-01-26 | 2010-07-08 | Ralph Cook | Viewing images with tilt control on a hand-held device |
US9507431B2 (en) | 2007-01-26 | 2016-11-29 | Apple Inc. | Viewing images with tilt-control on a hand-held device |
US10318017B2 (en) | 2007-01-26 | 2019-06-11 | Apple Inc. | Viewing images with tilt control on a hand-held device |
US8994644B2 (en) | 2007-01-26 | 2015-03-31 | Apple Inc. | Viewing images with tilt control on a hand-held device |
US8948788B2 (en) | 2008-05-28 | 2015-02-03 | Google Inc. | Motion-controlled views on mobile computing devices |
US20090325607A1 (en) * | 2008-05-28 | 2009-12-31 | Conway David P | Motion-controlled views on mobile computing devices |
US8847992B2 (en) * | 2008-08-22 | 2014-09-30 | Google Inc. | Navigation in a three dimensional environment using an orientation of a mobile device |
US20100045667A1 (en) * | 2008-08-22 | 2010-02-25 | Google Inc. | Navigation In a Three Dimensional Environment Using An Orientation Of A Mobile Device |
US20150052474A1 (en) * | 2009-07-13 | 2015-02-19 | Samsung Electronics Co., Ltd. | Scrolling method of mobile terminal and apparatus for performing the same |
US10082943B2 (en) * | 2009-07-13 | 2018-09-25 | Samsung Electronics Co., Ltd. | Scrolling method of mobile terminal and apparatus for performing the same |
US20110010659A1 (en) * | 2009-07-13 | 2011-01-13 | Samsung Electronics Co., Ltd. | Scrolling method of mobile terminal and apparatus for performing the same |
US9487388B2 (en) | 2012-06-21 | 2016-11-08 | Nextinput, Inc. | Ruggedized MEMS force die |
US9493342B2 (en) | 2012-06-21 | 2016-11-15 | Nextinput, Inc. | Wafer level MEMS force dies |
US9032818B2 (en) | 2012-07-05 | 2015-05-19 | Nextinput, Inc. | Microelectromechanical load sensor and methods of manufacturing the same |
US10490166B2 (en) * | 2012-07-09 | 2019-11-26 | Blackberry Limited | System and method for determining a display orientation of a mobile device |
US20140009498A1 (en) * | 2012-07-09 | 2014-01-09 | Research In Motion Limited | System and method for determining a display orientation of a mobile device |
US9902611B2 (en) | 2014-01-13 | 2018-02-27 | Nextinput, Inc. | Miniaturized and ruggedized wafer level MEMs force sensors |
US10466119B2 (en) | 2015-06-10 | 2019-11-05 | Nextinput, Inc. | Ruggedized wafer level MEMS force sensor with a tolerance trench |
WO2018078214A1 (en) * | 2016-10-31 | 2018-05-03 | Nokia Technologies Oy | Controlling content displayed in a display |
EP3316117A1 (en) * | 2016-10-31 | 2018-05-02 | Nokia Technologies OY | Controlling content displayed in a display |
US10671940B2 (en) | 2016-10-31 | 2020-06-02 | Nokia Technologies Oy | Controlling display of data to a person via a display apparatus |
US10809873B2 (en) | 2016-10-31 | 2020-10-20 | Nokia Technologies Oy | Controlling content displayed in a display |
US11243125B2 (en) | 2017-02-09 | 2022-02-08 | Nextinput, Inc. | Integrated piezoresistive and piezoelectric fusion force sensor |
US11808644B2 (en) | 2017-02-09 | 2023-11-07 | Qorvo Us, Inc. | Integrated piezoresistive and piezoelectric fusion force sensor |
US11255737B2 (en) | 2017-02-09 | 2022-02-22 | Nextinput, Inc. | Integrated digital force sensors and related methods of manufacture |
US11946817B2 (en) | 2017-02-09 | 2024-04-02 | DecaWave, Ltd. | Integrated digital force sensors and related methods of manufacture |
US11604104B2 (en) | 2017-02-09 | 2023-03-14 | Qorvo Us, Inc. | Integrated piezoresistive and piezoelectric fusion force sensor |
US11221263B2 (en) | 2017-07-19 | 2022-01-11 | Nextinput, Inc. | Microelectromechanical force sensor having a strain transfer layer arranged on the sensor die |
US11423686B2 (en) | 2017-07-25 | 2022-08-23 | Qorvo Us, Inc. | Integrated fingerprint and force sensor |
US11243126B2 (en) | 2017-07-27 | 2022-02-08 | Nextinput, Inc. | Wafer bonded piezoresistive and piezoelectric force sensor and related methods of manufacture |
US11946816B2 (en) | 2017-07-27 | 2024-04-02 | Nextinput, Inc. | Wafer bonded piezoresistive and piezoelectric force sensor and related methods of manufacture |
US11609131B2 (en) | 2017-07-27 | 2023-03-21 | Qorvo Us, Inc. | Wafer bonded piezoresistive and piezoelectric force sensor and related methods of manufacture |
US11898918B2 (en) | 2017-10-17 | 2024-02-13 | Nextinput, Inc. | Temperature coefficient of offset compensation for force sensor and strain gauge |
US11579028B2 (en) | 2017-10-17 | 2023-02-14 | Nextinput, Inc. | Temperature coefficient of offset compensation for force sensor and strain gauge |
US11385108B2 (en) | 2017-11-02 | 2022-07-12 | Nextinput, Inc. | Sealed force sensor with etch stop layer |
US11965787B2 (en) | 2017-11-02 | 2024-04-23 | Nextinput, Inc. | Sealed force sensor with etch stop layer |
US11874185B2 (en) | 2017-11-16 | 2024-01-16 | Nextinput, Inc. | Force attenuator for force sensor |
US11698310B2 (en) | 2019-01-10 | 2023-07-11 | Nextinput, Inc. | Slotted MEMS force sensor |
US10962427B2 (en) | 2019-01-10 | 2021-03-30 | Nextinput, Inc. | Slotted MEMS force sensor |
Also Published As
Publication number | Publication date |
---|---|
RU2003136273A (en) | 2005-05-27 |
CN1714326A (en) | 2005-12-28 |
US20170315594A1 (en) | 2017-11-02 |
FI20011039A (en) | 2002-11-17 |
KR20040007571A (en) | 2004-01-24 |
US11301196B2 (en) | 2022-04-12 |
WO2002093331A1 (en) | 2002-11-21 |
US7607111B2 (en) | 2009-10-20 |
US20170192729A9 (en) | 2017-07-06 |
EP1410142A1 (en) | 2004-04-21 |
KR100671585B1 (en) | 2007-01-18 |
US20060129951A1 (en) | 2006-06-15 |
US20100020102A1 (en) | 2010-01-28 |
US9727095B2 (en) | 2017-08-08 |
US20100125818A1 (en) | 2010-05-20 |
FI20011039A0 (en) | 2001-05-16 |
RU2288512C2 (en) | 2006-11-27 |
DE60208995D1 (en) | 2006-04-13 |
FI117488B (en) | 2006-10-31 |
JP4175897B2 (en) | 2008-11-05 |
CN1714326B (en) | 2012-05-30 |
US20020175896A1 (en) | 2002-11-28 |
JP2004534302A (en) | 2004-11-11 |
EP1410142B1 (en) | 2006-02-01 |
ATE317138T1 (en) | 2006-02-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11301196B2 (en) | Method, device and program for browsing information on a display | |
US10318017B2 (en) | Viewing images with tilt control on a hand-held device | |
US10095316B2 (en) | Scrolling and zooming of a portable device display with device motion | |
EP3042275B1 (en) | Tilting to scroll | |
US7952561B2 (en) | Method and apparatus for controlling application using motion of image pickup unit | |
USRE42336E1 (en) | Intuitive control of portable data displays | |
US7271795B2 (en) | Intuitive mobile device interface to virtual spaces | |
US20020158908A1 (en) | Web browser user interface for low-resolution displays | |
KR20020038950A (en) | A system and method utilizing motion input for manipulating a display of data | |
JP2004271671A (en) | Image display apparatus and terminal device equipped with the same | |
US20080092083A1 (en) | Personal viewing device | |
Nourbakhsh et al. | A Motion Sensor-Based User Interface for Construction Drawings Navigation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: APPLE INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MOTIONIP, LLC;REEL/FRAME:027520/0201 Effective date: 20100726 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |