US20030187886A1 - Method and apparatus for simultaneous highlighting of a physical version of a document and an electronic version of a document - Google Patents
Method and apparatus for simultaneous highlighting of a physical version of a document and an electronic version of a document Download PDFInfo
- Publication number
- US20030187886A1 US20030187886A1 US10/400,977 US40097703A US2003187886A1 US 20030187886 A1 US20030187886 A1 US 20030187886A1 US 40097703 A US40097703 A US 40097703A US 2003187886 A1 US2003187886 A1 US 2003187886A1
- Authority
- US
- United States
- Prior art keywords
- document
- annotation
- electronic
- paper document
- highlighted
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/583—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/5846—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using extracted text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/93—Document management systems
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10—TECHNICAL SUBJECTS COVERED BY FORMER USPC
- Y10S—TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10S707/00—Data processing: database and file management or data structures
- Y10S707/99931—Database or file accessing
- Y10S707/99933—Query processing, i.e. searching
- Y10S707/99936—Pattern matching access
Definitions
- This invention relates to scanning of a physical version (e.g., paper) of a document and more particularly to recording annotations made to a physical version of a document onto an electronic version of the document.
- a physical version e.g., paper
- Annotations that are generally made to a paper document include a printed annotation, a written annotation, a colored marking, a sound, and an image.
- Colored markings include highlights that are applied over the top of a machine-printed text.
- Pen scanners are capable of recording marks made to a paper document.
- Pen scanners are electronic devices used to record the path followed by a pen when a user writes or prints text or images on a paper document.
- Such pen scanners may include a pen with a video camera focused on a desktop or a paper document, a pen having an imbedded gyroscope to facilitate recording pen movements, or a pen having a small video camera focused on the tip of the pen scanner.
- Pen scanners have not been used, for example, to scan colored markings such as highlights. Although highlighted marks may be captured by a video camera, conventional systems do not provide for automatically mapping highlighted marks onto an electronic document.
- Another disadvantage to these conventional systems is that a user must manually access an electronic version of a document to make additional annotations to the document. It is therefore desirable to have a system for automatically preserving annotations, such as, for example, highlighted marks, that are made to a paper document and provides effective ways of accessing an electronic version of a document and electronically annotating a document automatically.
- a method and apparatus for preserving highlighted marks made to a paper document comprises annotating a portion of a paper document by highlighting, performing recognition on text in the highlighted portion of the paper document, accessing an electronic version of the paper document from a memory based on recognition results, matching information in the electronic version, and storing an indication of at least one highlighted mark with the electronic version of the paper document.
- FIG. 1 is a flow diagram of one embodiment of a process for highlighting a portion of a document
- FIG. 2 is a more detailed flow diagram of one embodiment of a process for highlighting a portion of a document
- FIG. 3 illustrates representative chains of highlights and their associated values.
- FIG. 4 illustrates one embodiment of an annotation system that includes an annotation device and a multi-function machine
- FIG. 5 illustrates another embodiment of an annotation system
- FIG. 6 illustrates one embodiment of a computer system.
- a device that records annotations (e.g., highlighting) made to a paper document, performing recognition on the information in the annotated portion of the document, an electronic version of the paper document, and storing the annotations, such as highlighted marks placed on a paper document, with the electronic version of the document.
- annotations e.g., highlighting
- the present invention also relates to apparatus for performing the operations herein.
- This apparatus may be specially constructed for the required purposes, or it may comprise a general purpose computer selectively activated or reconfigured by a computer program stored in the computer.
- a computer program may be stored in a computer readable storage medium, such as, but is not limited to, any type of disk including floppy disks, optical disks, CD-ROMs, and magnetic-optical disks, read-only memories (“ROMs”), random access memories (“RAMs”), erasable programmable read only memories (“EPROMs”), magnetic or optical cards, or any type of media suitable for storing electronic instructions, and each coupled to a computer system bus.
- a machine-readable medium includes any mechanism for storing or transmitting information in a form readable by a machine (e.g., a computer).
- a machine-readable medium includes ROM; RAM; magnetic disk storage media; optical storage media; flash memory devices; electrical, optical, acoustical or other form of propagated signals (e.g., carrier waves, infrared signals, digital signals, etc.); etc.
- FIG. 1 illustrates one embodiment of the process for simultaneously highlighting paper and electronic document.
- the process is performed by processing logic that may comprise hardware (e.g., dedicated logic), software (such as runs on a general purpose computer system or a dedicated machine), or a combination of both.
- the process begins by processing logic performing a highlighting function on a paper document with a highlighting pen having a camera focused on its tip (processing block 101 ).
- Processing logic associated with the camera records what is highlighted prior to (or substantially simultaneously with) ink from the highlighting pen being placed on and obscuring the document (processing block 102 ).
- Processing logic identifies one or more documents that contain highlighted text and identifies locations within those documents where the highlights occur (processing block 103 ).
- Processing logic records the highlights on an electronic copy of the document (processing block 104 ).
- the process utilizes a database containing an electronic copy of the paper document to facilitate the recording (and storing) portion of the process.
- FIG. 2 illustrates a more detailed flow diagram of the process of mapping images of highlighted text onto electronic documents.
- the process is performed by processing logic that may comprise hardware (e.g., dedicated logic), software (such as runs on a general purpose computer system or a dedicated machine), or a combination of both.
- processing logic extracting features from the highlighted text (processing block 201 ).
- Feature extraction may include, for example, performing optical character recognition (OCR), performing word shape analysis, or performing an extraction of characteristics based on black and white components in the highlighted portion.
- OCR optical character recognition
- processing logic composes a query using the features (processing block 202 ).
- processing logic posts the query (processing block 203 ).
- posting the query comprises receiving the queries sequentially and time-stamping them.
- a time out value may be used to split a sequence of queries into groups, with each group being from the same document.
- Posting the query also comprises applying the query to an unedited index.
- processing logic obtains a list of documents that contain the query and locations within each document where the query occurred. In one embodiment, the locations within each document where the query occurred are expressed as byte offsets.
- processing logic After posting the query, processing logic performs results accumulation (processing block 204 ).
- results accumulation the lists of documents returned by queries in a group are “accumulated.” Ideally, each query would return a single document, but the presence of duplicates, versions, or just documents with common text strings will more than likely cause this not to be true.
- processing logic increments a counter for each document that contains a query and records the location(s) within the document where the query was found. The results are accumulated in a buffer.
- the documents with highest number of times different highlighted words are located in documents are most likely to have been the ones highlighted by the user.
- the number of times that highlighted words are located in documents is set by a threshold. If none are chosen (of achieved ⁇ threshold), then processing logic assumes that the user must have highlighted a document not in the database.
- the presence of a new document, and thus a new group, may be automatically indicated in a number of ways.
- a button on the pen may be pressed to indicate a new document.
- a time out value may be used to indicate a new document is being processed. In such a case, if the pen has not processed any new information and the time out period expires, then the end of a document, and start of a new document, is indicated.
- the pen can have a gyro that indicates that the pen has been in one location for a predetermined period of time, thereby indicating that the user has finished with the current document.
- processing logic After performing results accumulation, processing logic performs results ranking (processing logic 205 ).
- the ranking process assigns a score that measures the “sequentialness” of the sequence of highlights applied to a document. The document with the lowest score is most likely to be the one that contained the highlights.
- a dynamic programming solution is used to perform the results ranking.
- the dynamic programming solution computes a “chain” of highlights that is the preferred chain of highlights and computes a score.
- FIG. 3 shows the sequence of highlights applied to two documents A and B. The starting positions within the documents where the corresponding highlighted text occurs are shown. For example, in document A the second highlighted text phrase occurs in positions 25 , 36 , and 89 .
- the dynamic programming algorithm processes the highlighted phrases sequentially. At each step it selects the starting position closest to the starting position of the previous highlighted phrase but after it. If it cannot select such a phrase it backtracks and deselects the most recent previous phrase for which there is an alternative with a smaller starting position. It substitutes that alternative and continues. Search strategies such as this are commonly known in art as depth first.
- a score is assigned to each highlight as the distance backward from the starting position of the previous highlight.
- the “sequentialness” score is the sum of these values for a given document.
- the sequence 30 - 36 - 49 - 51 - 55 - 58 is chosen for document A. Its score is zero.
- the sequence 8 - 32 - 7 - 21 -NULL- 8 is chosen for document B. Its score is 25+100+13 or 138.
- a default value of 100 is assigned for the fifth highlight since it was not found in document B. In this case, document A is the most likely one that contained the highlights.
- the output of results ranking is a set of documents, a score, a preferred list of highlights for each document.
- processing logic After performing results ranking, processing logic performs a database update (processing logic 706 ).
- processing logic records in the document database, for each document identified by ranking, the preferred list of highlights together with the times when those highlights were originally applied.
- the pen used in the highlighting process can be online (connected “live with the database) or offline.
- the database can indicate to the pen when it locates a unique document (e.g., results accumulation has found a stable set of documents).
- images of highlighted text or features, or queries
- the accumulated data is later downloaded to a database update (results accumulation and ranking) process.
- the inverted index could also be present on the pen. This would allow for iterative feedback to the user as if the user is using the pen in online mode. Thus, the results accumulation process could be performed on the pen.
- the pen has on/off, highlight on/off and new document indication buttons.
- the new document button is clicked when a user starts highlighting a new document. For example, referring back to FIG. 2, processing tests whether the user selects the new document button (processing block 210 ). If the new document button hasn't been selected, then the results ranking processing remains inactive and loops back to the beginning of processing block 210 . If the new document button has been selected, processing logic begins to perform results ranking on the information from the previous document.
- Ambiguous highlights are highlights that are not on a preferred chain or that map onto a document not selected by results ranking. In one embodiment, these are also recorded together with their timestamp and a pointer to the documents that were chosen by results ranking.
- FIG. 4 illustrates one embodiment of an annotation system.
- annotation system 100 comprises annotation device 140 and multi-function machine 180 such as computer system 10 described in FIG. 5.
- multi-function machine 180 may comprise devices such as a portable palm pilot and other suitable devices.
- a physical document (e.g., paper document) 110 includes a colored annotation 130 such as a mark left by a highlighter.
- annotation device 140 includes highlighter 154 and scanner 150 which includes a video camera coupled to highlighter 154 by a clip. The video camera may be located close to paper document 110 such as within one inch of the distance for recording visual marks, text, or images on the paper document.
- annotation system 100 also includes microphone 160 for recording sound, marker tip 170 for placing highlighted marks on the paper document, and micro switch 152 for activating tip switch 155 .
- Tip switch 155 is used when a user wishes to a capture or “pick up” marks, images, patterns, words or other marks on a paper document.
- scanner 150 is attached to a frame capture card (not shown) in a multifunction machine such as computer system 10 illustrated in FIG. 6 and tip switch 155 is wired to the switch of an input device button.
- Annotation device 140 allows marks to be made on a paper document either through using a highlighter, a standard pen, or other suitable marking utensil. These marks or annotations may then be associated with electronic content and/or are assigned a meaning using processing such as, for example, optical character recognition (“OCR”) and/or word shape analysis.
- OCR and word shape analysis are used to scan in text, such as printed text or written text, and determine characters or other information in the scanned in text. That is, the OCR processing and the word shape processing process the text or image data.
- OCR processing determines the text that is being captured by the video camera.
- Word shape processing determines the shape of characters that have been highlighted. Once one or more characters are determined, the entire word may be identified. This data is optionally stored in data repository (block 330 ) as document identifiers. These functions are performed by a feature extraction processing block.
- FIG. 5 illustrates another embodiment of an annotation system 400 .
- annotation device 140 transmits data to application program interface 410 .
- Application program interface 410 is a connection point at which information is transmitted between annotation device 140 and application server 430 .
- Data from application program interface 410 is received by control module 440 .
- Control module 440 controls the operation of application program interface 410 and recording server 420 .
- Candidate generation 450 then receives the data from control module 440 .
- Candidate generation 450 takes the data received from control module 440 and generates a set of document candidates that may contain text or images that resembles the text or images scanned from the paper document by annotation device 140 .
- This data is then sent to approximate string match 460 that attempts to match the string of text scanned in with strings of text that are stored in candidate documents 470 . It will be appreciated that candidate generation 450 and approximate string match 460 are connected to database 470 to access various portions of documents, characters, images or other suitable visual indicia.
- annotations such as highlighted marks are identified may depend upon whether the annotation is stored in memory.
- an image may be stored in the memory of multifunction machine 180 as a hyperlink. Therefore, an image captured by annotation device 140 that substantially or precisely matches an image stored in the memory of multifunction machine 180 is recorded as being located, for example, in a certain section of a page of a document. Captured image data that is not recognized as a hyperlink mark can optionally have its corresponding image text extracted. This image data can then serve as a parameter to a command or otherwise be used as input data. This provides an advantage related to recording page numbers that allows annotation device 140 to provide an electronic-to-paper hyperlinking and indexing mechanism.
- the highlighted area may cover annotations, printed text, or other marks.
- Annotations may include written annotations, printed annotations, colored markings, images, sound, or other visual indicia.
- the highlighter such as a yellow highlighter may cover a colored marking such as blue highlighted area or a green marked area.
- the phrases from the scanned paper document may be randomly selected or selected based upon specified instructions. These phrases are then used as queries to the text index for the electronic documents stored in the database (e.g., a data repository). The number of documents returned by each query, individually and in common, may then be calculated. Based upon this information, it may be estimated the number of words, characters, images or other suitable objects needed to identify a unique electronic document. Table 1 indicates in general the performance of phrase matching using the number of phrases and number of words per phrase.
- Duplicate documents are those documents in which a percent of unique vocabulary in common exceeds a specified threshold. The amount of duplicate documents recognized may be& reduced by the design of the user interface on the highlighting scanner or at retrieval time. Additionally, the process may be performed such that if more than one version of an electronic document is stored into a data repository, the latest version of the electronic document is automatically accessed.
- the data repository receives the resulting data from optical character recognition processing. Upon receipt, the data repository sends the data to the dictionary (block 350 ) to determine whether the word at issue is within the dictionary. If it is, a match is indicated. If not, processing logic looks to the word shape analysis processing block 320 for the result of its operation on the highlighted word or image.
- annotation system 100 may use a variety of program instructions but in particular, the image capture, pattern recognition, and other functional components of the annotation system 100 are generally coded in C++ with a commercial OCR software package being used to recognize words picked up from the paper document.
- FIG. 6 illustrates one embodiment of a computer system that performs operations described above.
- computer system 10 comprises processor 17 , memory 18 , and bus 15 such as bus or a point-to-point link.
- Processor 17 is coupled to the memory 18 by bus 15 .
- a number of user input/output devices such as a keyboard 20 and a display 25 , are coupled to chip set (not shown) which is then connected to processor 17 .
- the chipset (not shown) is typically connected to processor 17 using a bus that is different from bus 15 .
- Processor 17 represents a central processing unit of any type of architecture (e.g., the Intel architecture, Hewlett Packard architecture, Sun Microsystems architecture, IBM architecture, hybrid architecture, etc.). In addition, processor 17 could be implemented on one or more chips.
- Memory 18 represents one or more mechanisms for storing data such as the number of times the second code is checked and the results of checking the second code. Memory 18 may include ROM, RAM, magnetic disk storage mediums, optical storage mediums, flash. memory devices, and/or other machine-readable mediums.
- bus 15 may comprise one or more buses (e.g., accelerated graphics port bus, peripheral component interconnect bus, industry standard architecture bus, X-Bus, video electronics standards association related to buses, etc.) and bridges (also termed as bus controllers).
- buses e.g., accelerated graphics port bus, peripheral component interconnect bus, industry standard architecture bus, X-Bus, video electronics standards association related to buses, etc.
- bridges also termed as bus controllers.
- Network 30 represents one or more network connections for transmitting data over a machine readable media. Certain operations could also be implemented on multiple computers connected through such a network.
- FIG. 6 also illustrates that memory 18 has stored therein data 35 and program instructions (e.g. software, computer program, etc.) 36 .
- Data 35 represents data stored in one or more of the formats described herein.
- Program instructions 36 represents the necessary code for performing any and/or all of the techniques described with reference to FIGS. 1 - 2 . It will be recognized by one of ordinary skill in the art that the memory 18 preferably contains additional software (not shown), which is not necessary to understanding the invention.
- FIG. 5 additionally illustrates that the processor 17 includes decoder 40 .
- Decoder 40 is used for decoding instructions received by processor 17 into control signals and/or microcode entry points. In response to these control signals and/or microcode entry points, decoder 40 performs the appropriate operations.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Databases & Information Systems (AREA)
- Library & Information Science (AREA)
- Data Mining & Analysis (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Business, Economics & Management (AREA)
- General Business, Economics & Management (AREA)
- Document Processing Apparatus (AREA)
Abstract
A method and an apparatus for simultaneous highlighting of a paper document is described. The device includes a highlighter. A scanner is used and configured to capture at least one highlighted mark place on a paper document. The scanner is coupled to a memory for storing electronic versions of documents. An electronic document is accessed when a portion of the electronic document matches a portion of the paper document.
Description
- This invention relates to scanning of a physical version (e.g., paper) of a document and more particularly to recording annotations made to a physical version of a document onto an electronic version of the document.
- Although individuals are increasingly using electronic documents in place of paper documents, some users still prefer to read and annotate paper documents for a variety of reasons. For example, some users are more comfortable using a paper document instead of an electronic document for editing purposes. Other reasons relate to the numerous advantages of using a paper document. Such advantages include the high-resolution associated with a paper document, the portability of the paper document, the ease of copying the paper document, and the ease of editing or adding annotations to a paper document. Additionally, annotating a paper document does not require a power source which is required to directly annotate an electronic document.
- Annotations that are generally made to a paper document include a printed annotation, a written annotation, a colored marking, a sound, and an image. Colored markings include highlights that are applied over the top of a machine-printed text.
- Pen scanners are capable of recording marks made to a paper document. Pen scanners are electronic devices used to record the path followed by a pen when a user writes or prints text or images on a paper document. Such pen scanners may include a pen with a video camera focused on a desktop or a paper document, a pen having an imbedded gyroscope to facilitate recording pen movements, or a pen having a small video camera focused on the tip of the pen scanner.
- Pen scanners, however, have not been used, for example, to scan colored markings such as highlights. Although highlighted marks may be captured by a video camera, conventional systems do not provide for automatically mapping highlighted marks onto an electronic document.
- Another disadvantage to these conventional systems is that a user must manually access an electronic version of a document to make additional annotations to the document. It is therefore desirable to have a system for automatically preserving annotations, such as, for example, highlighted marks, that are made to a paper document and provides effective ways of accessing an electronic version of a document and electronically annotating a document automatically.
- A method and apparatus for preserving highlighted marks made to a paper document is disclosed. In one embodiment, a method for annotating a paper document comprises annotating a portion of a paper document by highlighting, performing recognition on text in the highlighted portion of the paper document, accessing an electronic version of the paper document from a memory based on recognition results, matching information in the electronic version, and storing an indication of at least one highlighted mark with the electronic version of the paper document.
- The features, aspects, and advantages of the invention will become more thoroughly apparent from the following detailed description, appended claims, and accompanying drawings in which:
- FIG. 1 is a flow diagram of one embodiment of a process for highlighting a portion of a document;
- FIG. 2 is a more detailed flow diagram of one embodiment of a process for highlighting a portion of a document;
- FIG. 3 illustrates representative chains of highlights and their associated values.
- FIG. 4 illustrates one embodiment of an annotation system that includes an annotation device and a multi-function machine;
- FIG. 5 illustrates another embodiment of an annotation system; and
- FIG. 6 illustrates one embodiment of a computer system.
- In accordance with one embodiment of the invention, a device is disclosed that records annotations (e.g., highlighting) made to a paper document, performing recognition on the information in the annotated portion of the document, an electronic version of the paper document, and storing the annotations, such as highlighted marks placed on a paper document, with the electronic version of the document.
- Some portions of the detailed descriptions which follow are presented in terms of algorithms and symbolic representations of operations on data bits within a computer memory. These algorithmic descriptions and representations are the means used by those skilled in the data processing arts to most effectively convey the substance of their work to others skilled in the art. An algorithm is here, and generally, conceived to be a self-consistent sequence of steps leading to a desired result. The steps are those requiring physical manipulations of physical quantities. Usually, though not necessarily, these quantities take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared, and otherwise manipulated. It has proven convenient at times, principally for reasons of common usage, to refer to these signals as bits, values, elements, symbols, characters, terms, numbers, or the like.
- It should be borne in mind, however, that all of these and similar terms are to be associated with the appropriate physical quantities and are merely convenient labels applied to these quantities. Unless specifically stated otherwise as apparent from the following discussion, it is appreciated that throughout the description, discussions utilizing terms such as “processing” or “computing” or “calculating” or “determining” or “displaying” or the like, refer to the action and processes of a computer system, or similar electronic computing device, that manipulates and transforms data represented as physical (electronic) quantities within the computer system's registers and memories into other data similarly represented as physical quantities within the computer system memories or registers or other such information storage, transmission or display devices.
- The present invention also relates to apparatus for performing the operations herein. This apparatus may be specially constructed for the required purposes, or it may comprise a general purpose computer selectively activated or reconfigured by a computer program stored in the computer. Such a computer program may be stored in a computer readable storage medium, such as, but is not limited to, any type of disk including floppy disks, optical disks, CD-ROMs, and magnetic-optical disks, read-only memories (“ROMs”), random access memories (“RAMs”), erasable programmable read only memories (“EPROMs”), magnetic or optical cards, or any type of media suitable for storing electronic instructions, and each coupled to a computer system bus.
- The algorithms and displays presented herein are not inherently related to any particular computer or other apparatus. Various general purpose systems may be used with programs in accordance with the teachings herein, or it may prove convenient to construct more specialized apparatus to perform the required method steps. The required structure for a variety of these systems will appear from the description below. In addition, the present invention is not described with reference to any particular programming language. It will be appreciated that a variety of programming languages may be used to implement the teachings of the invention as described herein.
- A machine-readable medium includes any mechanism for storing or transmitting information in a form readable by a machine (e.g., a computer). For example, a machine-readable medium includes ROM; RAM; magnetic disk storage media; optical storage media; flash memory devices; electrical, optical, acoustical or other form of propagated signals (e.g., carrier waves, infrared signals, digital signals, etc.); etc.
- Overview
- The present invention provides for simultaneously highlighting paper and electronic documents. The results of the process are to map images of highlighted text onto electronic documents. FIG. 1 illustrates one embodiment of the process for simultaneously highlighting paper and electronic document. The process is performed by processing logic that may comprise hardware (e.g., dedicated logic), software (such as runs on a general purpose computer system or a dedicated machine), or a combination of both.
- Referring to FIG. 1, the process begins by processing logic performing a highlighting function on a paper document with a highlighting pen having a camera focused on its tip (processing block101). Processing logic associated with the camera records what is highlighted prior to (or substantially simultaneously with) ink from the highlighting pen being placed on and obscuring the document (processing block 102). Processing logic identifies one or more documents that contain highlighted text and identifies locations within those documents where the highlights occur (processing block 103). Processing logic records the highlights on an electronic copy of the document (processing block 104). The process utilizes a database containing an electronic copy of the paper document to facilitate the recording (and storing) portion of the process.
- FIG. 2 illustrates a more detailed flow diagram of the process of mapping images of highlighted text onto electronic documents. The process is performed by processing logic that may comprise hardware (e.g., dedicated logic), software (such as runs on a general purpose computer system or a dedicated machine), or a combination of both. Referring to FIG. 2, the process begins with processing logic extracting features from the highlighted text (processing block201). Feature extraction may include, for example, performing optical character recognition (OCR), performing word shape analysis, or performing an extraction of characteristics based on black and white components in the highlighted portion. Next, processing logic composes a query using the features (processing block 202). Using the query, processing logic posts the query (processing block 203).
- In one embodiment, posting the query comprises receiving the queries sequentially and time-stamping them. A time out value may be used to split a sequence of queries into groups, with each group being from the same document. Posting the query also comprises applying the query to an unedited index. In response to posting the query, processing logic obtains a list of documents that contain the query and locations within each document where the query occurred. In one embodiment, the locations within each document where the query occurred are expressed as byte offsets.
- After posting the query, processing logic performs results accumulation (processing block204). In results accumulation, the lists of documents returned by queries in a group are “accumulated.” Ideally, each query would return a single document, but the presence of duplicates, versions, or just documents with common text strings will more than likely cause this not to be true. In one embodiment, processing logic increments a counter for each document that contains a query and records the location(s) within the document where the query was found. The results are accumulated in a buffer.
- After processing the queries in a group, the documents with highest number of times different highlighted words are located in documents are most likely to have been the ones highlighted by the user. In one embodiment, the number of times that highlighted words are located in documents is set by a threshold. If none are chosen (of achieved<threshold), then processing logic assumes that the user must have highlighted a document not in the database.
- The presence of a new document, and thus a new group, may be automatically indicated in a number of ways. In one embodiment, a button on the pen may be pressed to indicate a new document. In another embodiment, a time out value may be used to indicate a new document is being processed. In such a case, if the pen has not processed any new information and the time out period expires, then the end of a document, and start of a new document, is indicated. In another embodiment, the pen can have a gyro that indicates that the pen has been in one location for a predetermined period of time, thereby indicating that the user has finished with the current document.
- After performing results accumulation, processing logic performs results ranking (processing logic205). The ranking process assigns a score that measures the “sequentialness” of the sequence of highlights applied to a document. The document with the lowest score is most likely to be the one that contained the highlights.
- In one embodiment, a dynamic programming solution is used to perform the results ranking. The dynamic programming solution computes a “chain” of highlights that is the preferred chain of highlights and computes a score. For example, FIG. 3 shows the sequence of highlights applied to two documents A and B. The starting positions within the documents where the corresponding highlighted text occurs are shown. For example, in document A the second highlighted text phrase occurs in
positions - A score is assigned to each highlight as the distance backward from the starting position of the previous highlight. The “sequentialness” score is the sum of these values for a given document. In FIG. 3, the sequence30-36-49-51-55-58 is chosen for document A. Its score is zero. The sequence 8-32-7-21-NULL-8 is chosen for document B. Its score is 25+100+13 or 138. A default value of 100 is assigned for the fifth highlight since it was not found in document B. In this case, document A is the most likely one that contained the highlights.
- In one embodiment, the output of results ranking is a set of documents, a score, a preferred list of highlights for each document.
- After performing results ranking, processing logic performs a database update (processing logic706). When performing the database update, processing logic records in the document database, for each document identified by ranking, the preferred list of highlights together with the times when those highlights were originally applied.
- An Exemplary Pen
- The pen used in the highlighting process can be online (connected “live with the database) or offline. When online, the database can indicate to the pen when it locates a unique document (e.g., results accumulation has found a stable set of documents). When offline, images of highlighted text (or features, or queries) are accumulated on the pen together with their creation timestamps. The accumulated data is later downloaded to a database update (results accumulation and ranking) process.
- The inverted index could also be present on the pen. This would allow for iterative feedback to the user as if the user is using the pen in online mode. Thus, the results accumulation process could be performed on the pen.
- In one embodiment, the pen has on/off, highlight on/off and new document indication buttons. The new document button is clicked when a user starts highlighting a new document. For example, referring back to FIG. 2, processing tests whether the user selects the new document button (processing block210). If the new document button hasn't been selected, then the results ranking processing remains inactive and loops back to the beginning of
processing block 210. If the new document button has been selected, processing logic begins to perform results ranking on the information from the previous document. - Ambiguous highlights are highlights that are not on a preferred chain or that map onto a document not selected by results ranking. In one embodiment, these are also recorded together with their timestamp and a pointer to the documents that were chosen by results ranking.
- An Exemplary System
- FIG. 4 illustrates one embodiment of an annotation system. Referring to FIG. 4,
annotation system 100 comprisesannotation device 140 andmulti-function machine 180 such ascomputer system 10 described in FIG. 5. It will be appreciated thatmulti-function machine 180 may comprise devices such as a portable palm pilot and other suitable devices. A physical document (e.g., paper document) 110 includes acolored annotation 130 such as a mark left by a highlighter. One embodiment ofannotation device 140 includeshighlighter 154 andscanner 150 which includes a video camera coupled tohighlighter 154 by a clip. The video camera may be located close topaper document 110 such as within one inch of the distance for recording visual marks, text, or images on the paper document. - In one embodiment,
annotation system 100 also includesmicrophone 160 for recording sound,marker tip 170 for placing highlighted marks on the paper document, and micro switch 152 for activating tip switch 155. Tip switch 155 is used when a user wishes to a capture or “pick up” marks, images, patterns, words or other marks on a paper document. In one embodiment,scanner 150 is attached to a frame capture card (not shown) in a multifunction machine such ascomputer system 10 illustrated in FIG. 6 and tip switch 155 is wired to the switch of an input device button. -
Annotation device 140 allows marks to be made on a paper document either through using a highlighter, a standard pen, or other suitable marking utensil. These marks or annotations may then be associated with electronic content and/or are assigned a meaning using processing such as, for example, optical character recognition (“OCR”) and/or word shape analysis. OCR and word shape analysis are used to scan in text, such as printed text or written text, and determine characters or other information in the scanned in text. That is, the OCR processing and the word shape processing process the text or image data. OCR processing determines the text that is being captured by the video camera. Word shape processing determines the shape of characters that have been highlighted. Once one or more characters are determined, the entire word may be identified. This data is optionally stored in data repository (block 330) as document identifiers. These functions are performed by a feature extraction processing block. - FIG. 5 illustrates another embodiment of an
annotation system 400. Referring to FIG. 5,annotation device 140 transmits data toapplication program interface 410.Application program interface 410 is a connection point at which information is transmitted betweenannotation device 140 andapplication server 430. Data fromapplication program interface 410 is received bycontrol module 440.Control module 440 controls the operation ofapplication program interface 410 andrecording server 420.Candidate generation 450 then receives the data fromcontrol module 440.Candidate generation 450 takes the data received fromcontrol module 440 and generates a set of document candidates that may contain text or images that resembles the text or images scanned from the paper document byannotation device 140. This data is then sent toapproximate string match 460 that attempts to match the string of text scanned in with strings of text that are stored in candidate documents 470. It will be appreciated thatcandidate generation 450 andapproximate string match 460 are connected todatabase 470 to access various portions of documents, characters, images or other suitable visual indicia. - The manner in which annotations such as highlighted marks are identified may depend upon whether the annotation is stored in memory. For example, an image may be stored in the memory of
multifunction machine 180 as a hyperlink. Therefore, an image captured byannotation device 140 that substantially or precisely matches an image stored in the memory ofmultifunction machine 180 is recorded as being located, for example, in a certain section of a page of a document. Captured image data that is not recognized as a hyperlink mark can optionally have its corresponding image text extracted. This image data can then serve as a parameter to a command or otherwise be used as input data. This provides an advantage related to recording page numbers that allowsannotation device 140 to provide an electronic-to-paper hyperlinking and indexing mechanism. - The highlighted area may cover annotations, printed text, or other marks. Annotations may include written annotations, printed annotations, colored markings, images, sound, or other visual indicia. Regarding the colored marking, the highlighter such as a yellow highlighter may cover a colored marking such as blue highlighted area or a green marked area.
- It will be appreciated that the phrases from the scanned paper document may be randomly selected or selected based upon specified instructions. These phrases are then used as queries to the text index for the electronic documents stored in the database (e.g., a data repository). The number of documents returned by each query, individually and in common, may then be calculated. Based upon this information, it may be estimated the number of words, characters, images or other suitable objects needed to identify a unique electronic document. Table1 indicates in general the performance of phrase matching using the number of phrases and number of words per phrase.
TABLE 1 Performance of Phrase Matching Number of Number of words per phrase phrases 3 4 5 6 1 29% 35% 38% 39% 65 52 28 27 2 42% 45% 45% 46% 21 16 10 9 3 47% 48% 49% 49% 9 8 6 6 4 50% 50% 52% 52% 6 6 5 4 - Alternatively, on the first pass through the process described herein, if more than one electronic document is identified, additional images or text from the paper document may be scanned in order to obtain a single unique document. Duplicate documents are those documents in which a percent of unique vocabulary in common exceeds a specified threshold. The amount of duplicate documents recognized may be& reduced by the design of the user interface on the highlighting scanner or at retrieval time. Additionally, the process may be performed such that if more than one version of an electronic document is stored into a data repository, the latest version of the electronic document is automatically accessed.
- The data repository receives the resulting data from optical character recognition processing. Upon receipt, the data repository sends the data to the dictionary (block350) to determine whether the word at issue is within the dictionary. If it is, a match is indicated. If not, processing logic looks to the word shape analysis processing block 320 for the result of its operation on the highlighted word or image.
- It will be appreciated that
annotation system 100 may use a variety of program instructions but in particular, the image capture, pattern recognition, and other functional components of theannotation system 100 are generally coded in C++ with a commercial OCR software package being used to recognize words picked up from the paper document. - FIG. 6 illustrates one embodiment of a computer system that performs operations described above. Referring to FIG. 6,
computer system 10 comprisesprocessor 17,memory 18, and bus 15 such as bus or a point-to-point link.Processor 17 is coupled to thememory 18 by bus 15. In addition, a number of user input/output devices, such as akeyboard 20 and adisplay 25, are coupled to chip set (not shown) which is then connected toprocessor 17. The chipset (not shown) is typically connected toprocessor 17 using a bus that is different from bus 15. -
Processor 17 represents a central processing unit of any type of architecture (e.g., the Intel architecture, Hewlett Packard architecture, Sun Microsystems architecture, IBM architecture, hybrid architecture, etc.). In addition,processor 17 could be implemented on one or more chips.Memory 18 represents one or more mechanisms for storing data such as the number of times the second code is checked and the results of checking the second code.Memory 18 may include ROM, RAM, magnetic disk storage mediums, optical storage mediums, flash. memory devices, and/or other machine-readable mediums. In one example, bus 15 may comprise one or more buses (e.g., accelerated graphics port bus, peripheral component interconnect bus, industry standard architecture bus, X-Bus, video electronics standards association related to buses, etc.) and bridges (also termed as bus controllers). - While this embodiment is described in relation to a single processor computer system, the operations described herein could be implemented in a multi-processor computer system. In addition to other devices, one or more of a
network 30 may be present.Network 30 represents one or more network connections for transmitting data over a machine readable media. Certain operations could also be implemented on multiple computers connected through such a network. - FIG. 6 also illustrates that
memory 18 has stored thereindata 35 and program instructions (e.g. software, computer program, etc.) 36.Data 35 represents data stored in one or more of the formats described herein.Program instructions 36 represents the necessary code for performing any and/or all of the techniques described with reference to FIGS. 1-2. It will be recognized by one of ordinary skill in the art that thememory 18 preferably contains additional software (not shown), which is not necessary to understanding the invention. - FIG. 5 additionally illustrates that the
processor 17 includesdecoder 40.Decoder 40 is used for decoding instructions received byprocessor 17 into control signals and/or microcode entry points. In response to these control signals and/or microcode entry points,decoder 40 performs the appropriate operations. - In the preceding detailed description, the invention is described with reference to specific embodiments thereof. It will, however, be evident that various modifications and changes may be made thereto without departing from the broader spirit and scope of the invention as set forth in the claims. The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense.
Claims (17)
1. A method comprising:
annotating a portion of a paper document by highlighting;
performing recognition on text in the highlighted portion of the paper document;
accessing an electronic version of the paper document from a memory based on recognition results matching information in the electronic version; and
storing an indication of at least one highlighted mark with the electronic version of the paper document.
2. The method of in claim 1 , further comprising:
identifying the electronic version of the paper document by matching at least one scanned image with an image in the electronic version of the document.
3. The method of claim 1 , wherein the annotation comprises one of a group that includes at least a printed annotation, a written annotation, a colored marking, a sound, a highlighted mark and an image.
4. The method of claim 1 , further comprising:
mapping the at least one highlighted mark onto the electronic version of the document.
5. The method of claim 3 , further comprising determining by a word shape analysis module at least one characteristic of a character in the annotation.
6. The method of claim 1 , further comprising recording audio concurrently with the annotation.
7. A system for preserving at least one highlighted mark on a paper document comprising:
a highlighter;
a scanner configured to capture at least one of a plurality of characters, an image, and at least one highlighted mark placed on a paper document by the highlighter;
a memory coupled to the scanner and the processor for storing an electronic version of the document with the highlighted mark, the memory permitting access to the electronic version of the paper document using a scanner to identify at least a portion of the electronic document.
8. The system of claim 7 , wherein the memory permits access to the electronic version of the document using one of optical character recognition and word shape analysis.
9. The system of claim 7 , wherein the annotation is one from a group that includes a printed annotation, a written annotation, a colored marking, a sound, and an image.
10. The system of claim 7 , wherein the scanner directly maps the highlighted mark onto the electronic version of the document.
11. The system of claim 7 , further comprising a word shape analysis module for determining at least one characteristic of a character in the printed document.
12. The system of claim 11 , wherein the word shape analysis module is coupled to a dictionary that matches at least one word shape.
13. The system of claim 7 , wherein the scanner is coupled to a microphone to capture audio information.
14. An apparatus comprising:
means for annotating at least a portion of a paper document;
means for accessing an electronic document based on recognition results of an annotated portion of the paper document matching a portion of the electronic document; and
means for recording the annotation with the electronic document.
15. The apparatus of claim 14 , wherein the annotation comprises at least one of a group that includes a printed annotation, a written annotation, a colored marking, a sound, a highlighted mark and an image.
16. An article of manufacture having one or more recordable media having a plurality of executable instructions stored thereon which, when executed by a processing device, cause the processing device to:
access an electronic document based on a match between information resulting from performing recognition performed on an annotated portion of a paper document; and
store an indication of the annotation with the electronic document.
17. The article of manufacture of claim 16 , wherein the annotation is at least one of a group including one of a printed annotation, a written annotation, a colored marking, a sound, and an image.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/400,977 US20030187886A1 (en) | 2000-09-01 | 2003-03-26 | Method and apparatus for simultaneous highlighting of a physical version of a document and an electronic version of a document |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/653,965 US6671684B1 (en) | 2000-09-01 | 2000-09-01 | Method and apparatus for simultaneous highlighting of a physical version of a document and an electronic version of a document |
US10/400,977 US20030187886A1 (en) | 2000-09-01 | 2003-03-26 | Method and apparatus for simultaneous highlighting of a physical version of a document and an electronic version of a document |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/653,965 Continuation US6671684B1 (en) | 2000-09-01 | 2000-09-01 | Method and apparatus for simultaneous highlighting of a physical version of a document and an electronic version of a document |
Publications (1)
Publication Number | Publication Date |
---|---|
US20030187886A1 true US20030187886A1 (en) | 2003-10-02 |
Family
ID=28455123
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/653,965 Expired - Lifetime US6671684B1 (en) | 2000-09-01 | 2000-09-01 | Method and apparatus for simultaneous highlighting of a physical version of a document and an electronic version of a document |
US10/400,977 Abandoned US20030187886A1 (en) | 2000-09-01 | 2003-03-26 | Method and apparatus for simultaneous highlighting of a physical version of a document and an electronic version of a document |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/653,965 Expired - Lifetime US6671684B1 (en) | 2000-09-01 | 2000-09-01 | Method and apparatus for simultaneous highlighting of a physical version of a document and an electronic version of a document |
Country Status (1)
Country | Link |
---|---|
US (2) | US6671684B1 (en) |
Cited By (97)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030050927A1 (en) * | 2001-09-07 | 2003-03-13 | Araha, Inc. | System and method for location, understanding and assimilation of digital documents through abstract indicia |
WO2006023937A2 (en) * | 2004-08-23 | 2006-03-02 | Exbiblio B.V. | A portable scanning device |
US20060085477A1 (en) * | 2004-10-01 | 2006-04-20 | Ricoh Company, Ltd. | Techniques for retrieving documents using an image capture device |
US20070296695A1 (en) * | 2006-06-27 | 2007-12-27 | Fuji Xerox Co., Ltd. | Document processing system, document processing method, computer readable medium and data signal |
US20080005101A1 (en) * | 2006-06-23 | 2008-01-03 | Rohit Chandra | Method and apparatus for determining the significance and relevance of a web page, or a portion thereof |
US20080016091A1 (en) * | 2006-06-22 | 2008-01-17 | Rohit Chandra | Method and apparatus for highlighting a portion of an internet document for collaboration and subsequent retrieval |
US20080118064A1 (en) * | 2006-11-22 | 2008-05-22 | Bhogal Kulvir S | Method, system, and program product for encrypting portions of a document using selective annotation |
EP1756704A4 (en) * | 2004-02-15 | 2009-04-29 | Exbiblio Bv | Publishing techniques for adding value to a rendered document |
US7747428B1 (en) * | 2003-09-24 | 2010-06-29 | Yahoo! Inc. | Visibly distinguishing portions of compound words |
US7812860B2 (en) | 2004-04-01 | 2010-10-12 | Exbiblio B.V. | Handheld device for capturing text from both a document printed on paper and a document displayed on a dynamic display device |
US20110081892A1 (en) * | 2005-08-23 | 2011-04-07 | Ricoh Co., Ltd. | System and methods for use of voice mail and email in a mixed media environment |
US7990556B2 (en) | 2004-12-03 | 2011-08-02 | Google Inc. | Association of a portable scanner with input/output and storage devices |
US8073263B2 (en) | 2006-07-31 | 2011-12-06 | Ricoh Co., Ltd. | Multi-classifier selection and monitoring for MMR-based image recognition |
US8086038B2 (en) | 2007-07-11 | 2011-12-27 | Ricoh Co., Ltd. | Invisible junction features for patch recognition |
US20120042288A1 (en) * | 2010-08-16 | 2012-02-16 | Fuji Xerox Co., Ltd. | Systems and methods for interactions with documents across paper and computers |
US8144921B2 (en) | 2007-07-11 | 2012-03-27 | Ricoh Co., Ltd. | Information retrieval using invisible junctions and geometric constraints |
US8156427B2 (en) | 2005-08-23 | 2012-04-10 | Ricoh Co. Ltd. | User interface for mixed media reality |
US8156115B1 (en) | 2007-07-11 | 2012-04-10 | Ricoh Co. Ltd. | Document-based networking with mixed media reality |
US8156116B2 (en) | 2006-07-31 | 2012-04-10 | Ricoh Co., Ltd | Dynamic presentation of targeted information in a mixed media reality recognition system |
US8176054B2 (en) | 2007-07-12 | 2012-05-08 | Ricoh Co. Ltd | Retrieving electronic documents by converting them to synthetic text |
US8179563B2 (en) | 2004-08-23 | 2012-05-15 | Google Inc. | Portable scanning device |
US8184155B2 (en) | 2007-07-11 | 2012-05-22 | Ricoh Co. Ltd. | Recognition and tracking using invisible junctions |
US8195659B2 (en) | 2005-08-23 | 2012-06-05 | Ricoh Co. Ltd. | Integration and use of mixed media documents |
US8201076B2 (en) | 2006-07-31 | 2012-06-12 | Ricoh Co., Ltd. | Capturing symbolic information from documents upon printing |
US8238609B2 (en) | 2007-01-18 | 2012-08-07 | Ricoh Co., Ltd. | Synthetic image and video generation from ground truth data |
US8261094B2 (en) | 2004-04-19 | 2012-09-04 | Google Inc. | Secure data gathering from rendered documents |
US8276088B2 (en) | 2007-07-11 | 2012-09-25 | Ricoh Co., Ltd. | User interface for three-dimensional navigation |
US8332401B2 (en) | 2004-10-01 | 2012-12-11 | Ricoh Co., Ltd | Method and system for position-based image matching in a mixed media environment |
US8335789B2 (en) | 2004-10-01 | 2012-12-18 | Ricoh Co., Ltd. | Method and system for document fingerprint matching in a mixed media environment |
US8346620B2 (en) | 2004-07-19 | 2013-01-01 | Google Inc. | Automatic modification of web pages |
US8369655B2 (en) | 2006-07-31 | 2013-02-05 | Ricoh Co., Ltd. | Mixed media reality recognition using multiple specialized indexes |
US8385660B2 (en) | 2009-06-24 | 2013-02-26 | Ricoh Co., Ltd. | Mixed media reality indexing and retrieval for repeated content |
US8385589B2 (en) | 2008-05-15 | 2013-02-26 | Berna Erol | Web-based content detection in images, extraction and recognition |
US8418055B2 (en) | 2009-02-18 | 2013-04-09 | Google Inc. | Identifying a document by performing spectral analysis on the contents of the document |
US8442331B2 (en) | 2004-02-15 | 2013-05-14 | Google Inc. | Capturing text from rendered documents using supplemental information |
US8447066B2 (en) | 2009-03-12 | 2013-05-21 | Google Inc. | Performing actions based on capturing information from rendered documents, such as documents under copyright |
US8489624B2 (en) | 2004-05-17 | 2013-07-16 | Google, Inc. | Processing techniques for text capture from a rendered document |
US8489987B2 (en) | 2006-07-31 | 2013-07-16 | Ricoh Co., Ltd. | Monitoring and analyzing creation and usage of visual content using image and hotspot interaction |
US8505090B2 (en) | 2004-04-01 | 2013-08-06 | Google Inc. | Archive of text captures from rendered documents |
US8510283B2 (en) | 2006-07-31 | 2013-08-13 | Ricoh Co., Ltd. | Automatic adaption of an image recognition system to image capture devices |
US8521737B2 (en) | 2004-10-01 | 2013-08-27 | Ricoh Co., Ltd. | Method and system for multi-tier image matching in a mixed media environment |
US8600989B2 (en) | 2004-10-01 | 2013-12-03 | Ricoh Co., Ltd. | Method and system for image matching in a mixed media environment |
US8600196B2 (en) | 2006-09-08 | 2013-12-03 | Google Inc. | Optical scanners, such as hand-held optical scanners |
US8620083B2 (en) | 2004-12-03 | 2013-12-31 | Google Inc. | Method and system for character recognition |
US8627196B1 (en) * | 2005-03-30 | 2014-01-07 | Amazon Technologies, Inc. | Recognizing an electronically-executable instruction |
US8676810B2 (en) | 2006-07-31 | 2014-03-18 | Ricoh Co., Ltd. | Multiple index mixed media reality recognition using unequal priority indexes |
US8781228B2 (en) | 2004-04-01 | 2014-07-15 | Google Inc. | Triggering actions in response to optically or acoustically capturing keywords from a rendered document |
US8825682B2 (en) | 2006-07-31 | 2014-09-02 | Ricoh Co., Ltd. | Architecture for mixed media reality retrieval of locations and registration of images |
US8838591B2 (en) | 2005-08-23 | 2014-09-16 | Ricoh Co., Ltd. | Embedding hot spots in electronic documents |
US8856108B2 (en) | 2006-07-31 | 2014-10-07 | Ricoh Co., Ltd. | Combining results of image retrieval processes |
US8868555B2 (en) | 2006-07-31 | 2014-10-21 | Ricoh Co., Ltd. | Computation of a recongnizability score (quality predictor) for image retrieval |
US8874504B2 (en) | 2004-12-03 | 2014-10-28 | Google Inc. | Processing techniques for visual capture data from a rendered document |
US8892495B2 (en) | 1991-12-23 | 2014-11-18 | Blanding Hovenweep, Llc | Adaptive pattern recognition based controller apparatus and method and human-interface therefore |
US8949287B2 (en) | 2005-08-23 | 2015-02-03 | Ricoh Co., Ltd. | Embedding hot spots in imaged documents |
US8990235B2 (en) | 2009-03-12 | 2015-03-24 | Google Inc. | Automatically providing content associated with captured information, such as information captured in real-time |
US9020966B2 (en) | 2006-07-31 | 2015-04-28 | Ricoh Co., Ltd. | Client device for interacting with a mixed media reality recognition system |
US9058331B2 (en) | 2011-07-27 | 2015-06-16 | Ricoh Co., Ltd. | Generating a conversation in a social network based on visual search results |
US9063953B2 (en) | 2004-10-01 | 2015-06-23 | Ricoh Co., Ltd. | System and methods for creation and use of a mixed media environment |
US9063952B2 (en) | 2006-07-31 | 2015-06-23 | Ricoh Co., Ltd. | Mixed media reality recognition with image tracking |
US9081799B2 (en) | 2009-12-04 | 2015-07-14 | Google Inc. | Using gestalt information to identify locations in printed information |
US9116890B2 (en) | 2004-04-01 | 2015-08-25 | Google Inc. | Triggering actions in response to optically or acoustically capturing keywords from a rendered document |
US20150242096A1 (en) * | 2003-04-18 | 2015-08-27 | International Business Machines Corporation | Enabling a visually impaired or blind person to have access to information printed on a physical document |
US9143638B2 (en) | 2004-04-01 | 2015-09-22 | Google Inc. | Data capture from rendered documents using handheld device |
US9171202B2 (en) | 2005-08-23 | 2015-10-27 | Ricoh Co., Ltd. | Data organization and access for mixed media document system |
US9176984B2 (en) | 2006-07-31 | 2015-11-03 | Ricoh Co., Ltd | Mixed media reality retrieval of differentially-weighted links |
US9268852B2 (en) | 2004-02-15 | 2016-02-23 | Google Inc. | Search engines and systems with handheld document data capture devices |
US9292617B2 (en) | 2013-03-14 | 2016-03-22 | Rohit Chandra | Method and apparatus for enabling content portion selection services for visitors to web pages |
US9323784B2 (en) | 2009-12-09 | 2016-04-26 | Google Inc. | Image search using text-based elements within the contents of images |
US9373029B2 (en) | 2007-07-11 | 2016-06-21 | Ricoh Co., Ltd. | Invisible junction feature recognition for document security or annotation |
US9384619B2 (en) | 2006-07-31 | 2016-07-05 | Ricoh Co., Ltd. | Searching media content for objects specified using identifiers |
US9405751B2 (en) | 2005-08-23 | 2016-08-02 | Ricoh Co., Ltd. | Database for mixed media document system |
US9530050B1 (en) * | 2007-07-11 | 2016-12-27 | Ricoh Co., Ltd. | Document annotation sharing |
US9535563B2 (en) | 1999-02-01 | 2017-01-03 | Blanding Hovenweep, Llc | Internet appliance system and method |
CN106354730A (en) * | 2015-07-16 | 2017-01-25 | 北京国双科技有限公司 | Method and device for recognizing webpage text repeated content in webpage analysis |
US20190034718A1 (en) * | 2017-07-27 | 2019-01-31 | Celant Innovations, LLC | Method and apparatus for analyzing defined terms in a document |
US10289294B2 (en) | 2006-06-22 | 2019-05-14 | Rohit Chandra | Content selection widget for visitors of web pages |
CN110168540A (en) * | 2017-01-09 | 2019-08-23 | 微软技术许可有限责任公司 | Capture annotation on an electronic display |
US10757148B2 (en) | 2018-03-02 | 2020-08-25 | Ricoh Company, Ltd. | Conducting electronic meetings over computer networks using interactive whiteboard appliances and mobile devices |
US10860985B2 (en) | 2016-10-11 | 2020-12-08 | Ricoh Company, Ltd. | Post-meeting processing using artificial intelligence |
US10866713B2 (en) | 2006-06-22 | 2020-12-15 | Rohit Chandra | Highlighting on a personal digital assistant, mobile handset, eBook, or handheld device |
US10884585B2 (en) | 2006-06-22 | 2021-01-05 | Rohit Chandra | User widget displaying portions of content |
US10909197B2 (en) | 2006-06-22 | 2021-02-02 | Rohit Chandra | Curation rank: content portion search |
US11030585B2 (en) | 2017-10-09 | 2021-06-08 | Ricoh Company, Ltd. | Person detection, person identification and meeting start for interactive whiteboard appliances |
US11062271B2 (en) | 2017-10-09 | 2021-07-13 | Ricoh Company, Ltd. | Interactive whiteboard appliances with learning capabilities |
US11080466B2 (en) | 2019-03-15 | 2021-08-03 | Ricoh Company, Ltd. | Updating existing content suggestion to include suggestions from recorded media using artificial intelligence |
US11120342B2 (en) | 2015-11-10 | 2021-09-14 | Ricoh Company, Ltd. | Electronic meeting intelligence |
US11263384B2 (en) | 2019-03-15 | 2022-03-01 | Ricoh Company, Ltd. | Generating document edit requests for electronic documents managed by a third-party document management service using artificial intelligence |
US11270060B2 (en) | 2019-03-15 | 2022-03-08 | Ricoh Company, Ltd. | Generating suggested document edits from recorded media using artificial intelligence |
US11288686B2 (en) | 2006-06-22 | 2022-03-29 | Rohit Chandra | Identifying micro users interests: at a finer level of granularity |
US11301532B2 (en) | 2006-06-22 | 2022-04-12 | Rohit Chandra | Searching for user selected portions of content |
US11307735B2 (en) | 2016-10-11 | 2022-04-19 | Ricoh Company, Ltd. | Creating agendas for electronic meetings using artificial intelligence |
US11392754B2 (en) * | 2019-03-15 | 2022-07-19 | Ricoh Company, Ltd. | Artificial intelligence assisted review of physical documents |
US11429685B2 (en) | 2006-06-22 | 2022-08-30 | Rohit Chandra | Sharing only a part of a web page—the part selected by a user |
US11573993B2 (en) | 2019-03-15 | 2023-02-07 | Ricoh Company, Ltd. | Generating a meeting review document that includes links to the one or more documents reviewed |
US11720741B2 (en) | 2019-03-15 | 2023-08-08 | Ricoh Company, Ltd. | Artificial intelligence assisted review of electronic documents |
US11763344B2 (en) | 2006-06-22 | 2023-09-19 | Rohit Chandra | SaaS for content curation without a browser add-on |
US11853374B2 (en) | 2006-06-22 | 2023-12-26 | Rohit Chandra | Directly, automatically embedding a content portion |
Families Citing this family (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6671684B1 (en) * | 2000-09-01 | 2003-12-30 | Ricoh Co., Ltd. | Method and apparatus for simultaneous highlighting of a physical version of a document and an electronic version of a document |
US7712028B2 (en) * | 2001-10-19 | 2010-05-04 | Xerox Corporation | Using annotations for summarizing a document image and itemizing the summary based on similar annotations |
CA2483042C (en) * | 2002-04-22 | 2016-06-14 | Nielsen Media Research, Inc. | Methods and apparatus to collect audience information associated with a media presentation |
US7137077B2 (en) * | 2002-07-30 | 2006-11-14 | Microsoft Corporation | Freeform encounter selection tool |
US7392469B1 (en) | 2003-05-19 | 2008-06-24 | Sidney Bailin | Non-intrusive commentary capture for document authors |
US9251519B1 (en) * | 2003-12-31 | 2016-02-02 | Google Inc. | Systems and methods for monetizing subscription and archival news content |
US9008447B2 (en) | 2004-04-01 | 2015-04-14 | Google Inc. | Method and system for character recognition |
US8081849B2 (en) | 2004-12-03 | 2011-12-20 | Google Inc. | Portable scanning and memory device |
US9799060B2 (en) * | 2004-04-01 | 2017-10-24 | Google Inc. | Content access with handheld document data capture devices |
US8713418B2 (en) | 2004-04-12 | 2014-04-29 | Google Inc. | Adding value to a rendered document |
JP4630777B2 (en) * | 2004-09-13 | 2011-02-09 | キヤノン株式会社 | Method, apparatus, computer program and storage medium for changing digital document |
AU2005209670B2 (en) * | 2004-09-13 | 2008-10-30 | Canon Kabushiki Kaisha | Modifying digital documents |
KR101212929B1 (en) * | 2004-09-27 | 2012-12-14 | 구글 인코포레이티드 | Secure data gathering from rendered documents |
JP4448051B2 (en) * | 2005-04-19 | 2010-04-07 | キヤノン株式会社 | Image reading apparatus and method |
US8584005B1 (en) * | 2006-06-28 | 2013-11-12 | Adobe Systems Incorporated | Previewing redaction content in a document |
US9805010B2 (en) | 2006-06-28 | 2017-10-31 | Adobe Systems Incorporated | Methods and apparatus for redacting related content in a document |
US8179556B2 (en) * | 2007-03-26 | 2012-05-15 | Xerox Corporation | Masking of text in document reproduction |
US10628519B2 (en) | 2016-07-22 | 2020-04-21 | Dropbox, Inc. | Enhancing documents portrayed in digital images |
WO2021226710A1 (en) * | 2020-05-12 | 2021-11-18 | Applied Publishing Concepts Inc. | System and method for associating online content with offline content |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US34476A (en) * | 1862-02-25 | Improvement in hose-couplings | ||
US5243149A (en) * | 1992-04-10 | 1993-09-07 | International Business Machines Corp. | Method and apparatus for improving the paper interface to computing systems |
US5659164A (en) * | 1992-11-05 | 1997-08-19 | Schmid; Edward | Method of and system for apparatus for two-way automatically creating, identifying, routing and storing digitally scanned documents |
US5692073A (en) * | 1996-05-03 | 1997-11-25 | Xerox Corporation | Formless forms and paper web using a reference-based mark extraction technique |
US5832474A (en) * | 1996-02-26 | 1998-11-03 | Matsushita Electric Industrial Co., Ltd. | Document search and retrieval system with partial match searching of user-drawn annotations |
US6369811B1 (en) * | 1998-09-09 | 2002-04-09 | Ricoh Company Limited | Automatic adaptive document help for paper documents |
US6665841B1 (en) * | 1997-11-14 | 2003-12-16 | Xerox Corporation | Transmission of subsets of layout objects at different resolutions |
US6671684B1 (en) * | 2000-09-01 | 2003-12-30 | Ricoh Co., Ltd. | Method and apparatus for simultaneous highlighting of a physical version of a document and an electronic version of a document |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE68928276T2 (en) | 1988-05-27 | 1998-01-15 | Kodak Ltd | DOCUMENT RECORDING AND PROCESSING IN A DATA PROCESSING SYSTEM |
USRE34476E (en) * | 1990-05-14 | 1993-12-14 | Norwood Donald D | Hybrid information management system for handwriting and text |
US5920694A (en) | 1993-03-19 | 1999-07-06 | Ncr Corporation | Annotation of computer video displays |
US5583980A (en) | 1993-12-22 | 1996-12-10 | Knowledge Media Inc. | Time-synchronized annotation method |
US5768607A (en) | 1994-09-30 | 1998-06-16 | Intel Corporation | Method and apparatus for freehand annotation and drawings incorporating sound and for compressing and synchronizing sound |
US5970455A (en) | 1997-03-20 | 1999-10-19 | Xerox Corporation | System for capturing and retrieving audio data and corresponding hand-written notes |
-
2000
- 2000-09-01 US US09/653,965 patent/US6671684B1/en not_active Expired - Lifetime
-
2003
- 2003-03-26 US US10/400,977 patent/US20030187886A1/en not_active Abandoned
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US34476A (en) * | 1862-02-25 | Improvement in hose-couplings | ||
US5243149A (en) * | 1992-04-10 | 1993-09-07 | International Business Machines Corp. | Method and apparatus for improving the paper interface to computing systems |
US5659164A (en) * | 1992-11-05 | 1997-08-19 | Schmid; Edward | Method of and system for apparatus for two-way automatically creating, identifying, routing and storing digitally scanned documents |
US5832474A (en) * | 1996-02-26 | 1998-11-03 | Matsushita Electric Industrial Co., Ltd. | Document search and retrieval system with partial match searching of user-drawn annotations |
US5692073A (en) * | 1996-05-03 | 1997-11-25 | Xerox Corporation | Formless forms and paper web using a reference-based mark extraction technique |
US6665841B1 (en) * | 1997-11-14 | 2003-12-16 | Xerox Corporation | Transmission of subsets of layout objects at different resolutions |
US6369811B1 (en) * | 1998-09-09 | 2002-04-09 | Ricoh Company Limited | Automatic adaptive document help for paper documents |
US6671684B1 (en) * | 2000-09-01 | 2003-12-30 | Ricoh Co., Ltd. | Method and apparatus for simultaneous highlighting of a physical version of a document and an electronic version of a document |
Cited By (133)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8892495B2 (en) | 1991-12-23 | 2014-11-18 | Blanding Hovenweep, Llc | Adaptive pattern recognition based controller apparatus and method and human-interface therefore |
US9535563B2 (en) | 1999-02-01 | 2017-01-03 | Blanding Hovenweep, Llc | Internet appliance system and method |
US20080027933A1 (en) * | 1999-10-20 | 2008-01-31 | Araha, Inc. | System and method for location, understanding and assimilation of digital documents through abstract indicia |
US20030050927A1 (en) * | 2001-09-07 | 2003-03-13 | Araha, Inc. | System and method for location, understanding and assimilation of digital documents through abstract indicia |
US20150242096A1 (en) * | 2003-04-18 | 2015-08-27 | International Business Machines Corporation | Enabling a visually impaired or blind person to have access to information printed on a physical document |
US10276065B2 (en) * | 2003-04-18 | 2019-04-30 | International Business Machines Corporation | Enabling a visually impaired or blind person to have access to information printed on a physical document |
US7747428B1 (en) * | 2003-09-24 | 2010-06-29 | Yahoo! Inc. | Visibly distinguishing portions of compound words |
US7702624B2 (en) * | 2004-02-15 | 2010-04-20 | Exbiblio, B.V. | Processing techniques for visual capture data from a rendered document |
US8214387B2 (en) | 2004-02-15 | 2012-07-03 | Google Inc. | Document enhancement system and method |
EP1756704A4 (en) * | 2004-02-15 | 2009-04-29 | Exbiblio Bv | Publishing techniques for adding value to a rendered document |
EP1759276A4 (en) * | 2004-02-15 | 2009-04-29 | Exbiblio Bv | Establishing an interactive environment for rendered documents |
EP1759281A4 (en) * | 2004-02-15 | 2009-04-29 | Exbiblio Bv | Adding information or functionality to a rendered document via association with an electronic counterpart |
US9268852B2 (en) | 2004-02-15 | 2016-02-23 | Google Inc. | Search engines and systems with handheld document data capture devices |
US7706611B2 (en) | 2004-02-15 | 2010-04-27 | Exbiblio B.V. | Method and system for character recognition |
US7707039B2 (en) | 2004-02-15 | 2010-04-27 | Exbiblio B.V. | Automatic modification of web pages |
US7742953B2 (en) | 2004-02-15 | 2010-06-22 | Exbiblio B.V. | Adding information or functionality to a rendered document via association with an electronic counterpart |
US8831365B2 (en) | 2004-02-15 | 2014-09-09 | Google Inc. | Capturing text from rendered documents using supplement information |
US8515816B2 (en) | 2004-02-15 | 2013-08-20 | Google Inc. | Aggregate analysis of text captures performed by multiple users from rendered documents |
US7818215B2 (en) | 2004-02-15 | 2010-10-19 | Exbiblio, B.V. | Processing techniques for text capture from a rendered document |
US7831912B2 (en) | 2004-02-15 | 2010-11-09 | Exbiblio B. V. | Publishing techniques for adding value to a rendered document |
US8442331B2 (en) | 2004-02-15 | 2013-05-14 | Google Inc. | Capturing text from rendered documents using supplemental information |
US8447144B2 (en) * | 2004-02-15 | 2013-05-21 | Google Inc. | Data capture from rendered documents using handheld device |
US8005720B2 (en) | 2004-02-15 | 2011-08-23 | Google Inc. | Applying scanned information to identify content |
US8019648B2 (en) | 2004-02-15 | 2011-09-13 | Google Inc. | Search engines and systems with handheld document data capture devices |
US9143638B2 (en) | 2004-04-01 | 2015-09-22 | Google Inc. | Data capture from rendered documents using handheld device |
US9514134B2 (en) | 2004-04-01 | 2016-12-06 | Google Inc. | Triggering actions in response to optically or acoustically capturing keywords from a rendered document |
US9116890B2 (en) | 2004-04-01 | 2015-08-25 | Google Inc. | Triggering actions in response to optically or acoustically capturing keywords from a rendered document |
US8505090B2 (en) | 2004-04-01 | 2013-08-06 | Google Inc. | Archive of text captures from rendered documents |
US9633013B2 (en) | 2004-04-01 | 2017-04-25 | Google Inc. | Triggering actions in response to optically or acoustically capturing keywords from a rendered document |
US8781228B2 (en) | 2004-04-01 | 2014-07-15 | Google Inc. | Triggering actions in response to optically or acoustically capturing keywords from a rendered document |
US7812860B2 (en) | 2004-04-01 | 2010-10-12 | Exbiblio B.V. | Handheld device for capturing text from both a document printed on paper and a document displayed on a dynamic display device |
US9030699B2 (en) | 2004-04-19 | 2015-05-12 | Google Inc. | Association of a portable scanner with input/output and storage devices |
US8261094B2 (en) | 2004-04-19 | 2012-09-04 | Google Inc. | Secure data gathering from rendered documents |
US8489624B2 (en) | 2004-05-17 | 2013-07-16 | Google, Inc. | Processing techniques for text capture from a rendered document |
US8799099B2 (en) | 2004-05-17 | 2014-08-05 | Google Inc. | Processing techniques for text capture from a rendered document |
US9275051B2 (en) | 2004-07-19 | 2016-03-01 | Google Inc. | Automatic modification of web pages |
US8346620B2 (en) | 2004-07-19 | 2013-01-01 | Google Inc. | Automatic modification of web pages |
WO2006023937A2 (en) * | 2004-08-23 | 2006-03-02 | Exbiblio B.V. | A portable scanning device |
US8179563B2 (en) | 2004-08-23 | 2012-05-15 | Google Inc. | Portable scanning device |
WO2006023937A3 (en) * | 2004-08-23 | 2008-06-26 | Exbiblio Bv | A portable scanning device |
US20060085477A1 (en) * | 2004-10-01 | 2006-04-20 | Ricoh Company, Ltd. | Techniques for retrieving documents using an image capture device |
US8335789B2 (en) | 2004-10-01 | 2012-12-18 | Ricoh Co., Ltd. | Method and system for document fingerprint matching in a mixed media environment |
US8332401B2 (en) | 2004-10-01 | 2012-12-11 | Ricoh Co., Ltd | Method and system for position-based image matching in a mixed media environment |
US8600989B2 (en) | 2004-10-01 | 2013-12-03 | Ricoh Co., Ltd. | Method and system for image matching in a mixed media environment |
US8521737B2 (en) | 2004-10-01 | 2013-08-27 | Ricoh Co., Ltd. | Method and system for multi-tier image matching in a mixed media environment |
US8489583B2 (en) * | 2004-10-01 | 2013-07-16 | Ricoh Company, Ltd. | Techniques for retrieving documents using an image capture device |
US9063953B2 (en) | 2004-10-01 | 2015-06-23 | Ricoh Co., Ltd. | System and methods for creation and use of a mixed media environment |
US8620083B2 (en) | 2004-12-03 | 2013-12-31 | Google Inc. | Method and system for character recognition |
US8953886B2 (en) | 2004-12-03 | 2015-02-10 | Google Inc. | Method and system for character recognition |
US8874504B2 (en) | 2004-12-03 | 2014-10-28 | Google Inc. | Processing techniques for visual capture data from a rendered document |
US7990556B2 (en) | 2004-12-03 | 2011-08-02 | Google Inc. | Association of a portable scanner with input/output and storage devices |
US8627196B1 (en) * | 2005-03-30 | 2014-01-07 | Amazon Technologies, Inc. | Recognizing an electronically-executable instruction |
US8195659B2 (en) | 2005-08-23 | 2012-06-05 | Ricoh Co. Ltd. | Integration and use of mixed media documents |
US9405751B2 (en) | 2005-08-23 | 2016-08-02 | Ricoh Co., Ltd. | Database for mixed media document system |
US8949287B2 (en) | 2005-08-23 | 2015-02-03 | Ricoh Co., Ltd. | Embedding hot spots in imaged documents |
US8156427B2 (en) | 2005-08-23 | 2012-04-10 | Ricoh Co. Ltd. | User interface for mixed media reality |
US9171202B2 (en) | 2005-08-23 | 2015-10-27 | Ricoh Co., Ltd. | Data organization and access for mixed media document system |
US20110081892A1 (en) * | 2005-08-23 | 2011-04-07 | Ricoh Co., Ltd. | System and methods for use of voice mail and email in a mixed media environment |
US8838591B2 (en) | 2005-08-23 | 2014-09-16 | Ricoh Co., Ltd. | Embedding hot spots in electronic documents |
US10866713B2 (en) | 2006-06-22 | 2020-12-15 | Rohit Chandra | Highlighting on a personal digital assistant, mobile handset, eBook, or handheld device |
US11301532B2 (en) | 2006-06-22 | 2022-04-12 | Rohit Chandra | Searching for user selected portions of content |
US10909197B2 (en) | 2006-06-22 | 2021-02-02 | Rohit Chandra | Curation rank: content portion search |
US10884585B2 (en) | 2006-06-22 | 2021-01-05 | Rohit Chandra | User widget displaying portions of content |
US8910060B2 (en) | 2006-06-22 | 2014-12-09 | Rohit Chandra | Method and apparatus for highlighting a portion of an internet document for collaboration and subsequent retrieval |
US11853374B2 (en) | 2006-06-22 | 2023-12-26 | Rohit Chandra | Directly, automatically embedding a content portion |
US10289294B2 (en) | 2006-06-22 | 2019-05-14 | Rohit Chandra | Content selection widget for visitors of web pages |
US11763344B2 (en) | 2006-06-22 | 2023-09-19 | Rohit Chandra | SaaS for content curation without a browser add-on |
US11748425B2 (en) | 2006-06-22 | 2023-09-05 | Rohit Chandra | Highlighting content portions of search results without a client add-on |
US20080016091A1 (en) * | 2006-06-22 | 2008-01-17 | Rohit Chandra | Method and apparatus for highlighting a portion of an internet document for collaboration and subsequent retrieval |
US11288686B2 (en) | 2006-06-22 | 2022-03-29 | Rohit Chandra | Identifying micro users interests: at a finer level of granularity |
US11429685B2 (en) | 2006-06-22 | 2022-08-30 | Rohit Chandra | Sharing only a part of a web page—the part selected by a user |
US20080005101A1 (en) * | 2006-06-23 | 2008-01-03 | Rohit Chandra | Method and apparatus for determining the significance and relevance of a web page, or a portion thereof |
US8661031B2 (en) * | 2006-06-23 | 2014-02-25 | Rohit Chandra | Method and apparatus for determining the significance and relevance of a web page, or a portion thereof |
US8418048B2 (en) * | 2006-06-27 | 2013-04-09 | Fuji Xerox Co., Ltd. | Document processing system, document processing method, computer readable medium and data signal |
US20070296695A1 (en) * | 2006-06-27 | 2007-12-27 | Fuji Xerox Co., Ltd. | Document processing system, document processing method, computer readable medium and data signal |
US8073263B2 (en) | 2006-07-31 | 2011-12-06 | Ricoh Co., Ltd. | Multi-classifier selection and monitoring for MMR-based image recognition |
US9063952B2 (en) | 2006-07-31 | 2015-06-23 | Ricoh Co., Ltd. | Mixed media reality recognition with image tracking |
US8156116B2 (en) | 2006-07-31 | 2012-04-10 | Ricoh Co., Ltd | Dynamic presentation of targeted information in a mixed media reality recognition system |
US8510283B2 (en) | 2006-07-31 | 2013-08-13 | Ricoh Co., Ltd. | Automatic adaption of an image recognition system to image capture devices |
US8489987B2 (en) | 2006-07-31 | 2013-07-16 | Ricoh Co., Ltd. | Monitoring and analyzing creation and usage of visual content using image and hotspot interaction |
US9020966B2 (en) | 2006-07-31 | 2015-04-28 | Ricoh Co., Ltd. | Client device for interacting with a mixed media reality recognition system |
US8369655B2 (en) | 2006-07-31 | 2013-02-05 | Ricoh Co., Ltd. | Mixed media reality recognition using multiple specialized indexes |
US9176984B2 (en) | 2006-07-31 | 2015-11-03 | Ricoh Co., Ltd | Mixed media reality retrieval of differentially-weighted links |
US8676810B2 (en) | 2006-07-31 | 2014-03-18 | Ricoh Co., Ltd. | Multiple index mixed media reality recognition using unequal priority indexes |
US8856108B2 (en) | 2006-07-31 | 2014-10-07 | Ricoh Co., Ltd. | Combining results of image retrieval processes |
US8868555B2 (en) | 2006-07-31 | 2014-10-21 | Ricoh Co., Ltd. | Computation of a recongnizability score (quality predictor) for image retrieval |
US8825682B2 (en) | 2006-07-31 | 2014-09-02 | Ricoh Co., Ltd. | Architecture for mixed media reality retrieval of locations and registration of images |
US9384619B2 (en) | 2006-07-31 | 2016-07-05 | Ricoh Co., Ltd. | Searching media content for objects specified using identifiers |
US8201076B2 (en) | 2006-07-31 | 2012-06-12 | Ricoh Co., Ltd. | Capturing symbolic information from documents upon printing |
US8600196B2 (en) | 2006-09-08 | 2013-12-03 | Google Inc. | Optical scanners, such as hand-held optical scanners |
US20080118064A1 (en) * | 2006-11-22 | 2008-05-22 | Bhogal Kulvir S | Method, system, and program product for encrypting portions of a document using selective annotation |
US8654973B2 (en) | 2006-11-22 | 2014-02-18 | International Business Machines Corporation | Method, system, and program product for encrypting portions of a document using selective annotation |
US8238609B2 (en) | 2007-01-18 | 2012-08-07 | Ricoh Co., Ltd. | Synthetic image and video generation from ground truth data |
US10192279B1 (en) | 2007-07-11 | 2019-01-29 | Ricoh Co., Ltd. | Indexed document modification sharing with mixed media reality |
US9530050B1 (en) * | 2007-07-11 | 2016-12-27 | Ricoh Co., Ltd. | Document annotation sharing |
US8156115B1 (en) | 2007-07-11 | 2012-04-10 | Ricoh Co. Ltd. | Document-based networking with mixed media reality |
US9373029B2 (en) | 2007-07-11 | 2016-06-21 | Ricoh Co., Ltd. | Invisible junction feature recognition for document security or annotation |
US8086038B2 (en) | 2007-07-11 | 2011-12-27 | Ricoh Co., Ltd. | Invisible junction features for patch recognition |
US8989431B1 (en) | 2007-07-11 | 2015-03-24 | Ricoh Co., Ltd. | Ad hoc paper-based networking with mixed media reality |
US8144921B2 (en) | 2007-07-11 | 2012-03-27 | Ricoh Co., Ltd. | Information retrieval using invisible junctions and geometric constraints |
US8184155B2 (en) | 2007-07-11 | 2012-05-22 | Ricoh Co. Ltd. | Recognition and tracking using invisible junctions |
US8276088B2 (en) | 2007-07-11 | 2012-09-25 | Ricoh Co., Ltd. | User interface for three-dimensional navigation |
US8176054B2 (en) | 2007-07-12 | 2012-05-08 | Ricoh Co. Ltd | Retrieving electronic documents by converting them to synthetic text |
US8385589B2 (en) | 2008-05-15 | 2013-02-26 | Berna Erol | Web-based content detection in images, extraction and recognition |
US8418055B2 (en) | 2009-02-18 | 2013-04-09 | Google Inc. | Identifying a document by performing spectral analysis on the contents of the document |
US8638363B2 (en) | 2009-02-18 | 2014-01-28 | Google Inc. | Automatically capturing information, such as capturing information using a document-aware device |
US9075779B2 (en) | 2009-03-12 | 2015-07-07 | Google Inc. | Performing actions based on capturing information from rendered documents, such as documents under copyright |
US8990235B2 (en) | 2009-03-12 | 2015-03-24 | Google Inc. | Automatically providing content associated with captured information, such as information captured in real-time |
US8447066B2 (en) | 2009-03-12 | 2013-05-21 | Google Inc. | Performing actions based on capturing information from rendered documents, such as documents under copyright |
US8385660B2 (en) | 2009-06-24 | 2013-02-26 | Ricoh Co., Ltd. | Mixed media reality indexing and retrieval for repeated content |
US9081799B2 (en) | 2009-12-04 | 2015-07-14 | Google Inc. | Using gestalt information to identify locations in printed information |
US9323784B2 (en) | 2009-12-09 | 2016-04-26 | Google Inc. | Image search using text-based elements within the contents of images |
US20120042288A1 (en) * | 2010-08-16 | 2012-02-16 | Fuji Xerox Co., Ltd. | Systems and methods for interactions with documents across paper and computers |
US9058331B2 (en) | 2011-07-27 | 2015-06-16 | Ricoh Co., Ltd. | Generating a conversation in a social network based on visual search results |
US9292617B2 (en) | 2013-03-14 | 2016-03-22 | Rohit Chandra | Method and apparatus for enabling content portion selection services for visitors to web pages |
CN106354730A (en) * | 2015-07-16 | 2017-01-25 | 北京国双科技有限公司 | Method and device for recognizing webpage text repeated content in webpage analysis |
US11983637B2 (en) | 2015-11-10 | 2024-05-14 | Ricoh Company, Ltd. | Electronic meeting intelligence |
US11120342B2 (en) | 2015-11-10 | 2021-09-14 | Ricoh Company, Ltd. | Electronic meeting intelligence |
US11307735B2 (en) | 2016-10-11 | 2022-04-19 | Ricoh Company, Ltd. | Creating agendas for electronic meetings using artificial intelligence |
US10860985B2 (en) | 2016-10-11 | 2020-12-08 | Ricoh Company, Ltd. | Post-meeting processing using artificial intelligence |
CN110168540A (en) * | 2017-01-09 | 2019-08-23 | 微软技术许可有限责任公司 | Capture annotation on an electronic display |
US10713482B2 (en) * | 2017-07-27 | 2020-07-14 | Celant Innovations, LLC | Method and apparatus for analyzing defined terms in a document |
US20190034718A1 (en) * | 2017-07-27 | 2019-01-31 | Celant Innovations, LLC | Method and apparatus for analyzing defined terms in a document |
US11030585B2 (en) | 2017-10-09 | 2021-06-08 | Ricoh Company, Ltd. | Person detection, person identification and meeting start for interactive whiteboard appliances |
US11645630B2 (en) | 2017-10-09 | 2023-05-09 | Ricoh Company, Ltd. | Person detection, person identification and meeting start for interactive whiteboard appliances |
US11062271B2 (en) | 2017-10-09 | 2021-07-13 | Ricoh Company, Ltd. | Interactive whiteboard appliances with learning capabilities |
US10757148B2 (en) | 2018-03-02 | 2020-08-25 | Ricoh Company, Ltd. | Conducting electronic meetings over computer networks using interactive whiteboard appliances and mobile devices |
US11392754B2 (en) * | 2019-03-15 | 2022-07-19 | Ricoh Company, Ltd. | Artificial intelligence assisted review of physical documents |
US11270060B2 (en) | 2019-03-15 | 2022-03-08 | Ricoh Company, Ltd. | Generating suggested document edits from recorded media using artificial intelligence |
US11573993B2 (en) | 2019-03-15 | 2023-02-07 | Ricoh Company, Ltd. | Generating a meeting review document that includes links to the one or more documents reviewed |
US11720741B2 (en) | 2019-03-15 | 2023-08-08 | Ricoh Company, Ltd. | Artificial intelligence assisted review of electronic documents |
US11263384B2 (en) | 2019-03-15 | 2022-03-01 | Ricoh Company, Ltd. | Generating document edit requests for electronic documents managed by a third-party document management service using artificial intelligence |
US11080466B2 (en) | 2019-03-15 | 2021-08-03 | Ricoh Company, Ltd. | Updating existing content suggestion to include suggestions from recorded media using artificial intelligence |
Also Published As
Publication number | Publication date |
---|---|
US6671684B1 (en) | 2003-12-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6671684B1 (en) | Method and apparatus for simultaneous highlighting of a physical version of a document and an electronic version of a document | |
US10769431B2 (en) | Handheld device for capturing text from both a document printed on paper and a document displayed on a dynamic display device | |
US6665490B2 (en) | Obtaining and using data associating annotating activities with portions of recordings | |
CN1332341C (en) | Information processing apparatus, method, storage medium and program | |
JP4366108B2 (en) | Document search apparatus, document search method, and computer program | |
US7991778B2 (en) | Triggering actions with captured input in a mixed media environment | |
US7672543B2 (en) | Triggering applications based on a captured text in a mixed media environment | |
US7639387B2 (en) | Authoring tools using a mixed media environment | |
US8179563B2 (en) | Portable scanning device | |
US9357098B2 (en) | System and methods for use of voice mail and email in a mixed media environment | |
US7812860B2 (en) | Handheld device for capturing text from both a document printed on paper and a document displayed on a dynamic display device | |
US6178417B1 (en) | Method and means of matching documents based on text genre | |
JP5529082B2 (en) | Acquiring data from rendered documents using handheld devices | |
US20070050341A1 (en) | Triggering applications for distributed action execution and use of mixed media recognition as a control input | |
US20110035656A1 (en) | Identifying a document by performing spectral analysis on the contents of the document | |
JP2006085733A (en) | Filing/retrieval device and filing/retrieval method | |
JPH0773190A (en) | Pictograph naming for pen base computer system | |
JP4785655B2 (en) | Document processing apparatus and document processing method | |
WO2006023937A2 (en) | A portable scanning device | |
JP2010536188A6 (en) | Acquiring data from rendered documents using handheld devices | |
US11755659B2 (en) | Document search device, document search program, and document search method | |
WO1997004409A1 (en) | File searching device | |
JP2001094711A (en) | Document image processing unit and document image processing method | |
JP7105500B2 (en) | Computer-implemented Automatic Acquisition Method for Element Nouns in Chinese Patent Documents for Patent Documents Without Intercharacter Spaces | |
JP4261831B2 (en) | Character recognition processing method, character recognition processing device, character recognition program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |