WO2018105373A1 - Information processing device, information processing method, and information processing system - Google Patents

Information processing device, information processing method, and information processing system Download PDF

Info

Publication number
WO2018105373A1
WO2018105373A1 PCT/JP2017/041758 JP2017041758W WO2018105373A1 WO 2018105373 A1 WO2018105373 A1 WO 2018105373A1 JP 2017041758 W JP2017041758 W JP 2017041758W WO 2018105373 A1 WO2018105373 A1 WO 2018105373A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
display
information
text
text information
Prior art date
Application number
PCT/JP2017/041758
Other languages
French (fr)
Japanese (ja)
Inventor
祐平 滝
真一 河野
邦世 大石
徹哉 浅山
Original Assignee
ソニー株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ソニー株式会社 filed Critical ソニー株式会社
Priority to JP2018554906A priority Critical patent/JP6950708B2/en
Priority to US16/349,731 priority patent/US11189289B2/en
Priority to KR1020197014972A priority patent/KR20190091265A/en
Priority to DE112017006145.8T priority patent/DE112017006145T5/en
Publication of WO2018105373A1 publication Critical patent/WO2018105373A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue

Definitions

  • the present technology relates to an information processing apparatus, an information processing method, and an information processing system, and more particularly, to an information processing apparatus, an information processing method, and an information processing system that can support natural conversation using voice recognition.
  • This technology has been made in view of these circumstances, and is a technology that supports natural conversation using voice recognition.
  • the information processing apparatus includes the voice acquisition unit that acquires the voice information of the first user input to the voice input device, and the acquired voice in the display device for the second user.
  • a display control unit that controls display of text information corresponding to the information, wherein the display control unit displays the text information on the display device or the input amount of the voice information input from the voice input device Based on at least one of the above, control is performed regarding the display amount of the text information.
  • An information processing method includes: an audio acquisition step of acquiring audio information of a first user input to an audio input device by the information processing device; A display control step for controlling display of text information corresponding to the acquired voice information in a display device for two users, wherein the display control step includes a display amount of the text information on the display device, Alternatively, control regarding the display amount of the text information is performed based on at least one of the input amounts of the speech information input from the speech input device.
  • a program that is one aspect of the present technology includes a voice input device that acquires voice information of a first user, a display control device that controls display of text information corresponding to the acquired voice information, and the display control device A display device that displays the text information for the second user in accordance with the control from the display device, the display control device being input from the display amount of the text information on the display device or the voice input device Control relating to the display amount of the text information is performed based on at least one of the input amounts of the voice information.
  • the input voice information of the first user is acquired, and display of text information corresponding to the acquired voice information on the display device for the second user is controlled.
  • control relating to the display amount of the text information is performed based on at least one of the display amount of the text information on the display device or the input amount of the speech information input from the speech input device.
  • natural conversation using voice recognition can be performed.
  • FIG. 1 shows a first configuration example of a conversation support apparatus according to an embodiment of the present technology, and shows a case where the conversation support apparatus 10 is formed as one casing.
  • the conversation support device 10 is for supporting a conversation between a person who does not have anxiety about hearing (hereinafter referred to as user A) and a person who has anxiety about hearing (hereinafter referred to as user B). .
  • user A a person who does not have anxiety about hearing
  • user B a person who has anxiety about hearing
  • the first user in one aspect of the present technology corresponds to the user A in the present configuration example
  • the second user in one aspect of the present technology corresponds to the user 2 in the present configuration example. May be.
  • the 1st user in one side of this art should just be a user who inputs a voice. That is, the first user (user who inputs voice) is not limited to a single subject (user), and may be a plurality of subjects (users).
  • the second user in one aspect of the present technology may be a user who visually recognizes the displayed utterance text, and is not limited to a single subject, and may be a plurality of subjects.
  • the utterance of user A is converted into text (hereinafter referred to as utterance text) by voice recognition processing, and the utterance text is displayed on the display unit 43 for user B.
  • utterance text text
  • voice recognition processing voice recognition processing
  • the user B can understand the utterance text (character information) corresponding to the utterance (voice information) of the user A.
  • the utterance text displayed on the display unit 43 is displayed until the user B finishes reading or a predetermined time elapses.
  • an image of the user B from an image captured by the imaging unit 41 or an utterance of the user B collected by the sound collecting unit 42 is used. Used.
  • the display unit 43 for the user B is provided with a display unit 22 for the user A (FIG. 2) on the back side, and the display unit 22 has the same display as the display unit 43, that is, the user A
  • the utterance text corresponding to the utterance is displayed. Thereby, the user A can confirm whether or not the user's utterance has been correctly recognized.
  • FIG. 2 is a block diagram illustrating an internal configuration example of the conversation support apparatus according to the embodiment of the present technology.
  • the conversation support device 10 includes a sound collection unit 21, a display unit 22, an operation input unit 23, an information processing unit 30, an imaging unit 41, a sound collection unit 42, a display unit 43, and an operation input unit 44.
  • the sound collection unit 21, the display unit 22, and the operation input unit 23 are provided mainly for the user A.
  • the sound collecting unit 21 collects the voice (utterance) spoken by the user A and supplies the corresponding speech signal to the information processing unit 30.
  • the display unit 22 displays a screen corresponding to the image signal supplied from the information processing unit 30 (for example, an image signal for displaying an utterance text corresponding to the utterance of the user A on the screen).
  • the operation input unit 23 receives various operations from the user A and notifies the information processing unit 30 of operation signals corresponding thereto.
  • the information processing unit 30 converts the speech signal supplied from the sound collection unit 21 into speech text by speech recognition processing. Further, the information processing unit 30 supplies an image signal for displaying the utterance text on the screen to the display unit 43. Details of the information processing unit 30 will be described later.
  • the imaging unit 41, the sound collection unit 42, the display unit 43, and the operation input unit 44 are provided mainly for the user B.
  • the imaging unit 41 images the user B and supplies the moving image signal obtained as a result to the information processing unit 30.
  • the sound collecting unit 42 collects the voice (speech) spoken by the user B and supplies the corresponding speech signal to the information processing unit 30.
  • the display unit 43 displays a screen corresponding to an image signal supplied from the information processing unit 30 for displaying the utterance text corresponding to the utterance of the user A on the screen.
  • the operation input unit 44 receives various operations from the user B and notifies the information processing unit 30 of operation signals corresponding thereto.
  • FIG. 3 shows a configuration example of functional blocks included in the information processing unit 30.
  • the information processing unit 30 includes a voice recognition unit 31, an image recognition unit 32, a misrecognition learning unit 33, an analysis unit 35, an editing unit 36, an additional writing learning unit 37, a display waiting list holding unit 38, a display control unit 39, and a feedback unit. 40.
  • the speech recognition unit 31 generates an utterance text by converting an utterance signal corresponding to the utterance of the user A supplied from the sound collection unit 21 into an utterance text by speech recognition processing, and supplies the utterance text to the analysis unit 35.
  • the speech recognition unit 31 converts an utterance signal corresponding to the utterance of the user B supplied from the sound collection unit 42 into an utterance text by voice recognition processing, and the utterance text represents a specific keyword indicating the user B's already read. (For example, “Yes”, “Yes”, “Okay”, “OK”, “Next”, etc. registered in advance) are detected, and the detection result is supplied to the display control unit 39.
  • the image recognition unit 32 Based on the moving image signal supplied from the imaging unit 41, the image recognition unit 32 performs a specific operation indicating the user B's reading (for example, nodding, watching the screen and then looking in a direction other than the screen). It detects and supplies a detection result to the display control part 39. FIG. Further, the image recognition unit 32 measures the distance between the user B and the display unit 43 based on the moving image signal supplied from the imaging unit 41 and notifies the display control unit 39 of the measurement result. The distance between the user B and the display unit 43 is used to set the character size of the utterance text displayed on the display unit 43. For example, the longer the distance between the user B and the display unit 43, the larger the character size is set.
  • the line-of-sight direction may be determined based on the direction of the wearable device, that is, the direction of the user B's head or body.
  • the direction of the wearable device can be determined based on position information acquired from a camera, an acceleration sensor, a gyro sensor, or the like provided in the wearable device.
  • the Purkinje image of the eyeball of the user B and the pupil center may be determined using an infrared camera and an infrared LED, and the line-of-sight direction of the user B may be determined based on these.
  • the misrecognition learning unit 33 edits input from the user A or the user B to the utterance text corresponding to the utterance of the user A that is the result of the speech recognition process (for example, an erase instruction operation, a recurrent speech instruction operation, NG Corresponding to the word registration instruction operation), misrecognized words included in the utterance text are registered in the misrecognition list 34.
  • the misrecognition learning unit 33 A recognition result (second candidate or the like) other than a misrecognized word (first candidate of recognition result) is requested.
  • the analysis unit 35 analyzes the speech text corresponding to the speech of the user A generated by the speech recognition unit 31, for example, by decomposing the speech text into parts of speech or extracting keywords.
  • the editing unit 36 controls the amount of text that specifies particles or the like that do not impair the meaning of the utterance text even if line breaks or page breaks are added or deleted as appropriate to the utterance text.
  • An edit process such as a process is performed and supplied to the display wait list holding unit 38. In the editing process, it may be considered that at least one of line feed, page break or text amount suppression processing is performed, and at least one of line feed, page break or text amount suppression processing may be omitted.
  • the editing unit 36 can supply a plurality of related utterance texts to a display waiting list holding unit 38 in a thread.
  • an icon corresponding to a thread waiting for display may be displayed while displaying the current thread.
  • Display objects indicating threads waiting to be displayed are not limited to icons, and may be appropriately set. According to such a configuration, it is possible to easily grasp how much the user B has finished reading the other party's utterance text. Moreover, according to such a structure, the user B can act to suppress the input amount of the user A based on the progress of the utterance text.
  • the editing unit 36 outputs a sentence of the utterance text based on the editing operation input by the user A using the operation input unit 23 with respect to the utterance text corresponding to the utterance of the user A displayed on the display unit 22. Controls the process of deleting, inserting utterance text corresponding to recurrent utterances, and registering NG words.
  • the editing unit 36 performs an additional operation (specifically, “?”) That is input by the user A using the operation input unit 23 to the utterance text corresponding to the utterance of the user A displayed on the display unit 22.
  • the operation of adding a symbol such as “?” To the utterance text is controlled based on the operation of adding a symbol such as “(question mark)”. Note that symbols, pictograms, emoticons, etc. other than “?” May be additionally recorded.
  • the editing unit 36 is based on an editing operation or an additional writing operation input by the user B using the operation input unit 44 with respect to the utterance text corresponding to the utterance of the user A displayed on the display unit 43. Editing processing can be performed. In other words, both the user A and the user B can perform an editing operation and an additional writing operation on the displayed utterance text corresponding to the utterance of the user A.
  • the additional writing learning unit 37 learns the additional writing operation input by the user A or the user B, and based on the learning result, even if there is no additional writing operation from the user A or the user B, the same symbol or the like is given to the same utterance text.
  • the editing unit 36 is controlled so as to be additionally written.
  • the display waiting list holding unit 38 displays the edited utterance text including at least one of line feed, page break, and text amount suppression processing (the text amount suppression processing may not be performed depending on the number of characters).
  • the information is registered in the display wait list in the sequence order, that is, in the order in which the user A speaks.
  • the utterance text registered in the display waiting list is read from the display control unit 39, it is deleted from the display waiting list.
  • the display control unit 39 reads out the utterance texts from the display waiting list in chronological order, generates an image signal for displaying the read out utterance text on the screen, and supplies it to the display unit 22 and the display unit 43. Further, the display control unit 39 displays a display amount of the utterance text currently displayed on the display unit 22 and the display unit 43, a detection result of a specific keyword that is supplied from the voice recognition unit 31 and represents the user B already read, The display amount of the utterance text on the display unit 22 and the display unit 43 is controlled based on the detection result of the specific action representing the read of the user B supplied from the image recognition unit 32. Further, the display control unit 39 sets the character size for displaying the utterance text according to the distance between the user B and the display unit 43.
  • the feedback control unit 40 is registered in the utterance speed of the user A, the length of the utterance of the user A, the amount of speech recognition characters per unit time, the amount of utterance text displayed on the display unit 43, and the display waiting list. Corresponding to the amount of utterance text, whether or not the user B has already read, the reading speed of the user B, etc., the utterance speed is increased (or decreased) for the user A who is the utterer by using character display or voice output. To control the feedback to notify the user, to notify the utterance, or to prompt the next utterance.
  • the feedback control unit 40 corresponds to the amount of utterance text displayed on the display unit 43, the amount of utterance text registered in the display waiting list, whether or not the user B has been read, the reading speed of the user B, etc. Then, feedback that prompts the user B to read the utterance text is controlled by using a character display or the like.
  • the above-described functional blocks included in the information processing unit 30 do not have to be housed in the same casing, and may be arranged in a distributed manner. Some or all of these functional blocks may be arranged on a server on the Internet, a so-called cloud network.
  • FIG. 4 illustrates a second configuration example of the conversation support apparatus according to the embodiment of the present technology.
  • the conversation support device 10 is configured as a system including a plurality of different electronic devices.
  • the connection between the plurality of electronic devices constituting the conversation support apparatus 10 may be wired connection or may use predetermined wireless communication (for example, Bluetooth (registered trademark), Wi-Fi (trademark), etc.). Good.
  • the conversation support device 10 includes a smartphone 50 used by the user A and a tablet PC (hereinafter referred to as a tablet) 60 used by the user B.
  • FIG. 5 shows a state in which the constituent elements of the conversation support apparatus 10 shown in FIG. 2 are divided into the smartphone 50 and the tablet PC 60.
  • the smartphone 50 among the components of the conversation support device 10, the sound collection unit 21, the display unit 22, the operation input unit 23, and the information processing unit 30 are realized by the smartphone 50.
  • a microphone, a display, a touch panel, and the like included in the smartphone 50 correspond to the sound collection unit 21 and the operation input unit 23, respectively.
  • An application program executed by the smartphone 50 corresponds to the information processing unit 30.
  • the imaging unit 41, the sound collection unit 42, the display unit 43, and the operation input unit 44 are realized by the tablet 60.
  • the camera, microphone, display, touch panel, and the like included in the tablet 60 correspond to the imaging unit 41, the sound collection unit 42, the display unit 43, and the operation input unit 44, respectively.
  • the speech recognition unit 31 among the functional blocks of the information processing unit 30 is arranged in a server 72 that can be connected via the Internet 71.
  • FIG. 6 illustrates a third configuration example of the conversation support apparatus according to the embodiment of the present technology.
  • the conversation support device 10 is configured as a system including a plurality of electronic devices.
  • the projector 80 that projects the video for displaying the utterance text on the smartphone 50 used by the user A and the position where the user B lying on the bed can see, for example, the wall or ceiling of the room. And a camera 110 arranged on the ceiling or the like.
  • FIG. 7 shows a state in which the constituent elements of the conversation support apparatus 10 shown in FIG. 2 are divided into a smartphone 50, a projector 80, and a camera 110.
  • the sound collection unit 21, the display unit 22, the operation input unit 23, and the information processing unit 30 are realized by the smartphone 50.
  • the imaging unit 41 and the sound collection unit 42 are realized by the camera 110.
  • the image sensor and the microphone included in the camera 110 correspond to the imaging unit 41 and the sound collecting unit 42, respectively.
  • the display unit 43 and the operation input unit 44 are realized by the projector 80.
  • the projection unit and the remote controller included in the projector 80 correspond to the display unit 43 and the operation input unit 44, respectively.
  • the voice recognition unit 31 among the functional blocks included in the information processing unit 30 is arranged in a server 72 that can be connected via the Internet 71.
  • FIG. 8 illustrates a fourth configuration example of the conversation support apparatus according to the embodiment of the present technology.
  • the conversation support apparatus 10 is configured as a system including a plurality of different electronic devices.
  • the fourth configuration example includes a neck microphone 100 used by the user A, a television receiver (hereinafter referred to as TV) 90 disposed at a position where the user A and the user B can see,
  • the camera 110 is mounted on the TV 90.
  • FIG. 9 shows a state in which the components of the conversation support device 10 shown in FIG. 2 are divided into a neck microphone 100, a TV 90, and a camera 110.
  • the sound collection unit 21 is realized by the neck microphone 100.
  • the neck microphone 100 may be provided with a speaker that outputs sound in addition to the sound collecting unit 21.
  • the imaging unit 41 and the sound collection unit 42 are realized by the camera 110.
  • the display unit 43 and the operation input unit 44 are realized by the TV 90.
  • the display and remote controller included in the TV 90 correspond to the display unit 43 and the operation input unit 44, respectively. It is assumed that the display and the remote controller included in the TV 90 also serve as the display unit 22 and the operation input unit 23 for the user A.
  • the voice recognition unit 31 among the functional blocks of the information processing unit 30 is arranged in a server 72 that can be connected via the Internet 71.
  • the conversation support device 10 can be configured as one electronic device, or can be configured as a system in which a plurality of electronic devices are combined.
  • the first to fourth configuration examples described above can be combined as appropriate.
  • a wearable device such as a clock-type terminal or a head-mounted display, a monitor for a PC (personal computer), or the like can be employed in addition to the above-described example.
  • FIG. 10 is a flowchart for explaining display wait list generation processing by the conversation support apparatus 10. This display waiting list generation process is repeatedly executed until the power is turned off after the conversation support device 10 is activated.
  • step S1 when the user A speaks, the sound is acquired by the sound collecting unit 21.
  • the sound collecting unit 21 converts the voice of the user A into an utterance signal and supplies it to the information processing unit 30.
  • step S ⁇ b> 2 in the information processing unit 30, the speech recognition unit 31 converts the speech signal corresponding to the speech of the user A into speech text by performing speech recognition processing.
  • step S3 the analysis unit 35 analyzes the utterance text corresponding to the utterance of the user A generated by the voice recognition unit 31.
  • step S4 the editing unit 36 performs an editing process including at least one of a line feed, a page break, or a text amount suppression process on the utterance text corresponding to the utterance of the user A based on the analysis result, The processed utterance text is supplied to the display waiting list holding unit 38.
  • step S5 the display waiting list holding unit 38 holds the edited utterance texts supplied from the editing unit 36 in chronological order. Thereafter, the process returns to step S1, and the subsequent steps are repeated.
  • FIG. 11 is a flowchart for explaining utterance text display processing by the conversation support apparatus 10.
  • the utterance text display process is repeatedly executed in parallel with the display wait list generation process described above until the power is turned off after the conversation support device 10 is activated.
  • step S11 the display control unit 39 determines whether or not the utterance text is currently displayed on the screens of the display units 22 and 43. If it is determined that it is displayed, the process proceeds to step S12. In step S12, the display control unit 39 determines whether or not a predetermined shortest display time has elapsed since the display of the currently displayed utterance text has started, and the shortest display time has elapsed. Wait until. If the shortest display time has elapsed, the process proceeds to step S13.
  • step S ⁇ b> 13 the display control unit 39 represents the detection result of the specific keyword representing the user B's read, which is supplied from the voice recognition unit 31, and the user B's read that is supplied from the image recognition unit 32. Based on the detection result of the specific action, it is determined whether or not the reading of the user B with respect to the displayed utterance text has been detected.
  • FIG. 12 shows an example of determination of the read detection of the user B in step S13.
  • the user B can detect when a specific operation representing a read such as a nod is detected from the image recognition result of a moving image obtained by capturing the user B, the user B can detect when a specific operation indicating the read is detected a predetermined number of times (for example, twice). It is estimated that the user B has been read, and it is determined that the user B has been read.
  • the conversation is progressing between the user A and the user B when the user A can be detected. It is estimated that B understands, and it is determined that the user B's read has been detected.
  • the read determination of the user B is not limited to the above-described example.
  • the user may arbitrarily add a specific keyword indicating read or a specific operation indicating read.
  • step S14 the display control unit 39 determines whether or not a predetermined longest display time has elapsed since the display of the currently displayed utterance text has started, and the longest display time has elapsed. The process returns to step S13 until steps S13 and S14 are repeated. Then, when the user B has been read or the longest display time has elapsed, the process proceeds to step S15.
  • step S ⁇ b> 15 the display control unit 39 reads the utterance texts from the display waiting list in time series order, generates an image signal for displaying the read utterance texts on the screen, and supplies the image signals to the display unit 22 and the display unit 43. .
  • the display control unit 39 reads the utterance texts from the display waiting list in time series order, generates an image signal for displaying the read utterance texts on the screen, and supplies the image signals to the display unit 22 and the display unit 43. .
  • the screens of the display unit 22 and the display unit 43 are already full of utterance text, the screen is scrolled, the utterance text that was displayed first disappears from the screen, and is newly read from the display wait list.
  • the issued utterance text is displayed on the screen.
  • step S11 If it is determined in step S11 that the utterance text is not currently displayed on the screens of the display units 22 and 43, steps S12 to S14 are skipped, and the process proceeds to step S15.
  • the display waiting list generation process and the utterance text display process are executed in parallel, so that the user A's utterance is presented to the user B as the utterance text, The display of the utterance text is advanced.
  • FIG. 13 shows a situation where, for example, a user A who is an elementary school student and a user B who is a mother have a conversation using the conversation support device 10.
  • the user A uttered without saying at a stretch that he said, “If you went to school yesterday, you were told to collect 10000 yen because you collected money for school trips.”
  • FIG. 14 shows a display example of the display unit 43 in the situation shown in FIG. 14A shows a state in which the editing process is not reflected
  • FIG. 14B shows a state in which line breaks and page breaks are reflected in the editing process
  • FIG. 14C shows a line break and page break. The page and the text amount suppression processing are all reflected.
  • the display unit 43 initially displays the utterance text that does not reflect the editing process as shown in A of FIG. 14. Is done. In this state, line breaks and page breaks are made regardless of the meaning and context, so it is difficult to read, and the numerical value (10000 yen in the case of the figure) is divided in the middle, so the numerical value can be misunderstood. There is sex.
  • line breaks and page breaks in the editing process are reflected, which are shown in FIG. 14B.
  • line breaks and page breaks are performed according to the meaning and context of the utterance text, making it easier to read and the effect of suppressing misunderstandings such as numerical values.
  • the displayed utterance text may be deleted from the screen.
  • the display of A of FIG. 14 may be returned.
  • the display may return to the display shown in FIG. 14B.
  • the display of FIG. 14B is displayed, and when the user B performs the first operation, the display of FIG.
  • the displayed utterance text may be erased from the screen. Thereafter, each time the user B performs the first operation again, the display may return to the display of C in FIG. 14, B in FIG. 14, or A in FIG. 14.
  • the editing process is reflected in the displayed utterance text in response to the operation by the user B.
  • the editing process is performed on the displayed utterance text in accordance with the operation by the user A. It is also possible to reflect.
  • at least one of the first operation, the second operation, or the third operation may be regarded as a predetermined operation in one aspect of the present technology.
  • FIG. 15 shows a situation where user A and user B have a conversation using the conversation support device 10. However, illustration of user B is omitted. In the case of the figure, it is assumed that the user A utters a relatively short sentence such as “Good morning”, “Tomorrow at 10 o'clock at the Shinagawa station”.
  • FIG. 16 shows a display example on the display 43 of the utterance text corresponding to the utterance of the user A shown in FIG.
  • the utterance text corresponding to the sentence is also displayed divided into short parts as shown in FIG.
  • the utterance text other than “Good morning” is displayed in a state in which the text amount suppression process for deleting the nouns and the verbs and deleting the particles is reflected. That is, in the text amount suppression process of this example, parts of speech that are less important for understanding the meaning and context of the utterance text are omitted as appropriate.
  • the wording to be omitted is not limited to the part of speech, and may be appropriately set by the user.
  • the particles may be displayed less prominently than nouns or verbs related to the meaning or context of the utterance text.
  • the utterance text may be displayed such that nouns, verbs, etc. stand out from particles, etc.
  • FIG. 17 shows a display example in which the size of the particles such as particles is made smaller than the nouns and verbs related to the meaning and context of the utterance text so that the nouns and verbs stand out.
  • the color of particles such as particles is light and the color of characters such as nouns and verbs is displayed darkly, the brightness of characters such as particles is low, and the brightness of characters such as nouns and verbs is reduced.
  • the line may be displayed higher, or the line of a character such as a particle may be thinned, and the line of a character such as a noun or a verb may be displayed thickly.
  • FIG. 18 shows a display example when the delete button 111 is provided corresponding to each utterance text displayed on the display unit 22 for the user A.
  • Each utterance text shown in FIG. 18 corresponds to the utterance to user A shown in FIG.
  • the utterance text can be deleted by operating the delete button 111.
  • the word that should be recognized as “Shinagawa” is misrecognized as “Sanagawa”, so that when user A who has found this misrecognition operates the delete button 111, The utterance text including “Jonagawa” is deleted. Then, the misrecognition learning unit 33 learns that the utterance text including “Self” is erased (registered in the misrecognition list 34).
  • the user A can delete the misrecognized utterance text or the utterance text corresponding to the wrong utterance.
  • the delete button 111 can also be provided on the display section 43 for the user B.
  • the user B can erase the utterance text that has been read, for example, by operating the erase button 111.
  • the fact is notified to the user A side.
  • the user A can confirm the read of the user B with respect to the erased utterance text.
  • the fact may be notified to the user B side.
  • This notification method may use screen display or audio output.
  • FIG. 19 shows a display example when a re-utterance button 112 is provided corresponding to each utterance text displayed on the display unit 22 for the user A. Note that each utterance text shown in FIG. 19 corresponds to the utterance to user A shown in FIG.
  • the user A when the user A finds misrecognition in the utterance text that is the voice recognition result of his / her utterance, the user can re-phrase (speak again) the utterance text by operating the recurrence button 112.
  • the word that should be recognized as “Shinagawa” is misrecognized as “Sanakawa”, so that the user A who has found this misrecognition operates the recurrence button 112.
  • the currently displayed “Gather to Jonagawa tomorrow at 10:00” will be the utterance text that is the speech recognition result of the recurrent speech (correctly recognized) In case “Tomorrow we will gather at Shinagawa at 10:00”). Further, it is learned by the misrecognition learning unit 33 that the utterance text including “Sanagawa” has been replaced (registered in the misrecognition list 34).
  • the user A operates the recurrence button 112 to replace the misrecognized utterance text or the display of the utterance text corresponding to the wrong utterance with the utterance text corresponding to the recurrence at the position. Can do.
  • the re-utterance button 112 can be provided on the display section 43 for the user B. In that case, in response to the user B operating the re-utterance button 112, the user A is notified so as to prompt the re-utterance.
  • This notification method may use screen display or audio output.
  • FIG. 20 shows a display example when an NG word registration button 113 is provided corresponding to each utterance text displayed on the display unit 22 for the user A.
  • each utterance text shown by FIG. 20 respond
  • NG word registration button 113 For example, if user A finds misrecognition in the utterance text that is the speech recognition result of his utterance and does not want the misrecognition result to appear again, it is registered as an NG word by operating the NG word registration button 113 can do.
  • the user A can register a word that is erroneously recognized and is not desired to be displayed again as an NG word.
  • the NG word registration button 113 can also be provided on the display section 43 for the user B. In that case, the user B can also register a word that he / she does not want to redisplay as an NG word by operating the NG word registration button 113.
  • FIG. 21 shows a display example when the append button 114 is provided corresponding to each utterance text displayed on the display unit 22 for the user A.
  • Each utterance text shown in FIG. 21 corresponds to the utterance to user A shown in FIG.
  • the display example of FIG. 21 shows a result of the operation of the add button 114, and “?” Is added to the utterance text “Drinking medicine for lunch today” corresponding to the utterance of the user A. Yes. In this case, the fact that “?” Is added to “Today's lunch is already taken” is registered in the additional writing learning unit 37.
  • the user A can add “?” To the utterance text by operating the add button 114.
  • the append button 114 can be provided on the display unit 43 for the user B. In that case, when the user B does not understand the meaning of the displayed utterance text or wants to know more detailed contents, the user B selects a word or the like included in the displayed utterance text, and then appends it. By operating the button 114, the user A can be inquired about the meaning of a word or the like.
  • the delete button 111, the recurrence button 112, the NG word registration button 113, and the postscript button 114 are displayed individually, but they may be displayed simultaneously.
  • a predetermined touch operation for example, when the operation input unit 23 is a touch panel, a double operation is performed in response to an erase instruction, a recurrent speech instruction, an NG word registration, and an additional write instruction. Operation, long tap operation, flick operation, etc.
  • a three-dimensional gesture operation performed by the user A or the user B may be assigned to the deletion instruction, the recurrent speech instruction, the NG word registration, and the additional recording instruction.
  • the touch operation may be regarded as a two-dimensional gesture operation.
  • the three-dimensional gesture operation may be performed using a controller included in the acceleration sensor or the gyro sensor, or may be performed using an image recognition result related to the user's operation.
  • these touch operations and three-dimensional gesture operations may be simply referred to as “gesture operations”.
  • a nodding operation of the user B an operation of shaking the head, and the like can be assigned as the gesture operation.
  • the gaze detection function is employed in the wearable device, a physical action according to the movement of the gaze of the user B with respect to the displayed utterance text may be learned as a gesture operation. According to such a configuration, it is possible to improve the accuracy of the already-read determination according to the gesture operation.
  • a predetermined magic word uttered by the user A or the user B may be assigned to the erasure instruction, the recurrent speech instruction, the NG word registration, and the additional writing instruction.
  • the display of the utterance text corresponding to the utterance can be stopped. Good.
  • the discontinuation of the display of the utterance text can include the discontinuation of the display of the text in the middle of the analysis, that is, the discontinuation of the display processing of the undisplayed text.
  • the display of the utterance text is to be stopped, one sentence immediately before the erasure instruction is performed may be collectively erased by analyzing text information. As a result, it is possible to cancel text information (such as after words or fillers) that the user A has unintentionally input.
  • the information processing unit 30 immediately follows the predetermined gesture or the predetermined magic word. The display of the voice input that is input to may be prohibited. Thereby, since the user A can arbitrarily select a state in which no utterance is transmitted, the display of an unintended utterance can be suppressed.
  • FIG. 22 shows an example of a usage situation when the conversation support apparatus 10 can be used by three or more people.
  • the conversation support device 10 is used to support the conversation between the users A1, A2 and A3 who are not concerned about hearing and the user B who is concerned about hearing.
  • Each of the users A1 to A3 has a smartphone 50 for the user A, and the utterance texts corresponding to the utterances collected by the smartphones 50 existing in a predetermined distance range are grouped. Are collectively displayed on the display unit 43.
  • each smartphone 50 outputs a predetermined sound wave to each other, and can be realized by collecting and analyzing the sound wave output by other than itself.
  • the smartphone 50 may be detected from an image obtained by the camera 110 installed on the ceiling, and the position of each smartphone 50 may be specified.
  • utterance texts corresponding to the utterances of the users A1 to A3 are displayed in time series, and the displayed utterance text is uttered by whom of the users A1 to A3.
  • the speaker mark 121 representing the speaker is displayed in association with each utterance text so that the user B can determine whether it is present.
  • FIG. 23 shows the direction in which the speaker is in the state where the user B looks at the display unit 43 as another method for indicating who the user A1 to A3 is speaking the displayed utterance text. Is displayed on the screen.
  • the utterance direction instruction mark is displayed on the right side of the screen of the display unit 43. 131 is displayed.
  • the relative directions of the users A1, A2, and A3 when the user B looks at the display unit 43 can be detected from an image obtained by the camera 110 installed on the ceiling, for example.
  • FIG. 24 shows a situation in which the user A and the user B facing each other across the table are using the conversation support device 10.
  • the projector 80 may collectively project the screen of the display unit 22 for the user A and the screen of the display unit 43 for the user B onto the table.
  • the screen of the display unit 22 for the user A is displayed in a direction that the user A can easily read
  • the screen of the display unit 43 for the user B is displayed in a direction that the user B can easily read.
  • FIG. 25 shows an example of feedback to the user A who is the speaker among the users who are using the conversation support device 10.
  • the feedback control unit 40 controls the user A who is the speaker, for example, “SlowSdown”, “The screen is full”. , “Please speak slowly”, “Please wait”, “Please divide once”, “Feed unread”, etc. feedback to inform you to slow down the speaking rate, text display and voice output using smartphone 50 etc. Is done by.
  • an indicator corresponding to the utterance speed of the user A and the length of the utterance break may be displayed on the screen, or an alarm sound or the like may be output.
  • the user A When the user A speaks at an optimum speed or segmentation for voice recognition or screen display, the user A is given points, and the user A performs some service according to the given points. Benefits and rankings may be obtained.
  • the conversation support device 10 is used for the purpose of supporting the conversation between the user A who is not anxious about hearing and the user B who is uneasy about hearing.
  • the present invention can be applied to applications that support conversations between people using different languages. In that case, a translation process may be performed after the voice recognition process.
  • the conversation support device 10 may capture the mouth when the user A speaks as a moving image, display the utterance text, and display the moving image of the user A's mouth.
  • the display of the utterance text and the motion of the moving image of the user A's mouth may be displayed in synchronization.
  • the conversation support device 10 can be used for learning lip reading, for example.
  • the conversation support device 10 may record the utterance of the user A and store the utterance text that is the voice recognition result in association with the utterance text so that the saved result can be reproduced and displayed later.
  • the series of processes described above can be executed by hardware or can be executed by software.
  • a program constituting the software is installed in the computer.
  • the computer includes, for example, a general-purpose computer capable of executing various functions by installing a computer incorporated in dedicated hardware and various programs.
  • the smartphone 50 in the second configuration example described above corresponds to the computer.
  • FIG. 26 is a block diagram showing an example of the hardware configuration of a computer that executes the above-described series of processing by a program.
  • a CPU Central Processing Unit
  • ROM Read Only Memory
  • RAM Random Access Memory
  • An input / output interface 205 is further connected to the bus 204.
  • An input unit 206, an output unit 207, a storage unit 208, a communication unit 209, and a drive 210 are connected to the input / output interface 205.
  • the input unit 206 includes a keyboard, a mouse, a microphone, and the like.
  • the output unit 207 includes a display, a speaker, and the like.
  • the storage unit 208 includes a hard disk, a nonvolatile memory, and the like.
  • the communication unit 209 includes a network interface and the like.
  • the drive 210 drives a removable medium 211 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory.
  • the CPU 201 loads the program stored in the storage unit 208 to the RAM 203 via the input / output interface 205 and the bus 204 and executes the program. A series of processing is performed.
  • the program executed by the computer 200 can be provided by being recorded in, for example, a removable medium 211 such as a package medium.
  • the program can be provided via a wired or wireless transmission medium such as a local area network, the Internet, or digital satellite broadcasting.
  • the program can be installed in the storage unit 208 via the input / output interface 205 by attaching the removable medium 211 to the drive 210.
  • the program can be received by the communication unit 209 via a wired or wireless transmission medium and installed in the storage unit 208.
  • the program can be installed in the ROM 202 or the storage unit 208 in advance.
  • the program executed by the computer 200 may be a program that is processed in time series in the order described in this specification, or a necessary timing such as when a call is made in parallel. It may be a program in which processing is performed.
  • the present technology can also have the following configurations.
  • a voice acquisition unit that acquires the voice information of the first user input to the voice input device;
  • a display control unit for controlling display of text information corresponding to the acquired voice information in a display device for a second user, The display control unit controls the display amount of the text information based on at least one of the display amount of the text information on the display device or the input amount of the voice information input from the voice input device.
  • Processing equipment (2) The information processing apparatus according to (1), wherein the display control unit suppresses a display amount of the text information when a display amount of the text information becomes a predetermined amount or more.
  • the information processing apparatus according to (1) or (2), wherein the display control unit suppresses a display amount of the text information by suppressing a display amount of a predetermined part of speech included in the text information.
  • the display control unit suppresses a display amount of the text information based on a predetermined operation by the first user or the second user. apparatus.
  • the predetermined operation includes a first operation by the first user or the second user, The information processing apparatus according to (4), wherein the display control unit deletes the display of the text information based on the first operation after suppressing the display amount of the text information.
  • the predetermined operation includes a second operation by the first user or the second user
  • the information processing apparatus according to (5) wherein after the display of the text information is erased, the display control unit displays the text information erased in the display device again based on the second operation. .
  • information indicating that an operation related to the text information has been performed is used as information indicating the first user or the second user.
  • the information processing apparatus according to any one of (1) to (7), further including a notification unit that notifies the other of the information.
  • the notification unit may notify the other of the first user or the second user, The information processing apparatus according to (8), which notifies that the display amount of the text information is suppressed.
  • the notification unit may notify the other of the first user or the second user, The information processing apparatus according to (8) or (9), wherein the information indicating that the display of the text information has been deleted is notified.
  • the notification unit When the second user performs an operation for requesting re-speech of the text information displayed on the display device, the notification unit performs a notification to prompt the re-speech to the first user. ) To (10). (12) When the second user performs an operation for requesting an inquiry about the text information displayed on the display device, the notification unit has received an inquiry about the text information from the first user.
  • the information processing apparatus according to any one of (8) to (11).
  • the display control unit suppresses a display amount of the text information on the display device based on a result of the second user's read detection based on at least one of the second user's utterance or action.
  • the information processing apparatus according to any one of 1) to (12).
  • the information processing apparatus according to any one of (1) to (13), wherein the display control unit stops displaying the text information on the display device based on at least one of the utterance or action of the first user. .
  • a feedback control unit that controls notification of feedback information to at least one of the first user and the second user based on at least one of a display amount of the text information and an input amount of the audio information on the display device;
  • the information processing apparatus according to any one of (1) to (14).
  • the feedback information is information that prompts the first user to change at least one of an utterance speed and an utterance break.
  • the information processing apparatus according to (15) or (16), wherein the feedback information is information that prompts the second user to read the text information displayed on the display device.
  • a voice recognition unit that converts the voice information of the first user into the text information;
  • the information processing apparatus according to any one of (1) to (17), wherein the voice recognition unit is provided inside the information processing apparatus or on a server connected via the Internet.
  • a voice input device for acquiring voice information of the first user;
  • a display control device for controlling display of text information corresponding to the acquired voice information;
  • a display device for displaying the text information for a second user in accordance with control from the display control device;
  • the display control device performs control related to a display amount of the text information based on at least one of a display amount of the text information on the display device or an input amount of the voice information input from the voice input device. Processing system.

Landscapes

  • Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • General Health & Medical Sciences (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The present technology relates to an information processing device, information processing method, and information processing system, with which it is possible to carry out a smooth and natural conversation with a hard-of-hearing person. An information processing device according to an aspect of the present technology comprises: a speech acquisition unit which acquires speech information of a first user which has been inputted into a speech input device; and a display control unit which, in a display device for a second user, controls a display of text information which corresponds to the acquired speech information. On the basis of a display quantity of the text information in the display device and/or of an input quantity of the speech information which has been inputted via the speech input device, the display control unit carries out control which relates to the display quantity of the text information in the display device. The present technology may be applied to a conversation assistance device which assists with a conversation by a hard-of-hearing person.

Description

情報処理装置、情報処理方法、および情報処理システムInformation processing apparatus, information processing method, and information processing system
 本技術は、情報処理装置、情報処理方法、および情報処理システムに関し、特に、音声認識を利用した自然な会話を支援できるようにした情報処理装置、情報処理方法、および情報処理システムに関する。 The present technology relates to an information processing apparatus, an information processing method, and an information processing system, and more particularly, to an information processing apparatus, an information processing method, and an information processing system that can support natural conversation using voice recognition.
 スマートフォン等を利用した音声認識アプリケーションプログラムとして、ユーザの発話をテキスト化して画面に表示する技術が知られている。さらに、音声認識結果のテキストを要約する技術も存在する(例えば、特許文献1参照)。 As a voice recognition application program using a smartphone or the like, a technique for converting a user's utterance into text and displaying it on a screen is known. Furthermore, there is a technique for summarizing the text of the speech recognition result (see, for example, Patent Document 1).
WO2014-061388WO2014-061388
 ただし、スマートフォン等を利用した音声認識アプリケーションプログラムの場合、テキスト化された文章の表示量に限界がある。このため、音声認識を利用したコミュニケーションの提供に関して改善が求められている。 However, in the case of a speech recognition application program that uses a smartphone or the like, there is a limit to the amount of text that can be displayed. For this reason, there is a need for improvement in providing communication using voice recognition.
 本技術はこのような状況に鑑みてなされたものであり、音声認識を利用した自然な会話を支援する技術である。 This technology has been made in view of these circumstances, and is a technology that supports natural conversation using voice recognition.
 本技術の一側面である情報処理装置は、音声入力装置に入力された第1のユーザの音声情報を取得する音声取得部と、第2のユーザのための表示装置における、取得された前記音声情報に対応するテキスト情報の表示を制御する表示制御部とを備え、前記表示制御部は、前記表示装置における前記テキスト情報の表示量、または前記音声入力装置から入力された前記音声情報の入力量の少なくとも一方に基づいて、前記テキスト情報の表示量に関する制御を行う。 The information processing apparatus according to an aspect of the present technology includes the voice acquisition unit that acquires the voice information of the first user input to the voice input device, and the acquired voice in the display device for the second user. A display control unit that controls display of text information corresponding to the information, wherein the display control unit displays the text information on the display device or the input amount of the voice information input from the voice input device Based on at least one of the above, control is performed regarding the display amount of the text information.
 本技術の一側面である情報処理方法は、情報処理装置の情報処理方法において、前記情報処理装置による、音声入力装置に入力された第1のユーザの音声情報を取得する音声取得ステップと、第2のユーザのための表示装置における、取得された前記音声情報に対応するテキスト情報の表示を制御する表示制御ステップとを含み、前記表示制御ステップは、前記表示装置における前記テキスト情報の表示量、または前記音声入力装置から入力された前記音声情報の入力量の少なくとも一方に基づいて、前記テキスト情報の表示量に関する制御を行う。 An information processing method according to one aspect of the present technology includes: an audio acquisition step of acquiring audio information of a first user input to an audio input device by the information processing device; A display control step for controlling display of text information corresponding to the acquired voice information in a display device for two users, wherein the display control step includes a display amount of the text information on the display device, Alternatively, control regarding the display amount of the text information is performed based on at least one of the input amounts of the speech information input from the speech input device.
 本技術の一側面であるプログラムは、第1のユーザの音声情報を取得する音声入力装置と、取得された前記音声情報に対応するテキスト情報の表示を制御する表示制御装置と、前記表示制御装置からの制御に従い、前記テキスト情報を第2のユーザのために表示する表示装置とを備え、前記表示制御装置が、前記表示装置における前記テキスト情報の表示量、または前記音声入力装置から入力された前記音声情報の入力量の少なくとも一方に基づいて、前記テキスト情報の表示量に関する制御を行う。 A program that is one aspect of the present technology includes a voice input device that acquires voice information of a first user, a display control device that controls display of text information corresponding to the acquired voice information, and the display control device A display device that displays the text information for the second user in accordance with the control from the display device, the display control device being input from the display amount of the text information on the display device or the voice input device Control relating to the display amount of the text information is performed based on at least one of the input amounts of the voice information.
 本技術の一側面においては、入力された第1のユーザの音声情報が取得され、第2のユーザのための表示装置における、取得された前記音声情報に対応するテキスト情報の表示が制御される。この表示制御では、前記表示装置における前記テキスト情報の表示量、または前記音声入力装置から入力された前記音声情報の入力量の少なくとも一方に基づいて、前記テキスト情報の表示量に関する制御が行われる。 In one aspect of the present technology, the input voice information of the first user is acquired, and display of text information corresponding to the acquired voice information on the display device for the second user is controlled. . In this display control, control relating to the display amount of the text information is performed based on at least one of the display amount of the text information on the display device or the input amount of the speech information input from the speech input device.
 本技術の一側面によれば、音声認識を利用した自然な会話を行うことができる。 According to one aspect of the present technology, natural conversation using voice recognition can be performed.
本技術を適用した会話支援装置の第1の構成例を示す図である。It is a figure which shows the 1st structural example of the conversation assistance apparatus to which this technique is applied. 本技術を適用した会話支援装置の内部構成例を示すブロック図である。It is a block diagram which shows the internal structural example of the conversation assistance apparatus to which this technique is applied. 情報処理部の機能ブロック図である。It is a functional block diagram of an information processing part. 本技術を適用した会話支援装置の第2の構成例を示す図である。It is a figure which shows the 2nd structural example of the conversation assistance apparatus to which this technique is applied. 会話支援装置の構成要素を第2の構成例に対応して区分したブロック図である。It is the block diagram which divided the component of the conversation assistance apparatus corresponding to the 2nd structural example. 本技術を適用した会話支援装置の第3の構成例を示す図である。It is a figure which shows the 3rd structural example of the conversation assistance apparatus to which this technique is applied. 会話支援装置の構成要素を第3の構成例に対応して区分したブロック図である。It is the block diagram which divided the component of the conversation assistance apparatus corresponding to the 3rd structural example. 本技術を適用した会話支援装置の第4の構成例を示す図である。It is a figure which shows the 4th structural example of the conversation assistance apparatus to which this technique is applied. 会話支援装置の構成要素を第4の構成例に対応して区分したブロック図である。It is the block diagram which divided the component of the conversation assistance apparatus corresponding to the 4th structural example. 表示待ちリスト生成処理を説明するフローチャートである。It is a flowchart explaining a display waiting list production | generation process. 発話テキスト表示処理を説明するフローチャートである。It is a flowchart explaining an utterance text display process. 既読判定の例を示す図である。It is a figure which shows the example of already-read determination. テキスト量抑制処理の具体例を説明するための図である。It is a figure for demonstrating the specific example of a text amount suppression process. テキスト量抑制処理の具体例を説明するための図である。It is a figure for demonstrating the specific example of a text amount suppression process. テキスト量抑制処理の具体例を説明するための図である。It is a figure for demonstrating the specific example of a text amount suppression process. テキスト量抑制処理の具体例を説明するための図である。It is a figure for demonstrating the specific example of a text amount suppression process. テキスト量抑制処理の具体例を説明するための図である。It is a figure for demonstrating the specific example of a text amount suppression process. 編集処理(消去)の具体例を説明するための図である。It is a figure for demonstrating the specific example of an edit process (erasing). 編集処理(再発話)の具体例を説明するための図である。It is a figure for demonstrating the specific example of an edit process (recurrence story). 編集処理(NGワード登録)の具体例を説明するための図である。It is a figure for demonstrating the specific example of an edit process (NG word registration). 編集処理(追記)の具体例を説明するための図である。It is a figure for demonstrating the specific example of an edit process (additional writing). 会話支援装置の応用例を説明するための図である。It is a figure for demonstrating the application example of a conversation assistance apparatus. 会話支援装置の応用例を説明するための図である。It is a figure for demonstrating the application example of a conversation assistance apparatus. 会話支援装置の応用例を説明するための図である。It is a figure for demonstrating the application example of a conversation assistance apparatus. 発話者に対するフィードバックの例を説明するための図である。It is a figure for demonstrating the example of the feedback with respect to a speaker. コンピュータの構成例を示すブロック図である。It is a block diagram which shows the structural example of a computer.
 以下、本技術を実施するための最良の形態(以下、実施の形態と称する)について、図面を参照しながら詳細に説明する。 Hereinafter, the best mode for carrying out the present technology (hereinafter referred to as an embodiment) will be described in detail with reference to the drawings.
 <本技術の実施の形態である会話支援装置の第1の構成例>
 図1は、本技術の実施の形態である会話支援装置の第1の構成例を示しており、会話支援装置10を1つの筐体として形成した場合を示している。
<First configuration example of conversation support device according to an embodiment of the present technology>
FIG. 1 shows a first configuration example of a conversation support apparatus according to an embodiment of the present technology, and shows a case where the conversation support apparatus 10 is formed as one casing.
 会話支援装置10は、聴力に不安を持たない人(以下、ユーザAと称する)と、聴力に不安を持つ人(以下、ユーザBと称する)との間の会話を支援するためのものである。なお、本技術の一側面における第1のユーザが、本構成例のユーザAに対応し、本技術の一側面における第2のユーザが、本構成例のユーザ2に対応するものと見做してもよい。ただし、本技術の一側面における第1のユーザは、音声を入力するユーザであればよい。すなわち、第1のユーザ(音声を入力するユーザ)は、単一の主体(ユーザ)に限定されず、複数の主体(ユーザ)であり得る。同様に、本技術の一側面における第2のユーザは、表示された発話テキストを視認するユーザであればよく、単一の主体に限定されず、複数の主体であり得る。 The conversation support device 10 is for supporting a conversation between a person who does not have anxiety about hearing (hereinafter referred to as user A) and a person who has anxiety about hearing (hereinafter referred to as user B). . It is assumed that the first user in one aspect of the present technology corresponds to the user A in the present configuration example, and the second user in one aspect of the present technology corresponds to the user 2 in the present configuration example. May be. However, the 1st user in one side of this art should just be a user who inputs a voice. That is, the first user (user who inputs voice) is not limited to a single subject (user), and may be a plurality of subjects (users). Similarly, the second user in one aspect of the present technology may be a user who visually recognizes the displayed utterance text, and is not limited to a single subject, and may be a plurality of subjects.
 具体的には、ユーザAの発話を音声認識処理によってテキスト(以下、発話テキストと称する)に変換し、その発話テキストをユーザB用の表示部43に表示する。この表示をユーザBが読むことにより、ユーザBは、ユーザAの発話(音声情報)に対応する発話テキスト(文字情報)を理解することができる。 Specifically, the utterance of user A is converted into text (hereinafter referred to as utterance text) by voice recognition processing, and the utterance text is displayed on the display unit 43 for user B. By reading this display, the user B can understand the utterance text (character information) corresponding to the utterance (voice information) of the user A.
 表示部43に表示された発話テキストは、ユーザBが読み終わるか、または、所定に時間が経過するまで表示される。 The utterance text displayed on the display unit 43 is displayed until the user B finishes reading or a predetermined time elapses.
 表示された発話テキストがユーザBによって読み終わったか否かの判定には、例えば、撮像部41によって撮像された画像からユーザBの画像や、集音部42によって集音されたユーザBの発話が用いられる。 In determining whether or not the displayed utterance text has been read by the user B, for example, an image of the user B from an image captured by the imaging unit 41 or an utterance of the user B collected by the sound collecting unit 42 is used. Used.
 なお、ユーザB用の表示部43に裏面側には、ユーザA用の表示部22(図2)が設けられており、表示部22には、表示部43と同様の表示、すなわち、ユーザAの発話に対応する発話テキストが表示される。これにより、ユーザAは、自身の発話が正しく音声認識されたか否かを確認することができる。 Note that the display unit 43 for the user B is provided with a display unit 22 for the user A (FIG. 2) on the back side, and the display unit 22 has the same display as the display unit 43, that is, the user A The utterance text corresponding to the utterance is displayed. Thereby, the user A can confirm whether or not the user's utterance has been correctly recognized.
 <本技術の実施の形態である会話支援装置の構成例>
 図2は、本技術の実施の形態である会話支援装置の内部の構成例を示すブロック図である。
<Configuration example of conversation support apparatus according to an embodiment of the present technology>
FIG. 2 is a block diagram illustrating an internal configuration example of the conversation support apparatus according to the embodiment of the present technology.
 会話支援装置10は、集音部21、表示部22、操作入力部23、情報処理部30、撮像部41、集音部42、表示部43、および操作入力部44を有する。 The conversation support device 10 includes a sound collection unit 21, a display unit 22, an operation input unit 23, an information processing unit 30, an imaging unit 41, a sound collection unit 42, a display unit 43, and an operation input unit 44.
 集音部21、表示部22、および操作入力部23は、主にユーザAに対応するために設けられている。 The sound collection unit 21, the display unit 22, and the operation input unit 23 are provided mainly for the user A.
 集音部21は、ユーザAが話す音声(発話)を集音し、それに対応する発話信号を情報処理部30に供給する。表示部22は、情報処理部30から供給される画像信号(例えば、ユーザAの発話に対応する発話テキストを画面上に表示するための画像信号)に対応する画面を表示する。操作入力部23は、ユーザAからの各種の操作を受け付けて、それに対応する操作信号を情報処理部30に通知する。 The sound collecting unit 21 collects the voice (utterance) spoken by the user A and supplies the corresponding speech signal to the information processing unit 30. The display unit 22 displays a screen corresponding to the image signal supplied from the information processing unit 30 (for example, an image signal for displaying an utterance text corresponding to the utterance of the user A on the screen). The operation input unit 23 receives various operations from the user A and notifies the information processing unit 30 of operation signals corresponding thereto.
 情報処理部30は、集音部21から供給される発話信号を音声認識処理によって発話テキストに変換する。また、情報処理部30は、発話テキストを画面上に表示するための画像信号を表示部43に供給する。なお、情報処理部30の詳細については後述する。 The information processing unit 30 converts the speech signal supplied from the sound collection unit 21 into speech text by speech recognition processing. Further, the information processing unit 30 supplies an image signal for displaying the utterance text on the screen to the display unit 43. Details of the information processing unit 30 will be described later.
 撮像部41、集音部42、表示部43、および操作入力部44は、主にユーザBに対応するために設けられている。 The imaging unit 41, the sound collection unit 42, the display unit 43, and the operation input unit 44 are provided mainly for the user B.
 撮像部41は、ユーザB等を撮像し、その結果得られる動画像信号を情報処理部30に供給する。集音部42は、ユーザBが話す音声(発話)を集音し、それに対応する発話信号を情報処理部30に供給する。表示部43は、情報処理部30から供給される、ユーザAの発話に対応する発話テキストを画面上に表示するための画像信号に対応する画面を表示する。操作入力部44は、ユーザBからの各種の操作を受け付けて、それに対応する操作信号を情報処理部30に通知する。 The imaging unit 41 images the user B and supplies the moving image signal obtained as a result to the information processing unit 30. The sound collecting unit 42 collects the voice (speech) spoken by the user B and supplies the corresponding speech signal to the information processing unit 30. The display unit 43 displays a screen corresponding to an image signal supplied from the information processing unit 30 for displaying the utterance text corresponding to the utterance of the user A on the screen. The operation input unit 44 receives various operations from the user B and notifies the information processing unit 30 of operation signals corresponding thereto.
 <情報処理部30が有する機能ブロックの構成例>
 図3は、情報処理部30が有する機能ブロックの構成例を示している。
<Configuration example of functional block of information processing unit 30>
FIG. 3 shows a configuration example of functional blocks included in the information processing unit 30.
 情報処理部30は、音声認識部31、画像認識部32、誤認識学習部33、解析部35、編集部36、追記学習部37、表示待ちリスト保持部38、表示制御部39、およびフィードバック部40を有する。 The information processing unit 30 includes a voice recognition unit 31, an image recognition unit 32, a misrecognition learning unit 33, an analysis unit 35, an editing unit 36, an additional writing learning unit 37, a display waiting list holding unit 38, a display control unit 39, and a feedback unit. 40.
 音声認識部31は、集音部21から供給されるユーザAの発話に対応する発話信号を音声認識処理によって発話テキストに変換することにより発話テキストを生成し、解析部35に供給する。 The speech recognition unit 31 generates an utterance text by converting an utterance signal corresponding to the utterance of the user A supplied from the sound collection unit 21 into an utterance text by speech recognition processing, and supplies the utterance text to the analysis unit 35.
 また、音声認識部31は、集音部42から供給されるユーザBの発話に対応する発話信号を音声認識処理によって発話テキストに変換し、該発話テキストがユーザBの既読を表す特定のキーワード(予め登録されている、例えば「はい」、「うん」、「わかった」、「了解」、「次」等)を検出し、検出結果を表示制御部39に供給する。 In addition, the speech recognition unit 31 converts an utterance signal corresponding to the utterance of the user B supplied from the sound collection unit 42 into an utterance text by voice recognition processing, and the utterance text represents a specific keyword indicating the user B's already read. (For example, “Yes”, “Yes”, “Okay”, “OK”, “Next”, etc. registered in advance) are detected, and the detection result is supplied to the display control unit 39.
 画像認識部32は、撮像部41から供給される動画像信号に基づき、ユーザBの既読を表す特定の動作(例えば、うなずき、画面を注視していた後に画面以外の方向を見る等)を検出し、検出結果を表示制御部39に供給する。また、画像認識部32は、撮像部41から供給される動画像信号に基づき、ユーザBと表示部43との距離を測定し、測定結果を表示制御部39に通知する。ユーザBと表示部43との距離は、表示部43に表示する発話テキストの文字サイズを設定するために用いられる。例えば、ユーザBと表示部43との距離が長いほど、大きな文字サイズに設定される。 Based on the moving image signal supplied from the imaging unit 41, the image recognition unit 32 performs a specific operation indicating the user B's reading (for example, nodding, watching the screen and then looking in a direction other than the screen). It detects and supplies a detection result to the display control part 39. FIG. Further, the image recognition unit 32 measures the distance between the user B and the display unit 43 based on the moving image signal supplied from the imaging unit 41 and notifies the display control unit 39 of the measurement result. The distance between the user B and the display unit 43 is used to set the character size of the utterance text displayed on the display unit 43. For example, the longer the distance between the user B and the display unit 43, the larger the character size is set.
 なお、後述するヘッドマウントディスプレイ等のウェアラブル機器を用いる場合、ウェアラブル機器の向き、すなわちユーザBの頭の向きや体の向きに基づいて視線方向が判定されてもよい。ウェアラブル機器の向きは、ウェアラブル機器に設けられたカメラ、加速度センサ、ジャイロセンサ等から取得される位置情報に基づいて判定され得る。また、赤外線カメラと赤外線LEDを利用して、ユーザBの眼球のプルキニエ像と瞳孔中心を判定し、これらに基づいてユーザBの視線方向を判定してもよい。 Note that when a wearable device such as a head-mounted display described later is used, the line-of-sight direction may be determined based on the direction of the wearable device, that is, the direction of the user B's head or body. The direction of the wearable device can be determined based on position information acquired from a camera, an acceleration sensor, a gyro sensor, or the like provided in the wearable device. Further, the Purkinje image of the eyeball of the user B and the pupil center may be determined using an infrared camera and an infrared LED, and the line-of-sight direction of the user B may be determined based on these.
 誤認識学習部33は、音声認識処理の結果であるユーザAの発話に対応する発話テキストに対し、ユーザAまたはユーザBから入力される編集操作(例えば、消去指示操作、再発話指示操作、NGワード登録指示操作)に対応し、その発話テキストに含まれる誤認識されている単語を誤認識リスト34に登録する。また、誤認識学習部33は、音声認識処理の結果であるユーザAの発話に対応する発話テキストに、誤認識リスト34に登録された単語が含まれている場合、音声認識部31に対して、誤認識されている単語(認識結果の第1候補)以外の認識結果(第2候補等)を要求する。 The misrecognition learning unit 33 edits input from the user A or the user B to the utterance text corresponding to the utterance of the user A that is the result of the speech recognition process (for example, an erase instruction operation, a recurrent speech instruction operation, NG Corresponding to the word registration instruction operation), misrecognized words included in the utterance text are registered in the misrecognition list 34. In addition, when the utterance text corresponding to the utterance of the user A, which is the result of the speech recognition process, includes a word registered in the misrecognition list 34, the misrecognition learning unit 33 A recognition result (second candidate or the like) other than a misrecognized word (first candidate of recognition result) is requested.
 解析部35は、音声認識部31によって生成されたユーザAの発話に対応する発話テキストを、例えば、品詞に分解したり、キーワードを抽出したりする等の解析を行う。 The analysis unit 35 analyzes the speech text corresponding to the speech of the user A generated by the speech recognition unit 31, for example, by decomposing the speech text into parts of speech or extracting keywords.
 編集部36は、解析部35による解析結果に基づき、発話テキストに対して、適宜、改行や改頁を加えたり、消去しても発話テキストが有する意味を損なわない助詞等を特定するテキスト量抑制処理を行ったりする編集処理を行い、表示待ちリスト保持部38に供給する。なお、編集処理では、改行、改頁またはテキスト量抑制処理のうち少なくとも一つが行われるものと見做されてよく、改行、改頁またはテキスト量抑制処理のうち少なくとも一つが省略されてもよい。 Based on the analysis result of the analysis unit 35, the editing unit 36 controls the amount of text that specifies particles or the like that do not impair the meaning of the utterance text even if line breaks or page breaks are added or deleted as appropriate to the utterance text. An edit process such as a process is performed and supplied to the display wait list holding unit 38. In the editing process, it may be considered that at least one of line feed, page break or text amount suppression processing is performed, and at least one of line feed, page break or text amount suppression processing may be omitted.
 また、編集部36は、関連している複数の発話テキストをスレッドにまとめて表示待ちリスト保持部38に供給するようにすることができる。この場合、現在のスレッドを表示しつつ、表示待ちのスレッドに対応するアイコンを表示するようにしてもよい。表示待ちのスレッドを示す表示オブジェクトはアイコンに限定されず、適宜設定がなされてよい。このような構成によれば、ユーザBが、どの程度相手の発話テキストを読み終えたのかを容易に把握することができる。また、このような構成によれば、発話テキストの進捗状況に基づいて、ユーザBがユーザAの入力量を抑制するよう働き掛けることができる。 The editing unit 36 can supply a plurality of related utterance texts to a display waiting list holding unit 38 in a thread. In this case, an icon corresponding to a thread waiting for display may be displayed while displaying the current thread. Display objects indicating threads waiting to be displayed are not limited to icons, and may be appropriately set. According to such a configuration, it is possible to easily grasp how much the user B has finished reading the other party's utterance text. Moreover, according to such a structure, the user B can act to suppress the input amount of the user A based on the progress of the utterance text.
 さらに、編集部36は、表示部22に表示されている、ユーザAの発話に対応する発話テキストに対し、操作入力部23を用いてユーザAが入力する編集操作に基づき、発話テキストの一文を消去したり、再発話に対応する発話テキストを挿入したり、NGワードを登録したりする処理を制御する。また、編集部36は、表示部22に表示されている、ユーザAの発話に対応する発話テキストに対し、操作入力部23を用いてユーザAが入力する追記操作(具体的には、「?(クエスチョンマーク)」等の記号を追加する操作)に基づき、発話テキストに「?」等の記号を追加する処理を制御する。なお、「?」以外の記号、絵文字、顔文字等を追記できるようにしてもよい。 Further, the editing unit 36 outputs a sentence of the utterance text based on the editing operation input by the user A using the operation input unit 23 with respect to the utterance text corresponding to the utterance of the user A displayed on the display unit 22. Controls the process of deleting, inserting utterance text corresponding to recurrent utterances, and registering NG words. In addition, the editing unit 36 performs an additional operation (specifically, “?”) That is input by the user A using the operation input unit 23 to the utterance text corresponding to the utterance of the user A displayed on the display unit 22. The operation of adding a symbol such as “?” To the utterance text is controlled based on the operation of adding a symbol such as “(question mark)”. Note that symbols, pictograms, emoticons, etc. other than “?” May be additionally recorded.
 なお、編集部36は、表示部43に表示されている、ユーザAの発話に対応する発話テキストに対し、操作入力部44を用いてユーザBが入力する編集操作や追記操作にも基づいて、編集処理を行うことができる。換言すれば、表示されている、ユーザAの発話に対応する発話テキストに対しては、ユーザAとユーザBの双方が、編集操作や追記操作を行うことができる。 The editing unit 36 is based on an editing operation or an additional writing operation input by the user B using the operation input unit 44 with respect to the utterance text corresponding to the utterance of the user A displayed on the display unit 43. Editing processing can be performed. In other words, both the user A and the user B can perform an editing operation and an additional writing operation on the displayed utterance text corresponding to the utterance of the user A.
 追記学習部37はユーザAまたはユーザBが入力する追記操作を学習し、学習結果に基づき、ユーザAまたはユーザBからの追記操作が無くても、同様の発話テキストに対して同様の記号等を追記するように編集部36を制御する。 The additional writing learning unit 37 learns the additional writing operation input by the user A or the user B, and based on the learning result, even if there is no additional writing operation from the user A or the user B, the same symbol or the like is given to the same utterance text. The editing unit 36 is controlled so as to be additionally written.
 例えば、ユーザAの発話に対応する発話テキスト「薬は飲んだの」に対して「?」の追加を指示する追記操作が学習された場合、発話テキスト「薬は飲んだの」に対し、ユーザAまたはユーザBからの追記操作が無くても「?」を追加して「薬は飲んだの?」と編集するように編集部36を制御する。 For example, when an append operation for instructing the addition of “?” Is learned for the utterance text “Drug has been drunk” corresponding to the utterance of user A, Even if there is no additional operation from A or user B, the editing unit 36 is controlled to add “?” And edit “Did you take medicine?”.
 表示待ちリスト保持部38は、改行、改頁、またはテキスト量抑制処理のうちの少なくとも一つを含む編集処理済の発話テキスト(文字数によってはテキスト量抑制処理が行われない場合もある)を時系列順、すなわち、ユーザAが発話した順に表示待ちリストに登録する。表示待ちリストに登録された発話テキストは、表示制御部39から読み出された場合、表示待ちリストから消去される。 The display waiting list holding unit 38 displays the edited utterance text including at least one of line feed, page break, and text amount suppression processing (the text amount suppression processing may not be performed depending on the number of characters). The information is registered in the display wait list in the sequence order, that is, in the order in which the user A speaks. When the utterance text registered in the display waiting list is read from the display control unit 39, it is deleted from the display waiting list.
 表示制御部39は、表示待ちリストから時系列順に発話テキストを読み出し、読み出した発話テキストを画面上に表示するための画像信号を生成して表示部22および表示部43に供給する。また、表示制御部39は、表示部22および表示部43に現在表示されている発話テキストの表示量、音声認識部31から供給される、ユーザBの既読を表す特定のキーワードの検出結果、画像認識部32から供給されるユーザBの既読を表す特定の動作の検出結果等に基づき、表示部22および表示部43における発話テキストの表示量を制御する。さらに、表示制御部39は、ユーザBと表示部43との距離に応じて発話テキストを表示する際の文字サイズを設定する。 The display control unit 39 reads out the utterance texts from the display waiting list in chronological order, generates an image signal for displaying the read out utterance text on the screen, and supplies it to the display unit 22 and the display unit 43. Further, the display control unit 39 displays a display amount of the utterance text currently displayed on the display unit 22 and the display unit 43, a detection result of a specific keyword that is supplied from the voice recognition unit 31 and represents the user B already read, The display amount of the utterance text on the display unit 22 and the display unit 43 is controlled based on the detection result of the specific action representing the read of the user B supplied from the image recognition unit 32. Further, the display control unit 39 sets the character size for displaying the utterance text according to the distance between the user B and the display unit 43.
 フィードバック制御部40は、ユーザAの発話速度、ユーザAの発話の長さ、単位時間当たりの音声認識文字量、表示部43に表示されている発話テキストの量、表示待ちリストに登録されている発話テキストの量、ユーザBの既読の有無、ユーザBの読み取り速度等に対応して、発話者であるユーザAに対し、文字表示や音声出力を用いて、発話速度を早く(または遅く)するように通知したり、発話を区切るように通知したり、次の発声を促したりするフィードバックを制御する。また、フィードバック制御部40は、表示部43に表示されている発話テキストの量、表示待ちリストに登録されている発話テキストの量、ユーザBの既読の有無、ユーザBの読み取り速度等に対応して、ユーザBに対し、文字表示等を用いて、発話テキストを読むように促すフィードバックを制御する。 The feedback control unit 40 is registered in the utterance speed of the user A, the length of the utterance of the user A, the amount of speech recognition characters per unit time, the amount of utterance text displayed on the display unit 43, and the display waiting list. Corresponding to the amount of utterance text, whether or not the user B has already read, the reading speed of the user B, etc., the utterance speed is increased (or decreased) for the user A who is the utterer by using character display or voice output. To control the feedback to notify the user, to notify the utterance, or to prompt the next utterance. Further, the feedback control unit 40 corresponds to the amount of utterance text displayed on the display unit 43, the amount of utterance text registered in the display waiting list, whether or not the user B has been read, the reading speed of the user B, etc. Then, feedback that prompts the user B to read the utterance text is controlled by using a character display or the like.
 なお、情報処理部30が有する上述した機能ブロックは、全てが同一の筐体に収められている必要はなく、分散して配置されていてもよい。また、これらの機能ブロックの一部または全てがインターネット上のサーバ、いわゆる、クラウドネットワーク上に配置されていてもよい。 Note that the above-described functional blocks included in the information processing unit 30 do not have to be housed in the same casing, and may be arranged in a distributed manner. Some or all of these functional blocks may be arranged on a server on the Internet, a so-called cloud network.
 <本技術の実施の形態である会話支援装置の第2の構成例>
 図4は、本技術の実施の形態である会話支援装置の第2の構成例を示している。該第2の構成例では、会話支援装置10が異なる複数の電子装置から成るシステムとして構成される。この場合、会話支援装置10を構成する複数の電子装置間の接続は、有線接続でもよいし、所定の無線通信(例えば、Bluetooth(登録商標)、Wi-Fi(商標)等)を用いてもよい。
<Second configuration example of conversation support device according to an embodiment of the present technology>
FIG. 4 illustrates a second configuration example of the conversation support apparatus according to the embodiment of the present technology. In the second configuration example, the conversation support device 10 is configured as a system including a plurality of different electronic devices. In this case, the connection between the plurality of electronic devices constituting the conversation support apparatus 10 may be wired connection or may use predetermined wireless communication (for example, Bluetooth (registered trademark), Wi-Fi (trademark), etc.). Good.
 該第2の構成例は、会話支援装置10が、ユーザAが用いるスマートフォン50と、ユーザBが用いるタブレット型PC(以下、タブレットと称する)60から構成される。 In the second configuration example, the conversation support device 10 includes a smartphone 50 used by the user A and a tablet PC (hereinafter referred to as a tablet) 60 used by the user B.
 図5は、図2に示された会話支援装置10の構成要素を、スマートフォン50とタブレット型PC60とに区分した状態を示している。 FIG. 5 shows a state in which the constituent elements of the conversation support apparatus 10 shown in FIG. 2 are divided into the smartphone 50 and the tablet PC 60.
 すなわち、会話支援装置10の構成要素のうち、集音部21、表示部22、操作入力部23、および情報処理部30は、スマートフォン50によって実現される。この場合、スマートフォン50が備えるマイクロフォン、ディスプレイ、およびタッチパネル等がそれぞれ、集音部21、および操作入力部23に相当する。また、スマートフォン50が実行するアプリケーションプログラムが情報処理部30に相当する。 That is, among the components of the conversation support device 10, the sound collection unit 21, the display unit 22, the operation input unit 23, and the information processing unit 30 are realized by the smartphone 50. In this case, a microphone, a display, a touch panel, and the like included in the smartphone 50 correspond to the sound collection unit 21 and the operation input unit 23, respectively. An application program executed by the smartphone 50 corresponds to the information processing unit 30.
 また、会話支援装置10の構成要素のうち、撮像部41、集音部42、表示部43、および操作入力部44は、タブレット60によって実現される。この場合、タブレット60が備えるカメラ、マイクロフォン、ディスプレイ、およびタッチパネル等がそれぞれ撮像部41、集音部42、表示部43、および操作入力部44に相当する。 Of the components of the conversation support device 10, the imaging unit 41, the sound collection unit 42, the display unit 43, and the operation input unit 44 are realized by the tablet 60. In this case, the camera, microphone, display, touch panel, and the like included in the tablet 60 correspond to the imaging unit 41, the sound collection unit 42, the display unit 43, and the operation input unit 44, respectively.
 ただし、図5の場合、情報処理部30が有する機能ブロックのうちの音声認識部31が、インターネット71を介して接続可能なサーバ72に配置されている。 However, in the case of FIG. 5, the speech recognition unit 31 among the functional blocks of the information processing unit 30 is arranged in a server 72 that can be connected via the Internet 71.
 <本技術の実施の形態である会話支援装置の第3の構成例>
 図6は、本技術の実施の形態である会話支援装置の第3の構成例を示している。該第3の構成例は、会話支援装置10が異なる複数の電子装置から成るシステムとして構成される。
<Third configuration example of conversation support device according to an embodiment of the present technology>
FIG. 6 illustrates a third configuration example of the conversation support apparatus according to the embodiment of the present technology. In the third configuration example, the conversation support device 10 is configured as a system including a plurality of electronic devices.
 すなわち、該第3の構成例は、ユーザAが用いるスマートフォン50と、ベッドに横たわるユーザBが見ることができる位置、例えば、部屋の壁や天井に、発話テキストを表示する映像を投影するプロジェクタ80と、天井などに配置されているカメラ110とから構成される。 That is, in the third configuration example, the projector 80 that projects the video for displaying the utterance text on the smartphone 50 used by the user A and the position where the user B lying on the bed can see, for example, the wall or ceiling of the room. And a camera 110 arranged on the ceiling or the like.
 図7は、図2に示された会話支援装置10の構成要素を、スマートフォン50と、プロジェクタ80と、カメラ110とに区分した状態を示している。 FIG. 7 shows a state in which the constituent elements of the conversation support apparatus 10 shown in FIG. 2 are divided into a smartphone 50, a projector 80, and a camera 110.
 すなわち、会話支援装置10の構成要素のうち、集音部21、表示部22、操作入力部23、および情報処理部30は、スマートフォン50によって実現される。 That is, among the components of the conversation support device 10, the sound collection unit 21, the display unit 22, the operation input unit 23, and the information processing unit 30 are realized by the smartphone 50.
 また、会話支援装置10の構成要素のうち、撮像部41、および集音部42は、カメラ110によって実現される。この場合、カメラ110が備えるイメージセンサ、およびマイクロフォンがそれぞれ撮像部41、および集音部42に相当する。 Of the components of the conversation support device 10, the imaging unit 41 and the sound collection unit 42 are realized by the camera 110. In this case, the image sensor and the microphone included in the camera 110 correspond to the imaging unit 41 and the sound collecting unit 42, respectively.
 会話支援装置10の構成要素のうち、表示部43、および操作入力部44は、プロジェクタ80によって実現される。この場合、プロジェクタ80が備える投影部、およびリモートコントローラがそれぞれ表示部43、および操作入力部44に相当する。 Among the components of the conversation support apparatus 10, the display unit 43 and the operation input unit 44 are realized by the projector 80. In this case, the projection unit and the remote controller included in the projector 80 correspond to the display unit 43 and the operation input unit 44, respectively.
 図7の場合も、情報処理部30が有する機能ブロックのうちの音声認識部31が、インターネット71を介して接続可能なサーバ72に配置されている。 Also in the case of FIG. 7, the voice recognition unit 31 among the functional blocks included in the information processing unit 30 is arranged in a server 72 that can be connected via the Internet 71.
 <本技術の実施の形態である会話支援装置の第4の構成例>
 図8は、本技術の実施の形態である会話支援装置の第4の構成例を示している。該第4の構成例は、会話支援装置10が異なる複数の電子装置から成るシステムとして構成される。
<Fourth configuration example of the conversation support device according to the embodiment of the present technology>
FIG. 8 illustrates a fourth configuration example of the conversation support apparatus according to the embodiment of the present technology. In the fourth configuration example, the conversation support apparatus 10 is configured as a system including a plurality of different electronic devices.
 すなわち、該第4の構成例は、ユーザAが用いる首掛けマイク100と、ユーザAおよびユーザBが見ることができる位置に配置されているテレビジョン受像機(以下、TVと称する)90と、TV90に載置されているカメラ110とから構成される。 That is, the fourth configuration example includes a neck microphone 100 used by the user A, a television receiver (hereinafter referred to as TV) 90 disposed at a position where the user A and the user B can see, The camera 110 is mounted on the TV 90.
 図9は、図2に示された会話支援装置10の構成要素を、首掛けマイク100と、TV90と、カメラ110とに区分した状態を示している。 FIG. 9 shows a state in which the components of the conversation support device 10 shown in FIG. 2 are divided into a neck microphone 100, a TV 90, and a camera 110.
 すなわち、会話支援装置10の構成要素のうち、集音部21は、首掛けマイク100によって実現される。なお、首掛けマイク100には、集音部21の他、音声を出力するスピーカが設けられていてもよい。 That is, among the components of the conversation support device 10, the sound collection unit 21 is realized by the neck microphone 100. Note that the neck microphone 100 may be provided with a speaker that outputs sound in addition to the sound collecting unit 21.
 会話支援装置10の構成要素のうち、撮像部41、および集音部42は、カメラ110によって実現される。 Among the components of the conversation support device 10, the imaging unit 41 and the sound collection unit 42 are realized by the camera 110.
 会話支援装置10の構成要素のうち、表示部43、および操作入力部44は、TV90によって実現される。この場合、TV90が備えるディスプレイ、およびリモートコントローラがそれぞれ表示部43および操作入力部44に相当する。なお、TV90が備えるディスプレイ、およびリモートコントローラは、ユーザA用の表示部22および操作入力部23を兼ねているものとする。 Among the components of the conversation support device 10, the display unit 43 and the operation input unit 44 are realized by the TV 90. In this case, the display and remote controller included in the TV 90 correspond to the display unit 43 and the operation input unit 44, respectively. It is assumed that the display and the remote controller included in the TV 90 also serve as the display unit 22 and the operation input unit 23 for the user A.
 図9の場合も、情報処理部30が有する機能ブロックのうちの音声認識部31が、インターネット71を介して接続可能なサーバ72に配置されている。 Also in the case of FIG. 9, the voice recognition unit 31 among the functional blocks of the information processing unit 30 is arranged in a server 72 that can be connected via the Internet 71.
 上述した第1乃至第4の構成例のように、会話支援装置10は、1つの電子装置として構成することができるし、複数の電子装置を組み合わせたシステムとして構成することもできる。なお、上述した第1乃至第4の構成例は、適宜組み合わせることが可能である。 As in the first to fourth configuration examples described above, the conversation support device 10 can be configured as one electronic device, or can be configured as a system in which a plurality of electronic devices are combined. The first to fourth configuration examples described above can be combined as appropriate.
 また、会話支援装置10をシステムとして構成する電子装置としては、上述した例の他、時計型端末やヘッドマウントディスプレイ等のウェアラブル機器、PC(パーソナルコンピュータ)用のモニタ等を採用することができる。 Further, as an electronic device that configures the conversation support device 10 as a system, a wearable device such as a clock-type terminal or a head-mounted display, a monitor for a PC (personal computer), or the like can be employed in addition to the above-described example.
 <会話支援装置10の動作>
 次に、会話支援装置10の動作について説明する。
<Operation of Conversation Support Device 10>
Next, the operation of the conversation support apparatus 10 will be described.
 図10は、会話支援装置10による表示待ちリスト生成処理を説明するフローチャートである。この表示待ちリスト生成処理は、会話支援装置10が起動された後、電源がオフとされるまで、繰り返して実行される。 FIG. 10 is a flowchart for explaining display wait list generation processing by the conversation support apparatus 10. This display waiting list generation process is repeatedly executed until the power is turned off after the conversation support device 10 is activated.
 ステップS1において、ユーザAが発話すると、その音声が集音部21によって取得される。集音部21は、ユーザAの音声を発話信号に変換して情報処理部30に供給する。ステップS2において、情報処理部30では音声認識部31が音声認識処理を行うことにより、ユーザAの発話に対応する発話信号を発話テキストに変換する。 In step S1, when the user A speaks, the sound is acquired by the sound collecting unit 21. The sound collecting unit 21 converts the voice of the user A into an utterance signal and supplies it to the information processing unit 30. In step S <b> 2, in the information processing unit 30, the speech recognition unit 31 converts the speech signal corresponding to the speech of the user A into speech text by performing speech recognition processing.
 ステップS3において、解析部35は、音声認識部31によって生成されたユーザAの発話に対応する発話テキストを解析する。ステップS4において、編集部36は、解析結果に基づき、ユーザAの発話に対応する発話テキストに対し、改行、改頁、またはテキスト量抑制処理のうちの少なくとも一つを含む編集処理を行い、編集処理済の発話テキストを表示待ちリスト保持部38に供給する。 In step S3, the analysis unit 35 analyzes the utterance text corresponding to the utterance of the user A generated by the voice recognition unit 31. In step S4, the editing unit 36 performs an editing process including at least one of a line feed, a page break, or a text amount suppression process on the utterance text corresponding to the utterance of the user A based on the analysis result, The processed utterance text is supplied to the display waiting list holding unit 38.
 ステップS5において、表示待ちリスト保持部38は、編集部36から供給された編集処理済の発話テキストを時系列順に保持する。この後、処理はステップS1に戻されて、それ以降が繰り返される。 In step S5, the display waiting list holding unit 38 holds the edited utterance texts supplied from the editing unit 36 in chronological order. Thereafter, the process returns to step S1, and the subsequent steps are repeated.
 図11は、会話支援装置10による発話テキスト表示処理を説明するフローチャートである。この発話テキスト表示処理は、上述した表示待ちリスト生成処理と並行し、会話支援装置10が起動された後、電源がオフとされるまで、繰り返して実行される。 FIG. 11 is a flowchart for explaining utterance text display processing by the conversation support apparatus 10. The utterance text display process is repeatedly executed in parallel with the display wait list generation process described above until the power is turned off after the conversation support device 10 is activated.
 ステップS11において、表示制御部39は、現在、表示部22および43の画面に発話テキストが表示されているか否かを判定する。表示されていると判定した場合、処理はステップS12に進められる。ステップS12において、表示制御部39は、現在表示されている発話テキストの表示が開始されてから、予め定められている最短表示時間が経過しているか否かを判定し、最短表示時間が経過するまで待機する。最短表示時間が経過した場合、処理はステップS13に進められる。 In step S11, the display control unit 39 determines whether or not the utterance text is currently displayed on the screens of the display units 22 and 43. If it is determined that it is displayed, the process proceeds to step S12. In step S12, the display control unit 39 determines whether or not a predetermined shortest display time has elapsed since the display of the currently displayed utterance text has started, and the shortest display time has elapsed. Wait until. If the shortest display time has elapsed, the process proceeds to step S13.
 ステップS13において、表示制御部39は、音声認識部31から供給される、ユーザBの既読を表す特定のキーワードの検出結果、および画像認識部32から供給される、ユーザBの既読を表す特定の動作の検出結果に基づき、表示されている発話テキストに対するユーザBの既読を検知できたか否かを判定する。 In step S <b> 13, the display control unit 39 represents the detection result of the specific keyword representing the user B's read, which is supplied from the voice recognition unit 31, and the user B's read that is supplied from the image recognition unit 32. Based on the detection result of the specific action, it is determined whether or not the reading of the user B with respect to the displayed utterance text has been detected.
 図12は、ステップS13におけるユーザBの既読検知の判定の例を示している。 FIG. 12 shows an example of determination of the read detection of the user B in step S13.
 例えば、ユーザBによる発話の音声認識結果から「うん」等の既読を表す特定のキーワードが検出された場合、それが検知できた時点でユーザBが理解したと推定し、ユーザBの既読を検知できたと判定する。 For example, when a specific keyword representing a read such as “Yes” is detected from the speech recognition result of the utterance by the user B, it is assumed that the user B understands when a specific keyword indicating the read is detected, and the user B has already read Is determined to have been detected.
 また例えば、ユーザBを撮像した動画像の画像認識結果から、うなずき等の既読を表す特定の動作が検出された場合、それが所定回数(例えば、2回)検知できた時点でユーザBが理解したと推定し、ユーザBの既読を検知できたと判定する。 In addition, for example, when a specific operation representing a read such as a nod is detected from the image recognition result of a moving image obtained by capturing the user B, the user B can detect when a specific operation indicating the read is detected a predetermined number of times (for example, twice). It is estimated that the user B has been read, and it is determined that the user B has been read.
 また例えば、ユーザBを撮像した動画像の画像認識結果から、ユーザBが画面(表示部43)を注視していた後に画面以外の方向を見る状態が検出された場合、それが所定時間継続した時点でユーザBが理解したと推定し、ユーザBの既読を検知できたと判定する。 Further, for example, when a state in which the user B looks at the direction other than the screen after the user B is gazing at the screen (display unit 43) is detected from the image recognition result of the moving image obtained by capturing the user B, this continues for a predetermined time. It is estimated that the user B understands at the time, and it is determined that the user B has been detected.
 また例えば、ユーザAによる発話の音声認識結果から、ユーザAが新たに発話したことが検出された場合、それが検知できた時点でユーザAとユーザBの間で会話が進行しており、ユーザBが理解したと推定し、ユーザBの既読を検知できたと判定する。 Further, for example, when it is detected from the voice recognition result of the utterance by the user A that the user A has newly uttered, the conversation is progressing between the user A and the user B when the user A can be detected. It is estimated that B understands, and it is determined that the user B's read has been detected.
 なお、ユーザBの既読判定は、上述した例に限られるものではない。例えば、既読を表す特定のキーワードや、既読を表す特定の動作については、ユーザが任意に追加できるようにしてもよい。 In addition, the read determination of the user B is not limited to the above-described example. For example, the user may arbitrarily add a specific keyword indicating read or a specific operation indicating read.
 図12に戻り、ステップS13において、ユーザBの既読を検知できなかった場合、処理はステップS14に進められる。ステップS14において、表示制御部39は、現在表示されている発話テキストの表示が開始されてから、予め定められている最長表示時間が経過しているか否かを判定し、最長表示時間が経過するまで処理をステップS13に戻してステップS13およびS14を繰り返す。そして、ユーザBの既読を検知するか、または、最長表示時間が経過した場合、処理はステップS15に進められる。 Returning to FIG. 12, if the user B has not been read in step S <b> 13, the process proceeds to step S <b> 14. In step S14, the display control unit 39 determines whether or not a predetermined longest display time has elapsed since the display of the currently displayed utterance text has started, and the longest display time has elapsed. The process returns to step S13 until steps S13 and S14 are repeated. Then, when the user B has been read or the longest display time has elapsed, the process proceeds to step S15.
 ステップS15において、表示制御部39は、表示待ちリストから時系列順に発話テキストを読み出し、読み出した発話テキストを画面上に表示するための画像信号を生成して表示部22および表示部43に供給する。このとき、表示部22および表示部43の画面が既に発話テキストで一杯である場合、画面がスクロールされて、最も先に表示されていた発話テキストが画面上から消え、新たに表示待ちリストから読み出された発話テキストが画面上に表示される。 In step S <b> 15, the display control unit 39 reads the utterance texts from the display waiting list in time series order, generates an image signal for displaying the read utterance texts on the screen, and supplies the image signals to the display unit 22 and the display unit 43. . At this time, if the screens of the display unit 22 and the display unit 43 are already full of utterance text, the screen is scrolled, the utterance text that was displayed first disappears from the screen, and is newly read from the display wait list. The issued utterance text is displayed on the screen.
 なお、ステップS11において、現在、表示部22および43の画面に発話テキストが表示されていないと判定された場合、ステップS12乃至S14はスキップされて、処理はステップS15に進められる。 If it is determined in step S11 that the utterance text is not currently displayed on the screens of the display units 22 and 43, steps S12 to S14 are skipped, and the process proceeds to step S15.
 この後、処理はステップS11に戻されて、それ以降が繰り返される。 After this, the process returns to step S11 and the subsequent steps are repeated.
 以上説明したように、表示待ちリスト生成処理と発話テキスト表示処理が平行して実行されることにより、ユーザAの発話が発話テキストとしてユーザBに提示され、ユーザBの既読を待って、順次、発話テキストの表示が進められる。 As described above, the display waiting list generation process and the utterance text display process are executed in parallel, so that the user A's utterance is presented to the user B as the utterance text, The display of the utterance text is advanced.
 <改行、改頁、またはテキスト量抑制処理のうちの少なくとも一つを含む編集処理の具体例>
 次に、編集部36による改行、改頁、またはテキスト量抑制処理のうちの少なくとも一つを含む編集処理の具体例について説明する。
<Specific example of editing processing including at least one of line feed, page break, or text amount suppression processing>
Next, a specific example of editing processing including at least one of line feed, page break, or text amount suppression processing by the editing unit 36 will be described.
 図13は、例えば、小学生であるユーザAと、その母親であるユーザBとが、会話支援装置10を用いて会話を行う場合の様子である。同図においては、ユーザAが「昨日ね学校に行ったら修学旅行のお金を集めるから10000円持ってきてくれって言われた」と一気に区切ることなく発声しているものとする。 FIG. 13 shows a situation where, for example, a user A who is an elementary school student and a user B who is a mother have a conversation using the conversation support device 10. In the figure, it is assumed that the user A uttered without saying at a stretch that he said, “If you went to school yesterday, you were told to collect 10000 yen because you collected money for school trips.”
 図14は、図13に示された状況における表示部43の表示例を示している。ただし、図14のAは編集処理が反映されていない状態を示し、図14のBは、編集処理のうち、改行、および改頁が反映された状態を示し、図14のCは改行、改頁、およびテキスト量抑制処理の全てが反映されている状態を示している。 FIG. 14 shows a display example of the display unit 43 in the situation shown in FIG. 14A shows a state in which the editing process is not reflected, FIG. 14B shows a state in which line breaks and page breaks are reflected in the editing process, and FIG. 14C shows a line break and page break. The page and the text amount suppression processing are all reflected.
 図13に示されるように、ユーザAが発声を区切ることなく一気に発生した場合、表示部43には、当初、図14のAに示されるように、編集処理が反映されていない発話テキストが表示される。この状態は、意味や文脈に拘わりなく改行、改頁されるので、読みづらく、また、数値(同図の場合、10000円)が途中で分断されるため、数値を誤って理解されてしまう可能性が有る。 As shown in FIG. 13, when the user A occurs at a stroke without dividing the utterance, the display unit 43 initially displays the utterance text that does not reflect the editing process as shown in A of FIG. 14. Is done. In this state, line breaks and page breaks are made regardless of the meaning and context, so it is difficult to read, and the numerical value (10000 yen in the case of the figure) is divided in the middle, so the numerical value can be misunderstood. There is sex.
 図14のAの表示に対し、ユーザBが第1の操作(例えば、画面をタップする操作)を行うと、編集処理のうちの改行、および改頁が反映されて、図14のBに示されるように、発話テキストの意味や文脈に応じた改行や改頁が行われるので、読み易くなり、数値などの誤理解を抑止できる効果が期待できる。 When the user B performs a first operation (for example, an operation of tapping the screen) with respect to the display of FIG. 14A, line breaks and page breaks in the editing process are reflected, which are shown in FIG. 14B. As described above, line breaks and page breaks are performed according to the meaning and context of the utterance text, making it easier to read and the effect of suppressing misunderstandings such as numerical values.
 図14のBの表示に対し、ユーザBが第2の操作(例えば、画面をダブルタップする操作)を行うと、さらに、テキスト量抑制処理が反映されて、図14のCに示されるように、発話テキストのテキスト量が、意味や文脈が損なわれることなく抑制される。よって、上述した効果に加えて、ユーザBの既読に要する時間を短縮できる効果を期待できる。 When user B performs a second operation (for example, an operation of double-tapping the screen) on the display of B in FIG. 14, the text amount suppression process is further reflected, as shown in C in FIG. 14. The text volume of the utterance text is suppressed without impairing the meaning or context. Therefore, in addition to the above-described effects, it is possible to expect an effect that the time required for the user B to read can be shortened.
 なお、図14のCの表示に対し、ユーザBが第3の操作(例えば、画面をスワイプする操作)を行うと、表示されている発話テキストが画面から消去されるようにしてもよい。 Note that when the user B performs a third operation (for example, an operation of swiping the screen) on the display of C in FIG. 14, the displayed utterance text may be deleted from the screen.
 また、図14のBの表示に対し、ユーザBが第1の操作を行った場合、図14のAの表示に戻るようにしてもよい。同様に、図14のCの表示に対し、ユーザBが第2の操作を行った場合、図14のBの表示に戻るようにしてもよい。 Further, when the user B performs the first operation with respect to the display of B of FIG. 14, the display of A of FIG. 14 may be returned. Similarly, when the user B performs the second operation on the display shown in FIG. 14C, the display may return to the display shown in FIG. 14B.
 あるいは、図14のAの表示に対し、ユーザBが第1の操作を行うと図14のBの表示となり、さらに、ユーザBが第1の操作を行うと図14のCの表示となり、さらに、ユーザBが第1の操作を行うと、表示されている発話テキストが画面から消去されるようにしてもよい。この後、再びユーザBが第1の操作を行う毎に、図14のC、図14のB、または図14のAの表示に戻るようにしてもよい。 Alternatively, when the user B performs the first operation with respect to the display of FIG. 14A, the display of FIG. 14B is displayed, and when the user B performs the first operation, the display of FIG. When the user B performs the first operation, the displayed utterance text may be erased from the screen. Thereafter, each time the user B performs the first operation again, the display may return to the display of C in FIG. 14, B in FIG. 14, or A in FIG. 14.
 なお、上述した説明では、ユーザBによる操作に対応して、表示されている発話テキストに編集処理を反映するようにしたが、ユーザAによる操作に応じて、表示されている発話テキストに編集処理を反映することも可能である。また、第1の操作、第2の操作、または第3の操作の少なくとも一つが、本技術の一側面における所定の操作として見做されてもよい。 In the above description, the editing process is reflected in the displayed utterance text in response to the operation by the user B. However, the editing process is performed on the displayed utterance text in accordance with the operation by the user A. It is also possible to reflect. In addition, at least one of the first operation, the second operation, or the third operation may be regarded as a predetermined operation in one aspect of the present technology.
 <テキスト量抑制処理を含む編集処理の他の具体例>
 次に、テキスト量抑制処理を含む編集処理の他の具体例について説明する。
<Other specific examples of editing processing including text amount suppression processing>
Next, another specific example of editing processing including text amount suppression processing will be described.
 図15は、ユーザAとユーザBが会話支援装置10を用いて会話を行う場合の様子である。ただし、ユーザBの図示は省略されている。同図の場合、ユーザAが「おはようございます」、「明日は10時に品川駅に集合ね」等と比較的短い文を区切って発声しているものとする。 FIG. 15 shows a situation where user A and user B have a conversation using the conversation support device 10. However, illustration of user B is omitted. In the case of the figure, it is assumed that the user A utters a relatively short sentence such as “Good morning”, “Tomorrow at 10 o'clock at the Shinagawa station”.
 図16は、図15に示されたユーザAの発話に対応する発話テキストの表示部43における表示例を示している。ユーザAに発話が比較的短い文を区切って発声している場合、それに対応する発話テキストも、同図に示されるように、短分毎に区切って表示される。なお、同図の場合、「おはようございます」以外の発話テキストは、名詞や動詞を残して助詞等を消去するテキスト量抑制処理が反映された状態が表示されている。すなわち、本具体例のテキスト量抑制処理において、発話テキストの意味や文脈の理解のための重要性が低い品詞が適宜省略される。なお、省略する文言は品詞に限られず、ユーザによって適宜設定されてもよい。 FIG. 16 shows a display example on the display 43 of the utterance text corresponding to the utterance of the user A shown in FIG. When the user A utters a sentence with a relatively short utterance, the utterance text corresponding to the sentence is also displayed divided into short parts as shown in FIG. In the case of the figure, the utterance text other than “Good morning” is displayed in a state in which the text amount suppression process for deleting the nouns and the verbs and deleting the particles is reflected. That is, in the text amount suppression process of this example, parts of speech that are less important for understanding the meaning and context of the utterance text are omitted as appropriate. Note that the wording to be omitted is not limited to the part of speech, and may be appropriately set by the user.
 なお、消去しても発話テキストの意味や文脈が損なわれない助詞等を消去する代わりに、助詞等を、発話テキストの意味や文脈に関わる名詞や動詞等よりも目立たないように表示してもよい。換言すれば、名詞や動詞等が助詞等よりも目立つように発話テキストを表示するようにしてもよい。 Instead of deleting particles that do not impair the meaning or context of the utterance text, the particles may be displayed less prominently than nouns or verbs related to the meaning or context of the utterance text. Good. In other words, the utterance text may be displayed such that nouns, verbs, etc. stand out from particles, etc.
 図17は、助詞等の文字サイズを、発話テキストの意味や文脈に関わる名詞や動詞等よりも小さくして、名詞や動詞等を目立たせるようにした表示例を示している。 FIG. 17 shows a display example in which the size of the particles such as particles is made smaller than the nouns and verbs related to the meaning and context of the utterance text so that the nouns and verbs stand out.
 また図示は省略するが、助詞等の文字の色を薄く、名詞や動詞等の文字の色を濃く表示するようにしたり、助詞等の文字の輝度を低く、名詞や動詞等の文字の輝度を高く表示するようにしたり、助詞等の文字の線を細く、名詞や動詞等の文字の線を太く表示するようにしたりしてもよい。 Although not shown in the figure, the color of particles such as particles is light and the color of characters such as nouns and verbs is displayed darkly, the brightness of characters such as particles is low, and the brightness of characters such as nouns and verbs is reduced. The line may be displayed higher, or the line of a character such as a particle may be thinned, and the line of a character such as a noun or a verb may be displayed thickly.
 上述したように、発話テキストの意味に影響を及ぼさない助詞等を目立たせず、発話テキストの意味に影響を及ぼす名詞や動詞等を目立つように表示すれば、ユーザBは、目立たない助詞等を読まず、目立つ名詞や動詞等を読むことになる。よって、発話テキストが有する意味を損なうことなく、ユーザBが既読するまでに要する時間を短縮することができる。 As described above, if a noun, a verb, or the like that affects the meaning of the utterance text is displayed conspicuously without indicating a particle or the like that does not affect the meaning of the utterance text, the user B can display an insignificant particle, etc. Without reading, you will read prominent nouns and verbs. Therefore, it is possible to shorten the time required for the user B to read without losing the meaning of the utterance text.
 <編集部36による編集処理の具体例>
 次に、画面上に表示された発話テキストに対するユーザによるボタン操作に対応した編集処理について説明する。
<Specific Example of Editing Process by Editing Unit 36>
Next, the editing process corresponding to the button operation by the user for the utterance text displayed on the screen will be described.
 図18は、ユーザA用の表示部22に表示される各発話テキストに対応して消去ボタン111を設けた場合の表示例を示している。なお、図18に示される各発話テキストは、図15に示されたユーザAに発話に対応するものである。 FIG. 18 shows a display example when the delete button 111 is provided corresponding to each utterance text displayed on the display unit 22 for the user A. Each utterance text shown in FIG. 18 corresponds to the utterance to user A shown in FIG.
 例えば、ユーザAが自身の発話の音声認識結果である発話テキストに誤認識を発見した場合、消去ボタン111を操作することにより、当該発話テキストを消去させることができる。 For example, when the user A finds misrecognition in the utterance text that is the voice recognition result of his / her utterance, the utterance text can be deleted by operating the delete button 111.
 図18の表示例の場合、本来であれば「品川」に認識されるべき単語が「自奈川」に誤認識されているので、この誤認識を発見したユーザAが消去ボタン111を操作すると、「自奈川」を含む発話テキストが消去される。そして、「自奈川」を含む発話テキストが消去されたことが誤認識学習部33に学習される(誤認識リスト34に登録される)。 In the case of the display example in FIG. 18, the word that should be recognized as “Shinagawa” is misrecognized as “Sanagawa”, so that when user A who has found this misrecognition operates the delete button 111, The utterance text including “Jonagawa” is deleted. Then, the misrecognition learning unit 33 learns that the utterance text including “Self” is erased (registered in the misrecognition list 34).
 すなわち、ユーザAは、消去ボタン111を操作することにより、誤認識された発話テキストや言い間違えた発話に対応する発話テキストを消去することができる。 That is, by operating the delete button 111, the user A can delete the misrecognized utterance text or the utterance text corresponding to the wrong utterance.
 なお、ユーザB用の表示部43にも消去ボタン111を設けることができる。その場合、ユーザBは、消去ボタン111を操作することにより、例えば、読み終わった発話テキストを消去することができる。 Note that the delete button 111 can also be provided on the display section 43 for the user B. In this case, the user B can erase the utterance text that has been read, for example, by operating the erase button 111.
 ユーザBによる消去ボタン111の操作によって発話テキストを消去した場合、その旨をユーザA側に通知するようにする。これにより、ユーザAは、消去された発話テキストに対するユーザBの既読を確認することができる。反対に、ユーザAによる消去ボタン111の操作によって発話テキストを消去した場合、その旨をユーザB側に通知するようにしてもよい。この通知の方法は、画面表示を用いてもよいし、音声出力を用いてもよい。 When the utterance text is erased by the operation of the erase button 111 by the user B, the fact is notified to the user A side. Thereby, the user A can confirm the read of the user B with respect to the erased utterance text. On the contrary, when the utterance text is erased by the operation of the erase button 111 by the user A, the fact may be notified to the user B side. This notification method may use screen display or audio output.
 図19は、ユーザA用の表示部22に表示される各発話テキストに対応して再発話ボタン112を設けた場合の表示例を示している。なお、図19に示される各発話テキストは、図15に示されたユーザAに発話に対応するものである。 FIG. 19 shows a display example when a re-utterance button 112 is provided corresponding to each utterance text displayed on the display unit 22 for the user A. Note that each utterance text shown in FIG. 19 corresponds to the utterance to user A shown in FIG.
 例えば、ユーザAが自身の発話の音声認識結果である発話テキストに誤認識を発見した場合、再発話ボタン112を操作することにより、当該発話テキストを言い直す(再度発話する)ことができる。 For example, when the user A finds misrecognition in the utterance text that is the voice recognition result of his / her utterance, the user can re-phrase (speak again) the utterance text by operating the recurrence button 112.
 図19の表示例の場合、本来であれば「品川」に認識されるべき単語が「自奈川」に誤認識されているので、この誤認識を発見したユーザAが再発話ボタン112を操作し、改めて「明日は10時に品川に集合ね」等と発声すると、現在表示されている「明日は10時に自奈川に集合ね」が、再発話の音声認識結果である発話テキスト(正しく認識された場合「明日は10時に品川に集合ね」)によって置換される。また、「自奈川」を含む発話テキストが置換されたことが誤認識学習部33に学習される(誤認識リスト34に登録される)。 In the case of the display example in FIG. 19, the word that should be recognized as “Shinagawa” is misrecognized as “Sanakawa”, so that the user A who has found this misrecognition operates the recurrence button 112. When you say “Gather to Shinagawa at 10:00 tomorrow” again, the currently displayed “Gather to Jonagawa tomorrow at 10:00” will be the utterance text that is the speech recognition result of the recurrent speech (correctly recognized) In case “Tomorrow we will gather at Shinagawa at 10:00”). Further, it is learned by the misrecognition learning unit 33 that the utterance text including “Sanagawa” has been replaced (registered in the misrecognition list 34).
 すなわち、ユーザAは、再発話ボタン112を操作することにより、誤認識された発話テキストや言い間違えた発話に対応する発話テキストの表示を、その位置に再発話に対応する発話テキストによって置換させることができる。 That is, the user A operates the recurrence button 112 to replace the misrecognized utterance text or the display of the utterance text corresponding to the wrong utterance with the utterance text corresponding to the recurrence at the position. Can do.
 なお、発話テキスト全体(いまの場合、「明日は・・・集合ね」)を再発話するのではなく、単語(例えば、自奈川)を選択し、その単語だけを再発話できるようにしてもよい。 Instead of re-speaking the entire utterance text (in this case, “Tomorrow is a set”), select a word (for example, Jonagawa) so that only that word can be re-spoken. Good.
 また、ユーザB用の表示部43にも再発話ボタン112を設けることができる。その場合、ユーザBが再発話ボタン112を操作したことに対応して、ユーザA側に再発話を促すように通知するようにする。この通知の方法は、画面表示を用いてもよいし、音声出力を用いてもよい。 Also, the re-utterance button 112 can be provided on the display section 43 for the user B. In that case, in response to the user B operating the re-utterance button 112, the user A is notified so as to prompt the re-utterance. This notification method may use screen display or audio output.
 図20は、ユーザA用の表示部22に表示される各発話テキストに対応してNGワード登録ボタン113を設けた場合の表示例を示している。なお、図20に示される各発話テキストは、図15に示されたユーザAに発話に対応するものである。 FIG. 20 shows a display example when an NG word registration button 113 is provided corresponding to each utterance text displayed on the display unit 22 for the user A. In addition, each utterance text shown by FIG. 20 respond | corresponds to the utterance to the user A shown by FIG.
 例えば、ユーザAが自身の発話の音声認識結果である発話テキストに誤認識を発見し、その誤認識結果を再度出現させたくない場合、NGワード登録ボタン113を操作することにより、NGワードとして登録することができる。 For example, if user A finds misrecognition in the utterance text that is the speech recognition result of his utterance and does not want the misrecognition result to appear again, it is registered as an NG word by operating the NG word registration button 113 can do.
 図20の表示例の場合、ユーザAの何らかの発話が「エロ漫画」と誤認識されて表示されており、この単語を2度と表示させたくないと思ったユーザAがNGワード登録ボタン113を操作すると、表示されている発話テキスト「エロ漫画」が消去され、「エロ漫画」がNGワードとして誤認識学習部33に登録される(誤認識リスト34に登録される)。 In the case of the display example in FIG. 20, some utterance of user A is misrecognized as “erotic comic” and displayed, and user A who does not want to display this word again uses NG word registration button 113. When the operation is performed, the displayed utterance text “erotic comic” is deleted, and “erotic comic” is registered as an NG word in the misrecognition learning unit 33 (registered in the misrecognition list 34).
 すなわち、ユーザAは、NGワード登録ボタン113を操作することにより、誤認識され、再表示させたくない単語をNGワードとして登録することができる。 That is, by operating the NG word registration button 113, the user A can register a word that is erroneously recognized and is not desired to be displayed again as an NG word.
 なお、ユーザB用の表示部43にもNGワード登録ボタン113を設けることができる。その場合、ユーザBもNGワード登録ボタン113を操作することにより、再表示させたくない単語をNGワードとして登録することができる。 Note that the NG word registration button 113 can also be provided on the display section 43 for the user B. In that case, the user B can also register a word that he / she does not want to redisplay as an NG word by operating the NG word registration button 113.
 図21は、ユーザA用の表示部22に表示される各発話テキストに対応して追記ボタン114を設けた場合の表示例を示している。なお、図21に示される各発話テキストは、図15に示されたユーザAに発話に対応するものである。 FIG. 21 shows a display example when the append button 114 is provided corresponding to each utterance text displayed on the display unit 22 for the user A. Each utterance text shown in FIG. 21 corresponds to the utterance to user A shown in FIG.
 例えば、ユーザAが自身の質問等の発話の音声認識結果である発話テキストに対して「?」を追記した方がユーザBの理解が高まると考えた場合、追記ボタン114を操作することにより、当該発話テキストに「?」を追記することができる。 For example, if the user A thinks that adding “?” To the utterance text that is the voice recognition result of the utterance of his / her own question or the like will increase the understanding of the user B, by operating the append button 114, “?” Can be added to the utterance text.
 図21の表示例は、既に追記ボタン114が操作された結果を示しており、ユーザAの発話に対応する発話テキスト「今日のお昼はもう薬飲んだ」に対して「?」が追記されている。この場合、「今日のお昼はもう薬飲んだ」に対して「?」が追記されたことが、追記学習部37に登録される。 The display example of FIG. 21 shows a result of the operation of the add button 114, and “?” Is added to the utterance text “Drinking medicine for lunch today” corresponding to the utterance of the user A. Yes. In this case, the fact that “?” Is added to “Today's lunch is already taken” is registered in the additional writing learning unit 37.
 すなわち、ユーザAは、追記ボタン114を操作することにより、発話テキストに対して「?」を追記することができる。 That is, the user A can add “?” To the utterance text by operating the add button 114.
 また、ユーザB用の表示部43にも追記ボタン114を設けることができる。その場合、ユーザBは、表示されている発話テキストの意味が分からなかったり、より詳細な内容を知りたかったりした場合などにおいて、表示されている発話テキストに含まれる単語等を選択した後、追記ボタン114を操作することにより、ユーザA側に単語等の意味を問い合わせることができる。 In addition, the append button 114 can be provided on the display unit 43 for the user B. In that case, when the user B does not understand the meaning of the displayed utterance text or wants to know more detailed contents, the user B selects a word or the like included in the displayed utterance text, and then appends it. By operating the button 114, the user A can be inquired about the meaning of a word or the like.
 なお、「?」以外の記号や絵文字、顔文字などを追記できるように、追記ボタン114が操作された場合、追記する記号等をユーザに選択させるようにしてもよい。 It should be noted that when the append button 114 is operated so that symbols other than “?”, Pictograms, emoticons, etc. can be additionally recorded, the user may select a symbol to be added.
 上述した説明では、消去ボタン111、再発話ボタン112、NGワード登録ボタン113、および追記ボタン114を、それぞれ個別に表示するようにしているが、これらが同時に表示されるようにしてもよい。 In the above description, the delete button 111, the recurrence button 112, the NG word registration button 113, and the postscript button 114 are displayed individually, but they may be displayed simultaneously.
 また、各ボタンを表示する代わりに、消去指示、再発話指示、NGワード登録、および追記指示に対して、所定のタッチ操作(例えば、操作入力部23がタッチパネルである場合、タップ操作、ダブルタップ操作、ロングタップ操作、フリック操作等)を割り当てるようにしてもよい。さらに、各ボタンを表示する代わりに、消去指示、再発話指示、NGワード登録、および追記指示に対して、ユーザAまたはユーザBが行う三次元的なジェスチャ操作を割り当てるようにしてもよい。ここで、タッチ操作は二次元的なジェスチャ操作として見做されてもよい。また、三次元的なジェスチャ操作は、加速度センサやジャイロセンサが有するコントローラを利用して行われてもよく、ユーザの動作に関する画像認識結果を利用して行われてもよい。本明細書中において、これらのタッチ操作および三次元的なジェスチャ操作を、単に“ジェスチャ操作”という場合がある。 In addition, instead of displaying each button, a predetermined touch operation (for example, when the operation input unit 23 is a touch panel, a double operation is performed in response to an erase instruction, a recurrent speech instruction, an NG word registration, and an additional write instruction. Operation, long tap operation, flick operation, etc.) may be assigned. Furthermore, instead of displaying each button, a three-dimensional gesture operation performed by the user A or the user B may be assigned to the deletion instruction, the recurrent speech instruction, the NG word registration, and the additional recording instruction. Here, the touch operation may be regarded as a two-dimensional gesture operation. Further, the three-dimensional gesture operation may be performed using a controller included in the acceleration sensor or the gyro sensor, or may be performed using an image recognition result related to the user's operation. In this specification, these touch operations and three-dimensional gesture operations may be simply referred to as “gesture operations”.
 なお、ウェアラブル機器、特にヘッドマウントディスプレイが用いられる場合、ジェスチャ操作としてユーザBのうなずく動作や首を振る動作などがジェスチャ操作として割り当てられ得る。ウェアラブル機器に視線検出の機能が採用される場合、表示された発話テキストに対するユーザBの視線の移動に応じた身体的動作をジェスチャ操作として学習してもよい。このような構成によれば、ジェスチャ操作に応じた既読判定の精度を高めることが可能になる。 Note that when a wearable device, particularly a head-mounted display, is used, a nodding operation of the user B, an operation of shaking the head, and the like can be assigned as the gesture operation. When the gaze detection function is employed in the wearable device, a physical action according to the movement of the gaze of the user B with respect to the displayed utterance text may be learned as a gesture operation. According to such a configuration, it is possible to improve the accuracy of the already-read determination according to the gesture operation.
 またさらに、各ボタンを表示する代わりに、消去指示、再発話指示、NGワード登録、および追記指示に対して、ユーザAまたはユーザBが発声する所定のマジックワードを割り当てるようにしてもよい。 Furthermore, instead of displaying each button, a predetermined magic word uttered by the user A or the user B may be assigned to the erasure instruction, the recurrent speech instruction, the NG word registration, and the additional writing instruction.
 さらに、ユーザAが発話の直後に、消去指示に割り当てられている所定のジェスチャを行うか、または所定のマジックワードを発声した場合、その発話に対応する発話テキストの表示を中止できるようにしてもよい。 Further, when the user A performs a predetermined gesture assigned to the erasure instruction or utters a predetermined magic word immediately after the utterance, the display of the utterance text corresponding to the utterance can be stopped. Good.
 ここで、発話テキストの表示の中止は、解析途中のテキストの表示の中止、すなわち、未表示のテキストの表示処理の中止を含み得る。また、発話テキストの表示を中止する場合、テキスト情報の解析により、消去指示が行われる直前の一文をまとめて消去してもよい。これにより、ユーザAが意図せず音声入力を行ってしまったテキスト情報(独り言やフィラー後など)をキャンセルすることが可能になる。また、発話の前に、ユーザAが音声入力を行わないことを示す所定のジェスチャ、または所定のマジックワードを発声した場合、情報処理部30は、当該所定のジェスチャ、または所定のマジックワードの直後に入力される音声入力の表示を禁止してもよい。これにより、ユーザAは、発話を伝えない状態を任意に選択することができるため、意図しない発話の表示を抑制することができる。 Here, the discontinuation of the display of the utterance text can include the discontinuation of the display of the text in the middle of the analysis, that is, the discontinuation of the display processing of the undisplayed text. When the display of the utterance text is to be stopped, one sentence immediately before the erasure instruction is performed may be collectively erased by analyzing text information. As a result, it is possible to cancel text information (such as after words or fillers) that the user A has unintentionally input. In addition, when the user A utters a predetermined gesture or a predetermined magic word indicating that voice input is not performed before the utterance, the information processing unit 30 immediately follows the predetermined gesture or the predetermined magic word. The display of the voice input that is input to may be prohibited. Thereby, since the user A can arbitrarily select a state in which no utterance is transmitted, the display of an unintended utterance can be suppressed.
 <会話支援装置10の応用例>
 次に、会話支援装置10の応用例について説明する。
<Application example of conversation support device 10>
Next, an application example of the conversation support apparatus 10 will be described.
 図22は、会話支援装置10を3人以上で使用できるようにした場合の使用状況の一例を示している。同図の場合、聴力に不安を持たないユーザA1,A2,A3と、聴力に不安を持つユーザBとの間の会話を支援するために会話支援装置10が使用されている。 FIG. 22 shows an example of a usage situation when the conversation support apparatus 10 can be used by three or more people. In the case of the figure, the conversation support device 10 is used to support the conversation between the users A1, A2 and A3 who are not worried about hearing and the user B who is worried about hearing.
 ユーザA1乃至A3は、それぞれユーザA用のスマートフォン50を有しており、所定の距離範囲に存在するスマートフォン50がグループ化されて、それぞれが集音した発声に対応する発話テキストが、ユーザB用の表示部43に一括して表示される。 Each of the users A1 to A3 has a smartphone 50 for the user A, and the utterance texts corresponding to the utterances collected by the smartphones 50 existing in a predetermined distance range are grouped. Are collectively displayed on the display unit 43.
 なお、所定の距離範囲に存在するスマートフォン50を検知する方法としては、例えば、各スマートフォン50が相互に所定の音波を出力し、自己以外が出力した該音波を集音、解析することで実現できる。また、例えば、天井に設置しているカメラ110によって得られる画像からスマートフォン50を検知し、各スマートフォン50の位置を特定するようにしてもよい。 In addition, as a method of detecting the smartphone 50 existing in a predetermined distance range, for example, each smartphone 50 outputs a predetermined sound wave to each other, and can be realized by collecting and analyzing the sound wave output by other than itself. . Further, for example, the smartphone 50 may be detected from an image obtained by the camera 110 installed on the ceiling, and the position of each smartphone 50 may be specified.
 ユーザB用の表示部43には、ユーザA1乃至A3の発話に対応する発話テキストが時系列順に表示されるが、表示されている発話テキストが、ユーザA1乃至A3の誰から発声されたものであるのかをユーザBが判別できるように、各発話テキストに対応付けて発話者を表す発話者マーク121を表示するようにする。 In the display section 43 for the user B, utterance texts corresponding to the utterances of the users A1 to A3 are displayed in time series, and the displayed utterance text is uttered by whom of the users A1 to A3. The speaker mark 121 representing the speaker is displayed in association with each utterance text so that the user B can determine whether it is present.
 図23は、表示されている発話テキストが、ユーザA1乃至A3の誰から発声されたものであるのかを表す他の方法として、ユーザBが表示部43を見た状態において、発話者がどの方向に居るのかを表す発話方向指示マーク131を画面上に表示している。 FIG. 23 shows the direction in which the speaker is in the state where the user B looks at the display unit 43 as another method for indicating who the user A1 to A3 is speaking the displayed utterance text. Is displayed on the screen.
 図23の場合、ユーザBが表示部43を見た状態において右側に居るユーザA3の発話に対応する発話テキストが画面上に表示されているので、表示部43の画面の右側に発話方向指示マーク131が表示される。 In the case of FIG. 23, since the utterance text corresponding to the utterance of the user A3 on the right side when the user B looks at the display unit 43 is displayed on the screen, the utterance direction instruction mark is displayed on the right side of the screen of the display unit 43. 131 is displayed.
 なお、ユーザBが表示部43を見た状態におけるユーザA1,A2,A3の相対的な方向は、例えば、天井に設置しているカメラ110によって得られる画像から検知することができる。 The relative directions of the users A1, A2, and A3 when the user B looks at the display unit 43 can be detected from an image obtained by the camera 110 installed on the ceiling, for example.
 図24は、テーブルをはさんで向かい合うユーザAとユーザBが会話支援装置10を使用している状況である。この場合、プロジェクタ80により、ユーザA用の表示部22の画面とユーザB用の表示部43の画面を一括して当該テーブルに投影するようにしてもよい。このとき、ユーザA用の表示部22の画面はユーザAが読み易い向きに、ユーザB用の表示部43の画面はユーザBが読み易い向きに表示するようにする。 FIG. 24 shows a situation in which the user A and the user B facing each other across the table are using the conversation support device 10. In this case, the projector 80 may collectively project the screen of the display unit 22 for the user A and the screen of the display unit 43 for the user B onto the table. At this time, the screen of the display unit 22 for the user A is displayed in a direction that the user A can easily read, and the screen of the display unit 43 for the user B is displayed in a direction that the user B can easily read.
 <発話者であるユーザAに対するフィードバック>
 図25は、会話支援装置10を使用しているユーザのうち、発話者であるユーザAに対するフィードバックの一例を示している。
<Feedback for user A who is a speaker>
FIG. 25 shows an example of feedback to the user A who is the speaker among the users who are using the conversation support device 10.
 例えば、表示部43における発話テキストの表示が一杯になった場合、フィードバック制御部40に制御により、発話者であるユーザAに対して、例えば「Slow down」、「画面が一杯になりました」、「ゆっくり話してください」、「待ってください」、「一度区切ってください」、「未読があります」等の発話速度を緩めるように通知するフィードバックが、スマートフォン50等を用いた文字表示や音声出力によって行われる。 For example, when the display of the utterance text on the display unit 43 is full, the feedback control unit 40 controls the user A who is the speaker, for example, “SlowSdown”, “The screen is full”. , "Please speak slowly", "Please wait", "Please divide once", "Feed unread", etc. feedback to inform you to slow down the speaking rate, text display and voice output using smartphone 50 etc. Is done by.
 なお、ユーザAの発話速度や発話の区切りの長短に応じたインジケータを画面に表示したり、アラーム音等を出力したりするようにしてもよい。 It should be noted that an indicator corresponding to the utterance speed of the user A and the length of the utterance break may be displayed on the screen, or an alarm sound or the like may be output.
 そして、ユーザAが、音声認識や画面表示に対して最適な速度や区切りで発話している場合には、ユーザAに対してポイントを付与し、付与されたポイントに応じてユーザAが何らかのサービス特典やランキングを得られるようにしてもよい。 When the user A speaks at an optimum speed or segmentation for voice recognition or screen display, the user A is given points, and the user A performs some service according to the given points. Benefits and rankings may be obtained.
 <他の応用例>
 本実施の形態においては、会話支援装置10を、聴力に不安を持たないユーザAと、聴力に不安を持つユーザBとの間の会話を支援する用途で用いるようにしたが、本技術は、例えば、使用する言語が異なる人どうしの会話を支援する用途に応用することができる。その場合、音声認識処理の後、翻訳処理を行うようにすればよい。
<Other application examples>
In the present embodiment, the conversation support device 10 is used for the purpose of supporting the conversation between the user A who is not anxious about hearing and the user B who is uneasy about hearing. For example, the present invention can be applied to applications that support conversations between people using different languages. In that case, a translation process may be performed after the voice recognition process.
 また、会話支援装置10により、ユーザAが発話するときの口元を動画像として撮像しておき、発話テキストを表示するとともに、ユーザAの口元の動画像を表示するようにしてもよい。この場合、発話テキストの表示と、ユーザAの口元の動画像の動きを同期させて表示するようにしてもよい。このようにした場合、会話支援装置10を、例えば、読唇術の学習の用いることができる。 Alternatively, the conversation support device 10 may capture the mouth when the user A speaks as a moving image, display the utterance text, and display the moving image of the user A's mouth. In this case, the display of the utterance text and the motion of the moving image of the user A's mouth may be displayed in synchronization. In this case, the conversation support device 10 can be used for learning lip reading, for example.
 また、会話支援装置10により、ユーザAの発話を録音するとともに、その音声認識結果である発話テキストを対応付けて保存するようにし、この保存結果を後程改めて再生、表示できるようにしてもよい。 Also, the conversation support device 10 may record the utterance of the user A and store the utterance text that is the voice recognition result in association with the utterance text so that the saved result can be reproduced and displayed later.
 さらに、会話支援装置10に対して、ユーザAによるリアルタイムの発話を入力するだけでなく、録音されている音声を入力できるようにしてもよい。 Furthermore, not only a real-time utterance by the user A but also a recorded voice may be input to the conversation support device 10.
 <情報処理部30の他の構成例>
 上述した一連の処理は、ハードウェアにより実行することもできるし、ソフトウェアにより実行することもできる。一連の処理をソフトウェアにより実行する場合には、そのソフトウェアを構成するプログラムが、コンピュータにインストールされる。ここで、コンピュータには、専用のハードウェアに組み込まれているコンピュータや、各種のプログラムをインストールすることで、各種の機能を実行することが可能な、例えば汎用のコンピュータなどが含まれる。上記した第2の構成例におけるスマートフォン50は、該コンピュータに相当する。
<Another configuration example of the information processing unit 30>
The series of processes described above can be executed by hardware or can be executed by software. When a series of processing is executed by software, a program constituting the software is installed in the computer. Here, the computer includes, for example, a general-purpose computer capable of executing various functions by installing a computer incorporated in dedicated hardware and various programs. The smartphone 50 in the second configuration example described above corresponds to the computer.
 図26は、上述した一連の処理をプログラムにより実行するコンピュータのハードウェアの構成例を示すブロック図である。 FIG. 26 is a block diagram showing an example of the hardware configuration of a computer that executes the above-described series of processing by a program.
 このコンピュータ200において、CPU(Central Processing Unit)201,ROM(Read Only Memory)202,RAM(Random Access Memory)203は、バス204により相互に接続されている。 In this computer 200, a CPU (Central Processing Unit) 201, a ROM (Read Only Memory) 202, and a RAM (Random Access Memory) 203 are connected to each other by a bus 204.
 バス204には、さらに、入出力インタフェース205が接続されている。入出力インタフェース205には、入力部206、出力部207、記憶部208、通信部209、およびドライブ210が接続されている。 An input / output interface 205 is further connected to the bus 204. An input unit 206, an output unit 207, a storage unit 208, a communication unit 209, and a drive 210 are connected to the input / output interface 205.
 入力部206は、キーボード、マウス、マイクロフォンなどよりなる。出力部207は、ディスプレイ、スピーカなどよりなる。記憶部208は、ハードディスクや不揮発性のメモリなどよりなる。通信部209は、ネットワークインタフェースなどよりなる。ドライブ210は、磁気ディスク、光ディスク、光磁気ディスク、又は半導体メモリなどのリムーバブルメディア211を駆動する。 The input unit 206 includes a keyboard, a mouse, a microphone, and the like. The output unit 207 includes a display, a speaker, and the like. The storage unit 208 includes a hard disk, a nonvolatile memory, and the like. The communication unit 209 includes a network interface and the like. The drive 210 drives a removable medium 211 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory.
 以上のように構成されるコンピュータ200では、CPU201が、例えば、記憶部208に記憶されているプログラムを、入出力インタフェース205およびバス204を介して、RAM203にロードして実行することにより、上述した一連の処理が行われる。 In the computer 200 configured as described above, for example, the CPU 201 loads the program stored in the storage unit 208 to the RAM 203 via the input / output interface 205 and the bus 204 and executes the program. A series of processing is performed.
 コンピュータ200(CPU201)が実行するプログラムは、例えば、パッケージメディア等としてのリムーバブルメディア211に記録して提供することができる。また、プログラムは、ローカルエリアネットワーク、インターネット、デジタル衛星放送といった、有線または無線の伝送媒体を介して提供することができる。 The program executed by the computer 200 (CPU 201) can be provided by being recorded in, for example, a removable medium 211 such as a package medium. The program can be provided via a wired or wireless transmission medium such as a local area network, the Internet, or digital satellite broadcasting.
 コンピュータ200では、プログラムは、リムーバブルメディア211をドライブ210に装着することにより、入出力インタフェース205を介して、記憶部208にインストールすることができる。また、プログラムは、有線または無線の伝送媒体を介して、通信部209で受信し、記憶部208にインストールすることができる。その他、プログラムは、ROM202や記憶部208に、あらかじめインストールしておくことができる。 In the computer 200, the program can be installed in the storage unit 208 via the input / output interface 205 by attaching the removable medium 211 to the drive 210. The program can be received by the communication unit 209 via a wired or wireless transmission medium and installed in the storage unit 208. In addition, the program can be installed in the ROM 202 or the storage unit 208 in advance.
 なお、コンピュータ200が実行するプログラムは、本明細書で説明する順序に沿って時系列に処理が行われるプログラムであってもよいし、並列に、あるいは呼び出しが行われたとき等の必要なタイミングで処理が行われるプログラムであってもよい。 Note that the program executed by the computer 200 may be a program that is processed in time series in the order described in this specification, or a necessary timing such as when a call is made in parallel. It may be a program in which processing is performed.
 なお、本技術の実施の形態は、上述した実施の形態に限定されるものではなく、本技術の要旨を逸脱しない範囲において種々の変更が可能である。 Note that the embodiments of the present technology are not limited to the above-described embodiments, and various modifications can be made without departing from the gist of the present technology.
 本技術は以下のような構成も取ることができる。
(1)
 音声入力装置に入力された第1のユーザの音声情報を取得する音声取得部と、
 第2のユーザのための表示装置における、取得された前記音声情報に対応するテキスト情報の表示を制御する表示制御部と
 を備え、
 前記表示制御部は、前記表示装置における前記テキスト情報の表示量、または前記音声入力装置から入力された前記音声情報の入力量の少なくとも一方に基づいて、前記テキスト情報の表示量に関する制御を行う
 情報処理装置。
(2)
 前記表示制御部は、前記テキスト情報の表示量が所定の量以上となった場合、前記テキスト情報の表示量を抑制する
 前記(1)に記載の情報処理装置。
(3)
 前記表示制御部は、前記テキスト情報に含まれる所定の品詞の表示量を抑制することによって、前記テキスト情報の表示量を抑制する
 前記(1)または(2)に記載の情報処理装置。
(4)
 前記表示制御部は、前記第1のユーザまたは前記第2のユーザによる所定の操作に基づいて、前記テキスト情報の表示量を抑制する
 前記(1)から(3)のいずれかに記載の情報処理装置。
(5)
 前記所定の操作は、前記第1のユーザまたは前記第2のユーザによる第1の操作を含み、
 前記表示制御部は、前記テキスト情報の表示量を抑制した後、前記第1の操作に基づいて、前記テキスト情報の表示を消去させる
 前記(4)に記載の情報処理装置。
(6)
 前記所定の操作は、前記第1のユーザまたは前記第2のユーザによる第2の操作を含み、
 前記表示制御部は、前記テキスト情報の表示を消去させた後、前記第2の操作に基づいて、前記表示装置において消去させた前記テキスト情報を再び表示させる
 前記(5)に記載の情報処理装置。
(7)
 前記表示制御部は、前記テキスト情報の解析結果に従い、前記テキスト情報の表示の改行または改頁の少なくとも一方を制御する
 前記(1)から(6)のいずれかに記載の情報処理装置。
(8)
 前記第1のユーザまたは前記第2のユーザの一方が前記テキスト情報に関する操作を行った場合、前記テキスト情報に関する操作が行われたことを示す情報を、前記第1のユーザまたは前記第2のユーザの他方に対して通知する通知部をさらに備える
 前記(1)から(7)のいずれかに記載の情報処理装置。
(9)
 前記通知部は、前記第1のユーザまたは前記第2のユーザの一方が、前記テキスト情報の表示量を抑制させる操作を行った場合、前記第1のユーザまたは前記第2のユーザの他方に、前記テキスト情報の表示量が抑制されたことを通知する
 前記(8)に記載の情報処理装置。
(10)
 前記通知部は、前記第1のユーザまたは前記第2のユーザの一方が、前記テキスト情報の表示を消去する操作を行った場合、前記第1のユーザまたは前記第2のユーザの他方に、前記テキスト情報の表示が消去されたことを通知する
 前記(8)または(9)に記載の情報処理装置。
(11)
 前記通知部は、前記第2のユーザが、前記表示装置に表示された前記テキスト情報の再発話を要求する操作を行った場合、前記第1のユーザに再発話を促す通知を行う
 前記(8)から(10)のいずれかに記載の情報処理装置。
(12)
 前記通知部は、前記第2のユーザが、前記表示装置に表示された前記テキスト情報に関する問い合わせを要求するための操作を行った場合、前記第1のユーザに前記テキスト情報に関する問い合わせがあったことを通知する
 前記(8)から(11)のいずれかに記載の情報処理装置。
(13)
 前記表示制御部は、前記第2のユーザの発声または動作の少なくとも一方に基づく前記第2のユーザの既読検知の結果に基づいて、前記表示装置における前記テキスト情報の表示量を抑制する
 前記(1)から(12)のいずれかに記載の情報処理装置。
(14)
 前記表示制御部は、前記第1のユーザの発声または動作の少なくとも一方に基づき、前記表示装置における前記テキスト情報の表示を中止する
 前記(1)から(13)のいずれかに記載の情報処理装置。
(15)
 前記表示装置における前記テキスト情報の表示量または前記音声情報の入力量の少なくとも一方に基づき、前記第1のユーザまたは前記第2のユーザの少なくとも一方に対するフィードバック情報の通知を制御するフィードバック制御部をさらに備える
 前記(1)から(14)のいずれかに記載の情報処理装置。
(16)
 フィードバック情報は、前記第1のユーザに対して、発話速度、または発話区切りの少なくとも一方を変更するように促す情報である
 前記(15)に記載の情報処理装置。
(17)
 フィードバック情報は、前記第2のユーザに対して、前記表示装置に表示された前記テキスト情報の読み取りを促す情報である
 前記(15)または(16)に記載の情報処理装置。
(18)
 前記第1のユーザの前記音声情報を前記テキスト情報に変換する音声認識部をさらに備え、
 前記音声認識部は、前記情報処理装置の内部、または、インターネットを介して接続するサーバ上に設けられている
 前記(1)から(17)のいずれかに記載の情報処理装置。
(19)
 情報処理装置の情報処理方法において、
 前記情報処理装置による、
 音声入力装置に入力された第1のユーザの音声情報を取得する音声取得ステップと、
 第2のユーザのための表示装置における、取得された前記音声情報に対応するテキスト情報の表示を制御する表示制御ステップと
 を含み、
 前記表示制御ステップは、前記表示装置における前記テキスト情報の表示量、または前記音声入力装置から入力された前記音声情報の入力量の少なくとも一方に基づいて、前記テキスト情報の表示量に関する制御を行う
 情報処理方法。
(20)
 第1のユーザの音声情報を取得する音声入力装置と、
 取得された前記音声情報に対応するテキスト情報の表示を制御する表示制御装置と、
 前記表示制御装置からの制御に従い、前記テキスト情報を第2のユーザのために表示する表示装置と
 を備え、
 前記表示制御装置は、前記表示装置における前記テキスト情報の表示量、または前記音声入力装置から入力された前記音声情報の入力量の少なくとも一方に基づいて、前記テキスト情報の表示量に関する制御を行う
 情報処理システム。
The present technology can also have the following configurations.
(1)
A voice acquisition unit that acquires the voice information of the first user input to the voice input device;
A display control unit for controlling display of text information corresponding to the acquired voice information in a display device for a second user,
The display control unit controls the display amount of the text information based on at least one of the display amount of the text information on the display device or the input amount of the voice information input from the voice input device. Processing equipment.
(2)
The information processing apparatus according to (1), wherein the display control unit suppresses a display amount of the text information when a display amount of the text information becomes a predetermined amount or more.
(3)
The information processing apparatus according to (1) or (2), wherein the display control unit suppresses a display amount of the text information by suppressing a display amount of a predetermined part of speech included in the text information.
(4)
The information processing unit according to any one of (1) to (3), wherein the display control unit suppresses a display amount of the text information based on a predetermined operation by the first user or the second user. apparatus.
(5)
The predetermined operation includes a first operation by the first user or the second user,
The information processing apparatus according to (4), wherein the display control unit deletes the display of the text information based on the first operation after suppressing the display amount of the text information.
(6)
The predetermined operation includes a second operation by the first user or the second user,
The information processing apparatus according to (5), wherein after the display of the text information is erased, the display control unit displays the text information erased in the display device again based on the second operation. .
(7)
The information processing apparatus according to any one of (1) to (6), wherein the display control unit controls at least one of a line feed or a page break of the text information display according to the analysis result of the text information.
(8)
When one of the first user or the second user performs an operation related to the text information, information indicating that an operation related to the text information has been performed is used as information indicating the first user or the second user. The information processing apparatus according to any one of (1) to (7), further including a notification unit that notifies the other of the information.
(9)
When one of the first user or the second user performs an operation of suppressing the amount of display of the text information, the notification unit may notify the other of the first user or the second user, The information processing apparatus according to (8), which notifies that the display amount of the text information is suppressed.
(10)
When one of the first user or the second user performs an operation of deleting the display of the text information, the notification unit may notify the other of the first user or the second user, The information processing apparatus according to (8) or (9), wherein the information indicating that the display of the text information has been deleted is notified.
(11)
When the second user performs an operation for requesting re-speech of the text information displayed on the display device, the notification unit performs a notification to prompt the re-speech to the first user. ) To (10).
(12)
When the second user performs an operation for requesting an inquiry about the text information displayed on the display device, the notification unit has received an inquiry about the text information from the first user. The information processing apparatus according to any one of (8) to (11).
(13)
The display control unit suppresses a display amount of the text information on the display device based on a result of the second user's read detection based on at least one of the second user's utterance or action. The information processing apparatus according to any one of 1) to (12).
(14)
The information processing apparatus according to any one of (1) to (13), wherein the display control unit stops displaying the text information on the display device based on at least one of the utterance or action of the first user. .
(15)
A feedback control unit that controls notification of feedback information to at least one of the first user and the second user based on at least one of a display amount of the text information and an input amount of the audio information on the display device; The information processing apparatus according to any one of (1) to (14).
(16)
The information processing apparatus according to (15), wherein the feedback information is information that prompts the first user to change at least one of an utterance speed and an utterance break.
(17)
The information processing apparatus according to (15) or (16), wherein the feedback information is information that prompts the second user to read the text information displayed on the display device.
(18)
A voice recognition unit that converts the voice information of the first user into the text information;
The information processing apparatus according to any one of (1) to (17), wherein the voice recognition unit is provided inside the information processing apparatus or on a server connected via the Internet.
(19)
In the information processing method of the information processing apparatus,
According to the information processing apparatus,
A voice acquisition step of acquiring voice information of the first user input to the voice input device;
A display control step for controlling display of text information corresponding to the acquired voice information in a display device for a second user,
The display control step performs control related to the display amount of the text information based on at least one of the display amount of the text information on the display device or the input amount of the voice information input from the voice input device. Processing method.
(20)
A voice input device for acquiring voice information of the first user;
A display control device for controlling display of text information corresponding to the acquired voice information;
A display device for displaying the text information for a second user in accordance with control from the display control device;
The display control device performs control related to a display amount of the text information based on at least one of a display amount of the text information on the display device or an input amount of the voice information input from the voice input device. Processing system.
 10 会話支援装置, 21 集音部, 22 表示部, 23 操作入力部, 30 情報処理部, 31 音声認識部, 32 画像認識部, 33 誤認識学習部, 34 誤認識リスト, 35 解析部, 36 編集部, 37 追記学習部, 38 表示待ちリスト保持部, 39 表示制御部, 40 フィードバック制御部, 41 撮像部, 42 集音部, 43 表示部, 44 操作入力部, 50 スマートフォン, 60 タブレット型PC, 80 プロジェクタ, 90 TV, 100 首掛けマイク, 110 カメラ, 111 消去ボタン, 112 再発話ボタン, 113 NGワード登録ボタン, 114 追記ボタン, 200 コンピュータ, 201 CPU 10 conversation support devices, 21 sound collection unit, 22 display unit, 23 operation input unit, 30 information processing unit, 31 voice recognition unit, 32 image recognition unit, 33 misrecognition learning unit, 34 misrecognition list, 35 analysis unit, 36 Editing unit, 37 additional learning unit, 38 display waiting list holding unit, 39 display control unit, 40 feedback control unit, 41 imaging unit, 42 sound collection unit, 43 display unit, 44 operation input unit, 50 smartphone, 60 tablet PC , 80 projector, 90 TV, 100 head microphone, 110 camera, 111 erase button, 112 recurrence button, 113 NG word registration button, 114 additional button, 200 computer, 201 CPU

Claims (20)

  1.  音声入力装置に入力された第1のユーザの音声情報を取得する音声取得部と、
     第2のユーザのための表示装置における、取得された前記音声情報に対応するテキスト情報の表示を制御する表示制御部と
     を備え、
     前記表示制御部は、前記表示装置における前記テキスト情報の表示量、または前記音声入力装置から入力された前記音声情報の入力量の少なくとも一方に基づいて、前記テキスト情報の表示量に関する制御を行う
     情報処理装置。
    A voice acquisition unit that acquires the voice information of the first user input to the voice input device;
    A display control unit for controlling display of text information corresponding to the acquired voice information in a display device for a second user,
    The display control unit controls the display amount of the text information based on at least one of the display amount of the text information on the display device or the input amount of the voice information input from the voice input device. Processing equipment.
  2.  前記表示制御部は、前記テキスト情報の表示量が所定の量以上となった場合、前記テキスト情報の表示量を抑制する
     請求項1に記載の情報処理装置。
    The information processing apparatus according to claim 1, wherein the display control unit suppresses a display amount of the text information when a display amount of the text information becomes a predetermined amount or more.
  3.  前記表示制御部は、前記テキスト情報に含まれる所定の品詞の表示量を抑制することによって、前記テキスト情報の表示量を抑制する
     請求項2に記載の情報処理装置。
    The information processing apparatus according to claim 2, wherein the display control unit suppresses a display amount of the text information by suppressing a display amount of a predetermined part of speech included in the text information.
  4.  前記表示制御部は、前記第1のユーザまたは前記第2のユーザによる所定の操作に基づいて、前記テキスト情報の表示量を抑制する
     請求項2に記載の情報処理装置。
    The information processing apparatus according to claim 2, wherein the display control unit suppresses a display amount of the text information based on a predetermined operation by the first user or the second user.
  5.  前記所定の操作は、前記第1のユーザまたは前記第2のユーザによる第1の操作を含み、
     前記表示制御部は、前記テキスト情報の表示量を抑制した後、前記第1の操作に基づいて、前記テキスト情報の表示を消去させる
     請求項4に記載の情報処理装置。
    The predetermined operation includes a first operation by the first user or the second user,
    The information processing apparatus according to claim 4, wherein the display control unit deletes the display of the text information based on the first operation after suppressing the display amount of the text information.
  6.  前記所定の操作は、前記第1のユーザまたは前記第2のユーザによる第2の操作を含み、
     前記表示制御部は、前記テキスト情報の表示を消去させた後、前記第2の操作に基づいて、前記表示装置において消去させた前記テキスト情報を再び表示させる
     請求項5に記載の情報処理装置。
    The predetermined operation includes a second operation by the first user or the second user,
    The information processing apparatus according to claim 5, wherein the display control unit displays the text information erased in the display device again based on the second operation after erasing the display of the text information.
  7.  前記表示制御部は、前記テキスト情報の解析結果に従い、前記テキスト情報の表示の改行または改頁の少なくとも一方を制御する
     請求項2に記載の情報処理装置。
    The information processing apparatus according to claim 2, wherein the display control unit controls at least one of a line feed or a page break of the text information display according to the analysis result of the text information.
  8.  前記第1のユーザまたは前記第2のユーザの一方が前記テキスト情報に関する操作を行った場合、前記テキスト情報に関する操作が行われたことを示す情報を、前記第1のユーザまたは前記第2のユーザの他方に対して通知する通知部をさらに備える
     請求項1に記載の情報処理装置。
    When one of the first user or the second user performs an operation related to the text information, information indicating that an operation related to the text information has been performed is used as information indicating the first user or the second user. The information processing apparatus according to claim 1, further comprising a notification unit that notifies the other of the information.
  9.  前記通知部は、前記第1のユーザまたは前記第2のユーザの一方が、前記テキスト情報の表示量を抑制させる操作を行った場合、前記第1のユーザまたは前記第2のユーザの他方に、前記テキスト情報の表示量が抑制されたことを通知する
     請求項8に記載の情報処理装置。
    When one of the first user or the second user performs an operation of suppressing the amount of display of the text information, the notification unit may notify the other of the first user or the second user, The information processing apparatus according to claim 8, notifying that a display amount of the text information is suppressed.
  10.  前記通知部は、前記第1のユーザまたは前記第2のユーザの一方が、前記テキスト情報の表示を消去する操作を行った場合、前記第1のユーザまたは前記第2のユーザの他方に、前記テキスト情報の表示が消去されたことを通知する
     請求項8に記載の情報処理装置。
    When one of the first user or the second user performs an operation of deleting the display of the text information, the notification unit may notify the other of the first user or the second user, The information processing apparatus according to claim 8, which notifies that the display of text information has been erased.
  11.  前記通知部は、前記第2のユーザが、前記表示装置に表示された前記テキスト情報の再発話を要求する操作を行った場合、前記第1のユーザに再発話を促す通知を行う
     請求項8に記載の情報処理装置。
    The notification unit, when the second user performs an operation for requesting a re-speech of the text information displayed on the display device, performs a notification that prompts the first user to re-speak. The information processing apparatus described in 1.
  12.  前記通知部は、前記第2のユーザが、前記表示装置に表示された前記テキスト情報に関する問い合わせを要求するための操作を行った場合、前記第1のユーザに前記テキスト情報に関する問い合わせがあったことを通知する
     請求項8に記載の情報処理装置。
    When the second user performs an operation for requesting an inquiry about the text information displayed on the display device, the notification unit has received an inquiry about the text information from the first user. The information processing apparatus according to claim 8.
  13.  前記表示制御部は、前記第2のユーザの発声または動作の少なくとも一方に基づく前記第2のユーザの既読検知の結果に基づいて、前記表示装置における前記テキスト情報の表示量を抑制する
     請求項1に記載の情報処理装置。
    The display control unit suppresses a display amount of the text information on the display device based on a result of the second user's read detection based on at least one of the second user's utterance or action. The information processing apparatus according to 1.
  14.  前記表示制御部は、前記第1のユーザの発声または動作の少なくとも一方に基づき、前記表示装置における前記テキスト情報の表示を中止する
     請求項1に記載の情報処理装置。
    The information processing apparatus according to claim 1, wherein the display control unit stops displaying the text information on the display device based on at least one of the first user's utterance or action.
  15.  前記表示装置における前記テキスト情報の表示量または前記音声情報の入力量の少なくとも一方に基づき、前記第1のユーザまたは前記第2のユーザの少なくとも一方に対するフィードバック情報の通知を制御するフィードバック制御部をさらに備える
     請求項1に記載の情報処理装置。
    A feedback control unit that controls notification of feedback information to at least one of the first user and the second user based on at least one of a display amount of the text information and an input amount of the audio information on the display device; The information processing apparatus according to claim 1.
  16.  フィードバック情報は、前記第1のユーザに対して、発話速度、または発話区切りの少なくとも一方を変更するように促す情報である
     請求項15に記載の情報処理装置。
    The information processing apparatus according to claim 15, wherein the feedback information is information prompting the first user to change at least one of an utterance speed and an utterance break.
  17.  フィードバック情報は、前記第2のユーザに対して、前記表示装置に表示された前記テキスト情報の読み取りを促す情報である
     請求項15に記載の情報処理装置。
    The information processing apparatus according to claim 15, wherein the feedback information is information that prompts the second user to read the text information displayed on the display device.
  18.  前記第1のユーザの前記音声情報を前記テキスト情報に変換する音声認識部をさらに備え、
     前記音声認識部は、前記情報処理装置の内部、または、インターネットを介して接続するサーバ上に設けられている
     請求項1に記載の情報処理装置。
    A voice recognition unit that converts the voice information of the first user into the text information;
    The information processing apparatus according to claim 1, wherein the voice recognition unit is provided inside the information processing apparatus or on a server connected via the Internet.
  19.  情報処理装置の情報処理方法において、
     前記情報処理装置による、
     音声入力装置に入力された第1のユーザの音声情報を取得する音声取得ステップと、
     第2のユーザのための表示装置における、取得された前記音声情報に対応するテキスト情報の表示を制御する表示制御ステップと
     を含み、
     前記表示制御ステップは、前記表示装置における前記テキスト情報の表示量、または前記音声入力装置から入力された前記音声情報の入力量の少なくとも一方に基づいて、前記テキスト情報の表示量に関する制御を行う
     情報処理方法。
    In the information processing method of the information processing apparatus,
    According to the information processing apparatus,
    A voice acquisition step of acquiring voice information of the first user input to the voice input device;
    A display control step for controlling display of text information corresponding to the acquired voice information in a display device for a second user,
    The display control step performs control related to the display amount of the text information based on at least one of the display amount of the text information on the display device or the input amount of the voice information input from the voice input device. Processing method.
  20.  第1のユーザの音声情報を取得する音声入力装置と、
     取得された前記音声情報に対応するテキスト情報の表示を制御する表示制御装置と、
     前記表示制御装置からの制御に従い、前記テキスト情報を第2のユーザのために表示する表示装置と
     を備え、
     前記表示制御装置は、前記表示装置における前記テキスト情報の表示量、または前記音声入力装置から入力された前記音声情報の入力量の少なくとも一方に基づいて、前記テキスト情報の表示量に関する制御を行う
     情報処理システム。
    A voice input device for acquiring voice information of the first user;
    A display control device for controlling display of text information corresponding to the acquired voice information;
    A display device for displaying the text information for a second user in accordance with control from the display control device;
    The display control device performs control related to a display amount of the text information based on at least one of a display amount of the text information on the display device or an input amount of the voice information input from the voice input device. Processing system.
PCT/JP2017/041758 2016-12-05 2017-11-21 Information processing device, information processing method, and information processing system WO2018105373A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
JP2018554906A JP6950708B2 (en) 2016-12-05 2017-11-21 Information processing equipment, information processing methods, and information processing systems
US16/349,731 US11189289B2 (en) 2016-12-05 2017-11-21 Information processing device, information processing method, and information processing system
KR1020197014972A KR20190091265A (en) 2016-12-05 2017-11-21 Information processing apparatus, information processing method, and information processing system
DE112017006145.8T DE112017006145T5 (en) 2016-12-05 2017-11-21 INFORMATION PROCESSING DEVICE, INFORMATION PROCESSING METHOD AND INFORMATION PROCESSING SYSTEM

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US201662430000P 2016-12-05 2016-12-05
US62/430,000 2016-12-05
JP2017-074369 2017-04-04
JP2017074369 2017-04-04

Publications (1)

Publication Number Publication Date
WO2018105373A1 true WO2018105373A1 (en) 2018-06-14

Family

ID=62491200

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2017/041758 WO2018105373A1 (en) 2016-12-05 2017-11-21 Information processing device, information processing method, and information processing system

Country Status (1)

Country Link
WO (1) WO2018105373A1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2020039014A (en) * 2018-08-31 2020-03-12 株式会社コロプラ Program, information processing device, and method
JP2020126195A (en) * 2019-02-06 2020-08-20 トヨタ自動車株式会社 Voice interactive device, control device for voice interactive device and control program
JP2022056592A (en) * 2020-09-30 2022-04-11 本田技研工業株式会社 Conversation support device, conversation support system, conversation support method, and program
WO2022270456A1 (en) * 2021-06-21 2022-12-29 ピクシーダストテクノロジーズ株式会社 Display control device, display control method, and program
JP7517366B2 (en) 2021-08-16 2024-07-17 株式会社リコー Voice recording management system, voice recording management device, voice recording management method and program
WO2024150633A1 (en) * 2023-01-13 2024-07-18 ソニーグループ株式会社 Information processing device, information processing method and information processing program

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH09127459A (en) * 1995-11-02 1997-05-16 Canon Inc Display device provided with gaze detection system
US6172685B1 (en) * 1997-11-24 2001-01-09 Intel Corporation Method and apparatus for increasing the amount and utility of displayed information
JP2008097104A (en) * 2006-10-06 2008-04-24 Sharp Corp Device for exchanging message information and its operation method
JP2013235556A (en) * 2012-05-07 2013-11-21 Lg Electronics Inc Method for displaying text associated with audio file, and electronic device implementing the same
JP2014164692A (en) * 2013-02-27 2014-09-08 Yahoo Japan Corp Document display device, document display method and document display program
JP2015069600A (en) * 2013-09-30 2015-04-13 株式会社東芝 Voice translation system, method, and program
WO2016103415A1 (en) * 2014-12-25 2016-06-30 日立マクセル株式会社 Head-mounted display system and operating method for head-mounted display device

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH09127459A (en) * 1995-11-02 1997-05-16 Canon Inc Display device provided with gaze detection system
US6172685B1 (en) * 1997-11-24 2001-01-09 Intel Corporation Method and apparatus for increasing the amount and utility of displayed information
JP2008097104A (en) * 2006-10-06 2008-04-24 Sharp Corp Device for exchanging message information and its operation method
JP2013235556A (en) * 2012-05-07 2013-11-21 Lg Electronics Inc Method for displaying text associated with audio file, and electronic device implementing the same
JP2014164692A (en) * 2013-02-27 2014-09-08 Yahoo Japan Corp Document display device, document display method and document display program
JP2015069600A (en) * 2013-09-30 2015-04-13 株式会社東芝 Voice translation system, method, and program
WO2016103415A1 (en) * 2014-12-25 2016-06-30 日立マクセル株式会社 Head-mounted display system and operating method for head-mounted display device

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2020039014A (en) * 2018-08-31 2020-03-12 株式会社コロプラ Program, information processing device, and method
JP2020126195A (en) * 2019-02-06 2020-08-20 トヨタ自動車株式会社 Voice interactive device, control device for voice interactive device and control program
JP7120060B2 (en) 2019-02-06 2022-08-17 トヨタ自動車株式会社 VOICE DIALOGUE DEVICE, CONTROL DEVICE AND CONTROL PROGRAM FOR VOICE DIALOGUE DEVICE
JP2022056592A (en) * 2020-09-30 2022-04-11 本田技研工業株式会社 Conversation support device, conversation support system, conversation support method, and program
JP7369110B2 (en) 2020-09-30 2023-10-25 本田技研工業株式会社 Conversation support device, conversation support system, conversation support method and program
WO2022270456A1 (en) * 2021-06-21 2022-12-29 ピクシーダストテクノロジーズ株式会社 Display control device, display control method, and program
JP7517366B2 (en) 2021-08-16 2024-07-17 株式会社リコー Voice recording management system, voice recording management device, voice recording management method and program
WO2024150633A1 (en) * 2023-01-13 2024-07-18 ソニーグループ株式会社 Information processing device, information processing method and information processing program

Similar Documents

Publication Publication Date Title
WO2018105373A1 (en) Information processing device, information processing method, and information processing system
JP6710740B2 (en) Providing suggested voice-based action queries
CN106463114B (en) Information processing apparatus, control method, and program storage unit
US10592095B2 (en) Instantaneous speaking of content on touch devices
US20210193146A1 (en) Multi-modal interaction between users, automated assistants, and other computing services
WO2016103988A1 (en) Information processing device, information processing method, and program
US11462213B2 (en) Information processing apparatus, information processing method, and program
US11200893B2 (en) Multi-modal interaction between users, automated assistants, and other computing services
US20120260176A1 (en) Gesture-activated input using audio recognition
KR102193029B1 (en) Display apparatus and method for performing videotelephony using the same
WO2019107145A1 (en) Information processing device and information processing method
WO2016152200A1 (en) Information processing system and information processing method
JP6950708B2 (en) Information processing equipment, information processing methods, and information processing systems
WO2017175442A1 (en) Information processing device and information processing method
US12125486B2 (en) Multi-modal interaction between users, automated assistants, and other computing services
KR20140111574A (en) Apparatus and method for performing an action according to an audio command
CN117971154A (en) Multimodal response
WO2015156011A1 (en) Information processing device, information processing method, and program
US11150923B2 (en) Electronic apparatus and method for providing manual thereof
US11430429B2 (en) Information processing apparatus and information processing method
JPWO2020116001A1 (en) Information processing device and information processing method
KR101508444B1 (en) Display device and method for executing hyperlink using the same
JP5613102B2 (en) CONFERENCE DEVICE, CONFERENCE METHOD, AND CONFERENCE PROGRAM
JP2019179081A (en) Conference support device, conference support control method, and program
WO2020158218A1 (en) Information processing device, information processing method, and program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17879093

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2018554906

Country of ref document: JP

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 20197014972

Country of ref document: KR

Kind code of ref document: A

122 Ep: pct application non-entry in european phase

Ref document number: 17879093

Country of ref document: EP

Kind code of ref document: A1