WO2019085585A1 - Device control processing method and apparatus - Google Patents

Device control processing method and apparatus Download PDF

Info

Publication number
WO2019085585A1
WO2019085585A1 PCT/CN2018/100489 CN2018100489W WO2019085585A1 WO 2019085585 A1 WO2019085585 A1 WO 2019085585A1 CN 2018100489 W CN2018100489 W CN 2018100489W WO 2019085585 A1 WO2019085585 A1 WO 2019085585A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
photo
level
sound
emotion
Prior art date
Application number
PCT/CN2018/100489
Other languages
French (fr)
Chinese (zh)
Inventor
刘质斌
王九飚
周文斌
石秋成
王红霞
王琳
Original Assignee
格力电器(武汉)有限公司
珠海格力电器股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 格力电器(武汉)有限公司, 珠海格力电器股份有限公司 filed Critical 格力电器(武汉)有限公司
Publication of WO2019085585A1 publication Critical patent/WO2019085585A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/28Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
    • H04L12/2803Home automation networks
    • H04L12/2805Home Audio Video Interoperability [HAVI] networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/63Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/28Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
    • H04L12/2803Home automation networks
    • H04L12/2816Controlling appliance services of a home automation network by calling their functionalities
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/28Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
    • H04L12/2803Home automation networks
    • H04L2012/2847Home automation networks characterised by the type of home appliance used
    • H04L2012/2849Audio/video appliances

Definitions

  • the present application relates to the field of smart homes, and in particular to a device control processing method and apparatus.
  • the embodiment of the present application provides a device control processing method and device, so as to at least solve the technical problem that the home system in the related art cannot meet the user's demand for the intelligence degree of the home system.
  • a device control processing method including: acquiring information of a user, where the information includes at least one of: a photo of the user captured by an imaging device, through audio a voice of the user received by the device; using a model to evaluate an emotional level of the user corresponding to the information, wherein the model is trained according to a plurality of sets of data, each set of data in the plurality of sets of data Each includes: a photo and/or a sound of the user, and a label for identifying an emotional level represented by the photo and/or sound; transmitting a control command according to the emotional level, wherein the control command is used to indicate the device Perform the scheduled operation.
  • the method further includes: transmitting a photo and/or a sound of the user to another user; acquiring the other user as The label of the user's photo and/or sound added.
  • acquiring the tag added by the other user for the photo and/or sound of the user includes at least one of: sending a photo and/or a sound of the user, and a plurality of emotion levels that can be selected to Receiving, by the other user, an emotion level selected by the other user from the plurality of emotion levels as the tag; acquiring an evaluation of the photo and/or voice of the user by the other user, from the evaluation The emotion level is extracted as the tag, wherein the evaluation includes at least one of: an evaluation of a natural language, an evaluation of a voice.
  • the method further includes: after obtaining the photo and/or sound of the user, according to the photo of the user and/or Or the voice makes a question to the user; the emotion level corresponding to the photo and/or sound of the user is extracted according to the user's answer to the question questioned.
  • sending the control command according to the emotion level comprises: sending the control command if the emotion level matches a predetermined level, wherein the control command is used to control the device to perform the following operations At least one of: playing music corresponding to the emotion level, playing a video corresponding to the emotion level.
  • a device control processing apparatus including: a first acquiring unit, configured to acquire information of a user, where the information includes at least one of the following: capturing by the imaging device a photo of the user, the voice of the user received through the audio device; an evaluation unit for evaluating a mood level of the user corresponding to the information using a model, wherein the model is trained according to multiple sets of data Obtaining, each of the plurality of sets of data includes: a photo and/or a sound of the user, and a label for identifying an emotional level represented by the photo and/or the sound; the first sending unit, And a method for transmitting a control command according to the emotion level, wherein the control command is used to instruct the device to perform a predetermined operation.
  • the device further includes: a second sending unit, configured to send the photo and/or sound of the user to other users before evaluating the emotion level of the user corresponding to the information by using the model a second obtaining unit, configured to acquire the label added by the other user for the photo and/or sound of the user.
  • a second sending unit configured to send the photo and/or sound of the user to other users before evaluating the emotion level of the user corresponding to the information by using the model
  • a second obtaining unit configured to acquire the label added by the other user for the photo and/or sound of the user.
  • the second obtaining unit includes at least one of: a first sending module, configured to send a photo and/or a sound of the user, and a plurality of selectable multiple emotion levels to the other user; Determining, by the other user, an emotion level selected from the plurality of emotion levels as the label; an extracting module, configured to obtain an evaluation of the photo and/or sound of the user by the other user, and extracting from the evaluation
  • the emotion level is the tag, wherein the evaluation includes at least one of: an evaluation of a natural language, an evaluation of a voice.
  • the device further includes: a questioning unit, configured to: after obtaining the photo and/or sound of the user, using the model to evaluate the emotion level of the user corresponding to the information, according to the The user's photo and/or voice makes a question to the user; and an extracting unit is configured to extract an emotion level corresponding to the photo and/or sound of the user according to the user's answer to the question questioned.
  • a questioning unit configured to: after obtaining the photo and/or sound of the user, using the model to evaluate the emotion level of the user corresponding to the information, according to the The user's photo and/or voice makes a question to the user
  • an extracting unit is configured to extract an emotion level corresponding to the photo and/or sound of the user according to the user's answer to the question questioned.
  • the first sending unit includes: a second sending module, configured to send the control command if the emotion level matches a predetermined level, where the control command is used to control the device Performing at least one of the following operations: playing music corresponding to the emotion level, and playing a video corresponding to the emotion level.
  • a storage medium comprising a stored program, wherein the program performs the device control processing method according to any one of the above.
  • processor configured to run a program, wherein the program is executed to perform the device control processing method according to any one of the above.
  • the information of the user may be acquired, where the information includes at least one of the following: a photo of the user captured by the imaging device, a voice of the user received by the audio device, and a user corresponding to the model evaluation information.
  • the tag transmits a control command according to the emotion level, wherein the control command is used to instruct the device to perform a predetermined operation.
  • the device control processing method provided by the embodiment of the present application achieves the purpose of controlling the smart home system according to the acquired user's emotion, and achieves the technical effect of letting the user experience the happiness brought by the modern technology and improving the quality of life, thereby solving the problem.
  • the home system does not satisfy the technical problem of the user's demand for the intelligence degree of the home system, and the user experience is improved.
  • FIG. 1 is a flowchart of a device control processing method according to an embodiment of the present application.
  • FIG. 2 is a flow chart of an adjustment mechanism of a smart home system according to an embodiment of the present application
  • FIG. 3 is a schematic diagram of a device control apparatus according to an embodiment of the present application.
  • Pixel is the smallest unit that can be displayed on the computer screen. It is used to represent the unit of the image. It refers to the array of horizontal and vertical pixels that can be displayed. The more pixels in the screen, the higher the resolution of the image, the more the image is. Delicate and realistic.
  • Pixel refers to the value of a pixel.
  • Binarization Most of the pictures taken by the camera are color images.
  • the color image contains a huge amount of information.
  • the color picture is processed first, so that the picture only has foreground information.
  • With the background information you can simply define the foreground information as black and the background information as white. This is the binarization graph.
  • Neural network algorithm refers to the process of reasoning according to logic rules. It first converts information into concepts and represents user symbols. Then, according to symbolic operations, it performs logical reasoning in serial mode; this process can be written as serial instructions. Let the computer execute.
  • Voiceprint It is a sound wave spectrum that is displayed with electro-acoustic instruments and carries speech information.
  • Voiceprint recognition It is a kind of biometric technology. Also known as speaker recognition, there are two types, namely speaker recognition and speaker confirmation. Different tasks and applications use different voiceprint recognition techniques. For example, when narrowing the scope of criminal investigation, it may be necessary to identify the technology. Confirm the technology.
  • the following embodiments may be used in various electrical devices, and the types of the electrical devices are not specifically limited, including but not limited to: a washing machine, an air conditioner, a refrigerator, etc., and the above various electrical devices constitute the smart home system in the embodiment of the present application.
  • the embodiments of the present application are described in detail below.
  • a method embodiment of a device control processing method is provided, and it should be noted that the steps shown in the flowchart of the accompanying drawings may be executed in a computer system such as a set of computer executable instructions, and Although the logical order is shown in the flowcharts, in some cases the steps shown or described may be performed in a different order than the ones described herein.
  • FIG. 1 is a flowchart of a device control processing method according to an embodiment of the present application. As shown in FIG. 1 , the device control processing method includes the following steps:
  • Step S102 Acquire information of the user, where the information includes at least one of the following: a photo of the user captured by the imaging device, and a voice of the user received by the audio device.
  • one or more cameras may be installed in the user's home for taking a photo of the user.
  • the setting position of the camera is not specifically limited, and may include, but is not limited to, a user.
  • the door, ceiling, etc. of each room in the home can be used to collect photos of the user through cameras installed in different locations.
  • the user's photo is taken by using the one or more cameras described above, the user may be photographed every predetermined time period, and then the emotion of the user in the image may be analyzed.
  • the category of the captured image is not specifically limited in the embodiment of the present application, and may include, but is not limited to, a black and white image (grayscale image) and a color image (RGB image).
  • the information in the image can be analyzed according to the binarized image processing manner. Specifically, when analyzing, multiple pixel points in the image can be compared with the pixel position in the historical image to determine the existence. The pixel points of the difference are then distinguished from the pixel points where the difference exists, and the user's information can be extracted from the image captured by the imaging device.
  • one or more audio devices may be installed in the user's home for receiving the voice of the user.
  • the installation location of the audio device is not performed.
  • the specific limitation may include, but is not limited to, a position at a doorway, a ceiling, and the like of each room in the user's home.
  • the audio device may be installed at a position where the user is frequently active and the position is equal to the height of the human body.
  • the voice device includes a voice model library, where the voice model library stores voiceprints of each member of the family.
  • Each member of a family can issue a voice to the voice device, and the voice device can perform feature extraction to store voiceprints of different members into the voice model library.
  • the voice device can extract the feature of the member's voice, obtain the voiceprint of the member, and then match the voiceprint of the member with the voiceprint stored in the voice model library. And then identifying the family member corresponding to the voiceprint, and then obtaining the information corresponding to the member.
  • Step S104 using the model to evaluate the emotional level of the user corresponding to the information, wherein the model is obtained according to the plurality of sets of data, each of the plurality of sets of data includes: a photo and/or a sound of the user, and is used to identify The photo and/or voice represents the level of the emotional level of the label.
  • the above model may be an image captured by the camera during a predetermined period of time, a voice of the user within a predetermined period of time received by the audio device, and a label for identifying an emotional level represented by the photo and/or sound. Learned by training.
  • Step S106 sending a control command according to the emotion level, wherein the control command is used to instruct the device to perform a predetermined operation.
  • the model when the smart home system is running, the model can be used to evaluate the user's information to obtain the corresponding user's emotional level, and then send a control command to the smart home system according to the evaluated emotional level.
  • the device control processing method provided by the embodiment of the present application achieves the purpose of controlling the smart home system according to the acquired user's emotion, and achieves the technical effect of letting the user experience the happiness brought by the modern technology and improving the quality of life, thereby solving the problem.
  • the home system does not satisfy the technical problem of the user's demand for the intelligence degree of the home system, and the user experience is improved.
  • the camera captures the user's current facial expressions and limb movements, and combines image recognition technology to record the user's facial expressions and limb movements through the neural network.
  • the algorithm compares, judges, feedbacks, and learns; at the same time, the sound sensor is used to receive the user's voice, and the user's voice changes are recorded, and the neural network algorithm is used to compare, judge, feedback, and learn.
  • f facial expression, limb movement, sound ... external stimulation
  • independent variables facial expressions, limb movements, sounds etc.
  • dependent variable the current emotional level of the user; here different degrees of facial expressions, physical movements, sounds... external stimuli, etc., corresponding to different current emotional levels of the user.
  • the label used to identify the emotional level represented by the photo and/or sound can be obtained from multiple aspects.
  • the device control processing method may further include: sending the user's photo and/or voice to other users; Get tags added by other users to the user's photos and/or sounds.
  • acquiring a tag added by other users for the user's photo and/or sound may include at least one of: transmitting the user's photo and/or sound, and a plurality of selectable emotion levels to other users; receiving Other users select the emotion level from the plurality of emotion levels as a label; obtain other users' evaluations of the user's photos and/or sounds, and extract the emotion level as a label from the evaluation, wherein the evaluation includes at least one of the following: natural language Evaluation, speech evaluation.
  • the smart home system will send the user's photo or voice to the user's relatives and friends (that is, other users in the context), and the user's friends and relatives will receive the user's photo or sound and
  • the emotions of the user's historical time period are compared, and then the user's photos and/or sounds are tagged, for example, the user is depressed due to work pressure, or is troubled by unhappy things outside, etc.
  • the user's friends and relatives can also directly evaluate the received photo and/or voice of the user, and the smart home system extracts the emotion level as a label from the evaluation of the user's relatives and friends.
  • the evaluation may include, but is not limited to, an evaluative text (ie, a natural language evaluation in context) sent by the user's friends and relatives, and an evaluative speech (ie, an evaluation of the voice in the context).
  • the device control processing method may further include: after obtaining the photo and/or sound of the user, asking the user according to the photo and/or sound of the user.
  • the emotion level corresponding to the user's photo and/or voice is extracted according to the user's answer to the question questioned.
  • the smart home system can obtain the user's emotional level through a conversational manner.
  • the smart home system After receiving the user's photo and/or voice, the smart home system will ask the user a question, if the smart home system asks a question: "Today "How the mood”; the user replies: "The work pressure is too big, more irritating”; then the smart home system will extract the user's corresponding emotional level from the user's answer.
  • the smart home system can also perform self-correction in order to better understand the user and serve the user. For example, if the user's emotional level is obtained by the user's friends and relatives, and the user's emotional level is happy by the conversational manner, the smart home system combines the above-mentioned user's emotional level obtained from other people and The level of emotions obtained from the users themselves through dialogue, judgment of deviations, and corrections, continuous learning and improvement.
  • sending the control command according to the emotion level may include: sending a control command when the emotion level matches the predetermined level, wherein the control command is used to control the device to perform at least the following operations One: Play music corresponding to the emotional level, and play the video corresponding to the emotional level.
  • the user opens the smart home system, which observes and records the daily life of the user, interacts with the user, compares, judges, feeds back, and learns the user's emotions; wherein the communication mode may include but is not limited to: voice and system dialogue Such as inner monologues, transcripts, facial expressions, and body movements.
  • voice and system dialogue Such as inner monologues, transcripts, facial expressions, and body movements.
  • the smart home system will adjust the user's emotions according to their own decisions through external means (for example, playing music, videos, etc.).
  • the above-mentioned smart home system can also find out the cause of the user's negative emotion by querying, and then can take targeted measures to alleviate the emotion of the user. For example, when the historical home mood similar to the current mood of the user is stored in the smart home system, the historical solution corresponding to the historical emotion may be searched in the smart home system, and then the historical solution may be referred to or directly used to alleviate the user. Emotions. Wherein, in the above smart home system, when the historical mood similar to the current mood of the user is not found, the smart home system does not store a reference solution for solving the current mood of the user.
  • the smart home system can The network searches for similar emotions that the user's current emotional similarity reaches a certain threshold, and searches for a similar emotion solution on the network, and then the smart home system can refer to the current mood of the user by referring to the solution searched on the network.
  • Mitigation measures. 2 is a flowchart of an adjustment mechanism of a smart home system according to an embodiment of the present application. Specifically, as shown in FIG. 2, the smart home system may display facial expressions that may be presented by the user during use or before leaving the factory. Actions, sounds, etc. are stored in the smart home system.
  • the above information may be recorded, and then the facial expression previously stored in the smart home system may be The comparison of the body movements and the sounds, the judgment and the processing of the learning, the solution for the user to alleviate the user's emotions, for example, the corresponding music or video can be played, thereby alleviating the user's emotions.
  • FIG. 3 is a schematic diagram of a device control device according to an embodiment of the present application. As shown in FIG. 3, the device control device includes: first acquiring The unit 31, the evaluation unit 33 and the first transmitting unit 35. The device control device will be described in detail below.
  • the first obtaining unit 31 is configured to acquire information of the user, where the information includes at least one of the following: a photo of the user captured by the imaging device, and a voice of the user received by the audio device.
  • the evaluation unit 33 is connected to the first obtaining unit 31 for using the emotion level of the user corresponding to the model evaluation information, wherein the model is obtained according to the plurality of sets of data, and each set of data in the plurality of sets of data includes : The user's photo and/or sound, and a label that identifies the level of emotion that the photo and/or sound represents.
  • the first sending unit 35 is connected to the evaluation unit 33 for transmitting a control command according to an emotion level, wherein the control command is used to instruct the device to perform a predetermined operation.
  • the first acquiring unit 31 is configured to acquire information of the user, where the information includes at least one of the following: a photo of the user captured by the camera device, and received by the audio device.
  • the evaluation unit 33 is configured to use the model to evaluate the user's emotional level corresponding to the information, wherein the model is obtained according to the plurality of sets of data, each of the plurality of sets of data includes: the user's photo And/or a sound, and a label for identifying an emotional level represented by the photo and/or sound;
  • the first transmitting unit 35 is configured to send a control command according to the emotion level, wherein the control command is used to instruct the device to perform a predetermined operation.
  • the device control device provided by the embodiment of the present application achieves the purpose of controlling the smart home system according to the acquired user's emotion, and achieves the technical effect of letting the user experience the happiness brought by the modern technology to improve the quality of life, thereby solving the related In the technology, the home system can not meet the technical problem of the user's demand for the intelligence degree of the home system, and the user experience is improved.
  • the device control processing device further includes: a second sending unit, configured to send the photo and/or sound of the user to the user before using the model to evaluate the emotional level of the user corresponding to the information Other users; a second obtaining unit, configured to acquire tags added by other users for photos and/or sounds of the user.
  • the second obtaining unit includes at least one of the following: a first sending module, configured to send a photo and/or a sound of the user, and multiple emotion levels that can be selected to other users; Receiving, by the other user, an emotion level selected from a plurality of emotion levels as a label; an extraction module, configured to obtain an evaluation of the user's photo and/or sound by the other user, and extracting an emotion level as a label from the evaluation, wherein the evaluation includes the following At least one: evaluation of natural language, evaluation of speech.
  • the device control processing device further includes: a questioning unit, configured to: after acquiring the photo and/or sound of the user, using the model to evaluate the emotional level of the user corresponding to the information, according to The user's photo and/or sound asks the user; the extracting unit is configured to extract the emotional level corresponding to the user's photo and/or sound according to the user's answer to the question questioned.
  • a questioning unit configured to: after acquiring the photo and/or sound of the user, using the model to evaluate the emotional level of the user corresponding to the information, according to The user's photo and/or sound asks the user
  • the extracting unit is configured to extract the emotional level corresponding to the user's photo and/or sound according to the user's answer to the question questioned.
  • the first sending unit includes: a second sending module, configured to send a control command when the emotion level matches the predetermined level, where the control command is used to control the device to perform the following At least one of the operations: playing music corresponding to the emotional level, playing a video corresponding to the emotional level.
  • a storage medium comprising a stored program, wherein the program executes the device control processing method of any of the above.
  • processor configured to run a program, wherein the program is executed to execute the device control processing method of any one of the above.
  • the disclosed technical contents may be implemented in other manners.
  • the device embodiments described above are only schematic.
  • the division of the unit may be a logical function division.
  • there may be another division manner for example, multiple units or components may be combined or may be Integrate into another system, or some features can be ignored or not executed.
  • the mutual coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection through some interface, unit or module, and may be electrical or otherwise.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of the embodiment.
  • each functional unit in each embodiment of the present application may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit.
  • the above integrated unit can be implemented in the form of hardware or in the form of a software functional unit.
  • the integrated unit if implemented in the form of a software functional unit and sold or used as a standalone product, may be stored in a computer readable storage medium.
  • a computer readable storage medium A number of instructions are included to cause a computer device (which may be a personal computer, server or network device, etc.) to perform all or part of the steps of the methods described in various embodiments of the present application.
  • the foregoing storage medium includes: a U disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a removable hard disk, a magnetic disk, or an optical disk, and the like. .

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Automation & Control Theory (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Evolutionary Biology (AREA)
  • Child & Adolescent Psychology (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Hospice & Palliative Care (AREA)
  • Psychiatry (AREA)
  • Computational Linguistics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Acoustics & Sound (AREA)
  • User Interface Of Digital Computer (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The present application discloses a device control processing method and apparatus. The device control method comprises: acquiring information of a user, the information of the user comprising at least one of: a photograph of the user captured by a photographing device and a voice of the user received by an audio device; using a model to evaluate an emotional level of the user corresponding to the information of the user, the model being obtained by training according to a plurality of sets of data, each of the plurality of sets of data comprising: a photograph and/or a voice of the user and a tag for identifying an emotional level representing the photograph and/or the voice; and sending a control command according to the emotional level, the control command being used to instruct a device to perform a predetermined operation. The present application addresses the technical issue that home systems in the related art cannot meet requirements of users for an intelligence degree of the home systems.

Description

设备控制处理方法及装置Device control processing method and device
相关申请Related application
本申请要求2017年10月31日申请的,申请号为201711062745.0,名称为“设备控制处理方法及装置”的中国专利申请的优先权,在此将其全文引入作为参考。The present application claims the benefit of priority to the benefit of the benefit of the benefit of the benefit of the benefit of the benefit of the benefit of the benefit of the benefit of the benefit of the benefit of the benefit of the benefit of the benefit of the benefit of the benefit of the benefit of the benefit of the benefit of the benefit of the benefit of the benefit of the benefit of the disclosure.
技术领域Technical field
本申请涉及智能家居领域,具体而言,涉及一种设备控制处理方法及装置。The present application relates to the field of smart homes, and in particular to a device control processing method and apparatus.
背景技术Background technique
随着科学技术的发展,尤其是,人工智能的快速发展,使得人工智能快速地与各个领域结合,其中,人们在日常生活中能够感受到的就是各种家用设备的智能化。例如,用户可以利用手势对其进行控制,也可以通过语音实现对电视机的开启或者关闭。但是,目前市场上出售的家用设备的智能化程度,在一定程度上,并不能满足用户的需求。With the development of science and technology, especially the rapid development of artificial intelligence, artificial intelligence is quickly combined with various fields. Among them, people can feel the intelligence of various household devices in daily life. For example, the user can control it with a gesture, or turn the TV on or off by voice. However, the degree of intelligence of household equipment currently on the market does not meet the needs of users to a certain extent.
针对上述相关技术中家居系统并不能满足用户对家居系统的智能化程度的需求的问题,目前尚未提出有效的解决方案。In view of the above-mentioned related art, the problem that the home system does not satisfy the user's demand for the intelligence level of the home system has not yet proposed an effective solution.
发明内容Summary of the invention
本申请实施例提供了一种设备控制处理方法及装置,以至少解决相关技术中家居系统并不能满足用户对家居系统的智能化程度的需求的技术问题。The embodiment of the present application provides a device control processing method and device, so as to at least solve the technical problem that the home system in the related art cannot meet the user's demand for the intelligence degree of the home system.
根据本申请实施例的一个方面,提供了一种设备控制处理方法,包括:获取用户的信息,其中,所述信息包括以下至少之一:通过摄像设备拍摄到的所述用户的照片、通过音频设备接收到的所述用户的声音;使用模型评估所述信息对应的所述用户的情绪级别,其中,所述模型是根据多组数据训练得到的,所述多组数据中的每一组数据均包括:所述用户的照片和/或声音、以及用来标识该照片和/或声音所代表的情绪级别的标签;根据所述情绪级别发送控制命令,其中,所述控制命令用于指示设备进行预定操作。According to an aspect of the embodiments of the present application, a device control processing method is provided, including: acquiring information of a user, where the information includes at least one of: a photo of the user captured by an imaging device, through audio a voice of the user received by the device; using a model to evaluate an emotional level of the user corresponding to the information, wherein the model is trained according to a plurality of sets of data, each set of data in the plurality of sets of data Each includes: a photo and/or a sound of the user, and a label for identifying an emotional level represented by the photo and/or sound; transmitting a control command according to the emotional level, wherein the control command is used to indicate the device Perform the scheduled operation.
进一步地,在使用所述模型评估所述信息对应的所述用户的情绪级别之前,所述方法还包括:将所述用户的照片和/或声音发送给与其他用户;获取所述其他用户为所述用户的照片和/或声音添加的所述标签。Further, before using the model to evaluate the emotional level of the user corresponding to the information, the method further includes: transmitting a photo and/or a sound of the user to another user; acquiring the other user as The label of the user's photo and/or sound added.
进一步地,获取所述其他用户为所述用户的照片和/或声音添加的所述标签包括以下至 少之一:将所述用户的照片和/或声音,以及可以选择的多个情绪级别发送给所述其他用户;接收所述其他用户从所述多个情绪级别中选择出的情绪级别作为所述标签;获取所述其他用户对所述用户的照片和/或声音的评价,从所述评价中提取情绪级别作为所述标签,其中,所述评价包括以下至少之一:自然语言的评价、语音的评价。Further, acquiring the tag added by the other user for the photo and/or sound of the user includes at least one of: sending a photo and/or a sound of the user, and a plurality of emotion levels that can be selected to Receiving, by the other user, an emotion level selected by the other user from the plurality of emotion levels as the tag; acquiring an evaluation of the photo and/or voice of the user by the other user, from the evaluation The emotion level is extracted as the tag, wherein the evaluation includes at least one of: an evaluation of a natural language, an evaluation of a voice.
进一步地,在使用所述模型评估所述信息对应的所述用户的情绪级别之前,所述方法还包括:在获取到所述用户的照片和/或声音之后,根据所述用户的照片和/或声音向所述用户进行提问;根据所述用户对提问的问题的回答提取所述用户的照片和/或声音所对应的情绪级别。Further, before using the model to evaluate the emotional level of the user corresponding to the information, the method further includes: after obtaining the photo and/or sound of the user, according to the photo of the user and/or Or the voice makes a question to the user; the emotion level corresponding to the photo and/or sound of the user is extracted according to the user's answer to the question questioned.
进一步地,根据所述情绪级别发送所述控制命令包括:在所述情绪级别与预定级别相匹配的情况下,发送所述控制命令,其中,所述控制命令用于控制所述设备执行以下操作的至少之一:播放与所述情绪级别对应的音乐,播放与所述情绪级别对应的视频。Further, sending the control command according to the emotion level comprises: sending the control command if the emotion level matches a predetermined level, wherein the control command is used to control the device to perform the following operations At least one of: playing music corresponding to the emotion level, playing a video corresponding to the emotion level.
根据本申请实施例的另外一个方面,还提供了一种设备控制处理装置,包括:第一获取单元,用于获取用户的信息,其中,所述信息包括以下至少之一:通过摄像设备拍摄到的所述用户的照片、通过音频设备接收到的所述用户的声音;评估单元,用于使用模型评估所述信息对应的所述用户的情绪级别,其中,所述模型是根据多组数据训练得到的,所述多组数据中的每一组数据均包括:所述用户的照片和/或声音、以及用来标识该照片和/或声音所代表的情绪级别的标签;第一发送单元,用于根据所述情绪级别发送控制命令,其中,所述控制命令用于指示设备进行预定操作。According to another aspect of the embodiments of the present application, a device control processing apparatus is further provided, including: a first acquiring unit, configured to acquire information of a user, where the information includes at least one of the following: capturing by the imaging device a photo of the user, the voice of the user received through the audio device; an evaluation unit for evaluating a mood level of the user corresponding to the information using a model, wherein the model is trained according to multiple sets of data Obtaining, each of the plurality of sets of data includes: a photo and/or a sound of the user, and a label for identifying an emotional level represented by the photo and/or the sound; the first sending unit, And a method for transmitting a control command according to the emotion level, wherein the control command is used to instruct the device to perform a predetermined operation.
进一步地,所述装置还包括:第二发送单元,用于在使用所述模型评估所述信息对应的所述用户的情绪级别之前,将所述用户的照片和/或声音发送给与其他用户;第二获取单元,用于获取所述其他用户为所述用户的照片和/或声音添加的所述标签。Further, the device further includes: a second sending unit, configured to send the photo and/or sound of the user to other users before evaluating the emotion level of the user corresponding to the information by using the model a second obtaining unit, configured to acquire the label added by the other user for the photo and/or sound of the user.
进一步地,所述第二获取单元包括以下至少之一:第一发送模块,用于将所述用户的照片和/或声音,以及可以选择的多个情绪级别发送给所述其他用户;接收所述其他用户从所述多个情绪级别中选择出的情绪级别作为所述标签;提取模块,用于获取所述其他用户对所述用户的照片和/或声音的评价,从所述评价中提取情绪级别作为所述标签,其中,所述评价包括以下至少之一:自然语言的评价、语音的评价。Further, the second obtaining unit includes at least one of: a first sending module, configured to send a photo and/or a sound of the user, and a plurality of selectable multiple emotion levels to the other user; Determining, by the other user, an emotion level selected from the plurality of emotion levels as the label; an extracting module, configured to obtain an evaluation of the photo and/or sound of the user by the other user, and extracting from the evaluation The emotion level is the tag, wherein the evaluation includes at least one of: an evaluation of a natural language, an evaluation of a voice.
进一步地,所述装置还包括:提问单元,用于在使用所述模型评估所述信息对应的所述用户的情绪级别之前,在获取到所述用户的照片和/或声音之后,根据所述用户的照片和/或声音向所述用户进行提问;提取单元,用于根据所述用户对提问的问题的回答提取所述用户的照片和/或声音所对应的情绪级别。Further, the device further includes: a questioning unit, configured to: after obtaining the photo and/or sound of the user, using the model to evaluate the emotion level of the user corresponding to the information, according to the The user's photo and/or voice makes a question to the user; and an extracting unit is configured to extract an emotion level corresponding to the photo and/or sound of the user according to the user's answer to the question questioned.
进一步地,所述第一发送单元包括:第二发送模块,用于在所述情绪级别与预定级别 相匹配的情况下,发送所述控制命令,其中,所述控制命令用于控制所述设备执行以下操作的至少之一:播放与所述情绪级别对应的音乐,播放与所述情绪级别对应的视频。Further, the first sending unit includes: a second sending module, configured to send the control command if the emotion level matches a predetermined level, where the control command is used to control the device Performing at least one of the following operations: playing music corresponding to the emotion level, and playing a video corresponding to the emotion level.
根据本申请实施例的另外一个方面,还提供了一种存储介质,所述存储介质包括存储的程序,其中,所述程序执行上述任意一项所述的设备控制处理方法。According to still another aspect of the embodiments of the present application, there is also provided a storage medium, the storage medium comprising a stored program, wherein the program performs the device control processing method according to any one of the above.
根据本申请实施例的另外一个方面,还提供了一种处理器,其特征在于,所述处理器用于运行程序,其中,所述程序运行时执行上述任意一项所述的设备控制处理方法。According to still another aspect of the embodiments of the present application, there is also provided a processor, wherein the processor is configured to run a program, wherein the program is executed to perform the device control processing method according to any one of the above.
在本申请实施例中,可以获取用户的信息,其中,信息包括以下至少之一:通过摄像设备拍摄到的用户的照片、通过音频设备接收到的用户的声音;使用模型评估信息对应的用户的情绪级别,其中,模型是根据多组数据训练得到的,多组数据中的每一组数据均包括:用户的照片和/或声音、以及用来标识该照片和/或声音所代表的情绪级别的标签;根据情绪级别发送控制命令,其中,控制命令用于指示设备进行预定操作。通过本申请实施例提供的设备控制处理方法实现了根据获取的用户的情绪对上述智能家居系统进行控制的目的,达到了让用户体验现代科技带来的快乐提高生活质量的技术效果,进而解决了相关技术中家居系统并不能满足用户对家居系统的智能化程度的需求的技术问题,提升了用户体验。In the embodiment of the present application, the information of the user may be acquired, where the information includes at least one of the following: a photo of the user captured by the imaging device, a voice of the user received by the audio device, and a user corresponding to the model evaluation information. An emotional level in which the model is trained based on a plurality of sets of data, each of the sets of data including: a photo and/or sound of the user, and a level of emotion represented by the photo and/or sound. The tag transmits a control command according to the emotion level, wherein the control command is used to instruct the device to perform a predetermined operation. The device control processing method provided by the embodiment of the present application achieves the purpose of controlling the smart home system according to the acquired user's emotion, and achieves the technical effect of letting the user experience the happiness brought by the modern technology and improving the quality of life, thereby solving the problem. In the related art, the home system does not satisfy the technical problem of the user's demand for the intelligence degree of the home system, and the user experience is improved.
附图说明DRAWINGS
此处所说明的附图用来提供对本申请的进一步理解,构成本申请的一部分,本申请的示意性实施例及其说明用于解释本申请,并不构成对本申请的不当限定。在附图中:The drawings described herein are intended to provide a further understanding of the present application, and are intended to be a part of this application. In the drawing:
图1是根据本申请实施例的设备控制处理方法的流程图;1 is a flowchart of a device control processing method according to an embodiment of the present application;
图2是根据本申请实施例的智能家居系统的调节机制的流程图;以及2 is a flow chart of an adjustment mechanism of a smart home system according to an embodiment of the present application;
图3是根据本申请实施例的设备控制装置的示意图。FIG. 3 is a schematic diagram of a device control apparatus according to an embodiment of the present application.
具体实施方式Detailed ways
为了使本技术领域的人员更好地理解本申请方案,下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本申请一部分的实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都应当属于本申请保护的范围。The technical solutions in the embodiments of the present application are clearly and completely described in the following with reference to the accompanying drawings in the embodiments of the present application. It is an embodiment of the present application, and not all of the embodiments. All other embodiments obtained by a person of ordinary skill in the art based on the embodiments of the present application without departing from the inventive scope shall fall within the scope of the application.
需要说明的是,本申请的说明书和权利要求书及上述附图中的术语“第一”、“第二”等是用于区别类似的对象,而不必用于描述特定的顺序或先后次序。应该理解这样使用的 数据在适当情况下可以互换,以便这里描述的本申请的实施例能够以除了在这里图示或描述的那些以外的顺序实施。此外,术语“包括”和“具有”以及他们的任何变形,意图在于覆盖不排他的包含,例如,包含了一系列步骤或单元的过程、方法、系统、产品或设备不必限于清楚地列出的那些步骤或单元,而是可包括没有清楚地列出的或对于这些过程、方法、产品或设备固有的其它步骤或单元。It should be noted that the terms "first", "second" and the like in the specification and claims of the present application and the above-mentioned drawings are used to distinguish similar objects, and are not necessarily used to describe a specific order or order. It is to be understood that the data so used may be interchanged where appropriate, so that the embodiments of the present application described herein can be implemented in a sequence other than those illustrated or described herein. In addition, the terms "comprises" and "comprises" and "the" and "the" are intended to cover a non-exclusive inclusion, for example, a process, method, system, product, or device that comprises a series of steps or units is not necessarily limited to Those steps or units may include other steps or units not explicitly listed or inherent to such processes, methods, products or devices.
为方便用户理解本申请,下面对本申请实施例中涉及的部分术语或名词进行解释:To facilitate the user's understanding of the present application, some of the terms or nouns involved in the embodiments of the present application are explained below:
像素:是计算机屏幕上所能显示的最小单位,用来表示图像的单位,指可以显示出的水平和垂直像素的数组,屏幕中的像素点越多,画面的分辨率越高,图像就越细腻逼真。Pixel: is the smallest unit that can be displayed on the computer screen. It is used to represent the unit of the image. It refers to the array of horizontal and vertical pixels that can be displayed. The more pixels in the screen, the higher the resolution of the image, the more the image is. Delicate and realistic.
像素点:指像素的数值。Pixel: refers to the value of a pixel.
二值化:对摄像头拍摄的图片,大多数是彩色图像,彩色图像所含信息量巨大,对于图片的内容,可以简单的分为前景与背景,先对彩色图进行处理,使图片只有前景信息与背景信息,可以简单的定义前景信息为黑色,背景信息为白色,这就是二值化图了。Binarization: Most of the pictures taken by the camera are color images. The color image contains a huge amount of information. For the content of the picture, it can be simply divided into foreground and background. The color picture is processed first, so that the picture only has foreground information. With the background information, you can simply define the foreground information as black and the background information as white. This is the binarization graph.
神经网络算法:是指根据逻辑规则进行推理的过程,它是先将信息化成概念,并用户符号表示,然后,根据符号运算按串行模式进行逻辑推理;这一过程可以写成串行的指令,让计算机执行。Neural network algorithm: refers to the process of reasoning according to logic rules. It first converts information into concepts and represents user symbols. Then, according to symbolic operations, it performs logical reasoning in serial mode; this process can be written as serial instructions. Let the computer execute.
声纹:是用电声学仪器显示的携带有言语信息的声波频谱。Voiceprint: It is a sound wave spectrum that is displayed with electro-acoustic instruments and carries speech information.
声纹识别:是生物识别技术的一种。也称为说话人识别,有两种,即说话人辨认和说话人确认,不同的任务和应用会使用不同的声纹识别技术,如缩小刑侦范围时可能需要辨认技术,而银行交易时则需要确认技术。Voiceprint recognition: It is a kind of biometric technology. Also known as speaker recognition, there are two types, namely speaker recognition and speaker confirmation. Different tasks and applications use different voiceprint recognition techniques. For example, when narrowing the scope of criminal investigation, it may be necessary to identify the technology. Confirm the technology.
以下实施例可以用于各种电器设备中,对于电器设备的类型不做具体限定,包括但不限于:洗衣机、空调、冰箱等,上述各种电器设备组成了本申请实施例中的智能家居系统,下面对本申请实施例进行详细说明。The following embodiments may be used in various electrical devices, and the types of the electrical devices are not specifically limited, including but not limited to: a washing machine, an air conditioner, a refrigerator, etc., and the above various electrical devices constitute the smart home system in the embodiment of the present application. The embodiments of the present application are described in detail below.
根据本申请实施例,提供了一种设备控制处理方法的方法实施例,需要说明的是,在附图的流程图示出的步骤可以在诸如一组计算机可执行指令的计算机系统中执行,并且,虽然在流程图中示出了逻辑顺序,但是在某些情况下,可以以不同于此处的顺序执行所示出或描述的步骤。According to an embodiment of the present application, a method embodiment of a device control processing method is provided, and it should be noted that the steps shown in the flowchart of the accompanying drawings may be executed in a computer system such as a set of computer executable instructions, and Although the logical order is shown in the flowcharts, in some cases the steps shown or described may be performed in a different order than the ones described herein.
图1是根据本申请实施例的设备控制处理方法的流程图,如图1所示,该设备控制处理方法包括如下步骤:FIG. 1 is a flowchart of a device control processing method according to an embodiment of the present application. As shown in FIG. 1 , the device control processing method includes the following steps:
步骤S102,获取用户的信息,其中,信息包括以下至少之一:通过摄像设备拍摄到的用户的照片、通过音频设备接收到的用户的声音。Step S102: Acquire information of the user, where the information includes at least one of the following: a photo of the user captured by the imaging device, and a voice of the user received by the audio device.
在步骤S102中,一个方面,可以在用户的家里安装一个或多个摄像头,用于拍摄用 户的照片,在本申请实施例中对于摄像头的设置位置不做具体限定,可以包括但不限于:用户家里的每个房间的门口、天花板等,通过设置在不同位置的摄像头可以分别采集用户的照片。其中,在利用上述一个或多个摄像头拍摄用户的照片时,可以是每个预定时间段对用户进行拍摄,然后分析图像中用户的情绪。在本申请实施例中对拍摄的图像的类别不做具体限定,可以包括但不限于:黑白图像(灰度图像)、彩色图像(RGB图像)。在分析图像时,可以根据二值化图像处理方式分析图像中的信息,具体地,在分析时,可以对图像中的多个像素点与历史图像中的像素点位置进行比较,以确定出存在差异的像素点,然后将存在差异的像素点区分出来,进而可以从利用摄像设备拍摄得到的图像中提取出用户的信息。In step S102, in one aspect, one or more cameras may be installed in the user's home for taking a photo of the user. In the embodiment of the present application, the setting position of the camera is not specifically limited, and may include, but is not limited to, a user. The door, ceiling, etc. of each room in the home can be used to collect photos of the user through cameras installed in different locations. Wherein, when the user's photo is taken by using the one or more cameras described above, the user may be photographed every predetermined time period, and then the emotion of the user in the image may be analyzed. The category of the captured image is not specifically limited in the embodiment of the present application, and may include, but is not limited to, a black and white image (grayscale image) and a color image (RGB image). When analyzing the image, the information in the image can be analyzed according to the binarized image processing manner. Specifically, when analyzing, multiple pixel points in the image can be compared with the pixel position in the historical image to determine the existence. The pixel points of the difference are then distinguished from the pixel points where the difference exists, and the user's information can be extracted from the image captured by the imaging device.
另外一个方面,还可以在用户的家里安装有一个或者多个音频设备(也即是上下文中的声音传感器),用于接收用户的声音,在本申请实施例中对于音频设备的安装位置不做具体限定,可以包括但不限于:用户家里的每一个房间的门口处、天花板等位置,不过为了方便接收用户的语音,可以将上述音频设备安装在用户经常活动且位置与人体的身高相当的位置处。其中,上述语音设备中包含有语音模型库,该语音模型库中存储有一个家庭中的各个成员的声纹。其中,一个家庭中的每个成员可以对着该语音设备发出语音,该语音设备可以进行特征提取,将不同成员的声纹存储到语音模型库中。当该家庭中的某个成员发出语音时,该语音设备可以对该成员的语音进行特征提取,得到该成员的声纹,然后将该成员的声纹与语音模型库中存储的声纹进行匹配,然后识别出该声纹对应的家庭成员,进而得到该成员对应的信息。In another aspect, one or more audio devices (that is, sound sensors in the context) may be installed in the user's home for receiving the voice of the user. In the embodiment of the present application, the installation location of the audio device is not performed. The specific limitation may include, but is not limited to, a position at a doorway, a ceiling, and the like of each room in the user's home. However, in order to facilitate receiving the voice of the user, the audio device may be installed at a position where the user is frequently active and the position is equal to the height of the human body. At the office. The voice device includes a voice model library, where the voice model library stores voiceprints of each member of the family. Each member of a family can issue a voice to the voice device, and the voice device can perform feature extraction to store voiceprints of different members into the voice model library. When a member of the family emits a voice, the voice device can extract the feature of the member's voice, obtain the voiceprint of the member, and then match the voiceprint of the member with the voiceprint stored in the voice model library. And then identifying the family member corresponding to the voiceprint, and then obtaining the information corresponding to the member.
步骤S104,使用模型评估信息对应的用户的情绪级别,其中,模型是根据多组数据训练得到的,多组数据中的每一组数据均包括:用户的照片和/或声音、以及用来标识该照片和/或声音所代表的情绪级别的标签。Step S104, using the model to evaluate the emotional level of the user corresponding to the information, wherein the model is obtained according to the plurality of sets of data, each of the plurality of sets of data includes: a photo and/or a sound of the user, and is used to identify The photo and/or voice represents the level of the emotional level of the label.
其中,上述模型可以是利用摄像头在历史预定时间段中拍摄的图像,利用音频设备接收到的历史预定时间段内的用户的声音以及用来标识该照片和/或声音所代表的情绪级别的标签进行学习训练得到的。Wherein, the above model may be an image captured by the camera during a predetermined period of time, a voice of the user within a predetermined period of time received by the audio device, and a label for identifying an emotional level represented by the photo and/or sound. Learned by training.
步骤S106,根据情绪级别发送控制命令,其中,控制命令用于指示设备进行预定操作。Step S106, sending a control command according to the emotion level, wherein the control command is used to instruct the device to perform a predetermined operation.
通过上述步骤,在智能家居系统运行时,可以使用模型对用户的信息进行评估得到对应的用户的情绪级别,然后根据评估得到的情绪级别向智能家居系统发送控制命令。通过本申请实施例提供的设备控制处理方法实现了根据获取的用户的情绪对上述智能家居系统进行控制的目的,达到了让用户体验现代科技带来的快乐提高生活质量的技术效果,进而解决了相关技术中家居系统并不能满足用户对家居系统的智能化程度的需求的技术问 题,提升了用户体验。Through the above steps, when the smart home system is running, the model can be used to evaluate the user's information to obtain the corresponding user's emotional level, and then send a control command to the smart home system according to the evaluated emotional level. The device control processing method provided by the embodiment of the present application achieves the purpose of controlling the smart home system according to the acquired user's emotion, and achieves the technical effect of letting the user experience the happiness brought by the modern technology and improving the quality of life, thereby solving the problem. In the related art, the home system does not satisfy the technical problem of the user's demand for the intelligence degree of the home system, and the user experience is improved.
在本申请一个可选的实施例中,在智能家居系统的人工智能终端,利用摄像头捕捉用户当前的面部表情以及肢体动作,结合图像识别技术,进行用户面部表情以及肢体动作的记录,通过神经网络算法进行对比、判断、反馈、学习;同时,利用声音传感器接收用户的声音,并进行用户声音变化的记录,通过神经网络算法对比、判断、反馈、学习。通过观察记录用户日常生活表现出的面部表情、肢体动作、声音变化、判断、反馈、学习,得到用户在各种形态下的情绪状态方程,并以此为判别标准。其中,该情绪状态方程为:f(面部表情,肢体动作,声音......外界刺激)=用户(情绪级别),其中,自变量:面部表情,肢体动作,声音......外界刺激等;因变量:用户当前情绪级别;在这里不同程度的面部表情,肢体动作,声音......外界刺激等,对应用户不同的当前情绪级别。In an optional embodiment of the present application, in the artificial intelligence terminal of the smart home system, the camera captures the user's current facial expressions and limb movements, and combines image recognition technology to record the user's facial expressions and limb movements through the neural network. The algorithm compares, judges, feedbacks, and learns; at the same time, the sound sensor is used to receive the user's voice, and the user's voice changes are recorded, and the neural network algorithm is used to compare, judge, feedback, and learn. By observing and recording the facial expressions, limb movements, sound changes, judgments, feedbacks, and learnings exhibited by the daily life of the user, the emotional state equations of the user in various forms are obtained, and this is used as a criterion. Among them, the emotional state equation is: f (facial expression, limb movement, sound ... external stimulation) = user (emotional level), wherein, independent variables: facial expressions, limb movements, sounds..... External stimulus, etc.; dependent variable: the current emotional level of the user; here different degrees of facial expressions, physical movements, sounds... external stimuli, etc., corresponding to different current emotional levels of the user.
需要说明的是,用来标识该照片和/或声音所代表的情绪级别的标签可以从多个方面获取。It should be noted that the label used to identify the emotional level represented by the photo and/or sound can be obtained from multiple aspects.
一方面,在本申请一个可选的实施例中,在使用模型评估信息对应的用户的情绪级别之前,上述设备控制处理方法还可以包括:将用户的照片和/或声音发送给与其他用户;获取其他用户为用户的照片和/或声音添加的标签。On the one hand, in an optional embodiment of the present application, before using the model to evaluate the emotional level of the user corresponding to the information, the device control processing method may further include: sending the user's photo and/or voice to other users; Get tags added by other users to the user's photos and/or sounds.
在上述实施例中,获取其他用户为用户的照片和/或声音添加的标签可以包括以下至少之一:将用户的照片和/或声音,以及可以选择的多个情绪级别发送给其他用户;接收其他用户从多个情绪级别中选择出的情绪级别作为标签;获取其他用户对用户的照片和/或声音的评价,从评价中提取情绪级别作为标签,其中,评价包括以下至少之一:自然语言的评价、语音的评价。例如,在必要的情况下,智能家居系统会向用户的亲朋好友(也即是上下文中的其他用户)发送用户的照片或者声音,该用户的亲朋好友会对接收的该用户的照片或者声音与该用户历史时间段的情绪进行比较,然后为该用户的照片和/或声音添加标签,比如,该用户由于工作压力大而比较郁闷,或者,在外面遇到了不开心的事情而比较烦恼等等;当然,也可以通过该用户的亲朋好友对接收到的该用户的照片和/或声音直接做出评价,智能家居系统会从该用户的亲朋好友的评价中提取情绪级别作为标签,其中,上述评价可以包括但不限于:该用户的亲朋好友的发送的一段评价性文字(也即是上下文中的自然语言评价),一段评价性的语音(也即是上下文中的语音的评价)。In the above embodiment, acquiring a tag added by other users for the user's photo and/or sound may include at least one of: transmitting the user's photo and/or sound, and a plurality of selectable emotion levels to other users; receiving Other users select the emotion level from the plurality of emotion levels as a label; obtain other users' evaluations of the user's photos and/or sounds, and extract the emotion level as a label from the evaluation, wherein the evaluation includes at least one of the following: natural language Evaluation, speech evaluation. For example, if necessary, the smart home system will send the user's photo or voice to the user's relatives and friends (that is, other users in the context), and the user's friends and relatives will receive the user's photo or sound and The emotions of the user's historical time period are compared, and then the user's photos and/or sounds are tagged, for example, the user is depressed due to work pressure, or is troubled by unhappy things outside, etc. Of course, the user's friends and relatives can also directly evaluate the received photo and/or voice of the user, and the smart home system extracts the emotion level as a label from the evaluation of the user's relatives and friends. The evaluation may include, but is not limited to, an evaluative text (ie, a natural language evaluation in context) sent by the user's friends and relatives, and an evaluative speech (ie, an evaluation of the voice in the context).
另外一个方面,在使用模型评估信息对应的用户的情绪级别之前,该设备控制处理方法还可以包括:在获取到用户的照片和/或声音之后,根据用户的照片和/或声音向用户进行提问;根据用户对提问的问题的回答提取用户的照片和/或声音所对应的情绪级别。例如,上述智能家居系统可以通过对话式的方式获取用户的情绪级别,该智能家居系统接收到用 户的照片和/或声音之后,会向用户提出问题,假如,该智能家居系统提出问题:“今天心情如何”;用户回答:“工作压力太大了,比较烦躁”;那么该智能家居系统会从用户的回答中提取该用户对应的情绪级别。In another aspect, before using the model to evaluate the emotion level of the user corresponding to the information, the device control processing method may further include: after obtaining the photo and/or sound of the user, asking the user according to the photo and/or sound of the user. The emotion level corresponding to the user's photo and/or voice is extracted according to the user's answer to the question questioned. For example, the smart home system can obtain the user's emotional level through a conversational manner. After receiving the user's photo and/or voice, the smart home system will ask the user a question, if the smart home system asks a question: "Today "How the mood"; the user replies: "The work pressure is too big, more irritating"; then the smart home system will extract the user's corresponding emotional level from the user's answer.
在本申请一个可选的实施例中,为了更好地了解用户,服务用户,该智能家居系统还可以进行自校正。例如,通过用户的亲朋好友得到的用户的情绪级别为烦恼,而通过对话式的方式得到该用户的情绪级别为开心,那么该智能家居系统会结合上述从其他人员那里得到的用户的情绪级别以及通过对话的方式从用户自己那里得到的情绪级别,判断偏差,并进行修正,不断学习改进。In an optional embodiment of the present application, the smart home system can also perform self-correction in order to better understand the user and serve the user. For example, if the user's emotional level is obtained by the user's friends and relatives, and the user's emotional level is happy by the conversational manner, the smart home system combines the above-mentioned user's emotional level obtained from other people and The level of emotions obtained from the users themselves through dialogue, judgment of deviations, and corrections, continuous learning and improvement.
在本申请一个可选的实施例中,根据情绪级别发送控制命令可以包括:在情绪级别与预定级别相匹配的情况下,发送控制命令,其中,控制命令用于控制设备执行以下操作的至少之一:播放与情绪级别对应的音乐,播放与情绪级别对应的视频。In an optional embodiment of the present application, sending the control command according to the emotion level may include: sending a control command when the emotion level matches the predetermined level, wherein the control command is used to control the device to perform at least the following operations One: Play music corresponding to the emotional level, and play the video corresponding to the emotional level.
下面对本申请一个完整的实施例进行详细说明。A complete embodiment of the present application will be described in detail below.
首先,用户打开智能家居系统,该智能家居系统通过观察并记录用户的日常生活,与用户互动交流,对比、判断、反馈、学习用户情绪;其中,交流方式可以包括但不限于:语音与系统对话,如内心独白、文字记录、面部表情以及肢体动作等。在这里有一个条件就是,用户需要对该智能家居系统开放,尽可能地向该智能家居系统输入自己的比较全面的个人信息。当用户的情绪出现异常时,智能家居系统会根据自己的决策,通过外在的手段(例如,播放音乐、视频等),调节用户的情绪。另外,上述智能家居系统还可以通过询问的方式查找得到导致用户产生负面情绪的原因,进而可以有针对性地采取对该用户的情绪进行缓解的措施。例如,上述智能家居系统中存储有与用户当前情绪类似的历史情绪时,那么可以在该智能家居系统中查找该历史情绪对应的历史解决方案,进而可以参考或者直接采用该历史解决方案来缓解用户的情绪。其中,在上述智能家居系统中没有找到与用户当前的情绪类似的历史情绪时,上述智能家居系统当然也就不会存储有用于解决用户当前情绪的参考解决方案,此时,智能家居系统可以在网络上搜有与用户当前情绪相似度达到一定阈值的相似情绪,并在网络上搜索该相似情绪的解决方案,然后上述智能家居系统可以参考在网络上搜索到的解决方案对用户的当前情绪采取缓解措施。其中,图2是根据本申请实施例的智能家居系统的调节机制的流程图,具体如图2所示,上述智能家居系统在使用过程中或者是出厂前可以将用户可能呈现的面部表情,肢体动作以及声音等存储到该智能家居系统中。另外,在上述智能家居系统使用的过程中,在识别到用户的情绪变化、信息交流等用户反馈过去的信息时,可以将上述信息记录下来,然后与之前存储在智能家居系统中的面部表情,肢体动作以及声音进行对比,判断以及学习的处理,得到用户 缓解用户情绪的解决方案,例如可以播放对应的音乐或者视频等,从而来缓解用户的情绪。First, the user opens the smart home system, which observes and records the daily life of the user, interacts with the user, compares, judges, feeds back, and learns the user's emotions; wherein the communication mode may include but is not limited to: voice and system dialogue Such as inner monologues, transcripts, facial expressions, and body movements. One of the conditions here is that the user needs to be open to the smart home system and enter his or her more comprehensive personal information into the smart home system as much as possible. When the user's emotions are abnormal, the smart home system will adjust the user's emotions according to their own decisions through external means (for example, playing music, videos, etc.). In addition, the above-mentioned smart home system can also find out the cause of the user's negative emotion by querying, and then can take targeted measures to alleviate the emotion of the user. For example, when the historical home mood similar to the current mood of the user is stored in the smart home system, the historical solution corresponding to the historical emotion may be searched in the smart home system, and then the historical solution may be referred to or directly used to alleviate the user. Emotions. Wherein, in the above smart home system, when the historical mood similar to the current mood of the user is not found, the smart home system does not store a reference solution for solving the current mood of the user. At this time, the smart home system can The network searches for similar emotions that the user's current emotional similarity reaches a certain threshold, and searches for a similar emotion solution on the network, and then the smart home system can refer to the current mood of the user by referring to the solution searched on the network. Mitigation measures. 2 is a flowchart of an adjustment mechanism of a smart home system according to an embodiment of the present application. Specifically, as shown in FIG. 2, the smart home system may display facial expressions that may be presented by the user during use or before leaving the factory. Actions, sounds, etc. are stored in the smart home system. In addition, in the process of using the smart home system described above, when the user's emotion change, information exchange, and the like are used to feedback past information, the above information may be recorded, and then the facial expression previously stored in the smart home system may be The comparison of the body movements and the sounds, the judgment and the processing of the learning, the solution for the user to alleviate the user's emotions, for example, the corresponding music or video can be played, thereby alleviating the user's emotions.
根据本申请实施例的另外一个方面,还提供了一种设备控制处理装置,图3是根据本申请实施例的设备控制装置的示意图,如图3所示,该设备控制装置包括:第一获取单元31,评估单元33以及第一发送单元35。下面对该设备控制装置进行详细说明。According to another aspect of the embodiments of the present application, a device control processing device is further provided. FIG. 3 is a schematic diagram of a device control device according to an embodiment of the present application. As shown in FIG. 3, the device control device includes: first acquiring The unit 31, the evaluation unit 33 and the first transmitting unit 35. The device control device will be described in detail below.
第一获取单元31,用于获取用户的信息,其中,信息包括以下至少之一:通过摄像设备拍摄到的用户的照片、通过音频设备接收到的用户的声音。The first obtaining unit 31 is configured to acquire information of the user, where the information includes at least one of the following: a photo of the user captured by the imaging device, and a voice of the user received by the audio device.
评估单元33,与所述第一获取单元31连接,用于使用模型评估信息对应的用户的情绪级别,其中,模型是根据多组数据训练得到的,多组数据中的每一组数据均包括:用户的照片和/或声音、以及用来标识该照片和/或声音所代表的情绪级别的标签。The evaluation unit 33 is connected to the first obtaining unit 31 for using the emotion level of the user corresponding to the model evaluation information, wherein the model is obtained according to the plurality of sets of data, and each set of data in the plurality of sets of data includes : The user's photo and/or sound, and a label that identifies the level of emotion that the photo and/or sound represents.
第一发送单元35,与所述评估单元33连接,用于根据情绪级别发送控制命令,其中,控制命令用于指示设备进行预定操作。The first sending unit 35 is connected to the evaluation unit 33 for transmitting a control command according to an emotion level, wherein the control command is used to instruct the device to perform a predetermined operation.
在上述实施例中,在智能家居系统运行时,采用第一获取单元31,用于获取用户的信息,其中,信息包括以下至少之一:通过摄像设备拍摄到的用户的照片、通过音频设备接收到的用户的声音;评估单元33,用于使用模型评估信息对应的用户的情绪级别,其中,模型是根据多组数据训练得到的,多组数据中的每一组数据均包括:用户的照片和/或声音、以及用来标识该照片和/或声音所代表的情绪级别的标签;第一发送单元35,用于根据情绪级别发送控制命令,其中,控制命令用于指示设备进行预定操作。通过本申请实施例提供的设备控制装置实现了根据获取的用户的情绪对上述智能家居系统进行控制的目的,达到了让用户体验现代科技带来的快乐提高生活质量的技术效果,进而解决了相关技术中家居系统并不能满足用户对家居系统的智能化程度的需求的技术问题,提升了用户体验。In the above embodiment, when the smart home system is running, the first acquiring unit 31 is configured to acquire information of the user, where the information includes at least one of the following: a photo of the user captured by the camera device, and received by the audio device. The user's voice; the evaluation unit 33 is configured to use the model to evaluate the user's emotional level corresponding to the information, wherein the model is obtained according to the plurality of sets of data, each of the plurality of sets of data includes: the user's photo And/or a sound, and a label for identifying an emotional level represented by the photo and/or sound; the first transmitting unit 35 is configured to send a control command according to the emotion level, wherein the control command is used to instruct the device to perform a predetermined operation. The device control device provided by the embodiment of the present application achieves the purpose of controlling the smart home system according to the acquired user's emotion, and achieves the technical effect of letting the user experience the happiness brought by the modern technology to improve the quality of life, thereby solving the related In the technology, the home system can not meet the technical problem of the user's demand for the intelligence degree of the home system, and the user experience is improved.
在本申请一个可选的实施例中,该设备控制处理装置还包括:第二发送单元,用于在使用模型评估信息对应的用户的情绪级别之前,将用户的照片和/或声音发送给与其他用户;第二获取单元,用于获取其他用户为用户的照片和/或声音添加的标签。In an optional embodiment of the present application, the device control processing device further includes: a second sending unit, configured to send the photo and/or sound of the user to the user before using the model to evaluate the emotional level of the user corresponding to the information Other users; a second obtaining unit, configured to acquire tags added by other users for photos and/or sounds of the user.
在本申请一个可选的实施例中,第二获取单元包括以下至少之一:第一发送模块,用于将用户的照片和/或声音,以及可以选择的多个情绪级别发送给其他用户;接收其他用户从多个情绪级别中选择出的情绪级别作为标签;提取模块,用于获取其他用户对用户的照片和/或声音的评价,从评价中提取情绪级别作为标签,其中,评价包括以下至少之一:自然语言的评价、语音的评价。In an optional embodiment of the present application, the second obtaining unit includes at least one of the following: a first sending module, configured to send a photo and/or a sound of the user, and multiple emotion levels that can be selected to other users; Receiving, by the other user, an emotion level selected from a plurality of emotion levels as a label; an extraction module, configured to obtain an evaluation of the user's photo and/or sound by the other user, and extracting an emotion level as a label from the evaluation, wherein the evaluation includes the following At least one: evaluation of natural language, evaluation of speech.
在本申请一个可选的实施例中,该设备控制处理装置还包括:提问单元,用于在使用模型评估信息对应的用户的情绪级别之前,在获取到用户的照片和/或声音之后,根据用户的照片和/或声音向用户进行提问;提取单元,用于根据用户对提问的问题的回答提取用户 的照片和/或声音所对应的情绪级别。In an optional embodiment of the present application, the device control processing device further includes: a questioning unit, configured to: after acquiring the photo and/or sound of the user, using the model to evaluate the emotional level of the user corresponding to the information, according to The user's photo and/or sound asks the user; the extracting unit is configured to extract the emotional level corresponding to the user's photo and/or sound according to the user's answer to the question questioned.
在本申请一个可选的实施例中,第一发送单元包括:第二发送模块,用于在情绪级别与预定级别相匹配的情况下,发送控制命令,其中,控制命令用于控制设备执行以下操作的至少之一:播放与情绪级别对应的音乐,播放与情绪级别对应的视频。In an optional embodiment of the present application, the first sending unit includes: a second sending module, configured to send a control command when the emotion level matches the predetermined level, where the control command is used to control the device to perform the following At least one of the operations: playing music corresponding to the emotional level, playing a video corresponding to the emotional level.
根据本申请实施例的另外一个方面,还提供了一种存储介质,存储介质包括存储的程序,其中,程序执行上述任意一项的设备控制处理方法。According to still another aspect of the embodiments of the present application, there is also provided a storage medium comprising a stored program, wherein the program executes the device control processing method of any of the above.
根据本申请实施例的另外一个方面,还提供了一种处理器,其特征在于,处理器用于运行程序,其中,程序运行时执行上述任意一项的设备控制处理方法。According to still another aspect of the embodiments of the present application, there is also provided a processor, wherein the processor is configured to run a program, wherein the program is executed to execute the device control processing method of any one of the above.
上述本申请实施例序号仅仅为了描述,不代表实施例的优劣。The serial numbers of the embodiments of the present application are merely for the description, and do not represent the advantages and disadvantages of the embodiments.
在本申请的上述实施例中,对各个实施例的描述都各有侧重,某个实施例中没有详述的部分,可以参见其他实施例的相关描述。In the above-mentioned embodiments of the present application, the descriptions of the various embodiments are different, and the parts that are not detailed in a certain embodiment can be referred to the related descriptions of other embodiments.
在本申请所提供的几个实施例中,应该理解到,所揭露的技术内容,可通过其它的方式实现。其中,以上所描述的装置实施例仅仅是示意性的,例如所述单元的划分,可以为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,单元或模块的间接耦合或通信连接,可以是电性或其它的形式。In the several embodiments provided by the present application, it should be understood that the disclosed technical contents may be implemented in other manners. The device embodiments described above are only schematic. For example, the division of the unit may be a logical function division. In actual implementation, there may be another division manner, for example, multiple units or components may be combined or may be Integrate into another system, or some features can be ignored or not executed. In addition, the mutual coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection through some interface, unit or module, and may be electrical or otherwise.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。The units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of the embodiment.
另外,在本申请各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。In addition, each functional unit in each embodiment of the present application may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit. The above integrated unit can be implemented in the form of hardware or in the form of a software functional unit.
所述集成的单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的全部或部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可为个人计算机、服务器或者网络设备等)执行本申请各个实施例所述方法的全部或部分步骤。而前述的存储介质包括:U盘、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、移动硬盘、磁碟或者光盘等各种可以存储程序代码的介质。The integrated unit, if implemented in the form of a software functional unit and sold or used as a standalone product, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the present application, in essence or the contribution to the prior art, or all or part of the technical solution may be embodied in the form of a software product stored in a storage medium. A number of instructions are included to cause a computer device (which may be a personal computer, server or network device, etc.) to perform all or part of the steps of the methods described in various embodiments of the present application. The foregoing storage medium includes: a U disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a removable hard disk, a magnetic disk, or an optical disk, and the like. .
以上所述仅是本申请的优选实施方式,应当指出,对于本技术领域的普通技术人员来说,在不脱离本申请原理的前提下,还可以做出若干改进和润饰,这些改进和润饰也应视为本申请的保护范围。The above description is only a preferred embodiment of the present application, and it should be noted that those skilled in the art can also make several improvements and retouchings without departing from the principles of the present application. It should be considered as the scope of protection of this application.

Claims (12)

  1. 一种设备控制处理方法,其特征在于,包括:A device control processing method, comprising:
    获取用户的信息,其中,所述用户的信息包括以下至少之一:通过摄像设备拍摄到的所述用户的照片、通过音频设备接收到的所述用户的声音;Obtaining information of the user, where the information of the user includes at least one of the following: a photo of the user captured by the imaging device, and a voice of the user received by the audio device;
    使用模型评估所述用户的信息对应的所述用户的情绪级别,其中,所述模型是根据多组数据训练得到的,所述多组数据中的每一组数据均包括:所述用户的照片和/或声音、以及用来标识该照片和/或声音所代表的情绪级别的标签;Using a model to evaluate an emotional level of the user corresponding to the user's information, wherein the model is obtained according to a plurality of sets of data, each of the plurality of sets of data includes: a photo of the user And/or sound, and a label identifying the level of emotion represented by the photo and/or sound;
    根据所述情绪级别发送控制命令,其中,所述控制命令用于指示设备进行预定操作。And transmitting a control command according to the emotion level, wherein the control command is used to instruct the device to perform a predetermined operation.
  2. 根据权利要求1所述的方法,其特征在于,在使用所述模型评估所述用户的信息对应的所述用户的情绪级别之前,所述方法还包括:The method according to claim 1, wherein before the evaluating the emotion level of the user corresponding to the information of the user by using the model, the method further comprises:
    将所述用户的照片和/或声音发送给与其他用户;Sending photos and/or sounds of the user to other users;
    获取所述其他用户为所述用户的照片和/或声音添加的所述标签。Obtaining the tag added by the other user for the photo and/or sound of the user.
  3. 根据权利要求2所述的方法,其特征在于,获取所述其他用户为所述用户的照片和/或声音添加的所述标签包括以下至少之一:The method according to claim 2, wherein the obtaining the tag added by the other user for the photo and/or sound of the user comprises at least one of the following:
    将所述用户的照片和/或声音,以及可以选择的多个情绪级别发送给所述其他用户;接收所述其他用户从所述多个情绪级别中选择出的情绪级别作为所述标签;Sending a photo and/or sound of the user, and a plurality of emotion levels that can be selected to the other user; receiving an emotion level selected by the other user from the plurality of emotion levels as the tag;
    获取所述其他用户对所述用户的照片和/或声音的评价,从所述评价中提取情绪级别作为所述标签,其中,所述评价包括以下至少之一:自然语言的评价、语音的评价。Obtaining an evaluation of the photo and/or sound of the user by the other user, extracting an emotion level as the tag from the evaluation, wherein the evaluation includes at least one of the following: evaluation of natural language, evaluation of speech .
  4. 根据权利要求1所述的方法,其特征在于,在使用所述模型评估所述用户的信息对应的所述用户的情绪级别之前,所述方法还包括:The method according to claim 1, wherein before the evaluating the emotion level of the user corresponding to the information of the user by using the model, the method further comprises:
    在获取到所述用户的照片和/或声音之后,根据所述用户的照片和/或声音向所述用户进行提问;After obtaining the photo and/or sound of the user, asking the user according to the photo and/or sound of the user;
    根据所述用户对提问的问题的回答提取所述用户的照片和/或声音所对应的情绪级别。The emotion level corresponding to the photo and/or sound of the user is extracted according to the user's answer to the question questioned.
  5. 根据权利要求1所述的方法,其特征在于,根据所述情绪级别发送所述控制命令包括:The method according to claim 1, wherein the transmitting the control command according to the emotion level comprises:
    在所述情绪级别与预定级别相匹配的情况下,发送所述控制命令,其中,所述控制命令用于控制所述设备执行以下操作的至少之一:播放与所述情绪级别对应的音乐,播放与所述情绪级别对应的视频。Transmitting the control command when the emotion level matches a predetermined level, wherein the control command is used to control the device to perform at least one of: playing music corresponding to the emotion level, A video corresponding to the emotion level is played.
  6. 一种设备控制处理装置,其特征在于,包括:A device control processing device, comprising:
    第一获取单元,用于获取用户的信息,其中,所述用户的信息包括以下至少之一:通 过摄像设备拍摄到的所述用户的照片、通过音频设备接收到的所述用户的声音;a first acquiring unit, configured to acquire information of the user, where the information of the user includes at least one of the following: a photo of the user captured by the imaging device, and a voice of the user received by the audio device;
    评估单元,用于使用模型评估所述用户的信息对应的所述用户的情绪级别,其中,所述模型是根据多组数据训练得到的,所述多组数据中的每一组数据均包括:所述用户的照片和/或声音、以及用来标识该照片和/或声音所代表的情绪级别的标签;An evaluation unit, configured to use a model to evaluate an emotional level of the user corresponding to the information of the user, wherein the model is obtained according to a plurality of sets of data, each of the plurality of sets of data includes: a photo and/or sound of the user, and a label identifying the level of emotion represented by the photo and/or sound;
    第一发送单元,用于根据所述情绪级别发送控制命令,其中,所述控制命令用于指示设备进行预定操作。a first sending unit, configured to send a control command according to the sentiment level, wherein the control command is used to instruct the device to perform a predetermined operation.
  7. 根据权利要求6所述的装置,其特征在于,所述装置还包括:The device according to claim 6, wherein the device further comprises:
    第二发送单元,用于在使用所述模型评估所述用户的信息对应的所述用户的情绪级别之前,将所述用户的照片和/或声音发送给与其他用户;a second sending unit, configured to send the photo and/or voice of the user to other users before evaluating the emotion level of the user corresponding to the information of the user by using the model;
    第二获取单元,用于获取所述其他用户为所述用户的照片和/或声音添加的所述标签。And a second acquiring unit, configured to acquire the label added by the other user for the photo and/or sound of the user.
  8. 根据权利要求7所述的装置,其特征在于,所述第二获取单元包括以下至少之一:The apparatus according to claim 7, wherein said second acquisition unit comprises at least one of the following:
    第一发送模块,用于将所述用户的照片和/或声音,以及可以选择的多个情绪级别发送给所述其他用户;接收所述其他用户从所述多个情绪级别中选择出的情绪级别作为所述标签;a first sending module, configured to send a photo and/or a sound of the user, and a plurality of emotion levels that can be selected to the other user; and receive an emotion selected by the other user from the multiple emotion levels Level as the label;
    提取模块,用于获取所述其他用户对所述用户的照片和/或声音的评价,从所述评价中提取情绪级别作为所述标签,其中,所述评价包括以下至少之一:自然语言的评价、语音的评价。An extraction module, configured to obtain an evaluation of the photo and/or sound of the user by the other user, extracting an emotion level as the label from the evaluation, wherein the evaluation includes at least one of: natural language Evaluation, evaluation of speech.
  9. 根据权利要求6所述的装置,其特征在于,所述装置还包括:The device according to claim 6, wherein the device further comprises:
    提问单元,用于在使用所述模型评估所述用户的信息对应的所述用户的情绪级别之前,在获取到所述用户的照片和/或声音之后,根据所述用户的照片和/或声音向所述用户进行提问;a questioning unit, configured to: according to the photo and/or sound of the user, after acquiring the photo and/or sound of the user before using the model to evaluate the emotional level of the user corresponding to the information of the user Questioning the user;
    提取单元,用于根据所述用户对提问的问题的回答提取所述用户的照片和/或声音所对应的情绪级别。And an extracting unit, configured to extract an emotion level corresponding to the photo and/or the sound of the user according to the user's answer to the question questioned.
  10. 根据权利要求6所述的装置,其特征在于,所述第一发送单元包括:The apparatus according to claim 6, wherein the first sending unit comprises:
    第二发送模块,用于在所述情绪级别与预定级别相匹配的情况下,发送所述控制命令,其中,所述控制命令用于控制所述设备执行以下操作的至少之一:播放与所述情绪级别对应的音乐,播放与所述情绪级别对应的视频。a second sending module, configured to send the control command if the emotion level matches a predetermined level, where the control command is used to control the device to perform at least one of: playing and The music corresponding to the emotion level is played, and the video corresponding to the emotion level is played.
  11. 一种存储介质,其特征在于,所述存储介质包括存储的程序,其中,所述程序执行权利要求1至5中任意一项所述的设备控制处理方法。A storage medium, characterized in that the storage medium includes a stored program, wherein the program executes the device control processing method according to any one of claims 1 to 5.
  12. 一种处理器,其特征在于,所述处理器用于运行程序,其中,所述程序运行时执行权利要求1至5中任意一项所述的设备控制处理方法。A processor, wherein the processor is configured to execute a program, wherein the program is executed to perform the device control processing method according to any one of claims 1 to 5.
PCT/CN2018/100489 2017-10-31 2018-08-14 Device control processing method and apparatus WO2019085585A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201711062745.0A CN108039988B (en) 2017-10-31 2017-10-31 Equipment control processing method and device
CN201711062745.0 2017-10-31

Publications (1)

Publication Number Publication Date
WO2019085585A1 true WO2019085585A1 (en) 2019-05-09

Family

ID=62093587

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2018/100489 WO2019085585A1 (en) 2017-10-31 2018-08-14 Device control processing method and apparatus

Country Status (2)

Country Link
CN (1) CN108039988B (en)
WO (1) WO2019085585A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114446325A (en) * 2022-03-11 2022-05-06 平安普惠企业管理有限公司 Information pushing method and device based on emotion recognition, computer equipment and medium
CN115209048A (en) * 2022-05-19 2022-10-18 广东逸动科技有限公司 Image data processing method and device, electronic equipment and storage medium

Families Citing this family (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108039988B (en) * 2017-10-31 2021-04-30 珠海格力电器股份有限公司 Equipment control processing method and device
CN109118626B (en) * 2018-08-08 2022-09-13 腾讯科技(深圳)有限公司 Lock control method and device, storage medium and electronic device
KR20200035887A (en) * 2018-09-27 2020-04-06 삼성전자주식회사 Method and system for providing an interactive interface
CN109634129B (en) * 2018-11-02 2022-07-01 深圳慧安康科技有限公司 Method, system and device for realizing active care
CN109766776A (en) * 2018-12-18 2019-05-17 深圳壹账通智能科技有限公司 Operation executes method, apparatus, computer equipment and storage medium
CN109948780A (en) * 2019-03-14 2019-06-28 江苏集萃有机光电技术研究所有限公司 Aid decision-making method, device and equipment based on artificial intelligence
CN110096707B (en) * 2019-04-29 2020-09-29 北京三快在线科技有限公司 Method, device and equipment for generating natural language and readable storage medium
CN110197677A (en) * 2019-05-16 2019-09-03 北京小米移动软件有限公司 A kind of control method for playing back, device and playback equipment
CN110262413A (en) * 2019-05-29 2019-09-20 深圳市轱辘汽车维修技术有限公司 Intelligent home furnishing control method, control device, car-mounted terminal and readable storage medium storing program for executing
CN110491425A (en) * 2019-07-29 2019-11-22 恒大智慧科技有限公司 A kind of intelligent music play device
CN110412885A (en) * 2019-08-30 2019-11-05 北京青岳科技有限公司 A kind of household intelligent control system based on computer vision
JP7248615B2 (en) * 2020-03-19 2023-03-29 ヤフー株式会社 Output device, output method and output program
CN112631137A (en) * 2020-04-02 2021-04-09 张瑞华 Intelligent household control method and intelligent control equipment applied to biological feature recognition
CN113589697A (en) * 2020-04-30 2021-11-02 青岛海尔多媒体有限公司 Control method and device for household appliance and intelligent household appliance
CN112180747A (en) * 2020-09-28 2021-01-05 上海连尚网络科技有限公司 Method and equipment for adjusting intelligent household equipment
CN112464018A (en) * 2020-12-10 2021-03-09 山西慧虎健康科技有限公司 Intelligent emotion recognition and adjustment method and system
CN115047824A (en) * 2022-05-30 2022-09-13 青岛海尔科技有限公司 Digital twin multimodal device control method, storage medium, and electronic apparatus

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103024521A (en) * 2012-12-27 2013-04-03 深圳Tcl新技术有限公司 Program screening method, program screening system and television with program screening system
US20140192229A1 (en) * 2013-01-04 2014-07-10 Samsung Electronics Co., Ltd. Apparatus and method for providing user's emotional information in electronic device
CN106919821A (en) * 2015-12-25 2017-07-04 阿里巴巴集团控股有限公司 User authentication method and device
CN107272607A (en) * 2017-05-11 2017-10-20 上海斐讯数据通信技术有限公司 A kind of intelligent home control system and method
CN108039988A (en) * 2017-10-31 2018-05-15 珠海格力电器股份有限公司 Equipment control processing method and device

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103024521A (en) * 2012-12-27 2013-04-03 深圳Tcl新技术有限公司 Program screening method, program screening system and television with program screening system
US20140192229A1 (en) * 2013-01-04 2014-07-10 Samsung Electronics Co., Ltd. Apparatus and method for providing user's emotional information in electronic device
CN106919821A (en) * 2015-12-25 2017-07-04 阿里巴巴集团控股有限公司 User authentication method and device
CN107272607A (en) * 2017-05-11 2017-10-20 上海斐讯数据通信技术有限公司 A kind of intelligent home control system and method
CN108039988A (en) * 2017-10-31 2018-05-15 珠海格力电器股份有限公司 Equipment control processing method and device

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114446325A (en) * 2022-03-11 2022-05-06 平安普惠企业管理有限公司 Information pushing method and device based on emotion recognition, computer equipment and medium
CN115209048A (en) * 2022-05-19 2022-10-18 广东逸动科技有限公司 Image data processing method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN108039988B (en) 2021-04-30
CN108039988A (en) 2018-05-15

Similar Documents

Publication Publication Date Title
WO2019085585A1 (en) Device control processing method and apparatus
KR101803081B1 (en) Robot for store management
Chen et al. Hierarchical cross-modal talking face generation with dynamic pixel-wise loss
CN110291489B (en) Computationally efficient human identification intelligent assistant computer
CN106295313B (en) Object identity management method and device and electronic equipment
TWI661363B (en) Smart robot and human-computer interaction method
WO2021077382A1 (en) Method and apparatus for determining learning state, and intelligent robot
KR20100001928A (en) Service apparatus and method based on emotional recognition
TW201220216A (en) System and method for detecting human emotion and appeasing human emotion
US9661208B1 (en) Enhancing video conferences
CN109986553B (en) Active interaction robot, system, method and storage device
US11852357B2 (en) Method for controlling air conditioner, air conditioner
CN109241336A (en) Music recommendation method and device
WO2021200503A1 (en) Learning system and data collection device
KR20200012355A (en) Online lecture monitoring method using constrained local model and Gabor wavelets-based face verification process
JP2010224715A (en) Image display system, digital photo-frame, information processing system, program, and information storage medium
Błażek et al. An unorthodox view on the problem of tracking facial expressions
CN115867948A (en) Method for identifying hygiene condition of object and related electronic equipment
TW202303444A (en) Image processing based emotion recognition system and method
CN115988164A (en) Conference room multimedia control method, system and computer equipment
CN113591550B (en) Method, device, equipment and medium for constructing personal preference automatic detection model
JP2005199373A (en) Communication device and communication method
JP2021033359A (en) Emotion estimation device, emotion estimation method, program, information presentation device, information presentation method and emotion estimation system
Zhang et al. Quantification of advanced dementia patients’ engagement in therapeutic sessions: An automatic video based approach using computer vision and machine learning
Miao et al. Study of detecting behavioral signatures within DeepFake videos

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18872426

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18872426

Country of ref document: EP

Kind code of ref document: A1

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 02.10.2020)

122 Ep: pct application non-entry in european phase

Ref document number: 18872426

Country of ref document: EP

Kind code of ref document: A1