US20160066829A1 - Wearable mental state monitor computer apparatus, systems, and related methods - Google Patents
Wearable mental state monitor computer apparatus, systems, and related methods Download PDFInfo
- Publication number
- US20160066829A1 US20160066829A1 US14/578,039 US201414578039A US2016066829A1 US 20160066829 A1 US20160066829 A1 US 20160066829A1 US 201414578039 A US201414578039 A US 201414578039A US 2016066829 A1 US2016066829 A1 US 2016066829A1
- Authority
- US
- United States
- Prior art keywords
- wearer
- mental state
- processor
- eyewear
- computer
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
- A61B5/112—Gait analysis
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B3/00—Apparatus for testing the eyes; Instruments for examining the eyes
- A61B3/10—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions
- A61B3/11—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for measuring interpupillary distance or diameter of pupils
- A61B3/112—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for measuring interpupillary distance or diameter of pupils for measuring diameter of pupils
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0002—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0002—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network
- A61B5/0015—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network characterised by features of the telemetry system
- A61B5/0022—Monitoring a patient using a global network, e.g. telephone networks, internet
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0059—Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
- A61B5/0077—Devices for viewing the surface of the body, e.g. camera, magnifying lens
-
- A61B5/0402—
-
- A61B5/0476—
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/1032—Determining colour for diagnostic purposes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
- A61B5/1103—Detecting eye twinkling
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
- A61B5/1113—Local tracking of patients, e.g. in a hospital or private home
- A61B5/1114—Tracking parts of the body
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
- A61B5/1116—Determining posture transitions
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
- A61B5/1126—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb using a particular sensing technique
- A61B5/1128—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb using a particular sensing technique using image analysis
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/117—Identification of persons
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/117—Identification of persons
- A61B5/1171—Identification of persons based on the shapes or appearances of their bodies or parts thereof
- A61B5/1176—Recognition of faces
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/145—Measuring characteristics of blood in vivo, e.g. gas concentration, pH value; Measuring characteristics of body fluids or tissues, e.g. interstitial fluid, cerebral tissue
- A61B5/1455—Measuring characteristics of blood in vivo, e.g. gas concentration, pH value; Measuring characteristics of body fluids or tissues, e.g. interstitial fluid, cerebral tissue using optical sensors, e.g. spectral photometrical oximeters
- A61B5/14551—Measuring characteristics of blood in vivo, e.g. gas concentration, pH value; Measuring characteristics of body fluids or tissues, e.g. interstitial fluid, cerebral tissue using optical sensors, e.g. spectral photometrical oximeters for measuring blood gases
- A61B5/14552—Details of sensors specially adapted therefor
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/16—Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
- A61B5/165—Evaluating the state of mind, e.g. depression, anxiety
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/24—Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
- A61B5/316—Modalities, i.e. specific diagnostic methods
- A61B5/318—Heart-related electrical modalities, e.g. electrocardiography [ECG]
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/24—Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
- A61B5/316—Modalities, i.e. specific diagnostic methods
- A61B5/369—Electroencephalography [EEG]
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/40—Detecting, measuring or recording for evaluating the nervous system
- A61B5/4076—Diagnosing or monitoring particular conditions of the nervous system
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/42—Detecting, measuring or recording for evaluating the gastrointestinal, the endocrine or the exocrine systems
- A61B5/4261—Evaluating exocrine secretion production
- A61B5/4266—Evaluating exocrine secretion production sweat secretion
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/44—Detecting, measuring or recording for evaluating the integumentary system, e.g. skin, hair or nails
- A61B5/441—Skin evaluation, e.g. for skin disorder diagnosis
- A61B5/443—Evaluating skin constituents, e.g. elastin, melanin, water
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/48—Other medical applications
- A61B5/486—Bio-feedback
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/48—Other medical applications
- A61B5/4884—Other medical applications inducing physiological or psychological stress, e.g. applications for stress testing
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/68—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
- A61B5/6801—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
- A61B5/6802—Sensor mounted on worn items
- A61B5/6803—Head-worn items, e.g. helmets, masks, headphones or goggles
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
- A61B5/7246—Details of waveform analysis using correlation, e.g. template matching or determination of similarity
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7271—Specific aspects of physiological measurement analysis
- A61B5/7278—Artificial waveform generation or derivation, e.g. synthesising signals from measured signals
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B7/00—Instruments for auscultation
- A61B7/02—Stethoscopes
- A61B7/04—Electric stethoscopes
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B24/00—Electric or electronic controls for exercising apparatus of preceding groups; Controlling or monitoring of exercises, sportive games, training or athletic performances
- A63B24/0062—Monitoring athletic performances, e.g. for determining the work of a user on an exercise apparatus, the completed jogging or cycling distance
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/30—Authentication, i.e. establishing the identity or authorisation of security principals
- G06F21/31—User authentication
- G06F21/34—User authentication involving the use of external additional devices, e.g. dongles or smart cards
- G06F21/35—User authentication involving the use of external additional devices, e.g. dongles or smart cards communicating wirelessly
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/19—Sensors therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/197—Matching; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
- G06V40/23—Recognition of whole body movements, e.g. for sport training
- G06V40/25—Recognition of walking or running movements, e.g. gait recognition
-
- G—PHYSICS
- G07—CHECKING-DEVICES
- G07C—TIME OR ATTENDANCE REGISTERS; REGISTERING OR INDICATING THE WORKING OF MACHINES; GENERATING RANDOM NUMBERS; VOTING OR LOTTERY APPARATUS; ARRANGEMENTS, SYSTEMS OR APPARATUS FOR CHECKING NOT PROVIDED FOR ELSEWHERE
- G07C9/00—Individual registration on entry or exit
- G07C9/30—Individual registration on entry or exit not involving the use of a pass
- G07C9/32—Individual registration on entry or exit not involving the use of a pass in combination with an identity check
- G07C9/37—Individual registration on entry or exit not involving the use of a pass in combination with an identity check using biometric data, e.g. fingerprints, iris scans or voice recognition
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B21/00—Alarms responsive to a single specified undesired or abnormal condition and not otherwise provided for
- G08B21/02—Alarms for ensuring the safety of persons
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B21/00—Alarms responsive to a single specified undesired or abnormal condition and not otherwise provided for
- G08B21/02—Alarms for ensuring the safety of persons
- G08B21/04—Alarms for ensuring the safety of persons responsive to non-activity, e.g. of elderly persons
- G08B21/0407—Alarms for ensuring the safety of persons responsive to non-activity, e.g. of elderly persons based on behaviour analysis
- G08B21/0423—Alarms for ensuring the safety of persons responsive to non-activity, e.g. of elderly persons based on behaviour analysis detecting deviation from an expected pattern of behaviour or schedule
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B21/00—Alarms responsive to a single specified undesired or abnormal condition and not otherwise provided for
- G08B21/02—Alarms for ensuring the safety of persons
- G08B21/04—Alarms for ensuring the safety of persons responsive to non-activity, e.g. of elderly persons
- G08B21/0438—Sensor means for detecting
- G08B21/0461—Sensor means for detecting integrated or attached to an item closely associated with the person but not worn by the person, e.g. chair, walking stick, bed sensor
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B21/00—Alarms responsive to a single specified undesired or abnormal condition and not otherwise provided for
- G08B21/02—Alarms for ensuring the safety of persons
- G08B21/04—Alarms for ensuring the safety of persons responsive to non-activity, e.g. of elderly persons
- G08B21/0438—Sensor means for detecting
- G08B21/0476—Cameras to detect unsafe condition, e.g. video cameras
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B19/00—Teaching not covered by other main groups of this subclass
- G09B19/0092—Nutrition
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B5/00—Electrically-operated educational appliances
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B5/00—Electrically-operated educational appliances
- G09B5/06—Electrically-operated educational appliances with both visual and audible presentation of the material to be studied
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H20/00—ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance
- G16H20/40—ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance relating to mechanical, radiation or invasive therapies, e.g. surgery, laser therapy, dialysis or acupuncture
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H40/00—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
- G16H40/60—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices
- G16H40/63—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices for local operation
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H40/00—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
- G16H40/60—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices
- G16H40/67—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices for remote operation
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/20—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L63/00—Network architectures or network communication protocols for network security
- H04L63/08—Network architectures or network communication protocols for network security for authentication of entities
- H04L63/0861—Network architectures or network communication protocols for network security for authentication of entities using biometrical features, e.g. fingerprint, retina-scan
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2560/00—Constructional details of operational features of apparatus; Accessories for medical measuring apparatus
- A61B2560/04—Constructional details of apparatus
- A61B2560/0475—Special features of memory means, e.g. removable memory cards
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2562/00—Details of sensors; Constructional details of sensor housings or probes; Accessories for sensors
- A61B2562/02—Details of sensors specially adapted for in-vivo measurements
- A61B2562/0219—Inertial sensors, e.g. accelerometers, gyroscopes, tilt switches
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2562/00—Details of sensors; Constructional details of sensor housings or probes; Accessories for sensors
- A61B2562/02—Details of sensors specially adapted for in-vivo measurements
- A61B2562/0223—Magnetic field sensors
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2562/00—Details of sensors; Constructional details of sensor housings or probes; Accessories for sensors
- A61B2562/02—Details of sensors specially adapted for in-vivo measurements
- A61B2562/0257—Proximity sensors
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2576/00—Medical imaging apparatus involving image processing or analysis
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/02—Detecting, measuring or recording pulse, heart rate, blood pressure or blood flow; Combined pulse/heart-rate/blood pressure determination; Evaluating a cardiovascular condition not otherwise provided for, e.g. using combinations of techniques provided for in this group with electrocardiography or electroauscultation; Heart catheters for measuring blood pressure
- A61B5/0205—Simultaneously evaluating both cardiovascular conditions and different types of body conditions, e.g. heart and respiratory condition
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/02—Detecting, measuring or recording pulse, heart rate, blood pressure or blood flow; Combined pulse/heart-rate/blood pressure determination; Evaluating a cardiovascular condition not otherwise provided for, e.g. using combinations of techniques provided for in this group with electrocardiography or electroauscultation; Heart catheters for measuring blood pressure
- A61B5/024—Detecting, measuring or recording pulse rate or heart rate
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/05—Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves
- A61B5/053—Measuring electrical impedance or conductance of a portion of the body
- A61B5/0531—Measuring skin impedance
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/08—Detecting, measuring or recording devices for evaluating the respiratory organs
- A61B5/0816—Measuring devices for examining respiratory frequency
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
- A61B5/1118—Determining activity level
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7271—Specific aspects of physiological measurement analysis
- A61B5/7282—Event detection, e.g. detecting unique waveforms indicative of a medical condition
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61F—FILTERS IMPLANTABLE INTO BLOOD VESSELS; PROSTHESES; DEVICES PROVIDING PATENCY TO, OR PREVENTING COLLAPSING OF, TUBULAR STRUCTURES OF THE BODY, e.g. STENTS; ORTHOPAEDIC, NURSING OR CONTRACEPTIVE DEVICES; FOMENTATION; TREATMENT OR PROTECTION OF EYES OR EARS; BANDAGES, DRESSINGS OR ABSORBENT PADS; FIRST-AID KITS
- A61F2/00—Filters implantable into blood vessels; Prostheses, i.e. artificial substitutes or replacements for parts of the body; Appliances for connecting them with the body; Devices providing patency to, or preventing collapsing of, tubular structures of the body, e.g. stents
- A61F2/50—Prostheses not implantable in the body
- A61F2/76—Means for assembling, fitting or testing prostheses, e.g. for measuring or balancing, e.g. alignment means
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61F—FILTERS IMPLANTABLE INTO BLOOD VESSELS; PROSTHESES; DEVICES PROVIDING PATENCY TO, OR PREVENTING COLLAPSING OF, TUBULAR STRUCTURES OF THE BODY, e.g. STENTS; ORTHOPAEDIC, NURSING OR CONTRACEPTIVE DEVICES; FOMENTATION; TREATMENT OR PROTECTION OF EYES OR EARS; BANDAGES, DRESSINGS OR ABSORBENT PADS; FIRST-AID KITS
- A61F2/00—Filters implantable into blood vessels; Prostheses, i.e. artificial substitutes or replacements for parts of the body; Appliances for connecting them with the body; Devices providing patency to, or preventing collapsing of, tubular structures of the body, e.g. stents
- A61F2/50—Prostheses not implantable in the body
- A61F2/76—Means for assembling, fitting or testing prostheses, e.g. for measuring or balancing, e.g. alignment means
- A61F2002/7695—Means for testing non-implantable prostheses
-
- G—PHYSICS
- G02—OPTICS
- G02C—SPECTACLES; SUNGLASSES OR GOGGLES INSOFAR AS THEY HAVE THE SAME FEATURES AS SPECTACLES; CONTACT LENSES
- G02C11/00—Non-optical adjuncts; Attachment thereof
- G02C11/10—Electronic devices other than hearing aids
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2218/00—Aspects of pattern recognition specially adapted for signal processing
Definitions
- a computer-implemented method of assessing the mental state of a wearer of a wearable device comprises providing eyewear to the wearer that comprises at least one or more sensors coupled to the eyewear that are adapted to detect one or more characteristics of the wearer of the eyewear. The one or more characteristics are associated with the wearer's mental state.
- the system also receives one or more signals from the one or more sensors, wherein each of the one or more signals relates to at least one characteristic associated with the wearer.
- the at least one characteristic is selected from a group consisting of: (1) pupil size; (2) heart rate; (3) perspiration level; (4) respiration rate; (5) movement; and (6) brainwave activity.
- the system analyzes the one or more received signals to determine the at least one characteristic associated with the wearer.
- the system may then facilitate the determination of a mental state of the wearer based on the at least one characteristic, and associate the mental state of the wearer with an object and/or an activity.
- a computer-implemented method of assessing the mental state of a wearer of a wearable device comprises providing the wearer with eyewear comprising a front-facing camera and an eye-facing camera that are adapted to detect one or more characteristics of the wearer of the eyewear.
- the one or more characteristics of the wearer are associated with the wearer's mental state.
- the system receives one or more first images from the eye facing camera, wherein at least one of the one or more first images relates to at least one characteristic associated with the wearer.
- the system also analyzes, by a processor, the at least one or more first images to determine the at least one characteristic.
- the method includes facilitating determination of a mental state of the wearer based on the at least one characteristic.
- FIG. 1 is a block diagram of a Mental State Monitoring System in accordance with an embodiment of the present system.
- FIG. 2 is a block diagram of the Mental State Server of FIG. 1 .
- FIGS. 3A-3B depict a flowchart that generally illustrates various steps executed by a Mental State Monitoring Module according to a particular embodiment.
- FIG. 4 is a perspective view of eyewear that may serve as the Wearable Health Monitoring Device 156 of FIG. 1 .
- a wearable mental state monitoring system may, for example, be embodied in any suitable wearable device configured to monitor the mental state of a wearer.
- the system may, for example, be embodied as a pair of eyewear, as contact lenses, as a wristwatch, as a suitable piece of clothing (e.g., such as a suitable shirt, pair of pants, undergarment, compression sleeve, etc.), as footwear, as a hat, as a helmet, as an orthopedic cast, or any other suitable wearable item.
- a wearable mental state monitoring system embodied as a pair of eyewear may enable the system to access one or more (e.g., all five) of a wearer's senses (e.g., touch, sight, sound, smell, and taste) based at least in part on a proximity of the eyewear to the wearer's sensory systems (e.g., eyes, mouth, ears, nose) when worn by the wearer.
- a wearer's senses e.g., touch, sight, sound, smell, and taste
- the wearer's sensory systems e.g., eyes, mouth, ears, nose
- the system comprises one or more sensors configured to determine one or more attributes of the wearer's mental state.
- the one or more sensors may be coupled to the wearable device in any suitable way.
- the one or more sensors may be embedded into the wearable device, coupled to the wearable device, and/or operatively coupled to the wearable device.
- the one or more sensors may include, for example, one or more heart rate monitors, one or more electrocardiograms (EKG), one or more electroencephalograms (EEG), one or more pedometers, one or more thermometers, one or more transdermal transmitter sensors, one or more front-facing cameras, one or more eye-facing cameras, one or more microphones, one or more accelerometers, one or more gyroscopes, one or more blood pressure sensors, one or more pulse oximeters, one or more respiration rate sensors, one or more blood alcohol concentration (BAC) sensors, one or more near-field communication sensors, or any other suitable one or more sensors.
- EKG electrocardiograms
- EEG electroencephalograms
- pedometers one or more thermometers
- transdermal transmitter sensors one or more front-facing cameras
- one or more eye-facing cameras one or more microphones
- one or more accelerometers one or more gyroscopes
- one or more blood pressure sensors one or more pulse oximeters
- the system is configured to gather data, for example, using the one or more sensors, about the wearer (e.g., such as temperature, balance, heart rate, activity, activity levels, food eaten, medications taken, steps taken, position, movements, facial muscle movements, etc.).
- the wearer e.g., such as temperature, balance, heart rate, activity, activity levels, food eaten, medications taken, steps taken, position, movements, facial muscle movements, etc.
- the sensors sense the mental state of the wearer by monitoring certain characteristics of the wearer including changes in pupil size, heart rate, perspiration level, composition of the wearer's perspiration, respiration rate, movement, brainwave activity, and/or any other suitable characteristic.
- the system determines the wearer's mental state based on these characteristics. For instance, where the wearer has an increase in heart rate and/or perspiration level, the system may determine that the wearer is in a state of emotional stress. After determining the wearer's mental state, the system may notify the wearer or a third party (e.g., the wearer's physician) of the mental state. In particular embodiments, the system may also, or alternatively, save the mental state and related information to computer memory.
- the system may notify the wearer via the wearable device or through a notification sent to a mobile device associated with the wearer.
- the system may also provide the wearer with one or more suggestions on how to address the wearer's current mental state. For instance, when the wearer is in a state of emotional stress, the system may suggest that the user meditate, remove themselves from their current physical or social situation, execute one or more exercises, etc.
- the system may also (e.g., at least substantially simultaneously) capture one or more images of a person or other object located in close proximity to the wearer (e.g., using a camera, such as a forward-facing camera associated with eyewear worn by the wearer).
- the system may capture and analyze the image to determine whether the person or other object in the image caused the wearer's mental state. In other embodiments, the system may simply capture the image to determine what the user was viewing when they were under a state of emotional stress. If the system determines that the person or other object in the image caused the wearer's mental state, the system may notify the wearer (or other individual) of the association and may also provide the wearer suggested actions to address the wearer's mental state.
- the system may also capture one or more images to determine reoccurring mental states (e.g., mental state patterns) in the wearer's life. For instance, if on a first occasion when the wearer is speaking to a particular person, the wearer becomes emotionally stressed, the system may capture an image of the particular person. If later, on a second occasion when the wearer is speaking to the particular person, the wearer again becomes emotionally stressed, the system may again capture an image of the particular person. The system may compare the first image with the second image. If the system determines that the first and second images are of the same particular person, the system may notify the wearer that the wearer's mental state when speaking to the particular person is “stressed”. The system may then provide suggestions to the wearer on how to address the wearer's mental state when in close proximity to the particular person.
- mental states e.g., mental state patterns
- the system may use one or more similar techniques to determine that a user is typically in a particular mental state (e.g., relaxed, happy, emotionally stressed) when engaged in a particular activity (e.g., sailing, yoga, golf, commuting, a particular work activity).
- the system may be adapted to automatically determine that the wearer is engaged in a particular work activity based on information from one or more of the sensors that the wearer is currently wearing (e.g., eyewear based sensors).
- the system may use information from a forward facing camera installed in a pair of eyewear that the individual is wearing to determine that the individual is on a golf course (e.g., by detecting the presence of a hole, a pin, a fairway etc. . . . in one or more images taken by the forward-facing camera) or that the individual is driving (e.g., by detecting the presence of roadway “stripes” in an image from the forward-facing camera).
- the present systems and methods may be, for example, embodied as a computer system, a method, or a computer program product. Accordingly, various embodiments may be entirely hardware or a combination of hardware and software. Furthermore, particular embodiments may take the form of a computer program product stored on a computer-readable storage medium having computer-readable instructions (e.g., software) embodied in the storage medium. Various embodiments may also take the form of Internet-implemented computer software. Any suitable computer-readable storage medium may be utilized including, for example, hard disks, compact disks, DVDs, optical storage devices, and/or magnetic storage devices.
- the computer instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on a user's computer and partly on a remote computer, or entirely on the remote computer or server.
- the remote computer may be connected to the user's computer through any type of network, including but not limited to: a local area network (LAN); a wide area network (WAN); a cellular network; or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
- These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner such that the instructions stored in the computer-readable memory produce an article of manufacture that is configured for implementing the function specified in the flowchart block or blocks.
- the computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer-implemented process (e.g., method) such that the instructions that execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart block or blocks.
- FIG. 1 is a block diagram of a Mental State Monitoring System 100 according to particular embodiments.
- the Mental State Monitoring System 100 includes One or More Networks 115 , One or More Third Party Servers 50 , a Mental State Monitoring Server 120 that may, for example, be adapted to execute a Mental State Monitoring Module 300 , a Database 140 , One or More Remote Computing Devices 154 (e.g., such as a smart phone, a tablet computer, a wearable computing device, a laptop computer, a desktop computer, etc.), and One or More Wearable Health Monitoring Devices 156 , which may, for example, be embodied as one or more of eyewear, headwear, clothing, a watch, a hat, a helmet, a cast, an adhesive bandage, a piece of jewelry (e.g., a ring, earring, necklace, bracelet, etc.), or any other suitable wearable device.
- the one or more computer networks 115 facilitate communication between the One or More Third Party Servers 50
- the one or more networks 115 may include any of a variety of types of wired or wireless computer networks such as the Internet, a private intranet, a mesh network, a public switch telephone network (PSTN), or any other type of network (e.g., a network that uses Bluetooth or near field communications to facilitate communication between computing devices).
- the communication link between the One or More Remote Computing Devices 154 and the Mental State Monitoring Server 120 may be, for example, implemented via a Local Area Network (LAN) or via the Internet.
- LAN Local Area Network
- FIG. 2 illustrates a diagrammatic representation of the architecture for the Mental State Monitoring Server 120 that may be used within the Mental State Monitoring System 100 .
- the computer architecture shown in FIG. 2 may also represent the computer architecture for any one of the One or More Remote Computing Devices 154 , one or more Third Party Servers 50 , and one or more Health Monitoring Devices 156 shown in FIG. 1 .
- the Mental State Monitoring Server 120 may be suitable for use as a computer within the context of the Mental State Monitoring System 100 that is configured for determining a mental state of a wearer by detecting characteristics of the wearer using signals received from sensors coupled to the eyewear.
- the Mental State Monitoring Server 120 may be connected (e.g., networked) to other computing devices in a LAN, an intranet, an extranet, and/or the Internet as shown in FIG. 1 .
- the Mental State Monitoring Server 120 may operate in the capacity of a server or a client computing device in a client-server network environment, or as a peer computing device in a peer-to-peer (or distributed) network environment.
- the Mental State Monitoring Server 120 may be a desktop personal computing device (PC), a tablet PC, a set-top box (STB), a Personal Digital Assistant (PDA), a cellular telephone, a web appliance, a network router, a switch or bridge, or any other computing device capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that computing device.
- PC personal computing device
- PDA Personal Digital Assistant
- STB set-top box
- a Personal Digital Assistant a cellular telephone
- web appliance a web appliance
- network router a network router
- switch or bridge any other computing device capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that computing device.
- any other computing device capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that computing device.
- the term “computing device” shall also be interpreted to include any collection of computing devices that individually or jointly execute a set (or multiple sets) of instructions to
- An exemplary Mental State Monitoring Server 120 includes a processing device 202 , a main memory 204 (e.g., read-only memory (ROM), flash memory, dynamic random access memory (DRAM) such as synchronous DRAM (SDRAM) or Rambus DRAM (RDRAM), etc.), a static memory 206 (e.g., flash memory, static random access memory (SRAM), etc.), and a data storage device 218 , which communicate with each other via a bus 232 .
- main memory 204 e.g., read-only memory (ROM), flash memory, dynamic random access memory (DRAM) such as synchronous DRAM (SDRAM) or Rambus DRAM (RDRAM), etc.
- DRAM dynamic random access memory
- SDRAM synchronous DRAM
- RDRAM Rambus DRAM
- static memory 206 e.g., flash memory, static random access memory (SRAM), etc.
- SRAM static random access memory
- the processing device 202 represents one or more general-purpose or specific processing devices such as a microprocessor, a central processing unit (CPU), or the like. More particularly, the processing device 202 may be a complex instruction set computing (CISC) microprocessor, reduced instruction set computing (RISC) microprocessor, very long instruction word (VLIW) microprocessor, or processor implementing other instruction sets, or processors implementing a combination of instruction sets.
- the processing device 202 may also be one or more special-purpose processing devices such as an application specific integrated circuit (ASIC), a field programmable gate array (FPGA), a digital signal processor (DSP), network processor, or the like.
- the processing device 202 may be configured to execute processing logic 226 for performing various operations and steps discussed herein.
- the Mental State Monitoring Server 120 may further include a network interface device 208 .
- the Mental State Monitoring Server 120 may also include a video display unit 210 (e.g., a liquid crystal display (LCD) or a cathode ray tube (CRT)), an alpha-numeric input device 212 (e.g., a keyboard), a cursor control device 214 (e.g., a mouse), and a signal generation device 216 (e.g., a speaker).
- a video display unit 210 e.g., a liquid crystal display (LCD) or a cathode ray tube (CRT)
- an alpha-numeric input device 212 e.g., a keyboard
- a cursor control device 214 e.g., a mouse
- a signal generation device 216 e.g., a speaker
- the data storage device 218 may include a non-transitory computing device-accessible storage medium 230 (also known as a non-transitory computing device-readable storage medium, a non-transitory computing device-readable medium, or a non-transitory computer-readable medium) on which is stored one or more sets of instructions (e.g., the Mental State Monitoring Module 300 ) embodying any one or more of the methodologies or functions described herein.
- the one or more sets of instructions may also reside, completely or at least partially, within the main memory 204 and/or within the processing device 202 during execution thereof by the Mental State Monitoring Server 120 —the main memory 204 and the processing device 202 also constituting computing device-accessible storage media.
- the one or more sets of instructions may further be transmitted or received over a network 115 via a network interface device 208 .
- computing device-accessible storage medium 230 is shown in an exemplary embodiment to be a single medium, the term “computing device-accessible storage medium” should be understood to include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions.
- the term “computing device-accessible storage medium” should also be understood to include any medium that is capable of storing, encoding, or carrying a set of instructions for execution by the computing device and that causes the computing device to include any one or more of the methodologies of the present invention.
- the term “computing device-accessible storage medium” should accordingly be understood to include, but not be limited to, solid-state memories, optical and magnetic media, etc.
- a system is adapted to assess the mental state of a wearer of a wearable device.
- Various aspects of the system's functionality may be executed by certain system modules, including the Mental State Monitoring Module 300 .
- the Mental State Monitoring Module 300 is discussed in greater detail below.
- FIG. 3A is a flow chart of operations performed by an exemplary Mental State Monitoring Module 300 , which may, for example, run on the Mental State Monitoring Server 120 , or any suitable computing device (such as the One or More Health Monitoring Devices 156 or a suitable mobile computing device).
- the Mental State Monitoring Module 300 may assess a wearer's mental state and make suggestions to the wearer to address a particular mental state associated with the wearer.
- the system begins, in various embodiments, at Step 305 by providing eyewear comprising at least one or more sensors coupled to the eyewear.
- the system may do this by, for example: (1) facilitating delivery of the eyewear to an address associated with a particular individual; (2) facilitating distribution of the eyewear from a healthcare worker to the individual; and (3) placing an order of the eyewear from a third party for delivery to the individual.
- this step may be executed manually (e.g., by a human being) rather than a computer.
- the one or more sensors that are coupled to the eyewear are adapted to detect one or more characteristics of a wearer of the eyewear, wherein the one or more characteristics of the wearer are associated with the wearer's mental state.
- the sensors coupled to the eyewear or other health monitoring device may include, for example, one or more of the following: a heart rate monitor, an electrocardiogram (EKG), an electroencephalogram (EEG), a pedometer, a thermometer, a front-facing camera, an eye-facing camera, a microphone, an accelerometer, a gyroscope, a magnetometer, a blood pressure sensor, a pulse oximeter, a respiration rate sensor, a blood alcohol concentration (BAC) sensor, a skin conductance response sensor, a near-field communication sensor, or any other suitable sensor.
- the sensors coupled to the eyewear comprise an eye-facing camera, a front-facing camera, and a heart rate monitor
- the one or more sensors are coupled to a computing device that is associated with (e.g., embedded within, attached to) the eyewear or other health monitoring device.
- the eyewear or other health monitoring device comprises at least one processor, computer memory, suitable wireless communications components (e.g., a Bluetooth chip) and a power supply for powering the health monitoring device and/or the various sensors.
- the sensors may be physically coupled to the eyewear in any suitable way.
- the sensors may be embedded into the eyewear.
- the sensors may be positioned along the brow bar of the eyewear.
- the sensors may be positioned along the one or more of the temples of the eyewear.
- the sensors may be coupled to one or more of the lenses of the eyewear.
- the one or more sensors may be coupled to a Bluetooth device that is configured to transmit the one or more signals to a handheld wireless device, and the step of receiving one or more signals from the one or more sensors (discussed below in reference to Step 310 ) further comprises receiving the one or more signals from the wireless handheld device (e.g., via the Internet).
- one or more of the sensors may be detachable from the eyewear. For instance, if a wearer does not need a temperature sensor or other particular sensor, the sensor may be removed from the eyewear.
- the system continues, at Step 310 by receiving one or more signals from the one or more sensors, wherein each of the one or more signals relates to at least one characteristic associated with the wearer.
- the one or more signals that relate to the at least one characteristic associated with the wearer may include one or more signals that may be used to derive: (1) the wearer's a heart rate, (2) the wearer's heart rhythm; (3) a distance traveled by the wearer; (4) the wearer's body temperature; (5) one or more images associated with the wearer or the environment; (6) one or more sounds associated with the wearer's body or environment; (7) a speed traveled by the wearer; (8) the wearer's blood pressure; (9) the wearer's oxygen saturation level; (10) the wearer's brainwave activity (e.g., the location of the brainwave relative to the wearer's brain, the frequency of the brainwave, and the type of brainwave (e.g., gamma waves, beta waves, alpha waves, theta waves, and delta waves
- the system may receive a signal from an eye-facing camera associated with the eyewear that the wearer's brow is furrowed at the same time that the system receives a signal from the heart rate sensor that the wearer's heart rate is above a predetermined target heart rate.
- the system may store data related to the signals and/or data derived from this data for later review and use in determining the mental state of the wearer.
- the system may receive one or more of the above-referenced signals substantially automatically.
- the system may receive one or more of the signals on a substantially periodic basis (e.g., by the second, by the minute, hourly, daily, etc.). For example, the system may receive one or more signals every thirty seconds throughout the day.
- the system may receive one or more signals at least partially in response to receiving an indication from the wearer that the system should receive a signal. For instance, the wearer may speak a voice command to the wearable device requesting that the device take the wearer's blood pressure.
- the system may receive an indication from the wearer of when to have the system receive the signal.
- the system may receive an indication from the wearer to have the system conduct a brain scan of the user (e.g., receive at least one brainwave signal from the EEG) at 8:00 a.m. and at 2:00 p.m. on a particular day.
- the system may receive a request from the wearer to have a particular signal received from a particular sensor at the same time that the system receives a second particular signal from a second particular sensor.
- the system may, at least partially in response to receiving the increased respiration rate signal, also obtain an image of the wearer's eye from an eye-facing camera associated with the eyewear.
- the system receives a signal of an image captured by the eyewear. In various embodiments, the system receives a plurality of images captured by the eyewear. In particular embodiments, the system receives the image from the front-facing camera. In some embodiments, the system receives the image substantially automatically from the front-facing camera. In other embodiments, the system may receive the image in response to receiving an indication from the wearer to capture the image. For example, the system may receive a voice command from the wearer to capture the image. In various embodiments, the system may store the captured image in local or remote memory. In some embodiments, the image captured by the eyewear may be a video.
- the system may receive only one signal from a single sensor associated with the eyewear. In other embodiments, the system may receive a signal from a plurality of the sensors associated with the eyewear. In yet other embodiments, the system may receive multiple signals from one or more of the sensors. In various embodiments, the system may be configured to receive a first signal from a first sensor at the same time that it receives a second signal from a second sensor. For example, the system may be configured to receive an image signal from a front-facing camera associated with the eyewear at the same time that the system receives a heart rate signal from a heart rate sensor associated with the eyewear. As a further example, the system may be configured to simultaneously receive a signal from both an eye-facing camera and an EEG associated with the eyewear.
- the system analyzes the one or more received signals to determine at least one characteristic associated with the wearer. For example, the system may analyze the one or more received signals to determine that the at least one characteristic associated with the wearer is an increase or decrease in pupil size. In particular embodiments, the system may analyze the one or more received signals to determine that the at least one characteristic associated with the wearer is an increase or decrease in heart rate. In some embodiments, the system may analyze the one or more received signals to determine that the at least one characteristic associated with the wearer is an increase or decrease in the wearer's perspiration rate. In other embodiments, the system may analyze the one or more received signals to determine that the at least one characteristic associated with the wearer is an increase or decrease in respiration rate.
- the system may analyze the one or more received signals to determine that the at least one characteristic associated with the wearer is an increase or decrease in movement. In still other embodiments, the system may analyze the one or more received signals to determine that the at least one characteristic associated with the wearer is an increase or decrease in brainwave activity and/or frequency. In various embodiments, the system may analyze the one or more received signals to determine that the at least one characteristic associated with the wearer is a change in location of the wearer's brainwave activity. In yet other embodiments, the system may analyze the one or more received signals to determine that the at least one characteristic associated with the wearer is a change in wearer's the type of brainwave (e.g., a change from gamma waves to delta waves).
- the type of brainwave e.g., a change from gamma waves to delta waves.
- the system may analyze the one or more received signals to determine that the at least one characteristic associated with the wearer is an increase in heart rate in conjunction with an increase in pupil size. It should be understood from this disclosure that the system may analyze the one or more signals to determine that the at least one characteristic associated with the wearer is a combination characteristics that allow the system to determine the mental state of the wearer.
- the system may store the results of the analysis of the one or more received signals for later comparison with past and future analyses of the one or more received signals.
- the system may analyze a particular received signal at a particular time of day (e.g., morning, noon, night, etc.).
- the system may analyze a particular received signal at the same time that the system analyzes a second particular received signal.
- the system may analyze the front-facing camera signal when it analyzes the heart rate signal.
- the system may chart the analysis of the one or more received signals in a visual diagram. For example, the system may chart the changes in the wearer's heart rate in a diagram displayed e.g., on the lens of the eyewear or on a separate display screen associated with the eyewear.
- the system facilitates determination of a mental state of the wearer based on the at least one characteristic.
- the system may facilitate determination of the mental state of the wearer substantially automatically.
- the system may facilitate determination of the mental state of the wearer in response to receiving manual input of indication request from the wearer to determine the wearer's mental state. For example, the system may receive a voice command from the wearer requesting the system determine the mental state of the wearer at that time.
- the system may facilitate determination of the mental state of the wearer after comparing various signals to predetermined thresholds to establish that the signal is indicative of at least one characteristic.
- the system may facilitate determination of the mental state of the wearer at random.
- the system may determine the mental state of the wearer at different, randomly selected times throughout the day.
- the system may facilitate determination of the mental state of the wearer periodically throughout the day at specified times. For instance, the system may determine the mental state of the wearer at 8:00 a.m., 10:00 a.m., 12:00 p.m., 2:00 p.m., and 4:00 p.m. on a given day.
- the system may determine the mental state of the wearer at predetermined intervals of time. For instance, the system may determine the mental state of the wearer every 30 minutes or every two hours.
- the determination of the mental state of the wearer may be that the wearer's mental state is in a particular emotional state (e.g., happy, sad, anxious, calm, scared, angry, surprised, ashamed, envious, curious, relaxed, emotionally stressed, confused, moody, etc.).
- the determination of the mental state of the wearer may be that the wearer is under emotional stress (i.e., the wearer's mental state is “emotionally stressed”).
- the determination of the mental state of the wearer may include the wearer's experience of the wearer's current physical state (e.g., the wearer's experience of bodily pain, nausea, rapid heartbeat, etc.).
- the determination of the wearer's mental state may include determining that the wearer is responding mentally and/or physically to sustaining a concussion.
- the determination of the wearer's mental state may include a determination of the wearer's cognitive state (e.g., inability to concentrate, poor judgment, racing thoughts, constant worrying, etc.).
- the determination of the wearer's mental state may be that the wearer is having memory problems.
- the determination of the mental state of the wearer may include a mental state associated with a particular behavior (e.g., eating more or less, sleeping too much or too little, isolation from others, procrastinating, neglecting responsibilities, using alcohol or drugs, nervous habits or twitches, etc.).
- the determination of the mental state may be that the wearer is overly sleepy (which the system may determine, for example, by determining that the wearer is sleeping more than a predetermined amount of time per day).
- the system may determine the mental state of the wearer using a scaled rating. For example, where the system is determining a mental state such as stress of the wearer, the system may determine that the wearer's mental state (e.g., stress level) is low, moderate, or high. In some embodiments, the system may determine that the mental state of the wearer includes multiple mental states. For instance, the system may determine that the wearer is both emotionally stressed and anxious.
- a mental state such as stress of the wearer
- the system may determine that the wearer's mental state (e.g., stress level) is low, moderate, or high.
- the system may determine that the mental state of the wearer includes multiple mental states. For instance, the system may determine that the wearer is both emotionally stressed and anxious.
- the system may optionally associate the mental state of the wearer with at least one object (e.g., an animate object, such as a particular person or animal, or an inanimate object, such as a scary clown doll), at least one activity (e.g., bowling, playing tennis, driving, working on a computer, or running), at least one external condition (e.g., the user's current workload at work, the user's current credit rating, the user's marital or dating status), and/or at least one internal condition (the user's current weight, health, etc. . . . ).
- object e.g., an animate object, such as a particular person or animal, or an inanimate object, such as a scary clown doll
- at least one activity e.g., bowling, playing tennis, driving, working on a computer, or running
- at least one external condition e.g., the user's current workload at work, the user's current credit rating, the user's marital or dating status
- the system may associate the mental state of the wearer with both an animate object and an inanimate object (e.g., a person and an object—for example, the system may determine that the wearer is typically emotionally stressed when they see both the wearer's spouse and a checkbook), or any other combination of factors described herein.
- the system may associate the mental state of the wearer with both a particular object and an activity (for example, the system may determine that the wearer is typically relaxed when the wearer's spouse is present and the wearer is watching TV).
- the system may associate the mental state of the wearer with the user's external or internal context.
- the external or internal context may include one or more of the following for a particular time period: (1) the user's current health; (2) the user's general state of mind; (3) the user's current spending behavior; (4) the current weather conditions at the user's current location (e.g., pollen count, UV index, air quality, precipitation, wind speed and direction, barometric pressure, humidity, outdoor temperature, season, cloud levels); (5) one or more social media entries made by or received by the individuals at the particular time; (6) the user's credit rating at the particular time; (7) the user's employment status at the particular time; (8) the user's housing information for the particular time; (9) information from one or more e-mails at or around the particular time; (10) the user's exercise activity during the particular time; (11) indoor temperature within the wearer's home or workplace; and/or (12) any other external information that may be relevant to and/or have an impact
- the system may receive information from and/or regarding, for example: (1) the wearer's health records; (2) the wearer's genetics; (3) the wearer's family history; (4) one or more of the wearer's physical attributes; (5) the wearer's workout schedule; (6) one or more social media accounts associated with the wearer; (7) one or more of the wearer's social behaviors; (8) the wearer's arrest history; (9) the weather at a user's current location (e.g., as determined from a suitable weather service); (10) the user's calendar; (11) the user's spending behaviors; (12) the wearer's credit history; (13) the wearer's employment status and/or history; (14) the wearer's marital status; (15) the wearer's current residence; (16) an email account of the wearer; (17) the wearer's travel history; (18) the wearer's aspirations; (19) the wearer's goals; (20) dietary information for the wearer, etc.
- the wearer's health records
- the system may associate the wearer's mental state (e.g., “emotionally stressed”, “unhappy”, etc. . . . ) with the activity of a paying bills by: (1) identifying an entry on the wearer's electronic calendar that indicates that the wearer will be paying bills in a particular time slot on a particular day; (2) confirming that the user is paying bills in the particular time slot by identifying a checkbook in an image taken by the system's front facing camera during the particular time slot; and (3) determining the wearer's mental state during the time slot using any suitable technique, such as those described herein.
- the wearer's mental state e.g., “emotionally stressed”, “unhappy”, etc. . . .
- Similar techniques may be used to determine longer-lasting impacts of certain external or internal conditions on a wearer's mental state.
- the system may use any suitable technique to determine the wearer's general mental state (e.g., relatively happy, relatively sad, under relative stress) over a predetermined number of hours, days, weeks, or months, by comparing the wearer's aggregate (e.g., average) mental state for that time period with the wearer's aggregate mental state for one or more time periods (e.g., of similar length), or by comparing the wearer's aggregate mental state for the time period with typical mental states of the wearer or one or more other individuals.
- the wearer's general mental state e.g., relatively happy, relatively sad, under relative stress
- the wearer's aggregate mental state e.g., average mental state for that time period with the wearer's aggregate mental state for one or more time periods (e.g., of similar length)
- time periods e.g., of similar length
- the system may then access any suitable information (e.g., any of the types of object, activity, internal context or external context information described herein, or other information) to determine what may be contributing to the mental state (e.g., using any suitable data analysis techniques). For example, the system may determine that the wearer is typically relatively happy in weeks in which the wearer is exercising more than five times per week and has a balance of less than $1,000 on their personal credit card, and experiences relatively high stress levels in weeks in which the user is exercising less than two times per week.
- any suitable information e.g., any of the types of object, activity, internal context or external context information described herein, or other information
- the system may determine that the wearer is typically relatively happy in weeks in which the wearer is exercising more than five times per week and has a balance of less than $1,000 on their personal credit card, and experiences relatively high stress levels in weeks in which the user is exercising less than two times per week.
- the system may associate the mental state of the wearer with a person, object, activity, internal context, or external context using one or more signals received from the eyewear's sensors and/or any other suitable data received by the system.
- the system may associate the mental state of the wearer with a person, object, or activity by receiving a manual input from the wearer. For instance, the system may receive indication request from the wearer to associate a particular person with the wearer's current mental state. In these embodiments, the system may then capture an image of the person being viewed by the wearer, identify the person in the image and associate the current mental state of the wearer with the identified person.
- the system may also monitor the wearer's mental state when the wearer next encounters the identified person and track the wearer's mental state over time each time the wearer encounters the identified person. In this way, the system can alert the wearer if the identified person continually causes the wearer to experience the same or similar mental state each time the wearer encounters the identified person.
- the system may associate the mental state of the wearer with (e.g., an animate or inanimate object) or activity substantially automatically after (e.g., in response to) determining the mental state of the wearer.
- the system may be configured to associate the mental state of the wearer with an animate or inanimate object by examining a received image of a first object (e.g., a first person) located in the received image.
- the system determines the object in the received image by identifying the object located in the image using any suitable image recognition techniques.
- the system determines the object in the received image by comparing the image of the object with one or more stored images. At least partially in response to determining that the image of the object at least substantially matches a stored image, the system may identify the object as being the same as the known object (e.g., person or thing) in the stored image.
- the system may associate the mental state of the wearer with an object based on the proximity of the object to the wearer. For example, the system may use a front-facing camera to determine that a particular person is in close proximity to the wearer (e.g., in front of the wearer) when the wearer is experiencing a particular mental state such as stress.
- the proximity of a person to the wearer may be obtained by an electronic device on the person (e.g., a cellphone, an RFID tag, etc.).
- the system may associate the mental state of the wearer with an activity based on the wearer performing the activity. For example, the system may use the system's processor and front facing camera to determine, from images taken by the front-facing camera, that the user is climbing stairs. The system may then assess the wearer's mental state (e.g., in any suitable way described herein) and associate that that mental state with stair climbing. For example, the system may determine that the user is typically mentally relaxed when climbing stairs.
- the system stores, in memory, the association between the wearer's mental state and the object, activity, internal context, and/or external context causing the wearer's mental state.
- the system may store the association between the wearer's mental state and the object, activity, external or internal context substantially automatically after (e.g., in response to) making the association between the wearer's mental state and the object, activity, internal context and/or external context.
- the system may store the association after (e.g., in response to) receiving manual input from the wearer requesting that the system store the association.
- the system may store the association for a specified period of time.
- the system may store the association for a day, a month, a year, etc. in the system's Database 140 .
- the system may store the association on any suitable server, database, or device.
- the system may store the association on the Mental State Monitoring Server 120 .
- the system notifies the wearer or other suitable individual of the association of the wearer's mental state and the object, activity, internal context, and/or external context. For example, in some embodiments, the system may notify the wearer and the wearer's physician of the association of the wearer's mental state with the object in the received image and/or the activity that the wearer was engaged in while experiencing the mental state. In various embodiments, the system notifies the wearer of the association by displaying an image on the lens of the eyewear, or a display screen associated with the eyewear. In other embodiments, the system notifies the wearer of the association by communicating the association through an audio speaker to the wearer.
- the system notifies the wearer of the association by sending a notification to the wearer's mobile device.
- the system notifies the wearer of the association via an electronic communication such as an email or text message.
- the system may notify the wearer of a single association substantially immediately after (e.g., in response to) the system associates the wearer's mental state with a particular object.
- the system may notify the wearer of all associations made on a particular day (or within another particular time period).
- the system may notify the wearer of the association after (e.g., in response to) detecting a particular event. For example, the system may notify the wearer of the association after the system no longer detects the presence or proximity of a particular object. In some embodiments, the system may notify the wearer of the association after a particular period of time. For instance, the system may notify the wearer of an association one hour after the system associates the wearer's mental state with a particular object or activity. In still other embodiments, the system may notify the wearer of the association at a particular time of day. As an example, the system may notify the wearer of an association between the wearer's mental state and an object, activity, internal context, and/or external context at the end of the day.
- the system provides the wearer with one or more suggested actions to address the wearer's current mental state.
- the system may provide suggested actions to the wearer in any suitable way.
- the system may provide one or more suggested actions to the wearer by displaying an image on the lens of the eyewear, or on a display screen associated with the eyewear.
- the system may provide one or more suggested actions to the wearer by communicating through an audio speaker to the wearer.
- the system provides one or more suggested actions to the wearer by sending a notification to the wearer's mobile device.
- the system may provide one or more suggested actions to the wearer via an electronic communication, such as an email or text message.
- the system may provide a single suggested action to the wearer.
- the system may provide multiple suggested actions to the wearer.
- the system may provide suggested actions to the wearer to address the wearer's current mental state substantially immediately after the system notifies the wearer of the association between the wearer's mental state and the particular object, activity, internal context and/or external context.
- the system may provide one or more suggested actions to the wearer after (e.g., in response to) detecting a particular event. For example, the system may provide suggested actions to the wearer after the system no longer detects the presence or proximity of a particular object.
- the system may provide one or more suggested actions to the wearer after a particular period of time. For instance, the system may provide suggested actions to the wearer one hour after the system associates the wearer's mental state with the particular object.
- the system may provide one or more suggested actions to the wearer at a particular time of day. As an example, the system may provide one or more suggested actions to the wearer at the end of the day.
- the one or more suggested actions to address the wearer's current mental state may be one or more techniques for reducing stress.
- the suggested actions to address the wearer's current mental state may include, for example: (1) taking medication; (2) praying; (3) engaging in one or more yoga poses; (4) taking deep breaths; (5) avoiding a particular object (e.g., a particular person or thing) based on the association between the wearer's mental state and the object.
- the suggested actions to address the wearer's current mental state may include suggesting that the wearer visit links to websites containing information on the particular mental state of the wearer.
- the suggested actions to address the wearer's current mental state may include a listing of suggested applications on the wearer's mobile device. For instance, where the wearer's current mental state is in a stressed mental state, the system may suggest a yoga application on the wearer's mobile device to assist in improving the wearer's mental state. As a further example, if the system determines that the wearer typically experiences high levels of stress when the wearer is carrying high amounts of credit card debt, the system may suggest that the user reduce their debt levels.
- the system when executing the Mental State Monitoring Module 300 , may omit particular steps, perform particular steps in an order other than the order presented above, or perform additional steps not discussed directly above.
- various steps executed “in response to” a particular event occurring, or a particular condition being satisfied may also be executed “at least partially in response to” the particular event occurring or the particular condition being satisfied.
- eyewear 400 includes: (1) an eyewear frame 410 ; (2) a first temple 412 ; and (3) a second temple 414 . These various components are discussed in more detail below.
- the eyewear 400 is computerized and may serve as the wearable health monitoring device 156 of FIG. 1 .
- eyewear 400 in various embodiments, includes any suitable eyewear frame 410 configured to support one or more lenses 418 , 420 .
- the eyewear frame 410 has a first end 402 and a second end 404 .
- the eyewear frame 410 may be made of any suitable material such as metal, ceramic, polymers or any combination thereof.
- the eyewear frame 410 is configured to support the first and second lenses 418 , 420 about the full perimeter of the first and second lenses 418 , 420 .
- the eyewear frame 410 may be configured to support the first and second lenses 418 , 420 about only a portion of each respective lens.
- the eyewear frame 410 is configured to support a number of lenses other than two lenses (e.g., a single lens, a plurality of lenses, etc.).
- the lenses 418 , 420 may include prescription lenses, sunglass lenses, or any other suitable type of lens (e.g., reading lenses, non-prescription lenses), which may be formed from glass or polymers.
- the eyewear frame 410 includes a first and second nose pad 422 (not shown in figure), 424 , which may be configured to maintain the eyewear 400 adjacent the front of a wearer's face such that the lenses 418 , 420 are positioned substantially in front of the wearer's eyes while the wearer is wearing the eyewear 400 .
- the nose pads 422 , 424 may comprise a material that is configured to be comfortable when worn by the wearer (e.g., rubber, etc.).
- the nose pads may include any other suitable material (e.g., plastic, metal, etc.).
- the nose pads may be integrally formed with the frame 410 .
- the eyewear frame 410 includes a first and second hinge 426 , 428 that attach the first and second temples 412 , 414 to the frame first and second ends 402 , 404 , respectively.
- the hinges may be formed by any suitable connection (e.g., tongue and groove, ball and socket, spring hinge, etc.).
- the first hinge 426 may be welded to, or integrally formed with, the frame 410 and the first temple 412 and the second hinge 428 may be welded to, or integrally formed with, the frame 410 and the second temple 414 .
- the first temple 412 is rotatably connected to the frame 410 at a right angle to extend the first temple 412 substantially perpendicular, substantially parallel, or anywhere in between the right angle to the frame 410 .
- the first temple 412 has a first and second end 412 a , 412 b .
- the first temple 412 includes an earpiece 413 configured to be supported by a wearer's ear.
- the second temple 414 is rotatably connected to the frame 410 at a right angle to extend the second temple 414 substantially perpendicular, substantially parallel, or anywhere in between the right angle to the frame 410 .
- the second temple 414 has a first and second end 414 a , 414 b .
- the second temple 414 includes an earpiece 415 configured to be supported by a wearer's ear.
- the second temple 414 has one or more sensors 430 connected to the second temple 414 .
- the one or more sensors 430 may be coupled to the frame 410 , the first and second temples 412 , 414 , the first and second lenses 418 , 410 , or any other portion of the eyewear 400 in any suitable way.
- the one or more sensors 430 may be embedded into the eyewear 400 , coupled to the eyewear 400 , and/or operatively coupled to the eyewear 400 .
- the one or more sensors may be formed at any point along the eyewear 400 .
- a fingerprint reader may be disposed adjacent the first temple of the eyewear 400 .
- the one or more sensors may be formed in any shape.
- the one or more sensors may be formed on the inner (back) surface of the frame 410 , the first and second temples 412 , 414 , the first and second lenses 418 , 410 , or any other portion of the eyewear 400 .
- the one or more sensors may be formed on the outer (front) surface of the frame 410 , the first and second temples 412 , 414 , the first and second lenses 418 , 410 , or any other portion of the eyewear 400 .
- the one or more sensors 430 that are coupled to the eyewear (or other wearable device) are adapted to detect one or more characteristics of the eyewear or a wearer of the eyewear, wherein the one or more characteristics of the wearer are associated with the wearer's identity.
- the one or more sensors coupled to the eyewear or other health monitoring device may include, for example, one or more of the following: a near-field communication sensor, a Bluetooth chip, a GPS unit, an RFID tag (passive or active), a fingerprint reader, an iris reader, a retinal scanner, a voice recognition sensor, a heart rate monitor, an electrocardiogram (EKG), an electroencephalogram (EEG), a pedometer, a thermometer, a front-facing camera, an eye-facing camera, a microphone, an accelerometer, a magnetometer, a blood pressure sensor, a pulse oximeter, a skin conductance response sensor, any suitable biometric reader, or any other suitable sensor.
- the one or more sensors may include a unique shape, a unique code, or a unique design physically inscribed into the eyewear that may be readable by an individual or a remote computing device.
- the sensors coupled to the eyewear may include one or more electronic communications devices such as a near field communication sensor, a Bluetooth chip, an active RFID, and a GPS unit.
- the one or more sensors are coupled to a computing device that is associated with (e.g., embedded within, attached to) the eyewear or other wearable device.
- the eyewear or other wearable device comprises at least one processor, computer memory, suitable wireless communications components (e.g., a Bluetooth chip) and a power supply for powering the wearable device and/or the various sensors.
- the one or more sensors may be coupled to a Bluetooth device that is configured to transmit the one or more signals to a handheld wireless device, and the step of using the eyewear to confirm the identity of the wearer of the eyewear (discussed above in reference to Step 310 ) further comprises receiving the one or more signals from the wireless handheld device (e.g., via the Internet).
- the wireless handheld device e.g., via the Internet
- one or more of the sensors may be detachable from the eyewear. For instance, if a wearer does not need a temperature sensor or other particular sensor, the sensor may be removed from the eyewear.
- the wearer may put on the wearable device in the morning and continue to wear the device throughout the day.
- the system tracks the brainwave activity, including the types of brainwaves, of the wearer using the system's EEG.
- the system particularly tracks the gamma wave activity of the wearer, which may at the optimal range, for example, signify a high focused mental state of the wearer.
- the system may store the brainwave activity and chart the highs and lows of the wearer's brainwave activity throughout the day in order to determine the times throughout the day when the wearer is most focused. For example, where the wearer has relatively high gamma wave activity from 9:00 a.m.
- the system will notify the wearer that the wearer's highest cognitive functioning occurs from 9:00-11:00 a.m. and the wearer's lowest cognitive functioning occurs from 2:00-3:00 p.m. The wearer may then use this information to alter the wearer's activities so that the wearer is focused on important matters from 9:00-11:00 a.m. and less important matters from 2:00-3:00 p.m.
- the system in a particular example, will also track low-productivity states. This period of low-productivity may be measured, for example, by tracking the wearer's brainwave activity, the wearer's movements, the distance traveled by the wearer, and/or the wearer's speed of travel.
- the system may also capture images of the wearer's surroundings for use, for example, in determining one or more activities that the wearer is engaged in.
- the system may capture a plurality of images from the system's front-facing camera in order to determine that the wearer has been sitting in front of a television during those periods of time. The system may then track the amount of time that the wearer spent in the particular state and then notify the wearer and provide the wearer with one or more recommendations, such as to exercise rather than watch TV, or to exercise while watching TV.
- the system tracks the wearer's various mental states through a typical day via the wearable device (e.g., a pair of eyewear).
- the system tracks the wearer's mental state by monitoring the wearer's pupil size, heart rate, perspiration rate, respiration rate, movement, and brainwave activity.
- the system may also monitor potential causes of a certain change in the wearer's pupil size, heart rate, perspiration level, respiration rate, movement, and brainwave activity using a front-facing camera.
- the system captures an image of the person causing the high stress.
- the system may then present this image to the wearer with a suggestion to not interact in the future with the particular person.
- the system may also suggest other ways of interacting with the person in the future should avoidance be impractical. For instance, if the wearer works for the particular person that causes the wearer high stress, the system may suggest writing down what the wearer needs to speak to the person about prior to interacting with the person.
- the system may, for example, provide these suggestions to the wearer through a notification sent to the wearer's mobile device so that other people around the wearer are not made aware of the wearer's high stress.
- the system may monitor and track one or more specific behaviors and/or movements of the wearer in order to diagnose particular mental states. For example, the system may track the wearer's movements to determine whether a wearer frequently has one or more simple tics such as motor tics including, for example, eye blinking or other nervous eye movements, facial grimacing, shoulder shrugging, muscle twitches, head or shoulder jerking, or one or more complex tics such as facial grimacing combined with a head twist and a shoulder shrug.
- simple tics such as motor tics including, for example, eye blinking or other nervous eye movements, facial grimacing, shoulder shrugging, muscle twitches, head or shoulder jerking, or one or more complex tics such as facial grimacing combined with a head twist and a shoulder shrug.
- the system may notify the wearer or the wearer's physician that the wearer has one or more behavioral characteristics that are consistent with a neurological condition such as Tourette syndrome.
Abstract
A computer-implemented method of assessing the metal state of an individual by: (1) providing the individual with a wearable device (e.g., eyewear) that includes one or more sensors for assessing the mental state of the individual, (2) using information from one or more of the sensors to assess the mental state of the individual; and (3) informing the individual or a third party of the individual's mental state. In various embodiments, the method further involves using the wearable device to determine one or more environmental factors that are related to the individual's mental state. For example, the method may involve determining (e.g., from one or more images taken using the wearable device) that the individual is frequently in a stressed emotional state when a particular person is present, when the individual is engaged in a particular activity, and/or when the wearer experiences a certain internal or external context.
Description
- This application claims the benefit of U.S. Provisional Patent Application No. 62/046,406, filed Sep. 5, 2014, entitled, “Wearable Health Computer Apparatus, Systems, and Related Methods,” which is hereby incorporated herein by reference in its entirety.
- Being able to detect stress in a person and the cause of that stress based on personal and environmental factors is of importance to many people. Accordingly, there is a need for improved systems and methods for monitoring and detecting stress for an individual. Various embodiments of the present systems and methods recognized and address the foregoing considerations, and others, of prior art systems and methods.
- In general, in various embodiments, a computer-implemented method of assessing the mental state of a wearer of a wearable device comprises providing eyewear to the wearer that comprises at least one or more sensors coupled to the eyewear that are adapted to detect one or more characteristics of the wearer of the eyewear. The one or more characteristics are associated with the wearer's mental state. The system also receives one or more signals from the one or more sensors, wherein each of the one or more signals relates to at least one characteristic associated with the wearer. The at least one characteristic is selected from a group consisting of: (1) pupil size; (2) heart rate; (3) perspiration level; (4) respiration rate; (5) movement; and (6) brainwave activity. In various embodiments, the system analyzes the one or more received signals to determine the at least one characteristic associated with the wearer. The system may then facilitate the determination of a mental state of the wearer based on the at least one characteristic, and associate the mental state of the wearer with an object and/or an activity.
- In various embodiments, a computer-implemented method of assessing the mental state of a wearer of a wearable device comprises providing the wearer with eyewear comprising a front-facing camera and an eye-facing camera that are adapted to detect one or more characteristics of the wearer of the eyewear. The one or more characteristics of the wearer are associated with the wearer's mental state. The system receives one or more first images from the eye facing camera, wherein at least one of the one or more first images relates to at least one characteristic associated with the wearer. The system also analyzes, by a processor, the at least one or more first images to determine the at least one characteristic. The method includes facilitating determination of a mental state of the wearer based on the at least one characteristic.
- Various embodiments of systems and methods for assessing a user's mental state are described below. In the course of this description, reference will be made to the accompanying drawings, which are not necessarily drawn to scale and wherein:
-
FIG. 1 is a block diagram of a Mental State Monitoring System in accordance with an embodiment of the present system. -
FIG. 2 is a block diagram of the Mental State Server ofFIG. 1 . -
FIGS. 3A-3B depict a flowchart that generally illustrates various steps executed by a Mental State Monitoring Module according to a particular embodiment. -
FIG. 4 is a perspective view of eyewear that may serve as the Wearable HealthMonitoring Device 156 ofFIG. 1 . - Various embodiments will now be described more fully hereinafter with reference to the accompanying drawings. It should be understood that the invention may be embodied in many different forms and should not be construed as limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the invention to those skilled in the art. Like numbers refer to like elements throughout.
- A wearable mental state monitoring system, in various embodiments, may, for example, be embodied in any suitable wearable device configured to monitor the mental state of a wearer. The system may, for example, be embodied as a pair of eyewear, as contact lenses, as a wristwatch, as a suitable piece of clothing (e.g., such as a suitable shirt, pair of pants, undergarment, compression sleeve, etc.), as footwear, as a hat, as a helmet, as an orthopedic cast, or any other suitable wearable item. In a particular example, a wearable mental state monitoring system embodied as a pair of eyewear may enable the system to access one or more (e.g., all five) of a wearer's senses (e.g., touch, sight, sound, smell, and taste) based at least in part on a proximity of the eyewear to the wearer's sensory systems (e.g., eyes, mouth, ears, nose) when worn by the wearer.
- In various embodiments, the system comprises one or more sensors configured to determine one or more attributes of the wearer's mental state. The one or more sensors may be coupled to the wearable device in any suitable way. For instance, the one or more sensors may be embedded into the wearable device, coupled to the wearable device, and/or operatively coupled to the wearable device. The one or more sensors may include, for example, one or more heart rate monitors, one or more electrocardiograms (EKG), one or more electroencephalograms (EEG), one or more pedometers, one or more thermometers, one or more transdermal transmitter sensors, one or more front-facing cameras, one or more eye-facing cameras, one or more microphones, one or more accelerometers, one or more gyroscopes, one or more blood pressure sensors, one or more pulse oximeters, one or more respiration rate sensors, one or more blood alcohol concentration (BAC) sensors, one or more near-field communication sensors, or any other suitable one or more sensors. In particular embodiments, the system is configured to gather data, for example, using the one or more sensors, about the wearer (e.g., such as temperature, balance, heart rate, activity, activity levels, food eaten, medications taken, steps taken, position, movements, facial muscle movements, etc.).
- In various embodiments, the sensors sense the mental state of the wearer by monitoring certain characteristics of the wearer including changes in pupil size, heart rate, perspiration level, composition of the wearer's perspiration, respiration rate, movement, brainwave activity, and/or any other suitable characteristic. The system then determines the wearer's mental state based on these characteristics. For instance, where the wearer has an increase in heart rate and/or perspiration level, the system may determine that the wearer is in a state of emotional stress. After determining the wearer's mental state, the system may notify the wearer or a third party (e.g., the wearer's physician) of the mental state. In particular embodiments, the system may also, or alternatively, save the mental state and related information to computer memory. In various embodiments, the system may notify the wearer via the wearable device or through a notification sent to a mobile device associated with the wearer. The system may also provide the wearer with one or more suggestions on how to address the wearer's current mental state. For instance, when the wearer is in a state of emotional stress, the system may suggest that the user meditate, remove themselves from their current physical or social situation, execute one or more exercises, etc.
- In various embodiments, while the system is using one or more sensors (e.g., eyewear based sensors) to assess the mental state of the wearer, the system may also (e.g., at least substantially simultaneously) capture one or more images of a person or other object located in close proximity to the wearer (e.g., using a camera, such as a forward-facing camera associated with eyewear worn by the wearer). In various embodiments, the system may capture and analyze the image to determine whether the person or other object in the image caused the wearer's mental state. In other embodiments, the system may simply capture the image to determine what the user was viewing when they were under a state of emotional stress. If the system determines that the person or other object in the image caused the wearer's mental state, the system may notify the wearer (or other individual) of the association and may also provide the wearer suggested actions to address the wearer's mental state.
- In various embodiments, the system may also capture one or more images to determine reoccurring mental states (e.g., mental state patterns) in the wearer's life. For instance, if on a first occasion when the wearer is speaking to a particular person, the wearer becomes emotionally stressed, the system may capture an image of the particular person. If later, on a second occasion when the wearer is speaking to the particular person, the wearer again becomes emotionally stressed, the system may again capture an image of the particular person. The system may compare the first image with the second image. If the system determines that the first and second images are of the same particular person, the system may notify the wearer that the wearer's mental state when speaking to the particular person is “stressed”. The system may then provide suggestions to the wearer on how to address the wearer's mental state when in close proximity to the particular person.
- The system may use one or more similar techniques to determine that a user is typically in a particular mental state (e.g., relaxed, happy, emotionally stressed) when engaged in a particular activity (e.g., sailing, yoga, golf, commuting, a particular work activity). In various embodiments, the system may be adapted to automatically determine that the wearer is engaged in a particular work activity based on information from one or more of the sensors that the wearer is currently wearing (e.g., eyewear based sensors). For example, the system may use information from a forward facing camera installed in a pair of eyewear that the individual is wearing to determine that the individual is on a golf course (e.g., by detecting the presence of a hole, a pin, a fairway etc. . . . in one or more images taken by the forward-facing camera) or that the individual is driving (e.g., by detecting the presence of roadway “stripes” in an image from the forward-facing camera).
- As will be appreciated by one skilled in the relevant field, the present systems and methods may be, for example, embodied as a computer system, a method, or a computer program product. Accordingly, various embodiments may be entirely hardware or a combination of hardware and software. Furthermore, particular embodiments may take the form of a computer program product stored on a computer-readable storage medium having computer-readable instructions (e.g., software) embodied in the storage medium. Various embodiments may also take the form of Internet-implemented computer software. Any suitable computer-readable storage medium may be utilized including, for example, hard disks, compact disks, DVDs, optical storage devices, and/or magnetic storage devices.
- Various embodiments are described below with reference to block diagram and flowchart illustrations of methods, apparatuses, (e.g., systems), and computer program products. It should be understood that each block of the block diagrams and flowchart illustrations, and combinations of blocks in the block diagrams and flowchart illustrations, respectively, can be implemented by a computer executing computer program instructions. These computer program instructions may be loaded onto a general purpose computer, a special purpose computer, or other programmable data processing apparatus that can direct a computer or other programmable data processing apparatus to function in a particular manner such that the instructions stored in the computer-readable memory produce an article of manufacture that is configured for implementing the functions specified in the flowchart block or blocks.
- The computer instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on a user's computer and partly on a remote computer, or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including but not limited to: a local area network (LAN); a wide area network (WAN); a cellular network; or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
- These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner such that the instructions stored in the computer-readable memory produce an article of manufacture that is configured for implementing the function specified in the flowchart block or blocks. The computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer-implemented process (e.g., method) such that the instructions that execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart block or blocks.
-
FIG. 1 is a block diagram of a MentalState Monitoring System 100 according to particular embodiments. As may be understood from this figure, the MentalState Monitoring System 100 includes One orMore Networks 115, One or MoreThird Party Servers 50, a MentalState Monitoring Server 120 that may, for example, be adapted to execute a MentalState Monitoring Module 300, aDatabase 140, One or More Remote Computing Devices 154 (e.g., such as a smart phone, a tablet computer, a wearable computing device, a laptop computer, a desktop computer, etc.), and One or More WearableHealth Monitoring Devices 156, which may, for example, be embodied as one or more of eyewear, headwear, clothing, a watch, a hat, a helmet, a cast, an adhesive bandage, a piece of jewelry (e.g., a ring, earring, necklace, bracelet, etc.), or any other suitable wearable device. In particular embodiments, the one ormore computer networks 115 facilitate communication between the One or MoreThird Party Servers 50, the MentalState Monitoring Server 120,Database 140, One or MoreRemote Computing Devices 154, and the one or moreHealth Monitoring Devices 156. - The one or
more networks 115 may include any of a variety of types of wired or wireless computer networks such as the Internet, a private intranet, a mesh network, a public switch telephone network (PSTN), or any other type of network (e.g., a network that uses Bluetooth or near field communications to facilitate communication between computing devices). The communication link between the One or MoreRemote Computing Devices 154 and the MentalState Monitoring Server 120 may be, for example, implemented via a Local Area Network (LAN) or via the Internet. -
FIG. 2 illustrates a diagrammatic representation of the architecture for the MentalState Monitoring Server 120 that may be used within the MentalState Monitoring System 100. It should be understood that the computer architecture shown inFIG. 2 may also represent the computer architecture for any one of the One or MoreRemote Computing Devices 154, one or moreThird Party Servers 50, and one or moreHealth Monitoring Devices 156 shown inFIG. 1 . In particular embodiments, the MentalState Monitoring Server 120 may be suitable for use as a computer within the context of the MentalState Monitoring System 100 that is configured for determining a mental state of a wearer by detecting characteristics of the wearer using signals received from sensors coupled to the eyewear. - In particular embodiments, the Mental
State Monitoring Server 120 may be connected (e.g., networked) to other computing devices in a LAN, an intranet, an extranet, and/or the Internet as shown inFIG. 1 . As noted above, the MentalState Monitoring Server 120 may operate in the capacity of a server or a client computing device in a client-server network environment, or as a peer computing device in a peer-to-peer (or distributed) network environment. The MentalState Monitoring Server 120 may be a desktop personal computing device (PC), a tablet PC, a set-top box (STB), a Personal Digital Assistant (PDA), a cellular telephone, a web appliance, a network router, a switch or bridge, or any other computing device capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that computing device. Further, while only a single computing device is illustrated, the term “computing device” shall also be interpreted to include any collection of computing devices that individually or jointly execute a set (or multiple sets) of instructions to perform any one or more of the methodologies discussed herein. - An exemplary Mental
State Monitoring Server 120 includes aprocessing device 202, a main memory 204 (e.g., read-only memory (ROM), flash memory, dynamic random access memory (DRAM) such as synchronous DRAM (SDRAM) or Rambus DRAM (RDRAM), etc.), a static memory 206 (e.g., flash memory, static random access memory (SRAM), etc.), and adata storage device 218, which communicate with each other via abus 232. - The
processing device 202 represents one or more general-purpose or specific processing devices such as a microprocessor, a central processing unit (CPU), or the like. More particularly, theprocessing device 202 may be a complex instruction set computing (CISC) microprocessor, reduced instruction set computing (RISC) microprocessor, very long instruction word (VLIW) microprocessor, or processor implementing other instruction sets, or processors implementing a combination of instruction sets. Theprocessing device 202 may also be one or more special-purpose processing devices such as an application specific integrated circuit (ASIC), a field programmable gate array (FPGA), a digital signal processor (DSP), network processor, or the like. Theprocessing device 202 may be configured to executeprocessing logic 226 for performing various operations and steps discussed herein. - The Mental
State Monitoring Server 120 may further include anetwork interface device 208. The MentalState Monitoring Server 120 may also include a video display unit 210 (e.g., a liquid crystal display (LCD) or a cathode ray tube (CRT)), an alpha-numeric input device 212 (e.g., a keyboard), a cursor control device 214 (e.g., a mouse), and a signal generation device 216 (e.g., a speaker). - The
data storage device 218 may include a non-transitory computing device-accessible storage medium 230 (also known as a non-transitory computing device-readable storage medium, a non-transitory computing device-readable medium, or a non-transitory computer-readable medium) on which is stored one or more sets of instructions (e.g., the Mental State Monitoring Module 300) embodying any one or more of the methodologies or functions described herein. The one or more sets of instructions may also reside, completely or at least partially, within themain memory 204 and/or within theprocessing device 202 during execution thereof by the MentalState Monitoring Server 120—themain memory 204 and theprocessing device 202 also constituting computing device-accessible storage media. The one or more sets of instructions may further be transmitted or received over anetwork 115 via anetwork interface device 208. - While the computing device-
accessible storage medium 230 is shown in an exemplary embodiment to be a single medium, the term “computing device-accessible storage medium” should be understood to include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions. The term “computing device-accessible storage medium” should also be understood to include any medium that is capable of storing, encoding, or carrying a set of instructions for execution by the computing device and that causes the computing device to include any one or more of the methodologies of the present invention. The term “computing device-accessible storage medium” should accordingly be understood to include, but not be limited to, solid-state memories, optical and magnetic media, etc. - As noted above, a system, according to various embodiments, is adapted to assess the mental state of a wearer of a wearable device. Various aspects of the system's functionality may be executed by certain system modules, including the Mental
State Monitoring Module 300. The MentalState Monitoring Module 300 is discussed in greater detail below. - Mental State Monitoring Module
-
FIG. 3A is a flow chart of operations performed by an exemplary MentalState Monitoring Module 300, which may, for example, run on the MentalState Monitoring Server 120, or any suitable computing device (such as the One or MoreHealth Monitoring Devices 156 or a suitable mobile computing device). In particular embodiments, the MentalState Monitoring Module 300 may assess a wearer's mental state and make suggestions to the wearer to address a particular mental state associated with the wearer. - The system begins, in various embodiments, at
Step 305 by providing eyewear comprising at least one or more sensors coupled to the eyewear. In various embodiments, the system may do this by, for example: (1) facilitating delivery of the eyewear to an address associated with a particular individual; (2) facilitating distribution of the eyewear from a healthcare worker to the individual; and (3) placing an order of the eyewear from a third party for delivery to the individual. In other embodiments, this step may be executed manually (e.g., by a human being) rather than a computer. - In various embodiments, the one or more sensors that are coupled to the eyewear (or other health monitoring device) are adapted to detect one or more characteristics of a wearer of the eyewear, wherein the one or more characteristics of the wearer are associated with the wearer's mental state. In various embodiments, the sensors coupled to the eyewear or other health monitoring device may include, for example, one or more of the following: a heart rate monitor, an electrocardiogram (EKG), an electroencephalogram (EEG), a pedometer, a thermometer, a front-facing camera, an eye-facing camera, a microphone, an accelerometer, a gyroscope, a magnetometer, a blood pressure sensor, a pulse oximeter, a respiration rate sensor, a blood alcohol concentration (BAC) sensor, a skin conductance response sensor, a near-field communication sensor, or any other suitable sensor. In particular embodiments, the sensors coupled to the eyewear comprise an eye-facing camera, a front-facing camera, and a heart rate monitor
- In various embodiments, the one or more sensors are coupled to a computing device that is associated with (e.g., embedded within, attached to) the eyewear or other health monitoring device. In particular embodiments, the eyewear or other health monitoring device comprises at least one processor, computer memory, suitable wireless communications components (e.g., a Bluetooth chip) and a power supply for powering the health monitoring device and/or the various sensors.
- In particular embodiments, the sensors may be physically coupled to the eyewear in any suitable way. For example, in various embodiments, the sensors may be embedded into the eyewear. In some embodiments, the sensors may be positioned along the brow bar of the eyewear. In other embodiments, the sensors may be positioned along the one or more of the temples of the eyewear. In still other embodiments, the sensors may be coupled to one or more of the lenses of the eyewear. As noted above, the one or more sensors may be coupled to a Bluetooth device that is configured to transmit the one or more signals to a handheld wireless device, and the step of receiving one or more signals from the one or more sensors (discussed below in reference to Step 310) further comprises receiving the one or more signals from the wireless handheld device (e.g., via the Internet). In particular embodiments, one or more of the sensors may be detachable from the eyewear. For instance, if a wearer does not need a temperature sensor or other particular sensor, the sensor may be removed from the eyewear.
- The system continues, at
Step 310 by receiving one or more signals from the one or more sensors, wherein each of the one or more signals relates to at least one characteristic associated with the wearer. In particular embodiments, the one or more signals that relate to the at least one characteristic associated with the wearer may include one or more signals that may be used to derive: (1) the wearer's a heart rate, (2) the wearer's heart rhythm; (3) a distance traveled by the wearer; (4) the wearer's body temperature; (5) one or more images associated with the wearer or the environment; (6) one or more sounds associated with the wearer's body or environment; (7) a speed traveled by the wearer; (8) the wearer's blood pressure; (9) the wearer's oxygen saturation level; (10) the wearer's brainwave activity (e.g., the location of the brainwave relative to the wearer's brain, the frequency of the brainwave, and the type of brainwave (e.g., gamma waves, beta waves, alpha waves, theta waves, and delta waves)); (11) the wearer's pupil size; (12) the wearer's perspiration level; (13) the wearer's hydration level; (14) the wearer's respiration rate; (15) the number and/or cadence of steps taken by the wearer; (16) the movement of one or more of the wearer's facial muscles; (17) one or more biochemical changes within the wearer's body; (18) changes in the one or more characteristics of the wearer's skin (e.g., skin paleness or clamminess) and/or (19) any other suitable attribute of the wearer or the wearer's environment. For instance, the system may receive a signal from an eye-facing camera associated with the eyewear that the wearer's brow is furrowed at the same time that the system receives a signal from the heart rate sensor that the wearer's heart rate is above a predetermined target heart rate. In various embodiments, the system may store data related to the signals and/or data derived from this data for later review and use in determining the mental state of the wearer. - In particular embodiments, the system may receive one or more of the above-referenced signals substantially automatically. In various embodiments, the system may receive one or more of the signals on a substantially periodic basis (e.g., by the second, by the minute, hourly, daily, etc.). For example, the system may receive one or more signals every thirty seconds throughout the day. In other embodiments, the system may receive one or more signals at least partially in response to receiving an indication from the wearer that the system should receive a signal. For instance, the wearer may speak a voice command to the wearable device requesting that the device take the wearer's blood pressure. In various embodiments, the system may receive an indication from the wearer of when to have the system receive the signal. For example, the system may receive an indication from the wearer to have the system conduct a brain scan of the user (e.g., receive at least one brainwave signal from the EEG) at 8:00 a.m. and at 2:00 p.m. on a particular day. In particular embodiments, the system may receive a request from the wearer to have a particular signal received from a particular sensor at the same time that the system receives a second particular signal from a second particular sensor. For example, when the system receives a signal that indicates that the user's respiration rate has increased, the system may, at least partially in response to receiving the increased respiration rate signal, also obtain an image of the wearer's eye from an eye-facing camera associated with the eyewear.
- In some embodiments, the system receives a signal of an image captured by the eyewear. In various embodiments, the system receives a plurality of images captured by the eyewear. In particular embodiments, the system receives the image from the front-facing camera. In some embodiments, the system receives the image substantially automatically from the front-facing camera. In other embodiments, the system may receive the image in response to receiving an indication from the wearer to capture the image. For example, the system may receive a voice command from the wearer to capture the image. In various embodiments, the system may store the captured image in local or remote memory. In some embodiments, the image captured by the eyewear may be a video.
- In various embodiments, the system may receive only one signal from a single sensor associated with the eyewear. In other embodiments, the system may receive a signal from a plurality of the sensors associated with the eyewear. In yet other embodiments, the system may receive multiple signals from one or more of the sensors. In various embodiments, the system may be configured to receive a first signal from a first sensor at the same time that it receives a second signal from a second sensor. For example, the system may be configured to receive an image signal from a front-facing camera associated with the eyewear at the same time that the system receives a heart rate signal from a heart rate sensor associated with the eyewear. As a further example, the system may be configured to simultaneously receive a signal from both an eye-facing camera and an EEG associated with the eyewear.
- Next, at
Step 315, the system analyzes the one or more received signals to determine at least one characteristic associated with the wearer. For example, the system may analyze the one or more received signals to determine that the at least one characteristic associated with the wearer is an increase or decrease in pupil size. In particular embodiments, the system may analyze the one or more received signals to determine that the at least one characteristic associated with the wearer is an increase or decrease in heart rate. In some embodiments, the system may analyze the one or more received signals to determine that the at least one characteristic associated with the wearer is an increase or decrease in the wearer's perspiration rate. In other embodiments, the system may analyze the one or more received signals to determine that the at least one characteristic associated with the wearer is an increase or decrease in respiration rate. In yet other embodiments, the system may analyze the one or more received signals to determine that the at least one characteristic associated with the wearer is an increase or decrease in movement. In still other embodiments, the system may analyze the one or more received signals to determine that the at least one characteristic associated with the wearer is an increase or decrease in brainwave activity and/or frequency. In various embodiments, the system may analyze the one or more received signals to determine that the at least one characteristic associated with the wearer is a change in location of the wearer's brainwave activity. In yet other embodiments, the system may analyze the one or more received signals to determine that the at least one characteristic associated with the wearer is a change in wearer's the type of brainwave (e.g., a change from gamma waves to delta waves). In some of these embodiments, the system may analyze the one or more received signals to determine that the at least one characteristic associated with the wearer is an increase in heart rate in conjunction with an increase in pupil size. It should be understood from this disclosure that the system may analyze the one or more signals to determine that the at least one characteristic associated with the wearer is a combination characteristics that allow the system to determine the mental state of the wearer. - In various embodiments, the system may store the results of the analysis of the one or more received signals for later comparison with past and future analyses of the one or more received signals. In particular embodiments, the system may analyze a particular received signal at a particular time of day (e.g., morning, noon, night, etc.). In some embodiments, the system may analyze a particular received signal at the same time that the system analyzes a second particular received signal. For instance, the system may analyze the front-facing camera signal when it analyzes the heart rate signal. In particular embodiments, the system may chart the analysis of the one or more received signals in a visual diagram. For example, the system may chart the changes in the wearer's heart rate in a diagram displayed e.g., on the lens of the eyewear or on a separate display screen associated with the eyewear.
- At
Step 320, the system facilitates determination of a mental state of the wearer based on the at least one characteristic. In various embodiments, the system may facilitate determination of the mental state of the wearer substantially automatically. In particular embodiments, the system may facilitate determination of the mental state of the wearer in response to receiving manual input of indication request from the wearer to determine the wearer's mental state. For example, the system may receive a voice command from the wearer requesting the system determine the mental state of the wearer at that time. In still other embodiments, the system may facilitate determination of the mental state of the wearer after comparing various signals to predetermined thresholds to establish that the signal is indicative of at least one characteristic. In various embodiments, the system may facilitate determination of the mental state of the wearer at random. For instance, the system may determine the mental state of the wearer at different, randomly selected times throughout the day. In some embodiments, the system may facilitate determination of the mental state of the wearer periodically throughout the day at specified times. For instance, the system may determine the mental state of the wearer at 8:00 a.m., 10:00 a.m., 12:00 p.m., 2:00 p.m., and 4:00 p.m. on a given day. In other embodiments, the system may determine the mental state of the wearer at predetermined intervals of time. For instance, the system may determine the mental state of the wearer every 30 minutes or every two hours. - In other embodiments, the determination of the mental state of the wearer may be that the wearer's mental state is in a particular emotional state (e.g., happy, sad, anxious, calm, scared, angry, surprised, ashamed, envious, curious, relaxed, emotionally stressed, confused, moody, etc.). For instance, the determination of the mental state of the wearer may be that the wearer is under emotional stress (i.e., the wearer's mental state is “emotionally stressed”). In some embodiments, the determination of the mental state of the wearer may include the wearer's experience of the wearer's current physical state (e.g., the wearer's experience of bodily pain, nausea, rapid heartbeat, etc.). For example, the determination of the wearer's mental state may include determining that the wearer is responding mentally and/or physically to sustaining a concussion. In other embodiments, the determination of the wearer's mental state may include a determination of the wearer's cognitive state (e.g., inability to concentrate, poor judgment, racing thoughts, constant worrying, etc.). For example, the determination of the wearer's mental state may be that the wearer is having memory problems. In yet other embodiments, the determination of the mental state of the wearer may include a mental state associated with a particular behavior (e.g., eating more or less, sleeping too much or too little, isolation from others, procrastinating, neglecting responsibilities, using alcohol or drugs, nervous habits or twitches, etc.). For instance, the determination of the mental state may be that the wearer is overly sleepy (which the system may determine, for example, by determining that the wearer is sleeping more than a predetermined amount of time per day).
- In various embodiments, the system may determine the mental state of the wearer using a scaled rating. For example, where the system is determining a mental state such as stress of the wearer, the system may determine that the wearer's mental state (e.g., stress level) is low, moderate, or high. In some embodiments, the system may determine that the mental state of the wearer includes multiple mental states. For instance, the system may determine that the wearer is both emotionally stressed and anxious.
- After determining the wearer's mental state, at
Step 325 the system may optionally associate the mental state of the wearer with at least one object (e.g., an animate object, such as a particular person or animal, or an inanimate object, such as a scary clown doll), at least one activity (e.g., bowling, playing tennis, driving, working on a computer, or running), at least one external condition (e.g., the user's current workload at work, the user's current credit rating, the user's marital or dating status), and/or at least one internal condition (the user's current weight, health, etc. . . . ). In yet other embodiments, the system may associate the mental state of the wearer with both an animate object and an inanimate object (e.g., a person and an object—for example, the system may determine that the wearer is typically emotionally stressed when they see both the wearer's spouse and a checkbook), or any other combination of factors described herein. In particular embodiments, the system may associate the mental state of the wearer with both a particular object and an activity (for example, the system may determine that the wearer is typically relaxed when the wearer's spouse is present and the wearer is watching TV). - In various embodiments, the system may associate the mental state of the wearer with the user's external or internal context. In some embodiments, the external or internal context may include one or more of the following for a particular time period: (1) the user's current health; (2) the user's general state of mind; (3) the user's current spending behavior; (4) the current weather conditions at the user's current location (e.g., pollen count, UV index, air quality, precipitation, wind speed and direction, barometric pressure, humidity, outdoor temperature, season, cloud levels); (5) one or more social media entries made by or received by the individuals at the particular time; (6) the user's credit rating at the particular time; (7) the user's employment status at the particular time; (8) the user's housing information for the particular time; (9) information from one or more e-mails at or around the particular time; (10) the user's exercise activity during the particular time; (11) indoor temperature within the wearer's home or workplace; and/or (12) any other external information that may be relevant to and/or have an impact on the user's mental state. In order to determine this information, the system may receive information from and/or regarding, for example: (1) the wearer's health records; (2) the wearer's genetics; (3) the wearer's family history; (4) one or more of the wearer's physical attributes; (5) the wearer's workout schedule; (6) one or more social media accounts associated with the wearer; (7) one or more of the wearer's social behaviors; (8) the wearer's arrest history; (9) the weather at a user's current location (e.g., as determined from a suitable weather service); (10) the user's calendar; (11) the user's spending behaviors; (12) the wearer's credit history; (13) the wearer's employment status and/or history; (14) the wearer's marital status; (15) the wearer's current residence; (16) an email account of the wearer; (17) the wearer's travel history; (18) the wearer's aspirations; (19) the wearer's goals; (20) dietary information for the wearer, etc.
- For example, the system may associate the wearer's mental state (e.g., “emotionally stressed”, “unhappy”, etc. . . . ) with the activity of a paying bills by: (1) identifying an entry on the wearer's electronic calendar that indicates that the wearer will be paying bills in a particular time slot on a particular day; (2) confirming that the user is paying bills in the particular time slot by identifying a checkbook in an image taken by the system's front facing camera during the particular time slot; and (3) determining the wearer's mental state during the time slot using any suitable technique, such as those described herein.
- Similar techniques may be used to determine longer-lasting impacts of certain external or internal conditions on a wearer's mental state. For example, the system may use any suitable technique to determine the wearer's general mental state (e.g., relatively happy, relatively sad, under relative stress) over a predetermined number of hours, days, weeks, or months, by comparing the wearer's aggregate (e.g., average) mental state for that time period with the wearer's aggregate mental state for one or more time periods (e.g., of similar length), or by comparing the wearer's aggregate mental state for the time period with typical mental states of the wearer or one or more other individuals. The system may then access any suitable information (e.g., any of the types of object, activity, internal context or external context information described herein, or other information) to determine what may be contributing to the mental state (e.g., using any suitable data analysis techniques). For example, the system may determine that the wearer is typically relatively happy in weeks in which the wearer is exercising more than five times per week and has a balance of less than $1,000 on their personal credit card, and experiences relatively high stress levels in weeks in which the user is exercising less than two times per week.
- In various embodiments, the system may associate the mental state of the wearer with a person, object, activity, internal context, or external context using one or more signals received from the eyewear's sensors and/or any other suitable data received by the system. In various embodiments, the system may associate the mental state of the wearer with a person, object, or activity by receiving a manual input from the wearer. For instance, the system may receive indication request from the wearer to associate a particular person with the wearer's current mental state. In these embodiments, the system may then capture an image of the person being viewed by the wearer, identify the person in the image and associate the current mental state of the wearer with the identified person. In some such embodiments, the system may also monitor the wearer's mental state when the wearer next encounters the identified person and track the wearer's mental state over time each time the wearer encounters the identified person. In this way, the system can alert the wearer if the identified person continually causes the wearer to experience the same or similar mental state each time the wearer encounters the identified person. In other embodiments, the system may associate the mental state of the wearer with (e.g., an animate or inanimate object) or activity substantially automatically after (e.g., in response to) determining the mental state of the wearer.
- In various embodiments, the system may be configured to associate the mental state of the wearer with an animate or inanimate object by examining a received image of a first object (e.g., a first person) located in the received image. In particular embodiments, the system determines the object in the received image by identifying the object located in the image using any suitable image recognition techniques. In other embodiments, the system determines the object in the received image by comparing the image of the object with one or more stored images. At least partially in response to determining that the image of the object at least substantially matches a stored image, the system may identify the object as being the same as the known object (e.g., person or thing) in the stored image.
- In some embodiments, the system may associate the mental state of the wearer with an object based on the proximity of the object to the wearer. For example, the system may use a front-facing camera to determine that a particular person is in close proximity to the wearer (e.g., in front of the wearer) when the wearer is experiencing a particular mental state such as stress. In some embodiments, the proximity of a person to the wearer may be obtained by an electronic device on the person (e.g., a cellphone, an RFID tag, etc.).
- In various embodiments, the system may associate the mental state of the wearer with an activity based on the wearer performing the activity. For example, the system may use the system's processor and front facing camera to determine, from images taken by the front-facing camera, that the user is climbing stairs. The system may then assess the wearer's mental state (e.g., in any suitable way described herein) and associate that that mental state with stair climbing. For example, the system may determine that the user is typically mentally relaxed when climbing stairs.
- Continuing to Step 330, the system stores, in memory, the association between the wearer's mental state and the object, activity, internal context, and/or external context causing the wearer's mental state. In various embodiments, the system may store the association between the wearer's mental state and the object, activity, external or internal context substantially automatically after (e.g., in response to) making the association between the wearer's mental state and the object, activity, internal context and/or external context. In particular embodiments, the system may store the association after (e.g., in response to) receiving manual input from the wearer requesting that the system store the association. In various embodiments, the system may store the association for a specified period of time. For example, the system may store the association for a day, a month, a year, etc. in the system's
Database 140. In some embodiments, the system may store the association on any suitable server, database, or device. In particular embodiments, the system may store the association on the MentalState Monitoring Server 120. - Next, at
Step 335, the system notifies the wearer or other suitable individual of the association of the wearer's mental state and the object, activity, internal context, and/or external context. For example, in some embodiments, the system may notify the wearer and the wearer's physician of the association of the wearer's mental state with the object in the received image and/or the activity that the wearer was engaged in while experiencing the mental state. In various embodiments, the system notifies the wearer of the association by displaying an image on the lens of the eyewear, or a display screen associated with the eyewear. In other embodiments, the system notifies the wearer of the association by communicating the association through an audio speaker to the wearer. In some embodiments, the system notifies the wearer of the association by sending a notification to the wearer's mobile device. In particular embodiments, the system notifies the wearer of the association via an electronic communication such as an email or text message. In other embodiments, the system may notify the wearer of a single association substantially immediately after (e.g., in response to) the system associates the wearer's mental state with a particular object. In yet other embodiments, the system may notify the wearer of all associations made on a particular day (or within another particular time period). - In other embodiments, the system may notify the wearer of the association after (e.g., in response to) detecting a particular event. For example, the system may notify the wearer of the association after the system no longer detects the presence or proximity of a particular object. In some embodiments, the system may notify the wearer of the association after a particular period of time. For instance, the system may notify the wearer of an association one hour after the system associates the wearer's mental state with a particular object or activity. In still other embodiments, the system may notify the wearer of the association at a particular time of day. As an example, the system may notify the wearer of an association between the wearer's mental state and an object, activity, internal context, and/or external context at the end of the day.
- Continuing to Step 340, the system provides the wearer with one or more suggested actions to address the wearer's current mental state. In various embodiments, the system may provide suggested actions to the wearer in any suitable way. In various embodiments, the system may provide one or more suggested actions to the wearer by displaying an image on the lens of the eyewear, or on a display screen associated with the eyewear. In other embodiments, the system may provide one or more suggested actions to the wearer by communicating through an audio speaker to the wearer. In some embodiments, the system provides one or more suggested actions to the wearer by sending a notification to the wearer's mobile device. In particular embodiments, the system may provide one or more suggested actions to the wearer via an electronic communication, such as an email or text message. In still other embodiments, the system may provide a single suggested action to the wearer. In yet other embodiments, the system may provide multiple suggested actions to the wearer.
- In various embodiments, the system may provide suggested actions to the wearer to address the wearer's current mental state substantially immediately after the system notifies the wearer of the association between the wearer's mental state and the particular object, activity, internal context and/or external context. In other embodiments, the system may provide one or more suggested actions to the wearer after (e.g., in response to) detecting a particular event. For example, the system may provide suggested actions to the wearer after the system no longer detects the presence or proximity of a particular object. In some embodiments, the system may provide one or more suggested actions to the wearer after a particular period of time. For instance, the system may provide suggested actions to the wearer one hour after the system associates the wearer's mental state with the particular object. In still other embodiments, the system may provide one or more suggested actions to the wearer at a particular time of day. As an example, the system may provide one or more suggested actions to the wearer at the end of the day.
- In various embodiments, the one or more suggested actions to address the wearer's current mental state may be one or more techniques for reducing stress. In particular embodiments, the suggested actions to address the wearer's current mental state may include, for example: (1) taking medication; (2) praying; (3) engaging in one or more yoga poses; (4) taking deep breaths; (5) avoiding a particular object (e.g., a particular person or thing) based on the association between the wearer's mental state and the object. In other embodiments, the suggested actions to address the wearer's current mental state may include suggesting that the wearer visit links to websites containing information on the particular mental state of the wearer. In some embodiments, the suggested actions to address the wearer's current mental state may include a listing of suggested applications on the wearer's mobile device. For instance, where the wearer's current mental state is in a stressed mental state, the system may suggest a yoga application on the wearer's mobile device to assist in improving the wearer's mental state. As a further example, if the system determines that the wearer typically experiences high levels of stress when the wearer is carrying high amounts of credit card debt, the system may suggest that the user reduce their debt levels.
- In various embodiments, the system, when executing the Mental
State Monitoring Module 300, may omit particular steps, perform particular steps in an order other than the order presented above, or perform additional steps not discussed directly above. In should also be understood that various steps executed “in response to” a particular event occurring, or a particular condition being satisfied, may also be executed “at least partially in response to” the particular event occurring or the particular condition being satisfied. - As shown in
FIG. 4 ,eyewear 400, according to various embodiments, includes: (1) aneyewear frame 410; (2) afirst temple 412; and (3) asecond temple 414. These various components are discussed in more detail below. In particular embodiments, theeyewear 400 is computerized and may serve as the wearablehealth monitoring device 156 ofFIG. 1 . - Eyewear Frame
- Referring still to
FIG. 4 ,eyewear 400, in various embodiments, includes anysuitable eyewear frame 410 configured to support one ormore lenses eyewear frame 410 has afirst end 402 and asecond end 404. Theeyewear frame 410 may be made of any suitable material such as metal, ceramic, polymers or any combination thereof. In particular embodiments, theeyewear frame 410 is configured to support the first andsecond lenses second lenses eyewear frame 410 may be configured to support the first andsecond lenses eyewear frame 410 is configured to support a number of lenses other than two lenses (e.g., a single lens, a plurality of lenses, etc.). In particular embodiments, thelenses - The
eyewear frame 410 includes a first and second nose pad 422 (not shown in figure), 424, which may be configured to maintain theeyewear 400 adjacent the front of a wearer's face such that thelenses eyewear 400. In particular embodiments, thenose pads 422, 424 may comprise a material that is configured to be comfortable when worn by the wearer (e.g., rubber, etc.). In other embodiments, the nose pads may include any other suitable material (e.g., plastic, metal, etc.). In still other embodiments, the nose pads may be integrally formed with theframe 410. - The
eyewear frame 410 includes a first andsecond hinge second temples first hinge 426 may be welded to, or integrally formed with, theframe 410 and thefirst temple 412 and thesecond hinge 428 may be welded to, or integrally formed with, theframe 410 and thesecond temple 414. - First and Second Temples
- As shown in
FIG. 4 , thefirst temple 412, according to various embodiments, is rotatably connected to theframe 410 at a right angle to extend thefirst temple 412 substantially perpendicular, substantially parallel, or anywhere in between the right angle to theframe 410. Thefirst temple 412 has a first andsecond end second end 412 b, thefirst temple 412 includes anearpiece 413 configured to be supported by a wearer's ear. Similarly, thesecond temple 414, according to various embodiments, is rotatably connected to theframe 410 at a right angle to extend thesecond temple 414 substantially perpendicular, substantially parallel, or anywhere in between the right angle to theframe 410. Thesecond temple 414 has a first andsecond end second end 414 b, thesecond temple 414 includes anearpiece 415 configured to be supported by a wearer's ear. - Sensors
- In various embodiments, the
second temple 414 has one ormore sensors 430 connected to thesecond temple 414. In various embodiments, the one ormore sensors 430 may be coupled to theframe 410, the first andsecond temples second lenses eyewear 400 in any suitable way. For instance, the one ormore sensors 430 may be embedded into theeyewear 400, coupled to theeyewear 400, and/or operatively coupled to theeyewear 400. In various embodiments, the one or more sensors may be formed at any point along theeyewear 400. For instance, a fingerprint reader may be disposed adjacent the first temple of theeyewear 400. In various embodiments, the one or more sensors may be formed in any shape. In addition, the one or more sensors may be formed on the inner (back) surface of theframe 410, the first andsecond temples second lenses eyewear 400. In other embodiments, the one or more sensors may be formed on the outer (front) surface of theframe 410, the first andsecond temples second lenses eyewear 400. - In various embodiments, the one or
more sensors 430 that are coupled to the eyewear (or other wearable device) are adapted to detect one or more characteristics of the eyewear or a wearer of the eyewear, wherein the one or more characteristics of the wearer are associated with the wearer's identity. In various embodiments, the one or more sensors coupled to the eyewear or other health monitoring device may include, for example, one or more of the following: a near-field communication sensor, a Bluetooth chip, a GPS unit, an RFID tag (passive or active), a fingerprint reader, an iris reader, a retinal scanner, a voice recognition sensor, a heart rate monitor, an electrocardiogram (EKG), an electroencephalogram (EEG), a pedometer, a thermometer, a front-facing camera, an eye-facing camera, a microphone, an accelerometer, a magnetometer, a blood pressure sensor, a pulse oximeter, a skin conductance response sensor, any suitable biometric reader, or any other suitable sensor. In some embodiments, the one or more sensors may include a unique shape, a unique code, or a unique design physically inscribed into the eyewear that may be readable by an individual or a remote computing device. In particular embodiments, the sensors coupled to the eyewear may include one or more electronic communications devices such as a near field communication sensor, a Bluetooth chip, an active RFID, and a GPS unit. - In various embodiments, the one or more sensors are coupled to a computing device that is associated with (e.g., embedded within, attached to) the eyewear or other wearable device. In particular embodiments, the eyewear or other wearable device comprises at least one processor, computer memory, suitable wireless communications components (e.g., a Bluetooth chip) and a power supply for powering the wearable device and/or the various sensors.
- As noted above, the one or more sensors may be coupled to a Bluetooth device that is configured to transmit the one or more signals to a handheld wireless device, and the step of using the eyewear to confirm the identity of the wearer of the eyewear (discussed above in reference to Step 310) further comprises receiving the one or more signals from the wireless handheld device (e.g., via the Internet). In particular embodiments, one or more of the sensors may be detachable from the eyewear. For instance, if a wearer does not need a temperature sensor or other particular sensor, the sensor may be removed from the eyewear.
- Sense and Track Focused States
- In a particular example of a wearer using the Mental
State Monitoring Module 300 to monitor their mental state, the wearer may put on the wearable device in the morning and continue to wear the device throughout the day. During this time, the system tracks the brainwave activity, including the types of brainwaves, of the wearer using the system's EEG. The system particularly tracks the gamma wave activity of the wearer, which may at the optimal range, for example, signify a high focused mental state of the wearer. The system may store the brainwave activity and chart the highs and lows of the wearer's brainwave activity throughout the day in order to determine the times throughout the day when the wearer is most focused. For example, where the wearer has relatively high gamma wave activity from 9:00 a.m. until 11:00 a.m., and relatively high theta wave activity from 2:00 p.m. until 3:00 p.m., the system will notify the wearer that the wearer's highest cognitive functioning occurs from 9:00-11:00 a.m. and the wearer's lowest cognitive functioning occurs from 2:00-3:00 p.m. The wearer may then use this information to alter the wearer's activities so that the wearer is focused on important matters from 9:00-11:00 a.m. and less important matters from 2:00-3:00 p.m. - Sense and Track Low-Productivity States
- Similar to the system tracking high focused states, the system, in a particular example, will also track low-productivity states. This period of low-productivity may be measured, for example, by tracking the wearer's brainwave activity, the wearer's movements, the distance traveled by the wearer, and/or the wearer's speed of travel. The system may also capture images of the wearer's surroundings for use, for example, in determining one or more activities that the wearer is engaged in. For instance, in a period in which the wearer has not moved and has had a particular type of brainwave such as theta waves for an extended period of time, the system may capture a plurality of images from the system's front-facing camera in order to determine that the wearer has been sitting in front of a television during those periods of time. The system may then track the amount of time that the wearer spent in the particular state and then notify the wearer and provide the wearer with one or more recommendations, such as to exercise rather than watch TV, or to exercise while watching TV.
- Identify Periods of High Stress
- In a further particular example of a wearer using the Mental
State Monitoring Module 300 of the One or More WearableHealth Monitoring Devices 156 to monitor the wearer's mental state, the system tracks the wearer's various mental states through a typical day via the wearable device (e.g., a pair of eyewear). The system tracks the wearer's mental state by monitoring the wearer's pupil size, heart rate, perspiration rate, respiration rate, movement, and brainwave activity. The system may also monitor potential causes of a certain change in the wearer's pupil size, heart rate, perspiration level, respiration rate, movement, and brainwave activity using a front-facing camera. - For example, if the wearer encounters a particular person that causes the wearer high stress as indicated by a rapid heartbeat in the wearer and an increase in the wearer's perspiration rate, the system captures an image of the person causing the high stress. The system may then present this image to the wearer with a suggestion to not interact in the future with the particular person. The system may also suggest other ways of interacting with the person in the future should avoidance be impractical. For instance, if the wearer works for the particular person that causes the wearer high stress, the system may suggest writing down what the wearer needs to speak to the person about prior to interacting with the person. The system may, for example, provide these suggestions to the wearer through a notification sent to the wearer's mobile device so that other people around the wearer are not made aware of the wearer's high stress.
- Monitor Behavior and Movements
- In another example, the system may monitor and track one or more specific behaviors and/or movements of the wearer in order to diagnose particular mental states. For example, the system may track the wearer's movements to determine whether a wearer frequently has one or more simple tics such as motor tics including, for example, eye blinking or other nervous eye movements, facial grimacing, shoulder shrugging, muscle twitches, head or shoulder jerking, or one or more complex tics such as facial grimacing combined with a head twist and a shoulder shrug. After monitoring the wearer's movements for a predetermined period of time (e.g., a day, a week, a month, etc.), the system may notify the wearer or the wearer's physician that the wearer has one or more behavioral characteristics that are consistent with a neurological condition such as Tourette syndrome.
- Many modifications and other embodiments of the invention will come to mind to one skilled in the art to which this invention pertains, having the benefit of the teaching presented in the foregoing descriptions and the associated drawings. Therefore, it is to be understood that the invention is not to be limited to the specific embodiments disclosed and that modifications and other embodiments are intended to be included within the scope of the appended claims. Although specific terms are employed herein, they are used in a generic and descriptive sense only and not for the purposes of limitation.
Claims (22)
1. A computer-implemented method of assessing the mental state of a wearer of eyewear comprising one or more sensors coupled to the eyewear, the one or more sensors being adapted to detect one or more characteristics of the wearer of the eyewear, wherein the one or more characteristics are associated with the wearer's mental state, the method comprising:
a. receiving, by a processor, one or more signals from the one or more sensors, wherein each of the one or more signals relates to at least one characteristic associated with the wearer, the at least one characteristic being selected from a group consisting of:
i. pupil size,
ii. heart rate,
iii. perspiration rate,
iv. respiration rate,
v. physical movement, and
vi. brainwave activity;
b. analyzing, by a processor, the one or more received signals to determine the at least one characteristic;
c. facilitating, by a processor, determination of a mental state of the wearer based, at least in part, on the at least one characteristic; and
d. associating, by a processor, the mental state of the wearer with at least one of one or more stimuli selected from a group consisting of: (1) an object; (2) an activity; (3) an internal context associated with the wearer; and (4) an external context associated with the wearer.
2. The computer-implemented method of claim 1 , further comprising the step of, providing a wireless transmitter coupled to the one or more sensors, the wireless transmitter being configured to transmit the one or more signals to a handheld computing device, wherein the step of receiving one or more signals from the one or more sensors further comprises, receiving the one or more signals from the handheld computing device via the Internet.
3. The computer-implemented method of claim 1 , wherein the one or more sensors comprises at least one sensor selected from a group consisting of:
a. an eye facing camera;
b. a forward facing camera;
c. an electrocardiogram sensor;
d. a heart rate monitor;
e. a microphone;
f. a skin conductance response sensor;
g. an accelerometer; and
h. a gyroscope.
4. The computer-implemented method of claim 3 , wherein the eyewear comprises an eye facing camera, a forward facing camera, and a heart rate monitor.
5. The computer-implemented method of claim 3 , further comprising the step of:
a. determining, by a processor, that the wearer of the eyewear is under emotional stress;
b. in response to determining that the wearer of the eyewear is under emotional stress, capturing, by a processor, an image by the forward facing camera;
c. receiving, by a processor, the image;
d. identifying an object from within the received image; and
e. facilitating notification of the wearer of the association of the wearer's emotionally stressed state and the object in the image.
6. The computer-implemented method of claim 5 , wherein the object is a particular individual.
7. The computer-implemented method of claim 6 , wherein the step of identifying the particular individual comprises using one or more face recognition techniques to determine that a human is present within the image.
8. The computer-implemented method of claim 6 , wherein the step of identifying the particular individual comprises using one or more face recognition techniques to determine the identity of the particular individual.
9. The computer-implemented method of claim 5 , wherein the step of determining that the wearer of the eyewear is under emotional stress further comprises:
a. receiving, by a processor, an image of the wearer's pupil from the eye facing camera;
b. determining, by a processor, from the received image the size of the wearer's pupil;
c. comparing, by a processor, the size of the wearer's pupil to a predefined measurement of the wearer's pupil; and
d. at least partially in response to determining that the wearer's pupil size exceeds the predefined measurement of the wearer's pupil determining, by a processor, that the wearer is in a state of emotional stress.
10. The computer-implemented method of claim 5 , wherein the step of determining that the wearer of the eyewear is under emotional stress further comprises:
a. determining, by a processor, the respiration rate of the wearer from the received one or more signals;
b. comparing the determined respiration rate to a predefined respiration rate level; and
c. in response to determining that the respiration rate exceeds the predefined respiration rate level, determining, by a processor, that the wearer is under emotional stress.
11. The computer-implemented method of claim 3 , further comprising the step of:
a. receiving, by a processor, an image captured by the eyewear;
b. after receiving the image, identifying, by a processor, an object in the received image; and
c. notifying, by a processor, the wearer of the association of the wearer's mental state and the object in the received image.
12. The computer-implemented method of claim 11 , wherein the object is a person.
13. The computer-implemented method of claim 3 , further comprising the steps of:
a. notifying, by a processor, the wearer of the eyewear of their mental state; and
b. providing, by a processor, to the wearer one or more suggested actions to address their current mental state.
14. The computer-implemented method of claim 13 , wherein:
a. the wearer's mental state is one of emotional stress; and
b. the one or more suggested actions comprise one or more techniques for reducing stress.
15. A computer-implemented method of assessing the mental state of a wearer of eyewear comprising one or more sensors coupled to the eyewear, the one or more sensors being adapted to detect one or more characteristics of the wearer of the eyewear, wherein the one or more characteristics are associated with the wearer's mental state, the method comprising:
a. receiving, by a processor, one or more first images from the eye facing camera, wherein at least one of the one or more first images relates to at least one characteristic associated with the wearer;
b. analyzing, by a processor, the at least one or more first images to determine the at least one characteristic; and
c. facilitating, by a processor, determination of a mental state of the wearer based on the at least one characteristic.
16. The computer-implemented method of claim 15 , further comprising the steps of:
a. receiving, by a processor, one or more second images from the front facing camera;
b. determining, by a processor, the mental state of wearer based at least in part on one or more wearer characteristics selected from a group consisting of:
i. the wearer's pupil size,
ii. the wearer's respiration rate,
iii. the wearer's perspiration rate,
iv. the paleness of the wearer's skin, and
v. the wearer's brain activity level;
c. associating, by a processor, the mental state of the wearer with a factor selected from a group consisting of:
i. an object in close proximity to the wearer; and
ii. an activity in which the wearer is participating in.
17. The computer-implemented method of claim 16 , wherein the step of associating the mental state of the wearer with the factor further comprises the steps of:
a. detecting, by a processor, the object in the one or more second images; and
b. storing, in memory, the association of the wearer's mental state with the object.
18. The computer-implemented method of claim 16 , wherein the step of associating the mental state of the wearer with the factor further comprises the steps of:
a. using the one or more second images to determine an activity that the wearer was engaged in when the one or more images were taken; and
b. storing, in memory, the association of the wearer's mental state with the activity.
19. The computer-implemented method of claim 16 , wherein the eyewear further comprises a near-field communication sensor that is configured to detect the presence of the at least one of the factor, the method further comprising the steps of:
a. receiving, by a processor, a signal from the near-field communication sensor;
b. determining, by a processor, the identity of the factor from the received signal; and
c. storing, by a processor, the identity of the factor and the mental state of the wearer in memory.
20. The computer-implemented method of claim 16 , wherein the step of determining the mental state of the wearer further comprises:
a. measuring, by a processor, a pupil size of the wearer;
b. measuring, by a processor, a respiration rate level of the wearer;
c. comparing, by a processor,
i. the pupil size of the wearer to a predefined pupil size; and
ii. comparing the respiration rate of the wearer to a predefined respiration rate; and
d. determining, by a processor, that the wearer is in a stressful state in response to determining that the measured pupil size exceeds the predefined pupil size and the measured respiration rate exceeds the predefined respiration rate.
21. A system for assessing the mental state of a wearer of a wearable device comprising:
a. at least one processor; and
b. memory operatively coupled to the at least one processor,
wherein the at least one processor is configured to:
i. receive one or more signals from a device selected from a group consisting of: a helmet and eyewear, the device having one or more sensors embedded therein, wherein each of the one or more signals relates to at least one characteristic associated with the wearer, the at least one characteristic selected from a group consisting of:
1. pupil size,
2. heart rate,
3. perspiration rate,
4. respiration rate,
5. movement, and
6. brainwave activity;
ii. analyze the one or more received signals to determine the at least one characteristic;
iii. determine a mental state of the wearer based at least in part on the at least one characteristic;
iv. associate the mental state of the wearer with one or more stimuli selected from a group consisting of: (1) an object; (2) an activity; (3) an internal context of the wearer; and (4) an external context of the wearer.
22. The system of claim 22 , wherein the at least one processor is further configured to monitor at least one of the wearer's pupil size and brainwave activity over time to determine whether the wearer has sustained a concussion.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/578,039 US20160066829A1 (en) | 2014-09-05 | 2014-12-19 | Wearable mental state monitor computer apparatus, systems, and related methods |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201462046406P | 2014-09-05 | 2014-09-05 | |
US14/578,039 US20160066829A1 (en) | 2014-09-05 | 2014-12-19 | Wearable mental state monitor computer apparatus, systems, and related methods |
Publications (1)
Publication Number | Publication Date |
---|---|
US20160066829A1 true US20160066829A1 (en) | 2016-03-10 |
Family
ID=55436366
Family Applications (12)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/550,406 Active 2037-11-20 US10448867B2 (en) | 2014-09-05 | 2014-11-21 | Wearable gait monitoring apparatus, systems, and related methods |
US14/562,454 Active 2034-12-15 US9795324B2 (en) | 2014-09-05 | 2014-12-05 | System for monitoring individuals as they age in place |
US14/578,039 Abandoned US20160066829A1 (en) | 2014-09-05 | 2014-12-19 | Wearable mental state monitor computer apparatus, systems, and related methods |
US14/588,122 Active 2035-02-18 US9649052B2 (en) | 2014-09-05 | 2014-12-31 | Systems, apparatus, and methods for using eyewear, or other wearable item, to confirm the identity of an individual |
US14/610,439 Active 2037-03-18 US10307085B2 (en) | 2014-09-05 | 2015-01-30 | Wearable physiology monitor computer apparatus, systems, and related methods |
US14/610,589 Abandoned US20160071423A1 (en) | 2014-09-05 | 2015-01-30 | Systems and method for monitoring an individual's compliance with a weight loss plan |
US14/610,628 Abandoned US20160070122A1 (en) | 2014-09-05 | 2015-01-30 | Computerized replacement temple for standard eyewear |
US15/594,898 Active 2035-02-02 US10188323B2 (en) | 2014-09-05 | 2017-05-15 | Systems, apparatus, and methods for using eyewear, or other wearable item, to confirm the identity of an individual |
US15/791,196 Abandoned US20180042523A1 (en) | 2014-09-05 | 2017-10-23 | System for monitoring individuals as they age in place |
US16/259,646 Active US10542915B2 (en) | 2014-09-05 | 2019-01-28 | Systems, apparatus, and methods for using a wearable device to confirm the identity of an individual |
US16/429,480 Active US10694981B2 (en) | 2014-09-05 | 2019-06-03 | Wearable physiology monitor computer apparatus, systems, and related methods |
US16/657,982 Abandoned US20200046260A1 (en) | 2014-09-05 | 2019-10-18 | Wearable gait monitoring apparatus, systems, and related methods |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/550,406 Active 2037-11-20 US10448867B2 (en) | 2014-09-05 | 2014-11-21 | Wearable gait monitoring apparatus, systems, and related methods |
US14/562,454 Active 2034-12-15 US9795324B2 (en) | 2014-09-05 | 2014-12-05 | System for monitoring individuals as they age in place |
Family Applications After (9)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/588,122 Active 2035-02-18 US9649052B2 (en) | 2014-09-05 | 2014-12-31 | Systems, apparatus, and methods for using eyewear, or other wearable item, to confirm the identity of an individual |
US14/610,439 Active 2037-03-18 US10307085B2 (en) | 2014-09-05 | 2015-01-30 | Wearable physiology monitor computer apparatus, systems, and related methods |
US14/610,589 Abandoned US20160071423A1 (en) | 2014-09-05 | 2015-01-30 | Systems and method for monitoring an individual's compliance with a weight loss plan |
US14/610,628 Abandoned US20160070122A1 (en) | 2014-09-05 | 2015-01-30 | Computerized replacement temple for standard eyewear |
US15/594,898 Active 2035-02-02 US10188323B2 (en) | 2014-09-05 | 2017-05-15 | Systems, apparatus, and methods for using eyewear, or other wearable item, to confirm the identity of an individual |
US15/791,196 Abandoned US20180042523A1 (en) | 2014-09-05 | 2017-10-23 | System for monitoring individuals as they age in place |
US16/259,646 Active US10542915B2 (en) | 2014-09-05 | 2019-01-28 | Systems, apparatus, and methods for using a wearable device to confirm the identity of an individual |
US16/429,480 Active US10694981B2 (en) | 2014-09-05 | 2019-06-03 | Wearable physiology monitor computer apparatus, systems, and related methods |
US16/657,982 Abandoned US20200046260A1 (en) | 2014-09-05 | 2019-10-18 | Wearable gait monitoring apparatus, systems, and related methods |
Country Status (4)
Country | Link |
---|---|
US (12) | US10448867B2 (en) |
EP (3) | EP3148435B1 (en) |
CA (3) | CA2953856A1 (en) |
WO (3) | WO2016037117A1 (en) |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160350138A1 (en) * | 2015-05-31 | 2016-12-01 | Roya Caroline SALAS | Biofeedback system |
CN106377271A (en) * | 2016-10-20 | 2017-02-08 | 中国矿业大学 | Wearable method and wearable device for monitoring and regulating physiology and psychology |
US9930102B1 (en) | 2015-03-27 | 2018-03-27 | Intuit Inc. | Method and system for using emotional state data to tailor the user experience of an interactive software system |
CN108401129A (en) * | 2018-03-22 | 2018-08-14 | 广东小天才科技有限公司 | Video call method, device, terminal based on Wearable and storage medium |
US20180241973A1 (en) * | 2017-02-21 | 2018-08-23 | Janet Newell | Video and audio recording system and method |
WO2018156992A1 (en) * | 2017-02-23 | 2018-08-30 | Miller Charles Robert Iii | Device and system for user context-cortical sensing and determination |
US10169827B1 (en) | 2015-03-27 | 2019-01-01 | Intuit Inc. | Method and system for adapting a user experience provided through an interactive software system to the content being delivered and the predicted emotional impact on the user of that content |
US20190167226A1 (en) * | 2017-12-04 | 2019-06-06 | International Business Machines Corporation | Infant gastrointestinal monitor |
US10332122B1 (en) | 2015-07-27 | 2019-06-25 | Intuit Inc. | Obtaining and analyzing user physiological data to determine whether a user would benefit from user support |
US10387173B1 (en) | 2015-03-27 | 2019-08-20 | Intuit Inc. | Method and system for using emotional state data to tailor the user experience of an interactive software system |
CN110447232A (en) * | 2017-03-31 | 2019-11-12 | 三星电子株式会社 | For determining the electronic equipment and its control method of user emotion |
US11331003B2 (en) | 2018-03-27 | 2022-05-17 | Samsung Electronics Co., Ltd. | Context-aware respiration rate determination using an electronic device |
Families Citing this family (118)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9900669B2 (en) | 2004-11-02 | 2018-02-20 | Pierre Touma | Wireless motion sensor system and method |
US10831982B2 (en) | 2009-10-14 | 2020-11-10 | Iplcontent, Llc | Hands-free presenting device |
US9968297B2 (en) * | 2012-06-14 | 2018-05-15 | Medibotics Llc | EEG glasses (electroencephalographic eyewear) |
US11612786B2 (en) * | 2012-08-31 | 2023-03-28 | Blue Goji Llc | System and method for targeted neurological therapy using brainwave entrainment with passive treatment |
US20140142442A1 (en) * | 2012-11-19 | 2014-05-22 | Judy Sibille SNOW | Audio Feedback for Medical Conditions |
US9400549B2 (en) | 2013-03-08 | 2016-07-26 | Chi Fai Ho | Method and system for a new-era electronic book |
WO2014179507A1 (en) * | 2013-04-30 | 2014-11-06 | White Chester | Body impact bracing apparatus |
US11918375B2 (en) | 2014-09-05 | 2024-03-05 | Beijing Zitiao Network Technology Co., Ltd. | Wearable environmental pollution monitor computer apparatus, systems, and related methods |
US10448867B2 (en) | 2014-09-05 | 2019-10-22 | Vision Service Plan | Wearable gait monitoring apparatus, systems, and related methods |
US10617342B2 (en) | 2014-09-05 | 2020-04-14 | Vision Service Plan | Systems, apparatus, and methods for using a wearable device to monitor operator alertness |
FR3028980B1 (en) * | 2014-11-20 | 2017-01-13 | Oberthur Technologies | METHOD AND DEVICE FOR AUTHENTICATING A USER |
US10215568B2 (en) | 2015-01-30 | 2019-02-26 | Vision Service Plan | Systems and methods for tracking motion, performance, and other data for an individual such as a winter sports athlete |
KR20160101497A (en) * | 2015-02-17 | 2016-08-25 | 삼성전자주식회사 | Wearable device and method for operating thereof |
US9886633B2 (en) * | 2015-02-23 | 2018-02-06 | Vivint, Inc. | Techniques for identifying and indexing distinguishing features in a video feed |
NZ773820A (en) * | 2015-03-16 | 2022-07-29 | Magic Leap Inc | Methods and systems for diagnosing and treating health ailments |
EP3274791A4 (en) * | 2015-03-27 | 2018-10-31 | Intel Corporation | Motion tracking using electronic devices |
EP3295371A4 (en) | 2015-05-11 | 2018-06-27 | Magic Leap, Inc. | Devices, methods and systems for biometric user recognition utilizing neural networks |
US10154129B2 (en) * | 2015-05-15 | 2018-12-11 | Polar Electro Oy | Wearable electronic apparatus |
US10799122B2 (en) | 2015-06-14 | 2020-10-13 | Facense Ltd. | Utilizing correlations between PPG signals and iPPG signals to improve detection of physiological responses |
US11154203B2 (en) | 2015-06-14 | 2021-10-26 | Facense Ltd. | Detecting fever from images and temperatures |
US11064892B2 (en) | 2015-06-14 | 2021-07-20 | Facense Ltd. | Detecting a transient ischemic attack using photoplethysmogram signals |
US10791938B2 (en) | 2015-06-14 | 2020-10-06 | Facense Ltd. | Smartglasses for detecting congestive heart failure |
US11103139B2 (en) | 2015-06-14 | 2021-08-31 | Facense Ltd. | Detecting fever from video images and a baseline |
US11103140B2 (en) | 2015-06-14 | 2021-08-31 | Facense Ltd. | Monitoring blood sugar level with a comfortable head-mounted device |
US20170225033A1 (en) * | 2015-06-23 | 2017-08-10 | Ipcomm Llc | Method and Apparatus for Analysis of Gait and to Provide Haptic and Visual Corrective Feedback |
DK3328276T3 (en) * | 2015-07-27 | 2023-09-18 | Massachusetts Inst Technology | APPARATUS IN CONNECTION WITH MONITORING METABOLISM |
US20170061823A1 (en) * | 2015-09-02 | 2017-03-02 | Hello Doctor Ltd. | System for tracking and monitoring personal medical data and encouraging to follow personalized condition-based profile and method thereof |
US11272864B2 (en) | 2015-09-14 | 2022-03-15 | Health Care Originals, Inc. | Respiratory disease monitoring wearable apparatus |
ES2830154T3 (en) | 2015-10-01 | 2021-06-03 | Dnanudge Ltd | Method, apparatus and system to transfer biological information safely |
US10861594B2 (en) | 2015-10-01 | 2020-12-08 | Dnanudge Limited | Product recommendation system and method |
JP6599473B2 (en) * | 2015-10-13 | 2019-10-30 | アルプスアルパイン株式会社 | Walking measurement device, walking measurement method, and program |
US10014967B2 (en) * | 2015-11-23 | 2018-07-03 | Huami Inc. | System and method for authenticating a broadcast device using facial recognition |
US10096383B2 (en) | 2015-11-24 | 2018-10-09 | International Business Machines Corporation | Performing a health analysis using a smart floor mat |
US10105095B2 (en) * | 2015-11-30 | 2018-10-23 | Oura Health Oy | Method and system for defining balance between physical activity and rest |
EP3178379A1 (en) * | 2015-12-09 | 2017-06-14 | Rythm | Method and device for bioelectric physiological signal acquisition and processing |
US10610146B1 (en) * | 2015-12-21 | 2020-04-07 | Dp Technologies, Inc. | Utilizing wearable devices in an internet of things environment |
US11094418B2 (en) * | 2015-12-31 | 2021-08-17 | Nokia Technologies Oy | Optimized biological measurement |
WO2017163227A1 (en) * | 2016-03-25 | 2017-09-28 | Randolph Andrae | User authentication using biometric information |
US10251597B2 (en) | 2016-04-21 | 2019-04-09 | Viavi Solutions Inc. | Health tracking device |
CA3021460A1 (en) * | 2016-04-22 | 2017-10-26 | Vioptix, Inc. | Determining absolute and relative tissue oxygen saturation |
US10955269B2 (en) | 2016-05-20 | 2021-03-23 | Health Care Originals, Inc. | Wearable apparatus |
US20170345274A1 (en) * | 2016-05-27 | 2017-11-30 | General Scientific Corporation | Neck posture recording and warning device |
US9955305B2 (en) | 2016-06-01 | 2018-04-24 | Tile, Inc. | User intervention based on proximity between tracking devices |
WO2018030734A1 (en) * | 2016-08-09 | 2018-02-15 | 주식회사 비플렉스 | 3d simulation method and apparatus |
WO2018027253A1 (en) * | 2016-08-11 | 2018-02-15 | Bloomfield Lochlan John | Health management system and method |
CN109690689A (en) | 2016-09-14 | 2019-04-26 | 豪夫迈·罗氏有限公司 | Digital biometric for progressive MS marks |
CN106510719B (en) * | 2016-09-30 | 2023-11-28 | 歌尔股份有限公司 | User gesture monitoring method and wearable device |
US10539549B2 (en) * | 2016-10-13 | 2020-01-21 | Worcester Polytechnic Institute | Mobile blood alcohol content and impairment sensing device |
WO2018072035A1 (en) * | 2016-10-20 | 2018-04-26 | Datapult, Inc. | Sickness prediction application system |
EP3534778B1 (en) | 2016-11-01 | 2022-04-20 | Polyvagal Science LLC | Systems for reducing sound sensitivities and improving auditory processing, behavioral state regulation and social engagement |
US11062175B2 (en) * | 2016-11-22 | 2021-07-13 | Japan Aerospace Exploration Agency | System, method, and program for estimating reduced attention state, and storage medium storing the same program |
GB201620638D0 (en) | 2016-12-05 | 2017-01-18 | Equi+Poise Ltd | A gait analysis system |
KR20180089803A (en) | 2017-02-01 | 2018-08-09 | 삼성전자주식회사 | Electronic apparatus and method for processing authentication |
US10973446B2 (en) * | 2017-02-13 | 2021-04-13 | David Schie | Device to extract physiological information and method therefor |
US11622716B2 (en) * | 2017-02-13 | 2023-04-11 | Health Care Originals, Inc. | Wearable physiological monitoring systems and methods |
EP3579751A1 (en) | 2017-02-13 | 2019-12-18 | Starkey Laboratories, Inc. | Fall prediction system and method of using same |
JP6894252B2 (en) * | 2017-02-16 | 2021-06-30 | 日本光電工業株式会社 | Sensor device and watching device |
SE541712C2 (en) * | 2017-02-22 | 2019-12-03 | Next Step Dynamics Ab | Method and apparatus for health prediction |
WO2018182159A1 (en) * | 2017-03-28 | 2018-10-04 | 문명일 | Smart glasses capable of processing virtual object |
CN107049338A (en) * | 2017-04-12 | 2017-08-18 | 河南工业大学 | A kind of medical use mood detection means communicated based on computer |
US11559252B2 (en) | 2017-05-08 | 2023-01-24 | Starkey Laboratories, Inc. | Hearing assistance device incorporating virtual audio interface for therapy guidance |
US11589807B2 (en) * | 2017-05-11 | 2023-02-28 | The Regents Of The University Of California | Biosensor for monitoring eyedrop usage compliance |
US10699247B2 (en) | 2017-05-16 | 2020-06-30 | Under Armour, Inc. | Systems and methods for providing health task notifications |
US10685585B2 (en) * | 2017-06-27 | 2020-06-16 | International Business Machines Corporation | Physical activity and dietary based services |
US10534203B2 (en) * | 2017-07-31 | 2020-01-14 | Snap Inc. | Near-field antenna for eyewear |
US11373450B2 (en) | 2017-08-11 | 2022-06-28 | Tectus Corporation | Eye-mounted authentication system |
US10832590B2 (en) * | 2017-09-13 | 2020-11-10 | At&T Intellectual Property I, L.P. | Monitoring food intake |
US11062572B1 (en) * | 2017-09-20 | 2021-07-13 | Amazon Technologies, Inc. | Visual indicator for head-mounted device |
US10073998B1 (en) * | 2017-11-29 | 2018-09-11 | The United States Of America As Representd By The Secretary Of The Navy | Multifunction wearable object identified glasses for the visually handicapped |
US11092998B1 (en) | 2018-01-12 | 2021-08-17 | Snap Inc. | Eyewear device with fingerprint sensor for user input |
CN108089326B (en) * | 2018-02-01 | 2023-12-26 | 北京七鑫易维信息技术有限公司 | Device suitable for being used with glasses |
US11183291B2 (en) | 2018-02-12 | 2021-11-23 | Zoe Limited | Generating personalized nutritional recommendations using predicted values of biomarkers |
US11915151B2 (en) | 2018-08-27 | 2024-02-27 | Zoe Limited | Accuracy of test data outside the clinic |
WO2019157450A1 (en) * | 2018-02-12 | 2019-08-15 | Cornell University | Methods and systems for concussion management using cold stimulus |
US11295860B2 (en) | 2018-02-12 | 2022-04-05 | Zoe Limited | Using at home measures to predict clinical state and improving the accuracy of at home measurements/predictions data associated with circadian rhythm and meal timing |
US11348479B2 (en) | 2018-05-23 | 2022-05-31 | Zoe Limited | Accuracy of measuring nutritional responses in a non-clinical setting |
US11183080B2 (en) | 2018-02-12 | 2021-11-23 | Zoe Limited | Generating predicted values of biomarkers for scoring food |
CA3091209C (en) * | 2018-03-01 | 2021-08-31 | Polyvagal Science LLC | Systems and methods for modulating physiological state |
US20210020295A1 (en) * | 2018-03-30 | 2021-01-21 | Hitachi, Ltd. | Physical function independence support device of physical function and method therefor |
US10424035B1 (en) * | 2018-05-16 | 2019-09-24 | Trungram Gyaltrul R. Sherpa | Monitoring conditions associated with remote individuals over a data communication network and automatically notifying responsive to detecting customized emergency conditions |
US10438479B1 (en) * | 2018-05-21 | 2019-10-08 | International Business Machines Corporation | Safety enhancement for persons under the care of others |
US11054638B2 (en) * | 2018-06-13 | 2021-07-06 | Reavire, Inc. | Tracking pointing direction of device |
US10121355B1 (en) * | 2018-06-26 | 2018-11-06 | Capital One Services, Llc | Condition-responsive wearable device for sensing and indicating proximity of an article with a specific characteristic |
US10922397B2 (en) | 2018-07-24 | 2021-02-16 | Dnanudge Limited | Method and device for comparing personal biological data of two users |
US10582897B2 (en) | 2018-07-24 | 2020-03-10 | Dnanudge Limited | Method and device for comparing personal biological data of two users |
US10722128B2 (en) | 2018-08-01 | 2020-07-28 | Vision Service Plan | Heart rate detection system and method |
US11380215B2 (en) * | 2018-08-30 | 2022-07-05 | Kyndryl, Inc. | Reward-based ecosystem for tracking nutritional consumption |
CN113168903A (en) * | 2018-09-12 | 2021-07-23 | 基因动力公司 | Product recommendation system and method |
WO2020061209A1 (en) * | 2018-09-18 | 2020-03-26 | Biointellisense, Inc. | Validation, compliance, and/or intervention with ear device |
CN109543546B (en) * | 2018-10-26 | 2022-12-20 | 复旦大学 | Gait age estimation method based on depth sequence distribution regression |
EP3895141B1 (en) | 2018-12-15 | 2024-01-24 | Starkey Laboratories, Inc. | Hearing assistance system with enhanced fall detection features |
US11638563B2 (en) | 2018-12-27 | 2023-05-02 | Starkey Laboratories, Inc. | Predictive fall event management system and method of using same |
EP3908969A1 (en) * | 2019-01-07 | 2021-11-17 | MetraLabs GmbH Neue Technologien und Systeme | Method and system for capturing the sequence of movement of a person |
US11150788B2 (en) * | 2019-03-14 | 2021-10-19 | Ebay Inc. | Augmented or virtual reality (AR/VR) companion device techniques |
US10811140B2 (en) | 2019-03-19 | 2020-10-20 | Dnanudge Limited | Secure set-up of genetic related user account |
US10699806B1 (en) | 2019-04-15 | 2020-06-30 | Dnanudge Limited | Monitoring system, wearable monitoring device and method |
US11156856B2 (en) | 2019-05-02 | 2021-10-26 | Tom Nepola | Eyewear with wearing status detector |
US11007406B2 (en) * | 2019-05-03 | 2021-05-18 | Xperience Robotics, Inc. | Wearable device systems and methods for guiding physical movements |
US11663927B2 (en) * | 2019-05-10 | 2023-05-30 | Tso-Cheng Chien | Food quantum tracking tools and methods related thereto |
ES1231504Y (en) * | 2019-05-14 | 2019-09-13 | I4Life Innovacion Y Desarrollos S L | Multifunctional Unlocking Device |
US11786694B2 (en) | 2019-05-24 | 2023-10-17 | NeuroLight, Inc. | Device, method, and app for facilitating sleep |
US11223939B1 (en) * | 2019-07-31 | 2022-01-11 | United Services Automobile Association (Usaa) | Environmental conditions monitoring system |
WO2021084903A1 (en) * | 2019-11-01 | 2021-05-06 | テルモ株式会社 | Image management system, wearable device, image management method, and image management program |
US11310322B2 (en) | 2019-11-21 | 2022-04-19 | Blackberry Limited | Method and system for pairing a chassis and container in an asset tracking system |
DE102020102281A1 (en) * | 2019-12-19 | 2021-06-24 | USound GmbH | Glasses with charging interface |
GB2590802A (en) | 2020-01-03 | 2021-07-07 | Dnanudge Ltd | Method and device for comparing personal biological data of two users |
US11948672B2 (en) * | 2020-02-27 | 2024-04-02 | Todd Martin | Mobile intelligent injury minimization system and method |
US10991190B1 (en) | 2020-07-20 | 2021-04-27 | Abbott Laboratories | Digital pass verification systems and methods |
CN112150767B (en) * | 2020-09-27 | 2021-11-23 | 河南城建学院 | Fatigue driving monitoring system based on Internet of things and computer |
KR102525485B1 (en) * | 2020-12-02 | 2023-04-25 | 임지호 | Self-authentication wearable device for the visually impaired |
CN112562260B (en) * | 2020-12-16 | 2022-08-09 | 浙江大华技术股份有限公司 | Anti-lost method and device |
DE102021104705A1 (en) | 2021-02-26 | 2022-09-01 | seiwo Technik GmbH | Method and system for recording the state of health of a person |
DE202021100976U1 (en) | 2021-02-26 | 2021-05-18 | seiwo Technik GmbH | System for recording the state of health of a person |
CN117561022A (en) * | 2021-07-02 | 2024-02-13 | 研究三角协会 | Systems, methods, and apparatus for detecting viral respiratory disease in pre-symptomatic and asymptomatic infected persons |
CN113457108B (en) * | 2021-07-07 | 2022-07-15 | 首都体育学院 | Cognitive characterization-based exercise performance improving method and device |
US20230028690A1 (en) * | 2021-07-23 | 2023-01-26 | Consumer Safety Technology Llc | Method and system of deploying a wearable transdermal based vehicle ignition interlock |
WO2023021509A1 (en) * | 2021-08-17 | 2023-02-23 | Brainwatch Tech Ltd. | Methods, systems, and devices for determining a status of brain and/or nerve functions of a patient |
US11806078B1 (en) | 2022-05-01 | 2023-11-07 | Globe Biomedical, Inc. | Tear meniscus detection and evaluation system |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030195398A1 (en) * | 2000-05-31 | 2003-10-16 | Kabushiki Kaisha Toshiba | Life support apparatus and method and method for providing advertisement information |
US20070273611A1 (en) * | 2004-04-01 | 2007-11-29 | Torch William C | Biosensors, communicators, and controllers monitoring eye movement and methods for using them |
US20100280336A1 (en) * | 2009-04-30 | 2010-11-04 | Medtronic, Inc. | Anxiety disorder monitoring |
US20110142413A1 (en) * | 2009-12-04 | 2011-06-16 | Lg Electronics Inc. | Digital data reproducing apparatus and method for controlling the same |
US20110169932A1 (en) * | 2010-01-06 | 2011-07-14 | Clear View Technologies Inc. | Wireless Facial Recognition |
US20120290401A1 (en) * | 2011-05-11 | 2012-11-15 | Google Inc. | Gaze tracking system |
US20130096397A1 (en) * | 2010-06-17 | 2013-04-18 | Nec Corporation | Sensitivity evaluation system, sensitivity evaluation method, and program |
US20130127980A1 (en) * | 2010-02-28 | 2013-05-23 | Osterhout Group, Inc. | Video display modification based on sensor input for a see-through near-to-eye display |
US20130317384A1 (en) * | 2012-05-25 | 2013-11-28 | Emotiv Lifesciences Inc. | System and Method for Instructing a Behavior Change in a User |
Family Cites Families (244)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3505879A (en) | 1968-07-31 | 1970-04-14 | Ford Motor Co | Electrical generator including components of an automotive vehicle mechanical speedometer |
US3548663A (en) | 1968-07-31 | 1970-12-22 | Ford Motor Co | Electrical generator including components of an automotive vehicle mechanical speedometer |
US3972038A (en) | 1975-03-28 | 1976-07-27 | Nasa | Accelerometer telemetry system |
US4100401A (en) | 1977-01-13 | 1978-07-11 | Tutt Eugene F | Calorie calculator-chronometer |
DE2709412A1 (en) | 1977-03-04 | 1978-09-07 | Max Baermann | Eddy current tachometer with temperature compensation |
US4195642A (en) | 1978-01-03 | 1980-04-01 | Beehive International | Wearable heart rate monitor |
GB1593839A (en) | 1978-05-26 | 1981-07-22 | Pringle R D | Performance testing device |
US4434801A (en) | 1980-04-30 | 1984-03-06 | Biotechnology, Inc. | Apparatus for testing physical condition of a self-propelled vehicle rider |
US4407295A (en) | 1980-10-16 | 1983-10-04 | Dna Medical, Inc. | Miniature physiological monitor with interchangeable sensors |
US4855942A (en) | 1987-10-28 | 1989-08-08 | Elexis Corporation | Pedometer and/or calorie measuring device and method |
US4878749A (en) | 1988-06-29 | 1989-11-07 | Mcgee James E | Protective eyewear with interchangeable decorative frames |
US4919530A (en) | 1989-01-25 | 1990-04-24 | Hyman Roger L | Eyeglass assembly |
US5670872A (en) | 1992-06-22 | 1997-09-23 | U.S. Philips Corporation | System and device with vertical and rotary wheel-velocity-measuring for determining vehicle displacement |
US5497143A (en) | 1993-06-24 | 1996-03-05 | Casio Computer Co., Ltd. | Electronic device for a vehicle |
US5422816A (en) | 1994-02-22 | 1995-06-06 | Trimble Navigation Limited | Portable personal navigation tracking system |
US5452480A (en) | 1994-04-15 | 1995-09-26 | Electric Eyewear Systems, Inc. | Ski goggles |
US7386401B2 (en) | 1994-11-21 | 2008-06-10 | Phatrat Technology, Llc | Helmet that reports impact information, and associated methods |
US5585871A (en) | 1995-05-26 | 1996-12-17 | Linden; Harry | Multi-function display apparatus |
US5746501A (en) | 1995-09-01 | 1998-05-05 | Chien; Tseng Lu | Portable object having a fastening band illuminated by a super thin lighting element |
US6183425B1 (en) | 1995-10-13 | 2001-02-06 | The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration | Method and apparatus for monitoring of daily activity in terms of ground reaction forces |
US5966680A (en) | 1996-02-15 | 1999-10-12 | Butnaru; Hanan | Motion sickness/vertigo prevention device and method |
US6790178B1 (en) | 1999-09-24 | 2004-09-14 | Healthetech, Inc. | Physiological monitor and associated computation, display and communication unit |
JP3608204B2 (en) | 1996-04-08 | 2005-01-05 | セイコーエプソン株式会社 | Exercise prescription support device |
US5976083A (en) | 1997-07-30 | 1999-11-02 | Living Systems, Inc. | Portable aerobic fitness monitor for walking and running |
US5891042A (en) | 1997-09-09 | 1999-04-06 | Acumen, Inc. | Fitness monitoring device having an electronic pedometer and a wireless heart rate monitor |
US6381482B1 (en) | 1998-05-13 | 2002-04-30 | Georgia Tech Research Corp. | Fabric or garment with integrated flexible information infrastructure |
US6013007A (en) | 1998-03-26 | 2000-01-11 | Liquid Spark, Llc | Athlete's GPS-based performance monitor |
US5931764A (en) | 1998-06-24 | 1999-08-03 | Viztec, Inc. | Wearable device with flexible display |
US7376238B1 (en) | 1998-09-18 | 2008-05-20 | Rivas Technologies International, Inc. | Pulse rate, pressure and heart condition monitoring glasses |
US6218958B1 (en) | 1998-10-08 | 2001-04-17 | International Business Machines Corporation | Integrated touch-skin notification system for wearable computing devices |
US6532298B1 (en) | 1998-11-25 | 2003-03-11 | Iridian Technologies, Inc. | Portable authentication device and method using iris patterns |
CN1297528A (en) | 1999-02-05 | 2001-05-30 | 柯蒂斯仪器股份有限公司 | Shaft sensor for angular velocity, torque and power |
US6736759B1 (en) | 1999-11-09 | 2004-05-18 | Paragon Solutions, Llc | Exercise monitoring system and methods |
US6431705B1 (en) | 1999-11-10 | 2002-08-13 | Infoeye | Eyewear heart rate monitor |
US7156809B2 (en) | 1999-12-17 | 2007-01-02 | Q-Tec Systems Llc | Method and apparatus for health and disease management combining patient data monitoring with wireless internet connectivity |
US7454002B1 (en) | 2000-01-03 | 2008-11-18 | Sportbrain, Inc. | Integrating personal data capturing functionality into a portable computing device and a wireless communication device |
US6513532B2 (en) | 2000-01-19 | 2003-02-04 | Healthetech, Inc. | Diet and activity-monitoring device |
JP2001297318A (en) | 2000-04-14 | 2001-10-26 | Omron Corp | Pedometer |
PT1285409E (en) | 2000-05-16 | 2005-08-31 | Swisscom Mobile Ag | PROCESS OF IDENTIFICATION OR BIOMETRIC AUTHENTICATION |
US6325507B1 (en) | 2000-06-02 | 2001-12-04 | Oakley, Inc. | Eyewear retention system extending across the top of a wearer's head |
DE10043797A1 (en) | 2000-09-06 | 2002-03-28 | Daimler Chrysler Ag | Integrated traffic monitoring system |
US20050054942A1 (en) | 2002-01-22 | 2005-03-10 | Melker Richard J. | System and method for therapeutic drug monitoring |
JP2004516462A (en) | 2000-12-15 | 2004-06-03 | ヌーリー テクノロジーズ リミテッド | Location based weather nowcast system and method |
US20020151810A1 (en) | 2001-04-16 | 2002-10-17 | Acumen, Inc. | Wrist-based fitness monitoring devices |
US6769767B2 (en) | 2001-04-30 | 2004-08-03 | Qr Spex, Inc. | Eyewear with exchangeable temples housing a transceiver forming ad hoc networks with other devices |
JP2003033328A (en) | 2001-07-19 | 2003-02-04 | Nippon Seimitsu Sokki Kk | Heart rate monitor and method for measuring heart rate |
JP2003242584A (en) | 2002-02-13 | 2003-08-29 | Seiko Instruments Inc | Wearable electronic equipment system and wearable electronic equipment |
KR101090667B1 (en) | 2002-04-22 | 2011-12-07 | 마시오 마크 아우렐리오 마틴스 애브리우 | Apparatus and method for measuring biological parameters |
US9153074B2 (en) | 2011-07-18 | 2015-10-06 | Dylan T X Zhou | Wearable augmented reality eyeglass communication device including mobile phone and mobile computing via virtual touch screen gesture control and neuron command |
GB2396421A (en) | 2002-12-16 | 2004-06-23 | Orange Personal Comm Serv Ltd | Head-worn device measuring brain and facial muscle activity |
AU2004263812B2 (en) | 2003-02-26 | 2009-10-08 | Marcio Marc Aurelio Martins Abreu | Apparatus and method for measuring biologic parameters |
US8109629B2 (en) | 2003-10-09 | 2012-02-07 | Ipventure, Inc. | Eyewear supporting electrical components and apparatus therefor |
US7192136B2 (en) | 2003-04-15 | 2007-03-20 | Howell Thomas A | Tethered electrical components for eyeglasses |
US7922321B2 (en) | 2003-10-09 | 2011-04-12 | Ipventure, Inc. | Eyewear supporting after-market electrical components |
US7500747B2 (en) | 2003-10-09 | 2009-03-10 | Ipventure, Inc. | Eyeglasses with electrical components |
US7792552B2 (en) | 2003-04-15 | 2010-09-07 | Ipventure, Inc. | Eyeglasses for wireless communications |
US7255437B2 (en) | 2003-10-09 | 2007-08-14 | Howell Thomas A | Eyeglasses with activity monitoring |
US7806525B2 (en) | 2003-10-09 | 2010-10-05 | Ipventure, Inc. | Eyeglasses having a camera |
US7380936B2 (en) | 2003-10-09 | 2008-06-03 | Ipventure, Inc. | Eyeglasses with a clock or other electrical component |
US8465151B2 (en) | 2003-04-15 | 2013-06-18 | Ipventure, Inc. | Eyewear with multi-part temple for supporting one or more electrical components |
US7500746B1 (en) | 2004-04-15 | 2009-03-10 | Ip Venture, Inc. | Eyewear with radiation detection system |
US7581833B2 (en) | 2003-10-09 | 2009-09-01 | Ipventure, Inc. | Eyewear supporting after-market electrical components |
US20050033200A1 (en) | 2003-08-05 | 2005-02-10 | Soehren Wayne A. | Human motion identification and measurement system and method |
US7059717B2 (en) | 2003-08-11 | 2006-06-13 | Bloch Nigel K | Eyeglasses with interchangable temple-members |
US10310296B2 (en) | 2003-10-09 | 2019-06-04 | Ingeniospec, Llc | Eyewear with printed circuit board |
US7438410B1 (en) | 2003-10-09 | 2008-10-21 | Ip Venture, Inc. | Tethered electrical components for eyeglasses |
US7677723B2 (en) | 2003-10-09 | 2010-03-16 | Ipventure, Inc. | Eyeglasses with a heart rate monitor |
FR2860700B1 (en) | 2003-10-10 | 2005-12-09 | Commissariat Energie Atomique | CROWN CONTROL DEVICE |
US8337013B2 (en) | 2004-07-28 | 2012-12-25 | Ipventure, Inc. | Eyeglasses with RFID tags or with a strap |
US8915588B2 (en) | 2004-11-02 | 2014-12-23 | E-Vision Smart Optics, Inc. | Eyewear including a heads up display |
US7793361B2 (en) | 2004-11-12 | 2010-09-14 | Nike, Inc. | Article of apparel incorporating a separable electronic device |
US20060115130A1 (en) | 2004-11-29 | 2006-06-01 | Douglas Kozlay | Eyewear with biometrics to protect displayed data |
US7457434B2 (en) | 2005-04-04 | 2008-11-25 | Massachusetts Eye & Ear Infirmary | Adaptively focusing extra-ocular vision prostheses |
US7400257B2 (en) | 2005-04-06 | 2008-07-15 | Rivas Victor A | Vital signals and glucose monitoring personal wireless system |
EP1877981A4 (en) * | 2005-05-02 | 2009-12-16 | Univ Virginia | Systems, devices, and methods for interpreting movement |
US8979295B2 (en) | 2005-05-17 | 2015-03-17 | Michael Waters | Rechargeable lighted glasses |
US20070052672A1 (en) | 2005-09-08 | 2007-03-08 | Swisscom Mobile Ag | Communication device, system and method |
US20070112287A1 (en) | 2005-09-13 | 2007-05-17 | Fancourt Craig L | System and method for detecting deviations in nominal gait patterns |
US11428937B2 (en) | 2005-10-07 | 2022-08-30 | Percept Technologies | Enhanced optical and perceptual digital eyewear |
US8696113B2 (en) | 2005-10-07 | 2014-04-15 | Percept Technologies Inc. | Enhanced optical and perceptual digital eyewear |
US20070081123A1 (en) | 2005-10-07 | 2007-04-12 | Lewis Scott W | Digital eyewear |
US9658473B2 (en) | 2005-10-07 | 2017-05-23 | Percept Technologies Inc | Enhanced optical and perceptual digital eyewear |
US7648463B1 (en) | 2005-12-15 | 2010-01-19 | Impact Sports Technologies, Inc. | Monitoring device, method and system |
GB0602127D0 (en) | 2006-02-02 | 2006-03-15 | Imp Innovations Ltd | Gait analysis |
US8188868B2 (en) | 2006-04-20 | 2012-05-29 | Nike, Inc. | Systems for activating and/or authenticating electronic devices for operation with apparel |
US7558622B2 (en) | 2006-05-24 | 2009-07-07 | Bao Tran | Mesh network stroke monitoring appliance |
US7539533B2 (en) | 2006-05-16 | 2009-05-26 | Bao Tran | Mesh network monitoring appliance |
US7543934B2 (en) | 2006-09-20 | 2009-06-09 | Ipventures, Inc. | Eyeglasses with activity monitoring and acoustic dampening |
WO2008073806A1 (en) | 2006-12-08 | 2008-06-19 | Sabic Innovative Plastics Ip B.V. | Active transdermal drug delivery system |
US8157730B2 (en) * | 2006-12-19 | 2012-04-17 | Valencell, Inc. | Physiological and environmental monitoring systems and methods |
CA2678109C (en) | 2007-01-15 | 2020-05-26 | Deka Products Limited Partnership | Device and method for food management |
JP2008198028A (en) | 2007-02-14 | 2008-08-28 | Sony Corp | Wearable device, authentication method and program |
US7987070B2 (en) | 2007-04-23 | 2011-07-26 | Dp Technologies, Inc. | Eyewear having human activity monitoring device |
US8944590B2 (en) | 2010-07-02 | 2015-02-03 | Mitsui Chemicals, Inc. | Electronic spectacle frames |
US7634379B2 (en) | 2007-05-18 | 2009-12-15 | Ultimate Balance, Inc. | Newtonian physical activity monitor |
US8750971B2 (en) | 2007-05-24 | 2014-06-10 | Bao Tran | Wireless stroke monitoring |
WO2008149172A1 (en) | 2007-06-07 | 2008-12-11 | Panagiotis Pavlopoulos | An eyewear comprising at least one display device |
US9254100B2 (en) | 2007-09-12 | 2016-02-09 | Cardiac Pacemakers, Inc. | Logging daily average metabolic activity using a motion sensor |
CN101801284B (en) | 2007-09-21 | 2012-10-03 | Tyco医疗健康集团 | Surgical device |
US8448846B2 (en) | 2007-11-18 | 2013-05-28 | Intel-Ge Care Innovations Llc | Medication recording device |
US8202148B2 (en) | 2007-12-03 | 2012-06-19 | Julius Young | Machine and method for caddying and golf instruction |
US20090195747A1 (en) | 2008-02-04 | 2009-08-06 | Insua Luisa M | Interchangeable eyeglass temples |
US20090227853A1 (en) | 2008-03-03 | 2009-09-10 | Ravindra Wijesiriwardana | Wearable optical pulse plethysmography sensors or pulse oximetry sensors based wearable heart rate monitoring systems |
US20120142443A1 (en) | 2008-03-17 | 2012-06-07 | Chris Savarese | Golf club apparatuses and methods |
CN101566874A (en) | 2008-04-24 | 2009-10-28 | 鸿富锦精密工业(深圳)有限公司 | Control device and electronic equipment using same |
CA2729183C (en) | 2008-07-18 | 2016-04-26 | Optalert Pty Ltd | Alertness sensing device |
US8011242B2 (en) | 2008-07-29 | 2011-09-06 | Garmin Switzerland Gmbh | System and device for measuring and analyzing forces applied by a cyclist on a pedal of a bicycle |
US20100042430A1 (en) | 2008-08-12 | 2010-02-18 | Irody Inc | System and method for collecting and authenticating medication consumption |
US20100045928A1 (en) | 2008-08-25 | 2010-02-25 | Tri-Specs, Inc. | Fashion eyewear frame that houses circuitry to effect wireless audio communication while providing extraneous background noise cancellation capability |
US20100136508A1 (en) | 2008-10-23 | 2010-06-03 | Damir Zekhtser | Meal Plan Management |
CN103119512A (en) | 2008-11-02 | 2013-05-22 | 大卫·乔姆 | Near to eye display system and appliance |
US8494507B1 (en) | 2009-02-16 | 2013-07-23 | Handhold Adaptive, LLC | Adaptive, portable, multi-sensory aid for the disabled |
WO2010096691A2 (en) | 2009-02-20 | 2010-08-26 | The Regents Of The University Of Colorado, A Body Corporate | Footwear-based body weight monitor and postural allocation, physical activity classification, and energy expenditure calculator |
JP5789199B2 (en) | 2009-02-25 | 2015-10-07 | ヴァレンセル,インコーポレイテッド | Headset and earbud |
US20130024211A1 (en) | 2009-04-09 | 2013-01-24 | Access Mobility, Inc. | Active learning and advanced relationship marketing and health interventions |
US8081082B2 (en) | 2009-05-27 | 2011-12-20 | International Business Machines Corporation | Monitoring patterns of motion |
US20100308999A1 (en) | 2009-06-05 | 2010-12-09 | Chornenky Todd E | Security and monitoring apparatus |
US8253561B2 (en) | 2009-06-10 | 2012-08-28 | Betty L. Bowers | Medication management apparatus and system |
US20100332571A1 (en) | 2009-06-30 | 2010-12-30 | Jennifer Healey | Device augmented food identification |
US10748447B2 (en) | 2013-05-24 | 2020-08-18 | Lincoln Global, Inc. | Systems and methods providing a computerized eyewear device to aid in welding |
US20130009907A1 (en) | 2009-07-31 | 2013-01-10 | Rosenberg Ilya D | Magnetic Stylus |
EP2484281A4 (en) * | 2009-09-30 | 2015-05-06 | Mitsubishi Chem Corp | Body movement signal information processing method, information processing system and information processing device |
US8303311B2 (en) | 2009-09-30 | 2012-11-06 | Forest Carl A | Sport personal coach system |
JP5504810B2 (en) | 2009-10-06 | 2014-05-28 | オムロンヘルスケア株式会社 | Walking posture determination device, control program, and control method |
US8605165B2 (en) | 2010-10-06 | 2013-12-10 | Ai Cure Technologies Llc | Apparatus and method for assisting monitoring of medication adherence |
US8290558B1 (en) | 2009-11-23 | 2012-10-16 | Vioptix, Inc. | Tissue oximeter intraoperative sensor |
FR2953284A1 (en) | 2009-12-02 | 2011-06-03 | Movea Sa | SYSTEM AND METHOD FOR DRIVER ASSISTANCE OF BIOMECHANIC DRIVE VEHICLE COMPRISING AT LEAST ONE WHEEL |
IT1397737B1 (en) | 2010-01-18 | 2013-01-24 | Giovanni Saggio | EQUIPMENT AND METHOD OF DETECTION, TRAINING AND TRAINING OF MOTOR ACTIVITIES |
US9229227B2 (en) | 2010-02-28 | 2016-01-05 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a light transmissive wedge shaped illumination system |
CN102906623A (en) | 2010-02-28 | 2013-01-30 | 奥斯特豪特集团有限公司 | Local advertising content on an interactive head-mounted eyepiece |
US20120206485A1 (en) | 2010-02-28 | 2012-08-16 | Osterhout Group, Inc. | Ar glasses with event and sensor triggered user movement control of ar eyepiece facilities |
US8568313B2 (en) | 2010-03-12 | 2013-10-29 | Rajendra Padma Sadhu | User wearable portable communication device for collection and transmission of physiological data |
US9652965B2 (en) | 2010-03-12 | 2017-05-16 | Rajendra Padma Sadhu | System and method for transmitting alerts and notifications to a user |
US20110224505A1 (en) | 2010-03-12 | 2011-09-15 | Rajendra Padma Sadhu | User wearable portable communicative device |
US8690750B2 (en) | 2010-05-14 | 2014-04-08 | Wesley W. O. Krueger | System and method for measuring and minimizing the effects of vertigo, motion sickness, motion intolerance, and/or spatial disorientation |
US8531355B2 (en) | 2010-07-23 | 2013-09-10 | Gregory A. Maltz | Unitized, vision-controlled, wireless eyeglass transceiver |
US20120029367A1 (en) | 2010-07-31 | 2012-02-02 | Hobeika Hind Louis | Heart rate waterproof measuring apparatus |
US9247212B2 (en) | 2010-08-26 | 2016-01-26 | Blast Motion Inc. | Intelligent motion capture element |
US8594971B2 (en) | 2010-09-22 | 2013-11-26 | Invensense, Inc. | Deduced reckoning navigation without a constraint relationship between orientation of a sensor platform and a direction of travel of an object |
US10216893B2 (en) | 2010-09-30 | 2019-02-26 | Fitbit, Inc. | Multimode sensor devices |
US8762102B2 (en) | 2010-09-30 | 2014-06-24 | Fitbit, Inc. | Methods and systems for generation and rendering interactive events having combined activity and location information |
US8712724B2 (en) | 2010-09-30 | 2014-04-29 | Fitbit, Inc. | Calendar integration methods and systems for presentation of events having combined activity and location information |
US8615377B1 (en) | 2010-09-30 | 2013-12-24 | Fitbit, Inc. | Methods and systems for processing social interactive data and sharing of tracked activity associated with locations |
US9241635B2 (en) | 2010-09-30 | 2016-01-26 | Fitbit, Inc. | Portable monitoring devices for processing applications and processing analysis of physiological conditions of a user associated with the portable monitoring device |
US8738323B2 (en) | 2010-09-30 | 2014-05-27 | Fitbit, Inc. | Methods and systems for metrics analysis and interactive rendering, including events having combined activity and location information |
US8738321B2 (en) | 2010-09-30 | 2014-05-27 | Fitbit, Inc. | Methods and systems for classification of geographic locations for tracked activity |
US8849610B2 (en) | 2010-09-30 | 2014-09-30 | Fitbit, Inc. | Tracking user physical activity with multiple devices |
EP2439580A1 (en) | 2010-10-01 | 2012-04-11 | Ophtimalia | Data exchange system |
KR101346661B1 (en) | 2010-11-15 | 2014-02-06 | 부경대학교 산학협력단 | Cosmetic composition for preventing skin aging comprising chitooligosaccharides |
JP2012113627A (en) | 2010-11-26 | 2012-06-14 | Terumo Corp | Portable terminal, calorie estimation method, and calorie estimation program |
US9113793B2 (en) | 2010-12-10 | 2015-08-25 | Rohm Co., Ltd. | Pulse wave sensor |
US20120169990A1 (en) | 2011-01-05 | 2012-07-05 | Burnstein Tracey E | Electronic eyewear and footwear |
US20120191016A1 (en) * | 2011-01-25 | 2012-07-26 | Harris Corporation | Gait based notification and control of portable devices |
US20120203310A1 (en) | 2011-02-04 | 2012-08-09 | Pugh Randall B | Spectacles for light therapy |
US9785242B2 (en) | 2011-03-12 | 2017-10-10 | Uday Parshionikar | Multipurpose controllers and methods |
US9317660B2 (en) | 2011-03-31 | 2016-04-19 | Adidas Ag | Group performance monitoring system and method |
US8911087B2 (en) | 2011-05-20 | 2014-12-16 | Eyefluence, Inc. | Systems and methods for measuring reactions of head, eyes, eyelids and pupils |
US9256711B2 (en) | 2011-07-05 | 2016-02-09 | Saudi Arabian Oil Company | Systems, computer medium and computer-implemented methods for providing health information to employees via augmented reality display |
US8184067B1 (en) | 2011-07-20 | 2012-05-22 | Google Inc. | Nose bridge sensor |
US9342610B2 (en) | 2011-08-25 | 2016-05-17 | Microsoft Technology Licensing, Llc | Portals: registered objects as virtualized, personalized displays |
TWI455342B (en) | 2011-08-30 | 2014-10-01 | Nat Univ Tsing Hua | Solar cell with selective emitter structure and manufacturing method thereof |
KR20130025675A (en) | 2011-09-02 | 2013-03-12 | 삼성전자주식회사 | User health monitoring system which comprises 3d glasses and display apparatus, and display apparatus and control method thereof |
US8941560B2 (en) | 2011-09-21 | 2015-01-27 | Google Inc. | Wearable computer with superimposed controls and instructions for external device |
US20150057512A1 (en) | 2011-11-16 | 2015-02-26 | Rijuven Corporation | Wearable heart failure monitor patch |
US20130138413A1 (en) | 2011-11-24 | 2013-05-30 | Auckland Uniservices Limited | System and Method for Determining Motion |
US20130157232A1 (en) | 2011-12-09 | 2013-06-20 | Joel Ehrenkranz | System and methods for monitoring food consumption |
US8540583B2 (en) | 2011-12-30 | 2013-09-24 | Nike, Inc. | System for tracking a golf ball and displaying an enhanced image of the golf ball |
US9141194B1 (en) | 2012-01-04 | 2015-09-22 | Google Inc. | Magnetometer-based gesture sensing with a wearable device |
US9529197B2 (en) | 2012-03-21 | 2016-12-27 | Google Inc. | Wearable device with input and output structures |
US9737261B2 (en) | 2012-04-13 | 2017-08-22 | Adidas Ag | Wearable athletic activity monitoring systems |
US9504414B2 (en) | 2012-04-13 | 2016-11-29 | Adidas Ag | Wearable athletic activity monitoring methods and systems |
US20130307670A1 (en) | 2012-05-15 | 2013-11-21 | Jonathan E. Ramaci | Biometric authentication system |
US9001427B2 (en) | 2012-05-30 | 2015-04-07 | Microsoft Technology Licensing, Llc | Customized head-mounted display device |
US20130329183A1 (en) | 2012-06-11 | 2013-12-12 | Pixeloptics, Inc. | Adapter For Eyewear |
US9005129B2 (en) | 2012-06-22 | 2015-04-14 | Fitbit, Inc. | Wearable heart rate monitor |
US8948832B2 (en) | 2012-06-22 | 2015-02-03 | Fitbit, Inc. | Wearable heart rate monitor |
US9599632B2 (en) | 2012-06-22 | 2017-03-21 | Fitbit, Inc. | Fitness monitoring device with altimeter |
US9035970B2 (en) | 2012-06-29 | 2015-05-19 | Microsoft Technology Licensing, Llc | Constraint based information inference |
US9579048B2 (en) | 2012-07-30 | 2017-02-28 | Treefrog Developments, Inc | Activity monitoring system with haptic feedback |
AU2013296579A1 (en) | 2012-07-30 | 2015-02-26 | Treefrog Developments, Inc. | Athletic monitoring |
WO2014021602A2 (en) | 2012-07-31 | 2014-02-06 | 인텔렉추얼디스커버리 주식회사 | Wearable electronic device and method for controlling same |
US9720231B2 (en) | 2012-09-26 | 2017-08-01 | Dolby Laboratories Licensing Corporation | Display, imaging system and controller for eyewear display device |
JP6021582B2 (en) | 2012-10-24 | 2016-11-09 | オリンパス株式会社 | Glasses-type wearable device and front part of glasses-type wearable device |
US10143405B2 (en) | 2012-11-14 | 2018-12-04 | MAD Apparel, Inc. | Wearable performance monitoring, analysis, and feedback systems and methods |
ITMI20121957A1 (en) | 2012-11-16 | 2014-05-17 | Marco Carrara | GLASSES WITH HIGH FLEXIBILITY OF USE |
US10045718B2 (en) | 2012-11-22 | 2018-08-14 | Atheer, Inc. | Method and apparatus for user-transparent system control using bio-input |
US20140218281A1 (en) | 2012-12-06 | 2014-08-07 | Eyefluence, Inc. | Systems and methods for eye gaze determination |
GB2496064B (en) | 2012-12-31 | 2015-03-11 | Nicholas Jamie Marston | Video camera shooting glasses |
ITMI20130024A1 (en) | 2013-01-10 | 2014-07-11 | Marco Carrara | METHOD OF ACQUISITION AND TREATMENT OF HEART RATE DATA |
US9520638B2 (en) | 2013-01-15 | 2016-12-13 | Fitbit, Inc. | Hybrid radio frequency / inductive loop antenna |
US20140204334A1 (en) | 2013-01-18 | 2014-07-24 | William Anthony Stoll | Bio-sensors in eyeglasses |
US9370302B2 (en) | 2014-07-08 | 2016-06-21 | Wesley W. O. Krueger | System and method for the measurement of vestibulo-ocular reflex to improve human performance in an occupational environment |
US9848776B2 (en) | 2013-03-04 | 2017-12-26 | Hello Inc. | Methods using activity manager for monitoring user activity |
US9500464B2 (en) | 2013-03-12 | 2016-11-22 | Adidas Ag | Methods of determining performance information for individuals and sports objects |
ES2900423T3 (en) | 2013-03-15 | 2022-03-16 | Percept Tech Inc | Enhanced optical and perceptual digital glasses |
US9851803B2 (en) | 2013-03-15 | 2017-12-26 | Eyecam, LLC | Autonomous computing and telecommunications head-up displays glasses |
WO2014144918A2 (en) | 2013-03-15 | 2014-09-18 | Percept Technologies, Inc. | Enhanced optical and perceptual digital eyewear |
US20140276096A1 (en) * | 2013-03-15 | 2014-09-18 | Bonutti Research, Inc. | Systems and methods for use in diagnosing a medical condition of a patient |
US9341526B2 (en) | 2013-04-01 | 2016-05-17 | Saris Cycling Group, Inc. | System for speed-based power calculation |
DE102013207064A1 (en) | 2013-04-19 | 2014-10-23 | Bayerische Motoren Werke Aktiengesellschaft | Method for selecting an information source for display on data glasses |
US20140324459A1 (en) | 2013-04-30 | 2014-10-30 | Hti Ip, L.L.C | Automatic health monitoring alerts |
US10930174B2 (en) | 2013-05-24 | 2021-02-23 | Lincoln Global, Inc. | Systems and methods providing a computerized eyewear device to aid in welding |
US20140375470A1 (en) | 2013-06-20 | 2014-12-25 | Chester Charles Malveaux | Wearable networked and standalone biometric sensor system to record and transmit biometric data for multiple applications |
US10512407B2 (en) | 2013-06-24 | 2019-12-24 | Fitbit, Inc. | Heart rate data collection |
ES2530421B1 (en) | 2013-07-30 | 2015-07-09 | Ion Eyewear, S.L. | ACCESSORY GLASSES FOR MOBILE DEVICES AND PC¿S |
FR3009270B1 (en) | 2013-07-31 | 2016-09-09 | Michelin & Cie | DEVICE AND METHOD FOR CONTROLLING THE POWER OF ASSISTANCE OF A POWER-ASSISTED VELO |
EP2833196B1 (en) | 2013-08-02 | 2016-03-16 | ESSILOR INTERNATIONAL (Compagnie Générale d'Optique) | A method of controlling a programmable ophthalmic lens device |
US9704412B2 (en) | 2013-08-26 | 2017-07-11 | John Andrew Wells | Biometric data gathering |
US20150065889A1 (en) | 2013-09-02 | 2015-03-05 | Life Beam Technologies Ltd. | Bodily worn multiple optical sensors heart rate measuring device and method |
US20150148632A1 (en) | 2013-11-26 | 2015-05-28 | David Alan Benaron | Calorie Monitoring Sensor And Method For Cell Phones, Smart Watches, Occupancy Sensors, And Wearables |
TWI548438B (en) | 2013-12-20 | 2016-09-11 | 岱宇國際股份有限公司 | Exercise device providing symmetry index |
US9595181B2 (en) | 2013-12-20 | 2017-03-14 | Invensense, Inc. | Wearable device assisting smart media application and vice versa |
WO2015099725A1 (en) | 2013-12-26 | 2015-07-02 | Intel Corporation | Secure transactions using a personal device |
WO2015095924A1 (en) | 2013-12-27 | 2015-07-02 | Koonung Heights Pty Ltd | A biofeedback, stress management and cognitive enhancement system |
US11298064B1 (en) | 2014-02-18 | 2022-04-12 | Orbital Research Inc. | Head-mounted physiological signal monitoring system, devices and methods |
EP3089658A4 (en) | 2014-02-24 | 2017-10-11 | Sony Corporation | Smart wearable devices and methods for acquisition of sensorial information from wearable devices to activate functions in other devices |
AU2015218578B2 (en) | 2014-02-24 | 2020-06-25 | Nedim T. Sahin | Systems, environment and methods for evaluation and management of autism spectrum disorder using a wearable data collection device |
US9031812B2 (en) * | 2014-02-27 | 2015-05-12 | Fitbit, Inc. | Notifications on a user device based on activity detected by an activity monitoring device |
FR3019346B1 (en) | 2014-03-31 | 2016-07-08 | Withings | METHOD FOR COUNTING STEPS PERFORMED BY A USER |
CN105893721A (en) | 2014-05-13 | 2016-08-24 | 陈威宇 | Adaptive skin care information prompt system and adaptive skin care prompt method |
US20150332149A1 (en) | 2014-05-15 | 2015-11-19 | Red Lozenge, Inc. | Tracking behavior and goal achievement |
US10478127B2 (en) | 2014-06-23 | 2019-11-19 | Sherlock Solutions, LLC | Apparatuses, methods, processes, and systems related to significant detrimental changes in health parameters and activating lifesaving measures |
TWI530276B (en) | 2014-07-08 | 2016-04-21 | 原相科技股份有限公司 | Biometric detection module with denoising function and biometric detection method thereof |
WO2016017997A1 (en) | 2014-07-31 | 2016-02-04 | Samsung Electronics Co., Ltd. | Wearable glasses and method of providing content using the same |
US20160041404A1 (en) | 2014-08-08 | 2016-02-11 | Marchon Eyewear, Inc. | Eyewear with interchangeable temples and brow bar |
HK1203120A2 (en) | 2014-08-26 | 2015-10-16 | 高平 | A gait monitor and a method of monitoring the gait of a person |
US20160066848A1 (en) | 2014-09-05 | 2016-03-10 | Vision Service Plan | Wearable environmental pollution monitor computer apparatus, systems, and related methods |
US10448867B2 (en) | 2014-09-05 | 2019-10-22 | Vision Service Plan | Wearable gait monitoring apparatus, systems, and related methods |
US10617342B2 (en) | 2014-09-05 | 2020-04-14 | Vision Service Plan | Systems, apparatus, and methods for using a wearable device to monitor operator alertness |
US20160117937A1 (en) | 2014-10-27 | 2016-04-28 | Bloom Technologies NV | System and method for providing biometric and context based messaging |
US9566033B2 (en) | 2014-11-03 | 2017-02-14 | Phillip Bogdanovich | Garment system with electronic components and associated methods |
KR102313220B1 (en) | 2015-01-09 | 2021-10-15 | 삼성전자주식회사 | Wearable device and method for controlling thereof |
US10548491B2 (en) | 2015-01-30 | 2020-02-04 | Koninklijke Philips N.V. | Photoplethysmography apparatus |
US20160223577A1 (en) | 2015-01-30 | 2016-08-04 | Vision Service Plan | Systems and methods for tracking motion of a bicycle or other vehicles |
US10349887B1 (en) | 2015-06-14 | 2019-07-16 | Facense Ltd. | Blood pressure measuring smartglasses |
US10398328B2 (en) | 2015-08-25 | 2019-09-03 | Koninklijke Philips N.V. | Device and system for monitoring of pulse-related information of a subject |
US9726904B1 (en) | 2015-09-29 | 2017-08-08 | Snap Inc. | Eyewear with conductive temple joint |
FR3043245B1 (en) | 2015-11-03 | 2017-10-27 | Stmicroelectronics Rousset | METHOD FOR READING AN EEPROM MEMORY AND CORRESPONDING DEVICE |
US9610476B1 (en) | 2016-05-02 | 2017-04-04 | Bao Tran | Smart sport device |
US20170255029A1 (en) | 2016-03-03 | 2017-09-07 | Vision Service Plan | Systems and methods for charging eyewear |
US20180064399A1 (en) | 2016-09-07 | 2018-03-08 | Heptagon Micro Optics Pte. Ltd. | Imaging systems including multi-tap demodulation pixels for biometric measurements |
EP3919969A1 (en) | 2016-09-22 | 2021-12-08 | Essilor International | Health monitoring device and wearing detection module for spectacles frame |
US20180206735A1 (en) | 2017-01-26 | 2018-07-26 | Microsoft Technology Licensing, Llc | Head-mounted device for capturing pulse data |
US10874305B2 (en) | 2018-01-15 | 2020-12-29 | Microsoft Technology Licensing, Llc | Sensor device |
-
2014
- 2014-11-21 US US14/550,406 patent/US10448867B2/en active Active
- 2014-12-05 US US14/562,454 patent/US9795324B2/en active Active
- 2014-12-19 US US14/578,039 patent/US20160066829A1/en not_active Abandoned
- 2014-12-31 US US14/588,122 patent/US9649052B2/en active Active
-
2015
- 2015-01-30 US US14/610,439 patent/US10307085B2/en active Active
- 2015-01-30 US US14/610,589 patent/US20160071423A1/en not_active Abandoned
- 2015-01-30 US US14/610,628 patent/US20160070122A1/en not_active Abandoned
- 2015-09-04 EP EP15838411.5A patent/EP3148435B1/en active Active
- 2015-09-04 EP EP15838234.1A patent/EP3189367B1/en active Active
- 2015-09-04 WO PCT/US2015/048656 patent/WO2016037117A1/en active Application Filing
- 2015-09-04 CA CA2953856A patent/CA2953856A1/en not_active Abandoned
- 2015-09-04 EP EP15839047.6A patent/EP3189371A4/en not_active Withdrawn
- 2015-09-04 WO PCT/US2015/048662 patent/WO2016037120A1/en active Application Filing
- 2015-09-04 WO PCT/US2015/048612 patent/WO2016037091A1/en active Application Filing
- 2015-09-04 CA CA2960425A patent/CA2960425A1/en not_active Abandoned
- 2015-09-04 CA CA2960429A patent/CA2960429A1/en not_active Abandoned
-
2017
- 2017-05-15 US US15/594,898 patent/US10188323B2/en active Active
- 2017-10-23 US US15/791,196 patent/US20180042523A1/en not_active Abandoned
-
2019
- 2019-01-28 US US16/259,646 patent/US10542915B2/en active Active
- 2019-06-03 US US16/429,480 patent/US10694981B2/en active Active
- 2019-10-18 US US16/657,982 patent/US20200046260A1/en not_active Abandoned
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030195398A1 (en) * | 2000-05-31 | 2003-10-16 | Kabushiki Kaisha Toshiba | Life support apparatus and method and method for providing advertisement information |
US20070273611A1 (en) * | 2004-04-01 | 2007-11-29 | Torch William C | Biosensors, communicators, and controllers monitoring eye movement and methods for using them |
US20100280336A1 (en) * | 2009-04-30 | 2010-11-04 | Medtronic, Inc. | Anxiety disorder monitoring |
US20110142413A1 (en) * | 2009-12-04 | 2011-06-16 | Lg Electronics Inc. | Digital data reproducing apparatus and method for controlling the same |
US20110169932A1 (en) * | 2010-01-06 | 2011-07-14 | Clear View Technologies Inc. | Wireless Facial Recognition |
US20130127980A1 (en) * | 2010-02-28 | 2013-05-23 | Osterhout Group, Inc. | Video display modification based on sensor input for a see-through near-to-eye display |
US20130096397A1 (en) * | 2010-06-17 | 2013-04-18 | Nec Corporation | Sensitivity evaluation system, sensitivity evaluation method, and program |
US20120290401A1 (en) * | 2011-05-11 | 2012-11-15 | Google Inc. | Gaze tracking system |
US20130317384A1 (en) * | 2012-05-25 | 2013-11-28 | Emotiv Lifesciences Inc. | System and Method for Instructing a Behavior Change in a User |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10387173B1 (en) | 2015-03-27 | 2019-08-20 | Intuit Inc. | Method and system for using emotional state data to tailor the user experience of an interactive software system |
US9930102B1 (en) | 2015-03-27 | 2018-03-27 | Intuit Inc. | Method and system for using emotional state data to tailor the user experience of an interactive software system |
US10169827B1 (en) | 2015-03-27 | 2019-01-01 | Intuit Inc. | Method and system for adapting a user experience provided through an interactive software system to the content being delivered and the predicted emotional impact on the user of that content |
US20160350138A1 (en) * | 2015-05-31 | 2016-12-01 | Roya Caroline SALAS | Biofeedback system |
US10332122B1 (en) | 2015-07-27 | 2019-06-25 | Intuit Inc. | Obtaining and analyzing user physiological data to determine whether a user would benefit from user support |
CN106377271A (en) * | 2016-10-20 | 2017-02-08 | 中国矿业大学 | Wearable method and wearable device for monitoring and regulating physiology and psychology |
US20180241973A1 (en) * | 2017-02-21 | 2018-08-23 | Janet Newell | Video and audio recording system and method |
WO2018156992A1 (en) * | 2017-02-23 | 2018-08-30 | Miller Charles Robert Iii | Device and system for user context-cortical sensing and determination |
US20210141453A1 (en) * | 2017-02-23 | 2021-05-13 | Charles Robert Miller, III | Wearable user mental and contextual sensing device and system |
CN110447232A (en) * | 2017-03-31 | 2019-11-12 | 三星电子株式会社 | For determining the electronic equipment and its control method of user emotion |
US20190167226A1 (en) * | 2017-12-04 | 2019-06-06 | International Business Machines Corporation | Infant gastrointestinal monitor |
CN108401129A (en) * | 2018-03-22 | 2018-08-14 | 广东小天才科技有限公司 | Video call method, device, terminal based on Wearable and storage medium |
US11331003B2 (en) | 2018-03-27 | 2022-05-17 | Samsung Electronics Co., Ltd. | Context-aware respiration rate determination using an electronic device |
Also Published As
Publication number | Publication date |
---|---|
US10694981B2 (en) | 2020-06-30 |
US20160070121A1 (en) | 2016-03-10 |
EP3189367A4 (en) | 2018-05-30 |
US20160071390A1 (en) | 2016-03-10 |
US20190298228A1 (en) | 2019-10-03 |
EP3189371A1 (en) | 2017-07-12 |
US9649052B2 (en) | 2017-05-16 |
WO2016037120A1 (en) | 2016-03-10 |
US10307085B2 (en) | 2019-06-04 |
CA2960429A1 (en) | 2016-03-10 |
US20160071423A1 (en) | 2016-03-10 |
US20180042523A1 (en) | 2018-02-15 |
US9795324B2 (en) | 2017-10-24 |
CA2960425A1 (en) | 2016-03-10 |
WO2016037117A1 (en) | 2016-03-10 |
US20160066820A1 (en) | 2016-03-10 |
CA2953856A1 (en) | 2016-03-10 |
US20170245757A1 (en) | 2017-08-31 |
EP3148435B1 (en) | 2023-10-18 |
EP3189367B1 (en) | 2024-04-10 |
EP3148435A4 (en) | 2018-01-17 |
US10448867B2 (en) | 2019-10-22 |
US20190159700A1 (en) | 2019-05-30 |
EP3189367A1 (en) | 2017-07-12 |
EP3148435A1 (en) | 2017-04-05 |
US10188323B2 (en) | 2019-01-29 |
US20160066847A1 (en) | 2016-03-10 |
US10542915B2 (en) | 2020-01-28 |
US20160070122A1 (en) | 2016-03-10 |
WO2016037091A1 (en) | 2016-03-10 |
US20200046260A1 (en) | 2020-02-13 |
EP3189371A4 (en) | 2018-05-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20160066829A1 (en) | Wearable mental state monitor computer apparatus, systems, and related methods | |
US20230297163A1 (en) | Monitoring a user of a head-wearable electronic device | |
US10617342B2 (en) | Systems, apparatus, and methods for using a wearable device to monitor operator alertness | |
US11123562B1 (en) | Pain quantification and management system and device, and method of using | |
US11064892B2 (en) | Detecting a transient ischemic attack using photoplethysmogram signals | |
US10136856B2 (en) | Wearable respiration measurements system | |
US10791938B2 (en) | Smartglasses for detecting congestive heart failure | |
US11341167B2 (en) | Data tagging | |
US10638938B1 (en) | Eyeglasses to detect abnormal medical events including stroke and migraine | |
US10524696B2 (en) | Virtual coaching based on respiration signals | |
DE102016118773A1 (en) | Systems for the acquisition of thermal measurements of the face | |
US20140121540A1 (en) | System and method for monitoring the health of a user | |
US20180103851A1 (en) | Detecting physiological responses based on thermal asymmetry of the face | |
US20220370757A1 (en) | Personalized sleep wellness score for treatment and/or evaluation of sleep conditions | |
US20220238220A1 (en) | Headset integrated into healthcare platform | |
US10085685B2 (en) | Selecting triggers of an allergic reaction based on nasal temperatures | |
US10136852B2 (en) | Detecting an allergic reaction from nasal temperatures | |
TWI670095B (en) | Smart sleep assistant system and method thereof | |
Ueda et al. | A Study on the Improvement for a Student’s Sleep-Phase with Wearable Devices | |
Chen et al. | EudaeSense | |
Cernisov et al. | A Doctoral Dissertation submitted to Keio University Graduate School of Media Design |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: VISION SERVICE PLAN, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SALES, JAY WILLIAM;KLOSINSKI, RICHARD CHESTER;WORKMAN, MATTHEW ALLEN;AND OTHERS;SIGNING DATES FROM 20141202 TO 20141204;REEL/FRAME:036288/0612 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |