Latest advances in capturing data on brain activity and eye movement are being combined to open up a host of ‘mindreading’ possibilities for the future.
These include the potential development of a system that can detect when drivers are in danger of falling asleep at the wheel.
The research has been undertaken at the University of Leicester with funding from the Engineering and Physical Sciences Research Council (EPSRC), and in collaboration with the University of Buenos Aires in Argentina.
The breakthrough involves bringing two recent developments in the world of technology together: high-speed eye tracking that records eye movements in unprecedented detail using cutting-edge infra-red cameras*; and high-density electroencephalograph** (EEG) technology that measures electrical brain activity with millisecond precision through electrodes placed on the scalp.
The research has overcome previous technological challenges which made it difficult to monitor eye movement and brain activity simultaneously. The team has done this by developing novel signal processing techniques.
This could be the first step towards a system that combines brain and eye monitoring to automatically alert drivers who are showing signs of drowsiness. The system would be built into the vehicle and connected unobtrusively to the driver, with the EEG looking out for brain signals that only occur in the early stages of sleepiness. The eye tracker would reinforce this by looking for erratic gaze patterns symptomatic of someone starting to feel drowsy and different from those characteristic of someone driving who is constantly looking out for hazards. Fatigue has been estimated to account for around 20 per cent of traffic accidents on the UK’s motorways.***
The breakthrough achieved by the University of Leicester could also ultimately be built on to deliver many other everyday applications in the years ahead. For example:
- Computer games of the future could dispense with the need for the player to physically interact with any type of console, mouse or other hand-operated system. Instead, eye movement and brain activity data would be collected and processed to indicate what action the player wants to take. By distinguishing the tiny differences in various types of brain activity, the EEG would identify the precise action the player desires (e.g. run, jump or throw), while the eye movement data would show exactly where on the screen the player was looking when they had this thought. This information could be combined to enable the correct action to occur. An unobtrusive headset would be all that would be required to capture the necessary data.
- People who have no arm functionality could move their wheelchairs simply through their eye movements. These movements could be tracked and the corresponding brain activity analysed to identify when these indicate a desire to move in a certain direction. This would then automatically activate a steering and propulsion mechanism that would drive the wheelchair to that place.
The Latest Streaming News: Brain and Eye Combined Monitoring updated minute-by-minute
Bookmark this page and come back often