A new paper claims that APOE ε4 carriers, a group previously described as having poorer long-term memory, actually have better visual working memory. On Twitter, Prof Dorothy Bishop raised the issue of multiple comparisons, or rather a lack thereof… Should the paper be #cancelled?!
A clash between two major figures in cognitive neuroscience came to a head yesterday, and it got a bit ugly. The dispute centers around a publicly posted peer review of a publicly posted manuscript. Although seemingly aligned with Open Science ideals, the public review prompted worries about ulterior motives and power dynamics in some researchers. In addition to being juicy drama, the events reveal that Open Science requires trust, tact, and integrity. This post summarises the things I’ve learned.
Testing children is less easy than testing adults, primarily because they lack the social inhibition to tell psychological researchers to go away with their super boring tests. This presents a problem in developmental research: How do you reach these kids?! We developed a bunch of iPad games to test the cognition of an entire classroom in one go. And it works!
In cognitive neuroscience, we’re interested in what guides human attention. We distinguish between influences from high-level cognition (e.g. current goals), and low-level visual features. There are highly sophisticated models of how visual features such as intensity, colour, and movement guide human attention. Computerised implementations of these models allow computers to mimic human eye movements. Turns out Taylor Swift’s amazing videos are an excellent example!
Tobii is a major player in the eye-tracking world, selling devices to customers in business and science. Today, Tobii has made a major step towards supporting open science by adding support for its new SDK in PyGaze (and by extension in OpenSesame). You can review the code on GitHub, download the package, or download a full WinPython version. Details below, and more info on downloading here.
Gazepoint is a relatively small player on the eye-tracking market. They sell two devices: the 60 Hz GP3 at a price of $695, and the 150 Hz GP3 HD at $1995 (both of those prices exclude VAT and shipping). Because of its relatively low price, the basic GP3 is an appealing model for researchers on a budget. As of today, PyGaze supports Gazepoint’s trackers through their OpenGaze API. Download the new code from GitHub, and have fun!
That is one confusing title! The point is this: When light reaches your eyes, you’re not immediately aware of that. It takes some time for your visual system to process the light, and to translate it into something the rest of your brain can work with. When that’s done, you consciously ‘see’. In a new paper, we show that the process of becoming aware of what you see, is affected by how large an object is. With an oversimplified example: If light bounces of a puppy, into your eyes, it takes a fraction of a second for you to become aware of the puppy. And it takes a fraction of a second longer if it’s a fat puppy.
This morning, the EyeTribe announced via an email to their customers that they would stop the development of their products. The particular reason is rather vague (“we’ve decided to go in a different direction with our technology“), and researchers across the board are not happy. The EyeTribe was the only real option for cheap eye tracking: It was great for demonstrations, for pupillometry and fixation control, it had a very elegant API, and the hardware was great for how much you paid for it. Best of all: It didn’t come with the restrictive licenses that almost all of the EyeTribe’s competitors use to milk their customers for more money. I, for one, am sad about the loss of this great company.