New technique could cut sensors required for terahertz imaging

Tuesday, 06 May, 2014

A new technique developed by MIT researchers is said to reduce the number of sensors required for terahertz or millimetre-wave imaging by a factor of 10, or even 100. The technique could also have implications for the design of new, high-resolution radar and sonar systems.

Like sonar or radar, terahertz imaging produces an image by comparing measurements across an array of sensors. Those arrays have to be very dense, since the distance between sensors is proportional to wavelength. Terahertz imaging, which is already familiar from airport security checkpoints, has a number of other promising applications - from explosives detection to collision avoidance in cars.

In the latest issue of IEEE Transactions on Antennas and Propagation, researchers in MIT’s Research Laboratory for Electronics describe the technique. In a digital camera, the lens focuses the incoming light so that light reflected by a small patch of the visual scene strikes a correspondingly small patch of the sensor array. In lower-frequency imaging systems, by contrast, an incoming wave - whether electromagnetic or, in the case of sonar, acoustic - strikes all of the sensors in the array. The system determines the origin and intensity of the wave by comparing its phase - the alignment of its troughs and crests - when it arrives at each of the sensors.

As long as the distance between sensors is no more than half the wavelength of the incoming wave, that calculation is fairly straightforward, a matter of inverting the sensors’ measurements. But if the sensors are spaced farther than half a wavelength apart, the inversion will yield more than one possible solution. Those solutions will be spaced at regular angles around the sensor array, a phenomenon known as ‘spatial aliasing’.

In most applications of lower-frequency imaging, however, any given circumference around the detector is usually sparsely populated. That’s the phenomenon that the new system exploits.

“Think about a range around you, like five feet,” says Gregory Wornell, the Sumitomo Electric Industries Professor in Engineering in MIT’s Department of Electrical Engineering and Computer Science and a co-author on the new paper. “There’s actually not that much at five feet around you. Or at 10 feet. Different parts of the scene are occupied at those different ranges, but at any given range, it’s pretty sparse. Roughly speaking, the theory goes like this: if, say, 10% of the scene at a given range is occupied with objects, then you need only 10% of the full array to still be able to achieve full resolution.”

The trick is to determine which 10% of the array to keep. Keeping every tenth sensor won’t work: it’s the regularity of the distances between sensors that leads to aliasing. Arbitrarily varying the distances between sensors would solve that problem, but it would also make inverting the sensors’ measurements - calculating the wave’s source and intensity - prohibitively complicated.

So Wornell and his co-authors - James Krieger, a former student of Wornell’s who is now at MIT’s Lincoln Laboratory, and Yuval Kochman, a former postdoc who is now an assistant professor at the Hebrew University of Jerusalem - instead prescribe a detector along which the sensors are distributed in pairs. The regular spacing between pairs of sensors ensures that the scene reconstruction can be calculated efficiently, but the distance from each sensor to the next remains irregular.

The researchers also developed an algorithm that determines the optimal pattern for the sensors’ distribution. In essence, the algorithm maximises the number of different distances between arbitrary pairs of sensors.

With his new colleagues at Lincoln Lab, Krieger has performed experiments at radar frequencies using a one-dimensional array of sensors deployed in a parking lot, which verified the predictions of the theory. Moreover, Wornell’s description of the sparsity assumptions of the theory - 10% occupation at a given distance means one tenth the sensors - applies to one-dimensional arrays. Many applications - such as submarines’ sonar systems - instead use two-dimensional arrays, and in that case, the savings compound: one tenth the sensors in each of two dimensions translates to one hundredth the sensors in the complete array

Related News

'Frozen smoke' sensors detect indoor pollutants

Researchers have developed a sensor made from 'frozen smoke' that uses artificial...

New sweat-resistant wearable robot sensor

Researchers have developed an electromyography (EMG) sensor technology that is unaffected by the...

Innovative robotic tech empowers pipe inspection

Researchers have demonstrated a novel method for inspecting large pipe structures using mobile...


  • All content Copyright © 2024 Westwick-Farrow Pty Ltd