Posts belonging to Category optics



Graphene, Not Glass, Is The Key To Better Optics

A lens just a billionth of a metre thick could transform phone cameras. Researchers at Swinburne University in Melbourne, Australia, have created ultra-thin lenses that cap an optical fibre, and can produce images with the quality and sharpness of much larger glass lenses.

Compared with current lenses, our graphene lens only needs one film to achieve the same resolution,” says Professor Baohua Jia, a research leader at Swinburne’s Centre for Micro-Photonics. “In the future, mobile phones could be much thinner, without having to sacrifice the quality of their cameras. Our lens also allows infrared light to pass through, which glass lenses don’t.”

Producing graphene can be costly and challenging, so Baohua and her colleagues used a laser to pattern layers of graphene oxide (graphene combined with oxygen). By then removing the oxygen, they produced low-cost, patterned films of graphene, a thousand times thinner than a human hair. “By patterning the graphene oxide film in this way, its optical and electrical properties can be altered, which allowed us to place them in different devices,” she says.

Warm objects give off infrared light, so mobile phones with graphene lenses could be used to scan for hotspots in the human body and help in the early identification of diseases like breast cancer. By attaching the lens to a fibre optic tip, endoscopes — instruments that are currently several millimetres wide—could be made a million times smaller. The team is also investigating graphene’s amazing properties for their potential use as supercapacitors, capable of storing very large amounts of energy, which could replace conventional batteries.

Baohua’s work on graphene lenses was published in Nature Communications.

Source: https://cosmosmagazine.com/

Computer Reads Body Language

Researchers at Carnegie Mellon University‘s Robotics Institute have enabled a computer to understand body poses and movements of multiple people from video in real time — including, for the first time, the pose of each individual’s hands and fingers. This new method was developed with the help of the Panoptic Studio — a two-story dome embedded with 500 video cameras — and the insights gained from experiments in that facility now make it possible to detect the pose of a group of people using a single camera and a laptop computer.

Yaser Sheikh, associate professor of robotics, said these methods for tracking 2-D human form and motion open up new ways for people and machines to interact with each other and for people to use machines to better understand the world around them. The ability to recognize hand poses, for instance, will make it possible for people to interact with computers in new and more natural ways, such as communicating with computers simply by pointing at things.

Detecting the nuances of nonverbal communication between individuals will allow robots to serve in social spaces, allowing robots to perceive what people around them are doing, what moods they are in and whether they can be interrupted. A self-driving car could get an early warning that a pedestrian is about to step into the street by monitoring body language. Enabling machines to understand human behavior also could enable new approaches to behavioral diagnosis and rehabilitation, for conditions such as autism, dyslexia and depression.

CLICK ON THE IMAGE TO ENJOY THE VIDEO

We communicate almost as much with the movement of our bodies as we do with our voice,” Sheikh said. “But computers are more or less blind to it.”

In sports analytics, real-time pose detection will make it possible for computers to track not only the position of each player on the field of play, as is now the case, but to know what players are doing with their arms, legs and heads at each point in time. The methods can be used for live events or applied to existing videos.

To encourage more research and applications, the researchers have released their computer code for both multi-person and hand pose estimation. It is being widely used by research groups, and more than 20 commercial groups, including automotive companies, have expressed interest in licensing the technology, Sheikh said.

Sheikh and his colleagues have presented reports on their multi-person and hand pose detection methods at CVPR 2017, the Computer Vision and Pattern Recognition Conference  in Honolulu.

Source: https://www.cmu.edu/

Optical Computer

Researchers at the University of Sydney (Australia) have dramatically slowed digital information carried as light waves by transferring the data into sound waves in an integrated circuit, or microchipTransferring information from the optical to acoustic domain and back again inside a chip is critical for the development of photonic integrated circuits: microchips that use light instead of electrons to manage data.

These chips are being developed for use in telecommunications, optical fibre networks and cloud computing data centers where traditional electronic devices are susceptible to electromagnetic interference, produce too much heat or use too much energy.

The information in our chip in acoustic form travels at a velocity five orders of magnitude slower than in the optical domain,” said Dr Birgit Stiller, research fellow at the University of Sydney and supervisor of the project.

It is like the difference between thunder and lightning,” she said.

This delay allows for the data to be briefly stored and managed inside the chip for processing, retrieval and further transmission as light wavesLight is an excellent carrier of information and is useful for taking data over long distances between continents through fibre-optic cables.

But this speed advantage can become a nuisance when information is being processed in computers and telecommunication systems.

Source: https://sydney.universty.au/

AR Smart Glasses, Next Frontier Of FaceBook

Facebook is hard at work on the technical breakthroughs needed to ship futuristic smart glasses that can let you see virtual objects in the real world. A patent application for a “waveguide display with two-dimensional scanner” was published on Thursday by three members from the advanced research division of Facebook’s virtual-reality subsidiary, Oculus.

The smart glasses being developed by Oculus will use a waveguide display to project light onto the wearer’s eyes instead of a more traditional display. The smart glasses would be able to display images, video, and work with connected speakers or headphones to play audio when worn.The display “may augment views of a physical, real-world environment with computer-generated elements” and “may be included in an eye-wear comprising a frame and a display assembly that presents media to a user’s eyes,” according to the filing.

By using waveguide technology, Facebook is taking a similar approach to Microsoft‘s HoloLens AR headset and the mysterious glasses being developed by the Google-backed startup Magic Leap.

One of the authors of the patent is, in fact, lead Oculus optical scientist Pasi Saarikko, who joined Facebook in 2015 after leading the optical design of the HoloLens at Microsoft.

While work is clearly being done on the underlying technology for Facebook‘s smart glasses now, don’t expect to see the device anytime soon. Michael Abrash, the chief scientist of Oculus, recently said that AR glasses won’t start replacing smartphones until as early as 2022.

Facebook CEO Mark Zuckerberg has called virtual and augmented reality the next major computing platform capable of replacing smartphones and traditional PCs. Facebook purchased Oculus for $2 billion in 2014 and plans to spend billions more on developing the technology.

Source: http://pdfaiw.uspto.gov/
A
ND
http://www.businessinsider.com

Use The Phone And See 3D Content Without 3D Glasses

RED, the company known for making some truly outstanding high-end cinema cameras, is set to release a smartphone in Q1 of 2018 called the HYDROGEN ONE. RED says that it is a standalone, unlocked and fully-featured smartphone “operating on Android OS that just happens to add a few additional features that shatter the mold of conventional thinking.” Yes, you read that right. This phone will blow your mind, or something – and it will even make phone calls.

In a press release riddled with buzzwords broken up by linking verbs, RED praises their yet-to-be smartphone with some serious adjectives. If we were just shown this press release outside of living on RED‘s actual server, we would swear it was satire. Here are a smattering of phrases found in the release.

Incredible retina-riveting display
Nanotechnology
Holographic multi-view content
RED Hydrogen 4-View content
Assault your senses
Proprietary H3O algorithm
Multi-dimentional audio

  • There are two models of the phone, which run at different prices. The Aluminum model will cost $1,195, but anyone worth their salt is going to go for the $1,595 Titanium version. Gotta shed that extra weight, you know?

Those are snippets from just the first three sections, of which there are nine. I get hyping a product, but this reads like a catalog seen in the background of a science-fiction comedy, meant to sound ridiculous – especially in the context of a ficticious universe.

Except that this is real life.

After spending a few minutes removing all the glitter words from this release, it looks like it will be a phone using a display similar to what you get with the Nintendo 3DS, or what The Verge points out as perhaps better than the flopped Amazon Fire Phone. Essentially, you should be able to use the phone and see 3D content without 3D glasses. Nintendo has already proven that can work, however it can really tire out your eyes. As an owner of three different Nintendo 3DS consoles, I can say that I rarely use the 3D feature because of how it makes my eyes hurt. It’s an odd sensation. It is probalby why Nintendo has released a new handheld that has the same power as the 3DS, but dropping the 3D feature altogether.

Anyway, back to the HYDROGEN ONE, RED says that it will work in tandem with their cameras as a user interface and monitor. It will also display what RED is calling “holographic content,” which isn’t well-described by RED in this release. We can assume it is some sort of mixed-dimensional view that makes certain parts of a video or image stand out over the others.

Source: http://www.red.com/
AND
http://www.imaging-resource.com/

Super-material Bends, Shapes And Focuses Sound Waves

These tiny 3D-printed bricks could one day allow people to create their own acoustics. That’s the plan of scientists from the universities of Bristol and Sussex. They’ve invented a metamaterial which bends and manipulates sound in any way the user wants. It’s helped scientists create what they call a ‘sonic alphabet‘.

CLICK ON THE IMAGE TO ENJOY THE VIDEO

We have discovered that you just need 16 bricks to make any type of sound that you can imagine. You can shape the sound just with 16 of them, just like you create any words with just 26 letters,” says Dr. Gianluca Memoli, researcher at Interact Lab at University of Sussex.

DIY kits like this, full of batches of the 16 aural letters, could help users create a sound library, or even help people in the same car to hear separate things.

With our device what you can have is you can strap a static piece on top of existing speakers and they can direct sound in two different directions without any overlap. So the passengers can hear completely different information from the driver,” explains Professor Sri Subramanian Interact Lab at University of Sussex. This technology is more than five years away, but smaller versions could be used to direct medical ultrasound devices far sooner.  “In a year we could have a sleeve that we can put on top of already existing projects in the market and make them just a little bit better. For example, we can have a sleeve that goes on top of ultrasound pain relieving devices that are used for therapeutic pain,” he adds.
Researchers say spatial sound modulators will one day allow us to perform audible tasks previously unheard of.

Source: http://www.sussex.ac.uk/

Virtual Images that Blend In And Interact With The Real-World

Avegant, a Silicon Valley startup that sells a pair of headphones equipped with a VR-like portable screen, is breaking into augmented reality. The company today announced that it’s developed a new type of headset technology powered by a so-called light field display.

Avegant ARCLICK ON THE IMAGE TO ENJOY THE VIDEO

The research prototype, which Avegant eventually plans on turning into a consumer product, is based on the company’s previous work with its Glyph projector. That device was a visor of sorts that floats a virtual movie screen in front of your eyes, and developing it gave Avegant insight into how to build an AR headset of its own.

Like Microsoft’s HoloLens and the supposed prototype from secretive AR startup Magic Leap, Avegant’s new headset creates virtual images that blend in and interact with the real-world environment. In a demo, the company’s wired prototype proved to be superior in key ways to the developer version of the HoloLens. Avegant attributes this not to the power of its tethered PC, but to the device’s light field display — a technology Magic Leap also claims to have developed, yet has never been shown off to the public.

The demo I experienced featured a tour of a virtual Solar System, an immersion within an ocean environment, and a conversation with a virtual life-sized human being standing in the same room. To be fair, Avegant was using a tethered and bulky headset that wasn’t all that comfortable, while the HoloLens developer version is a refined wireless device. Yet with that said, Avegant’s prototype managed to expand the field of view, so you’re looking through a window more the size of a Moleskine notebook instead of a pack of playing cards. The images it produced also felt sharper, richer, and more realistic.

In the Solar System demo, I was able to observe a satellite orbiting an Earth no larger than a bocce ball and identify the Big Red Spot on Jupiter. Avegant constructed its demo to show off how these objects could exist at different focal lengths in a fixed environment — in this case a converted conference room at the company’s Belmont, California office. So I was able to stand behind the Sun and squint until the star went out of focus in one corner of my vision and a virtual Saturn and its rings became crystal clear in the distance.

Source: http://www.theverge.com/

Nano-Implant Could Restore Sight

A team of engineers at the University of California San Diego (UC San Diego)  and La Jolla-based startup Nanovision Biosciences Inc. have developed the nanotechnology and wireless electronics for a new type of retinal prosthesis that brings research a step closer to restoring the ability of neurons in the retina to respond to light. The researchers demonstrated this response to light in a rat retina interfacing with a prototype of the device in vitro. The technology could help tens of millions of people worldwide suffering from neurodegenerative diseases that affect eyesight, including macular degeneration, retinitis pigmentosa and loss of vision due to diabetes.

Despite tremendous advances in the development of retinal prostheses over the past two decades, the performance of devices currently on the market to help the blind regain functional vision is still severely limited—well under the acuity threshold of 20/200 that defines legal blindness.

cortical neuronsPrimary cortical neurons cultured on the surface of an array of optoelectronic nanowires. Note the extensive neurite outgrowth and network formation

We want to create a new class of devices with drastically improved capabilities to help people with impaired vision,” said Gabriel A. Silva, one of the senior authors of the work and professor in bioengineering and ophthalmology at UC San Diego. Silva also is one of the original founders of Nanovision.

Power is delivered wirelessly, from outside the body to the implant, through an inductive powering telemetry system developed by a team led by Cauwenberghs.

The device is highly energy efficient because it minimizes energy losses in wireless power and data transmission and in the stimulation process, recycling electrostatic energy circulating within the inductive resonant tank, and between capacitance on the electrodes and the resonant tank. Up to 90 percent of the energy transmitted is actually delivered and used for stimulation, which means less RF wireless power emitting radiation in the transmission, and less heating of the surrounding tissue from dissipated power. For proof-of-concept, the researchers inserted the wirelessly powered nanowire array beneath a transgenic rat retina with rhodopsin P23H knock-in retinal degeneration.

The findings are published in a recent issue of the Journal of Neural Engineering.

Source: http://ucsdnews.ucsd.edu/

 

Nano-LED 1000 Times More Efficient

The electronic data connections within and between microchips are increasingly becoming a bottleneck in the exponential growth of data traffic worldwide. Optical connections are the obvious successors but optical data transmission requires an adequate nanoscale light source, and this has been lacking. Scientists at Eindhoven University of Technology (TU/e) now have created a light source that has the right characteristics: a nano-LED that is 1000 times more efficient than its predecessors, and is capable of handling gigabits per second data speeds.

NANO LEDWith electrical cables reaching their limits, optical connections like fiberglass are increasingly becoming the standard for data traffic. Over longer distances almost all data transmission is optical. Within computer systems and microchips, too, the growth of data traffic is exponential, but that traffic is still electronic, and this is increasingly becoming a bottleneck. Since these connections (‘interconnects’) account for the majority of the energy consumed by chips, many scientists around the world are working on enabling optical (photonic) interconnects. Crucial to this is the light source that converts the data into light signals which must be small enough to fit into the microscopic structures of microchips. At the same time, the output capacity and efficiency have to be good. Especially the efficiency is a challenge, as small light sources, powered by nano– or microwatts, have always performed very inefficiently to date.
The researchers in Eindhoven believe that their nano-LED is a viable solution that will take the brake off the growth of data traffic on chips. However, they are cautious about the prospects. The development is not yet at the stage where it can be exploited by the industry and the production technology that is needed still has to get off the ground.
The findings are reported in the online journal Nature Communications.

Source: https://www.tue.nl/

How To Fabricate The Hardest Diamond

The Australian National University (ANU) has led an international project to make a diamond that’s predicted to be harder than a jeweller’s diamond and useful for cutting through ultra-solid materials on mining sites. ANU Associate Professor Jodie Bradby said her team – including ANU PhD student Thomas Shiell and experts from RMIT, the University of Sydney and the United States – made nano-sized Lonsdaleite, which is a hexagonal diamond only found in nature at the site of meteorite impacts such as Canyon Diablo in the US.

diamond

This new diamond is not going to be on any engagement rings. You’ll more likely find it on a mining site – but I still think that diamonds are a scientist’s best friend. Any time you need a super-hard material to cut something, this new diamond has the potential to do it more easily and more quickly,” said Dr Bradby from the ANU Research School of Physics and Engineering.

Her research team made the Lonsdaleite in a diamond anvil at 400 degrees Celsius, halving the temperature at which it can be formed in a laboratory. “The hexagonal structure of this diamond’s atoms makes it much harder than regular diamonds, which have a cubic structure. We’ve been able to make it at the nanoscale and this is exciting because often with these materials ‘smaller is stronger‘.”

Lonsdaleite is named after the famous British pioneering female crystallographer Dame Kathleen Lonsdale, who was the first woman elected as a Fellow to the Royal Society.

The research is published in Scientific Reports.

Source: http://www.anu.edu.au/

Ultra Thin Nightvision Glasses Based On NanoPhotonics

Scientists from the Australian National University (ANU) have designed a nano crystal around 500 times smaller than a human hair that turns darkness into visible light and can be used to create light-weight night-vision glasses. Professor Dragomir Neshev from ANU said the new night-vision glasses could replace the cumbersome and bulky night-vision binoculars currently in use.

ultra-thin-nano-crystal-film_anu-1

The nano crystals are so small they could be fitted as an ultra-thin film to normal eye glasses to enable night vision,” said Professor Neshev from the Nonlinear Physics Centre within the ANU Research School of Physics and Engineering.

This tiny device could have other exciting uses including in anti-counterfeit devices in bank notes, imaging cells for medical applications and holograms.”

Co-researcher Dr Mohsen Rahmani said the ANU team’s achievement was a big milestone in the field of nanophotonics, which involves the study of behaviour of light and interaction of objects with light at the nano-scale.

nightvision-glasses

These semi-conductor nano-crystals can transfer the highest intensity of light and engineer complex light beams that could be used with a laser to project a holographic image in modern displays,” said Dr Rahmani, a recipient of the Australian Research Council (ARC) Discovery Early Career Researcher Award based at the ANU Research School of Physics and Engineering.

PhD student Maria del Rocio Camacho-Morales said the team built the device on glass so that light can pass through, which was critical for optical displays.

Source: http://www.anu.edu.au/

How To Generate Wonderful Colors

Colors are produced in a variety of ways. The best known colors are pigments. However, the very bright colors of the blue tarantula or peacock feathers do not result from pigments, but from nanostructures that cause the reflected light waves to overlap. This produces extraordinarily dynamic color effects.

blue-tarantulaScientists from Karlsruhe Institute of Technology (KIT) in Germany, in cooperation with international colleagues, have now succeeded in replicating nanostructures that generate the same color irrespective of the viewing angle.

In contrast to pigments, structural colors are non-toxic, more vibrant and durable. In industrial production, however, pigments have the drawback of being strongly iridescent, which means that the color perceived depends on the viewing angle. An example is the rear side of a CD. Hence, such colors cannot be used for all applications. Bright colors of animals, by contrast, are often independent of the angle of view. Feathers of the kingfisher always appear blue, no matter from which angle we look. The reason lies in the nanostructures: While regular structures are iridescent, amorphous or irregular structures always produce the same color. Yet, industry can only produce regular nanostructures in an economically efficient way. Radwanul Hasan Siddique, researcher at KIT in collaboration with scientists from USA and Belgium has now discovered that the blue tarantula does not exhibit iridescence in spite of periodic structures on its hairs. First, their study revealed that the hairs are multi-layered, flower-like structure. Then, the researchers analyzed its reflection behavior with the help of computer simulations. In parallel, they built models of these structures using nano-3D printers and optimized the models with the help of the simulations. In the end, they produced a flower-like structure that generates the same color over a viewing angle of 160 degrees. This is the largest viewing angle of any synthetic structural color reached so far.

Apart from the multi-layered structure and rotational symmetry, it is the hierarchical structure from micro to nano that ensures homogeneous reflection intensity and prevents color changes. Via the size of the “flower,” the resulting color can be adjusted, which makes this coloring method interesting for industry. “This could be a key first step towards a future where structural colorants replace the toxic pigments currently used in textile, packaging, and cosmetic industries,” says Radwanul Hasan Siddique of KIT’s Institute of Microstructure Technology, who now works at the California Institute of Technology. He considers short-term application in textile industry feasible. Dr. Hendrik Hölscher thinks that the scalability of nano-3D printing is the biggest challenge on the way towards industrial use. Only few companies in the world are able to produce such prints.

Source: http://www.kit.edu