Three new samples of future human-computer interaction gesture control technology routes

At present, the forward-looking test about the control of the interior of the car is unrelenting, Jaguar Land Rover is the most sci-fi, Ford is more accurate, and Toyota has chosen to be light and fun.

A few days ago, a message at the Frankfurt Motor Show gave us some hope for gesture control. The exhibitor's parts supplier Delphi showed its own gesture control technology. By using the camera to transmit and receive infrared light, the system can judge the gesture operation from the change of reflected light. At present, it is determined that the car company equipped with this technology and mass production is BMW. On the new 7 series listed on the domestic market on October 24 this year, buyers can enjoy this futuristic human-computer interaction technology.

However, each new technology product has to undergo many stages such as forward-looking tests, market research, sample display, test verification, etc., before finally reaching the final mass production. At present, the forward-looking test about the control of the interior of the car is unrestrained, and the news of mass production in the industry is only catching the wind. What are the reasons for the lack of scale of technology, and what obstacles and bottlenecks have been encountered in the intermediate stage of forward-looking to mass production? It seems to be a topic worth talking about.

Three days of gesture control case

In Xiaobian's view, technologies that are not necessarily mass-produced often point to new possibilities. In the existing case of gesture control, we found three interesting cases. Jaguar Land Rover is the most sci-fi, Ford has shown a high degree of precision, and Microsoft's project is light and fun.

Jaguar Land Rover and UltrahapaTIcs: tangible, invisible

This is a news not far from us. It turns out that Jaguar Land Rover invested in a very young British startup, UltrahapaTIcs. The startup's gesture and touch technology called ULT003 Mid-air touch has achieved very good results. Cool. The most interesting thing is that when you make touch operations in the air, your hands can feel tactile feedback similar to manipulating physical buttons.

The technical principle behind this is that there will be an ultrasonic transmitting device in the touch area. By varying the air pressure, a few high-voltage points of sound waves can be gathered at a specific position to generate a “touchable, invisible” interface. As long as the hand moves within a valid range, the device will track the change in your hand position and “fabricate” a virtual three-dimensional object in a suitable position. It can also be a “knob” or “button”.

ULT003 Mid-air touch display sample

Compared with the general touch operation, the button will move synchronously according to the position of your hand, without the need to look at the screen to find the touch point in the traditional gesture control. Moreover, the user can obtain specific information by touching, and the hand can truly feel the button force of the button and the texture of the touch to the plane, and the judgment of the operation effect is determined by the actual tactile feedback.

The whole process of interaction is like reading a novel intently and carefully, taking the water from the table and drinking it, and the sight does not need to be removed from the book, and the cup will “run” to the hand. There have been many people who have questioned whether gestures and touches still need to pay attention to whether the touch position is in the proper range. If the operation is successful, you should also glance at the screen to confirm. UltrashapaTIcs's “hand-to-hand” technology has eliminated these two concerns at one time.

Ford and Intel: Complex Face Recognition

The gesture control of Ford and Intel is part of the project called Mobii. In principle, the camera is used to recognize gestures. The car cloud has been introduced in the earlier article, using structured light technology. In simple terms, it is still the principle of infrared light reflection to determine the position, contour and motion of the object, but this kind of light source has the characteristics of point line surface, especially good at 3D reconstruction of a stereo image.

In the function presentation, the gesture control is combined with a simple voice command combination, and the gesture can be used to adjust the temperature or switch the sunroof while driving. This feature does not have much in the above video display, not technical limitations or bottlenecks, but because structured light technology appears to be somewhat overkill in gesture control, and Ford and Intel are targeting more complex facial recognition.

The face of the driver's seat is scanned by the camera in front of the driver to determine whether it is the certified owner and whether or not the vehicle system is activated. When the stranger makes further operations, the owner's mobile phone will receive a positive picture of the insider, which can be authorized on the mobile phone. Drivers who have already been certified can also activate the personalization settings saved in the system, in addition to a range of functions such as turning on the car system.

From the display of facial recognition technology, we can also discover the two advantages of this technology from the side. First, the way to verify your identity with your face can prove the level of sophistication of this technology. After all, the "seeing" person's face facial features and the clear hand shape change, reached a higher level of precision. Secondly, the data shows that if the system detects that there are other passengers in the car, it will close the page that reveals the privacy of the user, only the navigation interface is exposed. To some extent, the system used for identification can support multi-threaded jobs.

Toyota and Microsoft: four gesture poses

Microsoft's thinking about gesture control looks very interesting. The Internet company once considered introducing gesture control into the interior of the car, using only the built-in camera on our phone. Compared with the first two cases, this is a “post-installation” scheme for the interior environment.

This trend was discovered because of a patent application from Microsoft in 2012. In the patent content, three researchers built a gesture control system for the car entertainment information system. But these control actions are completely different from our usual raising, sliding, and rotating. They are like the picture below.

The index finger against the lips means that the volume of the mobile phone is lowered, and the hand is turned into a trumpet to indicate that the phone is used to make a call. A similar thinking chin movement means "I want to retrieve information", and the like action represents "agree".

When using this system, you need to put the smartphone on the top of the dashboard and use the built-in imagination of the phone to capture the gestures and posture movements of the car. It is reasonable to speculate that the function behind this function should be Image recognition technology. At the software level, the three researchers set the system to use gesture control with speech recognition, which is more common in human-computer interaction in the car.

This gesture control solution has not yet seen the landing case, you still can't see these strange control gestures in the car. The information also shows that in the project of Microsoft's concept car Smart Insect, which was co-operated with Toyota in 2013, the pre-installed version of the gesture control function has been implemented with the front-mounted camera. All gestures resume normal lifts and are used to operate secondary control functions in the car, such as entertainment information systems, navigation or to make calls to phones connected to the car.

Why is gesture control delayed?

The thoughts of the above three cases are relatively unpredictable, and for the final commercialization time, each family did not give a definite date. The news of Jaguar Land Rover's investment in UltrahapaTIcs took place this month. Ford and Toyota's technology show has been wavering as early as a few years ago, but the latest information has not been seen. For a considerable period of time, these three technologies have remained at the stage of “can’t see but can’t buy”.

In fact, with regard to gesture control, in addition to the BMW that has been able to show off the time of the product, there are many car companies aiming at the mass production target. At the CES at the beginning of this year, Volkswagen also showed a TOF-based gesture control technology with a Golf R Touch, and the same technology as the BMW 7 Series, but the first production model is said to be the Audi A8 in 2018. Mercedes-Benz has been the first to showcase the DICE (Dynamic & Intuitive Control Experience) in 2012. Although the sample has been misjudged, it has been improved to the latest version in a few years, but it is equipped with a new generation of Mercedes-Benz E-Class and S-Class. The future listing may be expected.

In addition to the demo show at the ABB camp, other companies have shown a strong interest in gesture control technology from start to finish. It is said that in the past few years, Ford, Honda, and Hyundai have applied for patents on gesture control. GM also works with university researchers and wants to make some names in this area. But at present, it seems that only one BMW is the only one in this technology.

From the endless stream of case studies, there seems to be no absolute threshold at the technical level. Most of the existing gesture control technologies use infrared technology, radar detection and image recognition to achieve functions. Therefore, a question worth pondering is why the gesture control has been delayed, and what is the status of this human-computer interaction technology in the future car? In the view of car cloud, the reasons may be as follows:

1. The limited possibility of expression under the premise of safety hinders the gesture control technology from entering the mass production model. One obvious trend is that gesture control is often used in conjunction with other human-computer interaction features. For example, in Ford and Intel projects, gesture control needs to be operated with voice control. In Xiaobian's view, the purpose of this setting is to simplify the operation process and ensure the safety of the driving environment.

In this case, the starting point of the combination of voice and gesture control is to use gesture control to visually indicate the target of the action. This part of the function often requires complicated analysis resources in the analysis stage of voice control. What makes the target do something is naturally not a gesture that can be said clearly.

2. The lack of precision in function execution and insufficient smoothness also restrict the popularity of gesture control to a certain extent. If you summarize the above cases and the gesture control that appears at this stage, it is not difficult to find that the application of gesture control in the car is mainly focused on secondary control, focusing on the entertainment information system and the control of in-vehicle facilities such as skylights and air conditioners. It does not involve driving control.

Between two related operations, it is likely that the control target will be misjudged because of the limitations of the interior space. You just want to turn down the volume a few squares, but accidentally turn on the wiper. In addition, between the different instruction actions of the same target, it is necessary to establish a clear distinction, and the setting of the start and stop actions also has a complicated logical relationship. Car cloud found that Audi displayed the gesture control in CES, and set the right hand flat pressing action in the control area as the starting point of the operation function. However, once the right hand leaves the active area, it needs to re-execute the push-down function, which is a redundant operation that has to be retained in practical applications.

The interaction goal between people and cars is to get as close as possible to the form of communication in daily life, and even the communication between people and cars becomes less noticeable. Therefore, gesture control should also pay attention to the strategy of simplifying and simplifying, and use the gestures and postures that are naturally exposed to play a role in the most advantageous aspects. Mass production cannot be blind, and it is a serious matter to use force on the blade.

LED Pool Light

Led Pool Light is a much required lighting solution. Used for swimming pools, water sports fields, ocean activity areas and so on, LEDER brings you a series of LED pool lights, which will meet your requirements for activities in the water.

For all kinds of water sports and water-related activities, safety measures must be taken first. In this regard, pool lights need to be installed adequately and in strategic locations. The safety and well-being of people who engage in water activities cannot be overemphasized. Meanwhile, LED Underwater Lighting is known for its energy efficiency and durability, making it a smart choice for many spaces such as swimming pools and aquatic theme parks.

LED Pool Light

Floating Pool Lights,Swimming Pool Lights,Solar Pool Lights ,Above Ground Pool Lights

JIANGMEN LEDERLIGHT LIGHTING Co.,LTD , https://www.lederlight.com