Tuesday, October 11, 2016

Depth study change visual examples search

Lei feng's network (search for "Lei feng's network" public attention) by writer Deng Qili, second-year computer Department of Harbin Institute of technology Shenzhen Graduate School, graduate, tutor for "city of scholars," Professor Xu. Research interest for deep learning and computer vision. First was awarded the 2015 Alibaba large-scale image search contest second prize, it ranked third.

Summary

In recent years, the depth of learning in computer vision tasks, have made a major breakthrough, one of the important factors is the strong nonlinear capabilities to understand the image deeper information. Based on deep learning Visual examples of the simple search method summarized and generalized, hopes to inspire readers. Fendi

Objective FENDI iPhone 6+

Given a query containing an object picture, Visual Search task is an instance found in the photo gallery from the candidate those queries image contains a picture of the same object. Compared with the General image search, instance search search criteria more demanding--contain the same objects, like a dress, the same car, and so on. The issue has a very wide range of applications, such as product search, vehicle search and location based on image recognition and so on. For example, mobile image search is through the analysis of goods using mobile phone camera photo, find the same or similar items from the library.

However, in the actual scene, because of the attitude, interference factors such as lighting and background, so the two images that contain the same objects tend to vary greatly in appearance. From this point of view, the Visual examples should search for the essence of the question is what kind of image feature which contains images of the same object is similar in feature space.

In recent years, deep learning in computer vision tasks are achieved major breakthroughs, including Visual examples search tasks. Paper main on based on depth learning of instance search algorithm (following referred to for "depth instance search algorithm") for analysis and summary, articles is divided into four a part: first part summary has classic Visual instance search algorithm of general process; second part and third part respectively from two a aspects to introduced in recent years main of depth instance search algorithm; end to end of features learning method and based on CNN features of features coding method Part I will summarize in 2015 first Alibaba a large image competition (Alibaba Large-scale Image Search Challenge, ALISC) appears in the related methods, introduce some examples of practice can improve search performance techniques and methods.

Classic examples of Visual search algorithm flow

In depth learning popular zhiqian, typical of instance search algorithm General is divided into three a stage: first in given of image intensive to extraction local not variable features, then will these local not variable features further to coding for a compact of image said, last will query image and candidate image library in the of image for similar degrees calculation (based on second step get of image said), found those belongs to same instance of pictures.

1. local invariant features. Extracting local invariant features details about local area of an image, do not care about the global information, and light within the local area has some invariant, geometric transformation. This for instance searching makes perfect sense because the target objects can be accompanied by a geometric transformation occurs in any area of an image. In earlier work in many instances search method using SIFT features.

2. character encoding. Significance to local features further coding is two-fold: mining information between these local features, enhanced discrimination ability; a single compact index of the eigenvectors is easier to implement, improve search speed. Common methods of VLAD (vector of locally aggregated descriptors), Fisher Vectors,triangular embedding. In here, paper simple to introduced Xia VLAD method (in paper behind times appeared): a,) VLAD method first using k-means get contains k a center of code this, then each local features was assigned to away from it recently of Center points (we will this a step called hard-assignment, zhihou will related articles on this for improved), last will these local features with assigned of Center points Zhijian of residual poor tired and as eventually of image said. From above we can see that VLAD disorder – don't care about local properties of spatial location, so you can further decouple global space information, is very robust to geometric transform.

3. the similarity calculation. A direct approach is based on distance function calculates the distance between features, such as the Euclidean distance, cosine distance. Another is learning the appropriate distance function, such as LMNN, ITML and measurement methods.

Summary: classic examples of Visual search algorithm performance is limited by the characteristics of hand-crafted said. In instances when the application deep learning search tasks, mainly feature from the start, namely how to extract more distinguishing images features

Characteristics of end-to-end learning method

NetVLAD: CNN architecture for weakly supervised place recognition (CVPR 2016)

This article is from INRIA Relja against Arandjelovi and other people's work. The article concerns a specific application instance search--location awareness. Position recognition problem, given a query image by querying a large place markers for a data set, and then use those similar picture to evaluate queries image locations. First, using Google Street View where Time Machine for mass tagging data set, then presented a Convolutional neural network architecture, NetVLAD--, VLAD method embedded in the CNN network, and "end-to-end" of learning. The method shown in the following figure:

Depth study change visual examples search

Hard-assignment is not in the original VLAD micro (local characteristic of each assigned to its nearest the center point) and therefore cannot be directly embedded into the CNN network, and participate in the error back-propagation. This article's workaround is using the softmax function to the hard-assignment operation soft-assignment operation-use 1x1 convolution and softmax function to get the local feature belongs to probability/weight of each center point, and then assigns maximum probability weight/center point. Therefore NetVLAD contains three parameters can be studied, and 1x1 above which is the convolution of the parameters used to predict soft-assignment, expressed as the center point of each cluster. And then on the figure of VLAD in the core layer through the accumulation of residual action. Author gives us the following figure illustrate the advantages of NetVLAD compared to the original VLAD: (more flexibility--learn better cluster-center point)

Depth study change visual examples search

Another improvement for this article is Weakly supervised triplet ranking loss. The method to solve the problem of training data may contain noise, triplet ranking replace respectively with potential loss of positive and negative samples are samples (containing at least one sample, but not sure which one) and a clear set of negative samples. And when you train, bound queries image and sample concentration is most likely are samples of the characteristic distance between pictures than queries image with all negative characteristic distance between pictures in the sample set is smaller.

Deep Relative Distance Learning: Tell the Difference Between Similar Vehicles (CVPR 2016)

Next article focused on vehicle identification/search problem, from Hongye Liu, who works at Peking University. As shown in the following figure, this problem can also be seen as instances search tasks.

Depth study change visual examples search

And a lot of the depth of supervisory instance search methods, this article is an attempt to map the original picture to a European in the feature space, and makes the space, pictures gathered more of the same vehicle, rather than a similar vehicle pictures are even more far away. In order to achieve the effect, the commonly used method is by optimizing the triplet ranking loss, to train the CNN network. However, the authors found that the original triplet ranking loss, there are some problems, as shown in the following figure:

Depth study change visual examples search

For the same sample, triples will be on the left of the loss function adjustment and the triple to the right is ignored. Difference between the two lies in the choice of anchor is different, this led to instability while training. In order to overcome this problem, by using coupled clusters loss (CCL) to replace the triplet ranking loss. The loss function is characterized by the triples into a sample set and a set of negative samples and samples gathered in the sample, samples of negative samples in the sample set and those who are further apart, so as to avoid the negative impact of randomly selected samples of anchor. The effect of the loss function as shown in the following figure:

Depth study change visual examples search

Finally this article for the particularity of vehicle problems, and design coupled clusters above loss, a hybrid network architecture is designed, and built a vehicle database so as to provide the necessary training.

DeepFashion: Powering Robust Clothes Recognition and Retrieval with Rich Annotations (CVPR 2016)

Finally, this article was also published in CVPR 2016, clothes were introduced to identify and search, also is an instance with search-related tasks from the Ziwei Liu, who works at the Chinese University of Hong Kong. First of all, this article introduces a database of clothes called DeepFashion. The database contains over 800K dress pictures, 50 fine grained categories and 1000 properties, and can also provide an additional key to clothes clothes attitude/cross-cutting and cross-relationship (cross-pose/cross-domain pair correspondences), some specific examples shown in the following figure:

Depth study change visual examples search

In order to illustrate the effect of the database, the author presents a new depth of learning networks, FashionNet--through the key points of the joint forecast clothes and property, learning to be more differentiated features. The overall framework of the network looks like this:

Depth study change visual examples search

FashionNet forward the calculation process is divided into three phases: the first phase, a clothes picture blue branch of the input to the network, the key to predicting clothes are visible and the position of the point. The second phase, according to the forecast in the previous step's key points, the key pool (landmark pooling layer) get clothes of local features. The third phase, "FC6 global" global features and "FC6 local" local features are spliced together to form a "fc7_fusion", as the final image. FashionNet introduces four loss function, and uses an iterative training to optimize. These losses are as follows: return loss corresponds to a key position, softmax loss is visible corresponds to the key points and clothing category, cross-entropy loss function corresponds to property prediction and triples loss function corresponds to the similarities between the clothes. Authors from the clothing category, property prediction and clothing searches these three aspects, FashionNet compared to other methods, have achieved significantly better results.

Summary: when there are enough tagging data, deep learning can learn features and measurement functions. The idea behind it is based on a given metric, learning characteristic features in the metric space has the best judgement. End-to-end feature learning characteristics of main research interest is how to build a better representation and loss of function.

CNN features the character coding method

Depth of instances of the search algorithm described in the above section of this article, and focuses on data-driven end-to-end characteristics of learning methods and corresponding images to search for data sets. Next, we focus on another issue: when there is no search data related to these sets, how to extract a valid image characteristics. In order to overcome the lack of field data, pre-training model is a viable strategy in the CNN (CNN model of training data sets in other tasks, like ImageNet taxonomy data set) based on extracting characteristics of a layer map (feature map), encode it to be suitable for instance search feature. This section under related papers in recent years, some of the key methods (in particular, all CNN model in this section are based on the ImageNet set of training models for categorical data).

Multi-Scale Orderless Pooling of Deep Convolutional Activation Features (ECCV 2014)

This article was published in ECCV 2014, is from the University of North Carolina at Chapel Hill Yunchao Gong and others, including the University of Illinois at Urbana-Champaign Liwei Wang's work. Because the CNN feature missing geometric invariance limits the variable scene classification and matching. Author attributed the problem of global CNN feature contains too much information and therefore made a multi-scale orderless pooling (MOP-CNN)-CNN features VLAD method combined with the disorder.

MOP-CNN the main steps, CNN network is first seen as a "local character" extractor and extract image on multiple scales "local character", and with VLAD the size of each "feature" encoding for the scale image features, and finally all the scale image features to connect together to form the final image. Feature extraction framework as shown below:

Depth study change visual examples search

Author search on the classification and case test on two tasks, as shown in the following figure, proves that MOP-CNN than in a regular CNN global features a better classification and search results.

Depth study change visual examples search

Exploiting Local Features from Deep Networks for Image Retrieval (CVPR 2015 workshop)

This article was published in a CVPR 2015 on the workshop, from the University of Maryland, College Park people such as Joe Yue-Hei Ng. Recent studies showed that, compared to the linked layer's output, convolution of feature maps (feature map) is more suitable for instance search. This article describes how to use the convolution of feature maps into a "local character", and VLAD the encoding to use for the image feature. In addition, the authors also conducted a series of experiments to observe different convolution effect of signatures for instance search accuracy.

Depth study change visual examples search

Aggregating Deep Convolutional Features for Image Retrieval(ICCV 2015)

This article was published on the ICCV 2015 next, comes from the Moscow Institute of physics and technology Artem Babenko and Victor Lempitsky skolkovo technology college work. Can be seen from the above two articles, a lot of depth examples search method using chaotic-coded method. But these, including the VLAD,Fisher Vector coding method of computation is usually relatively large. In order to overcome this problem, this article we design a simpler and more efficient coding method--Sum pooing. Sum of pooling specific definition looks like this:

Depth study change visual examples search

Which is the convolution in the spatial location of local features (local feature extraction methods, consistent with the last article). Using sum after pooling, global features to further implementation of PCA and L2 normalized to get the final features. Fisher Vector,Triangulation embedding and author Max pooling and comparing these methods proved sum pooling method is not only simple, and better.

Where to Focus: Query Adaptive Matching for Instance Retrieval Using Convolutional Feature Maps (arXiv 1606.6811)

Finally this article posted on the arXiv, came from Australia, University of Queensland Jiewei Cao and other people's work. As mentioned at the beginning of this article, messy background for instance search is very large. In order to overcome this problem, this article based on the article on the method of sum-pooling, presents a known query Adaptive matching (QAM) methodology to calculate the similarity between images. The core of the method is to perform pooled operations on multiple areas of an image, and create more characters to express images. And then in the match, queries image will correspond to the characteristics of these regions and best match score as a similarity between two images. So the next question is how to build this area.

Author proposes two approaches--Feature Map Pooling and Overlapped Spatial Pyramid Pooling (OSPP), to get the base region of the image. And then through the base region continue to merge, in order to find the best similarity scores for goals, build out of the goal area. One of the most attractive part is that authors will throughout the merger process, into the solution of a problem. Graph below shows part of QAM method results and the corresponding image map.

Depth study change visual examples search

Summary: in some instances the search task, due to the lack of sufficient training, so you can't directly "end to end" to learn of image features. At this time, how to turn ready-made CNN feature coding for instance Search images that become a hot research area.

2015 first summarises Alibaba a large image contest

Introduction in recent years, some examples of the depth of the main search method when you are finished, in the next part of this paper will summarize in Alibaba a large search contest related methods that appear in the image, to introduce some practice, you can improve the performance of Visual examples search techniques and methods.

Large-scale image search image search contest by Ali Alibaba group sponsored by requiring participating teams from huge Photo Gallery to find out those queries image contains a picture of the same object. This competition provides the following two types of data used for training: training about 200W picture collection (category levels that correspond to the labels and attributes), 1417-verify queries image and the corresponding search results (for a total of approximately 10W). In tests, given 3567 queries image, teams need from approximately 300W picture evaluation focused (no label), pictures of people who meet the requirements, evaluation is based on the top 20 of the mAP (mean Average Precision).

Briefly introduce our method for--Multi-level Image Representation for Instance Retrieval, which made the third. Many methods are used at the end of a convolution or retrieves the link layer characteristics, due to the high level of features have lost a lot of detail (for a deeper network, more serious loss), so the instance search is not very accurate, as shown in the following figure, the overall outline is similar, but the details are far.

FENDI iPhone 6+

Depth study change visual examples search

In order to overcome this problem, we will feature maps of the different layers in the CNN network (feature map) for integration, not only using the high-level features of semantic information, as well as details of the low-level features information, instance search more accurate. As shown in the following figure, our experiment is based on the GoogLeNet-22 network, for the last 8 characters (from the Inception of 3B to Inception 5B), first use the largest pool of these feature maps of different scales are sampled (converted to a character of the same), and use of these sampling results further convolution. And then these characteristics make weighted linear (convolution), on the basis of using sum pooling to get the final image. In training, we are based on the training data provided by optimizing triplet based on cosine distance ranking loss to end-to-end learning these characteristics. During testing, you can directly use cosine distances between features to measure the similarity of the image.

Depth study change visual examples search

Addition, reference Yu training SVM classification device Shi using has difficult points sample mining of thought, we of method first in Qian to calculation Shi, calculation current training batch in the all potential ternary group of loss (from current training batch in the selected two Zhang same category of pictures and a Zhang different category of pictures constitute potential ternary group), then found those "difficult" of ternary Group (more big of loss), last in reverse calculation Shi, using these "difficult" of ternary group for errors spread, to made better of training effect.

Next a brief summary of related methods that other teams. Characteristics of end-to-end learning methods, in addition to triplet ranking loss,Contrastive loss (corresponds to the Siamese network) is also a common loss function. In addition, there are worthy of our attention, you can improve search performance dramatically:

(A) mining

In supervised machine learning methods, more data can mean a higher degree of accuracy. Team in the calculations come from the Chinese Academy of Sciences, under the ImageNet-feature of the training model, clustering on the training set at the category level, and then through the threshold, more out with the figure, and with plans to train in the CNN network, learning feature. This method is simple, and you can improve search performance dramatically.

(B) target detection

Case retrieval in the complex final search performance is directly affected by the background noise. First, so a lot of teams try to use object detection (such as faster-rcnn) to locate areas of interest, and then further learning features, compare similarities. In addition, when there are no bounding box when training data, weak supervision is an effective method to locate the target.

(C) the first pool features and erjiechi characters fusion

Erjiechi method of capturing images of second-order statistical variables, covariance, can often achieve better search accuracy. Team led by Professor Li Peihua from Dalian in the CNN network based on the characteristics of first-order characteristics of pools and erjiechi integration, achieving very good results.

(D) joint feature and property prediction

This method and the DeepFashion mentioned in the third part of this article is similar to learning characteristics and prediction of properties of the picture (missions), resulting in more differentiated features.

Lei Feng network Note: authorized to Lei feng was released by deep learning Forum in this paper, please indicate the author and the source, no deletion of content.

Thursday, October 6, 2016

Verizon acquisition of addiction After Yahoo bought two car data company

Verizon acquisition of addiction? After Yahoo bought two car data company

Traditional mobile communication market is becoming saturated, many telecom operators are beginning to look to mobile communications markets in emerging areas. Last week, Verizon has spent $ 4.8 billion acquisition of Yahoo's digital media, advertising, and e-mail and other core business, intention to develop their Internet strategies. From the latest news, Verizon's ambitions may be more than that.

According to foreign media reports, Verizon also recently acquired a privately owned automobile to the Internet of things technologies Telogis, but prices were not disclosed. In addition, the news last week that Verizon will be the $ 2.4 billion acquisition of motor Fleetmatics GPS data-tracing company. Just Cavalli iPhone 5S case

More than Verizon, as the United States's second-largest telecommunications carrier, Verizon's main rival, AT&T seems to be depending on vehicle behavior of a potential market. In February this year, AT&T has been with Volkswagen, Porsche and Audi car company signed a cooperation agreement, they will be in the United States within a localized networking services provide technical support. Just Cavalli iPhone case

Just Cavalli iPhone case

According to industry analysts, telecom operators have to look at the car market, largely because the latter can create new network layout based on carrier current market value. Many express logistics enterprises, their vehicles and staff management is usually the car over a wireless network requires access to the database, so as to achieve real-time monitoring and scheduling purposes. This is obviously capable of integration with the operator's existing business, and add a stable service subscriptions for its revenue.

According to ABI Research Analyst Susan Beardslee's guess, after acquiring Fleetmatics and Telogis, Verizon is likely going to buy a data security or software technology company in order to protect a huge vehicle data, industry insiders believe Israel Internet security company Argus and software management company Movimento is more likely.

Monday, August 29, 2016

Declassified intelligent beauty and dynamic response of the technology behind

Lei feng's network: this paper from Tencent, Lei feng's network (search for "Lei feng's network", public interest) has been authorized to do so. Introduces the three part of face recognition systems, and mobile phone cameras can automatically facial depth analysis why.

This is a "face" of the era, a few words about the human face, the most well known is face recognition. The technology in the finance, insurance, education, security and other areas of activity, become the stars in the field of AI technology. Micro letter before the public, focuses on face recognition, this paper mainly introduces some quiet support for face recognition technology. To learn more about face recognition technology can be found in the application deep learning in face recognition--grandmother model "evolution"

In General, a complete face-recognition system consists of three main components, namely, face matching for face detection and face recognition. Lines of the three actions: face detection face location found in the image, then face matching face found on the eyes, nose, mouth, facial features, such as location, finally, face recognition feature and both face than to compute similarity, confirm the identity corresponding to the face.

                                           Ted Baker iPhone 5S case

Declassified: intelligent beauty and dynamic response of the technology behind the camera

Figure 1 face recognition processes

1. Introduction to face registration

Registration for face (Face Alignment), also known as facial feature points detection and location. Facial feature points different from the corner or SIFT features such as image feature points in General, facial feature points are usually a set of previously defined by manual (see Figure 2). According to different application scenarios, features a different number, for example 5, 68, 82 points.

Declassified: intelligent beauty and dynamic response of the technology behind the camera

 

Figure 2 common facial feature points detection and positioning of target detection

In addition plays a key role in human face recognition system, face matching technology in 3D face modeling, facial animation, facial analysis, face beautify and virtual makeup, face self-timer areas such as the dynamic response of a wide range of applications. Make a small, excellent facial registration tracking performance, mainstream mobile single-frame processing speed can reach less than 3ms, already in "every day p-dynamic response of self", "mobile QQ-a short video", "mobile video chat QQ-" "mobile phone Qzone-the dynamic response of the camera" scenarios, such as landing.

Declassified: intelligent beauty and dynamic response of the technology behind the camera

Figure 3 face beautify and virtual makeup Ted Baker phone cases

2. face registration status of research on traditional face registration

And others face similar changes in illumination, head pose, facial expression and so, will greatly affect the face and occlusion accuracy of registration. Face registration but also has its own characteristics, first describes the structure of the face feature points (outline and facial features), face stability of the structure is intact, and features fixed relative position and secondly, changes in posture and facial feature points of the head position changes. Traditional face registration and to keep trying to find a more accurate characterization of expression of this assemblage of both determine and change, and then select appropriate optimization method based on descriptor, which face feature points.

Most direct feature descriptor is used in color, grayscale, and different skin for each part of the face detection location. Slightly more complex you can choose from a variety of texture features, such as based on Haar texture features and Adaboost training cascade classifier for face registration. Above characterization not considering the relationship between points, and therefore does not have to maintain a reasonable face structure. Active shape models (Active Shape Models, ASM) and active appearance models (Active Appearance Model, AAM) can express both the texture and shape (shape) are two characters.

Shape features of both by the point distribution model (Point Distribution Model, PDM) to express it. Figure 4 to 600 people face the statistical distribution of facial feature points in the image map, red dot denotes the mean of the points. Texture characteristic of each feature of the ASM is, respectively, generated by calculating the points around the neighborhood information corresponding to each feature of the response (the Response Map). Delineated in blue area in Figure 5 is used to calculate the response, and red dots indicate the actual facial feature point location. AAM uses the whole face to describe texture by facial feature point location is transformed to the standard shape is obtained regardless of the shape of face texture, and is independent of shape based on the principal component analysis method for modeling facial texture.

Declassified: intelligent beauty and dynamic response of the technology behind the camera

Depth research on human face image registration

Starting from 2006, deep neural network has been in computer vision, speech recognition, and natural language processing and other fields have achieved unprecedented success, also face registration brought a breezy spring. Scholars are no longer making a face descriptors of the building complex. Academic industry accepted method for deep face registration there are two classes: cascade Convolutional network face registration (Cascade CNN) depth face registration and multitasking.

As shown in Figure 6, Cascade CNN consists of three levels, each level contains multiple convolution network. First-level estimate is given an initial position, on the basis of two fine adjusted after feature point location. Multi-task distribution, Brigadier General registration and other attributes related to face training simultaneously. Property contains the head pose associated with facial features, facial expressions, such as smiling mouth is likely to be open, positive face feature points symmetrically distributed. Multitasking helps enhance the feature point detection accuracy. However the speed of convergence of the different tasks have different and difficult, difficult to train. Currently offers two solutions for adjust the training process of different tasks: early termination rule (task-wise early stopping criterion) and dynamic control mechanism.

 

Declassified: intelligent beauty and dynamic response of the technology behind the camera

Figure 6 Cascade CNN network model

3. face different scenarios face registration registration

Academics face registration with each passing day, industry product applications technical requirements have become more sophisticated, and face different scenarios put forward different requirements of registration.

Face recognition business the core problem is semantic alignment between facial image pixels, namely face feature point location. Error description of the feature location results in extraction of face feature severely deformed, leading to identify performance degradation. In order to better support for face recognition, we increase the scope of change the face frame, in order to reduce dependence on the face detection frame size. Face feature point we choose five, ensures a certain degree of face structure description, and reduces the effect of registration error on face recognition.

 

Declassified: intelligent beauty and dynamic response of the technology behind the camera

Figure 7 for face recognition

Beauty takes facial feature points reach ultra high precision positioning, such as eye makeup eyeliner lashes only positioning is accurate enough in order to achieve the natural fit beauty effect. In order to provide precision, we use the cascade model, rough face facial features, facial for fine positioning.

 

Declassified: intelligent beauty and dynamic response of the technology behind the camera

Figure 8 intelligent beauty

Facial self portrait effect using mobile video, the registration process is strict. Traditional face matching technology does not have the capacity to determine whether tracking success, in order to avoid lost phenomena in the process of tracking (tracking to the inhuman face), you must rely on time-consuming face detection, face registration we face judgment features have been added, reduce reliance on face detection. We use elongated depth of neural networks, and compression using SVD decomposition model and algorithm acceleration, model size control in the 1M, the processing time on mainstream mobile phones only 3ms. Model size and speed are the industry's highest standards.

 

Ted Baker iPhone 5S case

Declassified: intelligent beauty and dynamic response of the technology behind the camera

1 face self-timer video special effects

Excellent face registration renewal

Excellent laboratory continuous follow up on technology trends, update versions. Face matching technology migration from traditional methods to the deep learning method, from the latest academic research results to the best choice, we went through several rounds of iterative update, did a lot of innovative and try it. Face registration version 1.0 was released in April 2013, rough face features, 4 months after version 2.0 of precision positioning was successfully published, and fun products. After version 3.0 accuracy greatly improved, while landing in beauty products. Version 4.0 using the deep learning method, precision has been further improved, with an average above the artificial level. In May this year we released the latest version 5.0 using depth multi-task learning method, speed and depth of network size has been greatly optimized mainstream mobile frame rates over 200, model 1M, and bring their own face judgment function. Mentioned in the introduction of face self-timer effect with the support of this version.

4. the follow-up research and development plan

Future we strive to enhance the user experience of applications fell, on the other hand is also actively exploring new scenarios. Currently face self-timer video registration and tracking defects still exist. To resolve this problem, improve the user experience depends on further ways of enhancing the human face registration of stability and accuracy. Other than applications referred to in this article has been, face matching technology can also be applied to intelligent access control system, financial core, live on the Internet industry, and many other fields. In new areas of application, studies how face matching technology to meet the new requirements is another issue that we will face.

Lei Feng network Note: this article by Lei feng's network, for reprint, please contact the original author and to indicate the source and author, no deletion of content.

Monday, August 22, 2016

Large print crowding the screen the bracelet of Gu Yu return to social business

On August 19, 2016, it was a normal Friday afternoon, about three points later, a bright green background color of the dialog box in your circle of friends crowding the screen. Hermes iPhone cover

Hermes iPhone cover

All these dialogs from a place called "characters" of the public: you want sent to the "characters", will generate a picture in the background, is returned to you. Friends of big company called seconds like science and technology, access to information to show that this is a company that was founded in the end of 15, mainly founder named Gu Dayu.

Familiar with screen readers may know of Lei Feng Gu Daewoo. 2013 as between sponsor Gu Dayu from Alibaba to leave, start building measuring data and pass data of Ghost in the shell technology, the main product is the Bong bracelet. Published later than a micro, which used to be major SNS products Alibaba against micro-letter; while Bong bracelet launch in 2014 are quickly started in the circles of science and technology, with design, automated algorithm and a good marketing style, Bong when new hardware is the star product of the circle.

Back to second great, this is the third time Daewoo Gu's business. This is the kind of product? Web on the night of the large crowding the screen a short interview of Lei Feng Gu Daewoo.

Characters are kind of product? Gu Dayu says this is what they figured out some time ago an interesting, spent about a day developed and popular than they expected. "Large print" pilot, the stress-test they even finished yet.

Second Chan-what is the main? About a month ago, seconds seconds Zambia launched their main product, and is a main sex video social networking applications. Lei feng's network (search for "Lei feng's network", public interest) edit, download a trial, second broadcast live close to the 1V1. Two random stranger video calling friend automatically after 3 minutes of talk, also set up within the second sowing "appreciation", as flowers in the live sports car, can recharge in Exchange for diamonds. Hermes iPhone case

Gu Dayu says "video social networking is a very interesting area, second seeding now online 3 weeks, progress is ideal. "But he gave no specific data on the second seeding now.

On the previous venture, Gu Yu recalled, at the end of last year, they adjust the Bong become a consumer products company, then he quit to redo a new product. Seconds like technology now has 17, main on the second seeding. Second seeding project currently taking place after the second round of funding.

Thursday, August 11, 2016

GAIR Summit strike

GAIR | Summit strike

On August 12, 2016, the "CCF-GAIR" Global Summit on artificial intelligence and Robotics in Shenzhen Futian Shangri-La Hotel officially opened. The main venue for the Conference has, outside the venue for the exhibition.

GAIR | Summit strike

Field event

GAIR | Summit strike GAIR | Summit strike

The Conference sponsored by the China computer society, Lei feng's network (search for "Lei feng's network", public interest) contractors, government guidance. As many as 1500 people participated in the event, they are fields of artificial intelligence, robotics, intelligent driving practitioners, researchers and experts in various media.

Dry before reports came out of the Summit, let us take a look at the off-site exhibition area. Exhibition brings together domestic and international leader in the field of artificial intelligence, robotics and intelligent handling of enterprise, including search dogs, 360, Jingdong intelligence.

Exhibition mainly products, search dog also offers interactive games, attracting many participants and onlookers.

Exhibition site:

GAIR | Summit strike

Dog man-machine war:

GAIR | Summit strike GAIR | Summit strike

Yang technology showcase the depth cameras, you can quickly identify the parcel size:

GAIR | Summit strike

Si-chi of the main voice intelligent interaction:

GAIR | Summit strike

--Automatic driving of the eye laser radar Velodyne YSL Galaxy S5 Case

More exhibitors:

GAIR | Summit strike GAIR | Summit strike GAIR | Summit strike

360 launched sprouting da child-robot:

GAIR | Summit strike

Intelligent speaker Jing dong Ding Dong:

GAIR | Summit strike

More products: YSL Galaxy S5 Case

YSL Galaxy S5 Case

GAIR | Summit strike GAIR | Summit strike GAIR | Summit strike

Official opening of the General Assembly on the morning of 12th, GAIR dry reports focus more on Lei Feng network Summit topic.

Thursday, August 4, 2016

Super God patent exposures iPhone 7 or dual camera offers optical zoom

Super God patent exposures, iPhone 7 or dual camera offers optical zoom

Recent United States Patent Office also exposed a super God of Apple patents, the design might appear in the next generation of its flagship iPhone 7. Moschino Galaxy S4 Cases

This is a pair of rear camera on the patent, but this time the feature is no longer HTC and a stack of domestic machine, called "focus after the shoot" stunt features. Rumors for so many years, iPhone finally has the potential to provide optical zoom, but this time with the dual focal length lens + algorithm synthesis route.

Moschino Galaxy S4 Cases

And crop photos of different digital zoom just algorithms, optical zoom can be lossless picture bigger, by moving optical lens, that can provide different focal length of the camera's field of vision. But lens structure is complex and bulky, so rarely on the phone using the optical zoom. But Apple's patent with dual lens "understatement" solves the problem of optical zoom and highlight the camera.

Super God patent exposures, iPhone 7 or dual camera offers optical zoom

"Double lens" the principle is very simple but very clever: these two are the main camera, but comes with different lens focal lengths, short (wide angle) lens to take a wide angle photo of the whole, and a telephoto (which can be roughly understood as a "telescope") lenses are responsible for to focus details photo shoot. Synthesis algorithm, is a focus of the last, non-destructive optical zoom magnification of the photos!

To the field of video, iPhone can shoot their own fun with picture-in-picture video. In addition, the dual-lens, and finally allow iPhone to video when you can shoot full pixels photo. Once this function is used in the iPhone 7, estimated the da an Android manufacturers a year ... ...

Super God patent exposures, iPhone 7 or dual camera offers optical zoom

via:appadvice

Tips Moschino Galaxy S4 Case

Super God patent exposures, iPhone 7 or dual camera offers optical zoom

Micro sweep sweep, author tips bar ~