WO2025011559A1 - Head size measurement method and apparatus, head grid generation method and apparatus, medium, and device - Google Patents
Head size measurement method and apparatus, head grid generation method and apparatus, medium, and device Download PDFInfo
- Publication number
- WO2025011559A1 WO2025011559A1 PCT/CN2024/104500 CN2024104500W WO2025011559A1 WO 2025011559 A1 WO2025011559 A1 WO 2025011559A1 CN 2024104500 W CN2024104500 W CN 2024104500W WO 2025011559 A1 WO2025011559 A1 WO 2025011559A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- head
- target object
- facial
- reference object
- circular reference
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/60—Analysis of geometric attributes
- G06T7/62—Analysis of geometric attributes of area, perimeter, diameter or volume
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/13—Edge detection
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
- G06V40/165—Detection; Localisation; Normalisation using facial parts and geometric relationships
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
Definitions
- the present disclosure relates to the field of artificial intelligence technology, and in particular to a head size measurement, head mesh generation method and device, computer storage medium and electronic equipment.
- the purpose of the present disclosure is to provide a head size measurement method, a 3D head mesh generation method and a device thereof, which at least to some extent overcome the problem of inaccurate head size determination in the related art.
- a head size measurement method comprising: acquiring a facial image of a target object, the facial image having a circular reference object at the forehead position of the target object; determining facial feature points of the target object; determining a facial local area including the circular reference object based on the facial feature points; extracting the circular reference object from the facial local area; and determining the head size of the target object based on the size of the circular reference object.
- determining the head size of the target object based on the size of the circular reference object comprises: determining the diameter of the circular reference object; determining the head size of the target object based on the diameter of the circular reference object; inch.
- the facial feature points include eye feature points
- the facial partial image including the circular reference object is a rectangular area including the circular reference object.
- extracting the circular reference object from the local facial area includes: using an image segmentation artificial neural network to segment the local facial area to obtain the circular reference object; using an edge detector to detect the edge of the circular reference object; using an ellipse fitting method to perform ellipse measurement on the edge of the circular reference object, thereby extracting the circular reference object.
- the image segmentation artificial neural network is U-Net or a fully convolutional network FCN, and/or the edge detector is a canny edge detector.
- the method further includes: dedistorting the facial image of the target object based on camera parameters of an image acquisition device, wherein the camera parameters include distortion coefficients and an intrinsic matrix, and the distortion coefficients include radial distortion coefficients and tangential distortion coefficients.
- the method further comprises: acquiring multi-view images of a preset chessboard at different angles and distances by the image acquisition device, thereby determining camera parameters of the image acquisition device.
- a 3D head mesh generation method comprising: acquiring multiple 2D face images of a target object, the 2D face images comprising a front face image, a right face side image, and a left face side image, the front face image having a circular reference object at a forehead position of the target object; reconstructing a 3D head shape of the target object based on the multiple 2D face images; determining a head size of the target object based on the front face image; and generating a 3D head mesh of the target object having the head size.
- determining the head size of the target object based on the frontal face image includes: determining the head size of the target object based on the frontal face image according to the above-mentioned head size measurement method.
- reconstructing the 3D head shape of the target object based on the multiple 2D face images includes:
- the encoder of the 3DMM based on the head shape and texture is used to reconstruct the 3D head shape of the target object based on the multiple 2D face images to obtain the reconstructed 3D head shape.
- the method further comprises: pre-training the head shape and texture based 3DMM based on training data.
- the pre-training of the 3DMM based on head shape and texture based on training data includes: generating a parameterized head based on the training data using a non-rigid iterative closest point (NICP) algorithm; aligning the parameterized head with a uniform size using a generalized Procter & Gamble analysis (GPA); and extracting principal components (PCs) from all parameterized head meshes using PCA to obtain the 3DMM based on head shape and texture.
- NCP non-rigid iterative closest point
- GPS generalized Procter & Gamble analysis
- PCs principal components
- a 3D head shape of the target object is reconstructed based on the multiple 2D face images using an encoder based on a 3DMM of head shape and texture
- obtaining the reconstructed 3D head shape comprises: inputting the multiple 2D face images into an artificial neural network encoder to predict 3D head mesh parameters, wherein the 3D head mesh parameters include 3DMM coefficients, lighting parameters, and camera parameters of the 3D head mesh, and minimizing the network loss function, wherein
- the network loss functions include: pixel-level loss Limg ( ⁇ ), perceptual identity loss Lide ( ⁇ ) of the reconstructed face image, facial edge loss Ledge ( ⁇ ) and projection feature point loss Llan ( ⁇ ).
- the method further comprises: extracting perceptual features used in perceptual identity loss calculation using a pre-trained face recognition network.
- a 3D head mesh use method is provided, which is applied to a 3D head mesh generated according to the above-mentioned 3D head mesh generation method, and includes: using the 3D head mesh for headwear design.
- the headgear includes headphones, glasses, a helmet, a mask, or a hat.
- a head size measurement device comprising: a facial image acquisition module, used to acquire a facial image of a target object, wherein the facial image has a circular reference object at the forehead position of the target object; a facial feature point determination module, used to determine the facial feature points of the target object; a local area determination module, used to determine a facial local area including the circular reference object based on the facial feature points; a reference object advance module, used to extract the circular reference object from the facial local area; and a head size determination module, used to determine a diameter of the circular reference object, and determine the head size of the target object based on the diameter of the circular reference object.
- a 3D head mesh generation device including: a face image acquisition module, used to acquire multiple 2D face images of a target object, the 2D face images including a front face image, a right face side image, and a left face side image, the front face image having a circular reference object at the forehead position of the target object; a head shape reconstruction module, used to perform 3D head shape reconstruction on the target object based on the multiple 2D face images; a head size determination module, used to determine the head size of the target object based on the front face image; and a head mesh generation module, used to generate a 3D head mesh with the head size of the target object.
- an electronic device comprising: a processor; and a memory for storing executable instructions of the processor; wherein the processor is configured to perform the above-mentioned head size measurement method, or 3D head mesh generation method by executing the executable instructions.
- a computer-readable storage medium on which a computer program is stored.
- the computer program is executed by a processor, the head size measurement method or the 3D head mesh generation method described above is implemented.
- the head size measurement method and 3D head mesh generation method provided by the embodiments of the present disclosure determine the head size of the target object based on the size of a circular reference object, and have a high accuracy rate in size determination.
- FIG1 is a schematic diagram showing the structure of a computer system in one embodiment of the present disclosure
- FIG2 shows a flow chart of a method for measuring head size in one embodiment of the present disclosure
- FIG3 shows a schematic diagram of a square mark in one embodiment of the present disclosure
- FIG4 shows a schematic diagram of a circular mark in one embodiment of the present disclosure
- FIG5A shows a flow chart of a circular reference object extraction method according to an embodiment of the present disclosure
- FIG5B shows a schematic diagram of a circular reference object measurement channel in one embodiment of the present disclosure
- FIG6A shows a flow chart of a method for generating a D head mesh in one embodiment of the present disclosure
- FIG6B shows a main workflow diagram of a method for generating a 3D head mesh according to an embodiment of the present disclosure
- FIG7 is a schematic diagram of a 3DMM of a human head in one embodiment of the present disclosure.
- FIG8 is a schematic diagram showing a 3D head shape reconstruction framework in one embodiment of the present disclosure.
- FIG. 9 illustrates measurement evaluation of different circular reference objects in one embodiment
- FIG10 shows a comparison of qualitative results of two methods using circle or square markings in one embodiment
- Figure 11 shows the application scenario of head-related products
- FIG12 is a schematic diagram of a head size measuring device according to an embodiment of the present disclosure.
- FIG13 is a schematic diagram of a 3D head mesh generation device in one embodiment of the present disclosure.
- FIG. 14 is a block diagram showing a structure of a computer device in one embodiment of the present disclosure.
- the solution provided by the present application uses a circular object as a measurement reference, and performs 3D head mesh reconstruction based on the head size determined by the circular reference object, and applies it to ergonomic design.
- a circular object as a measurement reference
- 3D head mesh reconstruction based on the head size determined by the circular reference object
- Artificial Intelligence is the theory, method, technology and application system that uses digital computers or machines controlled by digital computers to simulate, extend and expand human intelligence, perceive the environment, acquire knowledge and use knowledge to obtain the best results.
- artificial intelligence is a comprehensive technology in computer science that attempts to understand the essence of intelligence and produce a new intelligent machine that can respond in a similar way to human intelligence.
- Artificial intelligence is to study the design principles and implementation methods of various intelligent machines, so that machines have the functions of perception, reasoning and decision-making.
- AI basic technologies generally include sensors, dedicated AI chips, cloud computing, distributed storage, big data processing technology, operation/interaction systems, mechatronics, etc.
- AI software technologies mainly include computer vision technology, speech processing technology, natural language processing technology, and machine learning/deep learning.
- Computer vision is a science that studies how to make machines "see”. To put it more specifically, it refers to the use of cameras and computers to replace human eyes to identify, track and measure targets, and further perform graphic processing so that the computer processing becomes an image that is more suitable for human eye observation or transmission to instrument detection.
- computer vision studies related theories and technologies, and attempts to establish an artificial intelligence system that can obtain information from images or multi-dimensional data.
- Computer vision technology usually includes image processing, image recognition, image semantic understanding, image retrieval, OCR (Optical Character Recognition), video processing, video semantic understanding, video content/behavior recognition, three-dimensional object reconstruction, 3D (3-Dimension) technology, virtual reality, augmented reality, simultaneous positioning and map construction, and other technologies, as well as common biometric recognition technologies such as face recognition and fingerprint recognition.
- OCR Optical Character Recognition
- video processing video semantic understanding, video content/behavior recognition
- three-dimensional object reconstruction 3D (3-Dimension) technology
- virtual reality augmented reality
- simultaneous positioning and map construction and other technologies, as well as common biometric recognition technologies such as face recognition and fingerprint recognition.
- Machine Learning is a multi-disciplinary subject that involves probability theory, statistics, approximation theory, convex analysis, algorithm complexity theory and other disciplines. It specializes in studying how computers simulate or implement human learning behavior to acquire new knowledge or skills and reorganize existing knowledge structures to continuously improve their performance.
- Machine learning is the core of artificial intelligence and the fundamental way to make computers intelligent. Its applications are spread across all areas of artificial intelligence.
- Machine learning and deep learning usually include artificial neural networks, belief networks, reinforcement learning, transfer learning, inductive learning, and self-learning.
- artificial intelligence technology has been studied and applied in many fields, such as common smart homes, smart wearable devices, virtual assistants, smart speakers, smart marketing, unmanned driving, automatic driving, drones, robots, smart medical care, smart customer service, etc. I believe that with the development of technology, artificial intelligence technology will be applied in more fields and play an increasingly important role.
- FIG1 is a schematic diagram showing the structure of a computer system provided in one embodiment of the present disclosure.
- the system includes: a plurality of terminals 120 and a server (or server cluster) 140 .
- the terminal 120 may be a mobile terminal such as a mobile phone, a game console, a tablet computer, an e-book reader, smart glasses, an MP4 (Moving Picture Experts Group Audio Layer IV) player, a smart home device, an AR (Augmented Reality) device, a VR (Virtual Reality) device, or the terminal 120 may be a personal computer (PC), such as a laptop computer and a desktop computer.
- a mobile terminal such as a mobile phone, a game console, a tablet computer, an e-book reader, smart glasses, an MP4 (Moving Picture Experts Group Audio Layer IV) player, a smart home device, an AR (Augmented Reality) device, a VR (Virtual Reality) device, or the terminal 120 may be a personal computer (PC), such as a laptop computer and a desktop computer.
- PC personal computer
- the terminal 120 may be installed with an application program for providing and implementing the method disclosed herein.
- the terminal 120 is connected to the server 140 via a communication network.
- the communication network is a wired network or a wireless network. network.
- the server 140 is a server, or is composed of several servers, or is a virtualization platform, or is a cloud computing service center.
- the server 140 is used to provide background services for the application program that provides ....
- the server 140 undertakes the main computing work, and the terminal 120 undertakes the secondary computing work; or, the server 140 undertakes the secondary computing work, and the terminal 120 undertakes the main computing work; or, the terminal 120 and the server 140 adopt a distributed computing architecture for collaborative computing.
- the server 140 is used to store information such as the 3D head model established through training.
- the client of the application installed in different terminals 120 is the same, or the client of the application installed on the two terminals 120 is the client of the same type of application of different control system platforms.
- the specific form of the client of the application can also be different, for example, the application client can be a mobile client, a PC client or a World Wide Web (Web) client, etc.
- the number of the terminals 120 may be more or less. For example, there may be only one terminal, or there may be dozens or hundreds of terminals, or a greater number. The embodiment of the present application does not limit the number and device type of the terminals.
- the system may further include a management device (not shown in FIG. 1 ), which is connected to the server 140 via a communication network.
- a management device (not shown in FIG. 1 ), which is connected to the server 140 via a communication network.
- the communication network is a wired network or a wireless network.
- the above-mentioned wireless network or wired network uses standard communication technology and/or protocol.
- the network is usually the Internet, but it can also be any network, including but not limited to Local Area Network (LAN), Metropolitan Area Network (MAN), Wide Area Network (WAN), mobile, wired or wireless network, any combination of private network or virtual private network).
- technologies and/or formats including Hypertext Markup Language (HTML), Extensible Markup Language (XML), etc. are used to represent data exchanged through the network.
- conventional encryption technologies such as Secure Socket Layer (SSL), Transport Layer Security (TLS), Virtual Private Network (VPN), Internet Protocol Security (IPsec) can also be used to encrypt all or some links.
- SSL Secure Socket Layer
- TLS Transport Layer Security
- VPN Virtual Private Network
- IPsec Internet Protocol Security
- customized and/or dedicated data communication technologies may be used to replace or supplement the above-mentioned data communication technologies.
- FIG2 shows a flow chart of a head size measurement method in an embodiment of the present disclosure.
- the method provided in the embodiment of the present disclosure can be executed by any electronic device with computing and processing capabilities, such as the terminal 120 and/or the server 140 in FIG1 .
- the server 140 is used as the execution subject for example description.
- a facial image of a target object is acquired, and the facial image has a circular reference object at the forehead position of the target object.
- the target object is a user.
- the facial image uses a frontal facial image of the user.
- the user maintains a neutral expression (open eyes, close mouth) to avoid any expression effect in the 3D head reconstruction.
- the forehead position may include the middle position of the forehead.
- the circular reference object is a printed circular mark, for example, a black circular mark printed on white paper.
- the circular reference object may also be a reference object similar to a coin.
- the color of the circular reference object is not limited to black, and other colors that are easily distinguished from the skin color may also be used.
- the facial feature points of the target object include eye feature points, eyebrow feature points, or both.
- Eye feature points may include eye corner feature points and eyeball feature points; eyebrow feature points may be the center of the eyebrow.
- facial feature points may include nose feature points,
- determining a facial local area including a circular reference object based on facial feature points Based on the determined facial feature points, the position range of the circular reference object on the target object's face can be determined, thereby determining a facial local area including the circular reference object, which can be a rectangle or a square.
- the circular reference object can be extracted from the local facial region by an image segmentation network based on an artificial neural network, for example, a U-Net network or an FCN network based on a CNN network.
- the local facial region can be determined more accurately by facial feature points, and extracting a circular reference object from the local facial region can improve the accuracy and speed of processing.
- the circular reference object segmentation based on U-Net includes the following steps: Conv+Pooling downsampling the local facial area image; then Deconv deconvolution upsampling, cropping the previous low-level feature map, and performing feature fusion; then upsampling again. Repeat the above process until a feature map of a predetermined output size is obtained, and finally the output segment map is obtained through softmax.
- the advantages of the U-Net network include that only a small sample training set is required for training, it is a relatively small segmentation network, and the structure is simple. Through the previous local facial area segmentation, it can better adapt to the situation of U-net, and the segmentation accuracy is high.
- the head size of the target object can be determined based on the extracted size of the circular reference object.
- the size of the circular reference object can be the diameter, circumference, or area of the circular reference object.
- the local facial area is determined by facial feature points
- a circular reference object is extracted from the local facial area
- the head size of the target object is determined based on the size of the circular reference object, with high accuracy.
- the circular reference object is more suitable for the head area, has less distortion caused by deformation, and improves the measurement accuracy. Placing the circular reference object in the forehead area can reduce the deformation of the circular reference object.
- a checkerboard mark template (see, for example, FIG. 3 ) is printed for camera parameter measurement.
- a black checkerboard mark can be printed on A4 paper.
- OpenCV is used to calibrate the camera and calculate the camera parameters (i.e. distortion coefficients and intrinsic matrix) and undistort the captured image.
- the coefficients in this matrix include focal length (fx, fy) and optical center (cx, cy).
- a mobile phone is used as a representative tool to capture multi-view images (e.g., width and height: 3024 ⁇ 4025 pixels).
- a pre-designed chessboard (10 ⁇ 7 squares and 9 ⁇ 6 inner corners) with a total size of 210 ⁇ 297 mm is printed and then placed on a flat surface (e.g., a table).
- the size of all squares in the chessboard is the same (each square is 26.7 mm), and the size of the square is an important parameter for determining the actual distance in camera calibration.
- chessboard images are taken at different angles and distances, for example more than 10 images, to provide various square angles with different viewing angles for camera calibration.
- camera calibration is performed using, for example, OpenCV to calculate camera parameters (i.e., distortion coefficients and intrinsic matrices), and the captured images can be dedistorted according to the obtained camera parameters.
- FIG. 4 Use a black and white printer with A4 paper to print pre-designed circular markers for head size measurement.
- the printed template is shown in Figure 4.
- 54 circular markers can be cut from one sheet of printed paper.
- the circular markers can be attached to the forehead of the participant, etc., where these areas have a high degree of flatness.
- 3D head dimensions can be measured from 2D images using a black circle marker with a constant diameter (e.g., 15 mm) as a reference object.
- the target subject (participant) is asked to place this circle at a location on their forehead, such as the center of the forehead, which is relatively flatter than other facial areas.
- a frontal facial image with the marker attached to the participant's forehead is captured.
- two face images without the marker are captured from the left and right (e.g., from 10°–50°).
- the participants are instructed to maintain a neutral expression (eyes open, mouth closed) to avoid any expression effects in the 3D head reconstruction.
- an image dedistortion method can be applied to it based on the camera parameters determined previously.
- FIG. 5A shows a flow chart of a circular reference object extraction method in an embodiment of the present disclosure
- FIG. 5B shows a schematic diagram of a corresponding circular reference object measurement channel.
- a rectangular region is cropped on the facial image of the target object based on the detected facial feature points.
- the face region is cropped from the captured face image of the target object using a pre-trained artificial neural network (e.g., FAN) to detect facial feature points (e.g., the white points shown in FIG. 5B(a)). Based on the facial feature points, a facial rectangular area (for example, the rectangular area shown in FIG. 5B (a)) is determined, thereby extracting a local facial area for cropping.
- the cropped image is shown in FIG. 5B (b), for example, with a size of 256 ⁇ 256 pixels.
- an edge detector such as a canny edge detector, to filter the edges of the circular reference object to obtain the edges of the circular reference object, as shown in FIG. 5B(d) .
- the facial local area can be determined relatively accurately by using facial feature points, and the circular reference object is extracted from the facial local area, which can improve the accuracy and speed of processing.
- FIG. 6A shows a flowchart of a method for generating a 3D head mesh in one embodiment of the present disclosure.
- multiple 2D face images of the target object are obtained, the 2D face images including a front face image, a right face side image, and a left face side image, the front face image having a circular reference object at the forehead position of the target object.
- S604 reconstructing a 3D head shape of the target object based on multiple 2D face images.
- determining the head size of the target object based on the front face image Determine the head size of the target object based on the size of the circular reference object in the face proof image.
- S608 Generate a 3D head mesh with head size of the target object.
- the head size of the target object is determined based on the circular reference object in the frontal face image, thereby generating a 3D head mesh with accurate head size, which can meet high-precision size requirements and is applied to applications such as headwear design that have high requirements on head model size.
- FIG6B shows a main workflow diagram of a method for generating a 3D head mesh according to an embodiment of the present disclosure. As shown in FIG6B , the workflow consists of four parts:
- the 3D head shape reconstruction is described in detail below.
- 3D head shape reconstruction is implemented based on 3DMM.
- 3DMM is the basis for subsequent 3D head shape reconstruction.
- a large number of head shape 3DMMs are created based on the existing training data set.
- a head texture 3DMM model is also created.
- a specific example of creating a 3DMM model includes the following steps:
- the training dataset can include a head shape training dataset and a head texture training dataset.
- PCs Principal components
- FIG7 A schematic diagram of the generated 3DMM model of a human head is shown in FIG7 , which includes a head shape and a head texture model.
- Fig. 8 shows a schematic diagram of a 3D head shape reconstruction framework in one embodiment of the present disclosure.
- a CNN-based encoder e.g., ResNet50
- ResNet50 a CNN-based encoder
- parameters including a 3DMM coefficient of a head mesh (including head shape 801 and head texture 802 parameters), and a corresponding number of lighting and camera parameters 803, 804, and a fitting face image 806, a fitting face landmark 807, 808 are generated through a differentiable renderer 805, and the loss is determined in the ground-truth face images 809, the ground-truth face edge heatmap 810, and the ground-truth face landmark 811.
- the main network loss functions that need to be minimized include pixel-level loss Limg ( ⁇ ), perceptual identity loss Lide ( ⁇ ) of rendered face images, facial edge loss Ledge ( ⁇ ) and projected landmark loss Llan ( ⁇ ); a pre-trained face recognition network (Arcface) is used to extract perceptual features in the perceptual identity loss calculation.
- FIG9 shows an embodiment of the measurement evaluation of different circular reference objects to evaluate the measurement of different circular reference objects.
- the measurement method can accurately detect coins of different sizes and colors, for example, 1 Hong Kong dollar (diameter: 25.5 mm); 0.5 Hong Kong dollar (diameter: 22.5 mm); 0.1 Hong Kong dollar (diameter: 17.5 mm).
- the measurement method of the coarse to fine circular reference objects disclosed in the present invention can accurately and reliably detect reference objects with different sizes and colors, such as coins, and the measurement method is accurate, robust, and convenient.
- the head reconstruction method disclosed in the present invention has a more consistent facial contour and better reconstruction accuracy.
- Figure 10 shows a qualitative comparison of the two methods using circular or square markings in one embodiment. As can be seen from Figure 10, the method of the present disclosure using circular markings has higher accuracy than using square markings.
- FIG11 shows an application scenario of head-related products, where the top portion represents a solid color reconstructed head and the bottom portion represents a reconstructed head with texture.
- the head model generated by the 3D head mesh reconstruction method disclosed in the present invention has higher dimensional accuracy and is more suitable for ergonomic design, such as headwear design, and can be used for personalized customization (personalization), virtual try-on, and size selection of face and head-related products.
- Application scenarios of head-related products include, for example: (a) headphones; (b) sunglasses; (c) VR glasses; (d) masks; (e) helmets.
- FIG12 is a schematic diagram of a head size measuring device in one embodiment of the present disclosure. As shown in FIG12 , the head size measuring device includes:
- a facial image acquisition module 1201 is used to acquire a facial image of a target object, wherein the facial image has a circular reference object at the forehead of the target object;
- a facial feature point determination module 1202 is used to determine the facial feature points of the target object
- a local area determination module configured to determine a facial local area including the circular reference object based on the facial feature points
- a reference object advance module 1203, configured to extract the circular reference object from the facial local area
- the head size determination module 1204 is used to determine the diameter of the circular reference object, and determine the head size of the target object based on the diameter of the circular reference object.
- FIG13 is a schematic diagram of a 3D head mesh generation device in one embodiment of the present disclosure. As shown in FIG13 , the 3D head mesh generation device includes:
- a face image acquisition module 1301 is used to acquire multiple 2D face images of a target object, wherein the 2D face images include a front face image, a right face side image, and a left face side image, and the front face image has a circular reference object at the forehead position of the target object;
- a head shape reconstruction module 1302 is used to reconstruct a 3D head shape of the target object based on the multiple 2D face images;
- a head size determination module 1303, configured to determine the head size of the target object based on the front face image
- the head mesh generation module 1304 is used to generate a 3D head mesh having the head size of the target object.
- the 3DMM-based 3D head shape reconstruction method uses a CNN-based encoder with higher accuracy.
- a new feature point to edge loss is introduced to ensure the overall reconstructed shape accuracy.
- the experimental results determine the size and cost-effective number of input face images, which is very useful for practical applications.
- a circular reference object is used instead of a square mark.
- the circular reference object is less sensitive to deformation when attached to the face, and is more stable and easy to measure the head size.
- the use of coin-shaped objects is also common in daily life.
- the accurate head size measurement method based on U-Net is more stable and convenient.
- the technical solution method of the present disclosure is more suitable for ergonomic design.
- the electronic device 1400 according to this embodiment of the present invention is described below with reference to Fig. 14.
- the electronic device 1400 shown in Fig. 14 is only an example and should not bring any limitation to the functions and application scope of the embodiment of the present invention.
- the electronic device 1400 is presented in the form of a general-purpose computing device.
- the components of the electronic device 1400 may include, but are not limited to: at least one processing unit 1410, at least one storage unit 1420, and a bus 1430 connecting different system components (including the storage unit 1420 and the processing unit 1410).
- the storage unit stores a program code, which can be executed by the processing unit 1410, so that the processing unit 1410 performs the steps according to various exemplary embodiments of the present invention described in the above "Exemplary Method" section of this specification.
- the processing unit 1410 can perform the steps shown in Figures 2 and 6A.
- the storage unit 1420 may include a readable medium in the form of a volatile storage unit, such as a random access storage unit (RAM) 14201 and/or a cache storage unit 14202 , and may further include a read-only storage unit (ROM) 14203 .
- RAM random access storage unit
- ROM read-only storage unit
- the storage unit 1420 may also include a program/utility 14204 having a set (at least one) of program modules 14205, such program modules 14205 including but not limited to: an operating system, one or more application programs, other program modules, and program data, each of which or some combination may include an implementation of a network environment.
- program modules 14205 including but not limited to: an operating system, one or more application programs, other program modules, and program data, each of which or some combination may include an implementation of a network environment.
- Bus 1430 may represent one or more of several types of bus structures, including a memory unit bus or memory unit controller, a peripheral bus, an accelerated graphics port, a processing unit, or a local bus using any of a variety of bus architectures.
- the electronic device 1400 may also communicate with one or more external devices 1500 (e.g., keyboards, pointing devices, Bluetooth devices, etc.), may also communicate with one or more devices that enable a user to interact with the electronic device 1500, and/or communicate with any device that enables the electronic device 1400 to communicate with one or more other computing devices (e.g., routers, modems, etc.). Such communication may be performed via an input/output (I/O) interface 1450.
- the electronic device 1400 may also communicate with one or more networks (e.g., a local area network (LAN), a wide area network (WAN), and/or a public network, such as the Internet) via a network adapter 1460.
- networks e.g., a local area network (LAN), a wide area network (WAN), and/or a public network, such as the Internet
- the network adapter 1460 communicates with other modules of the electronic device 1400 via a bus 1430. It should be understood that, although not shown in the figure, other hardware and/or software modules may be used in conjunction with the electronic device 1500, including but not limited to: microcode, device drivers, redundant processing units, external disk drive arrays, RAID systems, tape drives, and data backup storage systems.
- the technical solution according to the implementation of the present disclosure can be embodied in the form of a software product, which can be stored in a non-volatile storage medium (which can be a CD-ROM, a USB flash drive, a mobile hard disk, etc.) or on a network, including several instructions to enable a computing device (which can be a personal computer, a server, a terminal device, or a network device, etc.) to execute the method according to the implementation of the present disclosure.
- a non-volatile storage medium which can be a CD-ROM, a USB flash drive, a mobile hard disk, etc.
- a computing device which can be a personal computer, a server, a terminal device, or a network device, etc.
- a computer-readable storage medium is also provided, on which a program product capable of implementing the above method of the present specification is stored.
- various aspects of the present invention can also be implemented in the form of a program product, which includes a program code, and when the program product is run on a terminal device, the program code is used to enable the terminal device to execute the steps according to various exemplary embodiments of the present invention described in the above "Exemplary Method" section of the present specification.
- a program product for implementing the above method according to an embodiment of the present invention is described, which can adopt a portable compact disk read-only memory (CD-ROM) and include program code, and can be run on a terminal device, such as a personal computer.
- a readable storage medium can be any tangible medium containing or storing a program, which can be used by or in combination with an instruction execution system, an apparatus or a device.
- the program product may be in any combination of one or more readable media.
- the readable medium may be a readable signal medium or a readable storage medium.
- the readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device or device, or any combination thereof. More specific examples of readable storage media (a non-exhaustive list) include: an electrical connection with one or more wires, a portable disk, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact A compact disk read only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the above.
- Computer readable signal media may include data signals propagated in baseband or as part of a carrier wave, in which readable program code is carried. Such propagated data signals may take a variety of forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the above. Readable signal media may also be any readable medium other than a readable storage medium, which may send, propagate, or transmit a program for use by or in conjunction with an instruction execution system, apparatus, or device.
- the program code embodied on the readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wired, optical cable, RF, etc., or any suitable combination of the foregoing.
- Program code for performing the operations of the present invention may be written in any combination of one or more programming languages, including object-oriented programming languages such as Java, C++, etc., and conventional procedural programming languages such as "C" or similar programming languages.
- the program code may be executed entirely on the user computing device, partially on the user device, as a separate software package, partially on the user computing device and partially on a remote computing device, or entirely on a remote computing device or server.
- the remote computing device may be connected to the user computing device through any type of network, including a local area network (LAN) or a wide area network (WAN), or may be connected to an external computing device (e.g., through the Internet using an Internet service provider).
- LAN local area network
- WAN wide area network
- Internet service provider e.g., AT&T, MCI, Sprint, EarthLink, etc.
- the technical solution according to the implementation of the present disclosure can be embodied in the form of a software product, which can be stored in a non-volatile storage medium (which can be a CD-ROM, a USB flash drive, a mobile hard disk, etc.) or on a network, including several instructions to enable a computing device (which can be a personal computer, a server, a mobile terminal, or a network device, etc.) to execute the method according to the implementation of the present disclosure.
- a non-volatile storage medium which can be a CD-ROM, a USB flash drive, a mobile hard disk, etc.
- a computing device which can be a personal computer, a server, a mobile terminal, or a network device, etc.
- the present disclosure is applicable to the field of artificial intelligence technology, and is used to solve the problem of inaccurate head size determination during 3D head/face reconstruction in related technologies, so as to achieve the effect of accurately determining the head size of a target object.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Software Systems (AREA)
- Evolutionary Computation (AREA)
- Geometry (AREA)
- Human Computer Interaction (AREA)
- Computing Systems (AREA)
- Artificial Intelligence (AREA)
- Medical Informatics (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Computational Linguistics (AREA)
- Databases & Information Systems (AREA)
- Biomedical Technology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Data Mining & Analysis (AREA)
- Computer Graphics (AREA)
- Image Analysis (AREA)
- Length Measuring Devices By Optical Means (AREA)
Abstract
Description
相关申请的交叉引用CROSS-REFERENCE TO RELATED APPLICATIONS
本公开要求于2023年07月10日提交的申请号为2023108333083、名称为“头部尺寸测量、头部网格生成方法及装置、介质及设备”的中国专利申请的优先权,该中国专利申请的全部内容通过引用全部并入本文。The present disclosure claims priority to Chinese patent application numbered 2023108333083, filed on July 10, 2023, and entitled “Head size measurement, head grid generation method and device, medium and equipment”, the entire contents of which are incorporated herein by reference.
本公开涉及人工智能技术领域,尤其涉及一种头部尺寸测量、头部网格生成方法及装置、计算机存储介质及电子设备。The present disclosure relates to the field of artificial intelligence technology, and in particular to a head size measurement, head mesh generation method and device, computer storage medium and electronic equipment.
当前,商用3D(三维)扫描仪是最广泛使用的获取3D物理头部模型的方法。然而,它们价格昂贵,而且不方便,这限制了它们的受欢迎程度和可获得性(accessibility)。因此,从2D图像重建3D模型既便宜又方便,更适合公众。Currently, commercial 3D scanners are the most widely used method for obtaining 3D physical head models. However, they are expensive and inconvenient, which limits their popularity and accessibility. Therefore, reconstructing 3D models from 2D images is both cheap and convenient and more suitable for the public.
已有许多成熟的3D头部/面部重建方法,它们主要集中在对物理尺寸精度要求不太高的应用中。相反,尺寸在人体工程学(ergonomic design)设计中非常重要。当前的方法无法导出具有足够的尺寸精度的3D头部/面部模型。There are many mature 3D head/face reconstruction methods, which mainly focus on applications where physical size accuracy is not very important. On the contrary, size is very important in ergonomic design. Current methods cannot derive 3D head/face models with sufficient size accuracy.
需要说明的是,在上述背景技术部分公开的信息仅用于加强对本公开的背景的理解,因此可以包括不构成对本领域普通技术人员已知的现有技术的信息。It should be noted that the information disclosed in the above background technology section is only used to enhance the understanding of the background of the present disclosure, and therefore may include information that does not constitute the prior art known to ordinary technicians in the field.
发明内容Summary of the invention
本公开的目的在于提供一种头部尺寸测量方法、3D头部网格生成方法及其装置,至少在一定程度上克服由于相关技术中头部尺寸确定不准确的问题。The purpose of the present disclosure is to provide a head size measurement method, a 3D head mesh generation method and a device thereof, which at least to some extent overcome the problem of inaccurate head size determination in the related art.
本公开的其他特性和优点将通过下面的详细描述变得显然,或部分地通过本公开的实践而习得。Other features and advantages of the present disclosure will become apparent from the following detailed description, or may be learned in part by the practice of the present disclosure.
根据本公开的一个方面,提供一种头部尺寸测量方法,包括:获取目标对象的面部图像,所述面部图像在所述目标对象的额头位置有圆形参考物体;确定所述目标对象的面部特征点;基于所述面部特征点确定包括所述圆形参考物体的面部局部区域;从所述面部局部区域中提取所述圆形参考物体;基于所述圆形参考物体的尺寸确定所述目标对象的头部尺寸。According to one aspect of the present disclosure, a head size measurement method is provided, comprising: acquiring a facial image of a target object, the facial image having a circular reference object at the forehead position of the target object; determining facial feature points of the target object; determining a facial local area including the circular reference object based on the facial feature points; extracting the circular reference object from the facial local area; and determining the head size of the target object based on the size of the circular reference object.
在一个实施例中,基于所述圆形参考物体的尺寸确定所述目标对象的头部尺寸包括:确定所述圆形参考物体的直径;基于所述圆形参考物体的直径确定所述目标对象的头部尺 寸。In one embodiment, determining the head size of the target object based on the size of the circular reference object comprises: determining the diameter of the circular reference object; determining the head size of the target object based on the diameter of the circular reference object; inch.
在一个实施例中,所述面部特征点包括眼部特征点,所述包括所述圆形参考物体的面部局部图像为包括所述圆形参考物体的矩形区域。In one embodiment, the facial feature points include eye feature points, and the facial partial image including the circular reference object is a rectangular area including the circular reference object.
在一个实施例中,所述从所述面部局部区域中提取所述圆形参考物体包括:使用图像分割人工神经网络对所述面部局部区域进行分割获得所述圆形参考物体;使用边缘检测器检测所述圆形参考物体的边缘;使用椭圆拟合方法对所述圆形参考物体的边缘进行椭圆测量,从而提取出所述圆形参考物体。In one embodiment, extracting the circular reference object from the local facial area includes: using an image segmentation artificial neural network to segment the local facial area to obtain the circular reference object; using an edge detector to detect the edge of the circular reference object; using an ellipse fitting method to perform ellipse measurement on the edge of the circular reference object, thereby extracting the circular reference object.
在一个实施例中,所述图像分割人工神经网络为U-Net、或全卷积网络FCN,和/或所述边缘检测器为canny边缘检测器。In one embodiment, the image segmentation artificial neural network is U-Net or a fully convolutional network FCN, and/or the edge detector is a canny edge detector.
在一个实施例中,该方法还包括:基于图像采集设备的相机参数对所述目标对象的面部图像进行去失真处理,其中,所述相机参数包括畸变系数和本征矩阵,所述畸变系数包括径向畸变系数和切向畸变系数。In one embodiment, the method further includes: dedistorting the facial image of the target object based on camera parameters of an image acquisition device, wherein the camera parameters include distortion coefficients and an intrinsic matrix, and the distortion coefficients include radial distortion coefficients and tangential distortion coefficients.
在一个实施例中,该方法还包括:通过所述图像采集设备以不同角度和距离采集预置棋盘的多视角图像,从而确定所述图像采集设备的相机参数。In one embodiment, the method further comprises: acquiring multi-view images of a preset chessboard at different angles and distances by the image acquisition device, thereby determining camera parameters of the image acquisition device.
根据本公开的另一方面,提供一种3D头部网格生成方法,包括:获取目标对象的多张2D人脸图像,所述2D人脸图像包括人脸正面图像和右侧人脸侧面图像和左侧人脸侧面图像,所述人脸正面图像在所述目标对象的额头位置有圆形参考物体;基于所述多张2D人脸图像对所述目标对象进行3D头部形状重建;基于所述人脸正面图像确定所述目标对象的头部尺寸;生成所述目标对象的具有头部尺寸的3D头部网格。According to another aspect of the present disclosure, a 3D head mesh generation method is provided, comprising: acquiring multiple 2D face images of a target object, the 2D face images comprising a front face image, a right face side image, and a left face side image, the front face image having a circular reference object at a forehead position of the target object; reconstructing a 3D head shape of the target object based on the multiple 2D face images; determining a head size of the target object based on the front face image; and generating a 3D head mesh of the target object having the head size.
在一个实施例中,所述基于所述人脸正面图像确定所述目标对象的头部尺寸包括:根据上述的头部尺寸测量方法基于所述人脸正面图像确定所述目标对象的头部尺寸。In one embodiment, determining the head size of the target object based on the frontal face image includes: determining the head size of the target object based on the frontal face image according to the above-mentioned head size measurement method.
在一个实施例中,所述基于所述多张2D人脸图像对所述目标对象进行3D头部形状重建包括:In one embodiment, reconstructing the 3D head shape of the target object based on the multiple 2D face images includes:
使用基于头部形状和纹理的3DMM的编码器基于所述多张2D人脸图像对所述目标对象进行3D头部形状重建,获得重建后的3D头部形状。The encoder of the 3DMM based on the head shape and texture is used to reconstruct the 3D head shape of the target object based on the multiple 2D face images to obtain the reconstructed 3D head shape.
在一个实施例中,该方法还包括:基于训练数据预先训练所述基于头部形状和纹理的3DMM。In one embodiment, the method further comprises: pre-training the head shape and texture based 3DMM based on training data.
在一个实施例中,所述基于训练数据预先训练所述基于头部形状和纹理的3DMM包括:基于所述训练数据使用非刚性迭代最近点(NICP)算法生成参数化头部;使用泛用型普氏(普鲁克)分析(GPA)将所述参数化头部与统一尺寸对齐;使用PCA从所有参数化头部网格中提取主成分(PC),从而获得所述基于头部形状和纹理的3DMM。In one embodiment, the pre-training of the 3DMM based on head shape and texture based on training data includes: generating a parameterized head based on the training data using a non-rigid iterative closest point (NICP) algorithm; aligning the parameterized head with a uniform size using a generalized Procter & Gamble analysis (GPA); and extracting principal components (PCs) from all parameterized head meshes using PCA to obtain the 3DMM based on head shape and texture.
在一个实施例中,使用基于头部形状和纹理的3DMM的编码器基于所述多张2D人脸图像对所述目标对象进行3D头部形状重建,获得重建后的3D头部形状包括:将所述多张2D人脸图像输入人工神经网络编码器以预测3D头部网格参数,所述3D头部网格参数包括3D头部网格的3DMM系数、照明参数和相机参数,以最小化的网络损失函数,所述 网络损失函数包括:像素级损失Limg(ξ)、重建的人脸图像的感知同一性损失Lide(ξ)、面部边缘损失Ledge(ξ)和投影特征点损失Llan(ξ)。In one embodiment, a 3D head shape of the target object is reconstructed based on the multiple 2D face images using an encoder based on a 3DMM of head shape and texture, and obtaining the reconstructed 3D head shape comprises: inputting the multiple 2D face images into an artificial neural network encoder to predict 3D head mesh parameters, wherein the 3D head mesh parameters include 3DMM coefficients, lighting parameters, and camera parameters of the 3D head mesh, and minimizing the network loss function, wherein The network loss functions include: pixel-level loss Limg (ξ), perceptual identity loss Lide (ξ) of the reconstructed face image, facial edge loss Ledge (ξ) and projection feature point loss Llan (ξ).
在一个实施例中,该方法还包括:使用预训练的人脸识别网络提取感知身份损失计算中使用的感知特征。In one embodiment, the method further comprises: extracting perceptual features used in perceptual identity loss calculation using a pre-trained face recognition network.
根据本公开的又一方面,提供一种3D头部网格使用方法,应用于根据上述的3D头部网格生成方法生成的3D头部网格,包括:将所述3D头部网格用于头饰设计。According to another aspect of the present disclosure, a 3D head mesh use method is provided, which is applied to a 3D head mesh generated according to the above-mentioned 3D head mesh generation method, and includes: using the 3D head mesh for headwear design.
在一个实施例中,所述头饰包括耳机、眼睛、头盔、口罩、或帽子。In one embodiment, the headgear includes headphones, glasses, a helmet, a mask, or a hat.
根据本公开的再一方面,提供一种头部尺寸测量装置,包括:面部图像获取模块,用于获取目标对象的面部图像,所述面部图像在所述目标对象的额头位置有圆形参考物体;面部特征点确定模块,用于确定所述目标对象的面部特征点;局部区域确定模块,用于基于所述面部特征点确定包括所述圆形参考物体的面部局部区域;参考物体提前模块,用于从所述面部局部区域中提取所述圆形参考物体;和头部尺寸确定模块,用于确定所述圆形参考物体的直径,基于所述圆形参考物体的直径确定所述目标对象的头部尺寸。According to another aspect of the present disclosure, a head size measurement device is provided, comprising: a facial image acquisition module, used to acquire a facial image of a target object, wherein the facial image has a circular reference object at the forehead position of the target object; a facial feature point determination module, used to determine the facial feature points of the target object; a local area determination module, used to determine a facial local area including the circular reference object based on the facial feature points; a reference object advance module, used to extract the circular reference object from the facial local area; and a head size determination module, used to determine a diameter of the circular reference object, and determine the head size of the target object based on the diameter of the circular reference object.
根据本公开的再一方面,提供一种3D头部网格生成装置,包括:人脸图像获取模块,用于获取目标对象的多张2D人脸图像,所述2D人脸图像包括人脸正面图像和右侧人脸侧面图像和左侧人脸侧面图像,所述人脸正面图像在所述目标对象的额头位置有圆形参考物体;头部形状重建模块,用于基于所述多张2D人脸图像对所述目标对象进行3D头部形状重建;头部尺寸确定模块,用于基于所述人脸正面图像确定所述目标对象的头部尺寸;头部网格生成模块,用于生成所述目标对象的具有头部尺寸的3D头部网格。According to another aspect of the present disclosure, a 3D head mesh generation device is provided, including: a face image acquisition module, used to acquire multiple 2D face images of a target object, the 2D face images including a front face image, a right face side image, and a left face side image, the front face image having a circular reference object at the forehead position of the target object; a head shape reconstruction module, used to perform 3D head shape reconstruction on the target object based on the multiple 2D face images; a head size determination module, used to determine the head size of the target object based on the front face image; and a head mesh generation module, used to generate a 3D head mesh with the head size of the target object.
根据本公开的再一个方面,提供一种如电子设备,包括:处理器;以及存储器,用于存储所述处理器的可执行指令;其中,所述处理器配置为经由执行所述可执行指令来执行上述的头部尺寸测量方法,或3D头部网格生成方法。According to another aspect of the present disclosure, there is provided an electronic device, comprising: a processor; and a memory for storing executable instructions of the processor; wherein the processor is configured to perform the above-mentioned head size measurement method, or 3D head mesh generation method by executing the executable instructions.
根据本公开的又一个方面,提供一种计算机可读存储介质,其上存储有计算机程序,所述计算机程序被处理器执行时实现上述的头部尺寸测量方法,或3D头部网格生成方法。According to another aspect of the present disclosure, a computer-readable storage medium is provided, on which a computer program is stored. When the computer program is executed by a processor, the head size measurement method or the 3D head mesh generation method described above is implemented.
本公开的实施例所提供的头部尺寸测量方法、3D头部网格生成方法,基于圆形参考物体的尺寸确定目标对象的头部尺寸,尺寸确定准确率高。The head size measurement method and 3D head mesh generation method provided by the embodiments of the present disclosure determine the head size of the target object based on the size of a circular reference object, and have a high accuracy rate in size determination.
应当理解的是,以上的一般描述和后文的细节描述仅是示例性和解释性的,并不能限制本公开。It is to be understood that the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the present disclosure.
此处的附图被并入说明书中并构成本说明书的一部分,示出了符合本公开的实施例,并与说明书一起用于解释本公开的原理。显而易见地,下面描述中的附图仅仅是本公开的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。The accompanying drawings herein are incorporated into the specification and constitute a part of the specification, illustrate embodiments consistent with the present disclosure, and together with the specification are used to explain the principles of the present disclosure. Obviously, the accompanying drawings described below are only some embodiments of the present disclosure, and for ordinary technicians in this field, other accompanying drawings can be obtained based on these accompanying drawings without creative work.
图1示出本公开一个实施例中计算机系统的结构示意图; FIG1 is a schematic diagram showing the structure of a computer system in one embodiment of the present disclosure;
图2示出本公开一个实施例中头部尺寸测量方法流程图;FIG2 shows a flow chart of a method for measuring head size in one embodiment of the present disclosure;
图3示出本公开一个实施例中方形标记示意图;FIG3 shows a schematic diagram of a square mark in one embodiment of the present disclosure;
图4示出本公开一个实施例中圆形标记示意图;FIG4 shows a schematic diagram of a circular mark in one embodiment of the present disclosure;
图5A示出本公开实施例中一种圆形参考物体提取方法流程图;FIG5A shows a flow chart of a circular reference object extraction method according to an embodiment of the present disclosure;
图5B示出本公开一个实施例中圆形参考物体测量通道示意图;FIG5B shows a schematic diagram of a circular reference object measurement channel in one embodiment of the present disclosure;
图6A示出本公开一个实施例中D头部网格生成方法的流程图;FIG6A shows a flow chart of a method for generating a D head mesh in one embodiment of the present disclosure;
图6B示出本公开一个实施例的生成3D头部网格方法的主要工作流程图;FIG6B shows a main workflow diagram of a method for generating a 3D head mesh according to an embodiment of the present disclosure;
图7示出本公开一个实施例中人类头部的3DMM的示意图;FIG7 is a schematic diagram of a 3DMM of a human head in one embodiment of the present disclosure;
图8示出本公开一个实施例中3D头部形状重建框架的示意图;FIG8 is a schematic diagram showing a 3D head shape reconstruction framework in one embodiment of the present disclosure;
图9示出一个实施例中不同圆形参考物体的测量评估;FIG. 9 illustrates measurement evaluation of different circular reference objects in one embodiment;
图10示出一个实施例中使用圆形或方形标记两种方法的定性结果比较;FIG10 shows a comparison of qualitative results of two methods using circle or square markings in one embodiment;
图11示出头部相关产品应用场景;Figure 11 shows the application scenario of head-related products;
图12示出本公开一个实施例中头部尺寸测量装置示意图;FIG12 is a schematic diagram of a head size measuring device according to an embodiment of the present disclosure;
图13示出本公开一个实施例中3D头部网格生成装置示意图;和FIG13 is a schematic diagram of a 3D head mesh generation device in one embodiment of the present disclosure; and
图14示出本公开一个实施例中计算机设备的结构框图。FIG. 14 is a block diagram showing a structure of a computer device in one embodiment of the present disclosure.
现在将参考附图更全面地描述示例实施方式。然而,示例实施方式能够以多种形式实施,且不应被理解为限于在此阐述的范例;相反,提供这些实施方式使得本公开将更加全面和完整,并将示例实施方式的构思全面地传达给本领域的技术人员。所描述的特征、结构或特性可以以任何合适的方式结合在一个或更多实施方式中。Example embodiments will now be described more fully with reference to the accompanying drawings. However, example embodiments can be implemented in a variety of forms and should not be construed as limited to the examples set forth herein; rather, these embodiments are provided so that the disclosure will be more comprehensive and complete and to fully convey the concepts of the example embodiments to those skilled in the art. The described features, structures, or characteristics may be combined in any suitable manner in one or more embodiments.
此外,附图仅为本公开的示意性图解,并非一定是按比例绘制。图中相同的附图标记表示相同或类似的部分,因而将省略对它们的重复描述。附图中所示的一些方框图是功能实体,不一定必须与物理或逻辑上独立的实体相对应。可以采用软件形式来实现这些功能实体,或在一个或多个硬件模块或集成电路中实现这些功能实体,或在不同网络和/或处理器装置和/或微控制器装置中实现这些功能实体。In addition, the accompanying drawings are only schematic illustrations of the present disclosure and are not necessarily drawn to scale. The same reference numerals in the figures represent the same or similar parts, and their repeated description will be omitted. Some of the block diagrams shown in the accompanying drawings are functional entities and do not necessarily correspond to physically or logically independent entities. These functional entities can be implemented in software form, or implemented in one or more hardware modules or integrated circuits, or implemented in different networks and/or processor devices and/or microcontroller devices.
本申请提供的方案,使用圆形物体作为测量参考,并基于该圆形参考物体确定的头部尺寸进行3D头部网格重建,并应用于人体工程学设计。为了便于理解,下面首先对本申请涉及到的几个名词进行解释。The solution provided by the present application uses a circular object as a measurement reference, and performs 3D head mesh reconstruction based on the head size determined by the circular reference object, and applies it to ergonomic design. For ease of understanding, several terms involved in the present application are first explained below.
人工智能(Artificial Intelligence,AI)是利用数字计算机或者数字计算机控制的机器模拟、延伸和扩展人的智能,感知环境、获取知识并使用知识获得最佳结果的理论、方法、技术及应用系统。换句话说,人工智能是计算机科学的一个综合技术,它企图了解智能的实质,并生产出一种新的能以人类智能相似的方式做出反应的智能机器。人工智能也就是研究各种智能机器的设计原理与实现方法,使机器具有感知、推理与决策的功能。Artificial Intelligence (AI) is the theory, method, technology and application system that uses digital computers or machines controlled by digital computers to simulate, extend and expand human intelligence, perceive the environment, acquire knowledge and use knowledge to obtain the best results. In other words, artificial intelligence is a comprehensive technology in computer science that attempts to understand the essence of intelligence and produce a new intelligent machine that can respond in a similar way to human intelligence. Artificial intelligence is to study the design principles and implementation methods of various intelligent machines, so that machines have the functions of perception, reasoning and decision-making.
人工智能技术是一门综合学科,涉及领域广泛,既有硬件层面的技术也有软件层面的 技术。人工智能基础技术一般包括如传感器、专用人工智能芯片、云计算、分布式存储、大数据处理技术、操作/交互系统、机电一体化等技术。人工智能软件技术主要包括计算机视觉技术、语音处理技术、自然语言处理技术以及机器学习/深度学习等几大方向。Artificial intelligence technology is a comprehensive discipline that covers a wide range of fields, including both hardware-level technology and software-level technology. Technology. AI basic technologies generally include sensors, dedicated AI chips, cloud computing, distributed storage, big data processing technology, operation/interaction systems, mechatronics, etc. AI software technologies mainly include computer vision technology, speech processing technology, natural language processing technology, and machine learning/deep learning.
计算机视觉技术(Computer Vision,CV)计算机视觉是一门研究如何使机器“看”的科学,更进一步的说,就是指用摄影机和电脑代替人眼对目标进行识别、跟踪和测量等机器视觉,并进一步做图形处理,使电脑处理成为更适合人眼观察或传送给仪器检测的图像。作为一个科学学科,计算机视觉研究相关的理论和技术,试图建立能够从图像或者多维数据中获取信息的人工智能系统。计算机视觉技术通常包括图像处理、图像识别、图像语义理解、图像检索、OCR(Optical Character Recognition,光学字符识别)、视频处理、视频语义理解、视频内容/行为识别、三维物体重建、3D(3-Dimension,三维)技术、虚拟现实、增强现实、同步定位与地图构建等技术,还包括常见的人脸识别、指纹识别等生物特征识别技术。Computer Vision Technology (CV) Computer vision is a science that studies how to make machines "see". To put it more specifically, it refers to the use of cameras and computers to replace human eyes to identify, track and measure targets, and further perform graphic processing so that the computer processing becomes an image that is more suitable for human eye observation or transmission to instrument detection. As a scientific discipline, computer vision studies related theories and technologies, and attempts to establish an artificial intelligence system that can obtain information from images or multi-dimensional data. Computer vision technology usually includes image processing, image recognition, image semantic understanding, image retrieval, OCR (Optical Character Recognition), video processing, video semantic understanding, video content/behavior recognition, three-dimensional object reconstruction, 3D (3-Dimension) technology, virtual reality, augmented reality, simultaneous positioning and map construction, and other technologies, as well as common biometric recognition technologies such as face recognition and fingerprint recognition.
机器学习(Machine Learning,ML)是一门多领域交叉学科,涉及概率论、统计学、逼近论、凸分析、算法复杂度理论等多门学科。专门研究计算机怎样模拟或实现人类的学习行为,以获取新的知识或技能,重新组织已有的知识结构使之不断改善自身的性能。机器学习是人工智能的核心,是使计算机具有智能的根本途径,其应用遍及人工智能的各个领域。机器学习和深度学习通常包括人工神经网络、置信网络、强化学习、迁移学习、归纳学习、式教学习等技术。Machine Learning (ML) is a multi-disciplinary subject that involves probability theory, statistics, approximation theory, convex analysis, algorithm complexity theory and other disciplines. It specializes in studying how computers simulate or implement human learning behavior to acquire new knowledge or skills and reorganize existing knowledge structures to continuously improve their performance. Machine learning is the core of artificial intelligence and the fundamental way to make computers intelligent. Its applications are spread across all areas of artificial intelligence. Machine learning and deep learning usually include artificial neural networks, belief networks, reinforcement learning, transfer learning, inductive learning, and self-learning.
随着人工智能技术研究和进步,人工智能技术在多个领域展开研究和应用,例如常见的智能家居、智能穿戴设备、虚拟助理、智能音箱、智能营销、无人驾驶、自动驾驶、无人机、机器人、智能医疗、智能客服等,相信随着技术的发展,人工智能技术将在更多的领域得到应用,并发挥越来越重要的价值。With the research and advancement of artificial intelligence technology, artificial intelligence technology has been studied and applied in many fields, such as common smart homes, smart wearable devices, virtual assistants, smart speakers, smart marketing, unmanned driving, automatic driving, drones, robots, smart medical care, smart customer service, etc. I believe that with the development of technology, artificial intelligence technology will be applied in more fields and play an increasingly important role.
本申请实施例提供的方案涉及人工智能的计算机视觉、机器学习等技术,具体通过如下实施例进行说明:The solution provided in the embodiments of the present application involves artificial intelligence computer vision, machine learning and other technologies, which are specifically described by the following embodiments:
图1示出本公开一个实施例中提供的计算机系统的结构示意图。该系统包括:若干个终端120和服务器(或服务器集群)140。FIG1 is a schematic diagram showing the structure of a computer system provided in one embodiment of the present disclosure. The system includes: a plurality of terminals 120 and a server (or server cluster) 140 .
终端120可以是手机、游戏主机、平板电脑、电子书阅读器、智能眼镜、MP4(MovingPicture Experts Group Audio Layer IV,动态影像专家压缩标准音频层面4)播放器、智能家居设备、AR(Augmented Reality,增强现实)设备、VR(Virtual Reality,虚拟现实)设备等移动终端,或者,终端120也可以是个人计算机(Personal Computer,PC),比如膝上型便携计算机和台式计算机等等。The terminal 120 may be a mobile terminal such as a mobile phone, a game console, a tablet computer, an e-book reader, smart glasses, an MP4 (Moving Picture Experts Group Audio Layer IV) player, a smart home device, an AR (Augmented Reality) device, a VR (Virtual Reality) device, or the terminal 120 may be a personal computer (PC), such as a laptop computer and a desktop computer.
其中,终端120中可以安装有用于提供实现本公开的方法的应用程序。The terminal 120 may be installed with an application program for providing and implementing the method disclosed herein.
终端120与服务器140之间通过通信网络相连。可选的,通信网络是有线网络或无线 网络。The terminal 120 is connected to the server 140 via a communication network. Optionally, the communication network is a wired network or a wireless network. network.
服务器140是一台服务器,或者由若干台服务器组成,或者是一个虚拟化平台,或者是一个云计算服务中心。服务器140用于为提供….的应用程序提供后台服务。可选地,服务器140承担主要计算工作,终端120承担次要计算工作;或者,服务器140承担次要计算工作,终端120承担主要计算工作;或者,终端120和服务器140之间采用分布式计算架构进行协同计算。The server 140 is a server, or is composed of several servers, or is a virtualization platform, or is a cloud computing service center. The server 140 is used to provide background services for the application program that provides .... Optionally, the server 140 undertakes the main computing work, and the terminal 120 undertakes the secondary computing work; or, the server 140 undertakes the secondary computing work, and the terminal 120 undertakes the main computing work; or, the terminal 120 and the server 140 adopt a distributed computing architecture for collaborative computing.
在一些可选的实施例中,服务器140用于存储通过训练建立的3D头部模型等信息。In some optional embodiments, the server 140 is used to store information such as the 3D head model established through training.
可选地,不同的终端120中安装的应用程序的客户端是相同的,或两个终端120上安装的应用程序的客户端是不同控制系统平台的同一类型应用程序的客户端。基于终端平台的不同,该应用程序的客户端的具体形态也可以不同,比如,该应用程序客户端可以是手机客户端、PC客户端或者全球广域网(World Wide Web,Web)客户端等。Optionally, the client of the application installed in different terminals 120 is the same, or the client of the application installed on the two terminals 120 is the client of the same type of application of different control system platforms. Based on the different terminal platforms, the specific form of the client of the application can also be different, for example, the application client can be a mobile client, a PC client or a World Wide Web (Web) client, etc.
本领域技术人员可以知晓,上述终端120的数量可以更多或更少。比如上述终端可以仅为一个,或者上述终端为几十个或几百个,或者更多数量。本申请实施例对终端的数量和设备类型不加以限定。Those skilled in the art will appreciate that the number of the terminals 120 may be more or less. For example, there may be only one terminal, or there may be dozens or hundreds of terminals, or a greater number. The embodiment of the present application does not limit the number and device type of the terminals.
可选的,该系统还可以包括管理设备(图1未示出),该管理设备与服务器140之间通过通信网络相连。可选的,通信网络是有线网络或无线网络。Optionally, the system may further include a management device (not shown in FIG. 1 ), which is connected to the server 140 via a communication network. Optionally, the communication network is a wired network or a wireless network.
可选的,上述的无线网络或有线网络使用标准通信技术和/或协议。网络通常为因特网、但也可以是任何网络,包括但不限于局域网(Local Area Network,LAN)、城域网(Metropolitan Area Network,MAN)、广域网(Wide Area Network,WAN)、移动、有线或者无线网络、专用网络或者虚拟专用网络的任何组合)。在一些实施例中,使用包括超文本标记语言(Hyper Text Mark-up Language,HTML)、可扩展标记语言(Extensible MarkupLanguage,XML)等的技术和/或格式来代表通过网络交换的数据。此外还可以使用诸如安全套接字层(Secure Socket Layer,SSL)、传输层安全(Transport Layer Security,TLS)、虚拟专用网络(Virtual Private Network,VPN)、网际协议安全(Internet ProtocolSecurity,IPsec)等常规加密技术来加密所有或者一些链路。在另一些实施例中,还可以使用定制和/或专用数据通信技术取代或者补充上述数据通信技术。Optionally, the above-mentioned wireless network or wired network uses standard communication technology and/or protocol. The network is usually the Internet, but it can also be any network, including but not limited to Local Area Network (LAN), Metropolitan Area Network (MAN), Wide Area Network (WAN), mobile, wired or wireless network, any combination of private network or virtual private network). In some embodiments, technologies and/or formats including Hypertext Markup Language (HTML), Extensible Markup Language (XML), etc. are used to represent data exchanged through the network. In addition, conventional encryption technologies such as Secure Socket Layer (SSL), Transport Layer Security (TLS), Virtual Private Network (VPN), Internet Protocol Security (IPsec) can also be used to encrypt all or some links. In other embodiments, customized and/or dedicated data communication technologies may be used to replace or supplement the above-mentioned data communication technologies.
下面,将结合附图及实施例对本示例实施方式中的头部尺寸测量和重建方法的各个步骤进行更详细的说明。The following will describe in more detail the various steps of the head size measurement and reconstruction method in this example implementation with reference to the accompanying drawings and embodiments.
图2示出本公开实施例中一种头部尺寸测量方法流程图。本公开实施例提供的方法可以由任意具备计算处理能力的电子设备执行,例如如图1中的终端120和/或服务器140。在下面的举例说明中,以服务器140为执行主体进行示例说明。FIG2 shows a flow chart of a head size measurement method in an embodiment of the present disclosure. The method provided in the embodiment of the present disclosure can be executed by any electronic device with computing and processing capabilities, such as the terminal 120 and/or the server 140 in FIG1 . In the following example description, the server 140 is used as the execution subject for example description.
如图2所示,S202,获取目标对象的面部图像,该面部图像在所述目标对象的额头位置有圆形参考物体。目标对象指的是用户。在一个实施例中,面部图像使用用户的正面面部图像。在一个实施例中,在面部图像捕获过程中,用户保持中性表情(睁开眼睛,闭上嘴巴),以避免3D头部重建中的任何表情效果。额头位置可以包括额头中间位置。在一 个实施例中,圆形参考物体是打印的圆形标记,例如,在白色纸张上打印的黑色圆形标记。圆形参考物体也可以是类似硬币的参考物体,圆形参考物体的颜色不限于黑色,也可以采用其他易于与皮肤颜色相区分的颜色。As shown in FIG. 2 , S202, a facial image of a target object is acquired, and the facial image has a circular reference object at the forehead position of the target object. The target object is a user. In one embodiment, the facial image uses a frontal facial image of the user. In one embodiment, during the facial image capture process, the user maintains a neutral expression (open eyes, close mouth) to avoid any expression effect in the 3D head reconstruction. The forehead position may include the middle position of the forehead. In one embodiment, the circular reference object is a printed circular mark, for example, a black circular mark printed on white paper. The circular reference object may also be a reference object similar to a coin. The color of the circular reference object is not limited to black, and other colors that are easily distinguished from the skin color may also be used.
S204,从面部图像中确定目标对象的面部特征点。在一个实施例中,目标对象的面部特征点包括眼睛特征点、眼眉特征点或者两者。眼睛的特征点可以包括眼角特征点、眼珠特征点;眼眉特征点可以是眼眉的中心位置。在一个实施例中,面部特征点可以包括鼻子特征点、S204, determining facial feature points of the target object from the facial image. In one embodiment, the facial feature points of the target object include eye feature points, eyebrow feature points, or both. Eye feature points may include eye corner feature points and eyeball feature points; eyebrow feature points may be the center of the eyebrow. In one embodiment, facial feature points may include nose feature points,
S206,基于面部特征点确定包括圆形参考物体的面部局部区域。基于确定的面部特征点,可以确定圆形参考物体在目标对象面部的位置范围,从而确定包括圆形参考物体的面部局部区域,该面部局部区域可以是矩形,或者方形。S206, determining a facial local area including a circular reference object based on facial feature points. Based on the determined facial feature points, the position range of the circular reference object on the target object's face can be determined, thereby determining a facial local area including the circular reference object, which can be a rectangle or a square.
S208,从面部局部区域中提取圆形参考物体。可以通过基于人工神经网络的图像分割网络从面部局部区域中提取圆形参考物体,例如,基于CNN网络的U-Net网络或者FCN网络。通过面部特征点可以比较准确地确定面部局部区域,而从面部局部区域中提取圆形参考物体,可以提高处理的准确度和速度。S208, extracting a circular reference object from the local facial region. The circular reference object can be extracted from the local facial region by an image segmentation network based on an artificial neural network, for example, a U-Net network or an FCN network based on a CNN network. The local facial region can be determined more accurately by facial feature points, and extracting a circular reference object from the local facial region can improve the accuracy and speed of processing.
在一个实施例中,基于U-Net进行圆形参考物体分割包括如下步骤:对面部局部区域图像进行Conv+Pooling下采样;然后Deconv反卷积进行上采样,裁剪(crop)之前的低层特征映射(feature map),进行特征融合;然后再次上采样。重复上述过程,直到获得输出预定大小的特征映射(feature map),最后经过softmax获得输出切割映射(output segment map)。U-Net网络的优点包括只需要小样本的训练集进行训练,是比较小的分割网络,结构简单。通过前面的面部局部区域分割,可以较好地适应U-net的情况,分割准确率高。In one embodiment, the circular reference object segmentation based on U-Net includes the following steps: Conv+Pooling downsampling the local facial area image; then Deconv deconvolution upsampling, cropping the previous low-level feature map, and performing feature fusion; then upsampling again. Repeat the above process until a feature map of a predetermined output size is obtained, and finally the output segment map is obtained through softmax. The advantages of the U-Net network include that only a small sample training set is required for training, it is a relatively small segmentation network, and the structure is simple. Through the previous local facial area segmentation, it can better adapt to the situation of U-net, and the segmentation accuracy is high.
S210,基于圆形参考物体的尺寸确定所述目标对象的头部尺寸。可以基于提取的圆形参考物体的尺寸确定目标对象的头部尺寸。圆形参考物体的尺寸可以是圆形参考物体的直径,周长,或者面积。S210, determining the head size of the target object based on the size of the circular reference object. The head size of the target object can be determined based on the extracted size of the circular reference object. The size of the circular reference object can be the diameter, circumference, or area of the circular reference object.
上述实施例中,通过面部特征点确定面部局部区域,在面部局部区域中提取圆形参考物体,基于圆形参考物体的尺寸确定目标对象的头部尺寸,准确率高。圆形参考物体相对于其他形状的参考物体,更适应头部区域,对变形引起的畸变较小,提高了测量准确度。将圆形参考物体置于额头区域,可以减少圆形参考物体的变形。In the above embodiment, the local facial area is determined by facial feature points, a circular reference object is extracted from the local facial area, and the head size of the target object is determined based on the size of the circular reference object, with high accuracy. Compared with reference objects of other shapes, the circular reference object is more suitable for the head area, has less distortion caused by deformation, and improves the measurement accuracy. Placing the circular reference object in the forehead area can reduce the deformation of the circular reference object.
需要注意的是,上述附图仅是根据本发明示例性实施例的方法所包括的处理的示意性说明,而不是限制目的。易于理解,上述附图所示的处理并不表明或限制这些处理的时间顺序。另外,也易于理解,这些处理可以是例如在多个模块中同步或异步执行的。It should be noted that the above figures are only schematic illustrations of the processes included in the method according to an exemplary embodiment of the present invention, and are not intended to be limiting. It is easy to understand that the processes shown in the above figures do not indicate or limit the time sequence of these processes. In addition, it is also easy to understand that these processes can be performed synchronously or asynchronously, for example, in multiple modules.
在介绍人脸图像捕获之前,先介绍相机参数测量的一种实现方式。Before introducing face image capture, a method for measuring camera parameters is first introduced.
在面部图像捕获之前打印棋盘标记模板(例如参见图3所示),用于相机参数测量。例如,可以在A4纸上打印黑色的棋盘标记。Before capturing the facial image, a checkerboard mark template (see, for example, FIG. 3 ) is printed for camera parameter measurement. For example, a black checkerboard mark can be printed on A4 paper.
通过这些棋盘图像,利用OpenCV进行相机标定,计算相机参数(即畸变系数和本征 矩阵),并对捕获的图像进行不失真处理。Through these chessboard images, OpenCV is used to calibrate the camera and calculate the camera parameters (i.e. distortion coefficients and intrinsic matrix) and undistort the captured image.
相机参数测量Camera parameter measurement
一些相机给捕获的图像带来比较明显的失真,包括径向失真和切向失真。可以进行相机校准以确定失真系数(例如,五个失真参数),然后消除捕获的图像的失真。此外,还测量了相机的内参矩阵K,用于消除图像失真和3D头部重建的透视转换步骤,该矩阵中的系数包括焦距(fx,fy)和光学中心(cx,cy),
Some cameras introduce significant distortion to the captured images, including radial distortion and tangential distortion. Camera calibration can be performed to determine the distortion coefficients (e.g., five distortion parameters), and then the distortion of the captured images is removed. In addition, the camera's intrinsic parameter matrix K is measured, which is used for image distortion removal and perspective transformation steps of 3D head reconstruction. The coefficients in this matrix include focal length (fx, fy) and optical center (cx, cy).
在一个实施例中,手机被用作捕获多视图图像(例如,宽度和高度:3024×4025像素)的代表工具。打印预先设计的棋盘(10×7个正方形和9个×6个内角),总尺寸为210×297毫米,然后放在平面上(例如,桌子)。棋盘中所有正方形的尺寸都相同(每个正方形尺寸为26.7毫米),正方形的尺寸是确定相机校准中实际距离的重要参数。In one embodiment, a mobile phone is used as a representative tool to capture multi-view images (e.g., width and height: 3024×4025 pixels). A pre-designed chessboard (10×7 squares and 9×6 inner corners) with a total size of 210×297 mm is printed and then placed on a flat surface (e.g., a table). The size of all squares in the chessboard is the same (each square is 26.7 mm), and the size of the square is an important parameter for determining the actual distance in camera calibration.
以不同的角度和距离拍摄棋盘的多张图像,例如多于10张,为相机校准提供具有不同视角的各种方角。使用这些棋盘图像,使用例如OpenCV执行相机校准,以计算相机参数(即失真系数和固有矩阵),可以根据获得相机参数对捕获的图像消除失真。Multiple images of the chessboard are taken at different angles and distances, for example more than 10 images, to provide various square angles with different viewing angles for camera calibration. Using these chessboard images, camera calibration is performed using, for example, OpenCV to calculate camera parameters (i.e., distortion coefficients and intrinsic matrices), and the captured images can be dedistorted according to the obtained camera parameters.
下面介绍圆形参考物体的准备的例子。An example of preparation of a circular reference object is described below.
使用带有A4纸的黑白打印机来打印预先设计的圆形标记,以进行头部尺寸测量。打印的模板如图4所示。可以从一张打印纸上剪下例如54个圆形标记。可以将圆形标记将贴在参与者的额头等部位,这些部位区域具有较高的平整度。Use a black and white printer with A4 paper to print pre-designed circular markers for head size measurement. The printed template is shown in Figure 4. For example, 54 circular markers can be cut from one sheet of printed paper. The circular markers can be attached to the forehead of the participant, etc., where these areas have a high degree of flatness.
下面简单说明如何使用圆形参考物体。Here is a brief description of how to use the circular reference object.
可以使用直径恒定(例如,15mm)的黑色圆圈标记作为参考对象,从而通过2D图像测量3D头部尺寸。目标对象(参与者)被要求把这个圆圈放在他们的额头位置,例如额头中心,额头的中心比其他面部区域相对平坦。首先,捕获了贴在参与者额头上的标记的正面面部图像。然后,从左边和右边(例如从10°–50°)捕获两张没有标记的人脸图像。在面部图像捕获过程中,参与者被指示保持中性表情(睁开眼睛,闭上嘴巴),以避免3D头部重建中的任何表情效果。捕获人脸图像后,可以根据前面确定的相机参数对其应用图像去失真方法。3D head dimensions can be measured from 2D images using a black circle marker with a constant diameter (e.g., 15 mm) as a reference object. The target subject (participant) is asked to place this circle at a location on their forehead, such as the center of the forehead, which is relatively flatter than other facial areas. First, a frontal facial image with the marker attached to the participant's forehead is captured. Then, two face images without the marker are captured from the left and right (e.g., from 10°–50°). During the facial image capture process, the participants are instructed to maintain a neutral expression (eyes open, mouth closed) to avoid any expression effects in the 3D head reconstruction. After the facial image is captured, an image dedistortion method can be applied to it based on the camera parameters determined previously.
图5A示出本公开实施例中一种圆形参考物体提取方法流程图,图5B示出对应的圆形参考物体测量通道示意图。FIG. 5A shows a flow chart of a circular reference object extraction method in an embodiment of the present disclosure, and FIG. 5B shows a schematic diagram of a corresponding circular reference object measurement channel.
如图5A所示,S502,基于检测到的面部特征点对目标对象的面部图像进行矩形区域裁剪。As shown in FIG5A , S502 , a rectangular region is cropped on the facial image of the target object based on the detected facial feature points.
在人脸图像预处理中,通过使用预训练的人工神经网络(例如,FAN)从捕获的目标对象得到面部图像中裁剪人脸区域来检测人脸面部特征点(例如图5B(a)中示出的白点。 基于面部特征点确定面部矩形区域(例如,图5B(a)中示出的矩形区域,从而提取出面部局部区域进行裁剪,裁剪后的图像参见图5B(b),例如大小为256×256像素。In face image preprocessing, the face region is cropped from the captured face image of the target object using a pre-trained artificial neural network (e.g., FAN) to detect facial feature points (e.g., the white points shown in FIG. 5B(a)). Based on the facial feature points, a facial rectangular area (for example, the rectangular area shown in FIG. 5B (a)) is determined, thereby extracting a local facial area for cropping. The cropped image is shown in FIG. 5B (b), for example, with a size of 256×256 pixels.
S504,使用预训练的U-Net进行准确的圆形参考区域分割获得圆形参考物体,例如图5B(c)中示出的黑色圆形对象。S504 , using the pre-trained U-Net to perform accurate circular reference area segmentation to obtain a circular reference object, such as the black circular object shown in FIG. 5B(c) .
S506,使用边缘检测器,例如canny边缘检测器,进行圆形参考物体边缘过滤,获得圆形参考物体的边缘,例如图5B(d)所示。S506 , use an edge detector, such as a canny edge detector, to filter the edges of the circular reference object to obtain the edges of the circular reference object, as shown in FIG. 5B(d) .
S508,基于圆形参考物体的边缘进行椭圆拟合,获得拟合的椭圆的直径,参见图5B(e)中所示的圆形参考物体。S508 , performing ellipse fitting based on the edge of the circular reference object to obtain the diameter of the fitted ellipse, see the circular reference object shown in FIG. 5B(e) .
上述实施例中,通过面部特征点可以比较准确地确定面部局部区域,而从面部局部区域中提取圆形参考物体,可以提高处理的准确度和速度。In the above embodiment, the facial local area can be determined relatively accurately by using facial feature points, and the circular reference object is extracted from the facial local area, which can improve the accuracy and speed of processing.
图6A示出本公开一个实施例中3D头部网格生成方法的流程图。FIG. 6A shows a flowchart of a method for generating a 3D head mesh in one embodiment of the present disclosure.
如图6A所示,S602,获取目标对象的多张2D人脸图像,所述2D人脸图像包括人脸正面图像和右侧人脸侧面图像和左侧人脸侧面图像,所述人脸正面图像在所述目标对象的额头位置有圆形参考物体。As shown in FIG6A , S602 , multiple 2D face images of the target object are obtained, the 2D face images including a front face image, a right face side image, and a left face side image, the front face image having a circular reference object at the forehead position of the target object.
S604,基于多张2D人脸图像对所述目标对象进行3D头部形状重建。S604: reconstructing a 3D head shape of the target object based on multiple 2D face images.
S606,基于人脸正面图像确定所述目标对象的头部尺寸。基于人脸证明图像中圆形参考物体的尺寸,确定目标对象的头部尺寸。S606, determining the head size of the target object based on the front face image. Determine the head size of the target object based on the size of the circular reference object in the face proof image.
S608,生成目标对象的具有头部尺寸的3D头部网格。S608: Generate a 3D head mesh with head size of the target object.
上述实施例中,基于人脸正面图像中的圆形参考物体确定目标对象的头部尺寸,从而生成具有准确的头部尺寸的3D头部网格,能够满足高精度的尺寸要求,应用于头饰设计等对头部模型尺寸要求高的应用。In the above embodiment, the head size of the target object is determined based on the circular reference object in the frontal face image, thereby generating a 3D head mesh with accurate head size, which can meet high-precision size requirements and is applied to applications such as headwear design that have high requirements on head model size.
图6B示出本公开一个实施例的生成3D头部网格方法的主要工作流程图。如图6B所示,该工作流程由4个部分组成:FIG6B shows a main workflow diagram of a method for generating a 3D head mesh according to an embodiment of the present disclosure. As shown in FIG6B , the workflow consists of four parts:
(1)2D人脸图像捕获;(1) 2D face image capture;
(2)3D头部形状重建(使用编码器和头部形状和纹理的3DMM);(2) 3D head shape reconstruction (using the encoder and 3DMM of head shape and texture);
(3)3D头部尺寸测量(通过使用U-Net测量圆形参考对象的直径);和(3) 3D head size measurement (by measuring the diameter of a circular reference object using U-Net); and
(4)重建头部网格的应用场景。(4) Application scenarios of reconstructing head mesh.
下面详细描述3D头部形状重建。The 3D head shape reconstruction is described in detail below.
在一个实施例中,基于3DMM实现3D头部形状重建。3DMM是后续3D头部形状重建的基础,根据已有的训练数据集创建大量的头部形状3DMM,此外,还创建头部纹理3DMM模型。一个创建3DMM模型的具体例子包括如下步骤:In one embodiment, 3D head shape reconstruction is implemented based on 3DMM. 3DMM is the basis for subsequent 3D head shape reconstruction. A large number of head shape 3DMMs are created based on the existing training data set. In addition, a head texture 3DMM model is also created. A specific example of creating a 3DMM model includes the following steps:
(1)使用非刚性迭代最近点(NICP)算法从训练数据集生成参数化头部,NICP算法可以采用例如Amberg B,Romdhani S,Vetter T.用于表面配准的最佳步长非刚性ICP算法。训练数据集可以包括头部形状训练数据集和头部纹理训练数据集。(1) Generate a parameterized head from a training dataset using a non-rigid iterative closest point (NICP) algorithm, which can be, for example, an optimal step-size non-rigid ICP algorithm for surface registration by Amberg B, Romdhani S, Vetter T. The training dataset can include a head shape training dataset and a head texture training dataset.
(2)使用泛化性普氏分析(GPA)将参数化头部与统一尺寸对齐; (2) aligning the parameterized heads to a uniform size using Generalized Platts Analysis (GPA);
(3)使用主分量分析(PCA)从所有参数化头网格中提取主成分(PC),从而建立3DMM模型。(3) Principal components (PCs) were extracted from all parameterized head meshes using principal component analysis (PCA) to build the 3DMM model.
生成的人类头部的3DMM模型的示意图如图7所示,期中包括头部形状和头部纹理模型。A schematic diagram of the generated 3DMM model of a human head is shown in FIG7 , which includes a head shape and a head texture model.
图8示出本公开一个实施例中3D头部形状重建框架的示意图。如图8所示,当输入多个人脸图像时,使用基于CNN的编码器(例如ResNet50)来预测参数,包括一个头部网格的3DMM系数(包括头部形状801和头部纹理802参数),以及相应数量的照明(lighting)和相机参数803、804,通过差分生成器(Differentiable Renderer)805,生成适配面部图像(Fitting face image)806,适配面部特征(Fitting face landmark)807、808,并于真值面部图像(Ground-truth Face Images)809、真值面部边缘热图(Ground-truth Face Edge Heatmap)810、真值面部特征(Ground-truth Face landmark)811确定损失。基于3DMM的形状重建方法的参数ξ包括αshp,αtex,γlight,k,γview,k,其中k=1,2,...,NI,NI是输入图像的数量。需要最小化的主要网络损失函数包括像素级损失Limg(ξ)、渲染人脸图像的感知身份损失Lide(ξ)、面部边缘损失Ledge(ξ)和投影地标损失Llan(ξ);使用预训练的人脸识别网络(Arcface)提取感知身份损失计算中的感知特征。Fig. 8 shows a schematic diagram of a 3D head shape reconstruction framework in one embodiment of the present disclosure. As shown in Fig. 8, when multiple face images are input, a CNN-based encoder (e.g., ResNet50) is used to predict parameters, including a 3DMM coefficient of a head mesh (including head shape 801 and head texture 802 parameters), and a corresponding number of lighting and camera parameters 803, 804, and a fitting face image 806, a fitting face landmark 807, 808 are generated through a differentiable renderer 805, and the loss is determined in the ground-truth face images 809, the ground-truth face edge heatmap 810, and the ground-truth face landmark 811. The parameters ξ of the 3DMM-based shape reconstruction method include α shp ,α tex ,γ light,k ,γ view,k , where k = 1,2,..., NI , NI is the number of input images. The main network loss functions that need to be minimized include pixel-level loss Limg (ξ), perceptual identity loss Lide (ξ) of rendered face images, facial edge loss Ledge (ξ) and projected landmark loss Llan (ξ); a pre-trained face recognition network (Arcface) is used to extract perceptual features in the perceptual identity loss calculation.
图9示出一个实施例中不同圆形参考物体的测量评估,以评估不同圆形参考物体的测量。该测量方法可以准确检测不同大小和颜色的硬币,例如,1港币(直径:25.5毫米);0.5港币(直径:22.5毫米);0.1港币(直径:17.5毫米)。实验表明,本公开的粗到细圆形参考物体的测量方法可以准确、可靠地检测具有不同尺寸和颜色的参考对象,例如硬币,测量方法准确,鲁棒,方便。从而使得本公开的头部重建方法具有更一致的面部轮廓和更好的重建准确性。FIG9 shows an embodiment of the measurement evaluation of different circular reference objects to evaluate the measurement of different circular reference objects. The measurement method can accurately detect coins of different sizes and colors, for example, 1 Hong Kong dollar (diameter: 25.5 mm); 0.5 Hong Kong dollar (diameter: 22.5 mm); 0.1 Hong Kong dollar (diameter: 17.5 mm). Experiments show that the measurement method of the coarse to fine circular reference objects disclosed in the present invention can accurately and reliably detect reference objects with different sizes and colors, such as coins, and the measurement method is accurate, robust, and convenient. As a result, the head reconstruction method disclosed in the present invention has a more consistent facial contour and better reconstruction accuracy.
图10示出一个实施例中使用圆形或方形标记两种方法的定性结果比较。从图10中可以看出,本公开使用圆形标记的方法比使用方形标记具有更高的准确性。Figure 10 shows a qualitative comparison of the two methods using circular or square markings in one embodiment. As can be seen from Figure 10, the method of the present disclosure using circular markings has higher accuracy than using square markings.
图11示出头部相关产品应用场景,其中,顶部表示纯色重建头部,底部表示带有纹理的重建的头部。通过本公开的3D头部网格重建方法生成的头部模型,具有更高的尺寸准确度,更适合人体工程学设计,例如头饰设计,可以用于个性化定制(个性化)、虚拟试穿、和面部和头部相关产品的尺寸选择。头部相关产品应用场景例如包括:(a)耳机;(b)太阳镜;(c)VR眼镜;(d)口罩;(e)头盔。FIG11 shows an application scenario of head-related products, where the top portion represents a solid color reconstructed head and the bottom portion represents a reconstructed head with texture. The head model generated by the 3D head mesh reconstruction method disclosed in the present invention has higher dimensional accuracy and is more suitable for ergonomic design, such as headwear design, and can be used for personalized customization (personalization), virtual try-on, and size selection of face and head-related products. Application scenarios of head-related products include, for example: (a) headphones; (b) sunglasses; (c) VR glasses; (d) masks; (e) helmets.
图12示出本公开一个实施例中头部尺寸测量装置示意图。如图12所示,该头部尺寸测量装置包括:FIG12 is a schematic diagram of a head size measuring device in one embodiment of the present disclosure. As shown in FIG12 , the head size measuring device includes:
面部图像获取模块1201,用于获取目标对象的面部图像,所述面部图像在所述目标对象的额头位置有圆形参考物体;A facial image acquisition module 1201 is used to acquire a facial image of a target object, wherein the facial image has a circular reference object at the forehead of the target object;
面部特征点确定模块1202,用于确定所述目标对象的面部特征点;A facial feature point determination module 1202 is used to determine the facial feature points of the target object;
局部区域确定模块,用于基于所述面部特征点确定包括所述圆形参考物体的面部局部区域; A local area determination module, configured to determine a facial local area including the circular reference object based on the facial feature points;
参考物体提前模块1203,用于从所述面部局部区域中提取所述圆形参考物体;和A reference object advance module 1203, configured to extract the circular reference object from the facial local area; and
头部尺寸确定模块1204,用于确定所述圆形参考物体的直径,基于所述圆形参考物体的直径确定所述目标对象的头部尺寸。The head size determination module 1204 is used to determine the diameter of the circular reference object, and determine the head size of the target object based on the diameter of the circular reference object.
图13示出本公开一个实施例中3D头部网格生成装置示意图。如图13所示,该3D头部网格生成装置包括:FIG13 is a schematic diagram of a 3D head mesh generation device in one embodiment of the present disclosure. As shown in FIG13 , the 3D head mesh generation device includes:
人脸图像获取模块1301,用于获取目标对象的多张2D人脸图像,所述2D人脸图像包括人脸正面图像和右侧人脸侧面图像和左侧人脸侧面图像,所述人脸正面图像在所述目标对象的额头位置有圆形参考物体;A face image acquisition module 1301 is used to acquire multiple 2D face images of a target object, wherein the 2D face images include a front face image, a right face side image, and a left face side image, and the front face image has a circular reference object at the forehead position of the target object;
头部形状重建模块1302,用于基于所述多张2D人脸图像对所述目标对象进行3D头部形状重建;A head shape reconstruction module 1302 is used to reconstruct a 3D head shape of the target object based on the multiple 2D face images;
头部尺寸确定模块1303,用于基于所述人脸正面图像确定所述目标对象的头部尺寸;A head size determination module 1303, configured to determine the head size of the target object based on the front face image;
头部网格生成模块1304,用于生成所述目标对象的具有头部尺寸的3D头部网格。The head mesh generation module 1304 is used to generate a 3D head mesh having the head size of the target object.
本公开实施例的方案中,创建了两个高分辨率的头部形状和纹理3DMM。基于3DMM的3D头部形状重建方法使用基于CNN的编码器具有更高的精度,在基于CNN的编码器中,引入了一种新的特征点到边缘损失,以确保整体重建的形状精度。实验结果确定了输入人脸图像的大小和具有成本效益的数量,这对于实际应用非常有用。In the solution of the disclosed embodiment, two high-resolution head shape and texture 3DMMs are created. The 3DMM-based 3D head shape reconstruction method uses a CNN-based encoder with higher accuracy. In the CNN-based encoder, a new feature point to edge loss is introduced to ensure the overall reconstructed shape accuracy. The experimental results determine the size and cost-effective number of input face images, which is very useful for practical applications.
本公开实施例的方案中,没有使用方形标记,而是使用圆形参考物体,圆形参考物体在附着在面部时对变形的敏感性较低,更稳定且易于进行头部尺寸测量。此外,使用硬币形物体在日常生活中也比较常见。采用基于U-Net的精确头部尺寸测量方法,更稳定,更方便。本公开的技术方案法更适合人体工程学设计。In the solution of the embodiment of the present disclosure, a circular reference object is used instead of a square mark. The circular reference object is less sensitive to deformation when attached to the face, and is more stable and easy to measure the head size. In addition, the use of coin-shaped objects is also common in daily life. The accurate head size measurement method based on U-Net is more stable and convenient. The technical solution method of the present disclosure is more suitable for ergonomic design.
所属技术领域的技术人员能够理解,本发明的各个方面可以实现为系统、方法或程序产品。因此,本发明的各个方面可以具体实现为以下形式,即:完全的硬件实施方式、完全的软件实施方式(包括固件、微代码等),或硬件和软件方面结合的实施方式,这里可以统称为“电路”、“模块”或“系统”。It will be appreciated by those skilled in the art that various aspects of the present invention may be implemented as a system, method or program product. Therefore, various aspects of the present invention may be specifically implemented in the following forms, namely: a complete hardware implementation, a complete software implementation (including firmware, microcode, etc.), or a combination of hardware and software, which may be collectively referred to herein as a "circuit", "module" or "system".
下面参照图14来描述根据本发明的这种实施方式的电子设备1400。图14显示的电子设备1400仅仅是一个示例,不应对本发明实施例的功能和使用范围带来任何限制。The electronic device 1400 according to this embodiment of the present invention is described below with reference to Fig. 14. The electronic device 1400 shown in Fig. 14 is only an example and should not bring any limitation to the functions and application scope of the embodiment of the present invention.
如图14所示,电子设备1400以通用计算设备的形式表现。电子设备1400的组件可以包括但不限于:上述至少一个处理单元1410、上述至少一个存储单元1420、连接不同系统组件(包括存储单元1420和处理单元1410)的总线1430。As shown in Fig. 14, the electronic device 1400 is presented in the form of a general-purpose computing device. The components of the electronic device 1400 may include, but are not limited to: at least one processing unit 1410, at least one storage unit 1420, and a bus 1430 connecting different system components (including the storage unit 1420 and the processing unit 1410).
其中,所述存储单元存储有程序代码,所述程序代码可以被所述处理单元1410执行,使得所述处理单元1410执行本说明书上述“示例性方法”部分中描述的根据本发明各种示例性实施方式的步骤。例如,所述处理单元1410可以执行如图2和图6A中所示的步骤。The storage unit stores a program code, which can be executed by the processing unit 1410, so that the processing unit 1410 performs the steps according to various exemplary embodiments of the present invention described in the above "Exemplary Method" section of this specification. For example, the processing unit 1410 can perform the steps shown in Figures 2 and 6A.
存储单元1420可以包括易失性存储单元形式的可读介质,例如随机存取存储单元(RAM)14201和/或高速缓存存储单元14202,还可以进一步包括只读存储单元(ROM)14203。 The storage unit 1420 may include a readable medium in the form of a volatile storage unit, such as a random access storage unit (RAM) 14201 and/or a cache storage unit 14202 , and may further include a read-only storage unit (ROM) 14203 .
存储单元1420还可以包括具有一组(至少一个)程序模块14205的程序/实用工具14204,这样的程序模块14205包括但不限于:操作系统、一个或者多个应用程序、其它程序模块以及程序数据,这些示例中的每一个或某种组合中可能包括网络环境的实现。The storage unit 1420 may also include a program/utility 14204 having a set (at least one) of program modules 14205, such program modules 14205 including but not limited to: an operating system, one or more application programs, other program modules, and program data, each of which or some combination may include an implementation of a network environment.
总线1430可以为表示几类总线结构中的一种或多种,包括存储单元总线或者存储单元控制器、外围总线、图形加速端口、处理单元或者使用多种总线结构中的任意总线结构的局域总线。Bus 1430 may represent one or more of several types of bus structures, including a memory unit bus or memory unit controller, a peripheral bus, an accelerated graphics port, a processing unit, or a local bus using any of a variety of bus architectures.
电子设备1400也可以与一个或多个外部设备1500(例如键盘、指向设备、蓝牙设备等)通信,还可与一个或者多个使得用户能与该电子设备1500交互的设备通信,和/或与使得该电子设备1400能与一个或多个其它计算设备进行通信的任何设备(例如路由器、调制解调器等等)通信。这种通信可以通过输入/输出(I/O)接口1450进行。并且,电子设备1400还可以通过网络适配器1460与一个或者多个网络(例如局域网(LAN),广域网(WAN)和/或公共网络,例如因特网)通信。如图所示,网络适配器1460通过总线1430与电子设备1400的其它模块通信。应当明白,尽管图中未示出,可以结合电子设备1500使用其它硬件和/或软件模块,包括但不限于:微代码、设备驱动器、冗余处理单元、外部磁盘驱动阵列、RAID系统、磁带驱动器以及数据备份存储系统等。The electronic device 1400 may also communicate with one or more external devices 1500 (e.g., keyboards, pointing devices, Bluetooth devices, etc.), may also communicate with one or more devices that enable a user to interact with the electronic device 1500, and/or communicate with any device that enables the electronic device 1400 to communicate with one or more other computing devices (e.g., routers, modems, etc.). Such communication may be performed via an input/output (I/O) interface 1450. Furthermore, the electronic device 1400 may also communicate with one or more networks (e.g., a local area network (LAN), a wide area network (WAN), and/or a public network, such as the Internet) via a network adapter 1460. As shown, the network adapter 1460 communicates with other modules of the electronic device 1400 via a bus 1430. It should be understood that, although not shown in the figure, other hardware and/or software modules may be used in conjunction with the electronic device 1500, including but not limited to: microcode, device drivers, redundant processing units, external disk drive arrays, RAID systems, tape drives, and data backup storage systems.
通过以上的实施方式的描述,本领域的技术人员易于理解,这里描述的示例实施方式可以通过软件实现,也可以通过软件结合必要的硬件的方式来实现。因此,根据本公开实施方式的技术方案可以以软件产品的形式体现出来,该软件产品可以存储在一个非易失性存储介质(可以是CD-ROM,U盘,移动硬盘等)中或网络上,包括若干指令以使得一台计算设备(可以是个人计算机、服务器、终端装置、或者网络设备等)执行根据本公开实施方式的方法。Through the description of the above implementation, it is easy for those skilled in the art to understand that the example implementation described here can be implemented by software, or by software combined with necessary hardware. Therefore, the technical solution according to the implementation of the present disclosure can be embodied in the form of a software product, which can be stored in a non-volatile storage medium (which can be a CD-ROM, a USB flash drive, a mobile hard disk, etc.) or on a network, including several instructions to enable a computing device (which can be a personal computer, a server, a terminal device, or a network device, etc.) to execute the method according to the implementation of the present disclosure.
在本公开的示例性实施例中,还提供了一种计算机可读存储介质,其上存储有能够实现本说明书上述方法的程序产品。在一些可能的实施方式中,本发明的各个方面还可以实现为一种程序产品的形式,其包括程序代码,当所述程序产品在终端设备上运行时,所述程序代码用于使所述终端设备执行本说明书上述“示例性方法”部分中描述的根据本发明各种示例性实施方式的步骤。In an exemplary embodiment of the present disclosure, a computer-readable storage medium is also provided, on which a program product capable of implementing the above method of the present specification is stored. In some possible implementations, various aspects of the present invention can also be implemented in the form of a program product, which includes a program code, and when the program product is run on a terminal device, the program code is used to enable the terminal device to execute the steps according to various exemplary embodiments of the present invention described in the above "Exemplary Method" section of the present specification.
描述了根据本发明的实施方式的用于实现上述方法的程序产品,其可以采用便携式紧凑盘只读存储器(CD-ROM)并包括程序代码,并可以在终端设备,例如个人电脑上运行。然而,本发明的程序产品不限于此,在本文件中,可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。A program product for implementing the above method according to an embodiment of the present invention is described, which can adopt a portable compact disk read-only memory (CD-ROM) and include program code, and can be run on a terminal device, such as a personal computer. However, the program product of the present invention is not limited thereto, and in this document, a readable storage medium can be any tangible medium containing or storing a program, which can be used by or in combination with an instruction execution system, an apparatus or a device.
所述程序产品可以采用一个或多个可读介质的任意组合。可读介质可以是可读信号介质或者可读存储介质。可读存储介质例如可以为但不限于电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。可读存储介质的更具体的例子(非穷举的列表)包括:具有一个或多个导线的电连接、便携式盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(EPROM或闪存)、光纤、便携式紧凑 盘只读存储器(CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。The program product may be in any combination of one or more readable media. The readable medium may be a readable signal medium or a readable storage medium. The readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device or device, or any combination thereof. More specific examples of readable storage media (a non-exhaustive list) include: an electrical connection with one or more wires, a portable disk, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact A compact disk read only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the above.
计算机可读信号介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了可读程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。可读信号介质还可以是可读存储介质以外的任何可读介质,该可读介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。Computer readable signal media may include data signals propagated in baseband or as part of a carrier wave, in which readable program code is carried. Such propagated data signals may take a variety of forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the above. Readable signal media may also be any readable medium other than a readable storage medium, which may send, propagate, or transmit a program for use by or in conjunction with an instruction execution system, apparatus, or device.
可读介质上包含的程序代码可以用任何适当的介质传输,包括但不限于无线、有线、光缆、RF等等,或者上述的任意合适的组合。The program code embodied on the readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wired, optical cable, RF, etc., or any suitable combination of the foregoing.
可以以一种或多种程序设计语言的任意组合来编写用于执行本发明操作的程序代码,所述程序设计语言包括面向对象的程序设计语言—诸如Java、C++等,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算设备上执行、部分地在用户设备上执行、作为一个独立的软件包执行、部分在用户计算设备上部分在远程计算设备上执行、或者完全在远程计算设备或服务器上执行。在涉及远程计算设备的情形中,远程计算设备可以通过任意种类的网络,包括局域网(LAN)或广域网(WAN),连接到用户计算设备,或者,可以连接到外部计算设备(例如利用因特网服务提供商来通过因特网连接)。Program code for performing the operations of the present invention may be written in any combination of one or more programming languages, including object-oriented programming languages such as Java, C++, etc., and conventional procedural programming languages such as "C" or similar programming languages. The program code may be executed entirely on the user computing device, partially on the user device, as a separate software package, partially on the user computing device and partially on a remote computing device, or entirely on a remote computing device or server. In cases involving a remote computing device, the remote computing device may be connected to the user computing device through any type of network, including a local area network (LAN) or a wide area network (WAN), or may be connected to an external computing device (e.g., through the Internet using an Internet service provider).
应当注意,尽管在上文详细描述中提及了用于动作执行的设备的若干模块或者单元,但是这种划分并非强制性的。实际上,根据本公开的实施方式,上文描述的两个或更多模块或者单元的特征和功能可以在一个模块或者单元中具体化。反之,上文描述的一个模块或者单元的特征和功能可以进一步划分为由多个模块或者单元来具体化。It should be noted that, although several modules or units of the device for action execution are mentioned in the above detailed description, this division is not mandatory. In fact, according to the embodiments of the present disclosure, the features and functions of two or more modules or units described above can be embodied in one module or unit. On the contrary, the features and functions of one module or unit described above can be further divided into multiple modules or units to be embodied.
此外,尽管在附图中以特定顺序描述了本公开中方法的各个步骤,但是,这并非要求或者暗示必须按照该特定顺序来执行这些步骤,或是必须执行全部所示的步骤才能实现期望的结果。附加的或备选的,可以省略某些步骤,将多个步骤合并为一个步骤执行,以及/或者将一个步骤分解为多个步骤执行等。In addition, although the steps of the method in the present disclosure are described in a specific order in the drawings, this does not require or imply that the steps must be performed in this specific order, or that all the steps shown must be performed to achieve the desired results. Additionally or alternatively, some steps may be omitted, multiple steps may be combined into one step, and/or one step may be decomposed into multiple steps, etc.
通过以上的实施方式的描述,本领域的技术人员易于理解,这里描述的示例实施方式可以通过软件实现,也可以通过软件结合必要的硬件的方式来实现。因此,根据本公开实施方式的技术方案可以以软件产品的形式体现出来,该软件产品可以存储在一个非易失性存储介质(可以是CD-ROM,U盘,移动硬盘等)中或网络上,包括若干指令以使得一台计算设备(可以是个人计算机、服务器、移动终端、或者网络设备等)执行根据本公开实施方式的方法。Through the description of the above implementation, it is easy for those skilled in the art to understand that the example implementation described here can be implemented by software, or by software combined with necessary hardware. Therefore, the technical solution according to the implementation of the present disclosure can be embodied in the form of a software product, which can be stored in a non-volatile storage medium (which can be a CD-ROM, a USB flash drive, a mobile hard disk, etc.) or on a network, including several instructions to enable a computing device (which can be a personal computer, a server, a mobile terminal, or a network device, etc.) to execute the method according to the implementation of the present disclosure.
本领域技术人员在考虑说明书及实践这里公开的发明后,将容易想到本公开的其它实施方案。本申请旨在涵盖本公开的任何变型、用途或者适应性变化,这些变型、用途或者适应性变化遵循本公开的一般性原理并包括本公开未公开的本技术领域中的公知常识或惯用技术手段。说明书和实施例仅被视为示例性的,本公开的真正范围和精神由所附的权利要求指出。 Those skilled in the art will readily appreciate other embodiments of the present disclosure after considering the specification and practicing the invention disclosed herein. This application is intended to cover any modification, use or adaptation of the present disclosure, which follows the general principles of the present disclosure and includes common knowledge or customary techniques in the art that are not disclosed in the present disclosure. The specification and examples are intended to be exemplary only, and the true scope and spirit of the present disclosure are indicated by the appended claims.
本公开适用于人工智能技术领域,用以解决相关技术中3D头部/面部重建时头部尺寸确定不准确的问题,达到准确确定目标对象的头部尺寸的效果。 The present disclosure is applicable to the field of artificial intelligence technology, and is used to solve the problem of inaccurate head size determination during 3D head/face reconstruction in related technologies, so as to achieve the effect of accurately determining the head size of a target object.
Claims (21)
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202310833308.3 | 2023-07-10 | ||
| CN202310833308.3A CN119295531A (en) | 2023-07-10 | 2023-07-10 | Head size measurement, head grid generation method and device, medium and equipment |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| WO2025011559A1 true WO2025011559A1 (en) | 2025-01-16 |
| WO2025011559A8 WO2025011559A8 (en) | 2025-03-06 |
Family
ID=94150318
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/CN2024/104500 Pending WO2025011559A1 (en) | 2023-07-10 | 2024-07-09 | Head size measurement method and apparatus, head grid generation method and apparatus, medium, and device |
Country Status (2)
| Country | Link |
|---|---|
| CN (1) | CN119295531A (en) |
| WO (1) | WO2025011559A1 (en) |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20190035149A1 (en) * | 2015-08-14 | 2019-01-31 | Metail Limited | Methods of generating personalized 3d head models or 3d body models |
| CN111563926A (en) * | 2020-05-22 | 2020-08-21 | 上海依图网络科技有限公司 | Method, electronic device, medium, and system for measuring physical size of object in image |
| CN113095149A (en) * | 2021-03-18 | 2021-07-09 | 西北工业大学 | Full-head texture network structure based on single face image and generation method |
| US20230154111A1 (en) * | 2021-11-15 | 2023-05-18 | Samsung Electronics Co., Ltd. | Method and apparatus for three-dimensional reconstruction of a human head for rendering a human image |
-
2023
- 2023-07-10 CN CN202310833308.3A patent/CN119295531A/en active Pending
-
2024
- 2024-07-09 WO PCT/CN2024/104500 patent/WO2025011559A1/en active Pending
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20190035149A1 (en) * | 2015-08-14 | 2019-01-31 | Metail Limited | Methods of generating personalized 3d head models or 3d body models |
| CN111563926A (en) * | 2020-05-22 | 2020-08-21 | 上海依图网络科技有限公司 | Method, electronic device, medium, and system for measuring physical size of object in image |
| CN113095149A (en) * | 2021-03-18 | 2021-07-09 | 西北工业大学 | Full-head texture network structure based on single face image and generation method |
| US20230154111A1 (en) * | 2021-11-15 | 2023-05-18 | Samsung Electronics Co., Ltd. | Method and apparatus for three-dimensional reconstruction of a human head for rendering a human image |
Also Published As
| Publication number | Publication date |
|---|---|
| CN119295531A (en) | 2025-01-10 |
| WO2025011559A8 (en) | 2025-03-06 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| EP3992919B1 (en) | Three-dimensional facial model generation method and apparatus, device, and medium | |
| EP4198814A1 (en) | Gaze correction method and apparatus for image, electronic device, computer-readable storage medium, and computer program product | |
| US12450688B2 (en) | Gaze correction method and apparatus for face image, device, computer-readable storage medium, and computer program product face image | |
| CN109859305B (en) | Three-dimensional face modeling and recognizing method and device based on multi-angle two-dimensional face | |
| EP3992918A1 (en) | Method for generating 3d expression base, voice interactive method, apparatus and medium | |
| WO2020103700A1 (en) | Image recognition method based on micro facial expressions, apparatus and related device | |
| WO2021196389A1 (en) | Facial action unit recognition method and apparatus, electronic device, and storage medium | |
| JP2020523665A (en) | Biological detection method and device, electronic device, and storage medium | |
| WO2022188697A1 (en) | Biological feature extraction method and apparatus, device, medium, and program product | |
| EP3928929A1 (en) | Auxiliary photographing device for movement disorder disease analysis, control method and apparatus | |
| CN114549557A (en) | Portrait segmentation network training method, device, equipment and medium | |
| CN108734078B (en) | Image processing method, image processing apparatus, electronic device, storage medium, and program | |
| CN108388889B (en) | Method and device for analyzing face image | |
| CN115994944B (en) | Training method of key point prediction model, three-dimensional key point prediction method and related equipment | |
| CN117079313A (en) | Image processing methods, devices, equipment and storage media | |
| CN118230394A (en) | A facial expression driving method and device based on 3D key points of human face | |
| CN111680577A (en) | Face detection method and device | |
| CN112580395A (en) | Depth information-based 3D face living body recognition method, system, device and medium | |
| CN116978081A (en) | Image processing method and device, storage medium, and program product | |
| CN116434253A (en) | Image processing method, device, equipment, storage medium and product | |
| KR20230053262A (en) | A 3D object recognition method based on a 2D real space image and a computer program recorded on a recording medium to execute the same | |
| WO2025011559A1 (en) | Head size measurement method and apparatus, head grid generation method and apparatus, medium, and device | |
| CN112883919A (en) | Glasses matching method and device based on face feature enhancement matching | |
| CN114120386A (en) | Face recognition method, device, equipment and storage medium | |
| CN111275754A (en) | Face acne mark proportion calculation method based on deep learning |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 24838811 Country of ref document: EP Kind code of ref document: A1 |