US12437382B2 - Device and computer-implemented method for evaluating a control of a generator for determining pixels of a synthetic image - Google Patents
Device and computer-implemented method for evaluating a control of a generator for determining pixels of a synthetic imageInfo
- Publication number
- US12437382B2 US12437382B2 US18/313,273 US202318313273A US12437382B2 US 12437382 B2 US12437382 B2 US 12437382B2 US 202318313273 A US202318313273 A US 202318313273A US 12437382 B2 US12437382 B2 US 12437382B2
- Authority
- US
- United States
- Prior art keywords
- latent
- latent code
- label map
- determining
- class
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0475—Generative networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/094—Adversarial learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/46—Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
- G06V10/467—Encoded features or binary features, e.g. local binary patterns [LBP]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/761—Proximity, similarity or dissimilarity measures
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/70—Labelling scene content, e.g. deriving syntactic or semantic representations
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30168—Image quality inspection
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30248—Vehicle exterior or interior
- G06T2207/30252—Vehicle exterior; Vicinity of vehicle
Definitions
- GAN controls may be based on visual inspection. Härkönen, E., Hertzmann, A., Lehtinen, J., Paris, S.’ “Ganspace: Discovering interpretable gan controls;” in: Advances in Neural Information Processing Systems (NeurIPS) (2020) discloses an example for such inspection.
- NeuroIPS Neural Information Processing Systems
- the device and the computer-implemented method for evaluating generator controls comprises providing a metric to evaluate an effectiveness of a generator control discovery method and determining the metric.
- the computer-implemented method for evaluating a control of a generator for determining pixels of a synthetic image wherein the generator is configured to determine pixels of the synthetic image from a first input comprising a label map and a first latent code, wherein the label map comprises a mapping of at least one class to at least one of the pixels, wherein the method comprises providing the label map and latent code, wherein the latent code comprises input data points in a latent space, providing the control, wherein the control comprises a set of directions for moving the latent code in the latent space, determining the first latent code depending on at least one input data point of the latent code that is moved in a first direction, wherein the first direction is selected from the set of directions, determining a distance between at least one pair of synthetic images, which are generated by the generator for different first inputs, wherein the different first inputs comprise the label map and vary by the first direction that is selected for determining the first latent code from the latent code.
- This distance provides a diversity
- the first input may comprise a class mask, wherein the class mask indicates the input data points to be moved in the first direction for a class, and the distance is determined depending on pixels of the synthetic images indicated by the class mask to be considered.
- the method may comprise determining the distance between pairs of synthetic images that are generated with the same label map and first direction and with varying latent code.
- a low consistency score implies that each class edits introduce consistent changes in an area of the synthetic image indicated by the class mask.
- the method may comprise determining an average of distances determined for different pairs, in particular determining a mean over these distances.
- the method may comprise determining distances for at least one class of the classes that the label map comprises and the average depending on the distances for the at least one class.
- the method may comprise synthesizing a plurality of the synthetic images with the same label map and testing a robustness or functionality of an application for processing digital images depending on distances or scores that are determined for the plurality of the synthetic images.
- the label map may be determined to emulate a digital image output of an image sensor, in particular a video, radar, LiDAR, ultrasonic, motion or thermal image sensor, wherein a control signal for a computer-controlled machine, in particular a robot, preferably a vehicle or a vehicle for autonomous driving, a domestic appliance, a power tool a manufacturing machine, a personal assistant, a device for automatic optical inspection, or an access system, is determined by a model depending on the synthetic image, in particular a model comprising an object detector configured for detecting objects in the synthetic image, or a classifier configured for classifying the synthetic image or objects in the synthetic image, or a segmenter, configured for segmenting the synthetic image, and wherein the model is trained depending on the synthetic image to determine the control signal and/or determining the label map and/or a selected first direction from the set of directions from user input detected by a graphical user interface, wherein a synthetic image is determined depending on the label map and/or latent code that is moved in the selected
- the device for evaluating a control of a generator for determining pixels of a synthetic image comprises at least one processor and at least one memory wherein the at least one memory is configured to store computer-readable instructions that when executed by the at least one processor cause the device to execute steps of the method according to one of the above-described methods and wherein the at least one processor is configured to execute the instructions.
- a computer program according to an example embodiment of the present invention may comprise computer-readable instructions that when executed by a computer cause the computer to perform the steps of the method of the present invention.
- This computer program has advantages that correspond to the advantages of the method of the present invention.
- FIG. 1 schematically depicts a device for determining pixels of a synthetic image, according to an example embodiment of the present invention.
- FIG. 3 schematically depicts a label map for the digital image, according to an example embodiment of the present invention.
- FIG. 4 schematically depicts a class mask for the digital image, according to an example embodiment of the present invention.
- FIG. 2 schematically depicts a digital image 202 .
- the digital image 202 may be a generated image.
- the digital image 202 comprises a first building 204 , a second building 206 , a third building 208 , a first car 210 , a second car 212 , and a third car 314 .
- the first building 204 is located on a left side of the digital image and the second building 206 is located on a right side of the digital image 202 .
- These buildings are located on opposite sides of a road 216 that extends from a lower left corner of the digital image 202 to the third building 208 which is located on an upper right corner of the digital image 202 .
- the first car 210 is located in the digital image 202 left of the second car 212 .
- the third car 214 is located in the digital image 202 right of the second car 210 .
- the first car 210 and the second car 212 are located closer to the first building 204 than the third car 212 .
- the mapping y in the example assigns the pixels of the digital image 202 that represent the first building 204 , the second building 206 and the third building 208 a first class 304 , building.
- the mapping y in the example assigns the pixels of the digital image 202 that represent the first car 210 , the second car 212 and the third car 214 a second class 306 , car.
- the mapping y in the example assigns the pixels of the digital image 202 that represent the walkway 220 a third class 308 , walkway.
- the mapping y in the example assigns the pixels of the digital image 202 that represent the tree 218 a fourth class 310 , tree.
- the mapping y in the example assigns the pixels of the digital image 202 that represent the street 216 a fifth class 312 , street.
- the class mask 402 may have the same or lower spatial dimensions as the label map 302 .
- the mapping M c may assign a group of pixels the same binary value in this case.
- FIG. 5 schematically depicts a first synthetic image 502 .
- the first synthetic image 502 comprises a first set of pixels 504 that correspond to the pixels that represent the street according to the class mask 402 and a second set of pixels 506 that correspond to the other pixels according to the class mask 402 .
- the synthetic image 502 may have the same special dimensions as the label map 302 , the class mask 402 and/or the digital image 202 .
- FIG. 7 schematically depicts a third synthetic image 702 .
- the third synthetic image 702 is based on another class mask that comprises a mapping M c that assigns the first binary value to pixels that according to the label map 302 represent the street or the building.
- the third synthetic image 702 comprises a first set of pixels 704 that correspond to the pixels that represent the street according to label map 302 and a second set of pixels 706 that corresponds to the pixels that represent the buildings according to the label map 302 and a third set of pixels 708 that correspond to the other pixels according to the label map 302 .
- the first set of pixels 704 that represents the street is different in color from the pixels of the digital image 202 that represent the street 216 .
- the second set of pixels 706 that represents the buildings is different in color from the pixels of the digital image 202 that represent the first building 204 , the second building 206 and the third building 308 .
- the third set of pixels 708 is unchanged compared to the digital image 202 .
- FIG. 8 schematically depicts a process 800 for evaluating generative adversarial network controls.
- Generative adversarial network controls are an example of a control of a generator for determining pixels of a synthetic image.
- the process 800 comprises determining pixels of a synthetic image 802 ′.
- the process 800 comprises providing a class mask 806 .
- the class mask 806 comprises a mapping of at least one class c ⁇ C to at least one of the pixels.
- the process 800 comprises providing a latent code 808 .
- the latent code comprises input data points 810 in a latent space.
- the latent code is for example sampled.
- the latent code 808 may be spatially aligned with the class mask 806 and/or the label map 804 and/or the synthetic images 802 , 802 ′.
- the class mask 806 is stored in an at least two-dimensional tensor
- the latent code 808 is stored in an at least three-dimensional tensor
- the label map 804 is stored in an at least three-dimensional tensor
- the synthetic image 802 or 802 ′ is stored in an at least three-dimensional tensor.
- the at least three-dimensional tensor for the synthetic image 802 or 802 ′ in one example has a dimension corresponding to a width of the synthetic image 802 or 802 ′ and a dimension corresponding to a height of the synthetic image 802 or 802 ′.
- the width of the synthetic image 802 or 802 ′ may be a given width 222 of the digital mage 202 .
- the height of the synthetic image 802 or 802 ′ may be a given height 224 of the digital image 202 .
- Spatially aligned in this context may mean that the tensor for the class mask 806 and/or the label map 804 comprises a dimension of the same size as the dimension corresponding to the width of the synthetic image 802 or 802 ′ and a dimension of the same size as the dimension corresponding to the height of the synthetic image 802 or 802 ′.
- the process 800 comprises determining the latent code 808 depending on the input data point 810 that is moved in the direction 812 .
- the process 800 comprises moving the input data point 810 if it is selected for moving by the class mask 806 , e.g. in case the first binary value, and otherwise not moving the input data point 810 .
- the process 800 comprises determining the synthetic image 802 or 802 ′ depending on an output of a generator 816 for an input that comprises the label map 804 and the class mask 806 and the latent code 808 .
- the generator 816 may be a generator of a generative adversarial network.
- the generator 816 may comprise another neural network, e.g. a diffusion model or a VQ model.
- the generator 816 is a well-trained generator G of a Semantic Image Synthesis, SIS, model.
- the generator G is configured to synthesize digital images from label maps and latent code.
- the generator G for example comprises a neural network.
- the generator G in this example comprises an input layer and an output layer and at least one intermediate layer l ⁇ L between the input layer and the output layer.
- the generator G and the latent code z is not restricted to a three-dimensional latent space.
- the latent code z may have more than three dimensions.
- To synthesize the synthetic image locally may result in changing only selected pixels, i.e. pixels that are indicated in the M c , compared to a synthetic image that results from synthesizing the synthetic image from the same latent code z without moving part of it in the first direction v k c .
- FIG. 9 schematically depicts a method for evaluating generative adversarial network controls.
- the method for discovering latent directions can be a different one and use a different optimization objective.
- the method for evaluating the control for the generator for determining pixels of the synthetic image x comprises a step 902 .
- the method comprises a step 904 .
- the method comprises a step 906 .
- the method comprises a step 910 .
- the method comprises determining the synthetic image depending on an output of the generator for the first input.
- the method comprises a step 912 .
- the method comprises determining a distance between at least one pair of synthetic images 802 , 802 ′, which are generated by the generator 816 for different first inputs.
- the method may comprise determining an average of distances determined for different pairs.
- a mean over these distances is determined.
- the different first inputs in one example comprise the label map 804 and vary by the first direction that is selected for determining the first latent code 808 from the latent code z. For example the specific diversity score Div or the class specific diversity score Div(c) is determined for these.
- the different first inputs in one example comprise the label map 804 and the class mask 806 and vary by the first direction that is selected for determining the first latent code 808 from the latent code z. For example the class specific disentanglement score Dis(c) is determined for these.
- the method may use mean Intersection-over-Union to assess an alignment of a generated synthetic image with a ground truth semantic label map, calculated via a pre-trained semantic segmentation network.
- the method may comprise synthesizing a plurality of the synthetic images with the same label map and with different first directions.
- the label map may be determined to emulate a digital image output of an image sensor, in particular a video, radar, LiDAR, ultrasonic, motion or thermal image sensor.
- the method may comprise determining a control signal for a computer-controlled machine by a model depending on the synthetic image.
- the model is in one example trained depending on the synthetic image to determine the control signal.
- the computer-controlled machine may be a robot.
- the robot may be a vehicle or a vehicle for autonomous driving, a domestic appliance, a power tool a manufacturing machine, a personal assistant, a device for automatic optical inspection, or an access system.
- the model may comprise an object detector configured for detecting objects in the synthetic image.
- the model may comprise a classifier configured for classifying the synthetic image or objects in the synthetic image.
- the model may comprise a segmenter, configured for segmenting the synthetic image.
- the method in one example comprises determining the label map from user input detected by a graphical user interface.
- the synthetic image is for example output by a display, in particular of the graphical user interface.
- a user may draft the label map and then modify the synthesized image using the learnt directions.
- the user may select the first direction from the set of directions.
- the synthetic image resulting from moving the latent code with the selected first direction is for example output.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Computation (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- General Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Health & Medical Sciences (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Computational Linguistics (AREA)
- Biomedical Technology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biophysics (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Quality & Reliability (AREA)
- Image Analysis (AREA)
Abstract
Description
x(v k)=G(z+αv k ,y)=F(h(z,v k ,y))
where
h(z,v k ,y)={G l(z,y)}l∈L
is the features or a chosen subset of features from the intermediate layers l∈L in the generator G(z,vk,y). The generator G and the latent code z is not restricted to a three-dimensional latent space. The latent code z may have more than three dimensions.
x(v k c)=G(z+αM c ⊙v k c ,y)=F(h(z,v k c ,y))
wherein ∥ ∥2 is the L2 Norm.
where K is the number of discovered latend directions and d is a distance between two synthetic images x(vk
where K is the number of discovered latend directions and d is the distance between two synthetic images x(z1,vk) and x(z2,vk), which are generated with the same label map y and latent direction, e.g. the first direction vk and with varying latend code z1,z2, e.g input noise. A low consistency score implies that each edit introduces consistent changes in the synthetic image.
where K is the number of discovered class specific latent directions and d is the distance between two synthetic images x(vk
wherein d is the distance between a synthetic image x which is generated with the label map y and unmoved latent code z, and a synthetic image x(vk c), which is generated with the label map y and latend code z which is moved with a class specific latent direction, e.g. the first direction vk c.
where K is the number of discovered class specific latend directions and d is the distance between two synthetic images x(z1,vk c) and x(z2,vk c), which are generated with the same label map y and class specific latent direction, e.g. the first direction vk c and with varying latent code z1, z2, e.g input noise. A low consistency score implies that each class edit introduces consistent changes in an area of the synthetic image indicated by the class mask Mc.
Claims (9)
Applications Claiming Priority (3)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| EP22175795 | 2022-05-27 | ||
| EP22175795.8 | 2022-05-27 | ||
| EP22175795.8A EP4283565A1 (en) | 2022-05-27 | 2022-05-27 | Device and computer-implemented method for evaluating a control of a generator for determining pixels of a synthetic image |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20230386004A1 US20230386004A1 (en) | 2023-11-30 |
| US12437382B2 true US12437382B2 (en) | 2025-10-07 |
Family
ID=81851295
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US18/313,273 Active 2044-01-10 US12437382B2 (en) | 2022-05-27 | 2023-05-05 | Device and computer-implemented method for evaluating a control of a generator for determining pixels of a synthetic image |
Country Status (3)
| Country | Link |
|---|---|
| US (1) | US12437382B2 (en) |
| EP (1) | EP4283565A1 (en) |
| CN (1) | CN117132979A (en) |
Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20190392304A1 (en) * | 2018-06-22 | 2019-12-26 | Insilico Medicine, Inc. | Mutual information adversarial autoencoder |
| US20200210755A1 (en) * | 2018-12-27 | 2020-07-02 | Bull Sas | Method of classification of images among different classes |
| US20220138897A1 (en) * | 2020-11-03 | 2022-05-05 | Adobe Inc. | Image Generation and Editing with Latent Transformation Detection |
| US20220261658A1 (en) * | 2021-02-18 | 2022-08-18 | Volkswagen Aktiengesellschaft | Apparatus, system and method for translating sensor label data between sensor domains |
| US20230053588A1 (en) * | 2021-08-12 | 2023-02-23 | Adobe Inc. | Generating synthesized digital images utilizing a multi-resolution generator neural network |
| US20230351566A1 (en) * | 2022-04-27 | 2023-11-02 | Adobe Inc. | Exemplar-based object appearance transfer driven by correspondence |
-
2022
- 2022-05-27 EP EP22175795.8A patent/EP4283565A1/en active Pending
-
2023
- 2023-05-05 US US18/313,273 patent/US12437382B2/en active Active
- 2023-05-26 CN CN202310609830.3A patent/CN117132979A/en active Pending
Patent Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20190392304A1 (en) * | 2018-06-22 | 2019-12-26 | Insilico Medicine, Inc. | Mutual information adversarial autoencoder |
| US20200210755A1 (en) * | 2018-12-27 | 2020-07-02 | Bull Sas | Method of classification of images among different classes |
| US20220138897A1 (en) * | 2020-11-03 | 2022-05-05 | Adobe Inc. | Image Generation and Editing with Latent Transformation Detection |
| US20220261658A1 (en) * | 2021-02-18 | 2022-08-18 | Volkswagen Aktiengesellschaft | Apparatus, system and method for translating sensor label data between sensor domains |
| US20230053588A1 (en) * | 2021-08-12 | 2023-02-23 | Adobe Inc. | Generating synthesized digital images utilizing a multi-resolution generator neural network |
| US20230351566A1 (en) * | 2022-04-27 | 2023-11-02 | Adobe Inc. | Exemplar-based object appearance transfer driven by correspondence |
Non-Patent Citations (12)
Also Published As
| Publication number | Publication date |
|---|---|
| CN117132979A (en) | 2023-11-28 |
| EP4283565A1 (en) | 2023-11-29 |
| US20230386004A1 (en) | 2023-11-30 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US11830246B2 (en) | Systems and methods for extracting and vectorizing features of satellite imagery | |
| Varghese et al. | ChangeNet: A deep learning architecture for visual change detection | |
| KR102114357B1 (en) | Method and device for constructing a table including information on a pooling type and testing method and testing device using the same | |
| US10410120B1 (en) | Learning method and testing method of object detector to be used for surveillance based on R-CNN capable of converting modes according to aspect ratios or scales of objects, and learning device and testing device using the same | |
| US10372968B2 (en) | Object-focused active three-dimensional reconstruction | |
| CN113052200A (en) | Sonar image target detection method based on yolov3 network | |
| JP2020038660A (en) | Learning method and learning device for detecting lane by using cnn, and test method and test device using the same | |
| KR102178469B1 (en) | Method and system for estimation of pedestrian pose orientation using soft target training based on teacher-student framework | |
| JP2020038661A (en) | Learning method and learning device for detecting lane by using lane model, and test method and test device using the same | |
| Lorsakul et al. | Traffic sign recognition for intelligent vehicle/driver assistance system using neural network on opencv | |
| CN119152335B (en) | UAV visual recognition method and device | |
| CN110705566B (en) | Multi-mode fusion significance detection method based on spatial pyramid pool | |
| KR20200095357A (en) | Learning method and learning device for heterogeneous sensor fusion by using merging network which learns non-maximum suppression | |
| CN109977968A (en) | A kind of SAR change detecting method of deep learning classification and predicting | |
| Naseer et al. | Multimodal Deep Learning Framework for Enhanced Semantic Scene Classification Using RGB-D Images | |
| JP7709857B2 (en) | DEVICE AND METHOD FOR TRAINING AND TESTING CLASSIFIERS - Patent application | |
| BenHajyoussef et al. | Recent advances on image edge detection | |
| CN116630245B (en) | A polyp segmentation method based on saliency map guidance and uncertainty semantic enhancement | |
| CN109493370A (en) | A kind of method for tracking target based on spatial offset study | |
| Lorsakul et al. | Traffic sign recognition using neural network on OpenCV: Toward intelligent vehicle/driver assistance system | |
| US12437382B2 (en) | Device and computer-implemented method for evaluating a control of a generator for determining pixels of a synthetic image | |
| CN119274167B (en) | A 3D target tracking method for multimodal autonomous driving based on spatiotemporal fusion | |
| US12505554B2 (en) | Device and computer-implemented method for determining pixels of a synthetic image | |
| Dubey et al. | Guidance System for Visually Impaired Persons Using Deep Learning and Optical Flow | |
| US20230386046A1 (en) | Device and computer-implemented method for determining pixels of a synthetic image |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| AS | Assignment |
Owner name: ROBERT BOSCH GMBH, GERMANY Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SCHOENFELD, EDGAR;KHOREVA, ANNA;BORGES, JULIO;SIGNING DATES FROM 20230525 TO 20230703;REEL/FRAME:064173/0230 |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |