Abstract:
A training system, a training method, and a recognition system are provided. The training method is used to train a neural network module including: an encoder module, a shared decoder module, a synthesis module, and a classification module. The training method includes performing in a training epoch: repeatedly executing: taking a training image from a training set as an input image, obtaining a first loss based on training feature images of the training image and the feature images corresponding to the training image, and obtaining a second loss based on a classification marker of the training image and a classification generated by the classification module in correspondence with the training image; and updating first parameters and second parameters based on an average value of all the first losses and an average value of all the second losses obtained in the preceding step and an update algorithm.
Abstract:
The present invention provides a compression method, wherein the compression method includes the steps of: setting a quantization function; setting a plurality of scaling ratios; receiving image data; for a block of a frame of the image data, using a conversion matrix to convert data of the block to generate a plurality of converted data; determining a specific scaling ratio from the plurality of scaling ratios according to the plurality of converted data; and using the quantization function to perform a quantization operation on a plurality of adjusted data to generate compressed data, wherein the plurality of adjusted data are generated according to the plurality of converted data and the specific scaling ratio.
Abstract:
A method for training a deep learning network for face recognition includes: utilizing a face landmark detector to perform face alignment processing on at least one captured image, thereby outputting at least one aligned image; inputting the at least one aligned image to a teacher model to obtain a first output vector; inputting the at least one captured image a student model corresponding to the teacher module to obtain a second output vector; and adjusting parameter settings of the student model according to the first output vector and the second output vector.
Abstract:
A motion image integration method includes acquiring a raw image, detecting a first motion region image and a second motion region image by using a motion detector according to the raw image, merging the first motion region image with the second motion region image for generating a motion object image according to a relative position between the first motion region image and the second motion region image, and cropping the raw image to generate a sub-image corresponding to the motion object image according to the motion object image. A range of the motion object image is greater than or equal to a total range of the first motion region image and the second motion region image. Shapes of the first motion region image, the second motion region image, and the motion object image are polygonal shapes.
Abstract:
An image contrast enhancement method and an apparatus thereof are disclosed, which calculate the degree of influencing the clarity according to the influence feature (e.g., heavy fog, dust, smoke, or etc.) in the image, and then adjust the brightness of the pixels corresponding to features of influencing the clarity according to the degree, thereby enhancing image contrast and removing phenomenon of influencing the clarity in the image.
Abstract:
A method for generating a target gain value of a wide dynamic range (WDR) operation is disclosed including: acquiring an average bright portion luminance corresponding to an average pixel luminance of a bright portion of a video image; acquiring an average dark portion luminance corresponding to an average pixel luminance of a dark portion of the video image; generating an initial gain value of the WDR operation according to a difference between the average bright portion luminance and the average dark portion luminance; and adjusting the initial gain value according to at least one of a color temperature and an exposure duration configuration value of the video image to generate the target gain value.
Abstract:
An image signal processing method includes: receiving an original color filter array (CFA) image and a pixel binned CFA image; computing a specific information of the pixel binned CFA image; and processing the original CFA image according to the specific information. The associated image signal processor includes an input terminal, an operating unit and a processing unit, wherein the input terminal is for receiving an original CFA image and a pixel binned CFA image, the operating unit is for computing a specific information of the pixel binned CFA image, and the processing unit is for processing the original CFA image according to the specific information and utilizing the pixel binned CFA image.
Abstract:
An image correction method arranged for processing an original image to obtain a corrected image includes steps: receiving the original image from an image sensor; regarding each pixel of the original image, calculating a horizontal distance and a vertical distance between the pixel and a reference point in the original image; determining a horizontal ratio parameter and a vertical ratio parameter according to the horizontal distance and the vertical distance between the pixel and the reference point in the original image; and performing an approximately non-linear regression calculation on the horizontal ratio parameter, the vertical ratio parameter and a coordinate of the pixel to obtain a position of the pixel in the corrected image.
Abstract:
A neural network system and a signal processing method are provided. The neural network system includes at least one processing unit and a neural network module. The signal processing method includes: inputting a neural network input to the neural network module by the processing unit to generate an input at a previous layer of each convolutional transformer layer; performing pointwise convolution on the input by a key embedding layer based on key convolutional kernels to output a key tensor; performing convolution on the input by a value embedding layer based on value convolutional kernels to output a value tensor; performing a convolution on the cascading tensor of a first tensor and the key tensor by an attention embedding layer based on attention convolution kernels to output an attention tensor; and outputting an output tensor based on the attention tensor and the value tensor by an output module.
Abstract:
The present invention provides a detection circuit including a neural network module and a calculation circuit is disclosed. The neural network module is configured to receive an image to generate an output tensor, wherein the output tensor includes position information of a specific object and distance adjustment information. The calculation circuit is coupled to the neural network module, and is configured to calculate an initial distance between an image capture device and the specific object according to the position information of the specific object, and generate an estimated distance according to the initial distance and the distance adjustment information.