-
公开(公告)号:US20230169309A1
公开(公告)日:2023-06-01
申请号:US17992775
申请日:2022-11-22
Applicant: DALIAN UNIVERSITY OF TECHNOLOGY
Inventor: Xin YANG , Xiaopeng WEI , Li ZHU , Xirong XU , Chenming DUAN
IPC: G06N3/042 , G06N3/0442
CPC classification number: G06N3/042 , G06N3/0442
Abstract: The present invention belongs to the technical field of knowledge graph, and provides a knowledge graph construction method for an ethylene oxide derivatives production process. According to data types and characteristics, data sources of the ethylene oxide derivatives production process are sorted and divided into three types: structural data, unstructured data and other types of data. An ontology layer and a data layer of a knowledge graph are constructed by combining top-down and bottom-up methods. A data-driven incremental ontology modeling method is proposed to ensure the expandability of the knowledge graph. For structural knowledge extraction, the safety of original data storage is ensured by means of virtual knowledge graph, and a new mapping mechanism is proposed to realize data materialization. For unstructured knowledge extraction, an entity extraction task is realized on the basis of a BERT-BiLSTM-CRF named entity recognition model by integrating a pre-training language model BERT.
-
公开(公告)号:US20220230322A1
公开(公告)日:2022-07-21
申请号:US17336702
申请日:2021-06-02
Applicant: DALIAN UNIVERSITY OF TECHNOLOGY
Inventor: Wen DONG , Xin YANG , Haiyang MEI , Xiaopeng WEI , Qiang ZHANG
Abstract: The invention belongs to scene segmentation's field in computer vision and is a depth-aware method for mirror segmentation. PDNet successively includes a multi-layer feature extractor, a positioning module, and a delineating module. The multi-layer feature extractor uses a traditional feature extraction network to obtain contextual features; the positioning module combines RGB feature information with depth feature information to initially determine the position of the mirror in the image; the delineating module is based on the image RGB feature information, combined with depth information to adjust and determine the boundary of the mirror. This method is the first method that uses both RGB image and depth image to achieve mirror segmentation in an image. The present invention has also been further tested. For mirrors with a large area in a complex environment, the PDNet segmentation results are still excellent, and the results at the boundary of the mirrors are also satisfactory.
-
公开(公告)号:US20240355140A1
公开(公告)日:2024-10-24
申请号:US18038611
申请日:2022-06-20
Applicant: DALIAN UNIVERSITY OF TECHNOLOGY
Inventor: Xin YANG , Xiaopeng WEI , Bo DONG , Haiwei ZHANG
CPC classification number: G06V40/20 , G06T7/20 , G06T7/70 , G06V10/764 , G06V10/7715 , G06V10/82 , G06V40/193 , G06T2207/30201
Abstract: The present invention belongs to the technical field of computer vision, and proposes a lightweight real-time emotion analysis method incorporating eye tracking. In the method, gray frames and event frames that have synchronized time are acquired through event-based cameras and respectively input to a frame branch and an event branch; the frame branch extracts spatial features by convolution operations, and the event branch extracts temporal features through conv-SNN blocks; the frame branch has a guide attention mechanism for the event branch; and the spatial features and the temporal features are integrated by fully connected layers, The final output is the average of the n fully connected layer outputs, which represents the final expression. The method can recognize the emotional expression of any stage in various complex light changing scenarios; and in the case of limited accuracy loss, the emotion recognition time is shortened to achieve “real-time” user emotion analysis.
-
4.
公开(公告)号:US20240295879A1
公开(公告)日:2024-09-05
申请号:US18464917
申请日:2023-09-11
Applicant: DALIAN UNIVERSITY OF TECHNOLOGY
Inventor: Xin YANG , Xuefeng YIN , Baocai YIN , Xiaopeng WEI
CPC classification number: G05D1/0274
Abstract: The present invention proposes an active scene mapping method based on constraint guidance and space optimization strategies, comprising a global planning stage and a local planning stage; in the global planning stage, the next exploration goal of a robot is calculated to guide the robot to explore a scene; and after the next exploration goal is determined, specific actions are generated according to the next exploration goal, the position of the robot and the constructed occupancy map in the local planning stage to drive the robot to go to a next exploration goal, and observation data is collected to update the information of the occupancy map. The present invention can effectively avoid long-distance round trips in the exploration process so that the robot can take account of information gain and movement loss in the exploration process, find a balance of exploration efficiency, and realize the improvement of active mapping efficiency.
-
公开(公告)号:US20240028036A1
公开(公告)日:2024-01-25
申请号:US18373623
申请日:2023-09-27
Applicant: DALIAN UNIVERSITY OF TECHNOLOGY
Inventor: Xin YANG , Xiaopeng WEI , Yang WANG , Qiang ZHANG
CPC classification number: G05D1/0088 , G05D1/0248 , G05D1/0214 , G06V10/82 , G06V10/80 , G06V20/58
Abstract: The present invention provides a robot dynamic obstacle avoidance method based on a multimodal spiking neural network. The present invention realizes a robot obstacle avoidance method in a dynamic environment by fusing laser radar data and processed event camera data and combining with the intrinsic learnable threshold of the spiking neural network for a scenario comprising dynamic obstacles. It solves the difficulty of failure of obstacle avoidance due to the difficulty in perceiving the dynamic obstacles in the obstacle avoidance task of a robot. The present invention helps the robot to fully perceive the static information and the dynamic information of the environment, uses the learnable threshold mechanism of the spiking neural network for efficient reinforcement learning training and decision making, and realizes autonomous navigation and obstacle avoidance in the dynamic environment. An event data enhanced model is combined to better adapt to the dynamic environment for obstacle avoidance.
-
6.
公开(公告)号:US20230236587A1
公开(公告)日:2023-07-27
申请号:US17992791
申请日:2022-11-22
Applicant: DALIAN UNIVERSITY OF TECHNOLOGY
Inventor: Xin YANG , Xiaopeng WEI , Li ZHU , Xirong XU , Zitao YIN
IPC: G05B23/02
CPC classification number: G05B23/0229 , G05B23/027 , G05B2223/02
Abstract: The present invention provides a data fusion and reconstruction method for fine chemical industry safety production based on a virtual knowledge graph. In view of the characteristics of fine chemical industry safety production data, such as a large amount of structured data, a multi-source heterogeneous database and a strong sequential logic, the present invention innovatively proposes a method of using a virtual knowledge graph to complete the fusion and reconstruction of a traditional database for fine chemical industry. The present invention fuses static structured knowledge in the field of fine chemical industry with a real-time dynamic database for chemical industry safety production in the concept of ontologies for the first time to organize time series data in the form of entities. In addition, the mapping rules of the existing OBDA system are improved based on a data set of the present invention.
-
公开(公告)号:US20220230324A1
公开(公告)日:2022-07-21
申请号:US17336748
申请日:2021-06-02
Applicant: DALIAN UNIVERSITY OF TECHNOLOGY
Inventor: Xin YANG , Haiyang MEI , Wen DONG , Xiaopeng WEI , Dengping FAN
Abstract: A method for segmenting a camouflaged object image based on distraction mining is disclosed. PFNet successively includes a multi-layer feature extractor, a positioning module, and a focusing module. The multi-layer feature extractor uses a traditional feature extraction network to obtain different levels of contextual features; the positioning module first uses RGB feature information to initially determine the position of the camouflaged object in the image; the focusing module mines the information and removes the distraction information based on the image RGB feature information and preliminary position information, and finally determines the boundary of the camouflaged object step by step. The method of the present invention introduces the concept of distraction information into the problem of segmentation of the camouflaged object and develops a new information exploration and distraction information removal strategy to help the segmentation of the camouflaged object image.
-
公开(公告)号:US20220212339A1
公开(公告)日:2022-07-07
申请号:US17564588
申请日:2021-12-29
Applicant: DALIAN UNIVERSITY OF TECHNOLOGY
Inventor: Xin YANG , Boyan WEI , Baocai YIN , Qiang ZHANG , Xiaopeng WEI , Zhenjun DU
IPC: B25J9/16
Abstract: The present invention belongs to the technical field of computer vision and provides a data active selection method for robot grasping. The core content of the present invention is a data selection strategy module, which shares the feature extraction layer of backbone main network and integrates the features of three receptive fields with different sizes. While making full use of the feature extraction module, the present invention greatly reduces the amount of parameters that need to be added. During the training process of the main grasp method detection network model, the data selection strategy module can be synchronously trained to form an end-to-end model. The present invention makes use of naturally existing labeled and unlabeled labels, and makes full use of the labeled data and the unlabeled data. When the amount of the labeled data is small, the network can still be more fully trained.
-
公开(公告)号:US20220148292A1
公开(公告)日:2022-05-12
申请号:US17257704
申请日:2020-03-13
Applicant: DALIAN UNIVERSITY OF TECHNOLOGY
Inventor: Xin YANG , Xiaopeng WEI , Qiang ZHANG , Haiyang MEI , Yuanyuan LIU
IPC: G06V10/774 , G06V10/77 , G06V10/80 , G06V10/70 , G06V10/776 , G06V10/44
Abstract: The invention discloses a method for glass detection in a real scene, which belongs to the field of object detection. The present invention designs a combination method based on LCFI blocks to effectively integrate context features of different scales. Finally, multiple LCFI combination blocks are embedded into the glass detection network GDNet to obtain large-scale context features of different levels, thereby realize reliable and accurate glass detection in various scenarios. The glass detection network GDNet in the present invention can effectively predict the true area of glass in different scenes through this method of fusing context features of different scales, successfully detect glass with different sizes, and effectively handle with glass in different scenes. GDNet has strong adaptability to the various glass area sizes of the images in the glass detection dataset, and has the highest accuracy in the field of the same type of object detection.
-
公开(公告)号:US20210216806A1
公开(公告)日:2021-07-15
申请号:US16963140
申请日:2020-05-13
Applicant: DALIAN UNIVERSITY OF TECHNOLOGY
Inventor: Xin YANG , Xiaopeng WEI , Qiang ZHANG , Yuhao LIU , Yu QIAO
Abstract: The invention belongs to the field of computer vision technology, and provides a fully automatic natural image matting method. For image matting of a single image, it is mainly composed of the extraction of high-level semantic features and low-level structural features, the filtering of pyramid features, the extraction of spatial structure information, and the late optimization of the discriminator network. The invention can generate accurate alpha matte without any auxiliary information, saving the time for scientific researchers to mark auxiliary information and the interaction time when users use it.
-
-
-
-
-
-
-
-
-