TECHNICAL SPECIFICATION MATCHING
    1.
    发明申请

    公开(公告)号:WO2022225806A1

    公开(公告)日:2022-10-27

    申请号:PCT/US2022/024995

    申请日:2022-04-15

    Abstract: Systems and methods are provided for detail matching. The method includes training a feature classifier (200) to identify technical features, and training a neural network model for a trained importance calculator (300) to calculate an importance value for each identified technical feature. The method further includes receiving a specification sheet (110) including a plurality of technical features, and receiving a plurality of descriptive sheets (120) each including a plurality of technical features. The method further includes identifying the technical features (130) in the specification sheet and the plurality of descriptive sheets using the trained feature classifier (200), and calculating an importance (140) for each identified technical feature using the trained feature importance calculator (300). The method further includes calculating a matching score (150) between the identified technical features of the specification sheet and the identified technical features of the plurality of descriptive sheets based on the importance of each identified technical feature.

    MULTI-FACETED KNOWLEDGE-DRIVEN PRE-TRAINING FOR PRODUCT REPRESENTATION LEARNING

    公开(公告)号:WO2022169656A1

    公开(公告)日:2022-08-11

    申请号:PCT/US2022/013982

    申请日:2022-01-27

    Abstract: A method for employing a knowledge-driven pre-training framework for learning product representation is presented. The method includes learning (1001) contextual semantics of a product domain by a language acquisition stage including a context encoder and two language acquisition tasks, obtaining (1003) multi-faceted product knowledge by a knowledge acquisition stage including a knowledge encoder, skeleton attention layers, and three heterogeneous embedding guided knowledge acquisition tasks, generating (1005) local product representations defined as knowledge copies (KC) each capturing one facet of the multi-faceted product knowledge, and generating (1007) final product representation during a fine-tuning stage by combining all the KCs through a gating network.

    COMPUTER CODE REFACTORING
    3.
    发明申请

    公开(公告)号:WO2022245590A1

    公开(公告)日:2022-11-24

    申请号:PCT/US2022/028535

    申请日:2022-05-10

    Abstract: Systems and methods are provided for automated computer code editing. The method includes training a code-editing neural network model (320) using a corpus of code editing data samples, including the pre-editing samples (110) and post-editing samples (120), and parsing (130) the pre-editing samples and post-editing samples into an Abstract Syntax Tree (AST). The method further includes using a grammar specification to transform (130) the AST tree into a unified Abstract Syntax Description Language (ASDL) graph for different programming languages, and using a gated graph neural network (GGNN) (320) to compute a vector representation (140, 150) for each node in the unified Abstract Syntax Description Language (ASDL) graph. The method further includes selecting and aggregating (160) support samples based on a query code with a multi-extent ensemble method, and altering the query code (170) iteratively using the pattern learned from the pre- and post-editing samples.

    SELF-LEARNING FRAMEWORK OF ZERO-SHOT CROSS-LINGUAL TRANSFER WITH UNCERTAINTY ESTIMATION

    公开(公告)号:WO2022240557A1

    公开(公告)日:2022-11-17

    申请号:PCT/US2022/025481

    申请日:2022-04-20

    Abstract: A method provided for cross-lingual transfer trains (230) a pre-trained multi-lingual language model based on a gold labeled training set in a source language to obtain a trained model. The method assigns (240) each sample in an unlabeled target language set to a silver label according to a model prediction by the trained model to obtain set of silver labels, and performs (250) uncertainty-aware label selection based on the silver label assigned to each sample according to the model prediction and the trained model to obtain selected silver labels. The method performs iterative training (260) on the selected labels by applying the selected silver labels in the target language set as training labels and re-training the trained model with the gold labels and the selected silver labels to obtain an iterative model, and performs (270) task-specific result prediction in target languages based on the iterative model to generate a final predicted result in target languages.

Patent Agency Ranking