Abstract:
Methods and systems for disentangled data generation include accessing (310) a dataset including pairs, each formed from a given input text structure and a given style label for the input text structures. An encoder is trained (360) to disentangle a sequential text input into disentangled representations, including a content embedding and a style embedding, based on a subset of the dataset, using an objective function that includes a regularization term that minimizes mutual information between the content embedding and the style embedding. A generator is trained (360) to generate a text output that includes content from the style embedding, expressed in a style other than that represented by the style embedding of the text input.
Abstract:
A computer-implemented method is provided for disentangled data generation. The method includes accessing (410), by a variational autoencoder, a plurality of supervision signals. The method further includes accessing (420), by the variational autoencoder, a plurality of auxiliary tasks that utilize the supervision signals as reward signals to learn a disentangled representation. The method also includes training (430) the variational autoencoder to disentangle a sequential data input into a time-invariant factor and a time- varying factor using a self-supervised training approach which is based on outputs of the auxiliary tasks obtained by using the supervision signals to accomplish the plurality of auxiliary tasks.
Abstract:
Systems and methods for predicting T-Cell receptor (TCR)-peptide interaction, including training a deep learning model for the prediction of TCR-peptide interaction by determining a multiple sequence alignment (MSA) for TCR-peptide pair sequences from a dataset of TCR-peptide pair sequences using a sequence analyzer, building TCR structures and peptide structures using the MSA and corresponding structures from a Protein Data Bank (PDB) using a MODELLER, and generating an extended TCR-peptide training dataset based on docking energy scores determined by docking peptides to TCRs using physical modeling based on the TCR structures and peptide structures built using the MODELLER. TCR-peptide pairs are classified and labeled as positive or negative pairs using pseudo-labels based on the docking energy scores, and the deep learning model is iteratively retrained based on the extended TCR-peptide training dataset and the pseudo- labels until convergence.
Abstract:
A method is provided for generating new binding peptides to MHC proteins includes training (430), by a processor device, a Generative Adversarial Network GAN having a generator and a discriminator only on a set of binding peptide sequences given training data including the set of binding peptide sequences and a set of non-binding peptide sequences. A GAN training objective includes the discriminator being iteratively updated to distinguish generated peptide sequences from sampled binding peptide sequences as fake or real and the generator being iteratively updated to fool the discriminator. The training includes optimizing (440) the GAN training objective while learning two projection vectors for a binding class with two cross-entropy losses. A first loss discriminating binding peptide sequences in the training data from non-binding peptide sequences in the training data. A second loss discriminating generated binding peptide sequences from non-binding peptide sequences in the training data.
Abstract:
Aspects of the present disclosure describe techniques for training a convolutional neural network using an inconsistent stochastic gradient descent (ISGD) algorithm. Training effort for training batches used by the ISGD algorithm are dynamically adjusted according to a determined loss for a given training batch which are classified into two sub states - well-trained or under-trained. The ISGD algorithm provides more iterations for under-trained batches while reducing iterations for well-trained ones.
Abstract:
Systems and methods for matching job descriptions with job applicants is provided. The method includes allocating each of one or more job applicants curriculum vitae (CV) into sections 320; applying max pooled word embedding 330 to each section of the job applicants CVs; using concatenated max-pooling and average-pooling 340 to compose the section embeddings into an applicants CV representation; allocating each of one or more job position descriptions into specified sections 220; applying max pooled word embedding 230 to each section of the job position descriptions; using concatenated max-pooling and average-pooling 240 to compose the section embeddings into a job representation; calculating a cosine similarity 250, 350 between each of the job representations and each of the CV representations to perform job-to-applicant matching; and presenting an ordered list of the one or more job applicants 360 or an ordered list of the one or more job position descriptions 260 to a user.
Abstract:
A computer-implemented method is provided for disentangled data generation. The method includes accessing (310), by a bidirectional Long Short-Term Memory (LSTM) with a multi-head attention mechanism, a dataset including a plurality of pairs each formed from a given one of a plurality of input text structures and given one of a plurality of style labels for the plurality of input text structures. The method further includes training (320) the bidirectional LSTM as an encoder to disentangle a sequential text input into disentangled representations comprising a content embedding and a style embedding based on a subset of the dataset. The method also includes training (350) a unidirectional LSTM as a decoder to generate a next text structure prediction for the sequential text input based on previously generated text structure information and a current word, from a disentangled representation with the content embedding and the style embedding.
Abstract:
Mobile phones and methods for mobile phone failure prediction include receiving respective log files from one or more mobile phone components, including at least one user application. The log files have heterogeneous formats. A likelihood of failure of one or more mobile phone components is determined based on the received log files by clustering the plurality of log files according to structural log patterns and determining feature representations of the log files based on the log clusters. A user is alerted to a potential failure if the likelihood of component failure exceeds a first threshold. An automatic system control action is performed if the likelihood of component failure exceeds a second threshold.
Abstract:
A system for binding peptide search for immunotherapy is presented. The system includes employing (101) a deep neural network to predict a peptide presentation given Major Histocompatibility Complex allele sequences and peptide sequences, training (103) a Variational Autoencoder (VAE) to reconstruct peptides by converting the peptide sequences into continuous embedding vectors, running (105) a Monte Carlo Tree Search to generate a first set of positive peptide vaccine candidates, running (105) a Bayesian Optimization search with the trained VAE and a Backpropagation search with the trained VAE to generate a second set of positive peptide vaccine candidates, using (107) a sampling from a Position Weight Matrix (sPWM) to generate a third set of positive peptide vaccine candidates, screening and merging (109) the first, second, and third sets of positive peptide vaccine candidates, and outputting (111) qualified peptides for immunotherapy from the screened and merged sets of positive peptide vaccine candidates.
Abstract:
A method for learning disentangled representations of videos is presented. The method includes feeding (1001) each frame of video data into an encoder to produce a sequence of visual features, passing (1003) the sequence of visual features through a deep convolutional network to obtain a posterior of a dynamic latent variable and a posterior of a static latent variable, sampling (1005) static and dynamic representations from the posterior of the static latent variable and the posterior of the dynamic latent variable, respectively, concatenating (1007) the static and dynamic representations to be fed into a decoder to generate reconstructed sequences, and applying (1009) three regularizes to the dynamic and static latent variables to trigger representation disentanglement. To facilitate the disentangled sequential representation learning, orthogonal factorization in generative adversarial network (GAN) latent space is leveraged to pre-train a generator as a decoder in the method.