Invention Grant
- Patent Title: Exploiting sparseness in training deep neural networks
- Patent Title (中): 在深层神经网络训练中利用稀疏性
-
Application No.: US13305741Application Date: 2011-11-28
-
Publication No.: US08700552B2Publication Date: 2014-04-15
- Inventor: Dong Yu , Li Deng , Frank Torsten Bernd Seide , Gang Li
- Applicant: Dong Yu , Li Deng , Frank Torsten Bernd Seide , Gang Li
- Applicant Address: US WA Redmond
- Assignee: Microsoft Corporation
- Current Assignee: Microsoft Corporation
- Current Assignee Address: US WA Redmond
- Agent Steve Wight; Carole Boelitz; Micky Minhas
- Main IPC: G06F15/18
- IPC: G06F15/18 ; G06N3/08

Abstract:
Deep Neural Network (DNN) training technique embodiments are presented that train a DNN while exploiting the sparseness of non-zero hidden layer interconnection weight values. Generally, a fully connected DNN is initially trained by sweeping through a full training set a number of times. Then, for the most part, only the interconnections whose weight magnitudes exceed a minimum weight threshold are considered in further training. This minimum weight threshold can be established as a value that results in only a prescribed maximum number of interconnections being considered when setting interconnection weight values via an error back-propagation procedure during the training. It is noted that the continued DNN training tends to converge much faster than the initial training.
Public/Granted literature
- US20130138589A1 EXPLOITING SPARSENESS IN TRAINING DEEP NEURAL NETWORKS Public/Granted day:2013-05-30
Information query