Invention Grant
- Patent Title: Dynamic sequencing of data partitions for optimizing memory utilization and performance of neural networks
-
Application No.: US17583499Application Date: 2022-01-25
-
Publication No.: US11722147B2Publication Date: 2023-08-08
- Inventor: Kent D. Cedola , Larry Marvin Wall , Boris Bobrov , George Petre , Chad Balling McBride , Amol Ashok Ambardekar
- Applicant: Microsoft Technology Licensing, LLC
- Applicant Address: US WA Redmond
- Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
- Current Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
- Current Assignee Address: US WA Redmond
- Agency: Newport IP, LLC
- Agent Han K. Gim
- Main IPC: H03M7/30
- IPC: H03M7/30 ; G06N3/04 ; G06N3/063 ; G06F12/0862 ; G06F9/46 ; G06F1/324 ; G06F3/06 ; G06F9/38 ; G06F12/08 ; G06F12/10 ; G06F15/80 ; G06F17/15 ; G06N3/049 ; G06N3/06 ; G06N3/08 ; G06N3/10 ; H04L45/02 ; H04L67/02 ; G06F9/30 ; H04L67/1001 ; G06F9/48 ; G06F12/02 ; G06N3/045 ; G06N3/065 ; G06F13/16 ; G06F1/3234 ; G06F13/28 ; H03M7/46 ; H04L45/50

Abstract:
Optimized memory usage and management is crucial to the overall performance of a neural network (NN) or deep neural network (DNN) computing environment. Using various characteristics of the input data dimension, an apportionment sequence is calculated for the input data to be processed by the NN or DNN that optimizes the efficient use of the local and external memory components. The apportionment sequence can describe how to parcel the input data (and its associated processing parameters—e.g., processing weights) into one or more portions as well as how such portions of input data (and its associated processing parameters) are passed between the local memory, external memory, and processing unit components of the NN or DNN. Additionally, the apportionment sequence can include instructions to store generated output data in the local and/or external memory components so as to optimize the efficient use of the local and/or external memory components.
Public/Granted literature
Information query
IPC分类: