Inventors:
- Menlo Park CA, US
Fan WU - Redwood City CA, US
Edith DALLARD - San Mateo CA, US
Linyan MEI - Heverlee, BE
Huseyin Ekin SUMBUL - San Francisco CA, US
International Classification:
G06N 3/063
G06N 3/04
G06N 3/08
Abstract:
A neural network accelerator includes a first memory device, a controller connected to the first memory device through a high-bandwidth (e.g., three-dimensional) interconnect, a configurable processing element (PE) array connected to the first memory device through a first data bus and including a two-dimensional (2D) array of PEs, a local memory connected to the controller and connected, through a second data bus, to the configurable PE array. The controller is configured to, during execution of a neural network (NN), dynamically configure the neural network accelerator for executing each NN layer of a plurality of NN layers of the neural network by selecting either weights of a weight tensor or input data of an input tensor of a tensor operation of the NN layer to store into the local memory, and configuring input and output connections of PEs in the 2D array of PEs for performing the tensor operation.