sensor data processing - stanford university · pdf filesensor data processing ... army night...

21
Sensor Data Processing Dr. Dan Hammerstrom Program Manager, MTO April 2014 Distribution A: Approved for Public Release; Distribution Unlimited

Upload: trandang

Post on 21-Mar-2018

219 views

Category:

Documents


3 download

TRANSCRIPT

Sensor Data Processing

Dr. Dan Hammerstrom Program Manager, MTO

April 2014

Distribution A: Approved for Public Release; Distribution Unlimited

2

•  We need to process massive amounts of sensor data of all different kinds •  Current computational approaches are very limited in processing such information •  This is particularly true in low power, embedded applications that have limited computational power and memory

•  Neural inspired architectures are being explored as one potential solution to this problem

•  The potential of neurocomputing remains underappreciated, since there are still so few successful applications that are built primarily from neural components

•  If we are going to leverage neural based computation, we need to better articulate the unique capabilities neural models bring to these applications

•  In this talk I will discuss the ”neural like” DARPA UPSIDE program as one potential solution to the sensor data crisis, I will also “hint” at possible approaches to the “Back End” problem

A Sensor Data Crisis

4/24/14

Front End Signal

Processing

Feature Extraction

Higher Order Feature

Extraction

Association Inference

Motor Control

Motor Control

Subprogram

Motor Control

Programs

Decision Making

The “Front End” UPSIDE The “Back End”

Distribution A: Approved for Public Release; Distribution Unlimited

PROGRAM STATUS

PROGRAM OVERVIEW PROGRAM OUTCOMES

Unconventional Processing of Signals for Intelligent Data Exploitation (UPSIDE)

Capability Objectives: •  Address critical gap between ISR sensor capabilities and the

onboard, real-time processing available for image data analysis, such as target recognition and tracking

•  Move beyond the inherent power constraints of scaled digital CMOS (Moore’s Law) to produce a new asymmetry in capability of future weapon systems

•  1,000X improvement in computing performance, 10,000X increase in power efficiency over traditional solutions

•  Produce systems capable of tracking >10,000 moving targets in real time (current state of art ~100’s)

Transition: •  USAF/AFRL, Army Night Vision Lab, MITRE

Exploit non-Boolean computation based on probabilistic inference and emerging non-digital / non-CMOS devices, to enhance the performance and power efficiency of real-time ISR systems and other power constrained applications

Emerging Devices such as the “Coupled Spin Torque Oscillator” provide pathway to 7 orders of magnitude in additional computing efficiency for real-time ISR Applications

Next Generation Persistent Video Surveillance: Performance requirements significantly exceed current computing capability (Track thousands of moving targets in real-time) mostly due to power constraints

Probabilistic inference through energy minimization:

Leveraging emerging devices for ISR processing:

3

Inference module for determining the most likely

feature from an energy minimizing matching process.

The degree of feature match is an estimate of the probability

that the input is a noisy version of the feature associated with

closest attractor basin or energy minimum.

BAE Systems Columbia Maryland University of Massachusetts Amherst, Massachusetts Johns Hopkins Baltimore, Maryland UCSB Santa Barbara, California Stony Brook University Long Island, New York SEMATECH Albany, New York

HRL Laboratories Malibu California Purdue University West Lafayette, Indiana University of Notre Dame South Bend, Indiana University of Pittsburgh Pittsburgh, Pennsylvania Intel Corp. Santa Clara, California NIST Gaithersburg, Maryland

University of Michigan Ann Arber Michigan Portland State University Portland, Oregon New Mexico Consortium, Los Alamos Lab New Mexico

University of Tennessee Knoxville Tennessee Oak Ridge National Laboratory Oak Ridge, Tennesse Stanford

PERFORMERS

Distribution A: Approved for Public Release; Distribution Unlimited

0.000001

0.00001

0.0001

0.001

0.01

0.1

1

10

100

1000

2000 2005 2010 2015 2020

Gig

a-Fe

atur

es*/

wat

ts

Year

4

Problem: The Computer Efficiency Gap Is Increasing

Data increasing, analysis and compression requirements increasing …

* Gabor Filter Requires 136 operations/pixel : J.S. Marques et al. “Pattern Recognition and Image Analysis” Second Iberian conference Vol. 1, pp. 335-342 (2005)

COTS Efficiency Ceiling

Processor Capability

Distribution A: Approved for Public Release; Distribution Unlimited

5

Solution: Non-Digital, Probabilistic Computing Reduces Hierarchy

Digital architectures are not well matched to feature extraction from sensor images

•  Images are inherently analog and low precision •  Digital algorithms are created to search for

image structures based on existing digital number crunching architectures

•  Digital abstractions limit data analysis •  100s to 1,000s of digital operations per image

activity •  Wasted energy in excess operations, data

movement and precision

Need new computing approaches matched to image processing

•  Use the physics of new emerging devices to extract features.

•  Data naturally represented in low precision and sparse form are more suitable to devices and efficient for data transfer

Multiple Cores

Transistors

Boolean Logic Gates

Functional Units

Processor

Software Layers (OS, drivers, …)

Programs

Larg

e Ef

ficie

ncy

Cost

General Purpose Computing Hierarchy

…. ….

…. ….

….

….

….

….

New Devices with analog behavior

Computational Model

Image Data

Eff

icie

ncy

Gai

n UPSIDE Reduced Hierarchy

…. ….

Computing directly with devices eliminates multiple layers of hierarchy/inefficiency

Distribution A: Approved for Public Release; Distribution Unlimited

6

•  “Distributed” best match association is implemented by a number of interconnected processing elements

•  Collectively these elements constitute the network state, and an “energy” can be defined and computed

•  By setting the connection strengths appropriately, the training vectors (our kernels) become local minima in the “energy” space

•  All kernels are compared simultaneously in one “energy minimizing” operation

•  The best match is to the kernel that is “closest” given our metric – it is also, under certain conditions, the most likely probabilistically (which is “inference”)

•  This is called an “attractor” memory model.

Distributed Best Match Association

attractor state

ener

gy

state x

attractor basin

1 ( , ) ( ) ( )2

E w i j y i y j= − ∑

Constraints are “soft” and can be represented as an energy “field” Which the network tries to minimize

Attractor Memory Model

Steps to Best Match Implementation:

Distribution A: Approved for Public Release; Distribution Unlimited

• Spontaneous synchronization corresponds to a stable minimum in an energy space

Key Concept: Association Via Emerging Devices: The Hardware is the Algorithm

• The system converges to the

“closest” energy minimum, which is generally the most likely vector in a Bayesian sense

Device Physics Interacting Nonlinear Systems Bayesian Inference

Nano-device Array

• Coupled devices operating at very low power levels

• The coupling coefficients create a controlled energy surface with specific local minima that models a probability distribution

EDs

Average

7

attractor basins

attractor states

Bayesian Parallel Processing

attractor state

ener

gy

state x

attractor basin

ED ED ED

ED ED ED

ED ED ED

ED ED ED

ED

ED

ED

ED

ED

ED

ED

ED

ED

ED

ED

ED

ED

ED

ED ED ED ED

Associative memories approximate probabilistic inference, Giga-ops Performance and micro-watts of power

Distribution A: Approved for Public Release; Distribution Unlimited

8

Solution: A New Computational Model and Implementation

New Paradigm – Non-Boolean, Probabilistic Computing

1.  Computing occurs by the physics of the devices (highly parallel)

2.  Devices perform the computational equivalent of hundreds of discrete digital operations

3.  The model can be configured into hierarchies that accomplish most of the computational work required by the application

Example: Find Features in Sensor Data (7x7 Gabor Edge Finding, 10 Giga-pixel Array)

Boolean Computation •  Processor: Intel 6 Core i7, GOPS: 6.7

•  1 inference is 140 operations/kernel, 24 kernels are compared / pixel

•  GOPs/watt: 0.1

•  Compute time = 7,700 sec

•  460 kilo-joules (60 watts for 7700 seconds)

Analog Direct Device Computation •  Processor: 10 X 10 Array of coupled oscillators Giga-

Inferences/sec = 400 (56k GOPS equivalent)

•  Compute time = 0.04 sec

•  430 milli-joules

Sensor Data: Active Edges located (in red)

Distribution A: Approved for Public Release; Distribution Unlimited

0.000001

0.00001

0.0001

0.001

0.01

0.1

1

10

100

1000

10000

100000

2000 2005 2010 2015 2020

Gig

a-Fe

atur

es*/

wat

ts

Year

9

UPSIDE: Performance Goals

UPSIDE: Mixed Signal CMOS

UPSIDE: Emerging Device

COTS Efficiency Ceiling

~100 Ops/nsec consuming µwatts

DoD Sensing Requirements (Notional)

Processor Capability

UPSIDE Goals: 3 orders of magnitude in throughput, 4 orders of magnitude in power efficiency, no loss in accuracy

Distribution A: Approved for Public Release; Distribution Unlimited

10

UPSIDE Program Tasks

UPSIDE Task 1 IM Development &

IPP Implementation

Image Processing Pipeline An application driver

UPSIDE Task 2 Mixed signal CMOS implementation of

computational model and system test bed

UPSIDE Task 3 Image processing

demonstration combining next-generation devices with

new computation model

Sensor/ Filter Classification

Maximum Entropy data reduction Feature Extraction

Object Classification

Intermediate – The Most Compute Intensive Back End Sensor Data

VCO_P

VCO_NIn

Out

•  Design and Fabricate a mixed signal representation of the Inference Module in state of the art MS CMOS

•  Implement a test bed system using MS CMOS

•  Validate against IPP simulation

•  Simulate the mapping of the Inference Module to specialized devices

•  Determine systems level performance-price

•  Show simple circuit operation

•  Recreate the traditional image processing pipeline (IPP) hierarchies of Inference Modules (IM)

Tadashi Shibata Intel Non-Boolean Project

CoFe2O4 epitaxial nanopillars fabricated on MgO, R. Comes, J. Lu, and S. Wolf, University of Virginia, unpublished

Filtering Feature Extraction

Complex Feature

Extraction

Distribution A: Approved for Public Release; Distribution Unlimited

•  Data become more “knowledge / context” intensive, containing both spatial and temporal information, as they move through the pipeline.

•  The “structure” of the data (high order correlations) carries the information •  Current computational approaches do not adequately represent complex spatial and

temporal data, limiting the ability to effectively perform complex recognition for important DoD tasks such as anomaly detection and scenario prediction

•  And we also need rapid learning

What About the Back End?

Front End Signal

Processing

Feature Extraction

Higher Order Feature

Extraction

Association Inference

Decision Making

The “Front End” UPSIDE Program The “Back End”

Sensor Output

Actionable Data/ Motor Control

11 Distribution A: Approved for Public Release; Distribution Unlimited

12

Can We Better Leverage Biological Computational Principles?

1) Sparse Distributed Representations

3) Hierarchy of memory regions

2) Sequence memory (spatial/temporal patterns)

5) All regions are sensory and motor

6) Attention

retina

cochlea

somatic

data stream

4) On-line learning from streaming data

Six principles have been proposed as fundamental to biological and machine intelligence.

Theoretical principles Simulation

Custom Silicon

Ref: Jeff Hawkins

•  Hierarchical Structure •  Builds rich models with many layers of abstraction

(“patterns of patterns”) •  Feed forward, feedback

•  On-Line Learning •  Adaptation of network based on acquired data

•  Sparse Distributed Representation •  Associative memory model for best match pattern retrieval improving

computational efficiency and scalability

•  Spatial/Temporal Sequence Memory •  Learns complex sequences of data predicting future inputs and

detecting anomalies

Key Ingredients to “Cortical” Algorithm, HTM

Engineer theoretical operating principles of neocortex in hardware.

Anatomy, Physiology

Distribution A: Approved for Public Release; Distribution Unlimited

HTM - Hierarchical Temporal Memory

13

Object Detection: Capture complex structure to allow recognizing parts and their relationships

Structure in the Data

Higher Complexity allows for differentiation of more object classes even when very similar •  “A Bike” vs “Klein Quantum Race” •  “A Car” vs “White 2013 Toyota Tundra” •  “A Person” vs “Joe Smith”

Distribution A: Approved for Public Release; Distribution Unlimited

14

•  “… the front end circuits of the brain … specialize in their own particular visual and auditory inputs, the rest of the brain converts these to random-access encodings in association areas throughout cortex. … these areas take initial sensory information and construct grammars

•  “These are not grammars of linguistic elements, they are grammatical organizations (nested, hierarchical, sequences of categories) of percepts – visual, auditory, …

•  “Processing proceeds by incrementally assembling these constructs … these grammars generate successively larger ‘proto-grammatical fragments,’ eventually constituting full grammars”

•  “They thus are not built in the manner of most hand-made grammars; they are statistically assembled, to come to exhibit rule-like behavior, of the kind expected for linguistic grammars

From Big Brain by Gary Lynch and Rick Granger (Palgrave McMillan 2008):

4/24/14 Distribution A: Approved for Public Release; Distribution Unlimited

15

Processing Video, “HTM” Algorithm

Preprocessed Input – 1 Frame

Input Frame - Casablanca

Input

Hierarchy Object Recognition: P(Bogart|X)

X=

Is this Humphrey Bogart? What is he doing?

Identifying Human Actions and Scenes

•  Using Hollywood 2 DataBase

Dataset with 12 classes of human actions and 1466 video clips.

•  Present sets of sequences (edge-filtered, binarized) to HTM model

•  Characterize/analyze features learned

•  Most current Machine Learning technique do not handle temporal data very well

•  When operational performance comparisons will have to be made with traditional video analysis techniques

Seedling performed by Neurithmic Systems Distribution A: Approved for Public Release; Distribution Unlimited

Mapping to Hardware

HTM requires hardware matched to the algorithm’s needs 1.  High Connectivity (more expensive than the computational requirements) 2.  Local Memory and Parameter Storage 3.  Simple, Low Precision Computation 4.  Configurable / Adaptable 5.  Sparse activity

Hierarchical Temporal Memory

Specialized HTM processor

Custom architectures can address HTM requirements:

•  High-risk exotic devices unnecessary •  Utilize conventional CMOS fabrication optimized for HTM

architecture/computational model •  Can benefit from latest advances in CMOS.

Conventional processors are a poor match:

•  Constrained: processor/memory partition, limited parallelism

•  Excessive: high precision, tiered caches, complex instruction sets, pipelines, etc.

Conventional Solutions

16 Distribution A: Approved for Public Release; Distribution Unlimited

17

Hardware Comparisons

ImageNet2012[1] Deep Learning Performance HTM Processor Estimate[2]

Platform power 490 watts 0.21 watts

Training Time 6 days 20 Minutes

Images / second 2.3 1000

Joules / image 212 0.0004

1.  Krizhevsky, Alex, Ilya Sutskever, and Geoff Hinton. “Imagenet Classification with Deep Convolutional Neural Networks.” In Advances in Neural Information Processing Systems 25, 1106–14, 2012. http://books.nips.cc/papers/files/nips25/NIPS2012_0534.pdf

2.  Zaveri et al., “Performance/Price Estimates for Cortical scale Hardware – A Design Space Exploration”, Journal of Neural Networks 24(2011)

1,000,000x less power

500x faster

Custom Digital Processor •  Assumed at 14nm, SRAM per processor node (PN) •  On-chip routing, realistic cortical model

Modular “Tiled” Architecture

400 PNs per chip 600 nodes / PN 0.21 Watt/chip

Distribution A: Approved for Public Release; Distribution Unlimited

18

Wide Area Motion Imagery (WAMI) •  Scenario Awareness and Prediction

Autonomous Vehicles •  Autonomous decision making •  Motion control •  Task Adaptation

Intelligent Signal Processing and Data Sharing •  Scalable Architecture: Soldier è Large systems •  Sensor Data Fusion and System Collaboration

ARGUS-IS Scenario Awareness

Application Driven

Large-scale DoD application Emphasis on data analysis and adaptation

Embedded neuro-processing for autonomous systems Emphasis on adaptive control and sensor data processing and reduced dependency on hand written code

Embedded On-Platform Computation ONR Large Displacement

Unmanned Underwater Vehicle (LDUUV)

North West Aerospace Alliance UAV - UK

Lockheed SMSS UAV

Distribution A: Approved for Public Release; Distribution Unlimited

19

Example DoD Application – ARGUS IS

© BAE Systems 2013

DoD application example: ARGUS (WAMI, 1.8 Gpixel sensor) seeks to track 100,000 moving objects in a next-generation wide area imaging (WAMI) system (currently, with analyst assistance they are tracking ~100 object simultaneously)

•  Requirement is to identify O(106) complete tracks per day (large urban area)

•  O(106) tracks requires O(104) analysts to process a single day’s data

•  Tracks are identified by combined spatial / temporal features and are a common example of complex objects

•  UPSIDE will increase the load on the back-end

Need to automate the pipeline, reducing burden on human analysts.

Need to mature algorithm(s) and develop scalable hardware to address realistic DoD applications.

ARGUS data with target/track overlay

Distribution A: Approved for Public Release; Distribution Unlimited

20

Complex Sensor Problems

•  Not possible with any neural algorithms today, nor with traditional techniques

•  Creates the capability of using learning systems to model and control complex systems.

Objective: Push the limits of algorithm scaling to introduce machine learning to the most challenging DoD sensor processing problems.

Wide Area Motion Imagery

Surveillance Imaging Scenario Awareness

Data Fusion from Different Sensors

Intelligent Signal Processing and Data Sharing

time

Tracking convoy of vehicles

•  Helps manage signal and system complexity by automating higher order relationships.

Distribution A: Approved for Public Release; Distribution Unlimited

21

•  Conceptually one can think of “computational intelligence” as a spectrum •  Among other things we need to do “small c” cognition before we move on to “Big C”

cognition •  And, while cognition remains our goal, I believe that it is possible to build very useful

systems with the technology we have now •  We have to build this field one brick at a time, moving incrementally to the right

Big “C” Cognition: humans

Small “c” cognition: most mammals

Computing Is currently

here

A long way!

ISP

Increasing Intelligence

Our Goal

A long way! A Rock The Krell

Distribution A: Approved for Public Release; Distribution Unlimited