middleware services for sensors systems in dynamic data...

22
Middleware Services for Sensors Systems in Dynamic Data-driven Oil Applications Manish Parashar The Applied Software Systems Laboratory ECE/CAIP, Rutgers University http://www.caip.rutgers.edu/TASSL (Ack: NSF, DoE)

Upload: others

Post on 25-Apr-2020

2 views

Category:

Documents


0 download

TRANSCRIPT

Middleware Services for Sensors Systems in Dynamic Data-driven Oil Applications

Manish ParasharThe Applied Software Systems Laboratory

ECE/CAIP, Rutgers Universityhttp://www.caip.rutgers.edu/TASSL

(Ack: NSF, DoE)

Outline

• Pervasive Grid Environments and Data-driven Management of Subsurface Geosystems– The Instrumented Oil Field

• Project AutoMate and Meteor: Middleware Services for Sensors Systems

• Deployments and Evaluations

• Concluding Remarks

Pervasive Computational Ecosystems and Dynamic Data Driven Applications

Components dynamically composed.

“WebServices”discovered &

invoked.

Resources discovered, negotiated, co-allocated on-the-fly. components

deployed

Experts query, configure resources

Experts interact and collaborate using ubiquitous and

pervasive portals

Applications& Services

Model A

Model BLaptop

PDA

ComputerScientist

Scientist

Resources

Computers, Storage,Instruments, ...

Data Archive &Sensors

DataArchives

Sensors, Non-Traditional Data

Sources

Experts mine archive, match

real-time data with history

Real-time data assimilation/injection

(sensors, instruments, experiments, data

archives),

Automated mining & matching

Components write into the archive

Experts monitor/interact with/interrogate/steer models (“what if”

scenarios,…). Application notifies experts of interesting events.

Pervasive Grid Environments – Unprecedented Opportunities• Pervasive Grids Environments

– Seamless, secure, on-demand access to and aggregation of, geographically distributed computing, communication and information resources

• Computers, networks, data archives, instruments, observatories, experiments, sensors/actuators, ambient information, etc.

– Context, content, capability, capacity awareness– Ubiquity and mobility

• Knowledge-based, information/data-driven, context/content-aware computationally intensive, pervasive applications– Symbiotically and opportunistically combine services/computations, real-

time information, experiments, observations, and to manage, control, predict, adapt, optimize, …

• Crisis management, monitor and predict natural phenomenon, monitor and manage engineering systems, optimize business processes

• A new paradigm ?– seamless access

• resources, services, data, information, expertise, …– seamless aggregation– seamless (opportunistic) interactions/couplings

Data-driven Management of Subsurface Geosystems: The Instrumented Oil Field (with UT-CSM, UT-IG, OSU, UMD, ANL)

Detect and track changes in data during production.Invert data for reservoir properties.Detect and track reservoir changes.

Assimilate data & reservoir properties intothe evolving reservoir model.

Use simulation and optimization to guide future production.

Data Driven

ModelDriven

Optimize

• Economic revenue

• Environmental hazard

• …

Based on the present subsurface knowledgeand numerical model

Improve numerical model

Plan optimal data acquisition

Acquire remote sensing data

Improve knowledge of subsurface to

reduce uncertainty

Update knowledge of modelMan

agem

ent d

ecis

ion

START

Dynamic Decision Dynamic Decision SystemSystem

Dynamic DataDynamic Data--Driven Driven Assimilation Assimilation

Data assimilation

Subsurface characterization

Experimental design

Dynamic, Data Driven Reservoir Management

LandfillsLandfills OilfieldsOilfields

ModelsModels SimulationSimulation

DataDataControlControlUndergroundPollution

UndergroundPollution

UnderseaReservoirsUnderseaReservoirs

Vision: Diverse Geosystems – Similar Solutions

Management of the Ruby Gulch Waste Repository (with UT-CSM, INL, OU)

– Flowmeter at bottom of dump– Weather-station– Manually sampled chemical/air

ports in wells– Approx 40K measurements/day

• Ruby Gulch Waste Repository/Gilt Edge Mine, South Dakota – ~ 20 million cubic yard of

waste rock– AMD (acid mine drainage)

impacting drinking water supplies

• Monitoring System– Multi electrode resistivity system

(523)• One data point every 2.4

seconds from any 4 electrodes – Temperature & Moisture sensors

in four wells

“Towards Dynamic Data-Driven Management of the Ruby Gulch Waste Repository,” M. Parashar, et al, DDDAS Workshop, ICCS 2006, Reading, UK, LNCS, Springer Verlag, Vol. 3993, pp. 384 – 392, May 2006.

Pervasive Grid Applications – Unprecedented Challenges: Uncertainty

• System Uncertainty– Very large scales– Ad hoc structures/behaviors

• p2p, hierarchical, etc, architectures– Dynamic

• entities join, leave, move, change behavior

– Heterogeneous• capability, connectivity, reliability,

guarantees, QoS– Lack of guarantees

• components, communication– Lack of common/complete

knowledge• number, type, location, availability,

connectivity, protocols, semantics, etc.

• Information Uncertainty– Availability, resolution, quality of

information– Devices capability, operation,

calibration– Trust in data, data models – Semantics

• Application Uncertainty– Dynamic behaviors

• space-time adaptivity– Dynamic and complex couplings

• multi-physics, multi-model, multi-resolution, ….

– Dynamic and complex (ad hoc, opportunistic) interactions

– Software/systems engineering issues

• Emergent rather than by design

Pervasive Grid Computing – Research Issues, Opportunities• Programming systems/models for data integration and runtime self-

management– components and compositions capable of adapting behavior, interactions

and information– correctness, consistency, performance, quality-of-service constraints

• Content-based asynchronous and decentralized discovery and access services– semantics, metadata definition, indexing, querying, notification

• Data management mechanisms for data acquisition and transport with real time, space and data quality constraints– high data volumes/rates, heterogeneous data qualities, sources – in-network aggregation, integration, assimilation, caching

• Runtime execution services that guarantee correct, reliable execution with predictable and controllable response time– data assimilation, injection, adaptation

• Security, trust, access control, data provenance, audit trails, accounting

Project AutoMate: Enabling Autonomic Applications

• Conceptual models and implementation architectures – programming systems based on popular programming models

• object, component and service based prototypes– content-based coordination and messaging middleware– amorphous and emergent overlays

• http://automate.rutgers.edu

Rudder

Coordination

Middlew

are

Sesam

e/DA

IS P

rotection Service

Autonomic Grid Applications

Programming SystemAutonomic Components, Dynamic Composition,

Opportunistic Interactions, Collaborative Monitoring/Control

Decentralized Coordination EngineAgent Framework,

Decentralized Reactive Tuple Space

Semantic Middleware ServicesContent-based Discovery, Associative Messaging

Content OverlayContent-based Routing Engine,

Self-Organizing Overlay

Ont

olog

y, T

axon

omy

Met

eor/S

quid

Con

tent

-bas

edM

iddl

ewar

e

Acc

ord

Prog

ram

min

gFr

amew

ork

Project Meteor: Middleware Stack

• A self-organizing tiered overlay network

• Content-based routing engine (Squid)– Flexible content-based routing and

querying with guarantees and bounded costs

– Decentralized information discovery

• Associative Rendezvous Messaging

– Symmetric post programming primitive– Content-based decoupled interaction

with programmable reactive behavior– Decentralized content-based in-

network aggregation (assimilation)

• Service-based API (WS)

AR messagingAR messaging

Content-based Routing

Self-organizing OverlaySelf-organizing Overlay

Pervasive Grid Environment

Pervasive Application

Agg

rega

tion

Agg

rega

tion

Project Meteor: Associative Rendezvous

• Content-based decoupled interaction with programmable reactive behaviors– Messages - (header, action, data)

• Symmetric post primitive: does not differentiating between interest/data

– Associative selection• match between interest and data

– Reactive behavior• Execute action field upon matching

• Decentralized in-network aggregation – Tries for back-propagating and

aggregating matching data items• Supports WS Notification standard

profile credentials

message context

TTL (time to live)

Action

[Data]

Header

• store• retrieve• notify• delete

Profile = list of (attribute, value) pairs:Examples:- Post <(sensor_type, temp.), (latitude, 10), (longitude, 20)>- Post <L, speed, 3600>, retrieve(AVERAGE, 300)

C1

C2

post (<p1, p2>, store, data)

notify_data(C2)

notify(C2)

(1)

(3)

(2) post(<p1, *>, notify_data(C2) )

<p1, p2> <p1, *>

match

Implementation/Deployment Overview

• Current implementation builds on JXTA– SquidTON, Squid, Comet and

Meteor layers are implemented as event-driven JXTA services

• Deployments include– Campus Grid @ Rutgers– ORBIT wireless testbed (400

nodes)• 400 802.11 radio nodes with

802.11 wireless connections

– PlanetLab wide-area testbed• At least one node selected

from each continent

ORBIT Testbed for Evaluation of Next Generation Wireless Networks: Phase 1: Indoor Grid

80 ft ( 20 nodes )

70 ft

m (

20 n

odes

)

Control switch

Data switch Application Servers

(User applications/ Delay nodes/

Mobility Controllers / Mobile Nodes)

Internet VPN Gateway / Firewall

Back-end servers

Front-endServers

Gigabit backboneVPN Gateway to

Wide-Area Testbed

SA1 SA2 SAP IS1 IS2 ISQ

RF/Spectrum Measurements Interference Sources

80 ft ( 20 nodes )

70 ft

m (

20 n

odes

)

Control switch

Data switch Application Servers

(User applications/ Delay nodes/

Mobility Controllers / Mobile Nodes)

Internet VPN Gateway / Firewall

Back-end servers

Front-endServers

Gigabit backboneVPN Gateway to

Wide-Area Testbed

SA1 SA2 SAP IS1 IS2 ISQ

RF/Spectrum Measurements Interference Sources

ORBIT Phase 2: Field Trial

• Requires ruggedized outdoor ready equipment (suggesting of the shelf technologies)

• Standard nodes used in dual role of mobile AP/mobile nodes deployed on busses

• Where possible connected to wired infrastructure; otherwise use of second radio interface for mesh type networking/wide area access

ORBIT: Hardware

512 MBRAM

Gigabit Ethernet(control)

GigabitEthernet

(data)

Intel/AtherosminiPCI802.11a/b/g

22.1Mhz

1 Ghz

pwr/resetvolt/temp

20 GBDISK

Serial Console110

VAC

RJ11 NodeIdBox+5v standby

PowerSupply

CPUVIA

C3 1Ghz

Intel/AtherosminiPCI802.11a/b/g

BluetoothUSB

CPURabbit Semi

RCM3700

10 BaseTEthernet

(CM)

Experimental Evaluation

• Data provided by Adolfo/Hector

• Aggregation services:– Complex queries issued from each of the three deployment environments

• Post(34-444,temp*, retrieve(count))

• Effectiveness of in-network aggregation– Aggregate range queries– Number of messages processed in the network

• Robustness in case of single peer failures

• Note that the performance of Squid and AR have been separately evaluated

Scalability of the Aggregation Services

Scalability of Meteor aggregation queries

0

500

1000

1500

2000

2500

3000

4 8 16 32 64

Number of peers

time

(ms) Wired LAN

Orbit Wireless Testbed

Wide-area PlanetLab

Aggregate query: post(<(100-300, 75-125), temperature>, retrieve(avg, 300))

• Experiences/issues with ORBIT deployment– Limitations of the MAC layer/protocol (802.11x)

• Channel interference, dropped messages, dropped beacons – Overlay structure is critical

• Hierarchical, multiplexed channels– Message sizes and volumes impact latency

• Average latency on 200 nodes (dedicated): ~2-3 seconds

Effectiveness of In-Network Aggregation

• Simulation results– 500 nodes

• AR aggregate message:– Post(<100-230, 35-120>,

retrieve(avg))

• Fig (a) number of messages with/without in-network aggregation

• Fig (b) actual number of messages per peer

Robustness of Aggregation Service

0

10

20

30

40

50

60

70

20 60 100

140

180

220

260

300

340

380

Time (milliseconds)

Num

ber o

f inc

lude

d re

sults

Time to aggregate all

Single node failure

Aggregate trie reconstruction

Reconstructed trie

• Assume single node failure– An intermediate node

fails at about 205 time ticks

– On-going aggregation operations are lost at this node

• About 100 ms required to correct the aggregation trie in a LAN environment

Conclusion

• Pervasive Computational Ecosystems & Next Generation Scientific Investigation– Knowledge-based, data and information driven, context-aware,

computationally intensive– Project AutoMate: Enabling Autonomic computational science

• Meteor middleware for sensor-based pervasive Grid environments– Unified programming abstractions for content-based decoupled interactions

and decentralized in-network aggregation• Recurrent and spatially constrained queries• Scalable and efficient trie-based design and implementation

– Deployments on LAN, Orbit and PlanetLab• Scalability, performance, robustness

• More Information, publications, software– www.caip.rutgers.edu/~parashar/– [email protected]