multi-agent exploration in unknown environments changchang wu nov 2, 2006

58
Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Upload: grant-harris

Post on 29-Jan-2016

222 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Multi-Agent Exploration in Unknown Environments

Changchang WuNov 2, 2006

Page 2: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Outline

• Why multiple robots• Design issues• Basic approaches

– Distributed– Centralized– Market-based

Page 3: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Why Multiple Robots

• Some tasks require a robot team

• Have potential to finish tasks faster

• Increase robustness w/ redundancy

• Compensate sensor uncertainty by merging overlapping information

• Multiple robots allow for more varied and creative solutions

Page 4: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

A Good Multi-Robot System Is:

• Robust: no single point of failure

• Optimized, even under dynamic conditions

• Quick to respond to changes

• Able to deal with imperfect communication

• Able to avoid robot interference

• Able to allocate limited resources

• Heterogeneous and able to make use of different robot skills

Page 5: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Basic Approaches

• Distributed – Every robot goes for itself

• Centralized– Globally coordinate all robots

• Market-based– Analogy To Real Economy

Page 6: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Distributed Methods

• Planning responsibility spread over team

• Each robot basically act independently• Robots use locally observable

information to coordinate and make their plans

Page 7: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Example: Frontier-Based Exploration Using Multiple Robots (Yamauchi 1998)

• A highly distributed approach

• Simple idea: To gain the most new information about the world, move to the boundary between open space and uncertainty territory

• Frontiers are the boundaries between open space and unexplored space

Page 8: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Occupancy Grid

• World is represented as grid• Each cell in the grid is assigned with a

probability of being already occupied/observed• The initial probability is all set to .5• Cell status can be Open (<0.5), Unknown

(=0.5) or Occupied (>0.5)• Bayesian rule is used to update cells by

merging information from each sensor reading (sonar)

Page 9: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Frontier Detection• Frontier = Boundary between open and unexplored

space.

• Any open cell adjacent to unknown cell is frontier edge cell.

• Frontier cells grouped into frontier regions based on adjacency.

• Accessible frontier = Robot can pass through opening.

• Inaccessible frontier = Robot cannot pass through opening.

Page 10: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Multi-Robot Navigation

• Simple algorithm: Each robot goes along the shortest obstacle free path to a frontier region

• Robots share a common map: All information obtained by any robot is available to all robots

• Robots are planning path independently • Use reactive strategy to avoid collisions• Robots may waste time for the same frontiers

Page 11: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

An Exploration Sequence

Page 12: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Distributed Methods: Pros & Cons

• Pros– Very robust. No single point failure– Fast response to dynamic conditions– Little or no communication is required– Easy….Little computation required

• Cons– Plans only based on local information– Solutions are often sub-optimal

Page 13: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Centralized Methods

• Robot team treated as a single “system” with many degrees of freedom

• A single robot or computer is the “leader”

• Leader plans optimal tasks for groups

• Group members send information to leader and carry out actions

Page 14: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Example: Arena (Jia 2004)

• Robots share a common map and only communicate with a leader

• Robots compete for resources by their efficiency

• leader greedily assigns the most efficient tasks

• Leader coordinate robots to handle interference

Page 15: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Background

• World representation– Occupancy grid

• Cost unit– Moving forward one step = Turning 45

degrees

• Cost overflow– Similar to minimum cost spanning tree– Easy to compute the shortest path– Easy to handle obstacle

Page 16: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Cost Overflow

Cost of 45° turning = Cost of one cell’s step

1 2

2 3

3

3

3

4

4

4

44

4

5

5 5 5 5

5

5

5

Direction

priority

6

6

6 6 6 6 6

7

7

7

77

8

8

8

8

88

8

9 9

9

9

9

9

9

10

10

10

10

10

10

11

11

11

11

11

11

11

12

12

12

12

12

12

12

12

Page 17: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Goal Candidates Detection

• A goal point P should satisfyi. P is passable (Mark the cells in warning range or

obstacles/Wall/Unknown cells as impassable)ii. Some unexplored cells lie in the circle with P as the

center and (R + K) as the radium, where R is the warning radius and K is usually 1

Robot cellpaths cellobservation cell

candidate goal

Page 18: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Goal Resource

• Reserved goal candidates– Robots obtained by competition

• Recessive goal candidates– The goal points in a given range

to a reserved goal point– This distance can be adjusted

Goal candidates

Recessive goals candidates

Page 19: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Path Resource

• Path resource is a time-space term

• For a given time, the cells close to any robot are marked off for safety

• Looks just like a widened path

• Basically a reactive strategy

goal

path

resource

Page 20: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Revenue and Utility

• Revenue – The expected gain of information that

robots observe at a goal point

• Utility used by many other approaches– Utility = revenue – cost

• Utility in this paper– Utility = Revenue / Cost– Better connected to purpose of smallest

cost– No need to care about unit conversion

Page 21: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Greedy Goal Selection

• Try to maximize the global utility

• Coordination: robots obtain goal and path resources exclusively

• Competition: repetitively select the pair of free agent and goal with highest utility

• Sub-optimal

Page 22: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Simple Algorithm

• Repeat until map is complete– Repeat #free robots times

1. Cost computation (Also make sure no interference with the busy robots)

2. Select the highest utility task (Compete)3. Mark off the associated robot and goal

points, and nearby goal points

Page 23: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

1st Competition:

2

2 2

2

2

2

2

2

2

2 2 2 2

2

2

2

2

2

2

3 3 3

3

3

3

3

3

3

3 3 3 3 3 3

3

3

3

3 3 34 4 4

4

4

4

4 4 4 4 4 4 4 4

1

1

1 1

1

1

1

1

1

11

55

5 55 5 5 5 55

6 6 6 6 6

6

6

6

6

6

Interval = 3 Competitor:

Page 24: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

1st Competition Result:

4 4

6 6

5 5

3

2

2

1

4

4

1

2 2

4 4

3 3

2 2

4

Interval = 3 Competitor:

Page 25: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

2nd Competition

1

2

1

2 2

3 3 3

4 4

4

4 4 4

44

1

11

1

1 1

Interval = 3 Competitor:

Satisfied:

2 2 2

2

2

2

2

2 2 2

3 3 3 3

3

3

3

3

3

3

3 3 3 3444 4 4 4

4

4

4

4

4

4

4 4 4 4 4

5 5 5 5 5

5 5 5

5

5

5 5

6

6

6

6 6 66 6

6

6

6

6

Page 26: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

2nd Competition Result

1

2

1

2 2

3 3 3

4 4

4

4 4 4

44

4

1

2

3

4

5

1

1

2

2

22

2

333

4 44

5 5 5

6 6 6

6

666

6

Interval = 3 Competitor:

Satisfied:

Page 27: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

3rd Competition

1

2

1

2 2

3 3 3

4 4

4

4 4 4

44

4

1

2

3

4

5

1

2

2

22

2

333

4 44

5 5 5

6 6 6

6

666

6 6

Competitor:

Satisfied:

Interval = 3

1

1 1

2

2

222

3 3 3 3

3

3

3

4 4 4 4 4

4

4

5 5 5 5 5

6 6

6

7 7

7

7

7

7

7

7

7

8

8

8

8

8

8

8

9

9

9

9

9

9

9

10

10

10

10

10

10

10

10

11 11

11

11

11

11

11

11

11

11 12

12

12

12

12

12

12

12

13

13

13

13

13

13

13

Page 28: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

3rd Competition Result

1

3 2

4

4

1

2

3

4

5

1

2

3

4

6 6 6

Competitor:

Satisfied:

Interval = 3

1

2

3

4 7

8

9 10 11 12

5 6

13

4

6

Page 29: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Planning Issues

• Do not transfer a reserved goal point to another free agent (unless necessary). Frequent change of tasks can cause localization error.

• Quit an assigned task when the goal point is unexpectedly observed by other robots

• Schedule at most one task for each agent

Page 30: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Possible Variations• Still keep busy agents in competition.

Remove the goal resources they win from competition.– This prevents those goal resources being

assigned to other agents– It is too early to burden a new task on a

robot who has not achieved it current task• No need to schedule them.

– New resources probably will be found when they reach the goals

Page 31: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Handling Failure of Planning

• It may fail to plan safe paths– When some robot get to a place where

• it is almost too close to other robot • it has no good space to detour

– And it choose to just wait there for other robots to move away, which is not known by other robots

• Avoidance of unexpected obstacle– Robots have simple reactive mechanism– Release resources and try to gain new task

Page 32: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Fail to plan safe paths

Competitor:

Satisfied:

Interval = 3

1

2

3 4 5 6 7 8

9

10

1

2

3410111213141516

collision

Page 33: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Reactive Mechanism

Competitor:

Satisfied:

Interval = 3

Page 34: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Exchange Tasks

Competitor:

Satisfied:

Interval = 3

1234

5

1

2

3

Page 35: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Some Statistics

Page 36: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Demo

Page 37: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Centralized Methods : Pros

• Leader can take all relevant information into account for planning

• Optimal s islution possible! • One can try different approximate

solutions to this problem

Page 38: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Centralized Methods: Cons

• Optimal solution is computationally hard– Intractable for more than a few robots

• Makes unrealistic assumptions:– All relevant info can be transmitted to leader

– This info doesn’t change during plan construction

• Vulnerable to malfunction of leader

• Heavy communication load for the leader

Page 39: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Market-Based Methods

• Based on market architecture• Each robot seeks to maximize individual

“profit”• Robots can negotiate and bid for tasks• Individual profit helps the common good• Decisions are made locally but effects

approach optimality– Preserves advantages of distributed

approach

Page 40: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Why Is This Good?

• Robust to changing conditions– Not hierarchical– If a robot breaks, tasks can be re-bid to others

• Distributed nature allows for quick response• Only local communication necessary• Efficient resource utilization and role adoption• Advantages of distributed system with

optimality approaching centralized system

Page 41: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Architecture• World is represented as a grid

– Squares are unknown (0), occupied (+), or empty (-)• Goals are squares in the grid for a robot to explore

– Goal points to visit are the main commodity exchanged in market

• For any goal square in the grid:– Cost based on distance traveled to reach goal– Revenue based on information gained by reaching goal

• R = (# of unknown cells near goal) x (weighting factor)

• Team profit = sum of individual profits– When individual robots maximize profit, the whole team

gains

Page 42: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Example World

Page 43: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Goal Selection Strategies

• Possible strategies:– Randomly select points, discard if

already visited– Greedy exploration:

•Choose goal point in closest unexplored region

– Space division by quadtree

Page 44: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Exploration Algorithm

Algorithm for each robot:1. Generate goals (based on goal selection

strategy)2. If OpExec (human operator) is reachable,

check with OpExec to make sure goals are new to colony

3. Rank goals greedily based on expected profit

4. Try to auction off /bid goals to each reachable robot

– If a bid is worth more than you would profit from reaching the goal yourself (plus a markup), sell it

Page 45: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Exploration Algorithm

5. Once all auctions are closed, explore highest-profit goal

6. Upon reaching goal, generate new goal points

– Maximum # of goal points is limited

7. Repeat this algorithm until map is complete

Page 46: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Bidding Example

• R1 auctions goal to R2

Page 47: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Expected vs. Real

• Robots make decisions based on expected profit– Expected cost and revenue based on

current map

• Actual profit may be different– Unforeseen obstacles may increase cost

• Once real costs exceed expected costs by some margin, abandon goal– Don’t get stuck trying for unreachable goals

Page 48: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Information Sharing

• If an auctioneer tries to auction a goal point already covered by a bidder:– Bidder tells auctioneer to update map– Removes goal point

• Robots can sell map information to each other– Price negotiated based on information gained– Reduces overlapping exploration

• When needed, OpExec sends a map request to all reachable robots– Robots respond by sending current maps– OpExec combines the maps by adding up cell

values

Page 49: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Advantages of Communication

• Low-bandwidth mechanisms for communicating aggregate information

• Unlike other systems, map info doesn’t need to be communicated repeatedly for coordination

Page 50: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

What Is a Robot Doing

• Goal generation and exploration• Sharing Information with other

robots• Report information to OpExec at

some frequency

Page 51: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Experimental Setup

• 4 or 5 robots– Equipped with

fiber optic gyroscopes

– 16 ultrasonic sensors

Page 52: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Experimental Setup

• Three test environments– Large room cluttered with obstacles– Outdoor patio, with open areas as well as walls and tables– Large conference room with tables and 100 people

wandering around• Took between 5 and 10 minutes to map areas

Page 53: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Experimental Results

Page 54: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Experimental Results

Page 55: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Experimental Results

• Successfully mapped regions• Performance metric (exploration efficiency):

– Area covered / distance traveled [m2 / m]– Market architecture improved efficiency over no

communication by a factor of 3.4

Page 56: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

Conclusion

• Market-based approach for multi-robot coordination is promising– Robustness and quickness of distributed system– Approaches optimality of centralized system– Low communication requirements

• Probably not perfect– Cost heuristics can be inaccurate– Much of this approach is still speculative

• Some pieces, such as leaders, may be too hard to do

Page 57: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

In Sum

• Distributed vs. centralized mapping• Distributed vs. centralized planning• Revenue/Cost vs. Revenue – Cost

• Often sub-optimal solutions• No common evaluation system for

comparisons

Page 58: Multi-Agent Exploration in Unknown Environments Changchang Wu Nov 2, 2006

References

• Yamauchi, B., "Frontier-Based Exploration Using Multiple Robots," In Proc. of the Second International Conference on Autonomous Agents (Agents98), Minneapolis, MN., 1998.

• Menglei Jia , Guangming Zhou ,Zonghai Chen, "Arena—an Architecture for Multi-Robot Exploration Combining Task Allocation and Path Planning,“ 2004

• Zlot, R., Stentz, A., Dias, M. B., and Thayer, S. “Multi-Robot Exploration Controlled By A Market Economy.” Proceedings of the IEEE International Conference on Robotics and Automation, 2002.

• http://voronoi.sbp.ri.cmu.edu/presentations/motionplanning2001Fall/FrontierExploration.ppt

• http://www.ai.mit.edu/courses/16.412J/lectures/advanced%20lecture_11.6.ppt

• http://mail.ustc.edu.cn/~jml/jml.files/Arena.ppt