AAU Student Projects - visit Aalborg University's student projects portal
A master's thesis from Aalborg University
Book cover


Guide Lego Robot Towards Intelligence

Authors

; ;

Term

4. term

Education

Publication year

2009

Abstract

Denne rapport beskriver implementeringen af et multiagent-system, der omsætter teorien om POMDP og I-POMDP til praksis. POMDP (delvist observeret Markov-beslutningsproblem) er en metode til at træffe beslutninger under usikkerhed og begrænset information, og I-POMDP udvider dette til situationer med flere interagerende agenter. I systemet repræsenteres agenterne med Lego Mindstorms, Wii Remotes bruges til positionering, grafiske modeller bygges i Hugin Expert, og systemet programmeres i konventionel C++. Fokus er på at implementere et udvalgt scenarie ud fra teorien og håndtere de praktiske problemer, der opstår undervejs. Blandt udfordringerne er: at designe en datastruktur, der effektivt kan rumme den eksponentielt voksende mængde data i et policy-træ (et diagram over mulige beslutninger og observationer); at konstruere en transitionsfunktion, som tager højde for usikkerhed i handlingers udfald; samt at vælge, repræsentere og implementere et miljø, som normalt ville være for stort til at understøtte længere tidshorisonter (hvor langt frem i tid planen rækker).

This report presents the implementation of a multi-agent system that puts POMDP and I-POMDP theory into practice. POMDP (Partially Observable Markov Decision Process) models decision-making under uncertainty and limited information, and I-POMDP extends this to interactions among multiple agents. In the system, agents are represented with Lego Mindstorms, Wii Remotes are used for positioning, graphical models are created in Hugin Expert, and the system is programmed in conventional C++. The focus is on implementing a chosen scenario using the theory and addressing the practical issues that arise. Key challenges include designing a data structure that efficiently handles the exponentially growing data in a policy tree (a diagram of possible decisions and observations), building a transition function that accounts for uncertainty in action outcomes, and choosing, representing, and implementing an environment that would normally be too large to support longer time horizons (how far ahead the plan extends).

[This abstract was generated with the help of AI]