Most of the state-of-the-art commercial simulation software mainly focuses on providing realistic animations and convincing artificial intelligence to avatars in the scenario. However, works on how to trigger the events and avatar reactions in the scenario in a natural and intuitive way are less noticed and developed. Typical events are usually triggered by predefined timestamps. Once the events are set, there is no easy way to interactively generate new events while the scene is running and therefore difficult to dynamically affect the avatar reactions. Based on this situation, we propose a framework to use human gesture as input to trigger events within a DI-Guy simulation scenario in real-time, which could greatly help users to control events and avatar reactions in the scenario. By implementing such a framework, we will be able to identify user’s intentions interactively and ensure that the avatars make corresponding reactions.
History
Event
Asia Pacific Simulation Technology and Training. Conference (17th : 2012 : Adelaide, South Asutralia)
Pagination
1 - 6
Publisher
[SimTecT]
Location
Adelaide, South Australia
Place of publication
[Adelaide, S.Aust.]
Start date
2012-06-18
End date
2012-06-21
ISBN-13
9780980809930
Publication classification
E1 Full written paper - refereed
Copyright notice
2012, SIMTECT
Title of proceedings
SimTecT 2012 : Simulation-integrated solutions : Proceedings of the Annual Asia Pacific Simulation Technology and Training Conference