The AIMM robot is an industrial mobile manipulator able to autonomously perceive and manipulate its environment. It was used for various scenarios (especially pre-assembly and logistics) in multiple projects and demos: FoF, EuRoC, Automatica 2018, TAPAS and Automatica 2016.
I am a core developer of AIMM’s software stack. My main responsibilities are creating software for autonomous task control and belief state modeling. We put a high effort in the system architecture of AIMM, as it is a complex, mobile robot with 13 DOFs (including an impedance controlled LWR 4+), four cameras systems (able to produce depth information using SGM) and laser scanners. It features six computers: One for low-level realtime control, one for navigation, one for object detection and scene reconstruction, one for path planning, one for data logging and processing and one for high level autonomy. On top of the multitude of software models we employ four different middlewares: ‘links and nodes’ for realtime control, ‘sensornet’ for high bandwidth sensor data, ‘ROS’ for third party library integration and data visualization, and Kuka’s Sunrise middleware. We use continuous integration to develop and release new software. Furthermore, we employ a powerful release and dependency management toolchain to track and deploy consistent software versions to AIMM.
My responsibilities: Autonomous Task Control, System Architecture, Task Execution Logging and Profiling, Semantic World Modeling