PGRs meeting and Research Presentations – April 2016

The monthly PGRs Research Presentations was held on Wed. 13th April, 2pm, Room MC3108.

This session we had the following presentations:

 

PGRs Monthly meeting_April2016  (Slides )

IMG_20160413_140241

Agenda

  • Speaker –>
  • A quick look at the new “PGRs Management System”,PGR-MS1

PGR-MS2

  • PGRs Blog.

PGR-Blog

  • Discussion of the activities plan.
  • Update and plan for the “Showcase Event”
  • Announcements, AOB, & closing

 

IMG_20160413_143607

Title: Life-long Spatio-temporal Exploration of Dynamic Environments: An overview.

By: Joao Santos

IMG_20160413_143555

 Abstract: The primary purpose of robotic exploration is to autonomously acquire a complete and precise model of the robot’s operational environment. To explore efficiently, the robot has to direct its attention to environment areas that are currently unknown. If the world was static, these areas would simply correspond to previously unvisited locations. In the case of dynamic environments, visiting all locations only once is not enough, because they may change over time. Thus, dynamic exploration requires that the environment locations are revisited and their (re-) observations are used to update a dynamic environment model. However, revisiting the individual locations with the same frequency and on a regular basis is not efficient because the environment dynamics will, in general, not be homegeneous, (i.e. certain areas change more often and the changes occur only at certain times).

Similarly to the static environment exploration, the robot should revisit only the areas whose states are unknown at the time of the planned visits. Thus, the robot has to use its environment model to predict the uncertainty of the individual locations over time and use these predictions to plan observations that from a theoretical point of view improve its knowledge about the world’s dynamics. Hence, the observations are scheduled in order to obtain information about the environment changes, which are mainly caused by human-activity. As a consequence, using schedules motivated by the changes in metric maps increases the chance to extract  dynamics that are essential for object learning and activity recording tasks.

 

 

 

 

 

Farewell and Thank You

At the end of the March’s meeting, we had a Farewell and Thank You for previous “PGRs Students Reps” and PGRs who have been helping in various activities, including Reading Group, trips, coffee mornings, and the showcase.

IMG_1606IMG_1608

 

  • PGRs Representatives:

–1st Students Rep:  Touseef Quraishi

2nd Students Reps:

  • Christian
  • Mohammadreza.

 

IMG_1614IMG_1611

PGRs who have been helping:

(George; showcase events & reading group – Saddam; Showcase Events & support PGRs meeting – Francesco; trips & showcase – Ibrahim; trips & showcase – Christian; regular presentations & reading group – Talal; Reading group – Hussein; showcase)

IMG_1627IMG_1623

 

Thanks to all, and best wishes for the years to come.

 

PGRs meeting and Research Presentations – March. 2016

The monthly PGRs Research Presentations was held on Wed. 9th March, 2pm, Room MC3108.

This session we had the following presentations:

Title: Facilitating Individualised Collaboration with Robots (FInCoR).

By: Peter Lightbody

AGEDNA:

  1. Speaker –>
  2. Break & refreshments
  3. Farewell & Thank you to previous PGRs Student Reps and those who have been actively helping & supporting the PGRs activities & community.
  4. Brief about the responsibilities and benefits of being a PGR Student Rep.      
  5. Reps Election. (By SU Rep.)
  6. Announcements, AOB, & closing

 

Abstract: Enabling a robot to seamlessly collaborate with a human counterpart requires a robot to not only identify human preferences, but also to adapt in order to decrease the likelihood of distress and discomfort for the human collaborator. This work presents the use of qualitative spacial relations, combined with hidden Markov models, to identify and learn the unique characteristics inherent in the way a person performs a task. By doing this in real-time, a collaborative robot will be able to adapt it’s behaviour in order to best accommodate the person intuitive way of performing a task.

 

 

 

 

 

 

 

PGRs meeting and Research Presentations – Feb. 2016

The monthly PGRs Research Presentations was held on Wed. 10th Feb., 2pm, Room MC3108.

This session we had the following presentations:

Title: Event-based Continuous STDP Learning using HMAX Model for Visual Pattern Recognition.
 By: Daqi Liu

Abstract: Ventral stream within the visual cortex plays an important role in form recognition and object representation. Understanding and modeling the processing mechanism of the ventral stream is quite significant and necessary for visual pattern recognition application. In our research, an event-based continuous spike timing dependent plasticity (STDP) learning method (ECS) using HMAX model for visual pattern recognition has been proposed. Through the proposed spiking encoding scheme, the spatiotemporal spiking pattern would be generated from the high-level features extracted from HMAX model and such spatiotemporal information conveys unique and distinguish selectivity to each input visual stimuli. The selectivity to the input visual stimuli will be emerged after the continuous learning using the proposed event-based STDP method. By incorporating background neural noise and time jitter into the input visual stimuli of the proposed method while adding nothing into the classic SVM algorithm, cross-validated experimental results on MNSIT handwritten character database show that the proposed ECS method still achieves a better performance even in such harsh conditions.  

 

 

 

 

 

 

 

PGRs meeting and Research Presentations – Nov. 2015

The monthly PGRs Research Presentations was held on Wed. 11th November, 2pm, Room MC3108.

This session we had the following presentations:

Title: “Profiling User Engagement with Promotional Social Media Content“. Title:   “A Bio-inspired Collision Detection Vision System Embedded for Autonomous Micro-robots  

By: Jamie Mahoney

By: Cheng Hu

Abstract: Organisations, retailers and brands have a long-established need to gain insight into the characteristics of their customers, users and online followers. Using Twitter as a case study, we describe a method of creating engagement profiles of users based on qualitative analyses of the social media content with which they have publicly engaged. By clustering these engagement profiles, we extend previous work by not only showing how people within a social graph can be clustered in useful ways, but also how these users are likely to engage with specific types of social media content – thus allowing for the creation of targeted social media content strategies. Our findings demonstrate that ‘traditional’ methods of social graph segmentation do not reflect groupings of similar users in terms of engagement behaviour, we also demonstrate that user engagement behaviours do not vary dramatically over time. We provide suggestions for how these findings might be used in the creation of effective strategies for companies, and organisations, wishing to issue promotional material via social media platforms. Abstract:  The vision system takes inspiration from locusts’ response in detecting fast approaching objects. Neurophysiological research suggested that locusts use a wide-field visual neuron called lobula giant movement detector (LGMD) to respond to imminent collisions. In this work, we present the implementation of selected neuron model by a low-cost ARM processor as part of a composite vision module. The developed system is able to perform image acquisition and processing independently. The vision module is placed on top of a micro-robot to initiate obstacle avoidance behaviour. Both simulation and real-world experiments were carried out to test the reliability and robustness of the vision system. The results of the performed experiments with different scenarios demonstrated the amenability of the developed bio-inspired vision system to be used as a low-cost embedded module in autonomous robots with high precision.

 

 

  • Then our usual catch-up agenda: