Ardhendu B.
Profile Summary
Reader (Associate Professor) and programme leader for computer science in the Department of Computer Science. I have served as an EHU Impact Fellow for REF panel B.
My area of expertise, Computer Vision, Deep Learning, Artificial Intelligence with applications to Healthcare
technology, Autonomous Systems and Smart Environments (e.g. Factory, Home, Vehicles, etc.)
My research focuses on computer vision, image/video analysis and pattern recognition.
Recent research is on recognising human activities from an egocentric view point (First-person-view) for assistive technologies and robotics applications. Developing multimodal-AI for health and social care applications.
Expertise
Reader & Researcher in Computer Science
> 5 Year Experience5/5
Electrical Engineering
> 5 Year Experience5/5
Employment History
Reader (Associate Professor) in Computer Vision & AI
Research and teaching Computer Vision and Artificial Intelligence
Research Fellow + Visiting Research Fellow
One of the most interesting challenges in modern science and engineering is the building of
intelligent machines which can interact with humans and perform useful tasks autonomously. In
light of this requirement, my research programme focuses on developing computational analogue
for basic human cognition and exploiting the strengths of computers to take full advantage of these
capabilities. Specifically, I am interested in intelligent systems for activity/behaviour analysis and
recognition that brings together artificial intelligence, computer vision, machine learning, insights
from psychology and physiology, algorithms and a great deal of computation.
PhD + RA + TA
The main idea behind this research was to investigate methods for building an efficient multimedia
system for document-based automatic indexing and retrieval of multimedia data captured from
multimodal environments such as meetings, conferences, etc. Both empirical image processing,
video segmentation methods and document analysis approaches are studied to bridge the
gap between temporal data and static information. In order to provide useful access points, the
captured audio-visual data were fragmented into reasonable distinct smaller segments using
projected documents, which provide meaningful semantic pointers because they appear at
a specific time, remain in visual focus for a definite duration and summarize the presenter’s
discourse at that time. A Visual Signature-based identification of low-resolution document images
is developed to link original electronic documents with the temporally segmented captured
multimedia data and image-based retrieval of multimedia information. Matching of signature
is based on both sequential as well as multi-level linear and non-linear fusion of various visual
features. The above-mentioned techniques prove the usefulness of documents as an additional
modality and natural interface, in interacting with multimedia data captured from multimodal
environments.
Education
Doctor of Philosophy (PhD)
Languages
English
Full Professional Proficiency