dc.contributor.advisor | Chen-Burger, Jessica | en |
dc.contributor.advisor | Fisher, Robert B. | en |
dc.contributor.author | Nadarajan, Gayathri | en |
dc.date.accessioned | 2011-02-02T11:32:34Z | |
dc.date.available | 2011-02-02T11:32:34Z | |
dc.date.issued | 2011 | |
dc.identifier.uri | http://hdl.handle.net/1842/4757 | |
dc.description.abstract | Traditional workflow systems have several drawbacks, e.g. in their inabilities to rapidly
react to changes, to construct workflow automatically (or with user involvement) and
to improve performance autonomously (or with user involvement) in an incremental
manner according to specified goals. Overcoming these limitations would be highly
beneficial for complex domains where such adversities are exhibited. Video processing
is one such domain that increasingly requires attention as larger amounts of images and
videos are becoming available to persons who are not technically adept in modelling
the processes that are involved in constructing complex video processing workflows.
Conventional video and image processing systems, on the other hand, are developed
by programmers possessing image processing expertise. These systems are tailored
to produce highly specialised hand-crafted solutions for very specific tasks, making
them rigid and non-modular. The knowledge-based vision community have attempted
to produce more modular solutions by incorporating ontologies. However,
they have not been maximally utilised to encompass aspects such as application context
descriptions (e.g. lighting and clearness effects) and qualitative measures.
This thesis aims to tackle some of the research gaps yet to be addressed by the
workflow and knowledge-based image processing communities by proposing a novel
workflow composition and execution approach within an integrated framework. This
framework distinguishes three levels of abstraction via the design, workflow and processing
layers. The core technologies that drive the workflow composition mechanism
are ontologies and planning. Video processing problems provide a fitting domain for
investigating the effectiveness of this integratedmethod as tackling such problems have
not been fully explored by the workflow, planning and ontological communities despite
their combined beneficial traits to confront this known hard problem. In addition, the
pervasiveness of video data has proliferated the need for more automated assistance
for image processing-naive users, but no adequate support has been provided as of yet.
A video and image processing ontology that comprises three sub-ontologies was
constructed to capture the goals, video descriptions and capabilities (video and image
processing tools). The sub-ontologies are used for representation and inference. In
particular, they are used in conjunction with an enhanced Hierarchical Task Network
(HTN) domain independent planner to help with performance-based selection of solution
steps based on preconditions, effects and postconditions. The planner, in turn,
makes use of process models contained in a process library when deliberating on the
steps and then consults the capability ontology to retrieve a suitable tool at each step. Two key features of the planner are the ability to support workflow execution (interleaves
planning with execution) and can perform in automatic or semi-automatic
(interactive) mode. The first feature is highly desirable for video processing problems
because execution of image processing steps yield visual results that are intuitive
and verifiable by the human user, as automatic validation is non trivial. In the semiautomaticmode,
the planner is interactive and prompts the user tomake a tool selection
when there is more than one tool available to perform a task. The user makes the tool
selection based on the recommended descriptions provided by the workflow system.
Once planning is complete, the result of applying the tool of their choice is presented
to the user textually and visually for verification. This plays a pivotal role in providing
the user with control and the ability to make informed decisions. Hence, the planner
extends the capabilities of typical planners by guiding the user to construct more
optimal solutions. Video processing problems can also be solved in more modular,
reusable and adaptable ways as compared to conventional image processing systems.
The integrated approach was evaluated on a test set consisting of videos originating
from open sea environment of varying quality. Experiments to evaluate the efficiency,
adaptability to user’s changing needs and user learnability of this approach were conducted
on users who did not possess image processing expertise. The findings indicate
that using this integrated workflow composition and execution method: 1) provides a
speed up of over 90% in execution time for video classification tasks using full automatic
processing compared to manual methods without loss of accuracy; 2) is more
flexible and adaptable in response to changes in user requests (be it in the task, constraints
to the task or descriptions of the video) than modifying existing image processing
programs when the domain descriptions are altered; 3) assists the user in selecting
optimal solutions by providing recommended descriptions. | en |
dc.language.iso | en | |
dc.publisher | The University of Edinburgh | en |
dc.relation.hasversion | G. Nadarajan, Y.-H. Chen-Burger, R. B. Fisher. “A Knowledge-Based Planner for Processing Unconstrained Underwater Videos”. In IJCAI Workshop on Learning Structural Knowledge From Observations (STRUCK 09), July 2009. | en |
dc.relation.hasversion | G. Nadarajan, A.Manataki, Y.-H. Chen-Burger. “Semantics-Based Process Support for Grid Applications”. Book Chapter in Nik Bessis (ed.): Grid Technology for Maximizing Collaborative Decision Management and Support: Advancing Effective Virtual Organizations, IGI Global, May 2009. | en |
dc.relation.hasversion | C. Spampinato, Y.-H. Chen-Burger, G. Nadarajan, R. B. Fisher. “Detecting, Tracking and Counting Fish in Low Quality Unconstrained Underwater Videos”. In International Conference on Computer Vision Theory and Applications (VISAPP 08), Jan 2008. | en |
dc.relation.hasversion | G. Nadarajan. “Planning for Automatic Video Processing using Ontology-Based Workflow”. Doctoral Consortium at International Conference on Automated Planning & Scheduling (ICAPS 07), Sept 2007. | en |
dc.relation.hasversion | G. Nadarajan and A. Renouf. “A Modular Approach for Automating Video Analysis”. In International Conference on Computer Analysis of Images and Patterns (CAIP 07), Aug 2007. | en |
dc.relation.hasversion | G. Nadarajan, Y.-H. Chen-Burger, J.Malone. “Semantic Grid Services for Video Analysis”. In International Workshop on Service Composition, Dec 2006. | en |
dc.relation.hasversion | G. Nadarajan, Y.-H. Chen-Burger, J.Malone. “Semantic-BasedWorkflow Composition for Video Processing in the Grid”. In IEEE/WIC/ACM International Conference on Web Intelligence (WI 06), Dec 2006. | en |
dc.subject | ontologies | en |
dc.subject | planning | en |
dc.subject | workflows | en |
dc.subject | knowledge-based vision | en |
dc.subject | semantics-based workflow composition | en |
dc.subject | ontology-based planning | en |
dc.subject | multiple executables | en |
dc.subject | video processing | en |
dc.title | Semantics and planning based workflow composition and execution for video processing | en |
dc.type | Thesis or Dissertation | en |
dc.type.qualificationlevel | Doctoral | en |
dc.type.qualificationname | PhD Doctor of Philosophy | en |