|
Area 1 - DATABASES AND INFORMATION SYSTEMS INTEGRATION
Area 2 - ARTIFICIAL INTELLIGENCE AND DECISION SUPPORT SYSTEMS
Area 3 - INFORMATION SYSTEMS ANALYSIS AND SPECIFICATION
Area 4 - Software Agents and Internet Computing
Area 1 - DATABASES AND INFORMATION SYSTEMS INTEGRATION
Title: |
O2PDGS:
AN APPROACH FOR UNDERSTANDING OBJECT ORIENTED PROGRAMS |
Author(s): |
Hamed
Al-Fawareh |
Abstract: |
In this paper, we provide a description of dependence
graphs for representing meaningful dependencies between components of
object-oriented programs. A formal description of the dependence relations
of interest is given before giving a representative illustration of
object-oriented program dependence graphs (O2PDGs). The paper also
discusses an approach for understanding object-oriented programs through
the use of O2PDGs. |
|
Title: |
ERP
SYSTEMS IMPLEMENTATION DETERMINANTS AND SUCCESS MEASURES IN CHINA: A CASE
STUDY APPROACH |
Author(s): |
Christy Cheung,
Zhe Zhang, Matthew Lee, Liang Zhang |
Abstract: |
With the growing intensive global competition and
integration of the world economy, manufacturing firms have to reduce
inventory level and operation costs, improve customer service to obtain
competitive advantage against their competitors. Manufacturing companies
are forced to adopt new methods to achieve the above objectives.
Enterprise resource planning (ERP) system is one of the most widely
accepted choices. AMR predicts the total ERP market will reach $66.6
billion by 2003, growing an estimated 32% annually over the next five
years. Significant benefits such as improved customer service, better
production scheduling, and reduced manufacturing costs can accrue from the
successful implementation of ERP (Ang et al, 1995). However, the
successful implementation rate is extremely low especially in China and
many firms didn’t achieve intended goals. Thus, it’s necessary for ERP
practitioners and researchers to investigate the reasons why the
implementation success rate of ERP systems in China is so low. Prior
studies mainly focus on critical success factors or single ERP
implementation success measure without theoretical support. This study
attempts to combine Ives, Hamilton, and Davis (1980) MIS research model
and DeLone & McLean’s (1992) IS success model to develop an ERP
implementation success model, identifying both generic and unique factors
that affect ERP systems implementation success in China and using multiple
ERP implementation success measures to assess whether an ERP
implementation is a success or failure. Through multiple case study
research method, more detailed information about ERP implementations could
be collected. Moreover, it solves problems of validity and reliability of
constructs occurring frequently in a single case study. The results of
this research can help ERP-related researchers, practitioners, and
companies to get more comprehension of ERP systems implementation issues
and the chance of ERP implementation success could be increased given
enough attention to these issues. |
|
Title: |
DATA
WAREHOUSING: A REPOSITORY MODEL FOR METADATA STORAGE AND RETRIEVAL BASED
ON THE HUMAN INFORMATION PROCESSING |
Author(s): |
Enrique Luna-Ramírez,
Félix García-Merayo, Covadonga Fernández-Baizán |
Abstract: |
The information on the creation, management and use of a
data warehouse is stored in what is called the metadata repository, making
this repository the single most important component of the data warehouse.
Accordingly, the metadata repository plays a fundamental role in the
construction and maintenance of the data warehouse, as well as for
accessing the data it stores. In this paper, we propose a repository model
conceived to store and retrieve the metadata of a corporation data
warehouse. With a view to achieving this objective, the model, composed of
an approach for modelling the repository structure and by a metamodel for
retrieving metadata, is based on the human information processing
paradigm. So, the model considers a series of distinctive functionalities
that can be built into a repository system to assure that it works
efficiently. These functionalities refer to the use of two memories for
storing the repository metadata and a set of structures and processes for
retrieving the information passing from one memory to another. One of the
memories in particular is used to store the most recurrent metadata in a
corporate environment, which can be rapidly retrieved with the help of the
above-mentioned structures and processes. These structures and processes
also serve to contextualise the information of a data warehouse according
to the projects or business areas to which it refers. |
|
Title: |
HOSPITAL
CASE RECORDS INFORMATION SYSTEM: CASE STUDY OF A KNOWLEDGE-BASED
PRODUCT |
Author(s): |
A. Neelameghan,
M. Vasudevan |
Abstract: |
Briefly discusses knowledge management and use of
knowledge-based products in enterprises. Enumerates the information
resources of a hospital and describes the design and development of a
patients’ case records system, specifically for a hospital specializing in
surgical cases of tumors of the central nervous system. Each case record
has data / information on over 150 attributes of patient, facility for
hypertext linking relevant images (CT scan, X-ray, NMR, etc.) and access
to electronic documents from other websites. The collaborative roles of
the hospital doctors and a consultant information specialist in the
development of the system are indicated. Output of a case record with
links to related CT scan pictures and a web document is presented as
example. Concludes mentioning the various uses of the system. |
|
Title: |
MODELS
FOR IMPLEMENTATION OF ONLINE REAL TIME IT-ENABLED SERVICE FOR ENTRY TO
PROFESSIONAL EDUCATION |
Author(s): |
Natesan
T.R, V. Rhymend Uthariaraj, George Washington .D.
|
Abstract: |
Any agency selecting candidates for admission to any
professional education has to administer a common entrance examination,
evaluate the responses and offer seats accoring to their merit. This task
has two parts viz., conduct of examination and admission. In this paper a
process oriented data model for the conduct of examination and admission
process has been developed and implemented, based on statistical and
mathematical models. The schedule for online real time registration for
the examination at various centres is based on a statistical model and the
centres for the conduct of counselling are selected based on a
mathematical programming model. This system has been implemented through
online real time distributed database with secured Virtual Private Network
(VPN) |
|
Title: |
STORAGE
OF COMPLEX BUSINESS RULES IN OBJECT DATABASES |
Author(s): |
Dalen Kambur,
Mark Roantree |
Abstract: |
True integration of large systems requires sharing of
information stored in databases beyond sharing of pure data: business
rules associated with this data must be shared also. This research focuses
on providing a mechanism for defining, storing and sharing business rules
across different information systems, in an area where existing
technologies are weak. In this paper, we present the pre-integration stage
where individual business rules are stored in the database for subsequent
exchange applications and information systems. |
|
Title: |
A
GRAPHICAL LANGUAGE FOR DEFINING VIEWS IN OBJECT ORIENTED
DATABASES |
Author(s): |
Elias Choueiri,
Marguerite Sayah |
Abstract: |
Within the framework of an Object Oriented Database
Graphical Query Environment for casual end users, a View Definition
Mechanism conceived for users who are expert in their application domain,
but not necessarily computer specialists, is proposed in this paper. In
this mechanism, a concentration is made on the strength of the graphical
view definition language and on the conviviality of the user interface.
The view definition language offers adaptation operations to the work
context and restructuring operations on both attributes and classes that
take into consideration the structure’s nesting and inheritance of the
database classes. The user interface conviviality is based on the
graphical visualization of the portion of the database schema that
represents the domain of interest for a user group, and on the use of the
graphical language for view definition. To eliminate crossings between
different links of the visualized composition hierarchy, a method for
graphical visualization is introduced. |
|
Title: |
A
TRANSPARENT CLIENT-SIDE CACHING APPROACH FOR APPLICATION SERVER
SYSTEMS |
Author(s): |
Daniel Pfeifer,
Zhenyu Wu |
Abstract: |
In recent years, application server technology has become
very popular for building complex but mission-critical systems. However,
the resulting solutions tend to suffer from serious performance and
scalability bottlenecks, because of their distributed nature and their
various software layers. This paper deals with the problem by presenting a
new approach about transparently caching results of a service interface's
read-only methods on the client side. Cache consistency is provided by a
descriptive cache invalidation model which may be specified by an
application programmer. As the cache layer is transparent to the server as
well as to the client code, it can be integrated with relatively low
effort even in systems that have already been implemented. Early
experimental results show that the approach is effective in improving a
server's response times and its transactional throughput. Roughly
speaking, the overhead for cache maintenance is small when compared to the
cost for method invocations on the server side. The cache's performance
improvements are dominated by the fraction of read method invocations and
the cache hit rate. Moreover, the cache can be smoothly integrated with
traditional caching strategies acting on other system layers (e. g.
caching of dynamic Web pages on a Web server). The presented approach as
well as the related prototype are not restricted to application server
scenarios but may be applied to any kind of interface-based software
layers. |
|
Title: |
EFFICIENT
STORAGE FOR XML DATABASES |
Author(s): |
Weiyi Ho,
Dave Elliman, Li Bai |
Abstract: |
The widespread activity involving the Internet and the Web
causes huge amount of electronic data to be generated everyday. This
includes, in particular, semi-structured textual data such as electronic
documents, computer programs, log files, transaction records, literature
citations, and emails. Storing and manipulating the data thus produced has
proven difficult. As conventional DBMSs are not suitable for handling
semi-structured data, there is a strong demand for systems that are
capable of handling large volumes of complex data in an efficient and
reliable way. The Extensible Markup Language (XML) provides such solution.
In this paper, we present the concept of ‘vertical view model’ and its
uses as a mapping mechanism for converting complex XML data to relational
database tables, and as a standalone data model for storing complex XML
data. |
|
Title: |
DATA
MANAGEMENT: THE CHALLENGE OF THE FUTURE |
Author(s): |
Alan Hodgett
|
Abstract: |
There has been an explosion in the generation of data in
organizations. Much of this data is both unstructured and decentralized.
This raises a number of issues for data management in organizations. This
paper reports on an investigation that was undertaken in Australia to
study the way in which organizations were dealing with the growth and
proliferation of data and are planning for the future. The results show a
high level of consciousness of the issues but indicate a prevalent
optimism that technology will continue to provide solutions to present and
future problems facing organizations. It appears that much magnetically
recorded data will inevitably be lost over the next few decades unless
positive actions are taken now to preserve the data. |
|
Title: |
TOWARDS
A TIMED-PETRI NET BASED APPROACH FOR THE SYNCHRONIZATION OF A MULTIMEDIA
SCENARIO |
Author(s): |
Abdelghani GHOMARI
|
Abstract: |
This article proposes a new approach for the
synchronization of a multimedia scenario based on a new class of
p-temporal Petri nets called p-RdPT+. One essential phase during the
synchronization of multimedia scenario is related to the characterization
of their logical and temporal structure. This structure is expressed
through a set of composition rules and synchronization constraints depend
on user interactions. An inconsistent situation is detected when some of
the constraints specified by the author can not be met during the
presentation. Hence, our approach permits verification of the
specification by temporal simulation of the Petri net automatically
generated or by analysing the graph of accessibility derived from the
generated p-RdPT+ model. |
|
Title: |
PLANNING
FOR ENTERPRISE COMPUTING SERVICES: ISSUES AND NECESSITIES ANALYZED
|
Author(s): |
Jason Tseng,
Emarson Victoria |
Abstract: |
While planning, simulation and modeling tools exist for
fields like network management and capacity/workload planning, little is
known about automated planning tools for computing services. Considering
the complexities and difficulties in deploying and managing computing
infrastructure and services, we need to examine their planning processes
instead, to augment existing enterprise management and planning solutions.
In this paper, we present the motivation and advantages of a planning tool
that automates the planning of computing services. This requires us to
consider the issues and problems in deploying and managing computing
services and their infrastructure. It allows us to understand why and how,
such a planning tool can be used to alleviate, if not eliminate some of
these problems. The planning tool works by actively abstracting properties
of actual computing components using an information model/framework and
formulating rules to analyze and automate the planning activity, using
only abstracted component representations. This will pave the way for
plans that closely reflect the actual computing environment, thus allowing
users to leverage the flexibility and virtualization in the planning
environment |
|
Title: |
EXTENDING
GROUPWARE FOR OLAP |
Author(s): |
Sunitha Kambhampati,
Daniel Ford, Vikas Krishna, Stefan Edlund |
Abstract: |
While applications built on top of groupware systems are
capable of managing mundane tasks such as scheduling and email, they are
not optimised for certain kinds of applications, for instance generating
aggregated summaries of scheduled activities. Groupware systems are
primarily designed with online transaction processing in mind, and are
highly focused on maximizing throughput when clients concurrently access
and manipulate information on a shared store. In this paper, we give an
overview and discuss some of the implementation details of a system that
transforms groupware Calendaring & Scheduling (C&S) data into a
relational OLAP database optimised for these kinds of analytical
applications. We also describe the structure of the XML documents that
carry incremental update information between the source groupware system
and the relational database, and show how the generic structure of the
documents enables us to extend the infrastructure to other groupware
systems as well. |
|
Title: |
REPCOM:
A CUSTOMISABLE REPORT GENERATOR COMPONENT SYSTEM USING XML-DRIVEN,
COMPONENT-BASED DEVELOPMENT APPROACH |
Author(s): |
Sai
Peck Lee, Chee Hoong Leong |
Abstract: |
It is undeniable that report generation is one of the most
important tasks in many companies regardless of the size of the company. A
good report generation mechanism can increase a company’s productivity in
terms of effort and time. This is more obvious in some startup companies,
which normally use some in-house report generators. Application
development could be complex and thus software developers might require
substantial efforts in maintaining application program code. In addition,
most of the report generators use a different kind of format to store the
report model. An application is no longer considered an enterprise-level
product if XML is not being used elsewhere. This paper introduces a
XML-driven and Component-based development approach to report generation
with the purpose of promoting portability, flexibility and genericity. In
this approach, report layout is specified using user-defined XML elements
together with queries that retrieve data from different databases. A
report is output as an HTML document, which can be viewed using an
Internet browser. This paper presents the approach using an example and
discusses the usage of the XML-driven report schema and how the proposed
reusable report engine of a customisable report generator component system
works to output an HTML report format. The customisable report generator
component system is implemented to support heterogeneous database
models |
|
Title: |
E-LEARNING
INFORMATION MANAGEMENT ISSUES IN XML-BASED MEDIATION |
Author(s): |
Boris Rousseau,
Eric Leray, Micheal O'Foghlu |
Abstract: |
The advancement in XML-based mediation has made a
significant impact on the area of E-Learning. Search engines have now been
provided with new ways to improve resource discovery and new tools to
customise resulting content. In the early days of XML, this work was
undertaken within the context of the European funded project GESTALT
(Getting Educational System Talk Across Leading Edge Technologies).
Building on this experience, new improvement came from the European funded
project GUARDIANS (Gateway for User Access to Remote Distributed
Information And Network Services). However, due to the lack of support for
native XML databases and XML querying languages, search facilities were
limited. This paper builds upon the achievements of both projects and
proposes a solution for XML querying in XQuery. |
|
Title: |
THE
KINDS OF IT SERVICES MOST APPROPRIATE FOR A PARTICULAR SOURCING
STRATEGY |
Author(s): |
Patrick Wall,
Larry Stapleton |
Abstract: |
IT processes and services often differ with regard to
which sourcing strategies suits them best. The significance of IT within
any given organization and the ability of that organization to provide an
efficient and innovative information system on its own often determines
what sourcing strategy it chooses. However, it is viewed as a better
strategy to identify certain IT processes that can be maintained
internally and then outsource those that the firm sees would be maintained
better by an external vendor. This paper identifies the most commonly
insourced, outsourced and selectively sourced IT activities and then asks
the question of why is this the case. |
|
Title: |
ERP
IMPLEMENTATION, CROSS-FUNCTIONALITY AND CRITICAL CHANGE FACTORS |
Author(s): |
Rolande Marciniak,
Redouane El Amrani, Frantz Rowe, Marc Bidan, Bénédicte
Geffroy-Maronnat |
Abstract: |
ERP (Enterprise Resource Planning) systems are
characterised by particular features such as functional coverage,
interdependent relationships, single database and standard management and
processing rules; all of which are capable of bringing about various
degrees of change within the company and, potentially, encourage a more
cross-functional overview of it. However, few quantitative studies have
been conducted to measure these effects. This is the background to this
paper, which studied 100 French companies to arrive at the following
assessment of ERP adoption. It then goes on to test the relationships
between the factors influencing the ERP lifecycle ((preparation
(organizational vision, process re-engineering), engineering (specific
developments), implementation strategy (functional coverage and speed)),
the perception of a more cross-functional overview of the company and,
more globally, the scope of the change this technology brings about within
the company. All these factors play significant roles, with functional
coverage appearing to be a particularly important consideration, which
should be addressed in future research. |
|
Title: |
LAB
INFORMATION MANAGEMENT SYSTEM FOR QUALITY CONTROL IN WINERIES |
Author(s): |
Manuel Urbano
Cuadrado, Maria Dolores Luque de Castro, Pedro Perez Juan |
Abstract: |
The great number of analysis necessary to carry out during
the wine production, as well as the storage, treatment and careful study
and discussion of the data these analyses provide is of paramount
importance for taking correct decisions for a better quality of both the
winery and the wine it produces. We describe a system devote to overall
management of information generate in the wine production processes. The
system based on otirntation to objects technology allows quality control
of the wine production in wineries and enables the integration of
semiautomated and automated analytical processes. |
|
Title: |
INFORMATION
SYSTEMS IN MEDICAL IMAGERY: CASE OF THE HOSPITAL OF BAB EL OUED |
Author(s): |
Abdelkrim MEZIANE
|
Abstract: |
The digital medical images got by the different existing
modalities, and processed by powerful computers, became a very powerful
means of diagnosis and economy. In Algeria, the patient is responsible of
the images which are delivered to him. These images are most of the time,
lost, not identified (name, date,…), or simply damaged for many reasons.
Doctors and radiologists are sometimes, if not most of the time, obliged
to ask the same patient to make the same radiography several times. The
Algerian park of medical images tools is not well known or exhaustively
assessed. The Algerian government reserves an important part of its budget
to health medical care. A part of this budget goes to complementary
medical tests, such as very expensive images paid by the taxpayer. Some
solutions do exist in order to reduce these costs, by investing a small
amount of money at the beginning. |
|
Title: |
SHIFTING
FROM LEGACY SYSTEMS TO A DATA MART AND COMPUTER ASSISTED INFORMATION
RESOURCES NAVIGATION FRAMEWORK |
Author(s): |
Nikitas Karanikolas, Christos Skourlas |
Abstract: |
Computer Assisted Information Resources Navigation (CAIRN)
was specified, in the past, as a framework that allows the end-users to
import and store full text and multimedia documents and then retrieve
information using Natural Language or field based queries. Our CAIRN
system is a general tool that has focused on medical information covering
the needs of physicians. Today, concepts related to Data Mining and Data
Marts have to be incorporated into such a framework. In this paper a
CAIRN-DAMM (Computer Assisted Medical Information Resources Navigation
& Diagnosis Aid Based On Data Marts & Data Mining) environment is
proposed and discussed. This integrated environment offers: document
management, multimedia documents retrieval, a Diagnosis–aid subsystem and
a Data Mart subsystem that permits the integration of legacy system’s
data. The diagnosis is based on the International Classification of
Diseases and Diagnoses, 9th revision (ICD-9). The document collection
stored in the CAIRN-DAMM system consists of data imported from the
Hospital Information System (HIS), laboratory tests extracted from the
Laboratory Information System (LIS), patient discharge letters,
ultrasound, CT and MRI images, statistical information, bibliography, etc.
There are also methods permitting us to propose, evaluate and organize in
a systematic way uncontrolled terms and to propose relationships between
these terms and ICD-9 codes. Finally, our experience from the use of the
tool for creating a Data Mart at the ARETEION University Hospital is
presented. Experimental results and a number of interesting observations
are also discussed. |
|
Title: |
ON
OPERATIONS TO CONFORM OBJECT-ORIENTED SCHEMAS |
Author(s): |
Alberto Abelló,
Elena Rodriguez, Elena Rodríguez, Marta Oliva,
José Samos, Fèlix Saltor, Eladio Garví |
Abstract: |
To build a Cooperative Information System from several
pre-existing heterogeneous systems, the schemas of these systems must be
integrated. Operations used for this purpose include conforming
operations, which change the form of a schema. In this paper, a set of
primitive conforming operations for Object-Oriented schemas are presented.
These operations are organized in matrixes according to the
Object-Oriented dimensions -Generalization/Specialization,
Aggregation/Decomposition- on which they operate. |
|
Title: |
A
MULTI-LEVEL ARCHITECTURE FOR DISTRIBUTED OBJECT BASES |
Author(s): |
Markus Kirchberg
|
Abstract: |
The work described in this article arises from two needs.
First, there is still a need for providing more sophisticated database
systems than just relational ones. Secondly, there is a growing need for
distributed databases. These needs are adressed by fragmenting schemata of
a generic object data model and providing an architecture for its
implementation. Key features of the architecture are the use of abstract
communicating agents to realize database transactions and queries, the use
of an extended remote procedure call to enable remote agents to
communicate with one another, and the use of multi-level transactions.
Linguistic reflection is used to map database schemata to the level of the
agents. Transparency for the users is achieved by using dialogue objects,
which are extended views on the database. |
|
Title: |
INVESTIGATING
THE EFFECTS OF IT ON ORGAISATIONAL DESIGN VARIABLES , TOWARDS A
THEORETICAL FRAMEWORK |
Author(s): |
Rahim Ghasemiyeh,
Feng Li |
Abstract: |
Over the past decades many papers have been published
about the effects of Information Technology (IT) on organisations. However
despite the facts that IT has become a fundamental variable for
organisational design very few studies have been done to explore this
vital issue in a systematic and convincing fashion. The small amount of
information and few theories available on the effects of IT on
organisational design is surprising. Also one major efficiency of previous
studies is the lack of empirical evidence. This has led researchers to
describe IT in general ways and resulted in different and very often
contradictory findings. Many researchers have become very concerned about
the shortfall of comprehensive study on organizational design and IT which
has been apparent for decades. One objective of this research is to fill
this gap. This study will investigate three questions, aiming to develop a
theoretical framework to evaluate the effects of IT on organisational
design,. What are the effects of IT on organisational design variables?
How IT influences organisational design variables? Which effects are
resulted from which IT technologies? These could be considered as the most
important features of this study, which are different with respect to
previous literature. |
|
Title: |
SERVICES
PROVIDERS’ PATTERNS FOR CLIENT/SERVER APPLICATIONS |
Author(s): |
Samar TAWBI,
Bilal CHEBARO |
Abstract: |
In this paper, we define two patterns that fall under the
category of the architectural patterns described in (Shaw, 1996), to
provide solutions for client-server applications. The first pattern
defines the structure of a client-server application by defining the
server's functionality in the form of standardized services, and the
second defines the structure of a service in this type of application. The
solution follows the patterns’ definition prototype used in (Gamma,
1995). |
|
Title: |
A
DISTRIBUTED JOB EXECUTION ENVIRONMENT USING ASYNCHRONOUS MESSAGING AND WEB
TECHNOLOGIES |
Author(s): |
Rod Fatoohi,
Nihar Gokhale |
Abstract: |
This is a project for developing an asynchronous approach
to distributed job execution of legacy code. A job execution environment
is a set of tools used to run jobs, generated to execute a legacy code,
and handles different input and output values for each run. Current job
execution and problem solving environments are mostly based on synchronous
messaging and customized API that needs to be ported to different
platforms. Here we are introducing an Internet-based job execution
environment using off-the-shelf J2EE (Java 2 Enterprise Edition)
components. The environment allows the execution of computational
algorithms utilizing standard Internet technologies such as Java, XML, and
asynchronous communication protocols. Our environment is based on
four-tier client/server architecture and uses Java messaging, for
inter-process communication, and XML fro job specification. It has been
tested successfully using several legacy simulation codes on pools of
Windows 2000 and Solaris systems. |
|
Title: |
DRUID:
COUPLING USER WRITTEN DOCUMENTS AND DATABASES |
Author(s): |
André
Flory, Frédérique Laforest, Youakim BADR |
Abstract: |
Most database applications capture their data using
graphical forms. Text fields have limited size and predefined types.
Although data in fields are associated with constrains, it should be
modeled in a suitable way to conform to a rigid schema. Unfortunately, too
much constrains on data are not convenient in human activities where most
activities are document-centric. In fact, documents become a natural way
for human production and consumption. Nowadays, an increased interest is
put on managing data with irregular structures, exchanging documents over
the net, and manipulating their contents as efficiently as with structured
data. In this paper, we introduce DRUID, a comprehensive document
capturing and wrapping system. It ensures flexible and well-adapted
information capture based on a Document User Interface and at the same
time information retrieval based on databases. DRUID relies on a wrapper
that transforms documents contents into relevant data. Also, it provides
an expressive specification language for end-users to write domain-related
extraction patterns. We validate our information system with a prototype
of different modules, the primary realization is promising for a wide
range of applications that use documents as a mean to store, exchange and
query information. |
|
Title: |
TOWARD
A FRAMEWORK FOR MANAGING INTERNET-ORIENTED DATABASE RESOURCES |
Author(s): |
Guozhou Zheng,
Chang Huang, Zhaohui Wu |
Abstract: |
The term “Grid” is used to describe those architectures
that manage the distributed resources across the Internet. This paper is
intended to introduce the Database Grid, an Internet oriented resource
management architecture for database resource. We identify the basic
requirements on database in two major application domains: e-science and
e-business. Next, we illustrate how a layered service architecture can
fulfil these emerging data sharing and data management requirements from
Grid computing application. We introduce a series of protocols to define
the proposed services. |
|
Title: |
A
FRAMEWORK FOR GENERATING AND MAINTAINING GLOBAL SCHEMAS IN HETEROGENEOUS
MULTIDATABASE SYSTEMS |
Author(s): |
Rehab Duwairi
|
Abstract: |
The problem of creating a global schema over a set of
heterogeneous databases is becoming more and more important due the
availability of multiple databases within organizations. The global schema
should provide a unified representation of local (possibly heterogeneous)
local schemas by analyzing them (to exploit their semantic contents),
resolving semantic and schematic discrepancies among them, and producing a
set of mapping functions that translate queries posed on the global schema
to queries posed on the local schemas. In this paper, we provide a general
framework that supports the integration of local schemas into a global
one. The framework takes into consideration the fact that local schemas
are autonomous and may evolve over time, which makes the definition of the
global schema obsolete. We define a set of integration operators that
integrates local schemas, based on the semantic relevance of their
classes, into a set of virtual classes that constitute the global schema.
We also define a set of modifications that can be applied to local schemas
as a consequence of their local autonomy. For every local modification, we
define a propagation rule that will automatically disseminate the effects
of that modification to the global schema without having to regenerate it
from scratch via integration. |
|
Title: |
A
SCALABLE DISTRIBUTED SEARCH ENGINE FOR INTRANET INFORMATION
RETRIEVAL |
Author(s): |
Minoru Uehara,
Minoru Udagawa, Yoshifumi Sakai, Hideki Mori,
Nobuyoshi Sato |
Abstract: |
Intranet information retrieval is very important for
corporations in business. They are trying to discover the useful knowledge
from hidden web pages by using data mining, knowledge discovery and so on.
In this process, search engine is useful. However, conventional search
engines, which are based on centralized architecture, are not suited for
intranet information retrieval because intranet information is frequently
updated. Centralized search engines take a long time to collect web pages
by crawler, robots and so on. So, we have developed a distributed search
engine, called Cooperative Search Engine (CSE), in order to retrieve fresh
information. In CSE, a local search engine located in each Web server
makes an index of local pages. And, a Meta search server integrates these
local search engines in order to realize a global search engine. In such a
way, the communication delay occurs at retrieval time. So, we have
developed several speedup techniques in order to realize fast retrieval.
As this result, we have succeeded in increasing the scalability of CSE. In
this paper, we describe speedup techniques and evaluate them. |
|
Title: |
A
WEB APPLICATION FOR ENGLISH-CHINESE CROSS LANGUAGE PATENT
RETRIEVAL |
Author(s): |
Wen-Yuan Hsiao,
Jiangping Chen, Elizabeth Liddy |
Abstract: |
This paper describes an English-Chinese cross language
patent retrieval system built on a commercial database management
software. The system makes use of various software products and lexical
resources for the purpose of helping English native speakers to search for
Chinese patent information. This paper reports the overall system design
and cross language information retrieval (CLIR) experiments conducted for
performance evaluation. The experimental results and the follow-up
analysis demonstrated that commercial database systems could be used as an
IR system with reasonable performance. Better performance could be
achieved if the translation resources were customized to the document
collection of the system, or more sophisticated translation disambiguation
strategies were applied. |
|
Title: |
TRIGGER-BASED
COMPENSATION IN WEB SERVICE ENVIRONMENTS |
Author(s): |
Randi Karlsen,
Thomas Strandenaes |
Abstract: |
In this paper we describe a technique for implementing
compensating transactions, based on the active database concept of
triggers. This technique enables specification and enforcement of
compensation logic in a manner that facilitates consistent and
semi-automatic compensation. A web service, with its loosely-coupled
nature and autonomy requirements, represents an environment well suited
for this compensation mechanism. |
|
Title: |
AN
ARCHITECTURE OF A SECURE DATABASE FOR NETWORKED COLLABORATIVE
ACTIVITIES |
Author(s): |
Akira
Baba, Michiharu Kudo, Kanta Matsuura, Kanta Matsuura
|
Abstract: |
Open network can be used for many purposes, e-commerce or
e-government, etc. Different from those conventional applications, we
consider networked collaborative activities, for example networked
research activities. This application might be very useful and research
activities could be significantly promoted. However, we must care about
many security problems. Among those problems, we focus on an architecture
of a secure database in this paper. The design of such an architecture is
not a trivial task, since the data sets in database could be composed of
wide range of data types, and each data type needs to satisfy its own
security properties, including not only security but also an appropriate
management of intellectual-property right, and so on. Thus, we design an
architecture of a secure database, considering data types and various
security operations. |
|
Title: |
USING
INFORMATION TECHNOLOGIES FOR MANAGING COOPERATIVE INFORMATION AGENT-BASED
SYSTEMS |
Author(s): |
Nacereddine ZAROUR,
Mahmoud BOUFAIDA, Lionel SEINTURIER |
Abstract: |
One of the most important problems encountered by the
cooperation among distributed infomation systems is that of heterogeneity
that is often not easy to deal with. This problem requires the use of the
best combination of software and hardware components for each
organization. However, the few suggested approaches for managing virtual
factories have not led to satisfaction. Along with motivating the
importance of such systems, this paper describes the major design goals of
agent-based architecture for supporting the cooperation of heterogeneous
information systems. It also shows how this architecture can be
implemented using the combination of XML and CORBA technologies. This
combination guarantees the interoperability of legacy systems regardless
respectiveley of their data models and platforms heterogeneity and,
therefore, improves the cooperation process. Examples are given from the
supply chains of manufacturing enterprises. |
|
Title: |
MODELING
A MULTIVERSION DATA WAREHOUSE: A FORMAL APPROACH |
Author(s): |
Tadeusz Morzy,
Robert Wrembel |
Abstract: |
A data warehouse is a large centralized repository that
stores a collection of data integrated from external data sources (EDSs).
The purpose of building a data warehouse is: to provide an integrated
access to distributed and usually heterogeneous information, to provide a
platform for data analysis and decision making. EDSs are autonomous in
most of the cases. In a consequence, their content and structure change in
time. In order to keep the content of a data warehouse up to date, after
source data changed, various warehouse refreshing techniques have been
developed, mainly based on an incremental view maintenance. A data
warehouse will also need refreshing after a schema of an EDS changed. This
problem has, however, received little attention so far. Few approaches
have been proposed and they tackle the problem by using mainly temporal
extensions to a data warehouse. Such techniques expose their limitations
in multi–period quering. Moreover, in order to support predictions of
trends by decision makers what–if analysis is often required. For these
purposes, multiversion data warehouses seem to be very promising. In this
paper we propose a model of a multiversion data warehouse, and show our
prototype implementation of such a multiversion data warehouse. |
|
Title: |
TRADDING
PRECISION FOR TIMELINESS IN DISTRIBUTED REAL-TIME DATABASES |
Author(s): |
Bruno SADEG
|
Abstract: |
Many information systems need not to obtain complete or
exact answers to queries submitted via a DBMS (Database Management
System). Indeed, in certain real-time applications, incomplete results
obtained timely are more interesting than complete results obtained late.
When the applications are distributed, DBMSs on which these applications
are based have a main problem of managing the transactions (concurrency
control and commit processes). Since these processes must be done timely
(such as each transaction meets its deadline), committing transactions
timely seems to be the main issue. In this paper, we deal with the global
distributed transaction commit and the local concurrency control problems
in applications where transactions may be decomposed into a mandatory part
and an optional part. In our model, the means to determine these parts is
based on a weight parameter which is assigned to each subtransaction. It
is used to help the coordinator process to execute the commit phase when a
transaction is close to its deadline. An other parameter, the estimated
execution time, is used by each participant site in combination with the
weight to solve the possible conflicts that may occur between local
subtransactions. The mechanisms used to deal with these issues is called
RT-WEP (Real-Time-Weighted Early Prepare) protocol. Some simulation have
made to compare RT-WEP protocol with two other protocols designed to the
same purpose. The results have shown that RT-WEP protocol may be applied
efficiently in a distributed real-time context by allowing more
transactions to meet their deadlines. |
|
Title: |
A
MODEL-DRIVEN APPROACH FOR ITEM SYNCHRONIZATION AND UCCNET INTEGRATION IN
LARGE E-COMMERCE ENTERPRISE SYSTEMS |
Author(s): |
Santhosh Kumaran,
Fred Wu, Simon Cheng, Mathews Thomas,
Santhosh Kumaran, Amaresh Rajasekharan, Ying Huang
|
Abstract: |
The pervasive connectivity of the Internet and the
powerful architecture of the WWW are changing many market conventions and
creating a tremendous opportunity for conducting business on the Internet.
Digital marketplace business models and the advancement of Web related
standards are tearing down walls within and between different business
artifacts and entities at all granularities and at all levels, from
devices, operating systems and middleware to directory, data, information,
application, and finally the business processes. As a matter of fact,
business process integration (BPI), which entails the integration of all
the facets of business artifacts and entities, is emerging as a key IT
challenge. In this paper, we describe our effort in exploring a new
approach to address the complexities of BPI. More specifically, we study
how to use a solution template based approach for BPI and explore the
validity of this approach with a frequently encountered integration
problem, the item synchronization problem for large enterprises. The
proposed approach can greatly reduce the complexities of the business
integration task and reduce the time and amount of effort of the system
integrators. Different customers are deploying the described Item
Synchronization system. |
|
Title: |
DATA
POSITION AND PROFILING IN DOMAIN-INDEPENDENT WAREHOUSE CLEANING |
Author(s): |
Ajumobi Udechukwu,
Christie Ezeife |
Abstract: |
A major problem that arises from integrating different
databases is the existence of duplicates. Data cleaning is the process for
identifying two or more records within the database, which represent the
same real world object (duplicates), so that a unique representation for
each object is adopted. Existing data cleaning techniques rely heavily on
full or partial domain knowledge. This paper proposes a positional
algorithm that achieves domain independent de-duplication at the attribute
level. The paper also proposes a technique for field weighting through
data profiling, which, when used with the positional algorithm, achieves
domain-independent cleaning at the record level. Experiments show that the
positional algorithm achieves more accurate de-duplication than existing
algorithms. |
|
Title: |
OPTIMIZING
ACCESS IN A DATA INTEGRATION SYSTEM WITH CACHING AND MATERIALIZED DATA
|
Author(s): |
Bernadette Farias
Lóscio, Ana Carolina Salgado, Maria da Conceição Moraes Batista
|
Abstract: |
Data integration systems are planned to offer uniform
access to data from heterogeneous and distributed sources. Two basic
approaches have been proposed in the literature to provide integrated
access to multiple data sources. In the materialized approach, data are
previously accessed, cleaned, integrated and stored in the data warehouse
and the queries submitted to the integration system are evaluated in this
repository without direct access to the data sources. In the virtual
approach, the queries posed to the integration system are decomposed into
queries addressed directly to the sources. The data obtained from the
sources are integrated and returned to the user. In this work we present a
data integration environment to integrate data distributed on multiple web
data sources which combines features of both approaches supporting the
execution of virtual and materialized queries. Other distinguished feature
of our environment is that we also propose the use of a cache system in
order to answer the most frequently asked queries. All these resources are
put together with the goal of optimizing the overall query response
time. |
|
Title: |
GLOBAL
QUERY OPTIMIZATION BASED ON MULTISTATE COST MODELS FOR A DYNAMIC
MULTIDATABASE SYSTEM |
Author(s): |
Qiang Zhu
|
Abstract: |
Global query optimization in a multidatabase system (MDBS)
is a challenging issue since some local optimization information such as
local cost models may not be available at the global level due to local
autonomy. It becomes even more difficult when dynamic environmental
factors are taken into consideration. In our previous work, a qualitative
approach was suggested to build so-called multistate cost models to
capture the performance behavior of a dynamic multidatabase environment.
It has been shown that a multistate cost model can give a good cost
estimate for a query run in any contention state in the dynamic
environment. In this paper, we present a technique to perform query
optimization based on multistate cost models for a dynamic MDBS. Two
relevant algorithms are proposed. The first one selects a set of
representative system environmental states for generating an execution
plan with multiple versions for a given query at compile time, while the
second one efficiently determines the best version to invoke for the query
at run time. Experiments demonstrate that the proposed technique is quite
promising for performing global query optimization in a dynamic MDBS.
Compared with related work on dynamic query optimization, our approach has
an advantage of avoiding the high overhead for modifying or re-generating
an execution plan for a query based on dynamic run-time information. |
|
Title: |
A
DATA, COMPUTATION, KNOWLEDGE GRID THE CASE OF THE ARION SYSTEM |
Author(s): |
Spyros
Lalis, Manolis Vavalis, Kyriakos Kritikos, Antonis
Smardas, Dimitris Plexousakis, Marios Pitikakis,
Catherine Houstis, Vassilis Christophides |
Abstract: |
The ARION system provides basic e-services of search and
retrieval of objects in scientific collections, such as, datasets,
simulation models and tools necessary for statistical and/or visualization
processing. These collections may represent application software of
scientific areas, they reside in geographically disperse organizations and
constitute the system content. The user may invoke on-line computations of
scientific datasets when the latter are not found into the system. Thus,
ARION provides the basic infrastructure for accessing and deriving
scientific information in an open, distributed and federated system. |
|
Title: |
SCANNING
A LARGE DATABASE ONCE TO MINE ASSOCIATION RULES |
Author(s): |
Frank
Wang
|
Abstract: |
Typically 95% of the data in the transaction databases are
zero. When it comes to sparse, the performance quickly degrades due to the
heavy I/O overheads in sorting and merging intermediate results. In this
work, we first introduce a list representation in main memory for storing
and computing datasets. The sparse transaction dataset is compressed as
the empty cells are removed Accordingly we propose a ScanOnce algorithm
for association rule mining on the platform of list representation, which
just needs to scan the transaction database once to generate all the
possible rules. In contrast, the well-known Apriori algorithm requires
repeated scans of the databases, thereby resulting in heavy I/O accesses
particularly when considering large candidate datasets. Attributing to its
integrity in data structure, the complete itemset counter tree can be
stored in a (one-dimensional) vector without any missing gap, whose
direct-addressing capability ensures fast access to any counter. In our
opinion, this new algorithm using list representation economizes storage
space and accesses. The experiments show that this ScanOnce algorithm
beats classic Apriori algorithm for large problem sizes, by factors
ranging from 2 to more than 6. |
|
Title: |
INTEGRATION
OF DISTRIBUTED SOFTWARE PROCESS MODELS |
Author(s): |
Mohamed Ahmed-nacer,
Nabila Lardjane |
Abstract: |
Developing software-in-the-large involves many developers,
with experts in various aspects of software development and in various
aspects of the application area. This paper presents an approach to
integrate software process models in a distributed context. It is based on
the fusion of process fragments (components) defined with the UML notation
(Unified Modelling Language). The integration methodology presented allows
unifying the various fragments both at the static level as well as at the
dynamic level (behavioural). We consider various possible semantic
conflicts; formal definitions of the inter-fragments properties are
formulated and solutions for these conflicts are proposed. This
integration approach provides multiple solutions for the integration
conflicts and gives the possibility to improve and design new software
process models by a merging of reusable process fragments. |
|
Title: |
A
BITEMPORAL STORAGE STRUCTURE FOR A CORPORATE DATA WAREHOUSE |
Author(s): |
Alberto Abelló,
Carme Martín |
Abstract: |
This paper brings together two research areas, i.e. Data
Warehouses and Temporal Databases, involving representation of time.
Looking at temporal aspects within a data warehouse, more similarities
than differences between temporal databases and data warehouses have been
found. The first closeness between these areas consists in the possibility
of a data warehouse redefinition in terms of a bitemporal database. A
bitemporal storage mechanism is proposed along this paper. In order to
meet this goal, a temporal study of data sources is developed. Moreover,
we will show how Object-Oriented temporal data models contribute to add
the integration and subject-orientation that is required by a data
warehouse. |
|
Title: |
TOWARD
A DOCUMENTARY MEMORY |
Author(s): |
Christine JULIEN,
Max CHEVALIER, Kais Khrouf |
Abstract: |
An organisation must enable to share knowledge and
information within its employees to optimise their tasks. However, the
volume of information contained in documents represents a major importance
for these companies. Indeed, companies may be fully reactive to any new
information and must follow the fast evolution of spread information. So,
a documentary memory, which store this information and allow end-user to
access or analyse it, constitutes a necessity for every enterprise. We
propose, in this paper, the architecture of such a system, based on a
document warehouse, allowing the storage of relevant documents and their
exploitation via the techniques of information retrieval, factual data
interrogation and information multidimensional analysis. |
|
Title: |
DISTRIBUTED
OVERLOAD CONTROL FOR REAL-TIME REPLICATED DATABASE SYSTEMS |
Author(s): |
Samia Saad-Bouzefrane,
C. Kaiser |
Abstract: |
In order to meet their temporal constraints, current
applications such as Web-based services and electronic commerce use the
technique of data replication. To take the replication benefit, we need to
develop con-currency control mechanisms with high performance even when
the distributed system is overloaded. In this paper, we present a protocol
that uses a new notion called importance value which is associated with
each real-time transaction. Under conditions of overload, this value is
used to select the most important transactions with respect to the
application transactions in order to pursue their execution ; the other
transactions are aborted. Our protocol RCCOS (Replica Concurrency-Control
for Overloaded Systems) augments the protocol MIR-ROR, a concurrency
control protocol designed for firm-deadline applications operating on
replicated real-time databases in order to manage efficiently transactions
when the distributed system is overloaded. A platform has been developped
to measure the number of transactions that meet their deadlines when the
processor load of each site is controlled. |
|
Title: |
INCREMENTAL
HORIZONTAL FRAGMENTATION OF DATABASE CLASS OBJECTS |
Author(s): |
Christie
Ezeife, Pinakpani Dey |
Abstract: |
Horizontal fragments of a class in an object-oriented
database system contain subsets of the class extent or instance objects.
These fragments are created with a set of system input data consisting of
the application queries, their access frequencies, the object database
schema with components - class inheritance and class composition
hierarchies as well as instance objects of classes. When these system
input to the fragmentation process change enough to affect system
performance, a re-fragmentation is usually done from scratch. This paper
proposes an incremental re-fragmentation method that uses mostly the
updated part of input data and previous fragments to define new fragments
more quickly, saving system resources and making the data at distributed
sites more available for network and web access. |
|
Title: |
GEONIS
- FRAMEWORK FOR GIS INTEROPERABILITY |
Author(s): |
Leonid Stoimenov,
Slobodanka Djordjevic-Kajan |
Abstract: |
This paper presents research in Geographic Information
Systems interoperability. Also, paper describes our work in development,
introduces interoperability framework called GeoNis, which uses proposed
technologies to perform integration task between GIS applications and
legacy data sources over the Internet. Our approach provides integration
of distributed GIS data sources and legacy information systems in local
community environment. |
|
Title: |
BUSINESS
CHANGE IMPACTS ON SYSTEM INTEGRATION |
Author(s): |
Fabio Rollo,
Gabriele Venturi, Gerardo Canfora |
Abstract: |
Large organizations have disparate legacy systems,
applications, processes, and data sources, which interact by means of
various kinds of interconnections. Merging of companies can increase the
complexity of system integration, with the need to integrate applications
like Enterprise Resource Planning and Customer Relationship Management.
Even if sometimes these applications provide a kind of access to their
underlying data and business logic, Enterprise Application Integration
(EAI) is still a challenge. In this paper we analyse the needs that drive
EAI with the aim of identifying the features that EAI platforms must
exhibit to enable companies to compete in the new business scenarios. We
discuss the limitations of current EAI platforms and their evaluation
methods, mainly economies of scale and economies of scope, and argue that
a shift is needed towards the economies of learning model. Finally, we
outline an EAI architecture that addresses current limitations enabling
economies of learning. |
|
Title: |
TECHNICAL
USE QUALITY IN A UNIVERSITY ENTERPRISE RESOURCE PLANNING SYSTEM:
PERCEPTIONS OF RESPONSE TIME AND ITS STRATEGIC IMPORTANCE |
Author(s): |
Michelle Morley
|
Abstract: |
Enterprise Resource Planning Systems (ERPs) are large,
complex enterprise-wide information system that offer benefits of
integration and data-richness to organisations. This paper explores the
quality issue of response times, and the impact of poor response times on
the ability of the organisation studied to achieve their strategy. The
PeopleSoft ERP was implemented within the International Centre (for
international student recruitment and support) at an Australian
University, as part of a University-wide implementation. To achieve the
goal of increased international student enrolments, fast turnaround times
on student applications are critical. The ERP offers poor response times
and this makes it difficult for the International Centre to achieve high
conversion rates (from applications to enrolments) and hence reduces the
perceived value, or ‘business quality’ (Salmela 1997), of the system to
the organisation. The paper uses a quality model developed from Eriksson
and Toern’s (1990) SOLE model, Lindroos’ (1997) Use Quality and Salmela’s
(1997) Business Quality model. |
|
Title: |
INTEGRATING
AUTOMATION DESIGN INFORMATION WITH XML |
Author(s): |
Seppo Kuikka,
Mika Viinikkala |
Abstract: |
Due to the number of parties participating in the design
phase of an automation project, various design, engineering and
operational systems are needed. At the moment, the means to transfer
information from one system to another system, so that it can be further
processed or reused, are not efficient. An integration approach in which
XML technologies are utilized for implementing systems integration is
introduced. Data content of systems are defined by XML Schema instances.
XML messages containing automation design information are transformed
using transformation stylesheets employing a generic standard vocabulary.
Loosely coupled, platform independent, data content-oriented integration
is enabled by XML technologies. A case study that proceeds according to
the approach is also described. It consists of both a software prototype
responsible for communication and data content including XML Schema
instances and transformation stylesheets for the systems covered in the
study. It is found that XML technologies seem to be a part of the right
solution. However, some issues related to schema design and
transformations are problematic. If complex systems are integrated, XML
technologies alone are not sufficient. Future developments include a
general purpose web-service solution that is to answer questions that were
not dealt with by this case study. |
|
Title: |
IMPRECISION
BASED QUERIES OVER MATERIALIZED AND VIRTUAL INTEGRATED VIEWS |
Author(s): |
Alberto Trombetta,
Danilo Montesi |
Abstract: |
The Global-As-View approach to data integration has
focused on the (semi-automatic) definition of a global schema starting
from a given set of known information sources. In this paper, we
investigate how to employ concepts and techniques to model imprecision in
defining mappings between the global schema and the source schemas and to
answer queries posed over the global schema. We propose an extended
relational algebra using fuzzy sets for defining SQL-like query mappings.
Such mappings explicitly take into account the similarities between global
and source schemas to discard source data items with low similarity and to
express the relevance of different sources in populating the global
schema. In the case the global schema is not materialized, we propose a
query rewriting technique for expressing over the sources the queries
posed over the global schema |
|
Title: |
THE
HAMLET DILEMMA ON EXTERNAL DATA IN DATA WAREHOUSES |
Author(s): |
Mattias Strand,
Marcus Olsson |
Abstract: |
Data warehouses are currently given a lot of attention;
both by academics and practitioners, and the amount of literature
describing different aspects of data warehousing is ever-increasing. Much
of this literature is covering the characteristics and the origin of the
data in the data warehouse and the importance of external data is often
pinpointed. Still, the descriptions of external data are on a general
level and the extent of external data usage is not given much attention.
Therefore, in this paper, we describe the results of an interview study,
partly aimed at outlining the current usage of external data in data
warehouses. The study was directed towards Swedish data warehouse
developers and the results shows that the usage of external data in data
warehouses is not as frequent as expected. Only 58 % of the respondents
had been working in projects that had an objective of integrating external
data. Reasons given for rather low usage were problems on assuring the
quality of the external data and lack of data warehouse maturity amongst
the user organizations. |
|
Title: |
PERFORMANCE
IMPROVEMENT OF DISTRIBUTED DATABASE MANAGEMENT SYSTEMS |
Author(s): |
Josep
Maria Muixi, August Climent |
Abstract: |
Distributed databases offer a complete range of desirable
features: availability, reliability, and responsiveness. However, all of
these benefits are at the expense of some extra management; main issues
considered in literature as the base of a tuned distributed database
system could be data replication and synchronization, concurrency access,
distributed query optimization or performance improvement. Work presented
here tries to provide some clues to the last point considering an issue
which has not been taken enough into account under our humble opinion:
load balancing of these distributed systems. It is tried to be shown how
the right load balancing policy influences the performance of a
distributed database management system, and more concretely a
shared-nothing one. |
|
Title: |
EMPIRICAL
VALIDATION OF METRICS FOR UML STATECHART DIAGRAMS |
Author(s): |
David Miranda,
Marcela Genero, Mario Piattini |
Abstract: |
It is widely recognised that the quality of Object
Oriented Software Systems (OOSS) must be assessed from the early stages of
their development. OO Conceptual models are key artifacts produced at
these early phases, which cover not only static aspects but also dynamic
aspects. Therefore, focusing on quality aspects of conceptual models could
contribute to produce better quality OOSS. While quality aspects of
structural diagrams, such as class diagrams, have being widely researched,
the quality of behavioural diagrams such as statechart diagrams have been
neglected. This fact leaded us to define a set of metrics for measuring
their structural complexity. In order to gather empirical evidence that
the structural complexity of statechart diagrams are closed with their
understandability we carried out a controlled experiment in a previous
work. The aim of this paper is to present a replication of that
experiment. The findings obtained in the replication corroborate the
results of the first experiment in the sense that at some extent, the
number of transitions, the number of states and the number of activities
influence statechart diagrams understandability. |
|
Title: |
A
SOLUTION FOR CONTEXTUAL INTEGRATION BASED ON THE CALCULATION OF A SEMANTIC
DISTANCE |
Author(s): |
Fabrice JOUANOT,
Kokou Yétongnon, Nadine Cullot |
Abstract: |
To achieve the interoperation of heterogeneous data
sources with respect to their context and rich semantics keeps yet a real
challenge. Users need to integrate useful information and query coupled
data sources in a transparent way. We propose a solution to help the
integration of heterogeneous sources according to their context. We
present a model to define contextual information associated to local data
and a mechanism which uses this semantics to compare local contexts and
integrate relevant data. Our contextual integration approach, using a rule
based language, allows us to build virtual objects in a semi-automatic
way. They play roles of transparent interfaces for end-users. |
|
Title: |
DATA
WAREHOUSE – PROCESS TO DEVELOP |
Author(s): |
Prasad
N. Sivalanka , Rakesh Agarwal |
Abstract: |
Building a data warehouse involves complex details of
analysis and design of an enterprise-having wide decision support system.
Dimensional modeling can be used to design effective and usable data
warehouses. The paper highlights the steps in the implementation of data
warehouse in a client project. All the observations and phases mentioned
in this document are with reference to the project carried out for
medium-to-large multi-dimensional databases for a client in a controlled
test environment. The recommendations, conclusions and observations made
in this document may not be generalized for all cases unless verified and
tested. |
|
Title: |
CREATING
THE DOCSOUTH PUBLISHER |
Author(s): |
Tony Bull
|
Abstract: |
In this Case Study, a Systems Integration problem is
solved using Object-Oriented Perl, XML/XSLT, and Java. Over the last two
years, the world-renowned Digitization Project ‘Documenting the American
South’ has been slowly converting its SGML-based Legacy system to an
XML-centric system. As of September 2002, the “DocSouth Publisher” has
been the latest change in realizing the new XML environment. |
|
Title: |
A
COMPARISON OF DATABASE SYSTEMS FOR STORING XML DOCUMENTS |
Author(s): |
Roger Davies,
Miguel Mira da Silva, Rui Cerveira Nunes |
Abstract: |
As the need to store large quantities of increasingly
complex XML documents augments, the requirements for database products
that claim to support XML also increases. For example, it is no longer
acceptable to store XML documents without using indices for efficient
retrieval of large collections. In this paper we analyse the current
versions of products representing the three main approaches to XML
storage: native XML databases, XML support by relational databases, and
object-oriented databases with XML support. Several products are analysed
and compared, including performance tests. Our main conclusion is that the
market urgently needs a standard query language and API, analogous to SQL
and ODBC, which were probably the main drivers for the success of
relational databases. |
|
Title: |
AUTOMATED
DATA MAPPING FOR CROSS ENTERPRISE DATA INTEGRATION |
Author(s): |
Stefan Böttcher,
Sven Groppe |
Abstract: |
Currently, there are multiple different classifications
for product descriptions used in enterprise-internal applications and
cross-enterprise applications, e.g. E-procurement systems. A key problem
is to run applications developed for one catalogue on product descriptions
that are stored in a different classification. A common solution is that a
catalogue specialist manually maps different classifications onto each
other. Our approach avoids unnecessary manual work for mapping and
automatically generates mappings between different classifications
wherever possible. This allows us to run E-procurement applications on
different catalogues with a fairly reduced manual work needed for mapping,
what we consider to be an important step towards enterprise application
integration. |
|
Title: |
XML-BASED
OLAP QUERY PROCESSING IN A FEDERATED DATA WAREHOUSES |
Author(s): |
Wolfgang Essmayr,
Edgar Weippl, Johannes Huber, Oscar Mangisengi |
Abstract: |
Today, XML is the format of choice to implement
interoperability between systems. This paper addresses the XML-based query
processing for heterogeneous OLAP data warehouses in a federated
architecture. In our approach, XML, as an intermediary representation, can
be used as a basis for federated queries and queries for local OLAP data
warehouses, whereas XML DTD can be used for query language definition and
validation of a XML federated query. |
|
Title: |
THE
ENHANCED GREEDY INTERCHANGE ALGORITHM FOR THE SELECTION OF MATERIALIZED
VIEWS UNDER A MAINTENANCE COST CONSTRAINT IN DATA WAREHOUSES |
Author(s): |
Omar Karam,
Osman Ibrahim, Rasha Ismail, Mohamed El-Sharkawy |
Abstract: |
A Data Warehouse is a central repository of integrated
information available for the purpose of efficient decision support or
OLAP queries. One of the important decisions when designing a data
warehouse is the selection of views to materialize and maintain in a data
warehouse. The goal is to select an appropriate set of materialized views
so as to minimize the total query response time and the cost of
maintaining the selected views under the constraint of a given total view
maintenance time. In this paper, the maintenance cost is incorporated to
the Greedy Interchange Algorithm (GIA). The performance and behavior of
the Greedy Algorithm considering the maintenance costs (GAm) and the
proposed Greedy Interchange Algorithm considering maintenance cost (GIAm)
are examined through experimentation. The GIAm improves the results over
the GAm by 56.5%, 60.6% and 80% for different maintenance time constraints
100%, 75% and 40% of total maximum maintenance time. An enhancement to the
GIAm is proposed, the enhancement introduced depends on selecting a subset
of views to which the GIA is applied rather than all the views of a view
graph. This selection is based upon views dependencies and result in
substantial run time. |
|
Title: |
RANKING
AND SELECTING COMPONENTS TO BUILD SYSTEMS |
Author(s): |
Alberto Sillitti,
Paolo Predonzani, Giampiero Granatella, Tullio Vernazza,
Giancarlo Succi |
Abstract: |
Component-Based Software Engineering (CBSE) allows
developers to build systems using existing components. Developers need to
find the best set of components that implements most of required features.
Retrieving components manually can be very complicated and time expensive.
Tools that partially automate this task help developers to build better
systems with less effort. This paper proposes a methodology for ranking
and selecting components to build an entire system instead of retrieving
just a single component. This methodology was developed in the European
project CLARiFi (CLear And Reliable Information For Integration). |
|
Title: |
A
CASE STUDY FOR A QUERY-BASED WAREHOUSING TOOL |
Author(s): |
Rami Rifaieh,
Nabila Aicha Benharkat |
Abstract: |
Data warehousing is an essential element of decision
support. In order to supply a decisional database, meta-data is needed to
enable the communication between various function areas of the warehouse
and, an ETL tool (Extraction, Transformation, and Load) is needed to
define the warehousing process. The developers use a mapping guideline to
specify the ETL tool with the mapping expression of each attribute. In
this paper, we will define a model covering different types of mapping
expressions. We will use this model to create an active ETL tool. In our
approach, we use queries to achieve the warehousing process. SQL queries
will be used to represent the mapping between the source and the target
data. Thus, we allow DBMS to play an expanded role as a data
transformation engine as well as a data store. This approach enables a
complete interaction between mapping meta-data and the warehousing tool.
In addition, this paper investigates the efficiency of a Query-based data
warehousing tool. It describes a query generator for reusable and more
efficient data warehouse (DW) processing. Besides exposing the advantages
of this approach, this paper shows a case study based on real scale
commercial data to verify our tool features. |
|
Title: |
EXTENDING
TREE AUTOMATA TO MODEL XML VALIDATION UNDER ELEMENT AND ATTRIBUTE
CONSTRAINTS |
Author(s): |
D. Laurent,
D. Duarte, B. Bouchou, Mírian Halfeld Ferrari Alves
|
Abstract: |
Algorithms for validation play a crucial role in the use
of XML as the standard for interchanging data among heterogeneous
databases on the Web. Although much effort has been made for formalizing
the treatment of elements, attributes have been neglected. This paper
presents a validation model for XML documents that takes into account the
element and attribute constraints imposed by a given DTD. Our main
contribution is the introduction of a new formalism to deal with both
kinds of constraints. We deem that our formalism has interesting
characteristics: it allows dealing with finite trees with attributes and
elements, it is simple, since it is just an extension of regular tree
automata and it allows the construction of a deterministic automaton
having the same expression power as that of a DTD. Moreover, our formalism
can be implemented easily, giving rise to an efficient validation
method. |
|
Title: |
AN
ARCHITECTURAL FRAMEWORK FOR WEB APPLICATIONS |
Author(s): |
Stefan Jablonski,
Ilia Petrov, Christian Meiler |
Abstract: |
The Web has an ever-changing technological landscape.
Standards and techniques utilized for the implementation of Web
Applications as well as the platforms on which they are deployed are
subject to constant changes. In order to develop Web-Applications in a
structured and systematic manner regardless of this dynamics a clear
development methodology considering the flexibility and extensibility as
central goals is needed. This paper proposes a definition of the term
Web-Application and a conceptual architectural framework for
Web-Applications. Besides this some important characteristics of such a
framework will be investigated and a construction methodology will be
presented. |
|
Title: |
A
MINIMAL COVER FOR DECLARATIVE EXPRESSIONS |
Author(s): |
Margaret Miró,
Josep Miró |
Abstract: |
Descriptive knowledge about a multivalued data table or
Information System can be expressed in declarative form by means of a
binary Boolean based language. This paper presents a contribution to the
study of an arbitrary multivalued Information System by introducing a
non-binary array algebra that allows the treatment of multiple valued data
tables with systematic algebraic techniques. An Information System can be
described by severeal distinct, but equivalent, array expressions. Among
these, the all-prime-ar expression is singled out. The all-prime-ar
expression is a unique expression, although it is not necessarily minimum
in the number of prime-ars. Finally, a completely intensional technique
that determines a cover, a minimal prime-ar expression is presented. |
|
Title: |
INTEGRATING
DISTRIBUTED HETEROGENOUS DATABASES AND DISTRIBUTED GRID
COMPUTING |
Author(s): |
Tapio Niemi
|
Abstract: |
The aim of this paper is to present a middleware that
combines the flexibility of distributed heterogeneous databases with the
performance of local data access. The middleware will support both XML and
relational database paradigms and applies Grid security techniques. The
computing and database access facilities are implemented using Grid and
Java technologies. In our system, data can be accessed in the same way
independently of its location, storage system, and even storage format.
The system will also support distributed queries and transaction
management over heterogeneous databases. Our system can be utilised in
many applications related to storing, retrieval, and analysis of
information. Because of advanced security components, e-commerce is a
potentical application area, too. The implementation is based on the
principle that each node on the computing grid containing a database
contains also a Java agent. The database requests are first sent to the
agent which takes care of security tasks, possibly does some preprocessing
or translation to the query, and finally transmits the request to the
database system. The agents also take care of distributed transaction
management. The system does not have a special master but each agent has a
capability to handle distributed transactions by sending requests to other
agents. |
|
Title: |
FORMALIZING
TYPES WITH ULTIMATE CLOSURE FOR MIDDLEWARE TOOLS IN INFORMATION SYSTEMS
ENGINEERING |
Author(s): |
Brian
Nicholas Rossiter, David Nelson, Michael A Heather |
Abstract: |
A definition of types in an information system is given
from real-world abstractions through constructs employed for data and
function descriptions through data schema and definitions to the physical
data values held on disk. This four-level architecture of types is
considered from the real-world interpretation of the types and the
level-pairs between types: in terms of mappings between the types at each
level and formally in terms of a composition of functors, adjoints and
natural transformations across the various types. The theory suggests that
four levels are sufficient to provide ultimate closure for computational
types to construct information systems. The Godement calculus can be used
to compose mappings at different levels. Examples of information systems
are examined in terms of the four-level architecture including the
Information Resource Dictionary Standard (IRDS), the Grid, the semantic
web using data exchange languages such as XML/RDF and MOF/MDA with meta
objects in a model-driven architecture. Only the IRDS and MOF are
genuinely based on four levels. IRDS appears to be the more open at the
top level but does not support two-way mappings. |
|
Title: |
UPDATING
GIS DATA USING PERSONAL DIGITAL ASSISTANTS |
Author(s): |
Alexandre Sousa,
João Lopes, Henrique Silva |
Abstract: |
Geo-referenced data is acquired and then postponed into an
existing GIS.With the advent of mobile computing devices, namelly personal
Digital Assistants (PDAs), this integration task is sure to be avoided. We
tried to extend a PDA GIS displaying system (Mordomo) in order to allow
metadata update. This way, the task of updating geo-referenced data could
be done on-site, in the palce were the data is to be acquired, and the
integration in the GIS could be done automatically. In order to have the
system coping with many different applications, we decided to provide a
transformer from and to GML, the OGC proposed standard. |
|
Title: |
CONSTRAINTS
AND MULTIDIMENSIONAL DATABASES |
Author(s): |
Franck Ravat,
Faiza Ghozzi, Gilles Zurfluh, Olivier Teste |
Abstract: |
The model we define organises data in a constellation of
facts and dimensions with multiple hierarchies. In order to insure data
consistence and reliable data manipulation, we extend this constellation
model by intra- and inter-dimension constraints. The intra-dimension
constraints allow the definition of exclusions and inclusions between
hierarchies of same dimension. The inter-dimension constraints are related
to hierarchies of different dimensions. Also, we study effects of these
constraints on multidimensional operations. In order to validate the
solutions we provide, we depict integration of these constraints within
GEDOOH prototype. |
|
Title: |
CQSERVER:
AN EXAMPLE OF APPLYING A DISTRIBUTED OBJECT INFRASTRUCTURE FOR
HETEROGENEOUS ENTERPRISE COMPUTATION OF CONTINUAL QUERIES |
Author(s): |
Jennifer Leopold,
Tyler Palmer |
Abstract: |
The revolution in computing brought about by the Internet
is changing the nature of computing from a personalized computing
environment to a ubiquitous computing environment in which both data and
computational resources are network-distributed. Client-server
communications protocols permit parallel ad hoc queries of
frequently-updated databases, but they do not provide the functionality to
automatically perform continual queries to track changes in those data
sources through time. The lack of persistence of the state of data
resources requires users to repeatedly query databases and manually
compare the results of searches through time. To date, continual query
systems have lacked both external and internal scalability. Herein we
describe CQServer, a scalable, platform- and implementation-independent
system that uses a distributed object infrastructure for heterogeneous
enterprise computation of both content- and time-based continual
queries. |
|
Title: |
AN
INTEGRATED APPROACH FOR EXTRACTION OF OBJECTS FROM XML AND TRANSFORMATION
TO HETEROGENEOUS OBJECT ORIENTED DATABASES |
Author(s): |
Uzair Ahmad
|
Abstract: |
XML is widely used by the database management systems for
data representation and transportation of data. In this paper we focus on
the integration of latest W3C XML Schema specifications and hash maps for
performance and efficient retrieval of objects from XML documents and
transforming them into heterogeneous object oriented databases. Remaking
of XML-ized databases from sizeable XML document faces a limitation of
memory resources. Besides XML Schema incorporation, this research endeavor
also provides new options for the handling of large XML-ized database
document size. |
|
Title: |
CONSTRUCTING
FEDERATED ENTERPRISE SCHEMAS WITH CONCEPTUALIZED DATA WRAPPERS |
Author(s): |
Thiran Philippe
|
Abstract: |
The ubiquity of the Internet gives organizations the
possibility to form virtual alliances. This not only implies that the
business transactions must be linked, but also requires that business
applications are integrated to support them. In this paper, we present an
integral approach for blending modern business data requirements with
existing legacy data resources that offers techniques at both the
conceptual and the implementation level. Therefore, we depend on
access/integration reverse engineering technologies, including conceptual
wrappers and gateways. The reverse engineering strategy allows modernized
business data systems to co-exists with legacy repository systems. In
particular, the methodology aims at constructing a conceptual Federation
Enterprise Schema (FES) for supporting the complicated task of data
wrapper integration throughout the development cycle: from specification
down to the actual implementation. The FES model plays a pivot role in the
creation of the virtual alliances by representing an unified data view for
all participants. This unified data model serves as the foundation for the
actual integration of the wrapped legacy data systems, possibly with
modernized data systems. Thus, in contrast to other available approaches,
the FES is not developed from scratch, but instead, composed out of
pre-existing legacy wrappers. This methodology is validate by an
experimental prototype that is still under development and sits on to of
DB-MAIN. |
|
Title: |
HEURISTIC
METHOD FOR A REAL WORLD TRANSPORT |
Author(s): |
Meriam Kefi,
Khaled Ghédira |
Abstract: |
Within the framework of an international sporting
manifestation which involved 23 countries in 23 disciplines and gathered
not less than 15000 participators (VIP, disciplines and gathered not less
than 15000 participators (VIP, officials, athletes, judges, referees,
doctors, journalists, technicians, voluntaries), the central committee of
organization was obliged to automatize its activities and to distribute
them among 16 committees in order to guarantee especially the best
conditions of organization and safety. Thus, we were called to elaborate a
prototype dealing with the transport activity. |
|
Title: |
METADATA-DRIVEN
MODELING OF TOOL-INDEPENDENT DATA TRANSFORMATIONS |
Author(s): |
Heiko Tapken,
Arne Harren |
Abstract: |
Due to their analytically oriented and cleansed
integration of data from several operational and external data sources,
data warehouse systems serve as a substantial technical foundation for
decision support. Within the scope of our research we are seeking novel
solutions for handling data acquisition within such environments. In this
paper we present some aspects of our approach to data acquisition. We
briefly sketch our framework and outline the underlying process model. We
introduce an approach for toolindependent modeling of data
transformations at a logical design layer in detail including a partial
description of our metamodel and an introduction to the
transformation language TL2. |
|
Title: |
EXTENDED
PROTECTED DATABASES: A PRACTICAL IMPLEMENTATION |
Author(s): |
Steve
Barker, Paul Douglas |
Abstract: |
We show how logic programs may be used to protect secure
databases that are accessed via a web interface from the unauthorized
retrieval of positive and negative information, and from unauthorized
insert and delete requests. To achieve this protection, we use a deductive
database expressed in a form that is guaranteed to permit only authorized
access requests to be performed. The protection of the positive
information that may be retrieved from a database and the information that
may be inserted are treated in a uniform way as is the protection of the
negative information in the database and the information that may be
deleted. The approach we describe has a number of attractive technical
results associated with it, it enables access control information to be
seamlessly incorporated into a deductive database, and it enables security
information to be used to help to optimize the evaluation of access
requests. These properties are particularly useful in the context of a
database which is accessed via the internet, since this form of access
requires a practical access control method which is both powerful and
flexible. We describe our implementation of a web-server front-end to a
deductive database which incorporates our access authorization proposals.
|
|
Title: |
USABILITY
AND WEB SITE EVALUATION: QUALITY MODELS AND USER TESTING
EVALUATIONS |
Author(s): |
Francisco Montero
|
Abstract: |
As the Internet expands, and the amount of information
that we can find on the web grows along with it, the usability of the
pages gets more important. Many of the sites still get quite low
evaluations from participants when it came to certain aspects of
usability. This paper proposes a set of quantitative and qualitative
metrics under a usability-centred quality model and an usability testing
experiment where this model can be validated. But finally, usability tests
may do a great job by showing what is not working in a design, but do not
get caught in the trap of asking testers to suggest design improvements
because creating Web sites is easy, however, creating sites that truly
meet the needs and expectations of the wide range of online users is quite
another story. |
|
Title: |
COPLA:
A PLATFORM FOR EAGER AND LAZY REPLICATION IN NETWORED DATABASES |
Author(s): |
Francesc
Daniel Muñoz-Escoí, Jose Manuel Bernabeu-Auban,
Luis Irún-Briz, Hendrik Decker |
Abstract: |
COPLA is a software tool that provides an object-oriented
view of a network of replicated relational databases. It supports a range
of consistency protocols, each of which supports different consistency
modes. The resulting scenario is a distributed environment where
applications may start multiple database sessions, which may use different
consistency modes, according to their needs. This paper describes the
COPLA platform, its architecture, its support for database replication and
one of the consistency algorithms that have been implemented on it. A
system of this kind may be used in the development of applications for
companies that have several branch offices, such as banks, hypermarkets,
etc. In such settings, several applications typically use on-site
generated data in local branches, while other applications also use
information generated in other branches and offices. The services provided
by COPLA enable an efficient catering for both local and non-local data
querying and processing. |
|
Title: |
ONTOLOGIES:
SOLVING SEMANTIC HETEROGENEITY IN A FEDERATED SPATIAL DATABASE
SYSTEM |
Author(s): |
Villie
Morocho Zurita, Lluis Pérez Vidal |
Abstract: |
Information integration has been an important area of
research for many years, and the problem of integration of geographic data
has recently emerged. This paper presents an approach based on the use of
Ontologies for solving the problem of semantic heterogeneity in the
process of the construction of a \emph{Federated Schema} in the framework
of geographic data. We make use of a standard technology (OMT-G based UML,
XMI based XML, GML from OpenGIS). |
|
Title: |
OLAPWARE:
ONLINE ANALYTICAL PROCESSING MIDDLEWARE |
Author(s): |
Fernando
Souza, Valeria Times, Robson Fidalgo |
Abstract: |
This paper presents OLAPWare, which is a Java middleware
for providing OLAP Services complaint with the OLE DB for OLAP standard.
OLE DB for OLAP is an industrial standard for allowing interoperability
among OLAP tools. However, it is limited to the Windows/COM platform, even
when a Java API is used. OLAPWare aims to overcome this limitation by
allowing its Java clients to query the objects of a dimensional data cube
without depending on the chosen implementation platform. In addition,
OLAPWare can be used as a server to other applications requiring online
analytical processing, such as Geographical Information System and Data
Mining. |
|
Title: |
A
METHODOLOGICAL FRAMEWORK FOR BUSINESS MODELLING |
Author(s): |
Judith Barrios
Albornoz, Jonás Montilva Calderón |
Abstract: |
Globalisation phenomenon has created a very competitive
environment for modern business organisations. In order to survive and
continue being competitive in that environment, an organisation has to
adapt to it quickly with a minimal negative impact over its current ways
of working and organising. A business model contains the knowledge needed
not only to support managers’ decisions concerned with change and
adaptation, but to ensure the opportunity and relevance of the information
produced by the automated systems supporting them. The purpose of this
paper is to present a methodological framework for business modelling.
This framework allows its users to represent organisation’s elements from
different perspectives taking into account their relationships. A business
model is presented as a set of three interrelated models – the Business
Goals model, the Business Processes model, and the Information Systems
model. The main contribution of our paper is to make visible and explicit
the relationships among the three levels: goals, business processes and
information systems. These relationships are commonly hidden or implicit
in most business modelling methods. Our proposition has proven its
usefulness as a strategic management tool in two studies cases. |
|
Title: |
MODELLING
DATA WAREHOUSING MANAGEMENT IN ENTERPRISE PERFORMANCE |
Author(s): |
Alberto Carneiro
|
Abstract: |
This paper intends to contribute to a better understanding
of the process through which data warehouse (DW), information technology,
other technical tools, and organisation actors can contribute to
enterprises’ effectiveness facing the challenges that are continuously
happening in the new information technology domain. Firstly, it presents
some researchers’ opinions about the role of Data Warehousing Management
(DWM) in the decision-making process. Consequently, it sustains that a set
of variables influences the relationship between decision effectiveness
and a valuable utilisation of DWM’s potential skills. A conceptual model
for the optimisation of enterprises’ performance as a function of DWM is
suggested. |
|
Title: |
DIA:
DATA INTEGRATION USING AGENTS |
Author(s): |
Ulrich Schiel,
Philip Medcraft, Cláudio Baptista |
Abstract: |
The classic problem of information integration has been
addressed for a long time. The Semantic Web project is aiming to define an
infrastructure that enables machine understanding. This is a vision that
tackles the problem of semantic heterogeneity by using ontologies for
information sharing. Agents have an important role in this infrastructure.
In this paper we present a new solution, known as DIA (Data Integration
using Agents), for semantic integration using mobile agents and
ontologies. |
|
Title: |
DATA
WAREHOUSE REFRESHMENT MAINTAINING TEMPORAL CONSISTENCY |
Author(s): |
Araque Francisco
|
Abstract: |
The refreshment of a data warehouse is an important
process which determines the effective usability of the data collected and
aggregated from the sources. Indeed, the quality of data provided to the
decision makers depends on the capability of the data warehouse system to
convey in a reasonable time, from the sources to the data marts, the
changes made at the data sources. We present our current work related to:
maintaining the temporal consistency between the data extracted from
semi-structured information sources and the data loaded in the data
warehouse according to temporal data warehouse designer requirements; and
monitoring the web in accordance with the temporal requirements of the
data warehouse designer. We use different approaches to maintain temporal
coherency of data gathered from web sources; and wrappers extended with
temporal characteristics to keep temporal consistency. Besides we present,
an integrated database architecture in which data warehouses are part of
the database, extended in order to express temporal concepts. |
|
Title: |
PATTERNS
AND COMPONENTS TO CAPITALIZE AND REUSE A COOPERATIVE INFORMATION SYSTEM
ARCHITECTURE |
Author(s): |
Magali SEGURAN,
Vincent COUTURIER |
Abstract: |
The growth and variety of distributed information sources
imply a need to exchange and/or to share information extracted from
various and heterogeneous databases. Cooperation of legacy information
systems requires advanced architectures able to solve conflicts coming
from data heterogeneity: technical, syntactic, structural and semantic
conflicts. So, we propose a multi-level architecture based on
object-oriented and distributed artificial intelligence to solve these
conflicts. Thanks to cooperation patterns and components we propose to
capitalize knowledge from this architecture to reuse it to develop new
cooperative applications. |
|
Title: |
SOFTWARE
PROCESS IMPROVEMENT DEFINED |
Author(s): |
Ivan Aaen
|
Abstract: |
This paper argues in favor of the development of
explanatory theory on software process improvement. The last one or two
decades commitment to prescriptive approaches in software process
improvement theory may contribute to the emergence of a gulf dividing
theorists and practitioners. It is proposed that this divide be met by the
development of theory evaluating prescriptive approaches and informing
practice with a focus on the software process policymaking and process
control aspects of improvement efforts. |
|
Title: |
SANGAM:
A FRAMEWORK FOR MODELING HETEROGENEOUS DATABASE TRANSFORMATIONS |
Author(s): |
Kajal Claypool
|
Abstract: |
A broad spectrum of data is available on-line in distinct
heterogeneous sources, and stored under different formats. As the number
of systems that utilize the heterogeneous data sources grows, the
importance of data translation and conversion mechanisms increases
greatly. The goal of our work is a to design a framework that simplifies
the task of translation specification and execution. Translation
specification between the source and the target schema can be accomplished
via (1) the discovery of matches between the source and the target
schemata; (2) the application of a pre-defined translation templates; or
(3) via manual user specification. In this paper we present a {\em
flexible}, {\em extensible} and {\em re-usable} translation modeling
framework wherein users can (1) explicitly model the translations between
schemas; (2) compose translations from an existing library of modeled
translation patterns; (3) choose from a library of translation operators;
(4) generate translation models based on a match process; (5) edit such
translation models; and (5) for all of these translation models, choose
automated execution strategies that transform the source schema and data
to the desired target schema and data. In this paper, we present the
system architecture for such a translation modeling framework. |
|
Title: |
A
COMPONENT-BASED METHOD FOR DEVELOPING WEB APPLICATIONS |
Author(s): |
Jonas Montilva,
Judith Barrios |
Abstract: |
We describe, in this paper, a component-based software
engineering method for helping development teams to plan, organize,
control, and develop web applications. The method is described in terms of
three methodological elements: a product model that captures the
architectural characteristics of web applications, a team model that
describes the different roles to be played by the members of a team during
the development of web applications, and a process model that integrates
the managerial and technical activities that are required to develop
componentized web applications of high quality. The main features of the
model are its component-based approach that helps reduce costs and
development time; its ability to integrate managerial and development
processes into a unique process model; and its emphasis on business
modelling as a way of gaining a better understanding of the application
domain objectives, functions and requirements. |
|
Title: |
ENTERPRISE
MIDDLEWARE FOR SCIENTIFIC DATA |
Author(s): |
Judi Thomson
|
Abstract: |
We describe an enterprise middleware system that
integrates, from a user’s perspective, data located on disparate data
storage devices without imposing additional requirements upon those
storage mechanisms. The system provides advanced search capabilities by
exploiting a repository of metadata that describes the integrated data.
This search mechanism integrates information from a collection of XML
documents with diverse schema. Users construct queries using familiar
search terms, and the enterprise system uses domain representations and
vocabulary mappings to translate the user’s query, expanding the search to
include other potentially relevant data. The enterprise architecture
allows flexibility with respect to domain dependent processing of user
data and metadata. |
|
Title: |
RECURSIVE
PATCHING - AN EFFICIENT TECHNIQUE FOR MULTICAST VIDEO STREAMING |
Author(s): |
Jack
Y. B. Lee, Y. W. Wong |
Abstract: |
Patching and transition patching are two techniques
proposed to build efficient video-on-demand (VoD) systems. Patching works
by allowing a client to playback video data from a patching stream while
caching video data from another multicast video stream for later playback.
The patching stream can be released once video playback reaches the point
where the cached data begins, and playback continues via the cache and the
shared multicast channel for the rest of the session. Transition patching
takes this patching technique one step further by allowing a new client to
cache video data not only from a full-length multicast channel, but also
from a nearby in-progress patching channel as well to further reduce
resource consumption. This study further generalizes these patching
techniques into a recursive patching scheme where a new client can cache
video data recursively from multiple patching streams to further reduce
resource consumption. This recursive patching scheme unifies the existing
patching schemes as special cases. Simulation results show that it can
achieve significant reductions (e.g. 60%~80%) in startup latency at the
same load and with the same system resources. |
|
Title: |
DESIGN
OF A LARGE SCALE DATA STREAM RECORDER |
Author(s): |
Roger Zimmermann
|
Abstract: |
Presently, digital continuous media (CM) are well
established as an integral part of many applications. In recent years, a
considerable amount of research has focused on the efficient retrieval of
such media. Scant attention has been paid to servers that can record such
streams in real time. However, more and more devices produce direct
digital output streams. Hence, the need arises to capture and store these
streams with an efficient data stream recorder that can handle both
recording and playback of many streams simultaneously and provide a
central repository for all data. We propose a design for a large scale
data stream recorder. Our goal was to introduce a unified architecture
that integrates multi-stream recording and retrieval in a coherent manner.
The discussion raises practical issues such as support for multizone disk
drives, variable bit rate media, and disk drives that have a different
write than read bandwidth. We provide initial solutions for some issues
while others will need to be investigated further. |
|
Title: |
DATA
CLEANSING FOR FISCAL SERVICES: THE TAVIANO PROJECT |
Author(s): |
Antonella Longo,
Mario Bochicchio |
Abstract: |
Fiscal incomes are vital for Governments, both for central
and local agencies, therefore data quality policies and on-line fiscal
services will play a key role in the e-Government scenario. In the opinion
of authors, in fact, no matter how well an Agency implements innovative
services, poor data quality can destroy its utility and cost real money.
The original contribution of this paper is about the Taviano project, a
real experience of data quality management for on-line fiscal services in
Italy: as first, we introduce the architecture of the system used to clean
fiscal data. As second, we show how appropriate data analysis procedures
can reduce the need for clerical review (manual inspection implies higher
costs) of fiscal data. The proposed system is based on an innovative
variant of the well known LCS (Longest Common Subsequence) approximate
string matching algorithm. |
|
Title: |
ASSUMING
A ROADMAP STRATEGY FOR E-BUSINESS |
Author(s): |
Luis Borges
Gouveia, Feliz Ribeiro Gouveia |
Abstract: |
Current developments towards the adoption of e-business
practices within existing organisations show that a number of requirements
must be met before inside and outside satisfaction, integration and
success is achieved. In order to provide a clear and straightforward
adoption of e-business practices, a roadmap strategy is proposed based on
the accomplishment of a number of steps to provide a more stable
environment for conducting an electronic-based business. The paper
proposes a roadmap strategy based on organisation needs to have an inside
experience on topics such as technology management, information systems,
information management and knowledge management to approach e-business
practices. The discussion is made using the e-readiness concept. For this
discussion, electronic business is considered as the conduction of
business using mostly an electronic support to interact both between the
organisation and people involved: suppliers, customers and partners and
the organisation’s professionals. |
|
Title: |
TRY
AND PATCH: AN APPROACH TO IMPROVE THE TRUST IN SOFTWARE
COMPONENTS |
Author(s): |
Philippe Mauran,
Gerard Padiou, Pham Loc |
Abstract: |
We investigate the adaptability of components to their
client use. This goal implies to specify the user behavior to control the
effective use of components. Furthermore, this control may be completed by
carrying out the dynamic adaptation of components to increase the provided
service. Through an illustrative sample, we first define the problem of
use safety. Then, we propose an approach to insure this safety thanks to
the notion of profile. Lastly, a pattern is proposed for the
implementation of a such a safety service. |
|
Title: |
BUSINESS
MODELLING WITH UML: IMPLEMENTING CUSTOMER ORIENTED IT SYSTEMS |
Author(s): |
Ashok Ranchhod,
Calin Gurau, Pauline Wilcox |
Abstract: |
The World Wide Web has allowed companies to reach
customers in markets which were previously inaccessible, and to compete
efficiently with the traditional store based retailers (de Kare-Silver,
1998). However, the low barriers to entry, the size of the market and the
relatively low costs of on-line business activities have created a
situation of intense competition. The only answer to this situation is to
build a strong brand name and to obtain the customers' long-term loyalty
(Novo, 2001a). The Internet empowers the customer, by offering
accessibility and ease of communication to previously inaccessible
markets(globally) (Chaston, 2001). The Internet user has the opportunity
to switch suppliers with several mouse clicks, to compare price and
products on a worldwide basis and to select without external pressure the
best available offer. The classical offer of low price/high quality
product does not work properly on the Internet because the same offer may
be available to hundreds of other on-line retailers (Wundermann, 2001).
One of the main ways in which on line retailers can create competitive
advantage is by offering customer-firm satisfaction (by developing
customer relatonship strategies), in addition to product-related
satisfaction. The adoption of a customer-oriented strategy is referred to
as Customer Relationship Management (CRM). In the on-line retailing
environment the introduction and maintenance of CRM requires a complex
process of planning, analysis, strategy design and implementation. This
paper discusses the importance of the business modelling to support this
process in the digital retailing arena and advocates the use of Unified
Modelling Language (UML) as a standard modelling language to support
business modelling. |
|
Title: |
AN
ONTOLOGY-BASED APROACH FOR EXCHANGING DATA BETWEEN HETEROGENEOUS DATABASE
SYSTEMS |
Author(s): |
Yamine AIT-AMEUR,
Mourad Mimoune, Guy PIERRA |
Abstract: |
This paper presents an approach which allows data exchange
between heterogonous databases. It targets at simultaneously semantic and
structural heterogeneity. From the semantic point of view, this approach
proposes an ontology based approach. On the one hand this ontology can be
referenced by universal identifiers and acceded by queries; on the other
hand, it can be exchanged between heterogonous databases systems. From the
structural point of view, this approach is based on the use of a generic
meta-schema, formalised in the EXPRESS language, and allowing the exchange
of any instance of any database schema. Exchanged instances reference, as
much as needed, the global unique identifiers defined by the ontology.
However, the conversion of exchange files to the various target systems
can be achieved in a generic manner (e.g. independently of the particular
exchanged model). The interest of the EXPRESS language to achieve directly
such a program is presented as well. |
|
Title: |
AN
XML VIEW OF THE "WORLD" |
Author(s): |
Leonardo Mariani,
Emanuela Merelli, Ezio Bartocci |
Abstract: |
The paper presents "Any Input XML Output" (AIXO), a
general and flexible software architecture for wrappers. The architecture
has been designed to present data sources as collections of XML documents.
The use of XSLT as extraction language permits extensive reuse of
standards, tools and knowledge. A prototype developed in Java has been
effectively proven in several case studies. The tool has also been
successfully integrated as a wrapper service into BioAgent, a mobile agent
middleware specialized for use in the molecular biology domain. |
|
Title: |
EXPLOITATION
OF THE DATAWAREHOUSE AT THE SERVICE OF THE HOTELS: A PROPOSAL OF CLIENT
ORIENTATION |
Author(s): |
Rosario Berriel,
Antonia Gil, Isabel Sánchez, Zenona González |
Abstract: |
The tourist business´ are seeing their work conditioned by
the changes and transformations that are derived from the global
environment in which they operate. Facing this new situation, it is
necessary that they consider changes in the business methods and
substitute the focus of management by process with the management
orientated to the client. For this, they should focus on a strategy of
integration of the information which permits them to join the client to
the Value Chain of the Business. With this strategy, they could make use
of all the potential of tools like DataWarehouse and Customer Relationship
Management, to obtain knowledge of the clients and services adapted to the
demand. |
|
Title: |
ADAPTIVE
SOFTWARE QUALITY |
Author(s): |
Jeffrey Voas
|
Abstract: |
In this paper, I discuss what I believe is the grand
challenge facing the software quality research community: the ability to
accurately determine, in the very earliest stages of development, the
techniques that will be needed to achieve desired levels of non-functional
attributes such as: reliability, availability, fault tolerance,
testability, maintainability, performance, software safety, and software
security. I will further consider the associated technical and economic
tradeoffs that must be made in order to: (1) achieve these desired
qualities, and (2) to certify that these qualities will be exhibited when
the software is deployed. And I will also take into account the fact that
satisfying a particular level of each attribute requires specific cost
expenditures, some of these attributes conflict with each other, and when
the environment or usage profile of the software is modified, all
guarantees or claims of quality should be viewed suspiciously until
additional evidence is provided. |
|
Title: |
SOFTWARE
APPLICATION PACKAGES SELECTION: AN EVALUATION PROCESS BASED ON THE SPIRAL
MODEL |
Author(s): |
Claudine Toffolon,
Salem Dakhli |
Abstract: |
Cost overruns, late deliveries, poor quality, and users
resistance are examples of the seemingly intractable problems encountered
in the software development and maintenance activities, and related to the
“software crisis”. In particular, maintenance of existing software systems
results in visible and invisible application backlogs that create ill-will
between software users and software developers. To reduce the problems
related to application backlogs, two strategies have been proposed:
software productivity improvement and amount of work reduction. The use of
standard application packages permits implementing the second strategy.
Although software packages are available quickly and usually less
expensive then software developed in-house, the procurement of such
packages involves many risks. In this paper, we propose a tool evaluation
process based on the spiral model to cope with the software packages
selection. This process rests on the global software engineering model
elaborated by the authors in a previous work. It has been applied in a
French insurance company to select three categories of tools: a software
project management tool, a software development tool and a software
package to support the litigation department activity |
|
Title: |
A
FORMAL MODEL FOR A OBJECT-RELATIONAL DATABASES |
Author(s): |
Valéria Magalhães
Pequeno |
Abstract: |
This paper describes a new object-relational data model
that will be used for modeling the integrated view schema and the source
databases schemas of a data warehouse environment. The model distinguishes
between object classes and literal classes. Furthermore, it divides a
schema in structure schema and behaviour schema. The main contribution of
this work is to define a formal model for an object-relational data model
which is general enough to encompass the constructs of any object-oriented
data model and most value-oriented models. |
|
Title: |
QUEROM
: AN OBJECT-ORIENTED MODEL TO REWRITING QUERY USING VIEWS |
Author(s): |
Abdelhak Seriai
|
Abstract: |
We propose in this article an object-oriented approach to
rewriting queries using views. Our approach aims to mitigate certain
limitations of existing query rewriting approaches. Among these
limitations, the inconsideration of certain types of object-oriented
complex queries or the lack of uniformity of this approaches compared to
the object-oriented model. The proposed approach is based, on one hand, on
an object-oriented representation model of queries and, on other hand, on
the object-oriented classification mechanism to determine queries
containment. In fact, classes representing queries defining existing views
are organized in an heritance hierarchy. Then, classification in this
heritance hierarchy of a class representing a query is exploited to
generate possible rewritings of this query. |
|
Title: |
REPLICATION
MANAGEMENT IN DISTRIBUTED DATABASES |
Author(s): |
Dejan Damnjanovic,
Miodrag Stanic, Ivana Mijajlovic, Anastasija Kokanovic
|
Abstract: |
Data replication is basically defined as maintaining
copies of data. In a replicated database where copies of the same data are
stored on multiple sites, replication can provide faster data access and
fault tolerance. One of the main challenges in introducing replication is
maintaining consistency without affecting performance. Since sinhronous
replication tehniques violate system performance considerably, in a great
number of commercial database management systems, asinhronous replication
is implemented as a solution built into these systems. Data replication is
widely used in several application types which work with distributed
databases, such as data warehouses, mobile environments and large scale
systems. These systems, quite different by its nature, put different
requirements which impose various problems that replication tehnology has
to solve. In this paper, one replication framework is explained. The
framework is meant to present a basis for replication environment
configuration. Superposing over the solution offered by Oracle, a special
algorithm is developped. The algorithm aimes to solve the problem of data
updating on every site, with garanteed data consistency considering
avoided conflicts |
|
Title: |
A
DESIGN OF A DISTRIBUTED APPLICATION: UTILISING MICO IN A PROTOTYPE
OODB |
Author(s): |
Wei
Tse Chew |
Abstract: |
In distributed systems, objects are distributed in an
environment that utilises different hardware architecture, operation
system and programming languages. Communication between objects in a
heterogeneous distributed environment is accomplished via middleware.
Middleware software resides between the application and the operating
system thus hiding some underlying complexities of both the application
and operating system. The increase in the diversity of computer platforms
used in worldwide IT infrastructure has helped make middleware popular.
Today many organizations have used the Internet, which is a very large
distributed environment, to integrate the various different systems used
in their organizations. Hence there is a need for a standard specification
such as CORBA to describe the basic infrastructure required to support
distributed objects. The design process of a distributed application
consists of several steps which can be divided into three main groups,
which is the OODB section, the application that utilises the OODB and the
IDL that enables object to be transferred from one OODB to another. |
|
Title: |
DYNAMIC
SYSTEM OF MANAGEMENT AND BUSINESS PROCESSES |
Author(s): |
Arminda Guerra,
Eurico Ribeiro Lopes |
Abstract: |
Lots of valuable information hidden in industrial
environments is barely exploited, since we need abstract and high-level
information that is tailored to the user's needs (Staud et al, 1998). The
real value of information technology organization legacy systems consists
in the "accumulation of years of business rules, policies, expertise and
‘know-how’” embedded in the system. In many cases it may be necessary to
build and test a prototype to develop a good understanding of the system’s
needs and requirements (Jurison, 1999) (Christina, 2000). In this paper we
describe a system, which consolidates the database systems of the legacy
systems with the business process rules. Done this, information retrieval
will be easy by any business section independently of the legacy systems.
|
|
Area 2 - ARTIFICIAL INTELLIGENCE AND DECISION SUPPORT SYSTEMS
Title: |
THE
ESSENCE OF KNOWLEDGE MANAGEMENT |
Author(s): |
Marco Bettoni,
Sibylle Schneider |
Abstract: |
We contend in this presentation that more sustainable and
successful Knowledge Management (KM) solutions can be built by using the
principles of Knowledge Engineering (KE) to understand knowledge in a more
appropriate way. We will basically explore five aspects of practical
knowledge relevant for promoting the essential Human Factors (HF) involved
in KM tasks: the value and function of knowledge, the motor and mechanism
of knowledge, the two states and 3 conversions of individual knowledge,
the logic of experience (organisation of knowledge) and knowledge
processes (wheel of knowledge). We explain their consequences under the
form of five principles that we suggest could be used as leading criteria
for designing and evaluating KM solutions and systems in a new way more
appropriate for implementing successfully the old insight of the essential
role of people. |
|
Title: |
CONVENTIONAL
VERSUS INTERVAL CLUSTERING USING KOHONEN NETWORKS |
Author(s): |
Mofreh Hogo, Pawan Lingras, Miroslav Snorek |
Abstract: |
This paper provides a comparison between conventional and
interval set representations of clusters obtained using the Kohonen neural
networks. The interval set clustering is obtained using a modification of
the Kohonen algorithm based on the properties of rough sets. The paper
includes experimental results for a web usage mining application.
Clustering is one of the important functions in web usage mining. The
clusters and associations in web usage mining do not necessarily have
crisp boundaries. Researchers have studied the possibility of using fuzzy
sets in web mining clustering applications. Recent attempts have adapted
genetic algorithms, K-means clustering algorithm, and Kohonen neural
networks based on the properties of rough sets to obtain interval set
representation of clusters. The comparison between interval and
conventional clustering, provided in this paper, may be helpful in
understanding the usefulness of some of the non-conventional clustering
algorithms in certain data mining applications. |
|
Title: |
PARTIALLY
CONNECTED NEURAL NETWORKS FOR MAPPING PROBLEMS |
Author(s): |
Can Isik,
Sanggil Kang |
Abstract: |
In this paper, we use partially connected feedforward
neural networks (PCFNNs) for input-output mapping problems to avoid a
difficulty in determining epoch while fully connected feedforward neural
networks (FCFNNs) are being trained. PCFNNs can also, in some cases,
improve generalization. Our method can be applicable to real input-output
mapping problems such as blood pressure estimation and etc. |
|
Title: |
MAPPING
DESIGNS TO USER PERCEPTIONS USING A STRUCTURAL HMM: APPLICATION TO KANSEI
ENGINEERING |
Author(s): |
Jun Tan,
D. Bouchaffra |
Abstract: |
This paper presents a novel approach for mapping designs
to user perceptions. We show how this interaction can be expressed using
three classification techniques. We introduce a novel classifier called
"structural hidden Markov model" (SHMM) that enables to learn and predict
user perceptions. We have applied this approach to Kansei engineering in
order to map car external contours (shapes) to customer perceptions. The
accuracy ob-tained using the SHMM is 90%. This model has outperformed the
neural network and the k-nearest-neighbor classifiers. |
|
Title: |
IMPROVING
SELF-ORGANIZING FEATURE MAP (SOFM) TRAINING ALGORITHM USING K-MEANS
INITIALIZATION |
Author(s): |
Abdel-Badeeh Salem,
Mostafa Syiam, Ayad Fekry Ayad |
Abstract: |
Self-Organizing Feature map (SOFM) is a competitive neural
network in which neurons are organized in an l-dimensional lattice (grid)
representing the feature space. The principal goal of the SOFM is to
transform an incoming pattern of arbitrary dimension into a one- or two-
dimensional discrete map, and to perform this transformation adaptively in
a topologically ordered fashion. Usually, SOFM can be initialized using
random values for the weight vectors. This paper presents a different
approach for initializing SOFM. This approach depends on the K-means
algorithm as an initialization step for SOFM. The K-means algorithms is
used to select N 2 (the size of the feature map to be formed) cluster
centers from the data set. Then, depending on the interpattern distances,
the N 2 selected cluster centers are organized into an N x N array so as
to form the initial feature map. Later, the initial map will be fine-tuned
by the traditional SOFM algorithm. Two data sets are used to compare
between the proposed method and the traditional SOFM algorithm. The
comparison results indicated that: using the first data set, the proposed
method required 5,000 epochs to fine tune the map while the traditional
SOFM required 20,000 epochs (4 times faster). Using the second data set,
the traditional SOFM required 10,000 epochs while the proposed method
required only 1,000 epochs (10 times faster) |
|
Title: |
MODEL-BASED
NEURAL NETWORKS FOR BRAIN TUMOR DIAGNOSIS |
Author(s): |
A. Salem,
Safaa Amin, M. Tolba |
Abstract: |
This study aims to develop an intelligent neural network
based system to automatically detect and classify brain tumors from head
Magnetic Resonance Image (MRI) to help nonexperts doctors in diagnosing
Brain tumors. Three types of brain tumors have been investigated which are
acoustic neuroma tumor, which is a benign tumor occurring in the acoustic
canals, optical glioma which occurs in the optic nerve or in the area
connecting the two nerves and astrocytomas tumor. Two NN-based systems
were developed for brain tumor diagnosis. The first system uses the
Principal Component Analysis (PCA) for dimentionality reduction and
feature extraction that extract the global features of the MRI cases. The
second system uses manual and the expectation maximization segmentation
algorithm to extract the local features of the MRI cases.Then Multi-Layer
Perceptron (MLP) network is used for the classification of these features
that obtaind from the PCA and the segmentation. A comparision study is
made between the performance of MLP. Experimental results of real cases
shows that peak recognition rate of 100% is achieved using PCA and 96.7%
when applying the segmentation algorithm before the classification. |
|
Title: |
AGENTS
FOR HIGH-LEVEL PROCESS MANAGEMENT: THE RIGHT ACTIVITIES, PEOPLE AND
RESOURCES TO SATISFY PROCESS CONSTRAINT |
Author(s): |
John Debenham
|
Abstract: |
Multiagent systems are an established technology for
managing high-level business processes. High-level business processes are
considerably more complex to manage than production workflows. They are
opportunistic in nature whereas production workflows are routine. Each
stage in a high-level process usually has a well-defined sub-goal, but the
best way to achieve that sub-goal within value, time and cost constraints
may not be known for certain. To achieve each sub-goal, resources,
including human resources, must be found and brought together in an
appropriate way. Alternatives include face-to-face meetings, and email
exchanges. In a multiagent system for high-level process management each
player is assisted by a personal agent. The system manages goal-driven
sub-processes and manages the commitments that players make to each other.
These commitments will be to perform some task and to assume some level of
responsibility. The way in which the selection of tasks and the delegation
of responsibility is done attempts to reflect high-level corporate
principles and to ‘sit comfortably’ with the humans involved. Commitments
are derived through a process of inter-agent negotiation that considers
each individual’s constraints and performance statistics. The system has
been trialed on business process management in a university administrative
context. |
|
Title: |
A
COMPARISON OF AUSTRALIAN FINANCIAL SERVICE FAILURE MODELS:HYBRID NEURAL
NETWORKS, LOGIT MODELS AND DISCRIMINANT ANALYSIS |
Author(s): |
Juliana Yim,
Heather Mitchell |
Abstract: |
This study investigated whether two artificial neural
networks (ANNs), multilayer perceptron (MLP) and hybrid networks using
statistical and ANN approaches, can outperform traditional statistical
models for predicting Australian financial service failures one year prior
to the financial distress. The results suggest that hybrid neural networks
outperform all other models one and two years before failure. Therefore,
hybrid neural network model is a very promising tool for failure
prediction in terms of predictive accuracy. This supports the conclusion
that for researchers, policymakers and others interested in early warning
systems, hybrid networks would be useful. |
|
Title: |
THREE-DIMENSIONAL
OBJECT RECOGNITION USING SUPPORT VECTOR MACHINE NEURAL NETWORK BASED ON
MOMENT INVARIANT FEATURES |
Author(s): |
Doaa Hegazy,
Ashraf Ibrahim, Mohamed Said Abdel Wahaab, Sayed Fadel
|
Abstract: |
A novel scheme using a combination of moment invariants
and Support Vector Machine (SVM) network is proposed for recognition of
three-dimensional (3-D) objects from two-dimensional (2-D) views. The
moment invariants are used in the feature extraction process since they
are invariant to translation, rotation and scaling of objects. Support
Vector Machines (SVMs) have been recently proposed as a new technique for
pattern recognition. In the proposed scheme, SVM neural network, which
trained using the Kernel Adatron (KA) with Gaussian kernel, is used for
training (classification) and testing step. The proposed scheme is applied
to a database of 1440 different views for 20 complex 3-D objects and very
good results are achieved without adding noise to the test views. Using
noisy test data also yielded promising results. |
|
Title: |
A
QUALITY-OF-SERVICE-AWARE GENETIC ALGORITHM FOR THE SOURCE ROUTING IN
AD-HOC MOBILE NETWORKS |
Author(s): |
Said Ghoniemy,
Mohamed Hashem, Mohamed Hamdy |
Abstract: |
A QoS-aware delay-constrained unicast source routing
algorithm for ad-hoc networks based on a genetic algorithm is proposed in
this paper. The proposed algorithm is based on a new chromosomal encoding
which depends on the network links instead of nodes. Advantages of the
link-based encoding in the ad-hoc routing problem were studied. Promising
results have been obtained when the proposed algorithm was compared to
other routing algorithms. Results also showed that the proposed algorithm
shows a better performance for heavy QoS constraints the average delay
requirements and cost |
|
Title: |
SUPPORTING
STRATEGIC ALLIANCES THE SMART WAY |
Author(s): |
Iain Bitran,
Steffen Conn |
Abstract: |
The Network Economy forces managers to pursue
opportunities and engage competition through alliances and networks of
alliances. Managers and organisations must therefore nurture the skills
that successful alliance development and management require, and attain
the “partnering mindset” pertinent to this new industrial paradigm.
Studies indicate that alliance success remains an elusive aspiration for
the majority of organisations, with up to seventy percent failing to meet
their initial objectives. The SMART Project addresses this issue by
developing a systematic managerial method for strategic alliance formation
and management. This method provides the structure for a software-based
decision support system that includes extensive learning and support
materials for manager and business consultant training. Following a brief
introduction, this paper provides an overview of the concepts and issues
relating to strategic alliances and networks. Subsequently, the
requirements and functioning of the SMART System are described. Finally,
the future direction and validation strategy of the project are
relayed. |
|
Title: |
A
HYBRID APPROACH FOR HANDWRITTEN ARABIC CHARACTER RECOGNITION: COMBINING
SELF-ORGANIZING MAPS (SOMS) AND FUZZY RULES |
Author(s): |
E. Moneer,
Mohamed Hussien, Abdel-Badeeh Salem, Mostafa Syiam
|
Abstract: |
This paper presents a hybrid approach combining
self-organizing feature map (SOM) and fuzzy rules to develop an
intelligent system for handwritten Arabic character recognition. In the
learning phase, the SOM algorithm is used to produce prototypes which
together with the corresponding variances are used to determine fuzzy
regions and membership functions. Fuzzy rules are then generated by
learning from training characters. In the recognition phase, an input
character is classified by a fuzzy rule based classifier. An unknown
character is then re-classified by an SOM classifier. Experiments on a
database of 41,033 handwritten Arabic character (20,142 used for training
and 20,891 used for testing). The experimental results achieve a
classification rate 93.1%. |
|
Title: |
KNOWLEDGE
MANAGEMENT IN ENTERPRISES: A RESEARCH AGENDA |
Author(s): |
Konstantinos Karnezis,
Konstantinos Ergazakis |
Abstract: |
Knowledge Management is an emerging area, which is gaining
interest by both enterprises and academics. The effective implementation
of a KM strategy is considering as a “must” and as a precondition of
success for contemporary enterprises, as they enter the era of the
knowledge economy. However, the field of Knowledge Management has been
slow in formulating a universally accepted conceptual framework and
methodology, due to the many pending issues that have to be addressed.
This paper attempts to propose a novel taxonomy for Knowledge Management
research by co instantaneously presenting the current status with some
major themes of Knowledge Management research. The discussion presented on
these issues should be of value to researchers and practitioners. |
|
Title: |
AN
ALGORITHM FOR MINING MAXIMAL FREQUENT SETS BASED ON DOMINANCY OF
TRANSACTIONS |
Author(s): |
Srikumar Krishnamoorthy,
Bharat Bhasker |
Abstract: |
Several algorithms for mining maximal frequent sets have
been proposed in the recent past. These algorithms, mostly, follow the
bottom-up approach. In this paper, we present a top-down algorithm for
mining the maximal frequent sets. The proposed algorithm uses a concept of
dominancy factor of a transaction for limiting the search space. The
algorithm is especially efficient for longer patterns. We theoretically
model and compare the proposed algorithm with MaxMiner (an algorithm for
mining long patterns) and show it to be more efficient |
|
Title: |
THE
STRATEGIC AND OPERATIONAL ROLES OF MICROCOMPUTERS IN SMES: A PERCEPTUAL
GAP ANALYSIS |
Author(s): |
ZELEALEM TEMTIME
|
Abstract: |
Although strategic planning and information technology are
key concepts in management research, they have been widely used in
relation to only large firms. Only few studies attempted to examine the
perceptions of small and medium enterprises (hereafter, SMEs) about the
role of IT in strategy making. Moreover, these studies are of less
significance for developing countries as the definition and environment of
SMEs vary from developed to developing country. This article analyses the
strategic use of microcomputers and software packages in corporate
planning and decision-making in small and medium enterprises (hereafter,
SMEs). Data were collected from 44 SMEs from 3 cities in the Republic of
Botswana to study their perceptions about the use of computer-based
technology to solve managerial problems, and analysed using simple
descriptive statistics. The findings indicate that SMEs in Botswana
engaged in both strategic and operational planning activities. However,
microcomputers and software packages were used primarily for operational
and administrative tasks rather than for strategic planning. They perceive
that strategic planning is costly, time-consuming, and hence appropriate
for only large firms. The study also showed that firm size and strategic
orientation have direct and positive relation to the use of computer
technology for strategic decision making. The major implication of the
findings for future research has been identified and presented. |
|
Title: |
THE
USE OF NEUROFUZZY COMPUTABLE SYSTEM TO IDENTIFY PROMINENT BEHAVIOR
CHARACTERISTICS IN SUCCESSFUL ENTREPRENEURS |
Author(s): |
Rogério Bastos,
Angelita Ré, Lia Bastos |
Abstract: |
Ahead of small and medium companies there are individuals
responsible for the company’s process of creation and development. It is
of high importance to identify which characteristics and attributes that
contribute to determine the success of these entrepreneurs are. In the
present work, has been used a neurofuzzy computable system which permits
to identify prominent characteristics in individuals who got success in
their enterprises, considered so successful entrepreneurs. For that, a
research was taken among entrepreneurs of textile and furniture fields
from Santa Catarina State. |
|
Title: |
KNOWLEDGE
ACQUISITION THROUGH CASE-BASED ADAPTATION FOR HYDRAULIC POWER MACHINE
DESIGN |
Author(s): |
Chi-man VONG,
Yi-ping Li, Pak-kin WONG |
Abstract: |
Knowledge acquisition is the first but usually the most
important and difficult stage in building an intelligent decision-support
system. Existing intelligent systems for hydraulic system design use
production rules as its source of knowledge. However, this leads to
problems of knowledge acquisition and knowledge base maintenance. This
paper describes the application of CBR to hydraulic circuit design for
production machines, which helps acquiring knowledge and solving problems
by reusing this acquired knowledge (experience). A technique Case-Based
Adaptation (CBA) is implemented in the adaptation stage of CBR so that
adaptation becomes much easier. A prototype system has been developed to
verify the usefulness of CBR in hydraulic power machine design. |
|
Title: |
KNOWLEDGE
MANAGEMENT AND DATA CLASSIFICATION IN PELLUCID |
Author(s): |
Tung Dang,
Baltazar Frankovic |
Abstract: |
Abstract: The main aim of the Pellucid project is to
develop a platform based on the multi-agent technology for assisting
public employees in their organization. This paper deals with one of many
problems associated with building such a system. There is the problem of
classification and identification of required information for agent’s
performance. Pellucid agents use historical experience and information to
assist newly arriving employees, therefore searching for some specific
data from the database is a routine task that they have often to do. This
paper presents methods for encoding data and creating the database, so
that agents can have an easy access to the required information.
Furthermore, two methods applicable with every type of database for
classification and selection of historical information are
presented. |
|
Title: |
SCALING
UP INFORMATION UPDATES IN DISTRIBUTED CONDITION MONITORING |
Author(s): |
Sanguk Noh,
Paul Benninghoff |
Abstract: |
Monitoring complex conditions over multiple distributed,
autonomous information agents can be expensive and difficult to scale.
Information updates can lead to significant network traffic and processing
cost, and high update rates can quickly overwhelm a system. For many
applications, significant cost is incurred responding to changes at an
individual agent that do not result in a change to an overriding
condition. But often we can avoid much work of this sort by exploiting
application semantics. In particular, we can exploit constraints on
information change over time to avoid the expensive and frequent process
of checking for a condition that cannot yet be satisfied. We motivate this
issue and present a framework for exploiting the semantics of information
change in information agents. We partition monitored objects based on a
lower bound on the time until they can satisfy a complex condition, and
filter updates to them accordingly. We present and implement a simple
analytic model of the savings that accrue to our methods. Besides
significantly decreasing the workload and increasing the scalability of
distributed condition monitoring for many applications, our techniques can
appreciably improve the agents' response time between a condition
occurrence and its recognition. |
|
Title: |
A
WEB-BASED DECISION SUPPORT SYSTEM FOR TENDERING PROCESSES |
Author(s): |
Noor
Maizura Mohamad Noor, Brian Warboys, Nadia Papamichail
|
Abstract: |
A decision support system (DSS) is an interactive
computer-based system that helps decision makers utilise data and models
to solve complex and unstructured problems. Procurement is a decision
problem of paramount importance for any business. A critical and vital
procurement task is to select the best contractor during the tendering or
bidding process. This paper describes a Web-based DSS that aids decision
makers in choosing among competitive bids for building projects. The
system is based on a framework of a generic process approach and is
intended to be used as a general decision-making aid. The DSS is currently
being implemented as a research prototype in a process-support
environment. It coordinates the participants of tendering processes and
supports the submission, processing and evaluation of bids. A case study
is drawn from the construction business to demonstrate the applicability
of our approach. |
|
Title: |
ONE
APPROACH TO FUZZY EXPERT SYSTEMS CONSTRUCTION |
Author(s): |
Dmitry Vetrov,
Dmitry Kropotov |
Abstract: |
Some pattern recognition tasks contain expert information,
which can be expressed in the terms of linguistic rules. Theory of fuzzy
sets presents one of the most successive ways for using these rules.
However, in this case there appear two main problems of forming fuzzy sets
and generating fuzzy rules, which cannot be fully solved by expert in some
areas. These are two "weak points" which hold in the expansion of fuzzy
expert systems. The article below proposes one of possible solutions based
on the use of precedent information. |
|
Title: |
A
CAPABILITY MATURITY MODEL-BASED APPROACH TO THE MEASUREMENT OF SHARED
SITUATION AWARENESS |
Author(s): |
Edgar Bates
|
Abstract: |
Emerging technologies for decision aids offer the
potential for large volumes of data to be collected, processed, and
displayed without overloading users and has tremendous implications for
the ability of decision makers to approach total situation awareness and
achieving a dominant competitive advantage. In industry the measures of
effectiveness are clearly linked to performance in the marketplace, but in
the military measures of shared situational awareness generally lack the
analogous objective rigor. This paper, thus attempts to provide the
framework for assessing shared situational awareness using fundamental
system engineering and knowledge management paradigms. |
|
Title: |
THE
COMMUNIGRAM: MAKING COMMUNICATION VISIBLE FOR ENTERPRISE
MANAGEMENT |
Author(s): |
Piotr Lipinski,
Jerzy Korczak, Helwig Schmied, Kenneth Brown |
Abstract: |
The Communigram is a new methodological approach to
project and process management which illustrates the information flows in
the enterprise in a simple and intuitively comprehensible manner. It
complements currently existing information systems by providing a means to
plan organizational communication explic-itly such that the crucial
exchange of information may be suitably controlled. This considerably
improves the usefulness of information systems both in terms of
information transmission effectiveness and user ac-ceptance. In this
paper, the practical implementation of the Communigram in information
systems is de-scribed with some notes on technical details and on the
practical experience gained in its use. |
|
Title: |
THE
DESIGN AND IMPLEMENTATION OF IMPROVED INTELLIGENT ANSWERING
MODEL |
Author(s): |
Ruimin Shen,
Qun Su |
Abstract: |
Based on the analysis of the main technical problems in
the designs of the Intelligent Answering System, the traditional Answering
System Model and its working mechanism is provided. Based on the analysis
of the model, a Improved Intelligent Answering Model based on the data
generalization based on the patterns tree, association rule mining of
patterns, and the mergence and deletion of the rules based on the
knowledge tree is come up with and implemented. In the end, the
improvement of this model in intelligence is analyzed and proved with some
data in a experiment. |
|
Title: |
INTEGRATED
KNOWLEDGE BASED PROCESS IN MANUFACTURING ENVIRONMENT |
Author(s): |
jyoti K,
Dino Isa, Peter Blanchfield, V.P Kallimani |
Abstract: |
Abstract: Industries in Malaysia are facing the threat of
survival in this global competitive world. This factor is more evident in
small scale industries. They are unable to sustain due to the various
factors like expensive ,labour, Market fluctuations and the technology
additions. Hence to leverage the system there is a need of the structure
where in industry can expertise them selves by utilizing their own tacit
and explicit knowledge and for betterment and survival. This paper is
focused on the various factors in designing the knowledge platform in
manufacturing sector using the environments like J2EE, Artificial
Intelligence and prolog programming. Thus supporting the decisions taken
in the industry |
|
Title: |
ACT
E-SERVICE QUESTION ANSWERING SYSTEMS BASED ON FAQ CORPUS |
Author(s): |
Ben Chou,
Hou-Yi Lin, Yuei-Lin Chiang |
Abstract: |
World Wide Web (WWW) is a huge platform of information
interchange. Users can utilize search engine to search, interchange
information on the Internet. Nowadays, there are about 5 hundred millions
of web pages at least in the world. With information overloading
everywhere on the Internet, users are often swamped with keyword-based
search engine and waste much time on impertinent web pages because of the
keyword appearance in the pages. After several innovations of search
engine, search results are more and more precision and intelligent. In the
future, semantic processing and intelligent sifting and ranking
technologies are integrated into the third generation search engine. Thus,
it is useful for satisfying and closing to the needs users wanted. In this
research, we try to combine text mining, concept space, and some related
technologies to implement a search engine, which has an appropriate
capability of understanding natural language questions. And we will
demonstrate it with ACT e-Service. |
|
Title: |
TME:
AN XML-BASED INFORMATION EXTRACTION SYSTEM |
Author(s): |
Shixia Liu,
Liping Yang |
Abstract: |
Information extraction is a form of shallow text
processing that locates a specified set of relevant information in a
natural-language document. In this paper, a system—Template Match Engine
(TME) is developed to extract useful information from unlabelled texts.
The main feature of this system is that it describes the extraction task
by an XML template profile, which is more flexible than traditional
pattern match methods. The system first builds an initial template profile
by utilizing domain knowledge. Then the initial template profile is used
to extract information from electronic documents. This step produces some
feedback words by enlarging and analyzing the extracted information. Next,
this template profile is refined by the feedback words and concept
knowledge related to them. Finally, the refined profile is used to extract
specified information from electronic documents. The experiment results
show that TME system increases recall without loss of precision. |
|
Title: |
A
GENERAL KNOWLEDGE BASE FOR COMPARING DESCRIPTIONS OF KNOWLEDGE |
Author(s): |
Susanne Dagh,
Harald Kjellin |
Abstract: |
The complexity associated with managing knowledge bases
makes it necessary to use a simple syntax when formalising knowledge for a
knowledge base. If a large number of people contribute with descriptions
of objects to such a knowledge base and if it is necessary to make precise
comparisons between the objects of the knowledge base, then some important
requirements must be fulfilled; 1) It is necessary that all contributors
of knowledge descriptions perceive the knowledge in a similar way; 2) It
is crucial that the definitions in the descriptions are on the right level
of abstraction; 3) It must be easy for the contributors of knowledge
descriptions to create knowledge structures and also to remove them. We
propose principles for creating a general knowledge base that fulfils
these requirements. We constructed a prototype to test the principles. The
tests and inquiries showed that the prototype satisfies the requirements,
and thus our conclusion is that the proposed general knowledge base
facilitates comparisons of knowledge descriptions. |
|
Title: |
CONSTRAINT-BASED
CONTRACT NET PROTOCOL |
Author(s): |
Alexander Smirnov,
Nikolai Chilov, Tatiana Levashova, Michael Pashkin
|
Abstract: |
The paper describes and analyses a constraint-based
contract net protocol designed as a part of the being developed
KSNet-approach. This approach addresses the problem of knowledge logistics
and considers it as a problem of configuring a knowledge source network.
Utilizing intelligent agents is motivated by a distributed and scalable
nature of the problem. Made improvements to the contract net protocol
concern a formalism of agents’ knowledge representation and a scenario of
the agents’ interaction. For the agents’ knowledge representation and
manipulation a formalism of object-oriented constraint networks was
chosen. Modifications related to the interaction scenarios include
introduction of iterative negotiation, concurrent conformation of
proposals, extended set of available messages, additional role for agents
and agents’ ability to change their roles during scenarios. Examples of
the modifications are shown via UML diagrams. A short scenario at the end
of the paper illustrates advantages of the developed modifications. |
|
Title: |
SIMULATING
DATA ENVELOPMENT ANALYSIS USING NEURAL NETWORKS |
Author(s): |
Pedro Gouvêa
Coelho |
Abstract: |
This article studies the creation of efficiency
measurement structures of Decision-Making Units (DMUs) by using high-speed
optimisation modules, inspired in the idea of an unconventional Artificial
Neural Network (ANN) and numerical methods. In addition, the Linear
Programming Problem (LPP) inherent in the Data Envelopment Analysis (DEA)
methodology is transformed into an optimisation problem without
constraints, by using a pseudo-cost function, including a penalty term,
causing high cost every time one of the constraints is violated. The LPP
is converted into a differential equations system. A non-standard ANN
implements a numerical solution based on the gradient method. |
|
Title: |
SET-ORIENTED
INDEXES FOR DATA MINING QUERIES |
Author(s): |
Janusz Perek,
Zbyszko Krolikowski, Mikolaj Morzy |
Abstract: |
One of the most popular data mining methods is frequent
itemset and association rule discovery. Mined patterns are usually stored
in a relational database for future use. Analyzing discovered patterns
requires excessive subset search querying in large amount of database
tuples. Indexes available in relational database systems are not well
suited for this class of queries. In this paper we study the performance
of four different indexing techniques that aim at speeding up data mining
queries, particularly improving set inclusion queries in relational
databases. We investigate the performance of those indexes under varying
factors including the size of the database, the size of the query, the
selectivity of the query, etc. Our experiments show significant
improvements over traditional database access methods using standard B+
tree indexes. |
|
Title: |
USING
KNOWLEDGE ENGINEERING TOOL TO IDENTIFY THE SUBJECT OF A DOCUMENT -
RESEARCH RESULTS |
Author(s): |
Offer Drori
|
Abstract: |
Information databases today contain many millions of
electronic documents. Locating information on the Internet today is
problematic, due to the enormous number of documents it contains. Several
other studies have found that associating documents with a subject or list
of topics can improve locatability of information on the Internet [5] [6]
[7]. Effective cataloguing of information is performed manually, requiring
extensive resources. Consequently, most information is currently not
catalogued. This paper aims to present a software tool that automatically
locates the subject of a document and to show the results of a test
performed, using the software tool (TextAnalysis) specially developed for
this purpose |
|
Title: |
SUMMARIZING
MEETING MINUTES |
Author(s): |
Carla Lopo
|
Abstract: |
In this paper it is analyzed the problem of summarization,
and specifically the problem of summarization of meeting verbatim. In
order to solve it, it is proposed an approach that consists of structuring
the meeting data and complementary data related to the environment in
which the meeting is integrated. Then, the creation of possible summaries
is based in the identification of genre of summaries and SQL
queries. |
|
Title: |
ON
FAST LEARNING OF NEURAL NETWORKS USING BACK PROPAGATION |
Author(s): |
Kanad Keeni
|
Abstract: |
This study discusses the subject of training data
selection for neural networks using back propagation. We have made only
one assumption that there are no overlapping of training data belonging to
different classes, in other words the training data is
linearly/semi-linearly separable . Training data is analyzed and the data
that affect the learning process are selected based on the idea of
Critical points. The proposed method is applied to a classification
problem where the task is to recognize the characters A,C and B,D. The
experimental results show that in case of batch mode the proposed method
takes almost 1/7 of real and 1/10 of user training time required for
conventional method. On the other hand in case of online mode the proposed
method takes 1/3 of training epochs, 1/9 of real and 1/20 of user and 1/3
system time required for the conventional method. The classification rate
of training and testing data are the same as it is with the conventional
method. |
|
Title: |
A
PORTAL SYSTEM FOR PRODUCTION INFORMATION SERVICES |
Author(s): |
Yuan-Hung Chen,
Jyi-Shane Liu |
Abstract: |
Production data are usually voluminous, continuous, and
tedious. Human efforts to derive production information from raw data
often result in extra work loading, lagging, and errors. Undesirable
results may occur when related functional units are not integrated in
parallel with the same updated information. Therefore, successful
production information management must address two significant problems:
speed of information and effect of information. We propose a production
information portal (PIP) architecture to facilitate information derivation
efficiency and information utilization performance. The architecture is
developed by integrating concepts of data and information management,
event monitoring, configurable services, decision support, and information
portal. A rigorous system analysis and modelling process is conducted to
produce detailed specifications on functional modules, operation
procedures, and data/control flows. The utility of the architecture and
the prototype system was verified in a semiconductor fabrication domain
and was tested by actual users on real data from a world class
semiconductor company. |
|
Title: |
AN
EXPERT SYSTEM FOR PREVENTING AND CORRECTING BURDEN SLIPS, DROPS AND HANGS
IN A BLAST FURNACE |
Author(s): |
David Montes,
Raquel Blanco, Eugenia Diaz, Javier Tuya,
Faustino Obeso |
Abstract: |
This paper describes an expert system for preventing and
correcting burden slips, drops and hangs inside a blast furnace. The
system monitors and takes the decisions through the analysis and
evaluation of more than a hundred parameters considered as input
variables. The main difference between the system proposed here and a
classical diagnostic system is the coexistence of three different models
of behaviour: one based on a theoretical model of behaviour of
permeability, a second empirical model based on the considerations given
by the human experts, and a third model derived from the study of the real
behaviour observed in the furnace over time, obtained by means of the
study of historical files, using machine learning techniques. |
|
Title: |
PREDICTING
OF CUSTOMER DEFECTION IN ELECTRONIC COMMERCE:USING BACK-PROPAGATION NEURAL
NETWORKS |
Author(s): |
Ya-Yueh Shih
|
Abstract: |
Since the cost of retaining an existing customer is lower
than that of developing a new one, exploring potential customer defection
becomes an important issue in the fiercely competitive environment of
electronic commerce. Accordingly, this study used artificial neural
networks (ANNs) to predict customers’ repurchase intentions and thus avoid
defection based on a set of criteria of quality attributes satisfaction
and three beliefs in theory of planned behavior (TPB). The predicted
repurchase intentions found by utilizing ANNs was compared with
traditional analytic tools such as multiple discriminant analysis (MDA).
Finally, via T-test analysis indicated that predicted accuracy of ANNs is
better in both training and testing phases. |
|
Title: |
KNOWLEDGE
MANAGEMENT SYSTEMS FOR LEVERAGING ENTERPRISE DATA RESOURCES: TAXONOMY AND
KEY ISSUES |
Author(s): |
Mahesh
S. Raisinghani |
Abstract: |
With today’s emphasis on competitiveness, team-based
organizations, and responsiveness, top management cannot separate their
responsibilities between people management and traditional/e-business
management, since they are both interrelated in knowledge management
systems (KMS). Understanding how to manage under conditions of rapid
change is a critical skill in the knowledge economy. Today, work in
organizations of KMS is increasingly organized with teamwork-based;
instead of, the traditional organization charts. As the workforce becomes
increasingly diverse and global, it is important for top management to
recognize that diversity is a positive force for KMS. Today’s team based,
geographically dispersed employees are increasingly guided by a network of
values and tradition as part of an organizational culture in KMS. Managing
that culture and establishing those changed values are crucial KMS
management tasks. This paper explores, describes, and assesses the
integration, impact, and implications of KMS for theory and practice.
|
|
Title: |
CONTENT-BASED
REASONING IN INTELLIGENT MEDICAL INFORMATION SYSTEMS |
Author(s): |
Marek Ogiela
|
Abstract: |
This paper describes an innovative approach to the use of
linguistic methods of structural image analysis in intelligent systems of
visual data perception. They are directed at understanding medical images
and a deeper analysis of their semantic contents. This type of image
reasoning and understanding is possible owing to the use of especially
defined graph grammars enabling one both the correct recognition of
significant disease lesions and conducting a deeper analysis of the
discovered irregularities on various specific levels. The proposed
approach will be described on selected examples of images obtained in
radiological diagnosis. |
|
Title: |
KNOWLEDGE
BASE GRID: TOWARD GLOBAL KNOWLEDGE SHARING |
Author(s): |
Wu Zhaohui,
Xu Jiefeng, Wu Zhaohui |
Abstract: |
Grid technologies enable widespread sharing and
coordinated use of networked resources. Bringing knowledge into Grid can
be more challenging because in such settings, we encounter difficulties
such as standardization of knowledge representation, developing standard
protocols to support semantic interoperability, and developing methodology
to construct on-demand intelligent services. In this paper, we present an
open Knowledge Base Grid architecture that addresses these challenges. We
first discuss the requirements of knowledge representation in the
Internet, and then argue about the importance of developing standard
protocols in such a knowledgeable Internet, at last we present some
inference services which provide high level knowledge services such as
correlative semantic browsing, knowledge query, forward and backward
chaining inference etc. KB-Grid provides a platform for Distributed
Artificial Intelligence. |
|
Title: |
FACE
PATTERN RECOGNITION AND EXTRACTION FROM MULTIPLE PERSONS SCENE |
Author(s): |
Tetsuo Hattori
|
Abstract: |
A method for face recognition of acquaintance as a
subpattern in a given image is proposed. We consider that the face pattern
to be recognized in the input image is approximately an affine transformed
(rotated, enlarged and/or reduced, and translated) pattern of a registered
original one. In order to estimate the parameters of the affine
transformation, the method uses a Karhunen-Loeve (KL) expansion, spatial
correlation, and an approximate equation based on Taylor’s expansion of
affine transformation. In this paper, we deal with two types of pattern
representation: ordinary grey level representation and a normalized
gradient vector field (NGVF) one. The experimental result shows that our
method using NGVF representation is considerably effective. |
|
Title: |
EXTRACTION
OF FEELING INFORMATION FROM CHARACTERS USING A MODIFIED FOURIER
TRANSFORM |
Author(s): |
Tetsuo Hattori
|
Abstract: |
An automated feature extraction and evaluation method of
feeling information from printed and handwritten characters is proposed.
This method is based on image processing and pattern recognition
techniques. First, an input binarized pattern is transformed by a distance
transformation. Second, a two-dimensional vector field is composed from
the gradient of the distance distribution. Third, a divergence operator
extracts source and sink points from the field, and also the vectors on
those points. Fourth, the Fourier transform is done for the vector field
as a complex valued function. Differently from conventional methods, we
deal with the Fourier transform with Laplacian operated phase. Fifth,
applying the KL expansion method to the data of the complex vectors
obtained from some kinds of character fonts, we extract some common
feature vectors of each character font. Using those common vectors and
linear multiple regression model, an automated quantitative evaluation
system can be constructed. The experimental results show that our vector
field method using the combination of Fourier transform and KL expansion
is considerably more efficient in the discrimination of printed characters
(or fonts), comparing with conventional method using gray level (or
binarized) character pattern and KL expansion. Moreover, we obtain the
results that the evaluation system based on the regression model
comparatively meets well to the human assessment. |
|
Title: |
A
CONCEPTUAL MODEL FOR A MULTIAGENT KNOWLEDGE BUILDING SYSTEM |
Author(s): |
Barbro Back,
Adrian Costea, Tomas Eklund, Antonina Kloptchenko
|
Abstract: |
Financial decision makers are challenged by the access to
massive amounts of both numeric and textual financial information made
achievable by the Internet. They are in need of a tool that makes possible
rapid and accurate analysis of both quantitative and qualitative
information, in order to extract knowledge for decision making. In this
paper we propose a conceptual model of a knowledge-building system for
decision support based on a society of software agents, and data and text
mining methods. |
|
Title: |
BRIDGING
THE GAP BETWEEN SOCIAL AND TECHNICAL PROCESSES TO FACILITATE IT ENABLED
KNOWLEDGE DISSEMINATION |
Author(s): |
James Cunningham,
Yacine Rezgui, Brendan Berney, Elaine Ferneley |
Abstract: |
The need for organizations to encourage collaborative
working through knowledge sharing in order to better exploit their
intellectual capital is recognized. However, much of the work to date
suggests that despite the intuitive appeal of a collaborative approach
significant knowledge remains locked away. It has been argued that the
problem is both technological and cultural. Whilst technologically mature,
sophisticated information communication technologies (ICTs) exist,
providing a technological medium to support a collaborative culture in
which knowledge can be elicited, stored, shared and disseminated is still
elusive. This paper presents the work being undertaken as part of the IST
funded e-COGNOS project that is developing an open, model-based
infrastructure and a set of web-based tools that promote consistent
knowledge management within collaborative construction environments. The
e-COGNOS project has adopting an approach which moves away from the notion
of technology managing information and toward the idea of social processes
and technological tools evolving reciprocally – the notion of
co-construction. Within this co-construction metaphor the project is
developing a set of tools that mimic the social process of knowledge
discovery thus aiming to bridge the gap between social and technological
knowledge discovery and dissemination. |
|
Title: |
THE
DEVELOPMENT OF A PROTOTYPE OF AN ENTERPRISE MARKETING DECISION SUPPORT
SYSTEM |
Author(s): |
Junkang Feng,
Xi Wang, Fugen Song |
Abstract: |
Against the background of the increasing importance of
marketing decision making for the manufacturing enterprises and yet
relatively weak and insufficient research on systematic methodologies for
overall marketing decision making, we build up a model-based framework for
marketing decision making. The framework offers an approach of fusing
quantitative calculations with qualitative analysis for marketing decision
making. Our review of the literature on the architecture of a Decision
Support System (DSS) would seem to show that there exists a gap between
the theories of the architecture of a DSS, which consists of mainly a
database (DB), a model base (MB) and a knowledge base (KB), and the use of
this architecture in practical design and implementing a DSS. To fill this
gap, we put forward a notion of “Tri-Base Integration”, based upon which
we have developed and tested an innovative architecture for a DSS. We have
built a prototype of an Enterprise Marketing Decision Support System based
upon these ideas. This prototype would seem to have proven the feasibility
of our model-based framework for overall marketing decision making and our
innovative architecture for a DSS. |
|
Title: |
APPLICATION
OF NEURAL NETWORKS TO WATER TREATMENT: MODELING OF COAGULATION
CONTROL |
Author(s): |
M. Salem,
Hala Abdel-Gelil, L. Abdel All |
Abstract: |
Water treatment includes many complex phenomena, such as
coagulation and flocculation. These reactions are hard or even impossible
to control by conventional methods. The paper presents a new methodology
for determining the optimum coagulate dosage in water treatment process.
The methodology is based on a neural network based-model; the learning
process is implemented by using the Error Backpropagation algorithm using
raw water quality parameters as input. |
|
Title: |
USING
KNOWLEDGE DISCOVERY IN DATABASES TO IDENTIFY ANALYSIS PATTERNS |
Author(s): |
Paulo Engel,
Carolina Silva, Cirano Iochpe |
Abstract: |
Geographic information systems (GIS) are becoming more
popular, increasing the need of implementing geographic databases (GDB).
But the GDB design is not easy and requires experience in the task. To
support that, the use of analysis patterns has been proposed. Although
very promising, the use of analysis patterns in GDB design is yet very
restrict. The main problem is that patterns are based on specialists’
experience. In order to help and speed up the identification of new and
valid patterns, which are less dependent on specialists’ knowledge than
those now available, this paper proposes the identification of analysis
patterns on the basis of the process of knowledge discovery in databases
(KDD). |
|
Title: |
SEMANTIC
ANNOTATIONS AND SEMANTIC WEB USING NKRL (NARRATIVE KNOWLEDGE
REPRESENTATION LANGUAGE) |
Author(s): |
Gian Zarri
|
Abstract: |
We suggest that it should be possible to come closer to
the Semantic Web goals by using ‘semantic annotations’ that enhance the
traditional ontology paradigm by supplementing the ontologies of concepts
with ‘ontologies of events’. We present then some of the properties of
NKRL (Narrative Knowledge Representation Language), a conceptual modelling
formalism that makes use of ontologies of events to take into account the
semantic characteristics of those ‘narratives’ that represent a very large
percentage of the global Web information. |
|
Title: |
INDUCTION
OF TEMPORAL FUZZY CHAINS |
Author(s): |
Jose
Jesus Castro Sanchez, Luis Rodriguez Benitez, Luis Jimenez
Linares, Juan Moreno Garcia |
Abstract: |
The aim of this paper is to present an algorithm to induce
the Temporal Fuzzy Chains (TFCs) (Eurofuse 2002). TFCs are used to model
the dynamic systems in a linguistic manner. TFCs make use of two different
concepts: the traditional method to represent the dynamic systems named
state vectors, and the linguistic variables used in fuzzy logic. Thus,
TFCs are qualitative and represents the "temporal zones" using linguistic
states and linguistic transitions between the linguistic states. |
|
Title: |
THE
PROTEIN STRUCTURE PREDICTION MODULE OF THE PROT-GRID |
Author(s): |
Dimitrios
Frossyniotis, George Papadopoulos, Dimitrios Vogiatzis
|
Abstract: |
In this work, we describe the protein secondary structure
prediction module of a distributed bio-informatics system. Protein
databases contain over a million of sequenced proteins, however there is
structuring information for at most 2\% of that number. The challenge is
to reliably predict the structure based on classifiers. Our contribution
is the evaluation of architectures of multiple classifier systems on a
standard dataset (CB-396) containing protein sequencing information. We
compare the results of a single classifier system based on SVMs, as well
as with our version of an SVM based adaBoost algorithm and a novel fuzzy
multi-SVM classifier. |
|
Title: |
WITH
THE "DON'T KNOW" ANSWER IN RISK ASSESSMENT |
Author(s): |
Luigi Troiano,
Canfora Gerardo |
Abstract: |
Decision making often deals with incomplete and uncertain
information. Uncertainty concerns the level of confidence associated with
the value of a piece of information, while incompleteness derives from the
unavailability of data. Fuzzy numbers capture the uncertainty of
information, but they are not able to explicitly represent incompleteness.
In this paper we discuss an extension of fuzzy numbers, called fuzzy
numbers with indeterminateness, and show how they can be used to model
decision process involving incomplete information. In particular, the
paper focuses on the ``Don't Know'' answer to questionnaires and develops
an aggregation model that accounts for these type of answers. The main
contribution lies in the formalization of the interrelationships between
the risk of a decision and the incompleteness of the information on which
it is made. |
|
Title: |
FUZZY
INFERENCING IN WEB PAGE LAYOUT DESIGN |
Author(s): |
Abdul-Rahim Ahmad,
Otman Basir, Khaled Hassanein |
Abstract: |
The Web page layout design is a complex and ill-structured
problem where the evolving tasks, inadequate information processing
capabilities, cognitive biases and socio-emotional facets frequently
hamper the procurement of a superior alternative. An important aspect in
selection of a superior Web page layout design is the evaluation of its
fitness value. Automating the fitness evaluation of layouts seems to be a
significant step forward. It requires quantification of highly subjective
Web page design guidelines in the form of some fitness measure. The Web
usability and design guidelines come from experts who provide vague and
conflicting opinions. This paper proposes the exploitation of fuzzy
technology in modeling such subjective, vague, and uncertain Web usability
and design guidelines. |
|
Title: |
MAPPING
DOCUMENTS INTO CONCEPT DATABASES FOR THRESHOLD-BASED RETRIEVAL |
Author(s): |
REGHU
RAJ PATINHARE COVILAKAM, RAMAN S |
Abstract: |
The trajectory of topic description in text documents such
as news articles generally covers a small number of domain-specific
concepts. Domain-specific phrases are excellent indicators of these
concepts. Any form of representation of the concepts must invariably use
finite strings of some finite representation language. Then, the design of
a grammar with good selectivity and coverage is a viable solution to the
problem of content capturing. This paper deals with the design of such a
grammar for a small set of domains, which helps the representation of the
concepts using the relational framework. This paradigm throws light into
the possibility of denoting the text portion of web pages as a relational
database, which can facilitate information retrieval using simple SQL
queries obtained by translating a user's query. The advantage is that
highly relevant results can be retrieved by looking for a threshold value
in a specific attribute column. |
|
Title: |
A
NEW METHOD OF KNOWLEDGE CREATION FOR KNOWLEDGE ORGANIZATIONS |
Author(s): |
Mingshu Li,
Ying Dong |
Abstract: |
Knowledge creation is an interesting problem in knowledge
management (KM). Topic maps, especially XML Topic Map (XTM), is used to
organize information in a way that can be optimized for navigation. In
this paper, we adopt XTM as a new method to discuss the problem of
knowledge creation. Since XTM can be modeled as a formal hypergraph, we
study the problem based on XTM hypergraph. New XTM knowledge operations
have been designed for based on graph theories for knowledge creation.
Moreover, they have been implemented as a toolkit, and applied on our KM
platform. When applying the XTM knowledge operations, new knowledge can be
generated for knowledge organizations. The application of the operations
can fit users’ requests on the intelligent retrieval of the knowledge, or
on the analysis of the system knowledge structure. |
|
Title: |
AN
ARTIFICIAL NEURAL NETWORK BASED DECISION SUPPORT SYSTEM FOR
BUDGETING |
Author(s): |
Barbro Back,
Eija Koskivaara |
Abstract: |
This paper introduces an artificial neural network (ANN)
based decision support system for budgeting. The proposed system estimates
the future revenues and expenses of the organisation. We build models
based on four to six years’ monthly account values of a big organisation.
The monthly account values are regarded as a time-series and the target is
to predict the following year’s account values with the ANN. Thus, the
ANN’s output information is based on similar information on prior periods.
The prediction results are compared to the actual account values and to
the account values budgeted by the organisation. We found that ANN can be
used for modeling the dynamics of the account values on monthly basis and
for predicting the yearly account values. |
|
Title: |
A
DATA MINING METHOD TO SUPPORT DECISION MAKING IN SOFTWARE DEVELOPMENT
PROJECTS |
Author(s): |
José
Luis Álvarez-Macías |
Abstract: |
In this paper, we present a strategy to induce knowledge
as support decision making in Software Development Projects (SDP). The
motive of this work is to reduce the great quantity of SDP do not meet the
initial cost requirements, delivery date and the quality of the final
product. The main objective of this strategy is to support the manager in
the decision taking to establish the policies from management when
beginning a software project. Thus, we apply a data mining tool, called
ELLIPSES, on databases of SDP. The database are generated by means of the
simulation of a dynamic model for the management of SDP. ELLIPSES tool is
a new method oriented to discover knowledge according to the expert's
needs, by the detection of the most significant regions. The method
essence is found in an evolutionary algorithm that finds these regions one
after another. The expert decides which regions are significant and
determines the stop criterion. The extracted knowledge is offered through
two types of rules: quantitative and qualitative models. The tool also
offers a visualization of each rule by parallel coordinate systems. In
order to present this strategy, ELLIPSES is applied to a database which
has already been obtained by means of the simulation of a dynamic model on
a project concluded. |
|
Title: |
USABILITY
ISSUES IN DATA MINING SYSTEMS |
Author(s): |
Fernando Berzal
|
Abstract: |
When we build data mining systems, we should reflect upon
some design issues which are often overlooked in our quest for better data
mining techniques. In particular, we usually focus on algorithmic details
whose influence is minor when it comes to users’ acceptance of the systems
we build. This paper tries to highlight some of the issues which are
usually neglected and might have a major impact on our systems usability.
Solving some of the usability problems we have identified would certainly
add to the odds of successful data mining stories, improve user acceptance
and use of data mining systems, and spur renewed interest in the
development of new data mining techniques. Our proposal focuses on
integrating diverse tools into a framework which should be kept coherent
and simple from the user's point of view. Our experience suggests that
such a framework should include bottom-up dataset-building blocks to
describe input datasets, expert systems to propose suitable algorithms and
adjust their parameters, as well as visualization tools to explore data,
and communication and reporting services to share the knowledge discovered
from the massive amounts of data available in actual databases. |
|
Title: |
PLANNING
COOPERATIVE HOMOGENEOUS MULTIAGENT SYSTEMS USING MARKOV DECISION
PROCESSES |
Author(s): |
Bruno Scherrer,
François Charpillet, Iadine Chadès |
Abstract: |
This paper proposes a decision-theoric approach for
designing a set of situated agents so that they can solve a cooperative
problem. The approach we propose is based on reactive agents. Although
they do not negotiate, reactive agents can solve complex tasks such as
surrounding a mobile object : agents self-organize their activity through
the interaction with the environment. The design of each agent's behavior
results from solving a decentralized partially observable markov decision
process (DEC-POMDP). But, as solving a DEC-POMDP is NEXP-complete, we
propose an approximate solution to this problem based on both subjectivity
and empathy. An obvious advantage of the proposed approach is that we are
able to design agents' reactive policies considering features of a
cooperative problem (top-down conception) and not the opposite (down-top
conception). |
|
Title: |
AN
EFFICIENT PROCEDURE FOR ARTIFICIAL NEURAL NETWORKS RETRAINING |
Author(s): |
Razvan Matei,
Dumitru Iulian Nastac |
Abstract: |
The artificial neural networks (ANNs) ability to extract
significant information from an initial set of data allows both an
interpolation in the a priori defined points, and an extrapolation outside
the range bordered by the extreme points from the training set. The main
purpose of this paper is to establish how a viable ANN structure at a
previous moment of time could be re-trained in an efficient manner in
order to support modifications of the input-output function. To be able to
fulfill our goal, we use an anterior memory, scaled with a certain
convenient value. The evaluation of the computing effort involved in the
retraining of an ANN shows us that a good choice for the scaling factor
can substantially reduce the number of training cycles independent of the
learning methods. |
|
Title: |
PROMAIS:
A MULTI-AGENT DESIGN FOR PRODUCTION INFORMATION SYSTEMS |
Author(s): |
Lobna Hsairi,
Khaled Ghédira, Faiez Gargouri |
Abstract: |
In the age of information proliferation and communication
technology advances, Cooperative Information System (CIS) technology
becomes a vital factor for production system design in every modern
enterprise. In fact, current production system must hold to new strategic,
economic and organizational structures in order to face new challenges.
Consequently, intelligent software based on agent technology emerges to
improve system design on the one hand, and to increase production
profitability and enterprise competitive position on the other hand. This
paper starts with an analytical description of logical and physical flows
dealt with manufacturing, then proposes one Production Multi-Agent
Information System (ProMAIS). ProMAIS is a collection of stationary and
intelligent agent-agencies with specialized expertises, interacting to
carry out the shared objectives: cost-effective production in promised
delay and adaptability to the changes. In order to bring ProMAIS’s dynamic
aspect out, interaction protocols are specially zoomed out by cooperation,
negotiation and Contract Net protocols. |
|
Title: |
TEXT
SUMMARIZATION: AN UPCOMING TOOL IN TEXT MINING |
Author(s): |
S. Raman,
M. Saravanan |
Abstract: |
As Internet’s user base expands at an explosive rate, it
provides great opportunities as well as grand challenges for text data
mining. Text Summarization is the core functional task of text mining and
text analysis, and it consists of condensing documents and also in a
coherent order. This paper discusses the application of term distribution
models to text summarization for the extraction of key sentences based on
the identification of term patterns from the collection. The evaluation of
the results is based on the human-generated summaries as a point of
reference. Our system outperforms the other auto-summarizers considered at
different percentage a level of summarization, and the final summary is
close to the intersection of the frequently occurring sentences found in
the human-generated summaries at 40% summarization level. |
|
Title: |
AUTOMATION
OF CORE DESIGN OPTIMIZATION IN BWR |
Author(s): |
Yoko Kobayashi
|
Abstract: |
This paper deals with the application of evolutionary
algorithm and multi-agents algorithm to the information system in a
nuclear industry. The core design of a boiling water reactor (BWR) is a
hard optimization problem with nonlinear multi objective functions and
nonlinear constrains. We have developed an integrative two-stage genetic
algorithm (GA) to the optimum core design of a BWR and have realized the
automation of a complex core design. In this paper, we further propose a
new algorithm for combinatorial optimization using multi-agents. We name
it as multi-agents algorithm (MAA). In order to improve the convergence
performance of the core design optimization of BWR, we introduce this new
algorithm to the first stage of the two-stage GA previously developed. The
performance of the new algorithm also compared with the conventional
two-stage GA. |
|
Title: |
LEARNING
BAYESIAN NETWORKS FROM NOISY DATA. |
Author(s): |
Mohamed BENDOU,
Paul MUNTEANU |
Abstract: |
This paper analyzes the effects of noise on learning
Bayesian networks from data. It starts with the observation that limited
amounts of noise may cause a significant increase of the complexity of
learned networks. We show that, unlike classical over-fitting which
affects other classes of learning methods, this phenomenon is
theoretically justified by the alteration of the conditional independence
relations between the variables and is beneficial for the predictive power
of the learned models. We also discuss a second effect of noise on
learning Bayesian networks: the instability of the structures learned from
DAG-unfaithful noisy data. |
|
Title: |
BUILDING
INTELLIGENT CREDIT SCORING SYSTEMS USING DECISION TABLES |
Author(s): |
Manu De
Backer, Rudy Setiono, Christophe Mues, Jan Vanthienen,
Bart Baesens |
Abstract: |
Accuracy and comprehensibility are two important criteria
when developing decision support systems for credit scoring. In this
paper, we focus on the second criterion and propose the use of decision
tables as an alternative knowledge visualization formalism which lends
itself very well to build intelligent and user-friendly credit scoring
systems. Starting from a set of propositional if-then rules extracted by a
neural network rule extraction algorithm, we develop decision tables and
demonstrate their efficiency and user-friendliness for 2 real-life credit
scoring cases. |
|
Title: |
EVALUATING
THE SURVIVAL CHANCES OF VERY LOW BIRTHWEIGHT BABIES |
Author(s): |
Anália Lourenço,
Ana Cristina Braga, Orlando Belo |
Abstract: |
Scoring systems that quantify neonatal mortality have an
important role in health services research, planning and clinical
auditing. They provide means to monitoring, in a more accurate and
reliable way, the quality of care among and within hospitals. The
classical analyses based on a simple comparison of mortality or dealing
with the newborns birthweight solely have proved to be insufficient. There
are a large number of variables that influence the survival of newborns
that must to be taken into account. From strictly physiological
information through more subjective data, concerning medical care, there
are many variables to attend to. Scoring systems try to embrace such
elements, providing more reliable comparisons of the outcome.
Notwithstanding, if a clinical score intends to gain widespread between
clinicians, it must be simple and accurate and use routine data. In this
paper, it is presented a neonatal mortality risk evaluation case study,
pointing out data specificities and how different data preparation
approaches (namely, feature selection) will affect the overall
outcome. |
|
Title: |
THE
USE OF NEURAL NETWORK AND DATABASE TECHNOLOGY TO REENGINEER THE TECHNICAL
PROCESS OF MONITORING COAL COMBUSTION EFFICIENCY |
Author(s): |
Farhi Marir
|
Abstract: |
Monitoring the combustion process for electricity
generation using coal as a primary resource, is of a major concern to the
pertinent industries, power generation companies in particular. The carbon
content of fly ash is indicative of the combustion efficiency. The
determination of this parameter is useful to characterise the efficiency
of coal burning furnaces. Traditional methods such as thermogrametric
analysis (TGA) and loss on ignition which are based on ash collection and
subsequent analysis, proved to be tediously difficult, time consuming and
costly. Thus, a need for a new technology was inevitable and needed to
monitor the process in a more efficient method yielding a better
exploitation of the resources at the expense of a low cost. The main aim
of this work is to introduce a new automated system which can be bolted
onto a furnace and work online. The system consists of three main
components, namely, a laser instrument for signal acquisition, a neural
network tool for training, learning and simulation, and a database system
for storage and retrieval. The components have been designed, adapted and
tuned to communicate for knowledge acquisition of this multidimensional
problem. The system has been tested for a range of coal ashes and proved
to be efficient . reliable, fast and cost effective. |
|
Title: |
A
KNOWLEDGE MANAGEMENT TOOL FOR A COLLABORATIVE E-PROJECT |
Author(s): |
Luc Lamontagne,
Tang-Ho Lê |
Abstract: |
In this paper, we provide an overview of our software tool
to exploit and interchange procedural knowledge represented as networks of
semi-structured units. First, we introduce the notion of Procedural
Knowledge Hierarchy; then we present the modeling of Procedural Knowledge
by our software. We claim that the “bottom-up” approach, that is being
carried out with this tool, is appropriate to gather new candidate terms
for the construction of a new domain ontology. We also argue that the KU
modeling together with a pivot KU structure (rather than individual
keywords) would contribute a solution to the search engine on the Web. We
detail the updating technique basing on the distributed tasks of an
e-project. We also discuss some ideas pertaining to the identity issue for
the web based on some space and time representation. |
|
Title: |
STRUCTURED
CONTEXTUAL SEARCH FOR THE UN SECURITY COUNCIL |
Author(s): |
Irineu Theiss,
Ricardo Barcia, Marcelo Ribeiro, Eduardo Mattos,
Andre Bortolon, Tania C. D. Bueno, Hugo Hoeschl |
Abstract: |
This paper presents a generic model of a methodology that
emphasises the use of information retrieval methods combined with the
Artificial Intelligence technique named CBR – Case-Based Reasoning. In
knowledge-based systems, this methodology allows the human knowledge to be
automatically indexed. This type of representation turns compatible the
user language with the language found in the data contained in the
knowledge base of the system, retrieving to the user more adequate answers
to his/her search question. The paper describes the Olimpo System, a
knowledge based system that enables to retrieve information from textual
files, which is similar to the search context described by the user in
natural language. For the development of the system, 300 Resolutions of
the UN Security Council available on the Internet were indexed. |
|
Title: |
APPLYING
FUZZY LOGIC AND NEURAL NETWORK FOR QUANTIFYING FLEXIBILITY OF SUPPLY
CHAINS |
Author(s): |
Bjørn Solvang,
Ziqiong Deng, Wei Deng Solvang |
Abstract: |
Fuzzy Logic (FL) is the method that deals with uncertainty
and vagueness in the model or description of the systems involved as well
as those in the variables. A fuzzy logic system is unique in that it is
able to handle numerical and linguistic knowledge, simultaneously. This is
precisely the method that we’ve looking for when the quantification of
supply chain flexibility has become an urgent task. This paper first
elaborates the necessity of quantification of supply chain flexibility.
Thereafter, a methodological framework for measurement of supply chain
flexibility is introduced with the purpose of providing the research
background of this paper. Fuzzy logic system is applied to quantify six
types of supply chain flexibility as each depends on both qualify and
quantify measures. Further, since the value of supply chain flexibility is
also decided by the degree that it depends on each type of flexibility and
the decision of these degrees needs the incorporation of expert knowledge,
we apply Artificial Neural Network (ANN) to conduct the task. |
|
Title: |
AN
APPROACH OF DATA MINING USING MONOTONE SYSTEMS |
Author(s): |
Rein Kuusik,
Grete Lind |
Abstract: |
This paper treats data mining as a part of the process
called knowledge discovery in databases (KDD in short), which consists of
particular data mining algorithms and, under some acceptable computational
efficiency limitations, produces a particular enumeration of patterns.
Pattern is an expression (in a certain language) describing facts in a
subset of facts. The data mining step is one of the most implemented steps
of the whole KDD process. Also the KDD process involves preparing data for
analysis and interpreting results found in data mining step. The main
approach to data mining and its main disadvantage is shown and new method,
called generator of hypotheses, and its base algorithm MONSA is
presented. |
|
Title: |
DEVELOPMENT
OF AN ORGANIZATIONAL SUBJECT |
Author(s): |
Chamnong Jungthirapanich,
Parkpoom Srithimakul |
Abstract: |
Due to the globalization of markets are competitive,
skillful employees are most wanted. Therefore, it reflects to high turn
over rate in each organization. This research would create the pattern to
retain the knowledge of those employees as called “the organizational
subject model”. This pattern captures inner capability of the employees
and develops to be the contents for the organization, then uses the
educational method transform these contents to be the subject which is
called “the organizational subject”. The organizational subject model is
the new strategy to retain the knowledge of the skillful employees. This
research also shows the statistical method to evaluate the efficiency and
the effectiveness of the organizational subject and the hypothesis testing
to evaluate the achievement of the organizational subject model. This
model saves the knowledge capital investment, time, and futhermore to
identify the unity of the organization. |
|
Title: |
MINING
VERY LARGE DATASETS WITH SUPPORT VECTOR MACHINE ALGORITHMS |
Author(s): |
François Poulet,
Thanh-Nghi Do |
Abstract: |
In this paper, we present new support vector machines
(SVM) algorithms that can be used to classify very large datasets on
standard personal computers. The algorithms have been extended from three
recent SVMs algorithms: least squares SVM classification, finite Newton
method for classification and incremental proximal SVM classification. The
extension consists in building incremental, parallel and distributed SVMs
for classification. Our three new algorithms are very fast and can handle
very large datasets. An example of the effectiveness of these new
algorithms is given with the classification into two classes of one
billion points in 10-dimensional input space in some minutes on ten
personal computers (800 MHz Pentium III, 256 Mb RAM, Linux). |
|
Title: |
EXTENSION
OF THE BOX-COUNTING METHOD TO MEASURE THE FRACTAL DIMENSION OF FUZZY
DATA |
Author(s): |
Antonio
B. Bailón |
Abstract: |
The box-counting is a well known method used to estimate
the dimension of a set of points that define an object. Those points are
expressed with exact numbers that don't reflect the uncertainty that
affects them in many cases. In this paper we propose an extension to the
box-counting method that allows the measure of the dimension of sets of
fuzzy points, i.e. sets of points affected by some degree of uncertainty.
The fuzzy box-counting method allows the extension of algorithms that use
the fractal dimension of sets of crisp points to enable them to work with
fuzzy data. |
|
Title: |
TRACKER:
A FRAMEWORK TO SUPPORT REDUCING REWORK THROUGH DECISION
MANAGEMENT |
Author(s): |
Andy Salter,
Phil Windridge, Alan Dix, Rodney Clarke,
Caroline Chibelushi, John Cartmell, Ian Sommerville,
Victor Onditi, Hanifa Shah, Devina Ramduny,
Amanda Queck, Paul Rayson, Bernadette Sharp,
Albert Alderson |
Abstract: |
The Tracker project is studying rework in systems
engineering projects. Our hypothesis is that providing decision makers
with information about previous relevant decisions will assist in reducing
the amount of rework in a project. We propose an architecture for the
flexible integration of the tools implementing the variety of theories and
models used in the project. The techniques include ethnographic analysis,
natural language processing, activity theory, norm analysis, and speech
and handwriting recognition. In this paper, we focus on the natural
language processing components, and describe experiments which demonstrate
the feasibility of our text mining approach. |
|
Title: |
EVALUATION
OF AN AGENT-MEDIATED COLLABORATIVE PRODUCTION PROTOCOL IN AN INSTRUCTIONAL
DESIGN SCENARIO |
Author(s): |
Ignacio Aedo,
Paloma Díaz, Juan Manuel Dodero |
Abstract: |
Distributed knowledge creation or production is a
collaborative task that needs to be coordinated. A multiagent architecture
for collaborative knowledge production tasks is introduced, where
knowledge-producing agents are arranged into knowledge domains or marts,
and a distributed interaction protocol is used to consolidate knowledge
that is produced in a mart. Knowledge consolidated in a given mart can be
in turn negotiated in higher-level foreign marts. As an evaluation
scenario, the proposed architecture and protocol are applied to facilitate
coordination during the creation of learning objects by a distributed
group of instructional designers. |
|
Title: |
SYMBOLIC
MANAGEMENT OF IMPRECISION |
Author(s): |
Mazen EL-SAYED,
Daniel PACHOLCZYK |
Abstract: |
This paper presents a symbolic model for handling nuanced
information like "John is very tall". The model presented is based on a
symbolic M-valued predicate logic. The first object of this paper has been
to present a new representation method for handling nuanced statements of
natural language and which contains linguistic modifiers. These modifiers
are defined in a symbolic way within a multiset context. The second object
has been to propose new Generalized Modus Ponens rules dealing with
nuanced statements. |
|
Title: |
LIVE-REPRESENTATION
PROCESS MANAGEMENT |
Author(s): |
Daniel
Corkill |
Abstract: |
We present the live-representation approach for managing
and working in complex, dynamic business processes. In this approach,
important aspects of business-process modeling, project planning, project
management, resource scheduling, process automation, execution, and
reporting are integrated into an detailed, on-line representation of
planned and executing processes. This representation provides a real-time
view of past, present, and anticipated process activities and resourcing.
Changes resulting from process dynamics are directly reflected in the live
representation, so that, at any point in time, the latest information
about process status and downstream expectations is available. Managers
can directly manipulate the live representation to change process
structure and execution. These changes are immediately propagated
throughout the environment, keeping managers and process participants in
sync with process changes. A fundamental aspect of the live-representation
approach is obtaining and presenting current and anticipated activities as
an intrinsic and organic aspect of each participant's daily activities. By
becoming an active partner in these activities, the environment provides
tangible benefits in keeping everyone informed and coordinated without
adding additional duties and distractions. Equally important are providing
individuals the flexibility to choose when and how to perform activities
and allowing them to provide informative details of their progress without
being intrusive into the details of their workdays. In this paper, we
describe the technical and humanistic issues associated with the
live-representation approach and summarize the experiences gained in
providing a commercial implementation used in the automotive and aerospace
industries. |
|
Title: |
MR-BRAIN
IMAGE SEGMENTATION USING GAUSSIAN MULTIRESOLUTION ANALYSIS AND THE EM
ALGORITHM |
Author(s): |
Mohammed A-Megeed,
Mohammed F. Tolba, Mostafa Gad, Tarek Gharib |
Abstract: |
We present a MR image segmentation algorithm based on the
conventional Expectation Maximization (EM) algorithm and the
multiresolution analysis of images. Although the EM algorithm was used in
MRI brain segmentation, as well as, image segmentation in general, it
fails to utilize the strong spatial correlation between neighboring
pixels. The multiresolution-based image segmentation techniques, which
have emerged as a powerful method for producing high-quality segmentation
of images, are combined here with the EM algorithm to overcome its
drawbacks and in the same time take its advantage of simplicity. Two data
sets are used to test the performance of the EM and the proposed Gaussian
Multiresolution EM, GMEM, algorithm. The results, which proved more
accurate segmentation by the GMEM algorithm compared to that of the EM
algorithm, are represented statistically and graphically to give deep
understanding. |
|
Title: |
EPISTHEME:
A SCIENTIFIC KNOWLEDGE MANAGEMENT ENVIRONMENT |
Author(s): |
Julia
Strauch, Jonice Oliveira, Jano Souza |
Abstract: |
Nowadays, researchers create and change information faster
than in the past. Although great part of this exchanging is made by
documental form, there is also a great informal or tacit knowledge
exchange in people interactions.For the success of a scientific activity,
it is necessary that researchers be provided with all necessary knowledge
to execute their tasks, to make decisions, collaborate with one another
and disseminate individual knowledge to transform it into organizational
knowledge. In this context, we propose the scientific knowledge management
environment called Epistheme. This environment having as goals: to help
the organizational knowledge management, to be a learning environment, to
facilitate the communication of people on the same domain research and to
unify different perspectives and expertise in a single environment. This
article shows the Epistheme framework with the modules of identification,
creation, validation, integration, acquisition and knowledge
dissemination. |
|
Title: |
A
PROCESS-CENTERED APPROACH FOR KDD APPLICATION MANAGEMENT |
Author(s): |
Karin Becker,
Karin Becker |
Abstract: |
KDD is the knowledge-intensive task consisting of complex
interactions, protracted over time, between a human and a (large)
database, possibly supported by a heterogeneous suite of tools. Managing
this complex process, its underlying activities, resources and results, is
a laborious and complex task. In this paper, we present a documentation
model to structure and organize information necessary to manage a KDD
application, based on the premise that documentation is important not only
for better managing efforts, resources, and results, but also to capture
and reuse project and corporate experiences. The documentation model is
very flexible, independent of the particular process methodology and tools
applied, and its use through a supporting environment allows the capture,
storage and retrieval of information at any desired detail level, thus
adaptable to any analyst profile or corporation policy. The approach
presented is based on process-oriented organizational memory information
systems, which aim at capturing the informal knowledge generated and used
during corporate processes. The paper presents the striking features of
the model, and discusses its use in a real case study. |
|
Title: |
A
HYBRID CASE-BASED ADAPTATION MODEL FOR THYROID CANCER DIAGNOSIS |
Author(s): |
Abdel-Badeeh M. Salem,
Khaled
A. Nagaty, Bassant Mohamed El
Bagoury |
Abstract: |
: Adaptation in Case-Based Reasoning (CBR) is a very
difficult knowledge-intensive task, especially for medical diagnosis. This
is due to the complexities of medical domains, which may lead to uncertain
diagnosis decisions. In this paper, a new hybrid adaptation model for
cancer diagnosis has been developed. It combines transformational and
hierarchical adaptation techniques with certainty factors (CF’s) and
artificial neural networks (ANN’s). The model consists of a hierarchy of
three phases that simulates the expert doctor reasoning phases for cancer
diagnosis, which are the Suspicion, the To-Be-Sure and the Stage phases.
Each phase uses the learning capabilities of a single ANN to learn the
adaptation knowledge for performing the main adaptation task. Our model
first formalizes the adaptation knowledge using IF-THEN transformational
rules and then maps the transformational rules into numeric or binary
vectors for training the ANN at each phase. The transformational rules of
the Suspicion phase encode assigned CF’s to reflect the expert doctors’
feelings of cancer suspicion. The model is applied to thyroid cancer
diagnosis and is tested with 820 patient cases, which are obtained from
the expert doctors in the National Cancer Institute of Egypt.
Cross-validation test has shown a very high diagnosis performance rate
that approaches 100% with error rate 0.53%. The hybrid adaptation model is
described in the context of a prototype namely: Cancer-C that is a hybrid
expert system, which integrates neural networks into the CBR cycle. |
|
Title: |
DYNAMICS
OF COORDINATION IN INTELLIGENT SOCIAL MULTI-AGENTS ON ARTIFICAL MARKET
MODEL |
Author(s): |
Junko SHIBATA,
Wataru SHIRAKI, Koji OKUHARA |
Abstract: |
We propose market selection problems in consideration of
agent's preference. The artificial market is based on Hogg-Huberman model
with reward mechanism. By using our model, agents can not only make use of
imperfect and delayed information but also take the preference of the
agent into account on market selection. Our model includes a conventional
model that the benefit is an only factor for selecting. Finally the
dynamical behaviors of our system are investigated numerically. From
results of simulation, it is shown that agent's preference and uncertainty
for market selection. |
|
Title: |
PARTIAL
ABDUCTIVE INFERENCE IN BAYESIAN NETWORKS BY USING PROBABILITY
TREES |
Author(s): |
Jose
A. Gámez |
Abstract: |
The problem of partial abductive inference in Bayesian
networks is, in general, more complex to solve than other inference
problems as probability/evidence propagation or total abduction. When join
trees are used as the graphical structure over which propagation will be
carried out, the problem can be decomposed into two stages: (1) to obtain
a join tree containing only the variables included in the explanation set,
and (2) to solve a total abduction problem over this new join tree. In De
Campos et al. (2002) different techniques are studied in order to approach
this problem, obtaining as a result that not always the methods which
obtain join trees with smaller size are also those requiring less CPU time
during the propagation phase. In this work we propose to use (exact and
approximate) {\em probability trees} as the basic data structure for the
representation of the probability distributions used during the
propagation. From our experiments, we observe how the use of exact
probability trees improves the efficiency of the propagation. Besides,
when using approximate probability trees the method obtain very good
approximations and the required resources decrease considerably. |
|
Title: |
ONTOLOGY
LEARNING THROUGH BAYESIAN NETWORKS |
Author(s): |
Mario Vento,
Francesco Colace, Pasquale Foggia, Massimo De Santo
|
Abstract: |
In this paper, we propose a method for learning ontologies
used to model a domain in the field of intelligent e-learning systems.
This method is based on the use of the formalism of Bayesian networks for
representing ontologies, as well as on the use of a learning algorithm
that obtains the corresponding probabilistic model starting from the
results of the evaluation tests associated with the didactic contents
under examination. Finally, we present an experimental evaluation of the
method using real world data |
|
Title: |
LOGISTICS
BY APPLYING EVOLUTIONARY COMPUTATION TO MULTICOMMODITY FLOW
PROBLEM |
Author(s): |
Koji OKUHARA,
Wataru SHIRAKI, Eri DOMOTO, Toshijiro TANAKA |
Abstract: |
In this paper, we propose an application of one of
evolutionary computation, genetic algorithm, to logistics in
multicommodity flow problem. We chose a multicommodity flow problem which
can evaluate its congestion by traffic arrival ratio in a link. In
simulation, we show that a proposed network control method using genetic
algorithm is superior to the usual method which makes a path selection by
the Dijkstra method and a traffic control by the gradient method. |
|
Title: |
TOOL
FOR AUTOMATIC LEARNING OF BAYESIAN NETWORKS FROM DATABASE: AN APPLICATION
IN THE HEALTH AREA |
Author(s): |
Cristiane Koehler
|
Abstract: |
The learning of Bayesian Networks process is composed of
two stages: learning topology and learning parameters associated to this
topology. Currently, one of the most important research in the Artificial
Intelligence area is the development of efficient inference techniques to
use in intelligent systems. However, the usage of such techniques need the
availability of a valid knowledge model. The necessity to extract
knowledge from databases is increasing exponentially. More and more, the
amount of information exceeds the analysis capacity by the traditional
methods that do not analyse the information under the knowledge focus. It
is necessary the development of new techniques and tools to extract
knowledge from databases. In this article, the concepts of Data Mining and
knowledge breakthrough based on the Bayesian Networks technology had been
used to extract valid models of knowledge. Some learning bayesian
algorithm were been studied, where problems were founded, mainly in the
generation of the topology of the network with all the available variable
in the database. The application domain of this research is the healht
area, it was attested that in the clinical practice, the experts think
only with the more important variables to the decision taking. Some
algorithms have been analysed, and finally, a new algorithm was considered
to extract bayesian models considering only the more relevant variables to
the construction of the network topology. |
|
Title: |
COMPUTER
GAMES AND ECONOMICS EXPERIMENTS |
Author(s): |
Kay-Yut Chen,
Ren Wu |
Abstract: |
HP Labs has developed a software platform, called MUMS,
for moderating economics games between human and/or robot participants.
The primary feature of this platform is a flexible scripting language that
allows a researcher to implement any economics games in a relative short
time. This scripting language eliminates the need to program low-level
functions such as networking, databases and interface components. The
scripts are description of games including definitions of roles, timing
rules, the game tree (in a stage format), input and output (with respect
to a role, not client software). Definitions of variables and the use of
common mathematical and logical operations are also allowed to provide
maximum flexibility in handling the logic of games. This platform has been
used to implement a wide variety of business related games including
variations of a retailer game with simulated consumers and complex
business rules, a double sided call market and negotiation in a
procurement scenario. These games are constructed to accurately simulate
HP business environments. Carefully calibrated experiments, with human
subjects whose incentives were controlled by monetary compensations, were
conducted to test how different business strategies result in different
market behavior. For example, the retailer game was used to test how the
market reacts to changes of HP's contract terms such as return policies.
Experiment results were used in major HP consumer businesses to make
policy decisions. |
|
Title: |
MINING
WEB USAGE DATA FOR REAL-TIME ONLINE RECOMMENDATION |
Author(s): |
Stephen Rees,
Mo Wang |
Abstract: |
A user's browser history contains a lot of information
about the relationship between web pages and users. If this information
can be fully exploited, it may provide better knowledge about the user's
online behaviours and can provide better customer services and site
performance. In this paper, an online recommendation model is proposed
based on the web usage data. A special data structure for storing the
discovered item sets is described. This data structure is especially
suitable for online real time recommendation systems. Users are first
classified using neural network algorithm. Then within each group,
association rules algorithm is employed to discover common user profiles.
In this process, users' interested web sections has been traced and
modeled. Multiple support levels for different types page views and
varying window sizes are also considered. Finally, a recommendation sets
are generated based on user's active session. A demo website is provided
to demonstrate the proposed model. |
|
Title: |
TEXT
MINING FOR ORGANIZATIONAL INTELLIGENCE |
Author(s): |
Hercules do
Prado, Edilberto Silva, Edilson Ferneda |
Abstract: |
In this article it is presented a case study on the
creation of organisational intelligence in a Brazilian news agency
(Radiobras) with the application of text mining tools. Departing from the
question about if Radiobras is fulfilling its social role, we construct an
analysis model based on the enormous volume of texts produced by its
journalists. CRISP-DM method was applied including the acquisition of the
news produced during 2001, preparation of this material, with the
cleansing and formatting of the archives, creation of a model of
clustering and the generation of many views. The views had been supplied
to the administration of the company allowing them to develop more
accurate self-knowledge. Radiobras is an important company of Brazilian
State, that spreads out the acts of the public administration and needs a
self evaluation based in the knowledge of its results. As any other
company, Radiobras is subject to the increasing requirement of
competitiveness imposed to the modern organisations. In this scene, the
generation and retention of organisational intelligence have been
recognised as a competitive differential that can lead to a more adequate
management of the businesses, including its relationship with customers
and in the adequacy of its structure of work. The importance of the
information for the elaboration of the knowledge and, conse-quently, the
synthesis of intelligence is widely recognised, and requires a proper
treatment adjusted to reach insights that can lead to the activation of
the mental processes that will lead to that synthesis. Many internal and
external views on the organisation can be built with the use of tools for
the extraction of patterns from a large amount of data, subsidising
decisively the managers in the decision making process. These views,
constructed to answer the specific questions, constitute knowledge in a
process of Organisational Learning that influences radically the way in
which the organisation is managed. The contributions of IT in this field
were developed, initially, aiming at the extraction of patterns from
transactional databases that contains well structured data. However,
considering that most of the information in the organisations are found in
textual form, recent developments allows the extraction of interesting
patterns from this type of data. Some patterns extracted in our case study
are: (i) the measure of production and geographic distribution of
Radiobras news, (ii) a survey of the most used words, (iii) the discovery
of the covering areas of the news, (iv) the evaluation of how the company
is fulfilling its role, in accordance with the subjects approached in its
news, and (v) the evaluation of the journalistic covering of the company.
|
|
Title: |
STAR
– A MULTIPLE DOMAIN DIALOG MANAGER |
Author(s): |
Márcio Mourão,
Nuno Mamede, Pedro Madeira |
Abstract: |
In this work we propose to achieve not only a dialogue
manager for a domain, but also the aggregation of multiple domains in the
same dialogue management system. With this in mind, we have developed a
dialogue manager that consists of five modules. One of them, called Task
Manager, deserves special attention. Each domain is represented by a
frame, which is in turn composed by slots and rules. Slots define the
domain data relationship, and rules define the system’s behavior. Rules
are composed by operators (logical, conditional, and relational) and
functions that can reference frame slots. The use of frames made possible
all the remaining modules of the dialogue manager to become domain
independent. This is, beyond any doubt, a step ahead in the design of
conversational systems. |
|
Title: |
REQUIREMENTS
OF A DECISION SUPPORT SYSTEM FOR CAPACITY ANALYSIS AND PLANNING IN
ENTERPRISE NETWORKS |
Author(s): |
Américo Azevedo,
Abailardo Moreira |
Abstract: |
Capacity analysis and planning is a key activity in the
provision of adequate customer service levels and the management of the
company’s operational performance. Traditional capacity analysis and
planning systems have become inadequate in the face of several emerging
manufacturing paradigms. One such paradigm is the production in
distributed enterprise networks, consisting of subsets of autonomous
production units within supply chains working in a collaborative and
coordinated way. In these distributed networks, capacity analysis and
planning becomes a complex task, especially because it is performed in a
heterogeneous environment where the performance of individual
manufacturing sites and of the network as a whole should be simultaneously
considered. Therefore, the use of information system solutions is
desirable in order to support effective and efficient planning decisions.
Nevertheless, it seems that there is a lack of a clear definition of the
most important requirements that must be met by supporting solutions. This
paper attempts to identify some general requirements of a decision support
system to be used for capacity analysis and planning in enterprise
networks. Adaptability of capacity models, computational efficiency,
monitoring mechanisms, support for distributed order promising, and
integration with other systems, are some important requirements
identified. |
|
Title: |
A
SUBSTRATE MODEL FOR GLOBAL GUIDANCE OF SOFTWARE AGENTS |
Author(s): |
Guy Gouardères,
Nicolas Guionnet |
Abstract: |
We try to understand how large groups of software agents
can be given the means to achieve global tasks, while their point of view
on the situation is only local (reduced to a neighbourhood.) To understand
the duality between local abilities and global constraints, we introduced
a formal model. We used it to evaluate the possibility of existence of an
absolute criteria, for a local agent, to detect global failure (in order
to change the situation). The study of a sample of examples led us to the
fact that such a criteria does not always exist. When it exists, it’s
often too global for local agents to apply (it demands too a large field
of view to be employed.) That’s why we left, for a moment, the sphere of
absolute criteria, to look for something more flexible. We propose a tool
of domain globalisation that is inspired by continuous physics phenomena:
If the domain is too partitioned, we can add to it, a propagation layer,
to let the agents access data concerning its global state. This layer can
be a pure simulation of wave or heat equations, or an exotic
generalisation. We applied the concept to a maze obstruction
problem. |
|
Title: |
APPLYING
CASE-BASED REASONING TO EMAIL RESPONSE |
Author(s): |
Luc Lamontagne
|
Abstract: |
In this paper, we describe a case-based reasoning approach
for the semi-automatic generation of responses to email messages. This
task poses some challenges from a case-based reasoning perspective
especially to the precision of the retrieval phase and the adaptation of
textual cases. We are currently developing an application for the Investor
relations domain. This paper discusses how some of the particularities of
the domain corpus, like the presence of multiple requests in the incoming
email messages, can be addressed by the insertion of natural language
processing techniques in different phases of the reasoning cycle. |
|
Title: |
THE
INOVATION PLANNING TASK FOR PRODUCTS AND SERVICES |
Author(s): |
Alfram Albuquerque,
Marcelo Barros, Agenor Martins, Edilson Ferneda |
Abstract: |
Innovation is crucial for business competitive
intelligence and knowledge-based society. In this context, companies use
to base their activities on the efficiency of their processes for
supporting innovation of prod-ucts and services. Knowledge-based systems
should leverage the innovation process and its planning by storing
internal and external user information. In this paper, the authors detail
this innovation process by presenting and discussing an architecture for
the task of user support oriented to the innovation planning process. The
proposed architecture is based on QFD – a methodology that translates the
client voice into engineering requisites for products and services. Our
methodological proposal increases efficiency on the base of the
integration of both knowledge-based processes (KBPs) and mechanical
processes (MPs) used to transform quality specification or requisites into
engineering requirements. |
|
Title: |
DECISIO:
A COLLABORATIVE DECISION SUPPORT SYSTEM FOR ENVIRONMENTAL
PLANNING |
Author(s): |
Julia Strauch,
Manuel de Castro, Jano de Souza |
Abstract: |
Environmental planning projects often face problems such
as: difficulties to manage spatial data as a component of the process,
lack of coordination of the different areas, difficulties of knowledge
access, badly defined decision processes, and absence of documentation of
the entire process and its relevant data. Our proposal is a web-based
system that provides a workflow tool to design and execute the decision
process and group decision support tools that help the decision makers in
finding similar solutions, analyzing and prioritizing alternatives and
helping the interaction among users. The main goals of the proposal are:
Document the environmental process and data, provide tools to support
collaboration, conflict management and alternative analysis and make
available previous successful and failure similar cases. These
functionalities have their human-computer interaction adapted to
incorporate spatial data manipulation and geo-reference. The tool is being
used in agro-meteorological projects with the purpose of improving the
effectiveness and efficiency of the decision process and its result,
maximize profit and preserving natural resources. |
|
Title: |
CLASSIFYING
DATABASES BY K-PROPAGATED SELF-ORGANIZING MAP |
Author(s): |
Takao Miura,
Taqlow Yanagida, Isamu Shioya |
Abstract: |
In this investigation, we discuss classifiers to databases
by means of Neural Network. Among others, we introduce k-propagated Self
Organizing Map (SOM) which involves learning mechanism of neighbors. And
we show the feasibility of this approach. Also we evaluate the tool from
the viewpoint of statistical tests. |
|
Title: |
MAKE
OR BUY EXPERT SYSTEM (MOBES): A KNOWLEDGE-BASED DECISION SUPPORT TOOL TO
MAXIMISE STRATEGIC ADVANTAGE |
Author(s): |
noornina dahlan,
ai pin lee, reginald theam kwooi see, teng hoon lau, eng
han gan |
Abstract: |
This paper presents a knowledge-based tool, which aids
strategic make or buy decisions that are key components in enhancing an
organization’s competitive position. Most companies have no firm basis for
evaluating the make or buy decision; thereby using inaccurate costing
analyses for sourcing strategies, which are directly responsible for the
flexibility, customer service quality, and the core competencies of an
organization. As a result, a prototype of the Make or Buy Expert System
(MOBES) with multi-attribute analytic capability is developed. The
proposed model comprises four main dimensions: identification and
weighting of performance category; analysing technical capability
category; comparison of retrieved internal and external technical
capability profiles, and analysis of supplier category. This model aims to
enable an organisation to enhance its competitiveness by improving its
decision making process as well as leveraging its key internal resources
to move further forward in its quest for excellence. |
|
Title: |
AGENT
TECHNOLOGY FOR DISTRIBUTED ORGANIZATIONAL MEMORIES: THE FRODO
PROJECT |
Author(s): |
Ludger
van Elst, Andreas Abecker, Ansgar Bernardi |
Abstract: |
Comprehensive approaches to knowledge management in modern
enterprise are confronted with scenarios which are heterogeneous,
distributed, and dynamic by nature. Pro-active satisfaction of information
needs across intra-organizational boundaries requires dynamic negotiation
of shared understanding and adaptive handling of changing and ad-hoc task
contexts. We present the notion of a Distributed Organizational Memory
(DOM) as a meta-information system with multiple ontology-based structures
and a workflow-based context representation. We argue that agent
technology offers the software basis which is necessary to realize DOM
systems. We sketch a comprehensive Framework for Distributed
Organizational Memories which enables the implementation of scalable DOM
solutions and supports the principles of agent-mediated knowledge
management. |
|
Title: |
USING
THE I.S. AS A (DIS)ORGANIZATION GAUGE |
Author(s): |
Pedro Araujo,
Pedro Mendes |
Abstract: |
The textile and garment industry in Portugal is undergoing
some struggles. In their origin is a lack of organization of many
companies. This situation, together with an increasing dynamics of the
products and the markets, considerably complicates decision-making and
information systems can be a precious aid. But contrary to academics,
managers must be shown evidence of the advantages of using information
technology. So, to help attain this objective, we propose the definition
of an index quantifying the level of disorganization of the productive
sector of the company. Continuously using the information system to
monitor this index allows managers to improve the performance of the
company's operations. |
|
Title: |
HELPING
USER TO DISCOVER ASSOCIATION RULES. A CASE IN SOIL COLOR AS AGGREGATION OF
OTHER SOIL PROPERTIES |
Author(s): |
Manuel Sanchez-Marañon,
Jose-Maria Serrano, Gabriel Delgado, Julio Calero,
Daniel Sanchez, Maria-Amparo Vila |
Abstract: |
As commercial and scientific databases size increases
dramatically with little control on the overall application of this huge
amount of data, knowledge discovery techniques are needed in order to
obtain relevant and useful information to be properly used later. Data
mining tools, as association rules and approximate dependencies, has been
proven as effective and useful when users are looking for implicit or
non-intuitive relations between data. The current and main disadvantage of
rule-extraction algorithms rests on the sometimes excessive number of
obtained results. Since human expert aid is needed in order to give an
interpretation to results, a very interesting task is to make easier the
expert's work. An user interface and a knowledge discovery management
system would provide a comfortable way to easily sort out rules, according
to their utility. An example of this necessity is shown in a case
involving soil color as aggregation of other soil properties and as a
interesting descriptor for soil-forming processes. |
|
Title: |
PRODUCTION
ACTIVITY CONTROL USING AUTONOMOUS AGENTS |
Author(s): |
Eric Gouardères,
Mahmoud Tchikou |
Abstract: |
The need of adaptability of production structures is
continuously increased due to decrease of product life cycle and increase
of the competition. The efficiency of a production system is now described
not only in term of time cycle, due date, inventory level, but also in
term of flexibility and reactivity in order to integrate the evolution of
the market. Current methods for real time control of production system do
not provide sufficient tools for an effective production activity control.
The origin of such a problem is at the level of existing control
structures. This work details the design of a production activity control
system based on distributed structure. The structure is based on the
distributed artificial intelligence concepts. After having introduced the
context and reasoning work, we describe the different parts of our
multi-agent model. Lastly, we illustrate this approach on a practical
example of production cell. |
|
Title: |
HUMAN
IRIS TEXTURE SEGMENTATION ALGORITHM BASED ON WAVELET THEORY |
Author(s): |
Taha El-Arief,
Nahla El-Haggar, M. Helal |
Abstract: |
Iris recognition is a new biometric technology that
exceptionally accurate that has stable and distinctive features for
personal identification. For iris classification it is important to
isolate the iris pattern by locating its inner (pupil) and outer (limbus)
boundaries. This paper presents a texture segmentation algorithm for
segmenting the Iris from the human eye in more accurate and efficient
manner. A quad tree wavelet transform is first constructed to extract the
texture feature. The fuzzy c-means (FCM) algorithm is then applied to the
quad tree with the coarse-to-fine approach. Finally, the results
demonstrate its potential usefulness. |
|
Title: |
AN
EXPERT SYSTEM FOR CREDIT MANAGEMENT FOLLOW-UP |
Author(s): |
Nevine Labib,
Ezzat Korany, Hamdy Latif, Mohamed Abderabu |
Abstract: |
Commercial risk assessment nowadays has become the major
concern of banks since they are faced with severe losses of unrecoverable
credit. The proposed system is an Expert System prototype for Credit
management follow-up. The system uses rule-based inference mechanism of
reasoning. The knowledge were obtained from Experts woring in six
commercial Egyptian banks. It starts following up the granted loan. If the
customer refrains from paying, it calculates his credit rating. If the
customer credit rating is bad, it analyzes the problem causes and reasons
and accordingly takes the suitable remedial action. When tested, the
system proved to be efficient. |
|
Title: |
APPLICATION
OF GROUP METHOD OF DATA HANDLING TO VIRTUAL ENVIRONMENT
SIMULATOR |
Author(s): |
Wataru SHIRAKI
|
Abstract: |
In this paper, we propose decision support system that
selects the most useful development plan for preservation of natural
environment and target species from two or more development plan. For the
purpose, after recognizing the environmental situation and the impact
among environmental factors where the species exist, we select a
sustainable development based on evaluation and prediction of environment
assessment by reconstructing dynamics in computer simulation. Then, we
present hybrid system using artificial life technology such as cellular
automaton and group method of data handling, which can apply to
environmental assessment. From results of numerical example, proposal
system approximates coefficients with sufficient accuracy if the structure
of a model is known, it was also shown that near future dynamics can be
predicted, even if the structure of a model is unknown. |
|
Title: |
AN
EFFICIENT CLASSIFICATION AND IMAGE RETRIEVAL ALGORITHM BASED ON ROUGH SET
THEORY |
Author(s): |
Jafar Mohammed,
Aboul Ella Hassanien |
Abstract: |
: With an enormous amount of image data stored in
databases and data warehouses, it is increasingly important to develop
powerful tools for analysis of such data and mining interesting knowledge
from it. In this paper, we study the classification problem of image
databases and provide an algorithm for classification and retrieval image
data in the context of Rough Set methodology. We present an efficient
distance function called quadratic which works more efficiently with
retrieval image data. We also demonstrate that by choosing the useful
subset of rules based on simple decision table, the algorithm have high
accuracy for classification. |
|
Title: |
USING
SPECIALIZED KNOWLEDGE IN AUTOMATED WEB DOCUMENT SUMMARIZATION |
Author(s): |
Zhiping Zheng
|
Abstract: |
Automated text summarization is a natural language
processing task to generate short, concise, and comprehensive descriptions
of essential content of documents. This paper is going to describe some
new features in a real-time automated web document summarization system
used in Seven Tones Search Engine, a search engine specialized in
linguistics and languages. The main feature of this system is to use
algorithms designed specifically for Web pages in a specific knowledge
domain to improve the quality of summarization. It also considers the
unique characteristics of search engines. In special, linguistics features
should be very important to linguistics document. The documents are
assumed either HTML or plain text. A good HTML parser will affect
summarization quality very much although it is not a part of summarization
algorithm. |
|
Title: |
A
NEW APPROCH OF DATA MINING |
Author(s): |
Stéphane Prost,
Claude Petit |
Abstract: |
This paper describe a trajectories classification
algorithm ( each trajectory is defined by a finite number of values), it
gives for each class of trajectories a characteristic trajectory: the
meta-trajectory. The pathological trajectories are removed by the
algorithm. Classes are built by an ascendant method. Two classes are
built, then three and so on, a partition containing n classes allow to
built a partition with n+1 classes. For each class a meta-trajectory is
determined ( for example the gravity centre). The number of classes
depends on the minimum number of trajectory by classes allowed and a
parameter given by the user, which is compared with the inter-classes
inertia gain, other dispersion may be chosen. |
|
Title: |
EXPERIENCE
MANAGEMENT IN THE WORK OF PUBLIC ORGANIZATIONS: THE PELLUCID
PROJECT |
Author(s): |
Simon LAMBERT,
Sabine DELAITRE, Gianni VIANO, Simona STRINGA |
Abstract: |
One of the major issues in knowledge management for public
organisations is organisational mobility of employees, that is the
continual movement of staff between departments and units. As a
consequence of this, the capture, capitalisation and reuse of experience
become very important. In the PELLUCID project, three general scenarios
have been identified from studies of the pilot application cases. They are
contact management, document management and critical timing management.
These scenarios are outlined, and a corresponding approach to experience
formalisation is described. Requirements are also set out on the technical
solution able to support experience management |
|
Title: |
USING
GRAMMATICAL EVOLUTION TO DESIGN CURVES WITH A GIVEN FRACTAL
DIMENSION |
Author(s): |
Manuel Alfonseca,
Alfonso Ortega, Abdel Dalhoum |
Abstract: |
Lindenmayer Grammars have been applied to represent
fractal curves. In this work, Grammatical Evolution is used to
automatically generate and evolve Lindenmayer Grammars that represent
curves with a fractal dimension that approximates a pre-defined required
value. For many dimensions, this is a non trivial task to be performed
manually. The procedure used parallels biological evolution, acting
through three different levels: a genotype (a vector of integers) subject
to random modifications in different generations), a protein-like
intermediate level (a Lindenmayer Grammar with a single rule, generated
from the genotype by applying a transformation algorithm) and a phenotype
(the fractal curve). |
|
Title: |
DETECTION
OF CARDIAC ARRHYTHMIAS BY NEURAL NETWORKS |
Author(s): |
Noureddine Belgacem,
F. Reguig, M. Chikh |
Abstract: |
The classification of heart beats is important for
automated arrhythmia monitoring devices. The study describes a neural
classifier for the identification ad detection of cardiac arrhythmias in
surface (Electrocardiogram) ECGs. Traditional features for the
classification task are extracted by analyzing the heart rate and
morphology of QRS complex and P wave of the ECG signal. The performance of
the classifier is evaluated on the MIT-BIH database. The method achieved a
sensitivity of 94.60% and a specificity of 96.49% in discrimination of six
classes. |
|
Area 3 - INFORMATION SYSTEMS ANALYSIS AND SPECIFICATION
Title: |
CODE
GENERATION FOR DISTRIBUTED SYSTEMS |
Author(s): |
Ralf Gitzel,
Markus Aleksy |
Abstract: |
Due to the complexity of distributed code as opposed to
the easy way the corresponding designs can be described graphically,
interest in code generators which create applications based on abstract
system descriptions, is high. An indicator for this are the many
commercial products. This paper aims at exploring the theoretical
foundations of code generation for distributed systems with regard to the
data structures and template language syntax. Several existing approaches
are analysed and a new hybrid-form data structure is proposed. The goal,
that this paper aims at, is an adaptable, middleware-independent way to
produce software with minimal human code. |
|
Title: |
SOFTWARE
DEVELOPMENT GUIDED BY MODELS - THE XIS UML PROFILE |
Author(s): |
Miguel Luz,
Alberto Silva |
Abstract: |
The UML is used to detail high level software
specifications that will be interpolated for XMI and XIS (XML Information
Systems) as interchange formats based on XML. UML and XML are expected to
be the next generation of modeling and data interchange standards
respectively. In this paper, we describe the UML Profile for XIS
architecture as a proposal for software development guided by UML models.
The XIS system is based on a multi-phase generative programming approach,
starting from high-level UML models till software artifacts (such as Java
code and SQL scripts), passing through different representations, namely
OMG’s XMI, and our (XIS) specific XML vocabulary. The main contribute of
this paper is the overview of the XIS system and the proposal and
discussion of the XIS UML profile. |
|
Title: |
KEY
ISSUES IN INFORMATION SYSTEMS AND SOFTWARE ENGINEERING - VIEWS FROM A
JOINT NETWORK OF PRACTITIONERS AND ACADEMICS |
Author(s): |
M. RAMAGE,
D. TARGETT, Kecheng Liu, R. HARRISON, D. AVISON,
K. BENNETT, R. BISHOP |
Abstract: |
SEISN (Software Engineering and Information Systems
Network), a research project supported by the British research council
EPSRC, aims to promote the understanding of the two research communities
and practitioners. The network focuses on the exchange of ideas and
enables these communities to clarify their beliefs and present
experiences, findings and views. This paper summaries the work of this
research network, and investigates where there is common ground between
the IS and SE communities and practitioners and where differences remain.
With discussions on the key issues, the paper shows the future directions
of research in software engineering and information systems. |
|
Title: |
SUPPORTING
DELIBERATION PROCESS MECHANISM IN SOFTWARE SYSTEMS DEVELOPMENT |
Author(s): |
OSMAN EBRAHIM,
RANAI ELGOHARY, AHMED HAMAD |
Abstract: |
A model for providing automated support for the
deliberation process inherent in the software requirements engineering is
proposed. The model provides the representation and formal mechanisms that
supports stakeholders in their evaluating the available alternatives to
choose among them, based on specified criteria that precedes the final
decision. The same mechanism is used to quantify and formalize independent
judgment of each stakeholder and then combines these individual judgments
into a decision that express the group final decisions. The database
capable of representing and encompassing this huge amount of the process
knowledge in a way close to the conceptual data model has been designed.
The model will also provide a representation mechanism for capturing and
recording design rationale. This can assist in design replay or
justification of decisions as well as an important history trails for
management references. The developed model is applied and validated on a
software requirements engineering on a case study at Air Traffic Control
system (ATC). |
|
Title: |
ANALYSIS
ON RELATION BETWEEN SERVICE PARAMETERS FOR SERVICE LEVEL MANAGEMENT AND
SYSTEM UTILIZATION |
Author(s): |
Norihisa Komoda,
Shoji Konno, Masaharu Akatsu |
Abstract: |
Accompanying the rise of IT service providers such as ASP
(Application Service Provider) and iDC (internet Data Center), it becomes
popular to define the quality of information systems as SLA (Service Level
Agreement). The true goal of SLA is to guarantee a valuable service to
users. Hence, it is desirable to identify service parameters that are
highly related to user satisfaction. The service parameters concerning
every system should be different. We focus attention on system utilization
to select the parameters. Our expectation is the following hypothesis; by
investigating the characteristic on system utilization, we can
statistically predict the parameters that are the most critical to the
satisfaction of the system users. In this paper, we examine parameters for
availability and responsiveness, which are known as two major factors in
SLA. First, we provide three parameters for availability and
responsiveness, respectively. Next, we prepare a questionnaire about
system utilization. To analyze the relation between service parameters and
system utilization, we had several experienced system engineers answer the
questions for each system they had developed. Quantification theory type
II is applied for the analysis and validity of our hypothesis is
demonstrated. We also clarify characteristics on system that emphasizes
each service parameter. |
|
Title: |
ALIGNING
AN ENTERPRISE SYSTEM WITH ENTERPRISE REQUIREMENTS: AN ITERATIVE
PROCESS |
Author(s): |
Pnina Soffer
|
Abstract: |
Aligning an off-the-shelf software package with the
business processes of the enterprise implementing it is one of the main
problems in the implementation of enterprise systems. The paper proposes
an iterative alignment process, which takes a requirement-driven approach.
It benefits from reusing business process design without being restricted
by predefined solutions and criteria. The process employs an automated
matching between a model of the enterprise requirements and a model of the
enterprise system capabilities. It identifies possible matches between the
two models and evaluates the gaps between them despite differences in
their completeness and detail level. Thus it provides the enterprise with
a set of feasible combinations of requirements that can be satisfied by
the system as a basis for making implementation decisions. The automated
matching is applied iteratively, until a satisfactory solution is found.
Object Process Methodology (OPM) is applied for modeling both the system
and the enterprise requirements, which are inputs for the automated
matching. The alignment process has been tested in an experimental study,
whose encouraging results demonstrate its ability to provide a
satisfactory solution to the alignment problem. |
|
Title: |
TRACKING
BUSINESS RULE EVOLUTION TO SUPPORT IS MAINTENANCE |
Author(s): |
Marko Bajec
|
Abstract: |
Business rules describe how organisations are doing
business. Their value has also been recognised within the information
system (IS) domain, mostly because of their ability to make applications
flexible and amenable to change. In this paper we argue that business
rules can be used as a link between organisations and their ISs. We show
that business rules originate in organisations and that many business
rules are explicitly or implicitly captured in enterprise models. We
advocate, based on research work, that if business rules are managed in an
appropriate manner they can help keeping IS aligned and consistent with
the business environment. In the paper we propose a business rule
management scenario for managing business rules from an organisational
perspective. The scenario recognises business rule management as an
interface between enterprise modelling and IS development and maintenance.
|
|
Title: |
DESIGN
AND REALIZATION OF POWER PLANT SUPERVISORY INFORMATION SYSTEM (SIS) BASED
ON INFI 90 |
Author(s): |
Guozhong Zhang,
Zhen Ye |
Abstract: |
To improve the management level of power plants and to
adapt to the requirements of market-oriented reform for electric
enterprises, a method of design and realization of a Supervisory
Information System (SIS) based on INFI 90 DCS for manufacturing management
in power plant is put forward in this paper. By adding a CIU to the INFI
net, the real-time data of manufacture process is retrieved to the
historical data platform through interface PCs, fibers and switches. The
whole system used OpsCon as interface driver, iHistorian as the historical
data platform, iFIX as the configuration software, and infoAgent as a tool
for assistant decision-making such as on-line historical data analysis,
equipment status monitoring and malfunction diagnosis, equipment
reliability and life management etc. Practice shows that the SIS makes
full use of resources of DCS and MIS, forms a synthetic automation system
integrated by DCS, SIS, and MIS, and realizes an automatic control
covering the whole process of electric production. |
|
Title: |
TOWARDS
A DEFINITION OF THE KEY-PROBLEMS IN INFORMATION SYSTEM EVOLUTION -
FORMULATING PROBLEMS TO BETTER ADDRESS INFORMATION SYSTEM
PROJECTS |
Author(s): |
Virginie Goepp,
François Kiefer |
Abstract: |
Through up the years, a lot of methods and approaches were
proposed in the information system design (ISD) field. In spite of this
variety and number of propositions, over 80 % of information system
projects fail (Clancy 1995). On the one hand, this situation seems to be
very surprising but on the other hand, this diversity of works let the
research area in a state of fragmentation. A basic problem is the lack of
consensus on the information system (IS) notion. However, according to
Alter, its comprehension is essential to better understand project
failures. So, we come back up to this notion and show that IS have to
fulfil contradictory roles linked to individual and collective aspects of
information. This contradiction is the starting point for establishing a
key-problem framework of IS in general. Indeed, the contradiction notion
is an integrating part of TRIZ, Cyrillic acronym for "Theory of Inventive
Problem Solving", which is efficient for better formalize and address
problems in technical system design. So, we analyse its potential
contributions for developing modular and contingent IS project approaches,
which are project success factors. Then, we apply the TRIZ approach on our
first contradiction in order to obtain the key-problem framework. This
one, based on three contradiction classes, is developed and presented.
Each class of contradiction is linked with the semiotics features of
information and enables to formalize the intrinsic problems of information
systems. The potential applications of such a framework are also
discussed. |
|
Title: |
AN
ENVIRONMENT FOR SOFTWARE DEVELOPMENT BASED ON A DISTRIBUTED COLLABORATIVE
MODEL |
Author(s): |
Angélica de
Antonio, Marco Villalobos |
Abstract: |
In this paper, we present a model for collaborative
software design and an environment, called Sinergia, that is being
constructed based on this model. We describe the different concepts and
components of the proposed model, as well as the functional features that
the environment under development includes. The Sinergia tool uses a
combination of technologies, such as distributed CORBA objects, Java
servlets or relational databases, that make it useful in the context of a
distributed multidisciplinary software development team. |
|
Title: |
DEONTIC
CONSTRAINTS: FROM UML CLASS DIAGRAM TO RELATIONAL MODEL |
Author(s): |
Pedro Ramos
|
Abstract: |
Sometimes, because of one atypical situation, an important
mandatory association between classes in a UML Class Diagram must be
replaced by an optional one. That semantic and functionality
impoverishment happens because the mandatory constraint must have a
boolean value. In this paper we propose the introduction of a deontic
constraint in the UML Class Diagram and its automatic repercussion in the
correspondent Relational Model. The deontic constraint allows the formal
representations of requirements that ideally should always be fulfilled,
but that can be violated in atypical situations. If the violable
requirement is explicitly represented it is possible to maintain both the
requirement and its violation and, consequently, recur to monitoring
procedures for violation warnings. We present our proposal in the general
context of automatically mapping from object models into relational ones.
We adopt a formal approach, based on predicate calculus, because, apart
its soundness properties, it is an easy and understandable way to
integrate both models and the transposition rules. |
|
Title: |
STRUCTURAL
CONFLICT AVOIDANCE IN COLLABORATIVE ONTOLOGY ENGINEERING |
Author(s): |
Ziv Hellman,
Ziv Hellman |
Abstract: |
Given the increasing importance of ontologies in
enterprise settings, mechanisms enabling users working simultaneously to
edit and engineer ontologies in a collaborative environment are required.
The challenges in preventing structural conflicts from arising from
simultaneous user editing of ontologies are not trivial, given the high
level of dependencies between concepts in ontologies. In this paper we
identify and classify these dependencies. Sophisticated ontology locking
mechanisms based on a graph depiction of the dependencies that are
sufficient for preventing structural conflicts arising in collaborative
settings are proposed. Applications of this research to the Semantic Web
are also considered. |
|
Title: |
BEYOND
END USERS COMPUTING |
Author(s): |
Michael Heng
|
Abstract: |
Three central problems in the development, introduction,
maintenance, operation and up-grading of computer-based information
systems are (1) users and IS designers fail to understand each other well,
(2) it takes too long to introduce IS into organizations, and (3)
maintenance requires dedication associated with a sense of ownership, and
upgrading requires continuous attention and in-depth knowledge of the
business and potentials of IT. These three problems are often tackled
separately. It is argued that based on the recent advances in IS
development tools, environment and method as well as increased IT
literacy, a new way of handling all these problems in an integrated way is
possible in certain type of organizational setting. The key idea is to
form a team from the users' group in a big organization who are computer
enthusiasts to function as the core of a continuing team to take over the
responsibilities of developing, introducing, maintaining, and upgrading
the IS. The approach is essentially a synthesis of the structured way of
building systems and the end user computing. Some problems related to this
approach would also be surveyed.The approach is also very much in the
spirit of the idea of growing systems in emergent organizations as
propounded by Treux, Baskerville and Klein (1999) in their CACM paper.
|
|
Title: |
USING
LOTOS IN WORKFLOW SPECIFICATION |
Author(s): |
Alessandro Longheu
|
Abstract: |
Complexity of business processes is getting higher and
higher, due to the rapid evolution of market and technologies, and to the
reduced time-to-market for new products. Moreover, it is also essential to
check workflow (WF) correctness, as well as to guarantee specific business
rules. This can be achieved using specific tools within workflow
management systems (WFMS), but a formal approach (mathematical-based) is a
more effective methodology to guarantee workflow requirements. Formal
description techniques (FDT) based on process algebra allow both to
formally describe WF at any level of abstraction, and to formally verify
properties as correctness and business rules. In this paper, we apply FDT
in production workflows specification using the LOTOS language. In
particular, we first show how most recurrent WF patterns can be described
in LOTOS, then showing an application example of how WFs can be defined
based on LOTOS patterns, allowing a more effective verification of
correctness / business rules. |
|
Title: |
ANALYSIS
OF LEVEL OF INVOLVEMENT OF SIX BEST PRACTICES OF RUP IN OOSP |
Author(s): |
Muhammad Saeed,
Faheem Ahmed |
Abstract: |
No one in the software industry can deny the overwhelming
importance of software process model to increase the quality of software
product. It is a unanimous consent that better quality software can be
produced through well defined and refined process model. Rational Unified
Process (RUP) has emerged as a leading software process model, which has
incorporated the best among the best industry practices in it. The
adaptation of these best practices has almost fulfilled the requirements
of an ideal software process model, which should be based upon the actual
practices and theoretical concerns of software engineering in the
development of software product. In this paper we analyzes RUP with some
other object oriented process models like Object-Oriented Software Process
(OOSP) with respect to level of incorporation of six best industrial
proven practices i.e., iterative development, manage requirements, use
component architecture, model visually, verify quality and control
changes. The analysis will give us a true picture of involvement of six
best practices in OOSP, which will ultimate enables us to perform a
comparative analysis of the two-process model with respect to performance
and quality. |
|
Title: |
A
TEMPORAL REASONING APPROACH OF COMMUNICATION BASED WORKFLOW
MODELLING |
Author(s): |
ANDRES AGUAYO,
ANTINO CARRILLO, SERGIO GALVEZ, ANTONIO GUEVARA, JOSE
L CARO |
Abstract: |
Implementation of formal techniques to aid the design and
implementation of workflow management systems (WfMS) is still required. We
believe that formal methods can be applied in the field of properties
demonstration of a workflow specification. This paper develops a
formalization of the workflow paradigm based on communication (speech-act
theory) by using a temporal logic, namely, the Temporal Logic of Actions
(TLA). This formalization provides the basic theoretical foundation for
the automated demonstration of the properties of a workflow map, its
simulation, and fine-tuning by managers. |
|
Title: |
REFACTORING
USE CASE MODELS: A CASE STUDY |
Author(s): |
Gregory Butler
|
Abstract: |
Refactoring is a behavior-preserving program
trasnformation. Our research shows that refactoring as a concept can be
broadened to apply to use case models to improve their understandability,
changeability, reusability and traceability. In this paper we describe a
metamodel for use case modeling in detail. Based on this metamodel we
define and categorize a list of use case refactorings. Then we present a
case study to illustrate the practical use of these refactorings. Several
examples are described to show different views on refactoring use case
models. |
|
Title: |
ON
THE SYSTEMIC ENTERPRISE ARCHITECTURE METHODOLOGY (SEAM) |
Author(s): |
Alain
Wegmann |
Abstract: |
For companies to be more competitive, they need to align
their business and IT resources. Enterprise Architecture is the discipline
whose purpose is to align more effectively the strategies of enterprises
together with their processes and their resources (business and IT).
Enterprise architecture is complex because it involves different types of
practitioners with different goals and practices during the lifecycle of
the required changes. Enterprise Architecture can be seen as an art and is
largely based on experience. But Enterprise Architecture does not have
strong theoretical foundations. As a consequence, it is difficult to
teach, difficult to apply, and does not have true computer-supported tool.
This lack of tool is unfortunate as such tools would make the discipline
much more practical. This paper presents how system sciences, by defining
the concept of the systemic paradigm, can provide these theoretical
foundations. It then gives a concrete example of the application of these
foundations by presenting the SEAM paradigm. With the systemic paradigm,
the enterprise architects can improve their understanding of the existing
methodologies, and thus find explanations for the practical problems they
encounter. With the SEAM paradigm, architects can use a methodology that
alleviates most of these practical problems and can be supported by a
tool. |
|
Title: |
THE
RELEVANCE OF A GLOBAL ACCOUNTING MODEL IN MULTI-SITE ERP
IMPLEMENTATIONS |
Author(s): |
Ksenca Bokovec,
Talib Damij |
Abstract: |
ERP systems and their processes are cross-functional. They
transform companies' practice from traditional functional and local
oriented environments to global operations, where they integrate
functions, processes and locations. They can support company-specific
processes in the framework of globally defined organisational structures
and procedures if properly implemented. This paper seeks to contribute to
the area of multi-site ERP implementations. A case study from several
companies in a large retail corporation is presented, focusing on the
global accounting model from the perspective of an ERP implementation
project. This case study analyses the most important elements of a
globally designed financial and management accounting model and their
'translation' to the structures and processes of the ERP system. Moreover,
It demonstrates the importance of the application methodology in early
project phases. Central standardisation and maintenance issues of the
global accounting model are also outlined. |
|
Title: |
IMPLEMENTING
USER CENTRED PARTNERSHIP DESIGN - CHANGE IN ATTITUDE MADE A
DIFFERENCE |
Author(s): |
Paul Maj,
Gurpreet Kohli, Anuradha Sutharshan |
Abstract: |
IT project success depends upon a number of factors. There
are many in the information systems discipline who believe that user
participation is necessary for successful development. This paper is
primarily concerned with end users and implements a method of
incorporating end user participation in all the phases of an IT project.
The proposed qualitative, case-based approach aims to achieve high level
of usability of the delivered system and to make sure that skills and
knowledge of the team are better used. This approach enables users to
better understand and accept the new systems as well as ensuring that the
final deliverable is really what the users required. Significantly this
new method required a change in attitude and perception of not only the
end users but also the IT development staff. This process involves
studying the user tasks better, make users define what they want, make
regular and early prototypes of the user interface, and user involvement
from start until the end of the project. The aim of this paper was to
identify the user centred factors involved in different stages of the
project and to understand how the steps involved could make a positive
difference to an organisation. This approach was implemented and evaluated
in a local government agency in Western Australia. The results were
impressive. The suggested User oriented approach was then implemented in 3
other projects in the same organisation and the approach had made a
positive difference. |
|
Title: |
A
THREE-DIMENSIONAL REQUIREMENTS ELICITATION AND MANAGEMENT DECISION-MAKING
SCHEME FOR THE DEVELOPMENT OF NEW SOFTWARE COMPONENTS |
Author(s): |
Andreas Andreou,
Andreas Zografos, George Papadopulos |
Abstract: |
Requirements analysis and general management issues within
the development process of new software components are addressed in this
paper, focusing on factors that result from requirements elicitation and
significantly affect management decisions and development activities. A
new methodology performing a certain form of requirements identification
and collection prior to developing new software components is proposed and
demonstrated, the essence of which lays on a three-entity model that
describes the relationship between different types of component
stakeholders: Developers, reusers and end-users. The model is supported by
a set of critical factors analysed in the context of three main directions
that orient the production of a new component, that is, the generality of
the services offered, the management approach and the characteristics of
the targeted market. The investigation of the three directions produces
critical success factors that are closely connected and interdependent.
Further analysis of the significance of each factor according to the
priorities set by component developers can provide a detail picture of
potential management implications during the development process and more
importantly can support management decisions related to if and how
development should proceed.
|
|
Title: |
DEFINING
STABILITY FOR COMPONENT INTEGRATION ASSESSMENT |
Author(s): |
Alejandra Cechich,
Mario Piattini |
Abstract: |
The use of commercial of-the-shelf (COTS) products as
elements of larger systems is becoming increasingly commonplace.
Component-Based Software Development is focused on assembling previously
existing components (COTS and other non-developmental items) into larger
systems, and migrating existing systems toward component approaches.
Ideally, most of the application developer’s time is spent integrating
components. We present an approach that can be used in the process of
establishing component integration’s quality as an important field to
resolving CBS quality problems – problems ranging from CBS quality
definition, measurement, analysis, and improvement to tools, methods and
processes. In this paper, we describe and illustrate the use of the first
phase of our approach to determine relevant perturbations when
incorporating a COTS component into given software. |
|
Title: |
AUGMENTATION
OF VIRTUAL OBJECT TO REAL ENVIRONMENT |
Author(s): |
Felix Kulakov
|
Abstract: |
A problem of immersion of an arbitrary
computer-synthesized virtual body into a real environment at considerable
distance from observer is considered. The problem under discussion refers
to so-called Augmented Reality which is a rapidly developing trend within
Virtual Reality. A virtual body in this case is an augmentation to the
reality. The problem has "visual" and "tactile-force" aspects. Advanced
approaches to realization of these aspects of immersion are
proposed. |
|
Title: |
SOFTWARE
ENGINEERING ENVIRONMENT FOR BUSINESS INFORMATION SYSTEMS |
Author(s): |
Alar Raabe
|
Abstract: |
There is a growing need to make business information
systems development cycle shorter, and independent of underlying
technologies. Model driven synthesis of software offers solutions to these
problems. In the article we describe a set of tools and methods,
applicable for synthesizing business software from technology independent
models. This method and these tools are distinguished by the use of
extended meta-models, which embody knowledge of problem domain and target
software architecture of synthesized software system, by the use of model
conversion rules described using the combined meta-model, and by the use
of reference models of problem domains and sub-domains, which are combined
and extended during the construction of descriptions of software system.
The difference of our method from other domain specific methods is the
separate step of solution domain analysis and the use of meta-model
extensions. Work presented in the article has been done in the context of
developing product-line architecture for insurance applications. |
|
Title: |
ANALYSING
SECURITY REQUIREMENTS OF INFORMATION SYSTEMS USING TROPOS |
Author(s): |
Abdullah Gani,
Gordon Manson, Paolo Giorgini, Haralambos Mouratidis
|
Abstract: |
Security is an important issue when developing complex
information systems, however very little work has been done in integrating
security concerns during the analysis of information systems. Current
methodologies fail to adequately integrate security and systems
engineering, basically because they lack concepts and models as well as a
systematic approach towards security. We believe that security should be
considered during the whole development process and it should be defined
together with the requirements specification. This paper introduces
extensions to the Tropos methodology to accommodate security. A
description of new concepts is given along with an explanation of how
these concepts are integrated to the current stages of Tropos. The above
is illustrated using an agent-based health and social care information
system as a case study. |
|
Title: |
CUSTOMIZING
WEB-BASED SYSTEMS WITH OBJECT-ORIENTED VIEWS |
Author(s): |
Markus Schett,
Renate Motschnig-Pitrik |
Abstract: |
Although views have proved their place in relational data
models, their role in customizing object-oriented (OO) systems has been
severely underestimated. This phenomenon occurs despite the fact that
views in the OO paradigm can be designed such that their functionality by
far exceeds that of their relational cousins. Based on research in OO
databases and on the Viewpoint Abstraction, the purpose of this paper is
to integrate views into UML, to sketch the implementation of the tool
RoseView, and to illustrate applications of views in web-based systems. We
argue that designing system increments or adaptations as view contexts
allows for full-fledged customized system versions without ever affecting
the code of the original application, meaning significant savings in
maintenance. Further, we introduce RoseView, a tool implemented to make
views available in UML and thus to extend OO languages by an essential
abstraction dimension. |
|
Title: |
AN
XML BASED ADMINISTRATION METHOD ON ROLE-BASED ACCESS CONTROL IN THE
ENTERPRISE ENVIRONMENT |
Author(s): |
Chang
N. Zhang, Chang Zhang |
Abstract: |
In the distributed computing environments, users would
like to share resources and communicate with each other to perform their
jobs more efficiently. It is important to keep resources and the
information integrity from the unexpected use by unauthorized users.
Therefore, there is a strong demand on the access control of distributed
shared resources in the past few years. Role-Based-Access-Control (RBAC)
has been introduced and has offered a powerful means of specifying access
control decisions. In this paper, we propose an object-oriented RBAC model
for distributed system (ORBAC), it efficiently represents the real world.
Though ORBAC is a good model, administration of ORBAC including building
and maintaining access control information remains a difficult problem.
This paper describes practical method that can be employed in a
distributed system for managing security policies using extensible Markup
Language (XML). Based on the XML ORBAC security policy, an intelligent
role assignment backtracking algorithm is also presented, the computation
complexity of the algorithm is O (N) where N is the number of roles in the
user’s authorized role set. |
|
Title: |
DERIVING
USE CASES FROM BUSINESS PROCESSES, THE ADVANTAGES OF DEMO |
Author(s): |
Boris Shishkov, Jan L.G. Dietz
|
Abstract: |
The mismatch between the business requirements and the
actual functionality of the delivered software application is considered
to be a crucial problem in modern software development. Solving this
problem means to find out how to consistently place the software
specification model on a previously developed business process model. If
considering in particular the UML-based software design, we need to answer
in this regard a fundamental question, namely – how to find all relevant
use cases, based on sound business process modeling? Adopting the business
process modeling as a basis for identification of use cases has been
studied from three perspectives – it has been studied how use cases could
be derived based on DEMO, Semiotic and Petri Net business process models.
The goal of the current paper is, by considering the mentioned achieved
results, to study and analyze the strengths of DEMO concerning the
derivation of use cases. This could be helpful not only for the
investigation of DEMO but also for the further activities directed towards
finding out the most appropriate way(s) of identifying use cases from
business processes. |
|
Title: |
REQUIREMENTS
ENGINEERING VERSUS LANGUAGE/ACTION PERSPECTIVE: DIFFERENT FACETS AND
POSSIBLE CONTRIBUTION |
Author(s): |
Joseph Barjis,
Tychon Galatonov |
Abstract: |
In today’s Requirements Engineering (RE) it’s often
assumed that the users of the system-to-be understand well enough their
business needs and the overall goals of the system, therefore the data
provided by them is of utmost importance for engineering system
requirements; this strategy is sometimes called the “waiter strategy”.
While it is often justified, there are approaches that question the
validity of this “waiter strategy”. One of them is the Language/Action
Perspective. The Language/Action Perspective (hereinafter LAP) is an
approach to communication analysis in organisational/business systems,
i.e. social systems where there exists information interchange between its
components, human beings and machines (collectively called actors), with
the ultimate goal of fulfilling a mission of the organisation. One of the
features of LAP is that in distinction to the “waiter strategy” approaches
it assumes that it is the deeds the actors perform that are of crucial
importance for understanding the nature of the processes in the system.
This paper presents an overview of some results as well as a new possible
approach to RE using LAP; the following methods are taken into
consideration: DEMO (Dynamic Essential Modelling of Organisations)
methodology, the Semiotics approach and Petri nets. |
|
Title: |
PRESCRIBED
IMPACTS AND IMPACTS STIPULATED FOR ICS |
Author(s): |
Virginie govaere
|
Abstract: |
The aim of this work is to present the consequences of
introducing communication and information systems (ICS) into a company for
the users or the organizations users. Consequently, this introduction want
to inform and to warn the designer of new technologies of the impacts of
their achievements. With the ICS, the exchange of information in a
relationship is never neutral, its ability to circulate is never natural,
and the fact of being able to exchange it, whatever the means and media
used or these quality, has no predictive value concerning the real
exchanges. Thus, an analysis of the applications in technical terms
(performances and functionalities available) is insufficient. Indeed, the
taking into account of the context in the broad sense is essential to
determine their real performance. This structuring aims at bring out a
difference between the performance offered by the designers of ICS and
that observed in real situation. |
|
Title: |
AN
INNOVATIVE APPROACH TO WIRELESS APPLICATIONS DEVELOPMENT:AN EXPLORATION OF
PRACTISE |
Author(s): |
Phillip
Olla |
Abstract: |
Due to the developing of enabling technologies such as
mobile data networks and various types of affordable mobile devices,
mobile computing has become widely accepted and applied for both consumer
and business initiatives, it is fuelling the new trend of information
systems development. There is evidence that the profile of systems
development on innovative projects is very different from that faced in
the past when the systems development methodologies were first promoted
(Sawyer, 2001). Therefore there is the need to move from the past which
has documented problems ‘Software crisis’ (Hoch, Roeding, Purket, &
Lindner, 2000) to the future, by deriving new methodological approaches
more appropriate to the needs of the current development environment
(Fitzgerald, 2000). This paper used Action Research to study an
organisation called the Mobile Application Development and Integration
Centre, which created an innovative approach to develop and deploy
wireless applications produced by independent third parties |
|
Title: |
SPECIFYING
A KNOWLEDGE MANAGEMENT SOLUTION FOR THE CONSTRUCTION INDUSTRY:THE E-COGNOS
PROJECT |
Author(s): |
Yacine Rezgui,
Matthew Wetherill, Abdul Samad Kazi |
Abstract: |
The paper focuses upon the contribution which adequate use
of the latest development in Information and Communication Technologies
can make to the enhancement, development and improvement of professional
expertise in the Construction domain. The paper is based on the e-COGNOS
project, which aims at specifying and developing an open model-based
infrastructure and a set of tools that promote consistent knowledge
management within collaborative construction environments. The specified
solution emerged from a comprehensive analysis of the business and
information / knowledge management practices of the project end-users, and
makes use of a Construction specific ontology that is used as a basis for
specifying adaptive mechanisms that can organise documents according to
their contents and interdependencies, while maintaining their overall
consistency. The e-Cognos web-based infrastructure will include services,
which allow the creation, capture, indexing, retrieval and dissemination
of knowledge. It also promotes the integration of third-party services,
including proprietary tools. The e-COGNOS approach will be tested and
evaluated through a series of field trials. This will be followed by the
delivery of business recommendations regarding the deployment of e-COGNOS
in the construction sector. The research is ongoing and supported by the
European Commission under the IST programme – Key Action II |
|
Title: |
RELATIONSHIP
SUPPORT IN OBJECT MODELS |
Author(s): |
Mohamed Dahchour,
Alain Pirotte |
Abstract: |
Relationships play a central role in information modeling.
However, most object models (programming languages and database systems)do
not provide a construct to deal with them as autonomous units. They merely
treat them as pointer-valued attributes and therefore confine them to a
second-class status. The paper defines the generic semantics of
relationships, addresses a set of requirements to be satisfied to properly
manage all kinds of relationships, surveys existing techniques for
representing relationships in object models, and compares them to each
others according to whether they satisfy the relationship requirements.
|
|
Title: |
OPCAT
– A BIMODAL CASE TOOL FOR OBJECT-PROCESS BASED SYSTEM
DEVELOPMENT |
Author(s): |
Arnon Sturm,
Iris Reinhartz-Berger, Dov Dori |
Abstract: |
CASE tools have spread at a lower pace than expected. The
main reasons for this are their limited support of a particular method,
high cost, lack of measurable returns, and unrealistic user expectations.
Although many CASE tools implement familiar methods, their model checking
and simulation capabilities are limited, if not inexistent, and the syntax
and semantics of their graphic notations may not be clear to novice users.
Object-Process CASE Tool (OPCAT), which supports system development using
Object-Process Methodology, meets the challenges of next generation CASE
tools by providing a complete integrated software and system development
environment. Based on two human cognition principles, OPCAT enables
balanced modeling of the structural and behavioral aspects of systems in a
single model through a bimodal visual-lingual representation. Due to this
intuitive dual notation, the resulting model is comprehensible to both
domain experts and system architects engaged in the development process.
Due to its formality, it OPCAT also provides a solid basis for
implementation generation and an advanced simulation tool, which animates
system objects, processes, and states in a balanced way, enabling a
complete simulation of system structure and behavior. This paper presents
OPCAT and demonstrates its unique features through a small case study of a
travel management information system. |
|
Title: |
RAPID
DEVELOPMENT OF PROCESS MODELING TOOLS |
Author(s): |
michele risi,
andrea de lucia, Gennaro Costagliola, genoveffa tortora,
rita francese |
Abstract: |
We present an approach for the rapid development and
evolution of visual environments for modelling distributed software
engineering processes. The definition of the process modeling language
takes into account the requirements of the customer that directly
participates in the development process. The development process is
supported by the VLDesk, an integrated set of grammar-based tools for the
definition and automatic generation of visual environments. The produced
visual environment enables an organization to quickly design distributed
process models and generate the corresponding XML code that specifies the
activities with its elements, including actors and artifacts produced and
the transitions expressed in the form of event-condition-action rules. In
this way the designed process model can be easily instantiated for a
specific project and enacted by any workflow engine supporting a
programmable event-condition-action paradigm. |
|
Title: |
BUILDING
CONCEPTUAL SCHEMAS BY REFINING GENERAL ONTOLOGIES: A CASE STUDY |
Author(s): |
Xavier de
Palol, Jordi Conesa, Antoni Olivé |
Abstract: |
The approach of deriving conceptual schemas from general
ontologies has not been analyzed in detail in the field of information
systems engineering. We believe that the potential benefits of that
approach makes its analysis worthwhile. This paper aims at contributing to
this analysis by means of a case study. The scope of the case study is
rather limited, but even so we believe that some general valid conclusions
can be drawn. The main result has been that deriving conceptual schemas by
refining a general ontology may require less effort than building them
from scratch, and may produce better schemas. On the other hand, an
organization may achieve a high level of integration and reuse, at the
conceptual level, if that organization builds all its conceptual schemas
as a refinement of a general ontology. Our conclusions are similar to
those originated in the development of object-oriented designs and
applications using frameworks. |
|
Title: |
MANAGING
THE COMPLEXITY OF EMERGENT PROCESSES |
Author(s): |
Igor Hawryszkiewycz
|
Abstract: |
Business processes in knowledge intensive environments
often emerge rather than following predefined steps. Such emergence can
result in disconnected activities, which result in complex interaction
structures that require ways to maintain awareness across the activities
and to coordinate the activities to a common goal. The paper suggests that
new ways are needed to both model emergent processes and support and
manage them using information technologies. The paper describes a
metamodel, which includes the commands to create initial processes and to
realize emergence. It then describes a prototype system that implements
these semantics and realizes the creation of initial structures and their
emergence and coordination. |
|
Title: |
OPEN
SOURCE SECURITY ANALYSIS - EVALUATING SECURITY OF OPEN SOURCE VS. CLOSED
SOURCE OPERATING SYSTEMS |
Author(s): |
Paulo
Rodrigues Trezentos, Carlos Serrão, Daniel Neves
|
Abstract: |
Open source software is becoming a major trend in the
software industry. Operating systems (OS), Internet servers and several
other software applications are available under this licensing conditions.
This article assesses the security of open source technology, namely the
Linux OS. Since a growing number of critical enterprise information
systems are starting to use Linux OS, this evaluation could be helpful to
them. To illustrate the fact that application security depends, above all,
on the security of the OS underneath, we present the case of a DRM
(Digital Rights Management) solution – MOSESOpenSDRM - implemented on top
of the Linux OS, in the scope of the EU MOSES IST RTD programme. Some of
conclusions hereby drawn are not compatible with some Microsoft funded
studies that point to the fact that open source OS’s are more insecure.
This main idea was firstly present by the authors in the Interactive
Broadcasting Workshop - IST concertation meeting hosted by the European
Commission in September 2002 (Brussels). |
|
Title: |
TRUSTED
AUTHENTICATION BETWEEN USER AND MACHINE |
Author(s): |
EunBae
Kong, Soyoung Doo, JongNyeo Kim |
Abstract: |
: Authentication is an important issue in computer system
connected to internet. This paper describes a method of providing a
trusted path between a user and a system using an access control
processing technique. The method of providing a trusted path between a
user and a system includes the step of determining whether access to
resources of the system will be permitted or refused on the basis of
access control rules and stored attributes set by a security administrator
in the secure database. Thereafter, the user is notified of permission for
or refusal of the access in accordance with the result of the
determination |
|
Title: |
A
FRAMEWORK FOR BUSINESS SIMULATOR: A FIRST EXPERIENCE |
Author(s): |
Ronan Champagnat
|
Abstract: |
This paper deals with a multi-agent based modeling of a
company in order to perform a simulation. The specificities of the
simulation purpose is that it concerns the economical aspects, but also
the production process. This implies that the model of the company must
represent the production processes. This paper focuses on the modeling of
a company and the analysis of the model. In order to automatically derive
a simulation model from a modeling of the company an UML meta-model has
been achieved and is presented. Then a validation of the components of the
simulator is presented. It allows to perform a validation of the nominal
comportment of agents. This paper is structured as follows: first,
starting from a description of the company a multi-agent model is derived;
then a meta-model for plant modeling is presented; and a validation of the
simulator is detailed; finally the requirements and objectives for a
business simulator is discussed. |
|
Title: |
PATTERN
BASED ANALYSIS OF EAI LANGUAGES - THE CASE OF THE BUSINESS MODELING
LANGUAGE |
Author(s): |
Petia Wohed,
Arthur ter Hofstede, Marlon Dumas, Erik Perjons |
Abstract: |
Enterprise Application Integration (EAI) is a challenging
area that is attracting growing attention from the software industry and
the research community. A landscape of languages and techniques for EAI
has emerged and is continuously being enriched with new proposals from
different software vendors and coalitions. However, little or no effort
has been dedicated to systematically evaluate and compare these languages
and techniques. The work reported in this paper is a first step in this
direction. It presents an in-depth analysis of a language, namely the
Business Modeling Language, specifically developed for EAI. The framework
used for this analysis is based on a number of workflow and communication
patterns. This framework provides a basis for evaluating the advantages
and drawbacks of EAI languages with respect to recurrent problems and
situations. |
|
Title: |
DEFENDING
ESSENTIAL PROCESSES |
Author(s): |
Albert Alderson
|
Abstract: |
The essential purpose of a program makes up only a small
part of the overall task. All of the complications in the program come
from addressing what can go wrong. Where the essential business processes
remain stable, close examination shows complex defensive mechanisms which
change as new threats to the business develop. Norms derive from modelling
social behaviour but are not absolute expressions of what will happen,
people may act counter to the behaviour described in the norm. Many norms
in business are concerned with defending against erroneous or illegal
behaviour of staff and third parties. This paper uses a case study to
illustrate the development of defensive norms and how these norms may be
used in designing processes. Essential business processes cannot be
improved by adding defensive norms but processes are usually more
effective where security norms are implemented preventing the breaking of
norms. |
|
Title: |
TESTING
COTS WITH CLASSIFICATION-TREE METHOD |
Author(s): |
Hareton Leung,
Prema Paramasivam |
Abstract: |
This paper presents a new test method for COTS based on
the classification-tree method. Information from the system specification
and the COTS specification is used to guide the selection of test input.
We can generate test cases that verify input outside the system
specification but within the scope of COTS do not cause problem to the
system, verify input required by the system specification and within the
scope of COTS specification is providing correct results and verify input
not within the scope of COTS specification is actually not required by the
system specification. This paper presents our test selection method with
the help of a case example. |
|
Title: |
ORGANIZATIONAL
MULTI-AGENT ARCHITECTURES FOR INFORMATION SYSTEMS |
Author(s): |
Stephane Faulkner,
Thanh Tung Do, Manuel Kolp |
Abstract: |
A Multi-Agent System (MAS) architecture is an organization
of coordinated autonomous agents that interact in order to achieve
particular, possibly common goals. Considering real-world organizations as
an analogy, this paper proposes MAS architectural patterns for information
systems which adopt concepts from organizational theories. The patterns
are modeled using the i* framework which offers the notions of actor, goal
and actor dependency, specified in Formal Tropos and evaluated with
respect to a set of software quality attributes, such as predictability or
adaptability. We conduct a comparison of organizational and conventional
architectures using an e-business information system case study. |
|
Title: |
A
HIGH RELIABILITY DESIGN FOR NFS SERVER SOFTWARE BY USING AN EXTENDED PETRI
NET |
Author(s): |
Yasunari Shidama,
Katsumi Wasaki, Shin'nosuke Yamaguchi |
Abstract: |
In this paper, we present a model design for Network File
System processes based on the Logical Coloured Petri Net(LCPN). The LCPN
is an extended Petri net which solves the problem of system description in
place/transition nets and coloured Petri nets proposed before. This
extension of Petri nets is suitable for designing complex control systems
and for discussing methods of evaluating such systems. In order to study
the behavior of the server system modeled with this net, we provide
simulations on a Java program. From this work, we confirmed that this
extended Petri net is an effective tool for modelling the file server
processes. |
|
Title: |
HOW
DIGITAL IS COMMUNICATION IN YOUR ORGANIZATION? A METRICS AND AN ANALYSIS
METHOD |
Author(s): |
Tero Päivärinta,
Pasi Tyrväinen |
Abstract: |
Novel innovations in the area of digital media are
changing the ways we communicate and organize. However, few practical
measures exist for analysing the digitalisation of organizational
communication as an intermediate factor in the initiatives to adopt new
information and communication technologies (ICT). Building upon the genre
theory of organizational communication, a categorization of communication
forms, and quantitative measures we suggest such metrics and a measurement
method. A case study applying them in an industrial organization suggests
the method and metrics to be applicable for quantifying how new
information systems affect to organizational communication as well as for
anticipating their digitalisation impact prior to the implementation. The
metrics provide a basis for further work on analysing correlation between
organizational performance and adoption of information and communication
technology. |
|
Title: |
A
TWO TIER, GOAL-DRIVEN WORKFLOW MODEL FOR THE HEALTHCARE DOMAIN |
Author(s): |
Eric Browne
|
Abstract: |
Workflow models define a set of tasks to be undertaken to
achieve a set of goals. Very often, the set of goals is not articulated
explicitly, let alone modelled in such a way as to link the workflow
schema(s) to the goal schema(s). In this paper, we introduce a two tier
model, which clearly delineates the higher level goals (business model)
from the lower level tasks (process model), whilst elucidating the
relationships between the two tiers. We utilise a goal-ontology to
represent the upper level (business model) and decompose this to an
extended Petri-Net model for the lower level workflow schema. The
modelling of business processes, and the management of subsequent changes,
both become an integral part of the workflow itself. Healthcare is a
domain where it is quite common for goals not to be realized, or not to be
realized fully, and where alterations to the goals have to be undertaken
on a case by case (instance-level) basis. Consequently any workflow schema
will need to include tasks that both assess the degree to which a goal has
been achieved, and also allow for new goals to be established, or for the
workflow to be altered. We term such workflow schemas self-managing.
|
|
Title: |
TWO
APPROACHES IN SYSTEM MODELING AND THEIR ILLUSTRATIONS WITH MDA AND
RM-ODP |
Author(s): |
Alain
Wegmann, Andrey Naumenko |
Abstract: |
We explain two approaches to the design of system modeling
frameworks and perform their comparative analysis. The analysis
familiarizes the reader with strengths and weaknesses of the approaches,
and thus helps to grasp the preferences for their practical applications.
The first of the approaches is illustrated with the example of Model
Driven Architecture (MDA), and the second – with the example of Reference
Model of Open Distributed Processing (RM-ODP). |
|
Title: |
AN
DISTRIBUTED WORKFLOW SYSTEM ON OGSA: WORKFLOW GRID SERVICES |
Author(s): |
Kai Wei,
Zhaohui Wu |
Abstract: |
The Open Grid Services Architecture (OGSA) tries to
address the challenges of integrating services across distributed,
heterogeneous, dynamic “virtual organization” formed from the disparate
resources within a single enterprise and/or from external resource sharing
and service provider relationships. In this paper, we attempt to bring
forwards a new model of workflow system with the Grid Services, which have
more valuable characteristics than the Web Services. During discussing the
model, we present some new concepts in distributed workflow systems, all
of which are dealing with Grid services within the OGSA framework. |
|
Title: |
AN
APPROACH TO DFD MODELLING |
Author(s): |
Katja Damij
|
Abstract: |
The objective of this work is to introduce an approach to
developing Data Flow Diagrams. The approach discussed enables the analyst
to create a DFD in an easy manner based on identifying the system’s
activities. The approach has three steps. The first step develops the
activity table. The second step transforms the activity table into an
elementary-level DFD. The third step deals with creating DFDs at work and
business process levels. The approach discussed represents a well-defined
algorithm, which leads the analyst through a few prescribed and very
simple steps to achieve the goal of DFD development. This approach is
independent of the systems analyst and his/her experience. |
|
Title: |
TOWARDS
A VIEW BASED UNIFIED MODELING LANGUAGE |
Author(s): |
Abdelaziz KRIOUILE,
Sophie EBERSOLD, Xavier CREGUT, Mahmoud NASSAR,
Bernard COULETTE |
Abstract: |
To model complex software systems, we propose a
user-centred approach based on an extension of UML (Unified Modeling
Language) called VUML. VUML provides the concept of multiviews class whose
goal is to store and deliver information according to user (final or not)
profile (viewpoint). A multiviews class consists of a base class (default
view) and a set of views corresponding to different viewpoints on the
class. Each view is defined as an extension of the default view. VUML
allows the dynamic change of viewpoint and offers mechanisms to manage
consistency among dependent views. In this paper, we focus on the static
architecture of a VUML modelling: the multiviews class and its use in a
VUML class diagram, static relations such as specialisation and dependency
in VUML. We give an extract of VUML's semantics thanks to an extension of
the UML metamodel. Finally, we give an overview of dynamic aspects and the
principle of an implementation pattern to generate object-oriented code
from a VUML model. |
|
Title: |
IDENTIFYING
PATTERNS OF WORKFLOW DESIGN RELYING ON ORGANIZATIONAL STRUCTURE
ASPECTS |
Author(s): |
Lucinéia Thom,
Cirano Iochpe |
Abstract: |
Modern organizations have demands of automation of their
business processes, once they are highly complex and need to be
efficiently executed. For this reason the number of systems based on
information technologies able to provide a better documentation,
standardization and co-ordination of the business process is increasing.
In this context, workflow technology has been quite efficient, mainly for
the automation of business processes. However, as it is an emergent
technology and in constant evolution, workflow presents some limitations.
One of the main limitations is the lack of techniques that guarantee
correction and efficiency to the workflow project in the phases of
requisite analysis and modeling. Taking into account these problems and
having accomplished some studies, we formulated the hypothesis that it is
possible to infer the specific workflow (sub)processes structure from
knowledge on structural aspects of the organization and vice-versa. We
made the verification of such hypothesis through the identification of a
set of dependency rules among the aspects of the organizational structure
and workflow (sub)processes. This paper presents the set of rules and the
description of the technique used for the identification. |
|
Title: |
A
NEW LOOK AT THE ENTERPRISE INFORMATION SYSTEM LIFE CYCLE - INTRODUCING THE
CONCEPT OF GENERATIONAL CHANGE |
Author(s): |
Jon Davis,
Stephan Chalup, Elizabeth Chang |
Abstract: |
This paper discusses the Enterprise Information System
(EIS) life cycle and the phases of the EIS development life cycle. It
details the stages in the EIS life cycle and the characteristics of phases
in the system development life cycle and explains where it differs from
traditional concepts of software engineering. In particular it defines the
concept of generation change and when it is applied to a system. It also
describes the nature of the rapid evolution of the EIS and how it results
in version or generational change of the system, and how the EIS
development life cycle involves a multitude of engineering processes, not
just one. This new perspective could lead to the generation of new EIS
development methodologies in business modelling, analysis, design, project
management and project estimation. |
|
Title: |
IPM:
AN INCREMENTAL PROCESS MODEL FOR DISTRIBUTED COMPONENT-BASED SOFTWARE
DEVELOPMENT |
Author(s): |
Antonio
Francisco do Prado, Eduardo Santana de Almeida, Calebe de
Paula Bianchini |
Abstract: |
In spite of recent and constant researches in the
Component-Based Development (CBD) area, there is still lack of patterns,
process model and methodologies that effectively support as much the
development “for reuse” as “with reuse”. Considering the accelerated
growth of the Internet over the last decade, where distribution has become
an essential non-functional requirement of most applications, the problem
becomes bigger. This paper proposes a novel Incremental Process Model
(IPM) that integrates the concepts of Component-Based Software Engineering
(CBSE), frameworks, patterns, and distribution. This process model is
divided into two stages: the development “for reuse”, and the development
“with reuse”. A CASE tool is the main mechanism to apply this process
model, supporting inclusively, the code generation of components and
applications. A distributed components environment is proposed for
accomplishing the results of this process model. Through a case study it
is shown how the process model works on a given problem domain. |
|
Title: |
A
DYNAMIC ROLE BASED ACCESS CONTROL MODEL FOR ADAPTIVE WORKFLOW MANAGEMENT
SYSTEMS |
Author(s): |
Dulce Domingos,
Pedro Veiga |
Abstract: |
Workflow management systems (WfMSs) support the
definition, management and execution of business processes inside
organisations. However, business processes are dynamic by nature, meaning
that they must cope with frequent changes. As a consequence we have
witnessed the development of new types of WfMSs, supporting adaptive
workflows. These systems have specific access control requirements, which
are not answered by traditional WfMSs access control models. In this paper
we present a new model to deal with dynamic role based access control
models for adaptive WfMSs, that adapts and extends the well-accepted role
base access control principles in two directions. One direction consists
in a modified interpretation of permissions by considering that the
objects of adaptive WfMSs (e.g. process definition, process instances,
activities and activity instances) and the operations performed on them
like execute, change and read have to cope with dynamic permission
updates. The function that maps a set of permissions into a role is
extended to support exiting relations between workflow components. The
second direction makes execution more adaptive to dynamic changes in
organizations, by separating the role model from the resource model and
supporting the definition of dynamic roles as functions that may access to
external resource information systems. Our model also adapts the RBAC
administrative model to support dynamic authorizations, that is,
authorizations that can be defined and modified at run-time. |
|
Title: |
A
THREE PERSPECTIVE APPROACH TO GROUPWARE IMPLEMENTATION QUALITY MANAGEMENT:
WITHIN AN AUSTRALIAN UNIVERSITY |
Author(s): |
Dion Gittoes
|
Abstract: |
Implementing groupware into organisations to support
communication and collaboration is highly problematic and often approached
from a single perspective. Groupware implementation management is
influenced by individual, socio-organisational and technical issues. High
quality implementation management leads to system success. Quality
literature is investigated to succinctly align the three perspective
approach of groupware design (Palen 1999) to a three perspective approach
of information systems (IS) quality (Salmela 1997). IS quality is
influenced by business integration quality, IS user quality and IS work
quality. A study of a groupware implementation highlights the need for a
synthesis of all three perspectives to manage implementation quality and
understand the adoption challenges groupware systems face. Investigating
IS quality from all three perspectives leads to a holistic understanding
of groupware implementation quality management. Groupware quality is
investigated from the user perspective, employing ethnographic techniques
in an interpretative case study of a Lotus Notes implementation, email and
electronic calendar, within an Australian University. |
|
Title: |
COMPUTING
MESSAGE DEPENDENCIES IN SYSTEM DESIGNS AND PROGRAMS |
Author(s): |
Leszek Maciaszek,
Bruc Lee Liong |
Abstract: |
This paper explains the metrics for the computation of
class dependencies and introduces new metrics for the computation of
message dependencies in system designs and programs. The metrics are used
in the design-coding cycle of software production. They define the quality
of an architectural solution. The system architecture assumed in this
paper is an extension of the MVC (Model-View-Controller) framework
nicknamed BCEMD (Boundary-Control-Entity-Mediator-DBInterface). The paper
demonstrates how the BCEMD framework minimizes object dependencies in
synchronous message passing. We compute message dependencies from a parsed
bytecode. The metrics are then reflected in UML models representing the
system design. The paper starts by briefly explaining the BCEMD
architecture and its advantages. We summarize our earlier paper to show
how the BCEMD approach minimizes the cumulative class dependency. We then
introduce the new metrics resulting in a cumulative message dependency for
the system. The metrics measure the complexity of program’s run-time
behavior. Each metric is defined, given an algorithm for its computation,
and it is then exemplified. We demonstrate how the new metrics reinforce
our claim that the BCEMD architecture delivers understandable,
maintainable and scalable software solutions. |
|
Title: |
INTRUSION
DETECTION BASED ON DATA MINING |
Author(s): |
Xuan
Dau Hoang, Peter Bertok, Jiankun Hu |
Abstract: |
Traditional computer misuse detection techniques can
identify known attacks efficiently, but perform very poorly in other
cases. Anomaly detection has the potential to detect unknown attacks,
however, it is a very challenging task since it is aimed at the detection
of unknown attacks without any priori knowledge about specific intrusions.
This technology is still at its early stage. Existing research in this
area focuses either on user activity (macro-level) or on program operation
(micro-level) but not on both simultaneously. In this paper, an attempt to
look at both concurrently is presented. Based on an intrusion detection
framework (Lee, 2001), we implemented a user anomaly detection system and
conducted several intrusion detection experiments by analysing macro-level
and micro-level activities. User behaviour modelling is based on data
mining; frequent episode algorithms are used to build the user’s normal
profiles. The experimental results have shown that the system can detect
anomalies and changes in the user’s normal working patterns
effectively. |
|
Title: |
AN
AUTHENTICATION SCHEME USING A SECRET SHARING TECHNIQUE |
Author(s): |
Mohamed Al-Ibrahim
|
Abstract: |
We introduce an authentication scheme based on Shamir
threshold secret sharing technique. The scheme, in general, is used for
authenticating peer-to-peer communication. In particular, it is used for
authenticating a host to join a multicast group. |
|
Title: |
A
UNIFIED TOOL FOR EDITING INFORMATION OF DIFFERENT LEVELS OF
ABSTRACTION |
Author(s): |
Alexander Kleschev,
Vassili Orlov |
Abstract: |
Ontology specification languages, ontologies of different
levels of abstraction, enterprise knowledge and data are useful in the
life cycle of enterprise information systems. Since methods of work with
these kinds of information are in a research stage, new experimental
models for representation of information are proposed. As a result, extra
efforts on developing editing tools that are intended for editing
information represented using new models are needed. These tools often
turn out to be mutually incompatible. Meanwhile, the mentioned kinds of
information are closely interrelated. This interrelation results in
additional efforts on providing the editing tools compatibility. Users of
the editing tools need essentially different intellectual support in the
editing process. As a result, considerable efforts and resources are spent
for development of experimental information editing tools that provide
respective classes of users with intellectual support and for establishing
the tools compatibility. This paper presents a model of a unified editing
tool that is intended for solving this problem. |
|
Title: |
PORTUGUESE
LOCAL E_GOVERNMENT |
Author(s): |
Sílvia
Dias, Luis Amaral, Leonel Santos |
Abstract: |
The Internet, the World Wide Web and electronic commerce
are transforming the way of doing business. These changes are impacting
every industry in our country, including local government. The Internet
offers a wide variety of opportunities to improve services to citizens and
to divulge information’s about the communities. In Portugal, the adherence
to the Internet by local government is increasing visibly, but much more
has to be done. In 1999 a first study was done in order to evaluate the
situation of e-government in our country, and two years passed a new study
was undertaken, this time in order to evaluate the evolution registered in
this area. In this paper we describe some conclusions achieved in these
studies comparing their evolution in these two years. |
|
Title: |
A
SYNTHESIS OF BUSINESS ROLE MODELS |
Author(s): |
Alain Wegmann,
Pavel Balabko |
Abstract: |
Modern Information and Communication Technology open a
door for innovations that can improve the functioning of companies. Many
innovations can come from the analysis of business processes. Today
modeling is widely used for the analysis of business processes. In these
work we propose a process modeling technique based on role modeling. To
specify a process where one business object may play several roles, a
synthesis operation (the composition of two base roles in a third role)
has to be specified. All role-based techniques have difficulties
specifying role synthesis: synthesis is never specified without the
description of actual messages passing between business roles. Such
implementation details complicate the understanding of the model and
semantics of synthesis become implicit. To specify a business process of a
complex system at a higher level of abstraction requires the proper
understanding of relationships between roles, when they are put together
in one common context. In this paper we define the concept of “synthesis
constraints” that shows relations between roles. Using “synthesis
constraints” allows a business modeler to make explicit his decisions
about how the synthesis is done in an abstract and implementation
independent way. This approach can be used for building a BPR case tool
that enables the discovery of new business processes by means of different
disassembling and assembling of roles. |
|
Title: |
AGGREGATING
EXPERT PROFILES FOR SER QUERING AID |
Author(s): |
Miguel Delgado,
Maria-Jose Martin Bautista, Daniel Sanchez, Maria
Amparo Vila |
Abstract: |
We present two different models to aggregate document
evaluations and user profiles in the field of Information Retrieval. The
main aim of this work is to discuss a general methodology to establish the
most relevant terms to characterize a given “topic” on an Information
Retrieval System.We start from a set of documents from which a set of
characteristic terms is selected, in such a way that the presence of any
term in each document is known and we want to establish the most
significant ones in order to select “relevant” documents about a given
“topic” Π. For that, some experts , are required to assess the set of
documents. By aggregating these assessments with the presence weight of
terms, a measurement of their relevance in relation with Π may be
obtained.The two presented models differ in the fact that the experts can
query with the same terms (an unique query) to the system or with
different terms (several queries). In each one of these cases, there are
two possibilities: first aggregate the opinions of the experts about the
documents, and then obtain the topic profile. The second possibility is to
generate the expert profiles first, and then aggregate these profiles to
obtain the topic profile.Several different situations arise according to
the form in which the experts state their opinion, as well as from the
approach to aggregate the opinions. An overview of these situations and a
general methodology to cope with them for our model is presented here.
|
|
Title: |
TOWARDS
THE ENTERPRISE ENGINEERING APPROACH FOR INFORMATION SYSTEM MODELLING
ACROSS ORGANISATIONAL AND TECHNICAL BOUNDARIES |
Author(s): |
Prima Gustiene,
Remigijus Gustas |
Abstract: |
Enterprise Engineering proved to be useful when a
generally accepted intentional description of information system is not
available. A blueprint of enterprise infrastructure provides a basis for
system analysis of the organizational and technical processes. It is
sometimes referred as enterprise architecture. The major effort of this
paper is the demonstration of how to bridge a gap among various levels
(syntactic, semantic and pragmatic) of enterprise engineering. Most
information system engineering methodologies are heavily centred on system
development issues at the implementation level. Thus, such methodologies
are restrictive in a way that a supporting technical system specification
can not be motivated or justified in the context of organizational process
models. Enterprise models provide a basis for gradual understanding of why
and how various technical system components come about. Some undesirable
qualities of enterprise engineering are sketched in this paper. |
|
Title: |
EXTENDING
UML FOR MODELLING QUERIES TO OBJECT-RELATIONAL DATABASES |
Author(s): |
Carmen Costilla,
Esperanza Marcos, Belén Vela, Esperanza Marcos |
Abstract: |
Object-Relational Databases (ORDB) have become as a main
alternative for the management of complex data and relationships. It is
also very common to access to these databases through the web. Besides,
new products versions integrate Object-Relational model along with XML
data management. The framework of this paper is MIDAS, an
Object-Relational and XML based Methodology for the design of Web
Information Systems. MIDAS uses UML as modelling language for the
definition of the whole system, extending and specialising it for the
definition of systems based on the recommended technology. In MIDAS we
have proposed an UML extension for the design of ORDB focused on the
structural issues of the system. Due to the importance of queries on every
information system, in this paper we extend that work for ORDB queries
based on UML |
|
Title: |
ANALYSING
REQUIREMENTS FOR CONTENT MANAGEMENT |
Author(s): |
Virpi
Lyytikäinen |
Abstract: |
The content to be managed in organisations is in textual
or multimedia formats. Major part of the content is, however, stored in
documents. In order to find out the needs of the people and organisations
producing and using the content a profound requirements analysis is
needed. In the paper, a method for the requirements analysis for content
management purposes is introduced. The new method combines different
techniques from two existing methods, which were used in various content
management development projects. The paper also describes a case study
where the new method is exploited. |
|
Title: |
USING
MODEL OF ENTERPRISE AND SOFTWARE FOR DEPLOYMENT OF ERP SOFTWARE |
Author(s): |
Franck DARRAS
|
Abstract: |
In the deployment of ERP (Enterprise Resource Planning)
software, a suitable modelling of needs will lead to a better analysis and
a specification of the need for good quality. This modelling can be the
key to success. The enterprise modelling develops a formal framework where
each element of the enterprise is identified and seeks to include all
viewpoints in the representation of the enterprise's operation. But, the
the diversity of the formalisms does not facilitate their use in
management of a project. The aim of this paper is to show the use of the
concepts of enterprise modelling, with a formalism close to software
engineering in order to ameliorate the analysis and the deployment of ERP
systems. |
|
Title: |
A
CONTEXT-AWARE USER-ADAPTIVE SUPPORTING SYSTEM FOR GOAL-ORIENTED
REQUIREMENTS ELICITATION PROCESS |
Author(s): |
Chao Li,
Han Liu, Jizhe Wang, Qing Wang, Mingshu Li |
Abstract: |
Goal-oriented requirements elicitation is recognized as an
important elicitation method from both research and industry. While in a
complex multi-user environment, many problems rise up in performing
goal-oriented requirements elicitation because of ignorance of user
participation support. We present a supporting system here to assist users
in taking part in goal-oriented requirements elicitation process. The
system, with its serious consideration of user factor, is believed to
offer better participation experiences for users in GRE process. |
|
Title: |
A
PEER-TO-PEER KNOWLEDGE SHARING APPROACH FOR A NETWORKED RESEARCH
COMMUNITY |
Author(s): |
Yang Tian
|
Abstract: |
Over the past few years, the interests in the potential of
peer-to-peer computing and the use of different approaches to knowledge
sharing, involving the development of networked communities has grown
rapidly. This paper investigates the potential that a peer-to-peer
community may have for effective and efficient knowledge sharing. It
starts with an introduction to networked communities and the supporting
knowledge sharing activities. A comparison between the centralized and the
decentralized approaches in supporting networked communities is made. A
case study using a networked Journal Club is discussed in detail,
including the design and implementation of the supporting peer-to-peer
prototype using JXTA as the developing platform. The paper concludes with
a discussion of the peer-to-peer architecture as the direction of future
knowledge sharing systems. |
|
Title: |
TELEWORK:
EMPLOYMENT OPPORTUNITIES FOR A DISABLED CITIZEN |
Author(s): |
Nelson Rocha,
Silvina Santana |
Abstract: |
Disable citizens have being considered as potential
beneficiaries of teleworking. However, the subject raises several
questions. Specifically, it is important to determine companies’
willingness to adopt this new work modality, the activities they will
consider to pass to external entities and the more appropriated model to
adopt, when talking about teleworkers with special needs. On the other
hand, it is necessary to determine and analyse perceptions and
expectations, in order to manage eventual resistances and provide
solutions liable to being adopted and used efficiently. This work reports
the results of a study designed to find answers to these questions. It
also allowed finding out the competences potential teleworkers need to
have, enabling the progress of training actions and the development of
insertion strategies adapted to the teleworkers and to the needs and
expectations of employing companies. |
|
Title: |
DISTRIBUTED
WORKFLOW MANAGEMENT IN OPEN GRID SERVICES ARCHITECTURE |
Author(s): |
Zhen Yu,
Zhaohui Wu |
Abstract: |
Vast resources in a grid can be managed flexibly and
effectively by workflow management systems. Here a structure of the
workflow management system in Open Grid Services Architecture is proposed.
In the structure, main components of conventional workflow management
systems are made into some high-level Grid Services and distributed in the
grid. Then those services compose a distributed workflow management
system, which can make full use of the workflow resources and ensure the
process executed efficiently and reliably in the grid The interfaces
required by those workflow services and some implementation details of the
system are discussed too |
|
Title: |
MODELLING
AND GENERATING BUSINESS-TO-BUSINESS APPLICATIONS USING AN ARCHITECTURE
DESCRIPTION LANGUAGE - BASED APPROACH |
Author(s): |
Ilham Alloui
|
Abstract: |
The emergence of Internet and the World Wide Web in
accordance with new technological advances led the organisations to seize
the opportunities offered by electronic business. In particular, the
opportunity to co-operate within the context of electronic (virtual or
networked) enterprises / communities / alliances, based on open networks
and current information and communication technologies. Among the various
kinds of electronic alliances we target are inter-organisational ones that
aim at co-operating to achieve clients’ orders while preserving the
autonomy of involved organisations and enabling concurrency of their
activities, flexibility of their negotiations and dynamic evolution of
their environment. Members of such alliances may either have similar
competencies or complementary ones. In this context, the paper presents a
software architecture-based approach to model and generate
business-to-business (B2B) applications that support decentralised and
dynamic electronic alliances. The approach is founded on modelling
alliance’s life-cycle using an Architecture Description Language (ADL)
called Zeta and generating an executable code from the description into a
target implementation environment called Process/Web. The benefits of such
approach are manifold: (i) using an ADL provides high-level abstractions
hiding implementation details, (ii) having a language means that several
life-cycle models can be defined and modified according to change
requirements, (iii) generating executable code from abstract models can be
done in several target implementation languages. The work presented is
being developed and validated within the framework of the X French
regional project. |
|
Title: |
STUDY
ON CHINESE ENTERPRISE E-READINESS INDEX AND APPLICATION |
Author(s): |
Jian Chen,
Yucun Tian, Yan Zhu |
Abstract: |
Information Industry has become the global main industry
in 21st century, which is the mainstay of national economy and powerful
drive of economy development. How to use information technology to enhance
the core competition ability are the most important factor of the national
and enterprise competitiveness. Since the enterprise is the foundation of
the national economy, the construction of e-enterprise must be speeded up
enormously for improvement of national economy informatization. Therefore,
many scholars and experts are investigating this area at present. The
surveys of nearly 100 Chinese typical enterprises are analyzed and several
new analyzing algorithms will be concerned in this paper. Through these
methods, a set of Chinese enterprise e-readiness index is put
forward. |
|
Title: |
INTERORGANIZATIONAL
WORKFLOW IN THE MEDICAL IMAGING DOMAIN |
Author(s): |
Schahram Dustdar
|
Abstract: |
Interorganizational Workflows are increasingly gaining
relevance in enterprise information systems, particularly in developing
internet-based applications. A process model has to be shared to enable
work items to be managed in different workflow engines. The
state-of-the-art for three interorganizational workflow models, capacity
sharing, case transfer, and loosely coupled model, is discussed in this
paper. Further the medical imaging domain made early progress in workflow
standardization and its main concepts and software components are
introduced. Key workflows and protocols of the medical imaging domain are
described. Next the interorganizational workflow models are applied to the
domain and advantages of certain models are pointed out. Finally the
required infrastructure for a Web-service based design is discussed,
conclusions for Web-service based implementations are made and further
research areas are concluded. |
|
Title: |
FUNCTIONAL
SIZE MEASUREMENT OF LAYERED CONCEPTUAL MODELS |
Author(s): |
Geert Poels
|
Abstract: |
This paper builds on previous work showing a way to map
the concepts used in object-oriented business domain modelling onto (a
subset of) the concepts used by the COSMIC Full Function Points
(COSMIC-FFP) functional size measurement method for modelling and sizing a
software system from the point of view of its functional user
requirements. In this paper we present a refined set of measurement rules
resulting from a careful revision of our previous proposal, based on
‘field trials’, feedback from function points experts and the forthcoming
transition from COSMIC-FFP version 2.1 to the ISO/IEC standard version
2.2. The main contribution of the paper is, however, an extended set of
rules to be used when applying COSMIC-FFP to multi-layer conceptual
models, including at least an enterprise layer and, on top of this, an
information system services layer. We also outline the approach that will
be used to further verify and validate the proposed measurement rules and
to evaluate their efficiency and effectiveness. |
|
Title: |
FÖDERAL:
MANAGEMENT OF ENGINEERING DATA USING A SEMISTRUCTURED DATA MODEL |
Author(s): |
Christoph Mangold,
Ralf Rantzau, Bernhard Mitschang |
Abstract: |
The Föderal system is a flexible repository for the
management, integration and modeling of product data. Current systems in
this domain employ object-oriented data models. Whereas this is adequate
for the management of product data, it proves insufficient for integration
and modeling. Present semistructured data models, however, are suited
ideally for integration, but data management and also modeling is a
problem. In this paper we describe our approach to narrow down the gap
between structured and semistructured data models. We present the Föderal
information system which employs a new semistructured data model and show
how this model can be employed in the context of management, integration,
and modeling of engineering data. |
|
Title: |
SCORING
WWW PAGES WITH SOCIAL CREDIBILITY IN A HYPERLINK ENVIRONMENT |
Author(s): |
Hidenari Kiyomitsu,
Junya Morisita, Tatsuya Kinugawa, Masami Hirabayashi,
Kazuhiro Ohtsuki, Shinzo Kitamura |
Abstract: |
In this paper, we propose an approach in web page scoring.
We introduce the evaluation based on the credibility of the page creators.
We assume all Web pages are latently evaluated on this criterion. We call
this evaluation as an evaluation by Social Credibility(SC). We propose a
dynamical scoring approach utilizing this evaluation of SC and analysis of
link structure by defining degree of recommendation to each link. We show
the convergence of the calculation based on our approach under certain
conditions. We also show the diversity of this evaluation utilizing SC
that is given externally by regarding the evaluation of SC and the
propagation of scores are independent. We experiment this approach and
discuss about the results. |
|
Title: |
PRODUCING
DTB FROM AUDIO TAPES |
Author(s): |
Luís Carriço,
Teresa Chambel, Nuno Guimarães, Carlos Duarte |
Abstract: |
This paper presents a framework for the conversion of
audiotape spoken books to full featured digital talking books. It is
developed within the context of the IPSOM project. The introduction of
search, cross-referencing and annotation mechanisms, with multimedia and
trough multimodal capabilities are considered. Different formats and
standards are taken into consideration, as well as different interaction
alternatives. The resulting digital talking books aim the visually
impaired community, but also situated applications and studies of
cognitive aspects. The framework is part of a larger setting enabling the
authoring, by reuse of and enrichment of multimedia units, of digital
multimedia and multimodal documents. |
|
Title: |
SOFTWARE
CONFEDERATIONS AND MANUFACTURING |
Author(s): |
Michal Zemlicka
|
Abstract: |
Modern information systems of large companies and other
human organizations (like state administration, complex health care
systems) must have a specific architecture called software confederation
(a peer-to-peer network of large autonomous software units behaving like
permanent services). Confederation architecture is a notion and technology
orthogonal to most popular object-orientation. They are used in different
problem or scale domains. For the confederative systems specific experts
are necessary. Such experts can be found among experts having positive
experience with manufacturing systems but not among the experts with
strong object-orientation. Some technical problems simplifying the design
of EAI are discussed and not yet solved issues are formulated. |
|
Title: |
IMPLEMENTING
A GENERIC COMPONENT-BASED FRAMEWORK FOR TELE-CONTROL
APPLICATIONS |
Author(s): |
Avraam Chimaris,
George Papadopoulos |
Abstract: |
In this paper, we design and implement a generic framework
of components that can be used for the realization of Tele-Control
applications. This category of applications focuses particularly on the
issues of managing distributed units on remote end-systems. Such
applications contain remote and administrative units that are connected
and exchange data and control messages. In the analysis of our framework,
we used UML for the specifications, analysis and presentation of system
operations. The distributed units of our framework are using XML messages
and TCP channels for exchanging data and control messages. We implement a
communication “protocol” that contains the basic messages that can be used
in Tele-Control Systems. Finally we are presenting two different
applications, which are implemented by reusing the generic components of
our framework. |
|
Title: |
LANGUAGES
AND MECHANISMS FOR SOFTWARE PROCESSES AND MANUFACTURING ENTERPRISE
PROCESSES: SIMILARITIES AND DIFFERENCES |
Author(s): |
Franck Theroude,
Selma Arbaoui, hervé Verjus |
Abstract: |
This paper tends to confront two wide and deep process
fields: software process and enterprise manufacturing process (called for
short, manufacturing processes). It will analyse them, present a state of
the art according to a set of process requirements and conclude with a
similarities and differences |
|
Title: |
DESIGNING
TOWARDS SUPPORTING AND IMPROVING CO-OPERATIVE ORGANISATIONAL WORK
PRACTICES |
Author(s): |
M.
Cecilia C. Baranauskas, Rodrigo Bonacin |
Abstract: |
Literature in CSCW and related fields has acknowledged the
importance of understanding the social context in which prospective
computational systems for the workplace will be embedded. The
Participatory Design approaches share these concerns and address several
techniques to commit the design process to considerations that take into
account people’s work practices and participation. Although the
participatory techniques provide mechanisms to capture important concepts
of the organisational context, results of these techniques are not well
represented by traditional methods of system modelling. Organisational
Semiotics understands the whole organisation as a semiotic system and
provides methods for considering the social aspects of organisations in
modelling and deriving the system. In this paper we propose an approach
which combines PD techniques and OS methods to design CSCW systems. The
approach is illustrated with Pokayoke: a system designed to support
problem solving in the context of a lean manufacturing organisation. |
|
Title: |
TOWARDS
ADAPTIVE USER INTERFACE GENERATION: ONE STEP CLOSER TO PEOPLE |
Author(s): |
María Lozano,
Antonio Fernández-Caballero, Francisco Montero,
Víctor López-Jaquero |
Abstract: |
User interface generation has become a Software
Engineering branch of increasing interest, probably due to the great
amount of money, time and effort used to develop user interfaces and the
increasing level of exigency of user requirements for usability (Nielssen,
1993) and accessibility (W3C, 2002) compliance interfaces. There are
different kinds of users, and that is a fact we cannot ignore. Human
society is full of diversity and that must be reflected in human-computer
interaction design. Thus, we need to engage users in a new kind of
interaction concept where user interfaces are tailored-made, and where
user interfaces are intelligent and adaptive. A new generation of
specification techniques are necessary to face these challenges
successfully. Model-based design has proved to be a powerful tool to
achieve these goals. A first step towards adaptive user interface
generation is introduced by means of the concept of connector applied to
model-based design of user interfaces. |
|
Title: |
TOWARDS
AN AGENT ARCHITECTURAL DESCRIPTION LANGUAGE FOR INFORMATION SYSTEMS
|
Author(s): |
Manuel Kolp,
Manuel Kolp |
Abstract: |
This paper identifies the foundations for an architectural
description language (ADL) to specify multi-agent system architectures for
information systems. We propose a set of system architectural concepts
based on the BDI agent model and existing classical ADLs. We then
conceptualize SKwyRL-ADL, aimed at capturing a “core” set of structural
and behavioral concepts, including relationships that are fundamental in
architecture description for BDI-MAS. We partially apply our ADL on a
peer-to-peer document sharing example. |
|
Title: |
E-COMMERCE
AUTHENTICATION: AN EFFECTIVE COUNTERMEASURES DESIGN MODEL |
Author(s): |
Victor Sawma
|
Abstract: |
Existing authentication models for e-commerce systems take
into account satisfying legitimate user requirements described in security
standards. Yet, the process of introducing countermeasures to block
malicious user requirements is ad hoc and relies completely on the
security designer expertise. This leads to expensive implementation life
cycles if defects related to the design model were discovered during the
system-testing phase. In this paper, we describe an authentication
countermeasures design model for e-commerce systems. This model includes
effective countermeasures against all known malicious user requirements
and attacks. The described model is preventive in nature and can be used
with other authentication models or can be implemented as a stand-alone
module for e-commerce systems. |
|
Title: |
USER
INTERFACE COMPLEXITY ASSESSMENT IN LARGE-SCALE SAFETY-CRITICAL
ENVIRONMENTS |
Author(s): |
Erman Coskun,
Martha Grabowski |
Abstract: |
In order to design an understandable and usable interface,
the human-computer interaction, computer-supported cooperative work,
psychology, cognitive sciences, and human factors disciplines have
developed methodologies and determined critical elements for successful
user interfaces. The importance of the user interface increases
particularly in safety-critical or mission-critical systems where the user
has time limitations within which to make correct decisions. User
interfaces for these type of systems should be well-designed, easy to
understand and use. Otherwise mishaps or accidents may occur and
consequences of accidents may include loss of human life, large financial
losses, and environmental damage. All this suggests that examining the
complexity of user interface in safety-critical large-scale systems is
important. In this study, we study user interface complexity in
safety-critical environments and report the results of a study conducted
with an Embedded Intelligent Real-Time System and its operators. |
|
Title: |
REAL-TIME
DATABASE MODELING CONSIDERING QUALITY OF SERVICE |
Author(s): |
Maria
Lígia Barbosa Perkusich, Angelo Perkusich |
Abstract: |
Recent research points the real-time database systems
(RTDB) as a key functional unit that contribute to the success of emergent
applications, such as electronic commerce, notice for demand, telephony
systems and on line trading. These research is motivated by the fact of
the dealing of these applications with great amount of data, beyond data
and transactions with timing constraints. Due to the high service demand,
much transactions may miss their deadlines. To address these problems, we
present a RTDB model considering quality of service (QoS) to support
guarantees the performance. A simulation study shows that our model can
achieve a significant performance improvement, in terms of deadlines miss
and accumulated maximum imprecision resultant of the negotiation between
the logical and temporal consistency. Furthermore, we shows model analysis
generated by the Design/CPN tool. |
|
Title: |
DISTRIBUTED
SOFTWARE DEVELOPMENT: TOWARD AN UNDERSTANDING OF THE RELATIONSHIP BETWEEN
PROJECT TEAM, USERS AND CUSTOMERS |
Author(s): |
Roberto Evaristo,
Jorge Audy, Rafael Prikladnicki |
Abstract: |
The objective of this paper is to propose a typology for
distributed software development comprising the relation between the three
main stakeholders: project team (developers, analysts, managers, testers,
system administrator, graphic artist, etc.), customers and users. With
this objective, we propose a set of criteria to define geographically
distributed environments. As a result, a model to define the distribution
level for an organization in a DSD (Distributed Software Development)
environment is presented. Such model is applied in two case studies and
its results are discussed. These cases studies involve two companies with
headquarters in the United States (U.S.) and a development unit in Brazil.
Data from two exploratory case studies are presented to support the
proposed model. Advantages of this representation as well as some aspects
of the increasing distribution of software development particularly in a
few Brazilian organizations are discussed. |
|
Title: |
DEVELOPING
DOCUMENT AND CONTENT MANAGEMENT IN ENTERPRISES USING A ‘GENRE
LENS’ |
Author(s): |
Anne Honkaranta
|
Abstract: |
A great deal of organizational information content is
organized and produced as text and stored, understood and managed as
documents – logical (sets of) content units meant for human comprehension.
In some occasions content needed by human actors can be smaller or larger
by it’s grain size than that of a document. Dynamic nature of digital
documents alongside with multiple technologies used for enacting them have
made a document as an object of analysis fuzzy thus causing possibility
that important information content can be overlooked within enterprise
document management systems and development initiatives. I argue that
enterprises need to find means for analyzing their information content
independently from technologies and media; whether content is identified
and managed as documents or not. For the purpose I introduce a theory of
genres - typified communicative actions characterized by similar substance
and form - as a ‘lens’ that can be operationalized for document and
content analysis. In the paper I discuss how genres can be employed for
document and content analysis by discussing the literature review I
carried out. Research literature shows that theory of genres has been
applied in multiple ways, such as for requirements analysis, identifying
documents used in enterprise workgroups along with their metadata,
analysis and design of information coordination, and so on. Multiple
metadata frameworks have also been developed for classifying the
communicational content within enterprises. Findings of the literature
review implicate that genre ‘lens’ can be used for document and content
analysis in multiple ways. |
|
Title: |
THE
IMPACT OF INFORMATION AND COMMUNICATION TECHNOLOGIES IN SMALL AND MEDIUM
ENTERPRISES |
Author(s): |
Silvina Santana
|
Abstract: |
This work shows part of the results of an empirical
investigation carried out in small and medium enterprises of the
Portuguese industry. One of it’s goals was to determine the impact that
the use of Information and Communication Technologies (ICT) may have in
the way the companies deal with the external environment and in
organizational factors like culture, strategy, management and leadership,
structure, people, processes, routines and procedures and financial
resources. The investigation followed a model for the study of
organizational learning, previously developed and involved 458 companies
of the District of Aveiro, the ones that had responded and affirmed to
hold and use ICT in a preliminary auscultation to all the companies of the
industry of this district (3057). Collected data was submitted to
procedures of multivariate data analysis, namely, Principal Components
Analysis and Cluster Analyses. |
|
Title: |
A
PROMISING MANIFESTO FOR E-BUSINESS VALUE MODELING PROBLEMS |
Author(s): |
Midori Takao,
Masao Matsumoto |
Abstract: |
Continuing to the Value Proposition Innovation method
presented on occasion of the ICEIS’02 panel on “Problems and Promises of
e-business”, this paper explores other crucial subject matters relevant to
the theme, say, what the root problem is out from many of the difficulties
encountered in recent e-business modeling projects. The series of surveys
undertaken by Japan’s Research Thrust IEICE Software Enterprise Modeling
identify that the less matured support for such dual-discipline as
e-business is the root cause that generates numerous problems. One break
through for this is to provide an evaluation framework that allows you to
make decision whether the underlined e-business modeling is beneficial.
The framework will become a key component that is essentially needed for
forming a feedback loop in “model and go” support in e-business. |
|
Title: |
SOFTWARE
PROTOTYPING CLASSIFICATION |
Author(s): |
Claudine
Toffolon, Salem Dakhli |
Abstract: |
Many academics and practitioners consider that software
prototyping is a solution to many symptoms of the software crisis. As
software prototyping may be costly and complex, many typologies have been
proposed in the literature to help understanding this approach. The main
weakness of such typologies is related to their technical orientation. In
this paper, we propose an alternative classification of software
prototyping which take into account all the aspects of software. Our
typology is based upon two frameworks proposed by the authors in a
previous work: the software dimensions theory and the software global
model. |
|
Title: |
REUSING
A TIME ONTOLOGY |
Author(s): |
H.
Sofia Pinto, Duarte Nuno Peralta, Nuno Mamede |
Abstract: |
Ontologies are becoming crucial in several disparate
areas, such as the Semantic Web or Knowledge Manage-ment. Ontology
building is still more of an art than an engineering task. None of the
available methodologies to build ontologies from scratch has been widely
accepted. One cost effective way of building ontologies is by means of
reuse. In this article we describe the development of an ontology of Time
by means of reuse, following an evolving prototyping life cycle. This
process involved several complex subprocesses: knowledge acquisition and
requirement specification using Natural Language techniques, reverse
engineering, knowledge representation translation, technical evaluation.
As far as we know, this is the first time that all these processes have
been combined together. We describe the techniques and best practices that
were successfully used. |
|
Title: |
IS
DEVELOPMENT WITH IS COMPONENTS |
Author(s): |
Slim TURKI,
Michel LEONARD |
Abstract: |
In this paper, we expose our vision of the component based
information systems (IS) development in a data-intensive application
context. A hyperclass is a large class, formed from a subset of conceptual
classes of the global schema of a database, forming a unit with a precise
semantics. This concept introduces a kind of modularity in the definition
and the management of a database schema and a powerful kind of
independence between the methods and the schema. We present our global
approach of reuse and give our definition of an IS component (ISC). For
us, an ISC is an autonomous IS, respecting a set of conformity rules. It
is defined through a static space, a dynamic space and an integrity rules
space. We use the hyperclass concept to implement the static space.
Applied to an ISC, it facilitates its handling when refined or
integrated. |
|
Title: |
SYSTEMS
DEVELOPMENT METHOD RATIONALE: A CONCEPTUAL FRAMEWORK FOR
ANALYSIS |
Author(s): |
Pär Ågerfalk,
Kai Wistrand |
Abstract: |
The rationale of their creators inherently influences
information systems development methods. Rationale, which is always based
on the creator’s values and assumptions about the problem domain,
motivates, implicit or explicit, the different modelling activities and
primitives prescribed by a method. The method, and hence its inherited
rationale, directs method users’ attention toward certain kinds of
phenomena and away from others. Today we see a trend towards standardizing
systems development in terms of standard modelling languages and standard
development processes. When using an existing (standard) method,
developers are forced to rely on the rationale of that particular method.
Sometimes, however, there are reasons to enhance the standard method to
reflect aspects of the world held as important by the method users – but
not emphasized by the method creator. Hence, there is a need to integrate
the rationale of the method users with that of the existing method. In
this paper, we investigate what method rationale is and how it can be
modelled and analysed. The paper employs a method engineering approach in
that it proposes method support for analysing, describing and integrating
method rationale – an emerging essential task for method engineers in a
world of standardization. |
|
Title: |
A
METHODOLOGY FOR THE INTEGRATION OF CSCW APPLICATIONS |
Author(s): |
Anne
James, Richard Gatward, Rahat Iqbal |
Abstract: |
Organizations rely on a wide variety of collaborative
applications in order to support their every day activities and to share
resources. The collaborative applications are typically designed from
scratch if the existing applications do not meet organizational needs.
This requires more budget and time. This paper reports on the integration
of existing collaborative applications or computer supported cooperative
work (CSCW) in order to support collaborative activities of organizations.
This will meet their requirements at low cost. This is a part of our
research towards investigating and developing an integrative framework for
CSCW applications. It will be flexible enough to accommodate the various
and varying needs of the organization community. We discuss different
types of integration model and interoperability in CSCW and consider
different models of CSCW systems. A framework for CSCW integration is
presented. A methodology based on this framework is also proposed |
|
Title: |
EXTRACTING
THE SOFTWARE ELEMENTS AND DESIGN PATTERNS FROM THE SOFTWARE
FIELD |
Author(s): |
Mikio Ohki,
Yasushi Kambayashi |
Abstract: |
Deriving the class structure of object-oriented software
has been studied intensively. We have proposed a methodology to divide the
conceptual model used in the object-oriented analysis into basic elements,
such as classes, attributes, methods, relations, and to define constraint
characteristics and constructing operations on each element. In the
methodology, we have applied the field theory in the quantum physics to
software and proposed the software field concepts (Ohki and Kambayashi,
2002a). Our thesis is that software is a kind of fields in which software
elements, such as methods and attributes, interact each other to produce
certain behavioral patterns. The methodology explains well the
characteristics of class libraries (Ohki and Kambayashi, 2002b). Once the
software elements are extracted from the software field, the methodology
allows constructing design patterns from the characteristics of the
elements (Ohki and Kambayashi, 2002a). Although we defined the extract
operations to elicit the software elements, we failed to show that those
operations have reasons and are correct (Ohki and Kambayashi, 2002a). In
order to overcome this problem, in this paper, we introduce the
distribution functions to represent the software elements, and to
formulate the interactions of the functions. Using the distribution
functions and the interactions between them, we have succeeded to suggest
how to extract the software elements from the software field, and how to
derive the design patterns by using the characteristics of the extract
elements. This paper first describes the basic concepts of the software
field, and then introduces the distribution functions to represent the
software elements. In the latter part of this paper describes that it is
applicable to derive typical design patterns. |
|
Title: |
FLEXIBILE
PROCESSES AND METHOD CONFIGURATION: OUTLINE OF A JOINT INDUSTRY-ACADEMIA
RESEARCH PROJECT |
Author(s): |
Kjell Möller,
Pär Ågerfalk, Kai Wistrand, Gregor Börjesson,
Fredrik Karlsson, Martin Elmberg |
Abstract: |
This paper outlines a joint industry-academia research
project in the area of method engineering. Founded in practical
experiences and emerging theoretical constructs, the project aims as
developing theories, methods and tools to support the adaptation,
integration and construction of method components for flexible
configuration of system development methods. By explicating the
possibilities of using method’s inherent rationale, the possibilities to
adopt rigorous methods (such as the Rational Unified Process) to comply
with increasing demands of flexibility will be exploited. The paper also
addresses the approach to technology transfer adopted in the project,
viewing the project as existing in two different intellectual spheres, one
academic and one industrial. The two spheres overlap in a space of
conceptualization and interpretation shared by the practitioners and
academic researchers involved. This way the project adopts an iterative
process of reflection and application, generating knowledge directly
applicable in industry as well as knowledge of theoretical and scientific
importance. |
|
Title: |
A
NEW USER-CENTERED DESIGN OF DISTRIBUTED COLLABORATION ENVIRONMENTS: THE
RÉCIPROCITÉ PROJECT |
Author(s): |
Alain Derycke,
Frédéric Hoogstoel, Xavier Le Pallec, Ludovic Collet
|
Abstract: |
Designing collaborative applications is a hard task.
Indeed, anticipating users' needs and help users understanding the process
of proposed services are more difficult due to the group dimension. The
Réciprocité project is a new way in designing collaborative applications.
It tries to reduce the two previous difficulties. In this paper, we
present strong points of our approach: Peer-to-Peer (P2P), full-XML
architecture, and tailorability mechanisms. |
|
Title: |
ASPECT-ORIENTED
SOFTWARE DEVELOPMENT: AN OVERVIEW |
Author(s): |
Isabel
Brito, Ana Moreira |
Abstract: |
Separation of concerns is a software engineering principle
that claims the clearly identification of all the elements that
participate in a system. There are some concerns such as security and
performance that cut across many other concerns. The classical approaches
do not support the modularisation and further integration of these
crosscutting concerns with the functional requirements of a system
producing spread and tangled representations (e.g. specifications, code)
that are difficult to understand, maintain and evolve. Aspect-Oriented
Programming (AOP) aims at handling these problems. Recently we have been
noticing a clear interest on propagating the AOP ideas and concepts to
earlier activities of the software development process. This paper gives
an overview of aspect-oriented software development, giving a special
emphasis to aspect-oriented requirements engineering |
|
Title: |
COMPLEMENTARY
ADOPTION AND OPTIMIZATION OF ENTERPRISE RESOURCE PLANNING
SYSTEMS |
Author(s): |
C.
Sophie Lee |
Abstract: |
Enterprise Resource Planning (ERP) systems emphasize on
the integrative and platform adoption of system software, instead of
piece-meal upgrades. However, technological changes can only bring limited
benefit to the organization if other factors in the organization -- such
as strategy and organizational structure -- are not changed in the
coordinated or complementary directions. Failure to understand the
complementarity between technology and organizational may cause low payoff
of IT investment. Customer Relationship Management (CRM) software is the
latest addition to the ERP family. It is rooted in relationship marketing
or customer relationship management, which emphasizes the need to build a
long term relationship with customers. Spending on CRM software has grown
6 times over the past years but customer satisfaction of American
consumers did not grow. An examination of the literature reveals that
current CRM model tends to focus more on customer “service” instead of
customer “relationship”. This study proposes to combine the American
Customer Satisfaction Index (ACSI) model and complementarity framework to
provide optimal design and adoption of CRM software. |
|
Title: |
INFORMATION
SYSTEM FAILURE: ANALYSIS AND COMMENTARY |
Author(s): |
John Wise,
Ata Nahouraii, Anthony Debons |
Abstract: |
Contemporary events of considerable significance to
national and public welfare suggest that information was a significant
force on the character and outcome of each event such as the Chechen
rebels held more than 700 hostages in the Dubrovku, Kremlin theater on
October 23, 2002, the terrorist attack of 9/11, the Challenger shuttle
accident, the explosion at Bhopal. The need to identify the success and
failures of information systems in these events is deemed to be relevant
to national and private interests. In 1986, an aggregation of
distinguished scholars met in Bad Windsheim, Federal Republic of Germany
to serve as lecturers in an Advanced Study Institute sponsored by the
NATO’s Science division. A number of issues were addressed which included
the prevailing methods used in the assessment of information system
failure, the organizational factors pertaining thereof, the role of human
cognitive variables, the capacity of the system to induce or resist
failure and many other socio-economic- political variables that were
considered pertinent to an understanding of information system failure.
The paper summarizes these dimensions of information system failure as
presented at the institute and in addition, comments on the importance on
such systems based on contemporary socio-political circumstances. |
|
Title: |
NEW
APPROACH TO TEST THE SAP SYSTEM DATA SECURITY |
Author(s): |
Jen-Hao Tu
|
Abstract: |
The SAP system is the most widely used ERP (Enterprise
Resource Planning) system in the world. There are thousands of seamlessly
linked components and subsystems. Conducting security tests in a
complicated ERP system is still a major challenge. Based on the study of
the SAP system data security testing at the author’s company, this paper
will discuss issues related to legal and regulatory requirements, IT
security governance and segregation of duty in order to meet these
emerging security challenges. A practical SAP data security framework is
proposed to link these requirements to the business units. AIS (Audit
Information System) was, originally, an integrated audit tool provided by
the SAP company to facilitate both the SAP system and the business audit
process. The functionality of AIS will be explored to ensure the tests
meet these security requirements in the SAP data security framework.
|
|
Area 4 - INTERNET COMPUTING AND ELECTRONIC COMMERCE
Title: |
THE
RESOURCE FRAMEWORK FOR MOBILE APPLICATIONS: ENABLING COLLABORATION BETWEEN
MOBILE USERS |
Author(s): |
Jörg Roth
|
Abstract: |
Mobile devices are getting more and more interesting for
several kinds of field workers such as sales representatives or
maintenance engineers. When in the field, mobile users often want to
collaborate with other mobile users or with stationary colleagues at home.
Most established collaboration concepts are designed for stationary
scenarios and often do not sufficiently support mobility. Mobile users are
only weakly connected to the communication infrastructure by wireless
networks. Small mobile devices like PDAs often do not have sufficient
computational power to handle effortful tasks to coordinate and
synchro-nize users. They have for example very limited user interface
capabilities and reduced storage capacity. In addition, mobile devices are
subject to other usage paradigms like stationary computers and often
turned on and off during a session. In this paper, we introduce a
framework for mobile collaborative applications based on so-called
resources. The resource framework leads to a straightforward functional
decomposition of the overall application. Our platform Pocket DreamTeam
provides a runtime infrastructure for applica-tion based on resources. We
demonstrate the resource concept with the help of two applications build
to top of the Pocket DreamTeam platform. |
|
Title: |
A
SEMANTIC FRAMEWORK FOR DISTRIBUTED APPLICATIONS |
Author(s): |
Liang-Jie Zhang,
Wei-Tek Tsai, Bing Li |
Abstract: |
The .XIN technology is a novel approach to build and
integrate existing distributed applications. The essence of a .XIN is
business logic descriptions. Based on the concept of .XIN, developers’
effort is minimized because their developing work is concentrated on
mapping business logic to .XINs. The adaptor layer is an interpreter that
translates .XINs into implementations of particular distributed domains.
This layer hides details of implementation techniques of distributed
applications. Moreover, applications built with .XIN can share their
services over Internet via RXC (Remote .XIN Call) and a remote .XIN-based
Service can be blended into a local .XIN-based application via RXI (Remote
.XIN Interchange). Finally, an object interface can be mapped to a .XIN
interface. With the support of this mapping, both non-.XIN applications
and .XIN applications have the same interface, .XIN interface. So it is
possible for them to share their respective services over the Internet.
This is also a new approach to integrate heterogeneous applications. The
technology of .XIN is a semantic framework for distributed
applications. |
|
Title: |
OPEN
TRADING - THE SEARCH FOR THE INFORMATION ECONOMY'S HOLY GRAIL |
Author(s): |
Graham Scriven
|
Abstract: |
This paper examines the concept of Open Trading,
establishing its crucial importance in achieving comprehensive benefits
for all trading partners as a result of the move towards the Information
Economy. The rationale for interoperability is also examined and placed in
perspective. The paper considers how Open Trading can be achieved and
suggests ten principles as a practical guide for both vendors and business
organisations. |
|
Title: |
EVALUATION
OF MAINFRAME COMPUTER SYSTEM USING WEB SERVICE ARCHITECTURE |
Author(s): |
Yukinori Kakazu,
Mitsuyoshi Nagao |
Abstract: |
In this paper, we propose a mainframe computer system
using a web service architecture in order to realize a mainframe computer
system that permits users to conveniently access it to perform flexible
information processing. The web service is a system architecture that
communicates among applications through the Internet by using the SOAP
(Simple Object Access Protocol). SOAP is a simple protocol based on XML
and HTTP. It has the advantages that the communication can be performed
beyond the firewall provided to promote network security and that it can
be used on various platforms. The web service architecture inherits these
advantages of SOAP. It is likely that an effective and convenient
mainframe computer system used over the Internet can be implemented by
using the web service architecture. Moreover, the implementation of the
proposed system can bring about new application model. Applications that
users can unconsciously use the mainframe computer system and which can
perform large-scale information processing can be implemented on
low-performance clients, such as mobile platforms, by realizing the
proposed system. In addition, the application combining the
high-performance libraries on a mainframe computer system can be
implemented on such a client. We report the construction of the proposed
system and confirm its effectiveness through a computational experiment.
The experimental result revealed that effective information processing
could be performed over the Internet by using the proposed system. |
|
Title: |
WHAT
IS THE VALUE OF EMOTION IN COMMUNICATION? IMPLICATIONS FOR USER CENTRED
DESIGN. |
Author(s): |
Robert Cox
|
Abstract: |
This research presents an investigation into the question
- what is the value of emotion in communication. To gain a greater
appreciation of this title, it is this paper’s intention to de-construct
the sentence into its component parts – namely its nouns; Value, Emotions
and Communications, and to study them in isolation to each other and as a
total construct. Further, the everyday use of communications technology
(i.e. e-mail, chat lines, mobile and fixed line telephones) has changed
human communication norms. To identify the significance of this change, an
investigation into the question of whether emotions continue to play an
important role in effective human-to-human communications is most likely
warranted. |
|
Title: |
COMBINING
WEB BASED DOCUMENT MANAGEMENT AND EVENT-BASED SYSTEMS - MUDS AND MOOS
TOGETHER WITH DMS FORM AN COOPERATIVE OPEN SOURCE KNOWLEDGE
SPACE |
Author(s): |
Thorsten
Hampel |
Abstract: |
The WWW has developed as the de facto standard for
computer based learning. However, as a server-centered approach it
confines readers and learners to passive non-sequential reading. Authoring
and web-publishing systems aim at supporting the authors' design process.
Consequently, learners' activities are confined to selecting and reading
(downloading documents) with almost no possibilities to structure and
arrange their learning spaces nor do that in a cooperative manner. This
paper presents a learner-centered – completely web-based – approach
through virtual knowledge rooms. Our technical framework allows us to
study different technical configurations within the traditional university
setting. Considering the systems design the concept of virtual knowledge
rooms is to combine event-based technology of virtual worlds with the
classical document management functions in a client-server framework.
Knowledge rooms and learning materials such as documents or multimedia
elements are represented as a fully object oriented model of objects,
attributes and access rights. We do not focus on interactive systems
managing individual access rights to knowledge bases, but rather on
cooperative management and structuring of distributed knowledge bases.
|
|
Title: |
USERS-TAILORED
E-BUSSINES THROUGH A THREE-LAYER PERSONALIZATION MODEL BASED ON
AGENTS |
Author(s): |
Irene Luque
Ruiz, Miguel Angel Gómez-Nieto, Gonzalo Cerruela Garcia,
Enrique López Espinosa |
Abstract: |
The explosion of Internet together with the advantages
that offers nowadays the electronic commerce, are provoking an important
growth of the Web sites devoted to the development of this activity; what
gives rise to each time be greater the quantity of information that
arrives to the users upon the products or services that offer said sites.
The users before so much information, which in some cases will interest it
and in other not, finish for not processing it. This situation has
provoked that the researchers try to seek solutions, among the ones that
emphasizes the use of Artificial Intelligence for solve this problem. With
this idea appears the personalization of Web sites, which has as objective
to provide to the user the information that he needs. In this paper a
personalization model in various levels is proposed, which applied to the
Business Virtual Centre portal (BVC) will try to personalize services,
information, as well as, the activities that will be able to carry out
each user in it. Personalization model is based on: stereotypes existing
in the system, information introduced by the user and the knowledge
extracted from the information generated by the user during its stay in
the BVC. |
|
Title: |
PERSONALIZATION
MEETS MASS CUSTOMIZATION - SUPPORT FOR THE CONFIGURATION AND DESIGN OF
INDIVIDUALIZED PRODUCTS |
Author(s): |
Martin Lacher,
Thomas Leckner, Michael Koch, Rosmary Stegmann |
Abstract: |
Using electronic media for customer interaction enables
enterprises to better serve customers by cost-efficiently offering
personalized services to all customers. In this paper we address the area
of providing help for customers in selecting or designing individualized
products (mass customization) by using personalization technologies. The
paper provides an introduction to the application area and presents a
system for supporting the customization and design of individualized
products. The support solution is presented and discussed from a process
(customer) point of view and from a system point of view. |
|
Title: |
E-COMMERCE
PAYMENT SYSTEMS - AN OVERVIEW |
Author(s): |
Pedro Fonseca,
Joaquim Marques, Carlos Serrao |
Abstract: |
Electronic Commerce is playing a growing importance on
modern Economy since it provides a commodity way for consumers to acquire
goods and services through electronic means – Internet and the WWW are the
most important. However, this new way of trade raises important problems
on the way payments are being made, and trust is one of the most important
one. This paper starts by presenting some of the complexities related to
Electronic Commerce payments in this New Economy, both on a consumer and
seller perspective. Next, differences between the traditional and
electronic payment systems are identified and how they both deal with the
identified complexities. Electronic payment systems (EPS) are then
identified referring the advantages presented to Electronic Commerce.
Finally, a comparative EPS table is presented identifying strong and week
points on each of the EPS and conclusions are drawn from this. |
|
Title: |
CONTENT
ANALYSIS OF ONLINE INTERRATER RELIABILITY USING THE TRANSCRIPT RELIABILITY
CLEANING PERCENTAGE (TRCP): A SOFTWARE ENGINEERING CASE STUDY |
Author(s): |
Peter Oriogun
|
Abstract: |
In this paper the author presents a case study of online
discourse by message unit using quantitative content analysis, with
particular emphasis on the author's proposed interrater agreement
percentage that will be referred to in this paper as Transcript
Reliability Cleaning Percentage (TRCP). The paper will examine the ratings
of participants' messages in terms of level of engagement within a
negotiation forum in line with the author's Negotiated Incremental
Architecture, Oriogun (2002) using the Win-Win Spiral Model, Boehm (1988).
The variables that the author investigated are, participation, and
interaction. The paper is divided into six sections, that will introduce
the rationale for the study, a brief introduction to the Negotiated
Incremental Architecture, followed by the study itself, we then define
what we means by Transcripts Reliability Cleaning Percentage (TRCP) of
online discourse using message unit, followed by the interpretation of
individual participant's result and finally the author will conclude with
a recommendation of a follow-on paper, using our SQUAD approach to online
posted messages. The SQUAD approach is a semi-structured categorisation of
online messages. The paper also discusses the reasons why there has been
very little research on interrater reliability with respect to content
analysis of online discourse, furthermore, a comparison is made of Cohen's
kappa value as reported in Rouke, Anderson, Garrison & Archer (2000)
and the author's proposed Transcript Reliability Cleaning Percentage
(TRCP). It is argued in this paper that the proposed Transcript
Reliability Cleaning Percentage (TRCP) will better enhance interrater
reliability (percentage agreement between coders) of the rating of online
transcripts. The author is suggesting that it is not possible under
certain circumstances to obtain 100% agreement between coders after
discussion. However, the author noted that this was achieved by, Hara,
Bonk & Angeli (2000). |
|
Title: |
ARCHCOLLECT:
A SET OF COMPONENTS TOWARDS WEB USERS’ INTERACTIONS |
Author(s): |
Julio Ferreira,
Edgar Yano, Joao Sobral, Joubert Castro,
Tiago Garcia, Rodrigo Pagliares |
Abstract: |
Abstract This paper describes an example of a system that
emphasizes web users’ interactions, called ArchCollect. One JavaScript
component and five Java components gather information coming only from the
user, independing onthe web application that will be monitored and on the
web server used to support it. This improves the portability of this
software and its capacity to deal with many web applications in a Data
Center at the same time, for example. The ArchCollect relational model,
which is composed by several tables, provides analyses, regarding factors
such as purchases, business results, the length of time spent to serve
each interaction, user, process, service or product. In this software,
data extraction and the data analysis are performed either by
personalization mechanisms provided by internal algorithms, or by any
commercial decision making tools focused on services, such as, OLAP, Data
Mining and Statistics, or by both. |
|
Title: |
INTEGRATION
OF OBJECT-ORIENTED FRAMEWORKS HAVING IDL AND RPC-BASED
COMMUNICATIONS |
Author(s): |
Debnath Mukherjee
|
Abstract: |
This paper proposes a software architecture to unify
disparate application frameworks that have Interface Definition Language
(IDL) and RPC-based communication between client and server, thus enabling
distributed computation using disparate frameworks. The architecture also
demonstrates how multiple inheritance from classes belonging to disparate
object-oriented frameworks is possible. |
|
Title: |
THE
SECURE TRUSTED AGENT PROXY SERVER ARCHITECTURE |
Author(s): |
Michelangelo Giansiracusa
|
Abstract: |
Concerns of malicious host system attacks against agents
have been a significant factor in the absence of investment in agent
technologies for e-commerce in the greater Internet. However, in this
paper, it can be seen that agent systems represent a natural evolution in
distributed system paradigms. As in other distributed systems, applying
traditional distributed systems security techniques and incorporating
trusted third-parties can discourage bad behaviour by remote systems. The
concept and properties of a trusted proxy server host as a 'middle-man'
host anonomising authenticated agent entities in agent itineraries is
introduced, along with its inherent benefits. It is hoped that this fresh
secure agent architectural offering will inspire further new directions in
tackling the very challenging malicious agent platform problem. |
|
Title: |
SECURE
SMART CARD-BASED ACCESS TO AN E-LEARNING PORTAL |
Author(s): |
Josef von
Helden, Ralf Bruns, Jürgen Dunkel |
Abstract: |
The purpose of the project OpenLearningPlatform is the
development of an integrated E-learning portal in order to support
teaching and learning at universities. Compared to other E-learning
systems the originality of the OpenLearningPlatform is the strong smart
card-based authentication and encryption that significantly enhances its
usefulness. The secure authentication of every user and the encryption of
the transmitted data are the prerequisites to offer personalized and
authoritative services, which could not be offered otherwise. Hence, the
smart card technology provides the basis for more advanced E-learning
services. |
|
Title: |
TOWARDS
WEB SITE USER'S PROFILE: LOG FILE ANALYSIS |
Author(s): |
Carlos
Alberto de Carvalho, Ivo Pierozzi Jr., Eliane
Gonçalves Gomes, Maria de Cléofas Faggion Alencar |
Abstract: |
The Internet is a remote, innovative, extremely dynamic
and widely accessible communication medium. As in all other human
communication formats, we observe the development and adoption of its own
language, inherent to its multimedia aspects. The Embrapa Satellite
Monitoring is using the Internet as a dissemination medium of its research
results and interaction with clients, partners and web site users for more
than one decade. In order to evaluate the web site usage and performance
of the e-communication system the Webalizer software has been used to
track and to calculate statistics based on web server log file analysis.
The objective of the study is to analyze the data and evaluate the
indicators related to requests origin (search string, country, time),
actions performed by users (entry pages, agents) and system performance
(error messages). It will help to remodel the web site design to improve
the interaction dynamics and also develop a customized log file analyser.
This tool would retrieve coherent and real information. |
|
Title: |
SCALABLE
AND FLEXIBLE ELECTRONIC IDENTIFICATION |
Author(s): |
David Shaw,
S. Maj |
Abstract: |
Verification of network service requests may be based on
widely available identification and authentication services. Complexity or
multiple access requirements may require access control artefacts such as
hardware based signature generators. Characteristics of artefact generated
signatures include security and repeatability. Electronic signatures used
in remote transactions need to be graded, flexible and scalable to permit
appropriate user selection. Further, inherent error detection may reduce
inadvertent errors and misconduct and aid arbitration. |
|
Title: |
A
SURVEY RESEARCH OF B2B IN IRAN |
Author(s): |
Javad Karimi
Asl |
Abstract: |
EC is relatively new concept in business domain( Wigand,
1997). While the consumer side of the Web explosion has been much touted,
it is the Business-to-Business (B2B) market that has quietly surpassed
expectations. This paper is based on a survey of 102 business managers (or
IS) in Iran and discusses the management practices, application, problems
and technical situations with regard to EC development in this country .In
this paper was evaluated the B2B situation in Iran. This paper discusses
about their business or Is manager’s experiences, and satisfaction with
current electronic- commerce (EC) solutions in use. The finding of this
paper are useful for both researchers and practitioners as they provide an
insight for critical management issues which engage both under development
countries'non-governmental organizations and policy makers. The result of
this study shows that there are more differences between conditions of EC
in developed and derdeveloping countries |
|
Title: |
WHITE
PAPER FOR FLOWAGENT PLATFORM |
Author(s): |
Wenjun Wang
|
Abstract: |
FlowAgent is the network platform to implement "Streamline
Bus" with Jini network technology. "Streamline Bus" is trying to solve
problems that prevent us to integrate different applications
cross-enterprises/ organizations; it realizes task-scheduling among
different applications through pre-defined task data requiring / providing
relations; it can provide automatically workload balance, dynamic fail
over and run-time data/performance tracking. One critical issue of
FlowAgent platform is how to define the internal format for the task
running/scheduling data, (1) let it provide the isolated applications the
request data for running, as while (2) control the flow through the
“Streamline service”. Base on "Streamline Bus", you can build large-scale
scheduling systems, that integrates applications of different business
fields. Systems based on “Streamline Bus” are in full-distributed model,
are very different from traditional “Workflow systems”, which depend on
centralized rule engine and has much limitations on the types of
application can be integrated. |
|
Title: |
A
MISSION-BASED MULTIAGENT SYSTEM FOR INTERNET APPLICATIONS |
Author(s): |
Glenn Jayaputera,
Seng Loke, Arkady Zaslavsky |
Abstract: |
Software agents have been one of the most active research
areas in the last decade. As a result, new agent technologies and concepts
are emerging. Mobile agent technology has been used in real life
environments, such as on-line auctions, supply chain, information
gatherings, etc. In most situations, mobile agents must be created and
carefully crafted to work together almost from scratch. We believe that
this is quite inefficient for application developers and users, and hence
propose a system for generating and coordinating agents based on the
notion of agent missions. The prototype system is called eHermes and its
architecture and components are discussed in the paper. |
|
Title: |
KNOWLEDGE
CONSTRUCTION IN E-LEARNING - DESIGNING AN E-LEARNING ENVIRONMENT |
Author(s): |
Lily Sun,
Kecheng Liu, Shirley Williams |
Abstract: |
In the traditional classroom, students tend to depend on
tutors for their motivation, direction, goal setting, progress monitoring,
self-assessment, and achievement. A fundamental limitation is that
students have little opportunity to conduct and manage their learning
activities which are important for knowledge construction. E-Learning
approaches and applications which are supported by pervasive technologies,
have brought in great benefits to the whole society, meanwhile it also has
raised many challenging questions. One of the issues of which researchers
and educators are fully aware is that technologies cannot drive a
courseware design for e-Learning. An effective and quality learning
requires an employment of appropriate learning theory and paradigms,
organisation of contents, as well as methods and techniques of delivery.
This paper will introduce our research work in design an e-Learning
environment with emphases on instructional design of courseware for
e-learning. |
|
Title: |
THE
FUTURE OF TELEPHONY: THE IP SOLUTION |
Author(s): |
Sílvia Fernandes
|
Abstract: |
Enterprises have begun to transform their working
environment to meet not only the business world of today but also the
business world of tomorrow. The working methods are more flexible than
ever before: some workers collaborate entirely from home and others work
in several different offices circulating between remote workplaces. In a
short time the way we work will be so radically different that working
will be just what we do and no more where we are. As globalisation becomes
a business reality and technology transforms communications, the world of
data transmission together with wireless networks has progressed a lot
instead of fixed and wire-line voice communications that have barely
changed. However tariffs are still based on time and distance even though
it does not make any sense in today’s global marketplace, in spite of the
reduced costs that have resulted from the deregulation process over public
telephone networks. |
|
Title: |
TOWARD
A CLASSIFICATION OF INTERNET SCIENTIFIC CONFERENCES |
Author(s): |
Abed Ouhelli,
Prosper Bernard, Michel Plaisent, Lassana Maguiraga
|
Abstract: |
Since 1980, the classification of scientific production
has been an constant concern for academics. Despite its growing importance
in the last decade, Internet has not been investigated as an autonomous
domain. This communication relates the efforts to develop a first
classification of themes based on calls for paper submitted to the ISWORLD
community in the last two years. The distribution of theme and sub-themes
is presented and compared. |
|
Title: |
WEB
NAVIGATION PATTERNS |
Author(s): |
Eduardo Marques,
Ana Cristina Bicharra Garcia |
Abstract: |
Many Internet service providers and online services
require you to manually enter information, such as your user name and
password, to establish a connection. With Scripting support for Dial-Up
Networking, you can write a script to automate this process. A script is a
text file that contains a series of commands, parameters, and expressions
required by your Internet service provider or online service to establish
the connection and use the service. You can use any text editor, such as
Microsoft Notepad, to create a script file. Once you've created your
script file, you can then assign it to a specific Dial-Up Networking
connection by running the Dial-Up Scripting Tool. |
|
Title: |
DYNAMICALLY
RECONSTRUCTIVE WEB SERVER CLUSTER USING A HIERARCHICAL GROUPING
MECHANISM |
Author(s): |
Myong-soon
Park, Sung-il Lim |
Abstract: |
The Internet is quickly growing and people who use the WWW
are increasing exponentially. So, companies which offer Web Service want
to service to clients during 365*24*60. Therefore they use the cluster
system for the availability and the performance. The previous works have
made the dispatcher do static position. So, if a node in the system is
failed the total system results in crash. We need to make it do dynamic
position as like SASHA (Scalable Application-Space Highly-Available)
Architecture. SASHA Architecture is composed of COTS, Application-Space
Software, Agent and Tokenbeat protocol for system administration. Because
it composes nodes in system by a virtual ring, the system administration’s
overhead happened. Our paper will propose improved fault Detection and
Reconfiguration performance in SASHA Architecture. |
|
Title: |
CUSTOMER
LOYALTY IN E-BUSINESS |
Author(s): |
Bel
G. Raggad, Jim Lawler |
Abstract: |
This study examines from simulation the effects of the
privacy sensitivity of customers, the personalization practices or
standards of retailers and the difficulty in locating favorable sites, on
the loyalty of consumers to a Web site. The key finding of the study is
that customer privacy sensitivity is a critical success factor that
significantly impacts loyalty to a retailer. Customers have higher loyalty
to sites that request the least information, while they have lower loyalty
to sites that request the most information. Web retailers considering
expanded personalization of products or services to customers, through
increased personal information, need to rethink their practices. The study
also found that difficulty in locating a favorable site is a success
factor that impacts retailer loyalty, and that customers have higher
loyalty to difficult to locate favorable sites on the Web. These findings
are important at a time when consumers are empowered with Web technology
to immediately shop competitor sites. The significance of privacy to
loyalty is a factor that needs to be considered seriously by retailers, if
they are to compete for loyal customers, and this study furnishes a
framework to effectively research loyalty, personalization and privacy on
the Web. |
|
Title: |
OPERATION-SUPPORT
SYSTEM FOR LARGE-SCALE SYSTEM USING INFORMATION TECHNOLOGY |
Author(s): |
Seiji Koide,
Riichiro Mizoguchi, Akio Gofuku |
Abstract: |
We are developing an operation support system for
large-scale system such as rocket launch using Information Technology. In
the project, we build a multi-media database that organizes diverse
information and data produced in designing, testing, and practical
launching, develop case-based and model-based trouble shooting algorithms
and systems that automatically detect anomaly and diagnose the causes
rapidly, and provide a fast networking environment that allows us to work
with experts in distance. The distributed collaborative environment in
which all of human operators and software agents can work collaboratively
is been developing by means of the Web servicing technology such as UDDI,
WSDL, and SOAP, and the Semantic Web technology such as RDF, RDFS, OWL,
and Topic Maps. This project was prepared with the contract under the
Japanese IT program of the Ministry of Education, Culture, Sports, and
Technology. |
|
Title: |
SIMULATION
STUDY OF TCP PERFORMANCE OVER MOBILE IPV4 AND MOBILE IPV6 |
Author(s): |
Jiankun Hu,
Damien Phillips |
Abstract: |
Mobile IPv6 (MIPv6) is a protocol to deal with mobility
for the next generation Internet (IPv6). However, the performance of MIPv6
has not yet been extensively investigated. Knowledge of how MIPv6 affects
TCP performance, especially in comparison with MIPv4, can provide
directions for further improvement. In this report, an intensive
simulation study of TCP performance over MIPv4 and MIPv6 has been
conducted. Simulation using the famous network simulator NS-2 will be used
to highlight differences when TCP is used in hybrid wireless environment,
over these two Mobile IP protocols. Initial simulation has shown a solid
improvement in performance for MIPv6 when IPv6 Route Optimisation features
are used. During the course of simulation, a consistent event causing
dropped TCP throughput was identified. Out of order arrival of packets
would occur when the mobile node initiated a handover. This out of order
arrival invokes TCP congestion control falsely which reduces throughput.
The difference in overall throughput of MIPv4 compared to MIPv6 is roughly
proportional to the difference in packet size attributed to IPv6's
increased header size. Another contribution of this work is to provide
modifications and new functions such as node processing time, to the NS-2
simulator to make such investigation possible. To the best of our
knowledge, no similar publication has been reported. |
|
Title: |
COLLABORATIVE
ENGINEERING PORTAL |
Author(s): |
KRISHNA
KUMAR RAMALEKSHMI SANKAR KUMAR, COLIN TAY, KHENG YEOW TAN,
STEVEN CHAN, YONGLIN LI, SAI KONG CHIN, ZIN
MYINT THANT, SAI KONG CHIN |
Abstract: |
The Collaborative Engineering Portal (CE-Portal) is
envisioned to be a comprehensive state-of-the-art infrastructure for
facilitating collaborative engineering over the Web. This system offers a
Web-based collaborative use of High Performance Computing and Networking
technology for product/process design that helps the enterprises to
shorten design cycles. This platform allows government professionals and
engineers to share information among themselves and to work together with
their private counterparts as a virtual project team. The Collaborative
Engineering portal is developed as a multi-tiered system implemented using
VNC and other Java technologies. In conclusions, we analyze strengths,
weaknesses, opportunities and threats of the approach. |
|
Title: |
A
SURVEY OF KNOWLEDGE BASE GRID FOR TRADITIONAL CHINESE MEDICINE |
Author(s): |
Jiefeng Xu,
Zhaohui Wu |
Abstract: |
Knowledge base gird is a kind of grid, which takes many
knowledge bases as its foundation and its knowledge sources. All these
knowledge sources follow a public ontology standard defined by standard
organization. Knowledge base grid has its own specific domain knowledge,
and so can be browsed at semantic level. It also supports correlative
browse and knowledge discovery. In this paper, we introduce a generic
knowledge base grid for Traditional Chinese Medicine. Its framework
consists of three main parts: Virtual Open Knowledge Base, Knowledge Base
Index, and Semantic Browser. We anatomize the implementation in detail.
Furthermore, knowledge presentation and services of knowledge base grid
are discussed. |
|
Title: |
TOWARDS
A SECURE MOBILE AGENT BASED M-COMMERCE SYSTEM |
Author(s): |
Ning Zhang,
Omaima Bamasak |
Abstract: |
It is widely agreed that mobile agent technology, with its
useful features, will provide the technical foundation for future
m-commerce applications, as it can overcome the wireless network
limitations of limited bandwidth, frequent disconnections and mobile
device's weaknesses. In order for mobile agents to be accepted as a
primary technology for enabling m-commerce, proper security mechanisms
must be developed to address the new security issues they bring to the
fore. The most challenging and difficult problem among them is the issue
of protecting mobile agents against malicious hosts. Although, to the best
of our knowledge, there is yet no general solution to this problem,
mechanisms that provide effective protection against specific attacks from
hostile hosts have been proposed. This paper has analysed the security
requirements for a mobile agent in the context of m-commerce, surveyed the
related work in relation to the requirements specified, and suggested the
development of a framework that provides confidentiality of data carried
by a mobile agent by using secret sharing scheme together with fair
exchange and non-repudiation services. |
|
Title: |
NON-REPUDIATION
AND FAIRNESS IN ELECTRONIC DATA EXCHANGE |
Author(s): |
Aleksandra Nenadic,
Ning Zhang |
Abstract: |
In this paper we discuss the two security issues:
non-repudiation and fairness in association with e-commerce applications.
In particular, these issues are addressed in the context of electronic
data exchange, which is one of the most commonly seen e-commerce
applications. In detail, this paper gives a survey of the approaches to
non-repudiation and fair electronic data exchange protocols. We
additionally discuss the current technologies that propose solutions to
these issues, and the emerging standards in the area of business data
formats and protocols for the exchange of such data. Finally, we discuss
the architecture layer at which to implement the protocols for
non-repudiation and fair data exchange. |
|
Title: |
SOMEONE:
A COOPERATIVE SYSTEM FOR PERSONALIZED INFORMATION EXCHANGE |
Author(s): |
Layda Agosto,
Laurence Vignollet, Pascal Bellec, Michel Plu |
Abstract: |
This paper presents a user-centric, social-media service:
SoMeONe. It's goal is to build an information exchange network using Web
informational networks. It should allow the construction of personal
knowledge bases whose quality is improved by collaboration. It tries to
increase the user's commitment by helping him to establish and to maintain
interesting interactions with enriching people. Although many users are
individualist, the rules we define for this media should encourage a
cooperative behaviour. The functionalities it offers are between a
bookmark management system and mailing lists. With SoMeONe users exchange
informartion with semantic addressing: they only need to annotate
information for being diffused to appropriate users. Each user interacts
only through a manually controlled contact network composed of known and
trusted users. However, to keep his contact network open, SoMeOne helps
each user to send information to new appropriate users. In return, the
user expects these users to send him new information as well. In
companies, where the Intranet is composed of huge amounts of heterogeneous
and diverse information, such collective behaviour should increase the
personal efficiency of each collaborator. Thus, SoMeONe provides some
solutions to some knowledge management problems particularly for companies
aware of the value of their social capital. |
|
Title: |
POTENTIAL
ADVANTAGES OF SEMANTIC WEB FOR INTERNET COMMERCE |
Author(s): |
Yuxiao Zhao
|
Abstract: |
Past decade saw much hype in the area of information
technology. The emerging of semantic Web makes us ask if it is another
hype. This paper focuses on its potential application in Internet commerce
and intends to answer the question to some degree. The contributions are:
first, we find and examine twelve potential advantages of applying
semantic Web for Internet commerce; second, we conduct a case study of
e-procurement in order to show its advantages for each process of
e-procurement; lastly, we identify critical research issues that may
transfer the potential advantages into tangible benefits. |
|
Title: |
BUSINESS
MODEL ANALYSIS APPLIED TO MOBILE BUSINESS |
Author(s): |
Giovanni Camponovo
|
Abstract: |
Mobile business is a young promising industry created by
the emergence of wireless data networks. Similar to other emerging
industries, it is characterized by a large number of uncertainties at
different levels, in particular concerning technology, demand and
strategy. This paper focuses on the strategic uncertainties, where a large
number of actors are trying a number of strategic approaches to position
themselves in the most favourable position in the value system. As a
consequence, they are experimenting with a number of innovating business
models. We argue that successful business models are likely to be the ones
that best address the economic peculiarities underlying this industry,
like mobility, network effects and natural monopolies. The paper presents
the principal classes of actors that will participate in the mobile
business industry and give an overview of their business models based on a
formalized ontology. |
|
Title: |
VOICEXML
APPLIED TO A WIRELESS COMMUNICATION SYSTEM |
Author(s): |
FRANK WANG
|
Abstract: |
The project aims to develop a wireless online
communication system (Wireless Messenger) to aid communication for
small-medium enterprises. By expressing automated voice services using
VoiceXML, a visual Web site is created in addition to the physical WML Web
site. This wireless system links an out-of-office mobile phone and an
in-house server. The functions of this system include posting and
notifying of messages internally, posting and notifying of notices,
setting and notifying of events, calendar reference modules and
administrative controls. |
|
Title: |
A
NEW SOLUTION FOR IMPEMENTATION OF A COLLABORATIVE BUSINESS PROCESS MODEL
|
Author(s): |
Takaaki Kamogawa,
Masao Matsumoto |
Abstract: |
This paper presents a Collaborative Business Process Model
based on a Synchronized Theory. The Cisco case of co-working with
suppliers is viewed in terms of business-process collaboration to identify
issues concerning collaboration with suppliers. The authors also discuss
past and present concepts of collaboration, and propose that it is
necessary to combine a synchronized theory with a collaborative business
process model. We propose a new solution for implementation of the
Collaborative Business Process Model from the viewpoint of open
infrastructure. |
|
Title: |
A
DESIGN PROCESS FOR DEPLOYING B2B E-COMMERCE |
Author(s): |
Youcef Baghdadi
|
Abstract: |
This paper emphasizes on architecture and design process
for developing applications to support B2B electronic commerce due to
their growth and difference from other categories of e-commerce in many
aspects. It first describes current architectures, reference models,
approaches and implementing technologies. It then proposes an architecture
with four abstraction levels: business process, decomposition and
coordination, supporting electronic commerce services, im-plementing
technology, and the interfaces between them. This abstraction aims to make
B2B e-commerce process-driven not technology-driven. Thus making business
process independent from the implementing technologies. Finally, a
five-steps design process in accordance with this architecture is
described. |
|
Title: |
AN
OBJECT ORIENTED IMPLEMENTATION OF BELIEF-GOAL-ROLE AGENTS |
Author(s): |
Walid
Chainbi |
Abstract: |
One of the most driving forces behind multi-agent systems
research and development is the Internet. Agents are populating the
Internet at an increasingly rapid pace. Unfortunately, they are almost
universally asocial. Accordingly, adequate agent concepts will be
essential for agents in such open environment. To address this issue, we
show in the first part of this paper that agents need to have
communication concepts and organization concepts. We argue that instead of
the usual approach of starting from a set of intentional states, the
intentional structure should be deduced in terms of interaction. To this
end, we come up with ontologies related to communication and organization.
The second part of this paper deals with a study which compares the agent
paradigm to the object paradigm. We also show the capabilities as well as
the limits of the object paradigm to deal with the agent paradigm. We
illustrate our work with the well known prey/predator game. |
|
Title: |
BUILDING
SUPPLY CHAIN RELATIONSHIPS WITH KNOWLEDGE MANAGEMENT: ENGINEERING TRUST IN
COLLABORATIVE SYSTEMS |
Author(s): |
John
Perkins, Ann-Karin Jorgensen, Lisa Barton, Sharon Cox
|
Abstract: |
Collaborative systems are essential components of
electronic supply chains. Barriers to collaboration are identified and a
preliminary model for evaluating its characteristic features is proposed.
Some features of knowledge management and knowledge management systems are
briefly reviewed and the application of these to the needs of
collaborative system evaluation is explored. A process for iterative
evaluation and review of collaborative system performance is proposed.
Finally, a case study in the retail industry is used to assess the
contribution of knowledge management concepts and systems to develop
improved e-commerce performance in collaborative value networks. |
|
Title: |
WIDAM
- WEB INTERACTION DISPLAY AND MONITORING |
Author(s): |
Hugo Gamboa,
Vasco Ferreira |
Abstract: |
In this paper we describe the design and implementation of
a system called Web Interaction Display and Monitoring (WIDAM). We have
developed a web based client-server application that offers several
services: (i) real time monitoring of the user interaction to be used in
synchronous playback (Synchronous Monitoring Service) (ii) real time
observation by other users (Synchronous Playback Service); (iii) storage
of the user interaction information in the server database (Recording
Service); (iv) retrieval and playback of a stored monitored interaction
(Asynchronous Playback Service). WIDAM allows the usage of an interaction
monitoring system directly over a web page, without the need of any
installation, using low bandwidth comparatively to image based remote
display systems. We discuss several applications of the presented system
like intelligent tutoring systems, usability analysis, system performance
monitoring, synchronous or asynchronous e-learning tools. |
|
Title: |
AN
AGENT-MEDIATED MARKETPLACE FOR TRANSPORTATION TRANSACTIONS |
Author(s): |
Alexis Lazanas,
Pavlos Moraitis, Nikos Karacapilidis |
Abstract: |
This paper reports on the development of an innovative
agent-mediated electronic marketplace, which is able to efficiently handle
transportation transactions of various types. Software agents of the
proposed system represent and act for any user involved in a
transportation scenario, while they cooperate and get the related
information in real-time mode. Our overall approach aims at the
development of a flexible framework that achieves efficient communication
among all parties involved, constructs the possible alternative solutions
and performs the required decision-making. The system is able to handle
the complexity that is inherent in such environments, which is mainly due
to the frequent need of finding a modular" transportation solution, that
is one that fragments the itinerary requested to a set of sub-routes that
may involve different transportation means (trains, trucks, ships,
airplanes, etc.). The system's agents cooperate upon well-specified
business models, thus being able to manage all the necessary freighting
and fleet scheduling processes in wide-area transportation networks. |
|
Title: |
ENGINEERING
MULTIAGENT SYSTEMS BASED ON INTERACTION PROTOCOLS: A COMPOSITIONAL PETRI
NET APPROACH |
Author(s): |
Sea Ling,
Seng Wai Loke |
Abstract: |
Multiagent systems are useful in distributed systems where
autonomous and flexible behaviour with decentralized control is
advantageous or necessary. To facilitate agent interactions in multiagent
systems, a set of interaction protocols for agents has been proposed by
the Foundation of Intelligent Physical Agents (FIPA). These protocols are
specified diagramatically in an extension of UML called AUML (Agent UML)
for agent communication. In this paper, we informally present a means to
translate these protocols to equivalent Petri net specifications. Our
Petri nets are compositional, and we contend that compositionality is
useful since multiagent systems and their interactions are inherently
modular, and so that mission-critical parts of a system can be analysed
separately. |
|
Title: |
ENHANCING
NEWS READING EXPERIENCE THROUGH PERSONALIZATION OF NEWS CONTENT AND
SERVICES USING INTELLIGENT AGENTS |
Author(s): |
Logandran Balavijendran,
Soon Nyean Cheong, Azhar Kassim Mustapha |
Abstract: |
One of the most common things we use the Internet for is
to read the news. But there is so much news catering for so many people,
that it often gets confusing and difficult to read what you want to read
about. This system uses an Intelligent Agent to guess what the user is
interested in and personalizes the news content. This is done by observing
the user and determining short-term and long-term interests. To further
enrich the experience, it provides features that allows the user to track
specific news events and receive instant alerts; summarize news so you can
take a quick look before committing yourself; find background information
to learn about the news; search and filter results according to the user
profile and also provides a smart download tool that makes viewing heavy
multimedia content practical without needing large bandwidth (by
exploiting the irregular nature of internet traffic and use). This agent
is designed to work of the News on Demand Kiosk Network[1] and designed
primarily in J2EE. |
|
Title: |
AN
INTERNET ENABLED APPROACH FOR MRO MODELS AND ITS IMPLEMENTATION |
Author(s): |
Dennis
F Kehoe, Zenon Michaelides, Peiyuan Pan |
Abstract: |
This paper presents an Internet enabled approach for MRO
applications based on the discussion on different MRO models and its
implementation architectures. This approach is focused on using e-business
philosophy and Internet technology to meet the requirements of MRO
services. The proposed e-MRO models are framework techniques. Different
system architectures for this new approach are described and available
technologies for system implementation are also presented. |
|
Title: |
A
NEW USER-ORIENTED MODEL TO MANAGE MULTIPLE DIGITAL CREDENTIALS |
Author(s): |
José Oliveira,
Augusto Silva, Carlos Costa |
Abstract: |
E-Commerce and Services are become a major commodity
reality. Aspects like electronic identification, authentication and trust
are core elements in referred web market areas. The use of electronic
credentials and the adoption of a unique worldwide-accepted digital
certificate stored in a smart card will provide a higher level of security
while allowing total mobility with secure transactions over the web. While
this adoption does not take place, the widespread use of digital
credentials will inevitably lead to each service client having to be in
possession of different electronic credentials needed for all the services
he uses. We present a new approach that provides a user-oriented model to
manage multiple electronic credential, based in utilization of only one
smart card per user as a basis for secure management of web-based
services, thus contributing for a more generalized use of the
technology. |
|
Title: |
INTELLIGENT
AGENTS SUPPORTED COLLABORATION IN SUPPLY CHAIN MANAGEMENT |
Author(s): |
Minhong WANG,
Huaiqing WANG, Huisong ZHENG |
Abstract: |
In today's global marketplace, individual firms no longer
compete as independent entities but rather as integral part of supply
chain links. This paper addresses the approach of applying the technology
of intelligent agent in supply chain management to cater for the
increasing demand on collaboration between supply chain partners. A
multi-agent framework for collaborative planning, forecasting and
replenishment in supply chain management is developed. With the concerns
for exception handling and flexible collaboration between partners, some
function are proposed in the system such as product activity monitoring,
negotiation within partners, supply performance evaluation, and
collaboration plan adjustment. |
|
Title: |
FIDES
- A FINANCIAL DECISION AID THAT CAN BE TRUSTED |
Author(s): |
Sanja Vranes,
Snezana Sucurovic, Violeta Tomasevic, Mladen Stanojevic,
Vladimir Simeunovic |
Abstract: |
FIDES is aimed at valuating investment projects in
accordance with the well-known UNIDO standard and making recommendations
on a preferable investment, based on multicriteria analysis of available
investment options. FIDES should provide a framework for analyzing key
financial indicators, using the discounted cash-flow technique, and also
allows for non-monetary factors to enter the multicriteria assessment
process, whilst retaining an explicit and relatively objective and
consistent set of evaluation conventions and clear decision criteria.
Moreover, since virtually every investment and financing decision,
involving allocation of resources under uncertain conditions, is
associated with considerable risk, FIDES should integrate the risk
management module. The basic principle governing risk management is
intuitive and well articulated, taking into account investor’s subjective
appetite for and aversion to risk, and the decision sensitivity to the
uncertainty and/or imprecision of input data. Thus, with FIDES, financial
analysts and decision-makers will be provided with effective modeling
tools in the absence of complete or precise information and the
significant presence of human involvement. The decision aid will be
implemented using multiple programming paradigms (Internet programming,
production rules, fuzzy programming, multicriteria analysis, etc.), using
a three-tier architecture as a backbone. Being Web based, the application
is especially convenient for large, geographically dispersed
corporations. |
|
Title: |
AGENT-BASED
GENERIC SERVICES AND THEIR APPLICATION FOR THE MOBILEWORKFORCE |
Author(s): |
Makram Bouzid
|
Abstract: |
In this paper we propose an architecture of agent-based
services for easy development of multi-agent applications. It is based on
the notion of service components, which can be installed (“plugged”) into
a communicative agent, and which can be composed in order to offer more
sophisticated services. This architecture was validated through the design
and development of a set of generic services for mobile workforce support,
within the European project LEAP. These generic services were also built
to develop two multi-agent applications that assist the mobile workers of
British Telecommunications and the German Automobile Club ADAC. Both have
been tested in real world conditions in UK and Germany. |
|
Title: |
AN
EXTENSIBLE TOOL FOR THE MANAGEMENT OF HETEROGENEOUS REPRESENTATIONS OF XML
DATA |
Author(s): |
Riccardo Torlone,
Marco Imperia |
Abstract: |
In this paper we present a tool for the management and the
exchange of structured data in XML format, described according to a
variety of formats and models. The tool is based on a novel notion of
``metamodel'' that embeds, on the one hand, the main primitives adopted by
different schema languages for XML and, on the other hand, the basic
constructs of traditional database conceptual models. The metamodel is
used as a level of reference for the translation between heterogeneous
data representations. The tool enables the users to deal, in a uniform
way, with various schema definition languages for XML (DTD, XML Schema and
others) and the ER model, as a representative of a traditional conceptual
model. The model translation facility allows the user to switch from one
representation to another and accounts for possible loss of information in
this operation. Moreover, the tool is easily extensible since new models
and translations can be added to the basic set in a natural way. The tool
can be used to support a number of involved e-Business activities like:
information exchange between different organizations, integration of data
coming from heterogeneous information sources, XML data design and
re-engineering of existing XML repositories. |
|
Title: |
USER
AUTHENTICATION FOR E-BUSINESS |
Author(s): |
James
P H Coleman |
Abstract: |
There are many factors that need to be addressed before
e-business is seen as a truly usable service to the ordinary customer. The
most well known factors are: · The speed of access to the Internet and
service providers · The cost of access to the Internet infrastructure. The poor quality of a large number of e-business/e-commerce web sites – in
particular aspects such as the interface, design … A less well-known, but
perhaps equally important factor is user authentication. User
authentication is the process whereby the Service Provider (SP) is able to
identify the person using the web site. This is normally done by a
username/password combination. User Authentication is important for the
SPs because if a product is ordered or a service is requested, then the
supplier needs to be reasonably confident that the order/request is valid,
and not a hoax. Unfortunately, the situation has arisen where a user who
is a frequent web user may have accounts with many different SPs, e.g.
Their bank, telephone company, ISP, superannuation/pension fund, insurance
company, government (often with different departments within the
Government) and so on. In these cases the SPs use a registration process
where the user has a username and password. It is unfortunately usually
the case that the username and password combinations are different between
sites. This is a deterrent to the whole registration process as you have
people with multiple registrations. There are many e-Gateway systems that
offer a single-point-of-logon, for example the e-Government within the UK
e-Government Project which aims to solve the problem at least within their
infrastructure. The very large private sector has no such mechanism. This
paper investigates current e-Gateway systems (including those where the
primary purpose is not necessarily user authentication) and proposes a
model for a more universal e-Gateway. |
|
Title: |
ON
ESTIMATING THE AMOUNT OF LEARNING MATERIALS A CASE STUDY |
Author(s): |
Matti Järvenmpää,
Pasi Tyrväinen, Ari Sievänen |
Abstract: |
E-learning has been studied as the means to apply digital
computers to educational purposes. Although the benefits of information
and communication technology are obvious in several cases, there still
exists a lack of convincing measures for the value of using computers in
education. This reflects the general difficulty in evaluating investments
on information systems, known as the "IT investment paradox" that has not
been solved so far. In this paper we approach the problem by estimating
the amount of teaching and learning material in a target organisation, a
university faculty. As expected, the volume of learning material dominates
the communication of the faculty forming about 95% of all communication
volume and 78% to 82% of communication when measured with other metrics.
Also the use of alternative communication forms used in the target
organisation was analysed quantitatively. The study also indicates, that
communication forms dominating the volume of communication are likely to
be highly organisation-specific. |
|
Title: |
E-COMMERCE
ENGINEERING: A SHORT VS LONG SOFTWARE PROCESS FOR THE DEVELOPMENT OF
E-COMMERCE APPLICATIONS |
Author(s): |
Andreas Andreou,
Stephanos Mavromoustakos, Chrysostomos Chrysostomou, George
Samaras , Andreas Pitsillides, Christos Schizas,
Costas Leonidou |
Abstract: |
The immediacy in developing e-commerce applications, the
quality of the services offered by these systems and the need for
continuous evolution are primary issues that must be fully analysed and
understood prior and during the development process. In this context, the
present work suggests a new development framework which aims at estimating
the level of complexity a certain e-commerce system encompasses and
driving the selection of a long or short software process in terms of time
and effort. The proposed framework utilizes a special form of Business
Process Re-engineering (BPR) to define and assess critical business and
organizational factors within small-to-medium enterprises (SMEs) whishing
to go e-commerce. This set of factors is enriched with other critical
issues belonging to the quality requirements of the system and to the
application type of services it aspires to offer. The set of critical
factors identified is used to estimate the average complexity level of the
system using numerical values to describe the contribution of each factor
to the overall complexity. The level of complexity estimated dictates the
adoption of either a short or a long version of the well-known WebE
process for analysing, designing and implementing the e-commerce system
required by an SME. |
|
Title: |
ARCHITECTURE
OF AUTOMATIC RECOMMENDATION SYSTEM IN E-COMMERCE |
Author(s): |
Rajiv Khosla,
Qiubang Li |
Abstract: |
Automatic recommendation system will become an
indispensable tool for customers to shop online. This paper proposes an
architecture of automatic recommendation system in e-commerce. The
response time of the system, which is the bottleneck of the system, is
overcome by high performance computing. The architecture has already
applied to an online banking system. |
|
Title: |
ELECTRONIC
JOB MARKETPLACES: A NEWLY ESTABLISHED MANIFESTATION OF
E-BUSINESS |
Author(s): |
Georgios Dafoulas,
Mike Turega, Athanasios Nikolaou |
Abstract: |
Finding suitable candidates for critical job posts is
currently an issue of concern for most organizations. Consideration of
cultural fit, experience, ability to adapt to the company’s marketplace
and ability to grow with the organisation all weigh heavily on the minds
of most human resource professionals. Since the mid-90s a significant
number of recruiting firms started exploiting the Internet mainly because
of its global nature that provides access to an unlimited pool of skills.
Optimistic estimations examine the Internet as a medium for conducting the
recruitment and selection process in an online environment. This paper
suggests developing an integrated Electronic Job Marketplace offering a
new service in the Internet Job Market: Online Interviewing for screening
candidate employees. In order to meet hiring objectives and control the
increasing cost of recruiting, organisations could implement an online
recruiting and selection process. The critical requirements of the new
model are: eliminating paperwork, improving time-to-hire, reducing
turnover, creating a resume and position-centric environment as well as
using the Internet as a recruitment and selection tool. |
|
Title: |
ONE-TO-ONE
PERSONALIZATION OF WEB APPLICATIONS USING A GRAPH BASED MODEL |
Author(s): |
Georg Sonneck,
Thomas Mück |
Abstract: |
Due to the maturity of current web technology, a large
fraction of non-technically oriented IT end users are confronted with
increasingly complex web applications. Such applications should help these
end users to fulfill their tasks in the most effective and efficient way.
Out of this perspective there is little doubt that personalization issues
play an important role in the era of web applications. Several approaches
already exist to support so{-}called {\em Adaptive Hypermedia Systems},
i.e., systems which are able to adapt their output behaviour to different
user categories. In this paper, we are focusing on those personalization
and customization issues of web applications raised by task driven {\em
user interaction} and give as example the interaction patterns caused by
different users of a financial advisor system. To achieve this goal we
propose, in a first step, a graph{-}based model representing the logical
structure of web applications, a fully extensible XML schema description
modelling the structure of the nodes in the graph and a document type
definition to store user profiles. In a second step, this basic model is
augmented by process graphs corresponding to specific business tasks the
web application can be used for, leading to a first form of
personalization by assigning a user to a process task. We then show in a
final step how matches between stored skills within the user profile and
the node descriptions can lead to one{-}to{-}one personalization of the
process graph. |
|
Title: |
AN
INVESTIGATION OF THE NEGOTIATION DOMAIN FOR ELECTRONIC COMMERCE
INFORMATION SYSTEMS |
Author(s): |
Zlatko Zlatev,
Pascal Eck, van |
Abstract: |
To fully support business cycles, information systems for
electronic commerce need to be able to conduct negotiation automatically.
In recent years, a number of general frameworks for automated negotiation
have been proposed. Application of such frameworks in a specific
negotiation situation entails selecting the proper framework and adapting
it to this situation. This selection and adaptation process is driven by
the specific characteristics of the situation. This paper presents a
systematic investigation of there characteristics and surveys a number of
frameworks for automated negotiation. |
|
Title: |
COLLABORATOR
- A COLLABORATIVE SYSTEM FOR HETEROGENEOUS NETWORKS AND DEVICES |
Author(s): |
Agostino Poggi,
Matteo Somacher, Socrates Costicoglou, Federico Bergenti
|
Abstract: |
This paper presents a software framework, called
Collaborator, to provide a shared workspace supporting the activities of
virtual teams. This system exploits seamless integration of standard Web
technologies with agent technologies, enhancing the classic Web
communication mechanisms to support synchronous sharing of applications,
and its use through emerging technologies such as: third generation of
mobile networks and terminals, and new generation of home appliances. The
system presented in the paper is the main result of an on-going European
research project Collaborator (IST-2000-30045) that aims at specifying and
developing a software distributed environment to support efficient
synchronous collaborative work between virtual teams, and will experiment
such an environment in the construction and telecommunication working
sectors. |
|
Title: |
SOFTWARE
AGENTS TO SUPPORT ADMINISTRATION IN ASYNCHRONOUS TEAM
ENVIRONMENTS |
Author(s): |
Roger Tagg
|
Abstract: |
Current economic pressures are causing severe problems for
many enterprises in maintaining service standards with shrinking
headcounts. Front-line workers have had to shoulder runaway workloads.
Software Agent technologies have been widely advocated as a solution, but
there are few reported success stories. In the author’s previous work, a
design was been proposed for a system to support front-line staff in a
team teaching environment. This system is based on a domain-specific
screen desktop with drop boxes supported by a number of types of agent.
This paper analyses the work these agents have to do and the technology
needed to support them. |
|
Title: |
IT
INFRASTRUCTURE FOR SUPPLY CHAIN MANAGEMENT IN COMPANY NETWORKS WITH SMALL
AND MEDIUM-SIZED ENTERPRISES |
Author(s): |
Marcel Stoer,
Joerg Nienhaus, Nils Birkeland, Guido Menkhaus |
Abstract: |
The current trend of extending supply chain management
beyond the company's wall focuses on the integration of suppliers and
consumers into a single information network. The objective is to optimize
costs and opportunities for everyone involved. However, small-sized
enterprises can rarely carry the high acquisition and introduction costs
of hardware and software. This reduces the attractiveness of the
small-sized enterprise as partner in a logistics and a production network.
This article presents a lean IT infrastructure that targets small-sized
enterprises. It allows flexible and configurable integration with the
Internet, ERP systems and the secure communication of supply chain
management data. |
|
Title: |
AGENTS-MIDDLEWARE
APPROACH FOR CONTEXT AWARENESS IN PERVASIVE COMPUTING |
Author(s): |
Karim Djouani,
Abdelghani CHIBANI, Yacine AMIRAT |
Abstract: |
With the emergence of wireless distributed systems,
embedded computing is becoming more pervasive. Users in continuous
transitions between handheld devices and fixed computers expect to
maintain the same QoS. Thus, applications need to become increasingly
autonomous by reducing interactions with users. The present paper caters
with user’s mobility, context-aware embedded applications, distributed
systems, and in the general case accesses to remote services through
embedded middleware. The context, in which exist such applications,
exhibits some constraints like: low bandwidth, frequent disconnections,
resources poor devices (low CPU speed, little memory, low battery power,
etc). The first objective of our work is to proof that agent paradigm and
technologies present a great potential to fully blossom in this new area.
This allows the building of new and more effective pervasive applications.
Our vision, beyond what it was given in middleware and agents for
pervasive computing research, is including the context-awareness
capability into the early-introduced agents-middleware approach. Thus, we
have proposed an agents-middleware architecture approach, which is FIPA
standard compliant. This approach is a logical suite of some transitions
in research results; from embedded middleware approaches to lightweight
agents’ platform approaches, and arriving finally to context-aware
agents-middleware approach. In this way, we present the usefulness of
context notion through two derived concepts: pervasive context and user
profile. Upon we have introduced tow specialized agents, within the
agents-middleware, that process by inferring meta-data that describe
context information extracted from sources like: sensors, user, system
resources, wireless network, etc. on top of this agents-middleware we can
build context-aware pervasive applications. We present also our ongoing
work and the future targeted applications by our approach. |
|
Title: |
TOXIC
FARM: A COOPERATIVE MANAGEMENT PLATFORM FOR VIRTUAL TEAMS AND ENTERPRISES
|
Author(s): |
Hala Skaf-Molli,
Pascal Molli, Pradeep Ray, Fethi Rabhi, Gerald Oster
|
Abstract: |
The proliferation of the Internet has revolutionized the
way people work together for business. People located at remote places can
collaborate across organizational and national boundaries. Although the
Internet provides the basic connectivity, researchers all over the world
are grappling with the problems of defining, designing and implementing
web services that would help people collaborate effectively in virtual
teams and enterprises. These problems are exacerbated by a number of
issues, such as coordination, communication, data sharing, mobility and
security. Hence there is a strong need for multiple services (to address
above issues) though an open cooperative management platform to support
the design and implementation of virtual teams and enterprises in this
dynamic business environment. This paper presents a cooperative management
platform called Toxic Farm for this purpose and discusses its application
in business applications. |
|
Title: |
LEARNING
USER PROFILES FOR INTELLIGENT SEARCH |
Author(s): |
Pasquale Lops,
Marco Degemmis |
Abstract: |
The recent evolution of e-commerce emphasized the need for
more and more receptive services to the unique and individual requests of
users. Personalization has became an important strategy in Business to
Consumer commerce, where a user explicitly wants the e-commerce site to
consider his own information such as preferences in order to improve
access to relevant products. By analyzing the information provided by a
customer, his browsing and purchasing history, a personalization system
could learn a customer's personal preferences and store them in a personal
profile used to provide an intelligent search support. In this work, we
propose a two-step profiles generation process: in the first step, the
system learns coarse-grained profiles in which the preferences are the
product categories the user is interested into. In the second step, the
profiles are refined by a probabilistic model of each preferred product
category, induced from the descriptions of the products the user likes.
Experimental results demonstrate the effectiveness of the strategy
proposed. |
|
Title: |
AGENT
COMMUNICATION CHANNELS: TRANSPORT MECHANISMS |
Author(s): |
Qusay Mahmoud
|
Abstract: |
Most of the work that has been done on agent communication
has concentrated on ontologies – Agent Communication Languages (ACLs) that
are used to describe objects that the agents manipulate. Little attention,
if any, has been given to agent communication channels – the transport
layer through which data is sent between agents. Here we describe the
different communication transport techniques that can be used to send data
between agents, and then we will compare and contrast the different
transport mechanisms. This is important as the way agents communicate can
have a significant effect on the performance of agent-based systems. |
|
Title: |
IMPLEMENTING
AN INTERNET-BASED VOTING - A PROJECT EXPERIENCE |
Author(s): |
Alexander Prosser,
Robert Krimmer, Robert Kofler |
Abstract: |
Worldwide research groups have developed remote electronic
voting systems using several different approaches with no legal basis. In
2001 the Austrian Parliament passed a law allowing electronic voting with
digital signatures for public elections. Besides these legal requirements,
an algorithm has to solve the basic technical problem, of how to identify
the user uniquely with still guaranteeing the anonymity of one’s vote and
further not to allow fraud by the election administration. In this paper
the authors give an experience report on the implementation of the first
phase of an algorithm that fulfills these requirements by strictly
separating the registration from the vote submission phase. |
|
Title: |
TOWARDS
THE ENTERPRISES INFORMATION INFRASTRUCTURE BASED ON COMPONENTS AND
AGENTS |
Author(s): |
Manuel Chi,
Ernesto German, Matias Alvarado, Leonid Sheremetov,
Miguel Contreras |
Abstract: |
Information infrastructure as the mean to bring together
software applications within the enterprise is the key component to enable
cooperation, information and knowledge exchange in an open distributed
environment. In this article, component and agent paradigms for the
integration of virtual enterprises are analyzed and the advantages and
drawbacks of the proposed solution are discussed. As an example of the
infrastructure as an integration of the both technologies, a Component
Agent Platform (CAP) that uses DCOM as a particular case of component
model for its implementation is described. Finally, we discuss the
interoperability issues of the proposed solution and outline the
directions of the future work. |
|
Title: |
GUARDIAN
KNOWLEDGE FARM AGENTS AND SECURITY ARCHITECTURES: WEB SERVICES, XML, AND
WIRELESS MAPPINGS |
Author(s): |
Britton Hennessey,
Girish Hullur, Mandy McPherson, George Kelley |
Abstract: |
This paper merges the BDIP (beliefs, desires, intentions,
and plans) rational agent model into the Jungian rational behavioral
model. It also defines the key framework design dimensions and classified
intelligences of knowledge farm network agents having the necessary
know-how to function as trust and security guardians. The paper presents
four practical example application mappings of the converged BDIP-Jungian
framework into (1) seven design principles of computer systems security,
(2) the web services security architecture, (3) the XML family systems
security architecture, and (4) the wireless security architecture. |
|
Title: |
ICS-
AN AGENT MEDIATED E-COMMERCE SYSTEM: ONTOLOGIES USAGE |
Author(s): |
Sofiane Labidi
|
Abstract: |
The Electronic Commerce has presented an exponential
growth in relation to the number of users and amount of commercial
transactions. Recent advances in Software Agent’s technology allow
agent-based electronic commerce where agents are entities acting
autonomously (or semi-autonomously) on behalf of companies or people in
negotiation into virtual environments. In this work, we propose the ICS
(an Intelligent Commerce System) as a B2B E-Commerce system based on
intelligent and mobile software agent’s technology following the OMG MASIF
standard. Three important features of ICS are emphasized here: the
e-commerce lifecycle approach, the user modeling, and a proposed ontology
for each phase of the lifecycle. |
|
Title: |
IMPLEMENTATON
OF MOBILE INFORMATION DEVICE PROFILE ON VIRTUAL LAB |
Author(s): |
Aravind
Kumar Alagia Nambi |
Abstract: |
The rate at which information is produced in today’s world
is mind-boggling. The information is changing by every minute and today’s
corporate mantra is not “knowledge is power” but “Timely knowledge is
power “. Millions of Dollars are won or lost due to information or lack of
it. Business executives and corporate managers push their technology
managers to provide information at the right time in the right form. They
want information on the go and want to be connected all the time to the
Internet or their corporate network. The rapid advancement of Technology
in the field of miniaturization and that of communications has introduced
a lot of roaming devices for people to connect through to the network like
laptop, PDA, mobile phones and many embedded devices. Programming for
these devices were cumbersome and limited since each device supported
their own standard I/O ports, screen resolution and had specific
configurations. The introduction of Java 2 Micro Edition (J2ME) has solved
this problem to some extent. J2ME is divided into configuration and
profiles, which provide specific information to a group of related
devices. Mobile phones can be programmed using J2ME. If the mobility
offered by the cellular phones combined with Electrical Engineering many
new uses can be found out for existing electrical machines. It will also
enable remote monitoring of electrical machines and the various parameters
involved in Electrical Engineering |
|
Title: |
GLI-BBS:
A GROUPWARE BASED ON GEOGRAPHICAL LOCATION INFORMATION FOR FIELD
WORKERS |
Author(s): |
Tatsunori Sasaki,
Naoki Odashima, Akihiro Abe |
Abstract: |
Geographical Location Information (GLI) is information
showing in which geographical position a person or an object is located.
Using digital maps and digital photographs, we have developed a GLI-based
Bulletin Board System (GLI-BBS), and we are promoting applications for
various public works in local communities. Fieldworkers who participate in
public works can use the GLI-BBS effectively to share information and to
form mutual agreement. As examples of concrete GLI-BBS applications, a
support system for road maintenance and management operations are taken up
to examine important points in operation. |
|
Title: |
SECURING
INTERNET SERVERS FROM SYN FLOODING |
Author(s): |
Riaz Mahmood
|
Abstract: |
Denial-of-Service (DoS) attacks utilize the
vulnerabilities present in current Internet protocols and target end
server machines with flood of bogus requests – thus blocking the services
to the legitimate users. In this paper a counter denial-of-service method
called Dynamic Ingress Filtering Algorithm (DIFA) is introduced. This
algorithm aims to remove the network peripheries inability to counter
spoof-based denial-of-service attacks originating from valid network
prefixes. Dynamic Ingress Filtering mechanism by virtue of its design,
gives protection against both types of spoof-based attacks, generating
from valid network prefixes and invalid network prefixes. This is because
of the reason that incoming traffic IP addresses change rate is compared
with predefined threshold time limit. If the addresses are changing
rapidly from a particular source – the packets arriving from that host are
not forwarded. Advantages of DIFA include design simplicity, scalability
and reasonable implementation costs |
|
Title: |
WEB
SERVICES SECURITY MODEL BASED ON TRUST |
Author(s): |
Luminita Vasiu
|
Abstract: |
The concept of Web services is the latest in the evolution
of ever more modular and distributed computing. Web services represent a
fairly simple extension to existing component models, such as Microsoft's
Component Object Model (COM) or Sun's Enterprise Java Bean (EJB)
specification .It is obvious that Web services have what it takes to
change something important in the distributed programming field. But,
until they do it developers will have some difficulties in figuring out
how to solve and eliminate problems that appear when trying to build
heterogeneous applications.In an open environment security is always an
issue. In order to overcome this problem the main challenge is to
understand and asses the risk involved in securing a Web-based service.
How do you guarantee the security of a bank transaction service? There are
efforts being made to develop security mechanisms for Web services.
Standards like SAML, XKMS, SOAP security will probably be used in the
future to guarantee protection for both the consumers and the services. In
this paper we analyse some security issues faced by Web services and
present a security model based on trust which supports more specific
models such as identity-based-security, access control lists. |
|
Title: |
A
MULTI-AGENT ARCHITECTURE FOR DYNAMIC COLLABORATIVE FILTERING |
Author(s): |
Gulden Uchyigit,
Keith Clark |
Abstract: |
Collaborative Filtering systems suggest items to a user
because it is highly rated by some other user with similar tastes.
Although these systems are achieving great success on web based
applications, the tremendous growth in the number of people using these
applications require performing many recommendations per sec-ond for
millions of users. Technologies are needed that can rapidly produce high
quality recommendations for large community of users. In this paper we
present an agent based approach to collaborative filtering where agents
work on behalf of their users to form shared “interest groups”, which is a
process of pre-clustering users based on their interest profiles. These
groups are dynamically updated to reflect the user’s evolving interests
over time. We further present a multi-agent based simulation of the
architecture as a means of evaluating the system. |
|
Title: |
POLICIES
COMPOSITION THROUGH GRAPHICAL COMPONENTS |
Author(s): |
Rui
Lopes, Vitor Roque, Jose Luis Oliveira |
Abstract: |
Policy based management have gained a crescent importance
in the two last years. New demands on internetworking, on services
specification, on QoS achievement and generically on network management
functionality, have driven this paradigm to a very important level. The
main idea is to provide services that allow specifying management and
operational rules in the same way people do business. Despite the main
focus of this technology has been associated with network management
solutions, its generality allows to extend these principles to any
business process inside an organization. In this paper we discuss the main
proposals in the field, namely the IETF/DMTF model, and we present a
proposal that allows the specification of policy rules through a
user-friendly and component-oriented graphical interface. |
|
Title: |
TOWARDS
AGENT BASED BUSINESS INFORMATION SYSTEMS AND PROCESS MANAGEMENT |
Author(s): |
Johann Sievering,
Jean-Henry Morin |
Abstract: |
Todays Business Information Systems and Business
Intelligence applications have become key instruments of corporate
management. They have evolved over time to a mature discipline within IT
departments. However, they appear to be slow at integrating emerging
technologies offering major improvements to trading partners in the global
networked ecosystem. The Internet is slow-ly evolving towards Peer-to-Peer
architectures and grid computing, Agent-Oriented Program-ming, Digital
Rights and Policy Management, trusted computing, ontologies and semantics.
These evolutions are setting the ground and requirements for the future of
corporate IT. This pa-per reports on current investigations and
developments on this issue making the case for the in-tegration of
emerging technologies in Business Information Systems. In particular,
mobile agents and peer-to-peer computing offer major advantages in terms
of technical architectures as well as a programming paradigm shift. We are
currently working on a framework addressing these issues towards Active
Business Objects. |
|
Title: |
ANALYSIS
OF BUSINESS TO BUSINESS ELECTRONIC MARKETS IN CHINA: THEORETICAL AND
PRACTICAL PERSPECTIVES |
Author(s): |
Jing Zhao
|
Abstract: |
In China, electronic markets (e-markets) are in the early
stages of development. It has unique characteristics in e-commerce
activities and market mechanisms, which are largely a function of the
current industry structure, financial infrastructure and organization
structure. This paper addresses an interactive e-market space view and
proposes the interactive e-commerce model for studying e-commerce
activities and strategies in e-markets of China. Building on this
theoretical insight the model draws attention to the e-commerce process in
which buyers and sellers, virtual market manager and its business partners
are linked and in which web-based communication and collaboration take
place, and to the adopted innovative market mechanisms. The e-commerce
process can be modelled by separating main business activities into four
phases designed to exploit business opportunities. The model is applied to
analyse one successful B2B Exchange in China. It offers an effective
approach to studying dynamic structure of transaction and a high
performance e-commerce strategy. Our research identifies the four lever of
e-market capability. These abilities imply e-market potential to achieving
and sustaining a new level of e-commerce strategy performance, and a more
competitive position in a rapidly changing B2B electronic market of
China. |
|
Title: |
MEMBERSHIP
PORTAL AND SERVICE PROVISIONING SYSTEM FOR AN INFRASTRUCTURE OF
HUBS:MANAGED E-HUB |
Author(s): |
Liang-Jie Zhang,
Henry Chang, Zhong Tian, Shun Xiang Yang, Ying
Nan Zuo, Jing Min Xu, Tian Chao |
Abstract: |
The goal of Managed e-Hub research prototype is to build a
common infrastructure of hubs so that businesses can develop B2B exchanges
meeting their business needs based on it. In this paper, an open and
extensible framework for Managed e-Hub is presented and the hub
fundamental services are discussed in detail as well. The service
provisioning system of Managed e-Hub not only provides a way of
integrating other services into the hub by means of service on-boarding
and subscription, but also provisions these services with their required
provisioning information. |
|
Title: |
APPLICATION
SCENARIOS FOR DISTRIBUTED MANAGEMENT USING SNMP EXPRESSIONS |
Author(s): |
Rui Lopes
|
Abstract: |
Management distribution is an, we can say, old topic in
terms of the number of proposed solutions and publications. Recently, the
DISMAN workgroup suggested a set of MIB modules to address this matter in
the context of SNMP. One of the DISMAN modules has the capability of using
expressions to perform decentralized processing of management information
– the Expression MIB. Although existing for some time now, its
capabilities are not very well known. In fact, other DISMAN MIBs, such as
the Schedule MIB and the Script MIB already got some attention in several
papers and are target of very solid work. There are hardly any papers
describing the Expression MIB and its functionality. This paper
contributes to eliminate this absence by describing our implementation
effort around it as well as some real world applications for it. |
|
Title: |
AGENTAPI:
AN API FOR THE DEVELOPMENT OF MANAGED AGENTS |
Author(s): |
Rui Lopes
|
Abstract: |
Managed agents, namely SNMP agents, costs too much to
develop, test and maintain. Although assuming simplicity since its
origins, the SNMP model has several intrinsic aspects that make the
development of management applications a complex task. However, there are
tools available which intend to simplify this process by generating
automatic code based on the management information definition.
Unfortunately, these tools are usually complicated to use and require a
strong background of programming experience and network management
knowledge. This paper describes an API for managed agent development which
also provides multiprotocol capabilities. Without changing the code, the
resulting agent can be managed by SNMP, web browsers, wap browsers, CORBA
or any other access method either simultaneously or individually. |
|
Title: |
AUTOMATIC
E-COMMERCE USING A MOBILE AGENTS MODEL |
Author(s): |
Francesco
Maria Raimondi, Salvatore Pennacchio |
Abstract: |
Electronic commerce business to business using mobile
agents is one of the most important future promise and also one good
result of global mobility code. As we will show, classic commerce model
and electronic commerce model both introduce advantages and disadvantages.
Electronic commerce through mobile agents has an objective that is to
eliminate defects and to arrange advantages of previous models. In
particular it takes cue of selling negotiation, in which it is necessary
to take decisions. |
|
Title: |
A
TIME ZONE BASED DYNAMIC CACHE REPLACEMENT POLICY |
Author(s): |
Srividya Gopalan,
Kanchan Sripathy, Sridhar Varadarajan |
Abstract: |
This paper proposes a time zone based novel cache
replacement policy, LUV, intended for web traffic in the context of a
hybrid cache management strategy. The LUV replacement policy is based on
ranking of web objects on a set of metrics intercepted by a proxy server.
Further, in order to maximize the hit rate, the cache replacement policy
makes use of immediate past access patterns for individual web objects
with respect to various time zones. |
|
Title: |
BOOTSTRAPPING
THE SEMANTIC WEB BY DEVELOPING A MULTI-AGENT SYSTEM TO FACILITATE ONTOLOGY
REUSE:A RESEARCH AGENDA |
Author(s): |
Abir
Qasem |
Abstract: |
Ontologies are basic components of the Semantic Web but
are difficult to build, and this acts as a bottleneck in the spread of the
Semantic Web. Reuse is seen as one of the solution to this problem. This
paper addresses the feasibility of a multi-agent system that will
automatically identify the appropriate reusable ontologies and thereby
greatly reduce the burden of its users. First, the area of automated
software component reuse is reviewed and borrowed from in order to develop
an appropriate framework. Next, a research agenda is proposed for
developing this type of multi agent system for ontology reuse. Finally it
is argued that the proposed multi-agent system will enable faster
deployment of the Semantic Web by making the ontology development process
efficient and developed ontologies, more robust and interoperable. This
use of agents may help to bootstrap the Semantic Web itself by leveraging
from the emerging Semantic Web architecture, and contributing to its
growth. |
|
Title: |
A
DYNAMIC AND SCALABLE AGENT-BASED APPROACH FOR KNOWLEDGE DISCOVERY : WEB
SITE EXPLORATION |
Author(s): |
Aurelio López
López, Alberto Méndez Torreblanca |
Abstract: |
The World Wide Web has become an open world of information
with a continuous growth.This dynamic nature is causing several
difficulties for discovering potentially useful Knowledge from the web.
The techniques of web mining and software agents can be combined for
resolving the problem. In this paper, we propose a dynamic and scalable
agent-based approach for knowledge discovery from specific web sites,
where the information is constantly added or eliminated, or its structure
is permanently modified.We also report preliminary results of the approach
for the exploration of web sites. |
|
Title: |
INTELLIGENT
SOFTWARE AGENTS IN THE KNOWLEDGE ECONOMY: |
Author(s): |
Mahesh
S. Raisinghani |
Abstract: |
Intelligent agent technology is emerging as one of the most
important and rapidly advancing areas in information systems and
e-business. There is a tremendous explosion in the development of
agent-based applications in a variety of fields such as electronic
commerce, supply chain management, resource allocation, intelligent
manufacturing, industrial control, information retrieval and filtering,
collaborative work, decision support, and computer games. While research
on various aspects of intelligent agent technology and its application is
progressing at a very fast pace, this is only the beginning. There are
still a number of issues that have to be explored in terms of agent
design, implementation, and deployment. For example, salient
characteristics of agents in different domains, formal approaches for
agent-oriented modeling, designing and implementing agent-oriented
information systems, agent collaboration and coordination, and
organizational impact of agent-based systems are some of the areas in need
of further research. The purpose of this paper is to identify and explore
the issues, opportunities, and solutions related to intelligent agent
modeling, design, implementation, and deployment. |
Page Updated on 27-05-2003
Copyright © Escola Superior de Tecnologia
de Setúbal, Instituto Politécnico de Setúbal |