Excellence in Research and Innovation for Humanity

International Science Index

Commenced in January 1999 Frequency: Monthly Edition: International Paper Count: 172

Automatic Removal of Ocular Artifacts using JADE Algorithm and Neural Network
The ElectroEncephaloGram (EEG) is useful for clinical diagnosis and biomedical research. EEG signals often contain strong ElectroOculoGram (EOG) artifacts produced by eye movements and eye blinks especially in EEG recorded from frontal channels. These artifacts obscure the underlying brain activity, making its visual or automated inspection difficult. The goal of ocular artifact removal is to remove ocular artifacts from the recorded EEG, leaving the underlying background signals due to brain activity. In recent times, Independent Component Analysis (ICA) algorithms have demonstrated superior potential in obtaining the least dependent source components. In this paper, the independent components are obtained by using the JADE algorithm (best separating algorithm) and are classified into either artifact component or neural component. Neural Network is used for the classification of the obtained independent components. Neural Network requires input features that exactly represent the true character of the input signals so that the neural network could classify the signals based on those key characters that differentiate between various signals. In this work, Auto Regressive (AR) coefficients are used as the input features for classification. Two neural network approaches are used to learn classification rules from EEG data. First, a Polynomial Neural Network (PNN) trained by GMDH (Group Method of Data Handling) algorithm is used and secondly, feed-forward neural network classifier trained by a standard back-propagation algorithm is used for classification and the results show that JADE-FNN performs better than JADEPNN.
Levenberg-Marquardt Algorithm for Karachi Stock Exchange Share Rates Forecasting

Financial forecasting is an example of signal processing problems. A number of ways to train/learn the network are available. We have used Levenberg-Marquardt algorithm for error back-propagation for weight adjustment. Pre-processing of data has reduced much of the variation at large scale to small scale, reducing the variation of training data.

A Self Supervised Bi-directional Neural Network (BDSONN) Architecture for Object Extraction Guided by Beta Activation Function and Adaptive Fuzzy Context Sensitive Thresholding
A multilayer self organizing neural neural network (MLSONN) architecture for binary object extraction, guided by a beta activation function and characterized by backpropagation of errors estimated from the linear indices of fuzziness of the network output states, is discussed. Since the MLSONN architecture is designed to operate in a single point fixed/uniform thresholding scenario, it does not take into cognizance the heterogeneity of image information in the extraction process. The performance of the MLSONN architecture with representative values of the threshold parameters of the beta activation function employed is also studied. A three layer bidirectional self organizing neural network (BDSONN) architecture comprising fully connected neurons, for the extraction of objects from a noisy background and capable of incorporating the underlying image context heterogeneity through variable and adaptive thresholding, is proposed in this article. The input layer of the network architecture represents the fuzzy membership information of the image scene to be extracted. The second layer (the intermediate layer) and the final layer (the output layer) of the network architecture deal with the self supervised object extraction task by bi-directional propagation of the network states. Each layer except the output layer is connected to the next layer following a neighborhood based topology. The output layer neurons are in turn, connected to the intermediate layer following similar topology, thus forming a counter-propagating architecture with the intermediate layer. The novelty of the proposed architecture is that the assignment/updating of the inter-layer connection weights are done using the relative fuzzy membership values at the constituent neurons in the different network layers. Another interesting feature of the network lies in the fact that the processing capabilities of the intermediate and the output layer neurons are guided by a beta activation function, which uses image context sensitive adaptive thresholding arising out of the fuzzy cardinality estimates of the different network neighborhood fuzzy subsets, rather than resorting to fixed and single point thresholding. An application of the proposed architecture for object extraction is demonstrated using a synthetic and a real life image. The extraction efficiency of the proposed network architecture is evaluated by a proposed system transfer index characteristic of the network.
A Text Clustering System based on k-means Type Subspace Clustering and Ontology

This paper presents a text clustering system developed based on a k-means type subspace clustering algorithm to cluster large, high dimensional and sparse text data. In this algorithm, a new step is added in the k-means clustering process to automatically calculate the weights of keywords in each cluster so that the important words of a cluster can be identified by the weight values. For understanding and interpretation of clustering results, a few keywords that can best represent the semantic topic are extracted from each cluster. Two methods are used to extract the representative words. The candidate words are first selected according to their weights calculated by our new algorithm. Then, the candidates are fed to the WordNet to identify the set of noun words and consolidate the synonymy and hyponymy words. Experimental results have shown that the clustering algorithm is superior to the other subspace clustering algorithms, such as PROCLUS and HARP and kmeans type algorithm, e.g., Bisecting-KMeans. Furthermore, the word extraction method is effective in selection of the words to represent the topics of the clusters.

Algebraic Approach for the Reconstruction of Linear and Convolutional Error Correcting Codes

In this paper we present a generic approach for the problem of the blind estimation of the parameters of linear and convolutional error correcting codes. In a non-cooperative context, an adversary has only access to the noised transmission he has intercepted. The intercepter has no knowledge about the parameters used by the legal users. So, before having acess to the information he has first to blindly estimate the parameters of the error correcting code of the communication. The presented approach has the main advantage that the problem of reconstruction of such codes can be expressed in a very simple way. This allows us to evaluate theorical bounds on the complexity of the reconstruction process but also bounds on the estimation rate. We show that some classical reconstruction techniques are optimal and also explain why some of them have theorical complexities greater than these experimentally observed.

Qualitative Parametric Comparison of Load Balancing Algorithms in Parallel and Distributed Computing Environment
Decrease in hardware costs and advances in computer networking technologies have led to increased interest in the use of large-scale parallel and distributed computing systems. One of the biggest issues in such systems is the development of effective techniques/algorithms for the distribution of the processes/load of a parallel program on multiple hosts to achieve goal(s) such as minimizing execution time, minimizing communication delays, maximizing resource utilization and maximizing throughput. Substantive research using queuing analysis and assuming job arrivals following a Poisson pattern, have shown that in a multi-host system the probability of one of the hosts being idle while other host has multiple jobs queued up can be very high. Such imbalances in system load suggest that performance can be improved by either transferring jobs from the currently heavily loaded hosts to the lightly loaded ones or distributing load evenly/fairly among the hosts .The algorithms known as load balancing algorithms, helps to achieve the above said goal(s). These algorithms come into two basic categories - static and dynamic. Whereas static load balancing algorithms (SLB) take decisions regarding assignment of tasks to processors based on the average estimated values of process execution times and communication delays at compile time, Dynamic load balancing algorithms (DLB) are adaptive to changing situations and take decisions at run time. The objective of this paper work is to identify qualitative parameters for the comparison of above said algorithms. In future this work can be extended to develop an experimental environment to study these Load balancing algorithms based on comparative parameters quantitatively.
An Approach to Adaptive Load Balancing for RFID Middlewares

Recently, there have been an increasing interest in RFID system and RFID systems have been applied to various applications. Load balancing is a fundamental technique for providing scalability of systems by moving workload from overloaded nodes to under-loaded nodes. This paper presents an approach to adaptive load balancing for RFID middlewares. Workloads of RFID middlewares can have a considerable variation according to the location of the connected RFID readers and can abruptly change at a particular instance. The proposed approach considers those characteristics of RFID middle- wares to provide an efficient load balancing.

Design Techniques and Implementation of Low Power High-Throughput Discrete Wavelet Transform Tilters for JPEG 2000 Standard
In this paper, the implementation of low power, high throughput convolutional filters for the one dimensional Discrete Wavelet Transform and its inverse are presented. The analysis filters have already been used for the implementation of a high performance DWT encoder [15] with minimum memory requirements for the JPEG 2000 standard. This paper presents the design techniques and the implementation of the convolutional filters included in the JPEG2000 standard for the forward and inverse DWT for achieving low-power operation, high performance and reduced memory accesses. Moreover, they have the ability of performing progressive computations so as to minimize the buffering between the decomposition and reconstruction phases. The experimental results illustrate the filters- low power high throughput characteristics as well as their memory efficient operation.
A Symbol by Symbol Clustering Based Blind Equalizer
A new blind symbol by symbol equalizer is proposed. The operation of the proposed equalizer is based on the geometric properties of the two dimensional data constellation. An unsupervised clustering technique is used to locate the clusters formed by the received data. The symmetric properties of the clusters labels are subsequently utilized in order to label the clusters. Following this step, the received data are compared to clusters and decisions are made on a symbol by symbol basis, by assigning to each data the label of the nearest cluster. The operation of the equalizer is investigated both in linear and nonlinear channels. The performance of the proposed equalizer is compared to the performance of a CMAbased blind equalizer.
Bridging the Communication Gap at NASA - A Case Study in Communities of Practice
Following the loss of NASA's Space Shuttle Columbia in 2003, it was determined that problems in the agency's organization created an environment that led to the accident. One component of the proposed solution resulted in the formation of the NASA Engineering Network (NEN), a suite of information retrieval and knowledge-sharing tools. This paper describes the implementation of communities of practice, which are formed along engineering disciplines. Communities of practice enable engineers to leverage their knowledge and best practices to collaborate and take information learning back to their jobs and embed it into the procedures of the agency. This case study offers insight into using traditional engineering disciplines for virtual collaboration, including lessons learned during the creation and establishment of NASA-s communities.
Learning and Relationships in the Cyberspace
The cyberspace is an instrument through which internet users could get new experiences. It could contribute to foster one-s own growth, widening cognitive, creative and communicative abilities and promoting relationships. In the cyberspace, in fact, it is possible to create virtual learning communities where internet users improve their interpersonal sphere, knowledge and skills. The main element of e-learning is the establishment of online relationships, that are often collaborative.
Knowledge Creation and Innovation in Classroom
The concepts of knowledge creation and innovation have a strong relationship but this relationship has not been examined systematically. This study examines the utilization of knowledge creation processes of the Theory of Knowledge Creation in Higher Education Institutions. These processes consist of socialization, externalization, combination and internalization. This study suggests that the utilization of these processes will give impacts on innovation in academic performance. A cross-sectional study was conducted using survey questionnaires to collect data of the utilization of knowledge creation processes and classroom-s innovation. The samples are Business Management students of a Malaysian Higher Education Institution. The results of this study could help Higher Education Institutions to enrich the learning process of students through knowledge creation and innovation.
Closing the Achievement Gap Within Reading and Mathematics Classrooms by Fostering Hispanic Students- Educational Resilience
While many studies have conducted the achievement gap between groups of students in school districts, few studies have utilized resilience research to investigate achievement gaps within classrooms. This paper aims to summarize and discuss some recent studies Waxman, Padr├│n, and their colleagues conducted, in which they examined learning environment differences between resilient and nonresilient students in reading and mathematics classrooms. The classes consist of predominantly Hispanic elementary school students from low-income families. These studies all incorporated learning environment questionnaires and systematic observation methods. Significant differences were found between resilient and nonresilient students on their classroom learning environments and classroom behaviors. The observation results indicate that the amount and quality of teacher and student academic interaction are two of the most influential variables that promote student outcomes. This paper concludes by suggesting the following teacher practices to promote resiliency in schools: (a) using feedback from classroom observation and learning environment measures, (b) employing explicit teaching practices; and (c) understanding students on a social and personal level.
Italians- Social and Emotional Loneliness: The Results of Five Studies
Subjective loneliness describes people who feel a disagreeable or unacceptable lack of meaningful social relationships, both at the quantitative and qualitative level. The studies to be presented tested an Italian 18-items self-report loneliness measure, that included items adapted from scales previously developed, namely a short version of the UCLA (Russell, Peplau and Cutrona, 1980), and the 11-items Loneliness scale by De Jong-Gierveld & Kamphuis (JGLS; 1985). The studies aimed at testing the developed scale and at verifying whether loneliness is better conceptualized as a unidimensional (so-called 'general loneliness') or a bidimensional construct, namely comprising the distinct facets of social and emotional loneliness. The loneliness questionnaire included 2 singleitem criterion measures of sad mood, and social contact, and asked participants to supply information on a number of socio-demographic variables. Factorial analyses of responses obtained in two preliminary studies, with 59 and 143 Italian participants respectively, showed good factor loadings and subscale reliability and confirmed that perceived loneliness has clearly two components, a social and an emotional one, the latter measured by two subscales, a 7-item 'general' loneliness subscale derived from UCLA, and a 6–item 'emotional' scale included in the JGLS. Results further showed that type and amount of loneliness are related, negatively, to frequency of social contacts, and, positively, to sad mood. In a third study data were obtained from a nation-wide sample of 9.097 Italian subjects, 12 to about 70 year-olds, who filled the test on-line, on the Italian web site of a large-audience magazine, Focus. The results again confirmed the reliability of the component subscales, namely social, emotional, and 'general' loneliness, and showed that they were highly correlated with each other, especially the latter two. Loneliness scores were significantly predicted by sex, age, education level, sad mood and social contact, and, less so, by other variables – e.g., geographical area and profession. The scale validity was confirmed by the results of a fourth study, with elderly men and women (N 105) living at home or in residential care units. The three subscales were significantly related, among others, to depression, and to various measures of the extension of, and satisfaction with, social contacts with relatives and friends. Finally, a fifth study with 315 career-starters showed that social and emotional loneliness correlate with life satisfaction, and with measures of emotional intelligence. Altogether the results showed a good validity and reliability in the tested samples of the entire scale, and of its components.
Efficiency in Urban Governance towards Sustainability and Competitiveness of City : A Case Study of Kuala Lumpur
Malaysia has successfully applied economic planning to guide the development of the country from an economy of agriculture and mining to a largely industrialised one. Now, with its sights set on attaining the economic level of a fully developed nation by 2020, the planning system must be made even more efficient and focused. It must ensure that every investment made in the country, contribute towards creating the desirable objective of a strong, modern, internationally competitive, technologically advanced, post-industrial economy. Cities in Malaysia must also be fully aware of the enormous competition it faces in a region with rapidly expanding and modernising economies, all contending for the same pool of potential international investments. Efficiency of urban governance is also fundamental issue in development characterized by sustainability, subsidiarity, equity, transparency and accountability, civic engagement and citizenship, and security. As described above, city competitiveness is harnessed through 'city marketing and city management'. High technology and high skilled industries, together with finance, transportation, tourism, business, information and professional services shopping and other commercial activities, are the principal components of the nation-s economy, which must be developed to a level well beyond where it is now. In this respect, Kuala Lumpur being the premier city must play the leading role.
Finding Pareto Optimal Front for the Multi- Mode Time, Cost Quality Trade-off in Project Scheduling
Project managers are the ultimate responsible for the overall characteristics of a project, i.e. they should deliver the project on time with minimum cost and with maximum quality. It is vital for any manager to decide a trade-off between these conflicting objectives and they will be benefited of any scientific decision support tool. Our work will try to determine optimal solutions (rather than a single optimal solution) from which the project manager will select his desirable choice to run the project. In this paper, the problem in project scheduling notated as (1,T|cpm,disc,mu|curve:quality,time,cost) will be studied. The problem is multi-objective and the purpose is finding the Pareto optimal front of time, cost and quality of a project (curve:quality,time,cost), whose activities belong to a start to finish activity relationship network (cpm) and they can be done in different possible modes (mu) which are non-continuous or discrete (disc), and each mode has a different cost, time and quality . The project is constrained to a non-renewable resource i.e. money (1,T). Because the problem is NP-Hard, to solve the problem, a meta-heuristic is developed based on a version of genetic algorithm specially adapted to solve multi-objective problems namely FastPGA. A sample project with 30 activities is generated and then solved by the proposed method.
Effect of Particle Gravity on the Fractal Dimension of Particle Line in three-dimensional Turbulent Flows using Kinematic Simulation
In this study, the dispersion of heavy particles line in an isotropic and incompressible three-dimensional turbulent flow has been studied using the Kinematic Simulation techniques to find out the evolution of the line fractal dimension. The fractal dimension of the line is found in the case of different particle gravity (in practice, different values of particle drift velocity) in the presence of small particle inertia with a comparison with that obtained in the diffusion case of material line at the same Reynolds number. It can be concluded for the dispersion of heavy particles line in turbulent flow that the particle gravity affect the fractal dimension of the line for different particle gravity velocities in the range 0.2 < W < 2. With the increase of the particle drift velocity, the fractal dimension of the line decreases which may be explained as the particles pass many scales in their journey in the direction of the gravity and the particles trajectories do not affect by these scales at high particle drift velocities.
Effect of Inertia on the Fractal Dimension of Particle Line in three-dimensional Turbulent Flows using Kinematic Simulation
The dispersion of heavy particles line in an isotropic and incompressible three-dimensional turbulent flow has been studied using the Kinematic Simulation techniques to find out the evolution of the line fractal dimension. In this study, the fractal dimension of the line is found for different cases of heavy particles inertia (different Stokes numbers) in the absence of the particle gravity with a comparison with the fractal dimension obtained in the diffusion case of material line at the same Reynolds number. It can be concluded for the dispersion of heavy particles line in turbulent flow that the particle inertia affect the fractal dimension of a line released in a turbulent flow for Stokes numbers 0.02 < St < 2. At the beginning for small times, most of the different cases are not affected by the inertia until a certain time, the particle response time τa, with larger time as the particles inertia increases, the fractal dimension of the line increases owing to the particles becoming more sensitive to the small scales which cause the change in the line shape during its journey.
Effect of a Linear-Exponential Penalty Functionon the GA-s Efficiency in Optimization of a Laminated Composite Panel
A stiffened laminated composite panel (1 m length × 0.5m width) was optimized for minimum weight and deflection under several constraints using genetic algorithm. Here, a significant study on the performance of a penalty function with two kinds of static and dynamic penalty factors was conducted. The results have shown that linear dynamic penalty factors are more effective than the static ones. Also, a specially combined linear-exponential function has shown to perform more effective than the previously mentioned penalty functions. This was then resulted in the less sensitivity of the GA to the amount of penalty factor.
An Agent Oriented Approach to Operational Profile Management
Software reliability, defined as the probability of a software system or application functioning without failure or errors over a defined period of time, has been an important area of research for over three decades. Several research efforts aimed at developing models to improve reliability are currently underway. One of the most popular approaches to software reliability adopted by some of these research efforts involves the use of operational profiles to predict how software applications will be used. Operational profiles are a quantification of usage patterns for a software application. The research presented in this paper investigates an innovative multiagent framework for automatic creation and management of operational profiles for generic distributed systems after their release into the market. The architecture of the proposed Operational Profile MAS (Multi-Agent System) is presented along with detailed descriptions of the various models arrived at following the analysis and design phases of the proposed system. The operational profile in this paper is extended to comprise seven different profiles. Further, the criticality of operations is defined using a new composed metrics in order to organize the testing process as well as to decrease the time and cost involved in this process. A prototype implementation of the proposed MAS is included as proof-of-concept and the framework is considered as a step towards making distributed systems intelligent and self-managing.
Fast Painting with Different Colors Using Cross Correlation in the Frequency Domain
In this paper, a new technique for fast painting with different colors is presented. The idea of painting relies on applying masks with different colors to the background. Fast painting is achieved by applying these masks in the frequency domain instead of spatial (time) domain. New colors can be generated automatically as a result from the cross correlation operation. This idea was applied successfully for faster specific data (face, object, pattern, and code) detection using neural algorithms. Here, instead of performing cross correlation between the input input data (e.g., image, or a stream of sequential data) and the weights of neural networks, the cross correlation is performed between the colored masks and the background. Furthermore, this approach is developed to reduce the computation steps required by the painting operation. The principle of divide and conquer strategy is applied through background decomposition. Each background is divided into small in size subbackgrounds and then each sub-background is processed separately by using a single faster painting algorithm. Moreover, the fastest painting is achieved by using parallel processing techniques to paint the resulting sub-backgrounds using the same number of faster painting algorithms. In contrast to using only faster painting algorithm, the speed up ratio is increased with the size of the background when using faster painting algorithm and background decomposition. Simulation results show that painting in the frequency domain is faster than that in the spatial domain.
Enhancement of Stereo Video Pairs Using SDNs To Aid In 3D Reconstruction

This paper presents the results of enhancing images from a left and right stereo pair in order to increase the resolution of a 3D representation of a scene generated from that same pair. A new neural network structure known as a Self Delaying Dynamic Network (SDN) has been used to perform the enhancement. The advantage of SDNs over existing techniques such as bicubic interpolation is their ability to cope with motion and noise effects. SDNs are used to generate two high resolution images, one based on frames taken from the left view of the subject, and one based on the frames from the right. This new high resolution stereo pair is then processed by a disparity map generator. The disparity map generated is compared to two other disparity maps generated from the same scene. The first is a map generated from an original high resolution stereo pair and the second is a map generated using a stereo pair which has been enhanced using bicubic interpolation. The maps generated using the SDN enhanced pairs match more closely the target maps. The addition of extra noise into the input images is less problematic for the SDN system which is still able to out perform bicubic interpolation.

An Efficient Technique for Extracting Fuzzy Rulesfrom Neural Networks

Artificial neural networks (ANN) have the ability to model input-output relationships from processing raw data. This characteristic makes them invaluable in industry domains where such knowledge is scarce at best. In the recent decades, in order to overcome the black-box characteristic of ANNs, researchers have attempted to extract the knowledge embedded within ANNs in the form of rules that can be used in inference systems. This paper presents a new technique that is able to extract a small set of rules from a two-layer ANN. The extracted rules yield high classification accuracy when implemented within a fuzzy inference system. The technique targets industry domains that possess less complex problems for which no expert knowledge exists and for which a simpler solution is preferred to a complex one. The proposed technique is more efficient, simple, and applicable than most of the previously proposed techniques.

Evaluating some Feature Selection Methods for an Improved SVM Classifier

Text categorization is the problem of classifying text documents into a set of predefined classes. After a preprocessing step the documents are typically represented as large sparse vectors. When training classifiers on large collections of documents, both the time and memory restrictions can be quite prohibitive. This justifies the application of features selection methods to reduce the dimensionality of the document-representation vector. Four feature selection methods are evaluated: Random Selection, Information Gain (IG), Support Vector Machine (called SVM_FS) and Genetic Algorithm with SVM (GA_FS). We showed that the best results were obtained with SVM_FS and GA_FS methods for a relatively small dimension of the features vector comparative with the IG method that involves longer vectors, for quite similar classification accuracies. Also we present a novel method to better correlate SVM kernel-s parameters (Polynomial or Gaussian kernel).

Kalman-s Shrinkage for Wavelet-Based Despeckling of SAR Images
In this paper, a new probability density function (pdf) is proposed to model the statistics of wavelet coefficients, and a simple Kalman-s filter is derived from the new pdf using Bayesian estimation theory. Specifically, we decompose the speckled image into wavelet subbands, we apply the Kalman-s filter to the high subbands, and reconstruct a despeckled image from the modified detail coefficients. Experimental results demonstrate that our method compares favorably to several other despeckling methods on test synthetic aperture radar (SAR) images.
Efficient System for Speech Recognition using General Regression Neural Network
In this paper we present an efficient system for independent speaker speech recognition based on neural network approach. The proposed architecture comprises two phases: a preprocessing phase which consists in segmental normalization and features extraction and a classification phase which uses neural networks based on nonparametric density estimation namely the general regression neural network (GRNN). The relative performances of the proposed model are compared to the similar recognition systems based on the Multilayer Perceptron (MLP), the Recurrent Neural Network (RNN) and the well known Discrete Hidden Markov Model (HMM-VQ) that we have achieved also. Experimental results obtained with Arabic digits have shown that the use of nonparametric density estimation with an appropriate smoothing factor (spread) improves the generalization power of the neural network. The word error rate (WER) is reduced significantly over the baseline HMM method. GRNN computation is a successful alternative to the other neural network and DHMM.
Near-Field Robust Adaptive Beamforming Based on Worst-Case Performance Optimization
The performance of adaptive beamforming degrades substantially in the presence of steering vector mismatches. This degradation is especially severe in the near-field, for the 3-dimensional source location is more difficult to estimate than the 2-dimensional direction of arrival in far-field cases. As a solution, a novel approach of near-field robust adaptive beamforming (RABF) is proposed in this paper. It is a natural extension of the traditional far-field RABF and belongs to the class of diagonal loading approaches, with the loading level determined based on worst-case performance optimization. However, different from the methods solving the optimal loading by iteration, it suggests here a simple closed-form solution after some approximations, and consequently, the optimal weight vector can be expressed in a closed form. Besides simplicity and low computational cost, the proposed approach reveals how different factors affect the optimal loading as well as the weight vector. Its excellent performance in the near-field is confirmed via a number of numerical examples.
Fast Wavelet Image Denoising Based on Local Variance and Edge Analysis
The approach based on the wavelet transform has been widely used for image denoising due to its multi-resolution nature, its ability to produce high levels of noise reduction and the low level of distortion introduced. However, by removing noise, high frequency components belonging to edges are also removed, which leads to blurring the signal features. This paper proposes a new method of image noise reduction based on local variance and edge analysis. The analysis is performed by dividing an image into 32 x 32 pixel blocks, and transforming the data into wavelet domain. Fast lifting wavelet spatial-frequency decomposition and reconstruction is developed with the advantages of being computationally efficient and boundary effects minimized. The adaptive thresholding by local variance estimation and edge strength measurement can effectively reduce image noise while preserve the features of the original image corresponding to the boundaries of the objects. Experimental results demonstrate that the method performs well for images contaminated by natural and artificial noise, and is suitable to be adapted for different class of images and type of noises. The proposed algorithm provides a potential solution with parallel computation for real time or embedded system application.
Representation of Coloured Petri Net in Abductive Logic Programming (CPN-LP) and Its Application in Modeling an Intelligent Agent

Coloured Petri net (CPN) has been widely adopted in various areas in Computer Science, including protocol specification, performance evaluation, distributed systems and coordination in multi-agent systems. It provides a graphical representation of a system and has a strong mathematical foundation for proving various properties. This paper proposes a novel representation of a coloured Petri net using an extension of logic programming called abductive logic programming (ALP), which is purely based on classical logic. Under such a representation, an implementation of a CPN could be directly obtained, in which every inference step could be treated as a kind of equivalence preserved transformation. We would describe how to implement a CPN under such a representation using common meta-programming techniques in Prolog. We call our framework CPN-LP and illustrate its applications in modeling an intelligent agent.

A New Implementation of PCA for Fast Face Detection
Principal Component Analysis (PCA) has many different important applications especially in pattern detection such as face detection / recognition. Therefore, for real time applications, the response time is required to be as small as possible. In this paper, new implementation of PCA for fast face detection is presented. Such new implementation is designed based on cross correlation in the frequency domain between the input image and eigenvectors (weights). Simulation results show that the proposed implementation of PCA is faster than conventional one.
Recognition Machine (RM) for On-line and Isolated Flight Deck Officer (FDO) Gestures
The paper presents an on-line recognition machine (RM) for continuous/isolated, dynamic and static gestures that arise in Flight Deck Officer (FDO) training. RM is based on generic pattern recognition framework. Gestures are represented as templates using summary statistics. The proposed recognition algorithm exploits temporal and spatial characteristics of gestures via dynamic programming and Markovian process. The algorithm predicts corresponding index of incremental input data in the templates in an on-line mode. Accumulated consistency in the sequence of prediction provides a similarity measurement (Score) between input data and the templates. The algorithm provides an intuitive mechanism for automatic detection of start/end frames of continuous gestures. In the present paper, we consider isolated gestures. The performance of RM is evaluated using four datasets - artificial (W TTest), hand motion (Yang) and FDO (tracker, vision-based ). RM achieves comparable results which are in agreement with other on-line and off-line algorithms such as hidden Markov model (HMM) and dynamic time warping (DTW). The proposed algorithm has the additional advantage of providing timely feedback for training purposes.
A Modular On-line Profit Sharing Approach in Multiagent Domains
How to coordinate the behaviors of the agents through learning is a challenging problem within multi-agent domains. Because of its complexity, recent work has focused on how coordinated strategies can be learned. Here we are interested in using reinforcement learning techniques to learn the coordinated actions of a group of agents, without requiring explicit communication among them. However, traditional reinforcement learning methods are based on the assumption that the environment can be modeled as Markov Decision Process, which usually cannot be satisfied when multiple agents coexist in the same environment. Moreover, to effectively coordinate each agent-s behavior so as to achieve the goal, it-s necessary to augment the state of each agent with the information about other existing agents. Whereas, as the number of agents in a multiagent environment increases, the state space of each agent grows exponentially, which will cause the combinational explosion problem. Profit sharing is one of the reinforcement learning methods that allow agents to learn effective behaviors from their experiences even within non-Markovian environments. In this paper, to remedy the drawback of the original profit sharing approach that needs much memory to store each state-action pair during the learning process, we firstly address a kind of on-line rational profit sharing algorithm. Then, we integrate the advantages of modular learning architecture with on-line rational profit sharing algorithm, and propose a new modular reinforcement learning model. The effectiveness of the technique is demonstrated using the pursuit problem.
Controller Design for Euler-Bernoulli Smart Structures Using Robust Decentralized FOS via Reduced Order Modeling
This paper features the modeling and design of a Robust Decentralized Fast Output Sampling (RDFOS) Feedback control technique for the active vibration control of a smart flexible multimodel Euler-Bernoulli cantilever beams for a multivariable (MIMO) case by retaining the first 6 vibratory modes. The beam structure is modeled in state space form using the concept of piezoelectric theory, the Euler-Bernoulli beam theory and the Finite Element Method (FEM) technique by dividing the beam into 4 finite elements and placing the piezoelectric sensor / actuator at two finite element locations (positions 2 and 4) as collocated pairs, i.e., as surface mounted sensor / actuator, thus giving rise to a multivariable model of the smart structure plant with two inputs and two outputs. Five such multivariable models are obtained by varying the dimensions (aspect ratios) of the aluminium beam. Using model order reduction technique, the reduced order model of the higher order system is obtained based on dominant Eigen value retention and the Davison technique. RDFOS feedback controllers are designed for the above 5 multivariable-multimodel plant. The closed loop responses with the RDFOS feedback gain and the magnitudes of the control input are obtained and the performance of the proposed multimodel smart structure system is evaluated for vibration control.
Churn Prediction: Does Technology Matter?
The aim of this paper is to identify the most suitable model for churn prediction based on three different techniques. The paper identifies the variables that affect churn in reverence of customer complaints data and provides a comparative analysis of neural networks, regression trees and regression in their capabilities of predicting customer churn.
Unified Fusion Approach with Application to SLAM
In this paper, we propose the pre-processor based on the Evidence Supporting Measure of Similarity (ESMS) filter and also propose the unified fusion approach (UFA) based on the general fusion machine coupled with ESMS filter, which improve the correctness and precision of information fusion in any fields of application. Here we mainly apply the new approach to Simultaneous Localization And Mapping (SLAM) of Pioneer II mobile robots. A simulation experiment was performed, where an autonomous virtual mobile robot with sonar sensors evolves in a virtual world map with obstacles. By comparing the result of building map according to the general fusion machine (here DSmT-based fusing machine and PCR5-based conflict redistributor considereded) coupling with ESMS filter and without ESMS filter, it shows the benefit of the selection of the sources as a prerequisite for improvement of the information fusion, and also testifies the superiority of the UFA in dealing with SLAM.
Shift Invariant Support Vector Machines Face Recognition System
In this paper, we present a new method for incorporating global shift invariance in support vector machines. Unlike other approaches which incorporate a feature extraction stage, we first scale the image and then classify it by using the modified support vector machines classifier. Shift invariance is achieved by replacing dot products between patterns used by the SVM classifier with the maximum cross-correlation value between them. Unlike the normal approach, in which the patterns are treated as vectors, in our approach the patterns are treated as matrices (or images). Crosscorrelation is computed by using computationally efficient techniques such as the fast Fourier transform. The method has been tested on the ORL face database. The tests indicate that this method can improve the recognition rate of an SVM classifier.
Layered Multiple Description Coding For Robust Video Transmission Over Wireless Ad-Hoc Networks
This paper presents a video transmission system using layered multiple description (coding (MDC) and multi-path transport for reliable video communications in wireless ad-hoc networks. The proposed MDC extends a quality-scalable H.264/AVC video coding algorithm to generate two independent descriptions. The two descriptions are transmitted over different paths to a receiver in order to alleviate the effect of unstable channel conditions of wireless adhoc networks. If one description is lost due to transmission erros, then the correctly received description is used to estimate the lost information of the corrupted description. The proposed MD coder maintains an adequate video quality as long as both description are not simultaneously lost. Simulation results show that the proposed MD coding combined with multi-path transport system is largely immune to packet losses, and therefore, can be a promising solution for robust video communications over wireless ad-hoc networks.
Many-Sided Self Risk Analysis Model for Information Asset to Secure Stability of the Information and Communication Service
Information and communication service providers (ICSP) that are significant in size and provide Internet-based services take administrative, technical, and physical protection measures via the information security check service (ISCS). These protection measures are the minimum action necessary to secure the stability and continuity of the information and communication services (ICS) that they provide. Thus, information assets are essential to providing ICS, and deciding the relative importance of target assets for protection is a critical procedure. The risk analysis model designed to decide the relative importance of information assets, which is described in this study, evaluates information assets from many angles, in order to choose which ones should be given priority when it comes to protection. Many-sided risk analysis (MSRS) grades the importance of information assets, based on evaluation of major security check items, evaluation of the dependency on the information and communication facility (ICF) and influence on potential incidents, and evaluation of major items according to their service classification, in order to identify the ISCS target. MSRS could be an efficient risk analysis model to help ICSPs to identify their core information assets and take information protection measures first, so that stability of the ICS can be ensured.
MJPEG Real-Time Transmission in Industrial Environments Using a CBR Channel
Currently, there are many local area industrial networks that can give guaranteed bandwidth to synchronous traffic, particularly providing CBR channels (Constant Bit Rate), which allow improved bandwidth management. Some of such networks operate over Ethernet, delivering channels with enough capacity, specially with compressors, to integrate multimedia traffic in industrial monitoring and image processing applications with many sources. In these industrial environments where a low latency is an essential requirement, JPEG is an adequate compressing technique but it generates VBR traffic (Variable Bit Rate). Transmitting VBR traffic in CBR channels is inefficient and current solutions to this problem significantly increase the latency or further degrade the quality. In this paper an R(q) model is used which allows on-line calculation of the JPEG quantification factor. We obtained increased quality, a lower requirement for the CBR channel with reduced number of discarded frames along with better use of the channel bandwidth.
Real-Time Specific Weed Recognition System Using Histogram Analysis
Information on weed distribution within the field is necessary to implement spatially variable herbicide application. Since hand labor is costly, an automated weed control system could be feasible. This paper deals with the development of an algorithm for real time specific weed recognition system based on Histogram Analysis of an image that is used for the weed classification. This algorithm is specifically developed to classify images into broad and narrow class for real-time selective herbicide application. The developed system has been tested on weeds in the lab, which have shown that the system to be very effectiveness in weed identification. Further the results show a very reliable performance on images of weeds taken under varying field conditions. The analysis of the results shows over 95 percent classification accuracy over 140 sample images (broad and narrow) with 70 samples from each category of weeds.
Mining Genes Relations in Microarray Data Combined with Ontology in Colon Cancer Automated Diagnosis System
MATCH project [1] entitle the development of an automatic diagnosis system that aims to support treatment of colon cancer diseases by discovering mutations that occurs to tumour suppressor genes (TSGs) and contributes to the development of cancerous tumours. The constitution of the system is based on a) colon cancer clinical data and b) biological information that will be derived by data mining techniques from genomic and proteomic sources The core mining module will consist of the popular, well tested hybrid feature extraction methods, and new combined algorithms, designed especially for the project. Elements of rough sets, evolutionary computing, cluster analysis, self-organization maps and association rules will be used to discover the annotations between genes, and their influence on tumours [2]-[11]. The methods used to process the data have to address their high complexity, potential inconsistency and problems of dealing with the missing values. They must integrate all the useful information necessary to solve the expert's question. For this purpose, the system has to learn from data, or be able to interactively specify by a domain specialist, the part of the knowledge structure it needs to answer a given query. The program should also take into account the importance/rank of the particular parts of data it analyses, and adjusts the used algorithms accordingly.
Hardware Implementation of Stack-Based Replacement Algorithms
Block replacement algorithms to increase hit ratio have been extensively used in cache memory management. Among basic replacement schemes, LRU and FIFO have been shown to be effective replacement algorithms in terms of hit rates. In this paper, we introduce a flexible stack-based circuit which can be employed in hardware implementation of both LRU and FIFO policies. We propose a simple and efficient architecture such that stack-based replacement algorithms can be implemented without the drawbacks of the traditional architectures. The stack is modular and hence, a set of stack rows can be cascaded depending on the number of blocks in each cache set. Our circuit can be implemented in conjunction with the cache controller and static/dynamic memories to form a cache system. Experimental results exhibit that our proposed circuit provides an average value of 26% improvement in storage bits and its maximum operating frequency is increased by a factor of two
Harnessing Replication in Object Allocation
The design of distributed systems involves the partitioning of the system into components or partitions and the allocation of these components to physical nodes. Techniques have been proposed for both the partitioning and allocation process. However these techniques suffer from a number of limitations. For instance object replication has the potential to greatly improve the performance of an object orientated distributed system but can be difficult to use effectively and there are few techniques that support the developer in harnessing object replication. This paper presents a methodological technique that helps developers decide how objects should be allocated in order to improve performance in a distributed system that supports replication. The performance of the proposed technique is demonstrated and tested on an example system.
Web Usability : A Fuzzy Approach to the Navigation Structure Enhancement in a Website System, Case of Iranian Civil Aviation Organization Website
With the proliferation of World Wide Web, development of web-based technologies and the growth in web content, the structure of a website becomes more complex and web navigation becomes a critical issue to both web designers and users. In this paper we define the content and web pages as two important and influential factors in website navigation and paraphrase the enhancement in the website navigation as making some useful changes in the link structure of the website based on the aforementioned factors. Then we suggest a new method for proposing the changes using fuzzy approach to optimize the website architecture. Applying the proposed method to a real case of Iranian Civil Aviation Organization (CAO) website, we discuss the results of the novel approach at the final section.
A Consistency Protocol Multi-Layer for Replicas Management in Large Scale Systems
Large scale systems such as computational Grid is a distributed computing infrastructure that can provide globally available network resources. The evolution of information processing systems in Data Grid is characterized by a strong decentralization of data in several fields whose objective is to ensure the availability and the reliability of the data in the reason to provide a fault tolerance and scalability, which cannot be possible only with the use of the techniques of replication. Unfortunately the use of these techniques has a height cost, because it is necessary to maintain consistency between the distributed data. Nevertheless, to agree to live with certain imperfections can improve the performance of the system by improving competition. In this paper, we propose a multi-layer protocol combining the pessimistic and optimistic approaches conceived for the data consistency maintenance in large scale systems. Our approach is based on a hierarchical representation model with tree layers, whose objective is with double vocation, because it initially makes it possible to reduce response times compared to completely pessimistic approach and it the second time to improve the quality of service compared to an optimistic approach.
A General Model for Acquiring Knowledge
In this paper, based on the work in [1], we further give a general model for acquiring knowledge, which first focuses on the research of how and when things involved in problems are made then describes the goals, the energy and the time to give an optimum model to decide how many related things are supposed to be involved in. Finally, we acquire knowledge from this model in which there are the attributes, actions and connections of the things involved at the time when they are born and the time in their life. This model not only improves AI theories, but also surely brings the effectiveness and accuracy for AI system because systems are given more knowledge when reasoning or computing is used to bring about results.
An Efficient Algorithm for Computing all Program Forward Static Slices
Program slicing is the task of finding all statements in a program that directly or indirectly influence the value of a variable occurrence. The set of statements that can affect the value of a variable at some point in a program is called a program backward slice. In several software engineering applications, such as program debugging and measuring program cohesion and parallelism, several slices are computed at different program points. The existing algorithms for computing program slices are introduced to compute a slice at a program point. In these algorithms, the program, or the model that represents the program, is traversed completely or partially once. To compute more than one slice, the same algorithm is applied for every point of interest in the program. Thus, the same program, or program representation, is traversed several times. In this paper, an algorithm is introduced to compute all forward static slices of a computer program by traversing the program representation graph once. Therefore, the introduced algorithm is useful for software engineering applications that require computing program slices at different points of a program. The program representation graph used in this paper is called Program Dependence Graph (PDG).
The Coverage of the Object-Oriented Framework Application Class-Based Test Cases
An application framework provides a reusable design and implementation for a family of software systems. Frameworks are introduced to reduce the cost of a product line (i.e., family of products that share the common features). Software testing is a time consuming and costly ongoing activity during the application software development process. Generating reusable test cases for the framework applications at the framework development stage, and providing and using the test cases to test part of the framework application whenever the framework is used reduces the application development time and cost considerably. Framework Interface Classes (FICs) are classes introduced by the framework hooks to be implemented at the application development stage. They can have reusable test cases generated at the framework development stage and provided with the framework to test the implementations of the FICs at the application development stage. In this paper, we conduct a case study using thirteen applications developed using three frameworks; one domain oriented and two application oriented. The results show that, in general, the percentage of the number of FICs in the applications developed using domain frameworks is, on average, greater than the percentage of the number of FICs in the applications developed using application frameworks. Consequently, the reduction of the application unit testing time using the reusable test cases generated for domain frameworks is, in general, greater than the reduction of the application unit testing time using the reusable test cases generated for application frameworks.
Generating State-Based Testing Models for Object-Oriented Framework Interface Classes
An application framework provides a reusable design and implementation for a family of software systems. Application developers extend the framework to build their particular applications using hooks. Hooks are the places identified to show how to use and customize the framework. Hooks define the Framework Interface Classes (FICs) and the specifications of their methods. As part of the development life cycle, it is required to test the implementations of the FICs. Building a testing model to express the behavior of a class is an essential step for the generation of the class-based test cases. The testing model has to be consistent with the specifications provided for the hooks. State-based models consisting of states and transitions are testing models well suited to objectoriented software. Typically, hand-construction of a state-based model of a class behavior is expensive, error-prone, and may result in constructing an inconsistent model with the specifications of the class methods, which misleads verification results. In this paper, a technique is introduced to automatically synthesize a state-based testing model for FICs using the specifications provided for the hooks. A tool that supports the proposed technique is introduced.
Generating Class-Based Test Cases for Interface Classes of Object-Oriented Black Box Frameworks
An application framework provides a reusable design and implementation for a family of software systems. Application developers extend the framework to build their particular applications using hooks. Hooks are the places identified to show how to use and customize the framework. Hooks define the Framework Interface Classes (FICs) and their possible specifications, which helps in building reusable test cases for the implementations of these classes. This paper introduces a novel technique called all paths-state to generate state-based test cases to test the FICs at class level. The technique is experimentally evaluated. The empirical evaluation shows that all paths-state technique produces test cases with a high degree of coverage for the specifications of the implemented FICs comparing to test cases generated using round-trip path and all-transition techniques.
HIV Modelling - Parallel Implementation Strategies
We report on the development of a model to understand why the range of experience with respect to HIV infection is so diverse, especially with respect to the latency period. To investigate this, an agent-based approach is used to extract highlevel behaviour which cannot be described analytically from the set of interaction rules at the cellular level. A network of independent matrices mimics the chain of lymph nodes. Dealing with massively multi-agent systems requires major computational effort. However, parallelisation methods are a natural consequence and advantage of the multi-agent approach and, using the MPI library, are here implemented, tested and optimized. Our current focus is on the various implementations of the data transfer across the network. Three communications strategies are proposed and tested, showing that the most efficient approach is communication based on the natural lymph-network connectivity.
Memory Leak Detection in Distributed System
Due to memory leaks, often-valuable system memory gets wasted and denied for other processes thereby affecting the computational performance. If an application-s memory usage exceeds virtual memory size, it can leads to system crash. Current memory leak detection techniques for clusters are reactive and display the memory leak information after the execution of the process (they detect memory leak only after it occur). This paper presents a Dynamic Memory Monitoring Agent (DMMA) technique. DMMA framework is a dynamic memory leak detection, that detects the memory leak while application is in execution phase, when memory leak in any process in the cluster is identified by DMMA it gives information to the end users to enable them to take corrective actions and also DMMA submit the affected process to healthy node in the system. Thus provides reliable service to the user. DMMA maintains information about memory consumption of executing processes and based on this information and critical states, DMMA can improve reliability and efficaciousness of cluster computing.
New Features for Specific JPEG Steganalysis

We present in this paper a new approach for specific JPEG steganalysis and propose studying statistics of the compressed DCT coefficients. Traditionally, steganographic algorithms try to preserve statistics of the DCT and of the spatial domain, but they cannot preserve both and also control the alteration of the compressed data. We have noticed a deviation of the entropy of the compressed data after a first embedding. This deviation is greater when the image is a cover medium than when the image is a stego image. To observe this deviation, we pointed out new statistic features and combined them with the Multiple Embedding Method. This approach is motivated by the Avalanche Criterion of the JPEG lossless compression step. This criterion makes possible the design of detectors whose detection rates are independent of the payload. Finally, we designed a Fisher discriminant based classifier for well known steganographic algorithms, Outguess, F5 and Hide and Seek. The experiemental results we obtained show the efficiency of our classifier for these algorithms. Moreover, it is also designed to work with low embedding rates (< 10-5) and according to the avalanche criterion of RLE and Huffman compression step, its efficiency is independent of the quantity of hidden information.

Reduction of Overheads with Dynamic Caching in Fixed AODV based MANETs

In this paper we show that adjusting ART in accordance with static network scenario can substantially improve the performance of AODV by reducing control overheads. We explain the relationship of control overheads with network size and request patterns of the users. Through simulation we show that making ART proportionate to network static time reduces the amount of control overheads independent of network size and user request patterns.

Bio-Inspired Generalized Global Shape Approach for Writer Identification
Writer identification is one of the areas in pattern recognition that attract many researchers to work in, particularly in forensic and biometric application, where the writing style can be used as biometric features for authenticating an identity. The challenging task in writer identification is the extraction of unique features, in which the individualistic of such handwriting styles can be adopted into bio-inspired generalized global shape for writer identification. In this paper, the feasibility of generalized global shape concept of complimentary binding in Artificial Immune System (AIS) for writer identification is explored. An experiment based on the proposed framework has been conducted to proof the validity and feasibility of the proposed approach for off-line writer identification.
Content and Resources based Mobile and Wireless Video Transcoding
Delivering streaming video over wireless is an important component of many interactive multimedia applications running on personal wireless handset devices. Such personal devices have to be inexpensive, compact, and lightweight. But wireless channels have a high channel bit error rate and limited bandwidth. Delay variation of packets due to network congestion and the high bit error rate greatly degrades the quality of video at the handheld device. Therefore, mobile access to multimedia contents requires video transcoding functionality at the edge of the mobile network for interworking with heterogeneous networks and services. Therefore, to guarantee quality of service (QoS) delivered to the mobile user, a robust and efficient transcoding scheme should be deployed in mobile multimedia transporting network. Hence, this paper examines the challenges and limitations that the video transcoding schemes in mobile multimedia transporting network face. Then handheld resources, network conditions and content based mobile and wireless video transcoding is proposed to provide high QoS applications. Exceptional performance is demonstrated in the experiment results. These experiments were designed to verify and prove the robustness of the proposed approach. Extensive experiments have been conducted, and the results of various video clips with different bit rate and frame rate have been provided.
Image Mapping with Cumulative Distribution Function for Quick Convergence of Counter Propagation Neural Networks in Image Compression
In general the images used for compression are of different types like dark image, high intensity image etc. When these images are compressed using Counter Propagation Neural Network, it takes longer time to converge. The reason for this is that the given image may contain a number of distinct gray levels with narrow difference with their neighborhood pixels. If the gray levels of the pixels in an image and their neighbors are mapped in such a way that the difference in the gray levels of the neighbor with the pixel is minimum, then compression ratio as well as the convergence of the network can be improved. To achieve this, a Cumulative Distribution Function is estimated for the image and it is used to map the image pixels. When the mapped image pixels are used the Counter Propagation Neural Network yield high compression ratio as well as it converges quickly.
Comparison of Reliability Systems Based Uncertainty
Stochastic comparison has been an important direction of research in various area. This can be done by the use of the notion of stochastic ordering which gives qualitatitive rather than purely quantitative estimation of the system under study. In this paper we present applications of comparison based uncertainty related to entropy in Reliability analysis, for example to design better systems. These results can be used as a priori information in simulation studies.
Addressing Security Concerns of Data Exchange in AODV Protocol

The Ad Hoc on demand distance vector (AODV) routing protocol is designed for mobile ad hoc networks (MANETs). AODV offers quick adaptation to dynamic link conditions; it is characterized by low memory overhead and low network utilization. The security issues related to the protocol remain challenging for the wireless network designers. Numerous schemes have been proposed for establishing secure communication between end users, these schemes identify that the secure operation of AODV is a bi tier task (routing and secure exchange of information at separate levels). Our endeavor in this paper would focus on achieving the routing and secure data exchange in a single step. This will facilitate the user nodes to perform routing, mutual authentications, generation and secure exchange of session key in one step thus ensuring confidentiality, integrity and authentication of data exchange in a more suitable way.

Performance Comparison and Analysis of Serial Concatenated Convolutional Codes
In this paper, the performance of three types of serial concatenated convolutional codes (SCCC) is compared and analyzed in additive white Gaussian noise (AWGN) channel. In Type I, only the parity bits of outer encoder are passed to inner encoder. In Type II and Type III, both the information bits and the parity bits of outer encoder are transferred to inner encoder. As results of simulation, Type I shows the best bit error rate (BER) performance at low signal-to-noise ratio (SNR). On the other hand, Type III shows the best BER performance at high SNR in AWGN channel. The simulation results are analyzed using the distance spectrum.
Architecture, Implementation and Application of Tools for Experimental Analysis
This paper presents an architecture to assist in the development of tools to perform experimental analysis. Existing implementations of tools based on this architecture are also described in this paper. These tools are applied to the real world problem of fault attack emulation and detection in cryptographic algorithms.
Implementing a Prototype System for Power Facility Management using RFID/WSN
Firstly, research and development on RFID focuses on manufacturing and retail sectors, because it can improve supply chain efficiency. But, now a variety of field is considered the next research area for Radio Frequency Identification (RFID). Although RFID is infancy, RFID technology has great potential in power industry to significantly reduce cost, and improve quality of power supply. To complement the limitation of RFID, we adopt the WSN (Wireless Sensor Network) technology. However, relevant experience is limited, the challenge will be to derive requirement from business practice and to determine whether it is possible or not. To explore this issue, we conduct a case study on implementing power facility management system using RFID/WSN in Korea Electric Power Corporation (KEPCO). In this paper we describe requirement from power industry. And we introduce design and implementation of the test bed.
Self-tuned LMS Algorithm for Sinusoidal Time Delay Tracking
In this paper the problem of estimating the time delay between two spatially separated noisy sinusoidal signals by system identification modeling is addressed. The system is assumed to be perturbed by both input and output additive white Gaussian noise. The presence of input noise introduces bias in the time delay estimates. Normally the solution requires a priori knowledge of the input-output noise variance ratio. We utilize the cascade of a self-tuned filter with the time delay estimator, thus making the delay estimates robust to input noise. Simulation results are presented to confirm the superiority of the proposed approach at low input signal-to-noise ratios.
Different Approaches for the Design of IFIR Compaction Filter

Optimization of filter banks based on the knowledge of input statistics has been of interest for a long time. Finite impulse response (FIR) Compaction filters are used in the design of optimal signal adapted orthonormal FIR filter banks. In this paper we discuss three different approaches for the design of interpolated finite impulse response (IFIR) compaction filters. In the first method, the magnitude squared response satisfies Nyquist constraint approximately. In the second and third methods Nyquist constraint is exactly satisfied. These methods yield FIR compaction filters whose response is comparable with that of the existing methods. At the same time, IFIR filters enjoy significant saving in the number of multipliers and can be implemented efficiently. Since eigenfilter approach is used here, the method is less complex. Design of IFIR filters in the least square sense is presented.

NonStationary CMA for Decision Feedback Equalization of Markovian Time Varying Channels
In this paper, we propose a modified version of the Constant Modulus Algorithm (CMA) tailored for blind Decision Feedback Equalizer (DFE) of first order Markovian time varying channels. The proposed NonStationary CMA (NSCMA) is designed so that it explicitly takes into account the Markovian structure of the channel nonstationarity. Hence, unlike the classical CMA, the NSCMA is not blind with respect to the channel time variations. This greatly helps the equalizer in the case of realistic channels, and avoids frequent transmissions of training sequences. This paper develops a theoretical analysis of the steady state performance of the CMA and the NSCMA for DFEs within a time varying context. Therefore, approximate expressions of the mean square errors are derived. We prove that in the steady state, the NSCMA exhibits better performance than the classical CMA. These new results are confirmed by simulation. Through an experimental study, we demonstrate that the Bit Error Rate (BER) is reduced by the NSCMA-DFE, and the improvement of the BER achieved by the NSCMA-DFE is as significant as the channel time variations are severe.
The Turkish Version of Inventory of the Dimensions of Emerging Adulthood(The IDEA)
Emerging Adulthood, the period during ages 18 to 25, is a new conceptualitation proposed by Arnett which is especially prevalent in the industrialized countries. Turkey is basically a developing country having a young population structure. Investigating the presence of such a life period in such a culture might be helpful in understanding educational and psychological needs of people who are in their twenties. With the aim of investigating Emerging Adulthood in Turkey, a well-known instrument (IDEA, 2003) was adapted to Turkish language and Turkish culture. The scale was administered to 296 participants between 15 and 34 ages and validity and reliability were conducted. Exploratory factor analysis revealed three subscales. Reliability coefficients of the scale (Cronbach a) was found as .69. Test-retest reliability coefficients was found for the scale as .81. Finally, “The IDEA" with 20 items was obtained to be used in the Turkish population. The instrument is ready to be administered among Turkish young people for the investigation of transition to adulthood, and whether such a emerging adulthood period really existed.
Swedish: Being or Becoming? Immigration, National Identity and the Democratic State

This article discusses superordinate national identity as a means for immigrants integration into democratic polities. It is suggested that a superordinate national identity perceived as inclusive, by immigrants and by the native population, would be conducive to such integration. Command of the dominant language of society is seen as most important of the inclusive criteria. Other such criteria are respect of the country's political institutions and feelings of belonging to the country where you live. The argument is supported by data, showing a majority in favour of inclusive criteria for 'Swedishness', from a recent study among 1000 secondary school students of 'Swedish' and non-'Swedish' backgrounds.

The Effect of a Free -Trade Agreement upon Agricultural Imports
A free-trade agreement is found to increase Thailand-s agricultural imports from New Zealand, despite the short span of time for which the agreement has been operational. The finding is described by autoregressive estimates that correct for possible unit roots in the data. The agreement-s effect upon imports is also estimated while considering an error-correction model of imports against gross domestic product.
The Role of Faith-based Organizations in Building Democratic Process: Achieving Universal Primary Education in Sierra Leone

This paper aims to argue that religion and Faith-based Organizations (FBOs) contribute to building democratic process through the provision of education in Sierra Leone. Sierra Leone experienced a civil war from 1991 to 2002 and about 70 percent of the population lives in poverty. While the government has been in the process of rebuilding the nation, many forms of Civil Society Organizations (CSOs), including FBOs, have played a significant role in promoting social development. Education plays an important role in supporting people-s democratic movements through knowledge acquisition, spiritual enlightenment and empowerment. This paper discusses religious tolerance in Sierra Leone and how FBOs have contributed to the provision of primary education in Sierra Leone. This study is based on the author-s field research, which involved interviews with teachers and development stakeholders, notably government officials, Non-governmental Organizations (NGOs) and FBOs, as well as questionnaires completed by pupils, parents and teachers.

A Retrospective Analysis of a Professional Learning Community: How Teachers- Capacities Shaped It
The purpose of this paper is to describe the process of setting up a learning community within an elementary school in Ontario, Canada. The description is provided through reflection and examination of field notes taken during the yearlong training and implementation process. Specifically the impact of teachers- capacity on the creation of a learning community was of interest. This paper is intended to inform and add to the debate around the tensions that exist in implementing a bottom-up professional development model like the learning community in a top-down organizational structure. My reflections of the process illustrate that implementation of the learning community professional development model may be difficult and yet transformative in the professional lives of the teachers, students, and administration involved in the change process. I conclude by suggesting the need for a new model of professional development that requires a transformative shift in power dynamics and a shift in the view of what constitutes effective professional learning.
Principal Role and School Structure
This main purpose of the study reported here was to investigate the extent to which the form of school governance (particularly decision-making) had an impact upon the effectiveness of the school with reference to parental involvement, planning and budgeting, professional development of teachers, school facilities and resources, and student outcomes. Particular attention was given to decision-making within the governance arrangements. The study was based on four case studies of high schools in New South Wales, Australia including one government school, one independent Christian community school, one independent Catholic school, and one Catholic systemic school. The focus of the research was principals, teachers, parents, and students of four schools with varying governance structures. To gain a greater insight into the issues, the researchers collected information by questionnaire, semi-structured interview, and review of school key documents. This study found that it was not so much structure but the centrality of the school Principal and the way that the Principal perceived his/her roles in relation to others that impacted most on school governance.
The Wheel Garden: Project-Based Learning for Cross Curriculum Education

In this article, we discuss project-based learning in the context of a wheel garden as an instructional tool in science and mathematics education. A wheel garden provides multiple opportunities to teach across the curriculum, to integrate disciplines, and to promote community involvement. Grounded in the theoretical framework of constructivism, the wheel garden provides a multidisciplined educational tool that provides a hands-on, non-traditional arena for learning. We will examine some of the cultural, art, science, and mathematics connections made with this project.

Gendered Power Relations in the School:Construction of Schoolgirl Femininities in a Turkish High School

In this paper our aim is to explore the construction of schoolgirl femininities, drawing on the results of an ethnographic study conducted in a high school in Ankara, Turkey. In this case study which tries to explore the complexities of gender discourses, we were initially motivated by the questions that have been put forward by critical and feminist literature on education which emphasize the necessarily conflicting and partial nature of both reproduction and resistance and the importance of gendered power relations in the school context. Drawing on this paradigm our research tries to address to a more specific question: how are multiple schoolgirl femininities constructed within the context of gendered school culture, and especially in relation to hegemonic masculinity? Our study reveals that the general framework of multiple femininities is engendered by a tension between two inter-related positions. The first one is different strategies of accommodation and resistance to the gender-related problems of education. The second one is the school experience of girls which is conditioned by their differential position vis-à-vis the masculine resistance culture that is dominant in the school.

Creative Thinking Skill Approach Through Problem-Based Learning: Pedagogy and Practice in the Engineering Classroom
Problem-based learning (PBL) is one of the student centered approaches and has been considered by a number of higher educational institutions in many parts of the world as a method of delivery. This paper presents a creative thinking approach for implementing Problem-based Learning in Mechanics of Structure within a Malaysian Polytechnics environment. In the learning process, students learn how to analyze the problem given among the students and sharing classroom knowledge into practice. Further, through this course-s emphasis on problem-based learning, students acquire creative thinking skills and professional skills as they tackle complex, interdisciplinary and real-situation problems. Once the creative ideas are generated, there are useful additional techniques for tender ideas that will grow into a productive concept or solution. The combination of creative skills and technical abilities will enable the students to be ready to “hit-the-ground-running" and produce in industry when they graduate.
Teacher Education Reform and InternationalGlobalization Hegemony: Issues and Challengesin Turkish Teacher Education
Educational reforms are focused point of different nations. New reform movements generally claim that something is wrong with the current state of affairs, and that the system is deficient in its goals, its accomplishments and it is accused not being adopted into global changes all over the world. It is the same for Turkish education system. It is considered those recent reforms of teacher education in Turkey and the extent to which they reflect a response to global economic pressures. The paper challenges the view that such imposes are inevitable determinants of educational policy and argues that any country will need to develop its own national approach to modernizing teacher education in light of the global context and its particular circumstances. It draws on the idea of reflexive modernization developed by educators and discusses its implications for teacher education policy. The paper deals with four themes teacher education in last decade policy in Turkey; the shift away from the educational disciplines, the shift towards school-based approaches, and the emergence of more centralized forms of accountability of teacher competence.
The Way Classroom Functions: Another Hidden Curriculum to be Explored
This paper seeks to explore the actual classroom setting, to examine its role for students- learning, and attitude in the class. It presents a theoretical approach of the classroom as system to be explored and examines the concrete reality of Greek secondary education students, under the light of the above approach. Based on the findings of a quantitative and qualitative research, authors propose a rather ontological approach of the classroom and underline what the key-elements for such approach should be. The paper explores extensively the theoretical dimensions for the change of paradigm required and addresses the new issues to be considered.
Robust H8 Fuzzy Control Design for Nonlinear Two-Time Scale System with Markovian Jumps based on LMI Approach

This paper examines the problem of designing a robust H8 state-feedback controller for a class of nonlinear two-time scale systems with Markovian Jumps described by a Takagi-Sugeno (TS) fuzzy model. Based on a linear matrix inequality (LMI) approach, LMI-based sufficient conditions for the uncertain Markovian jump nonlinear two-time scale systems to have an H8 performance are derived. The proposed approach does not involve the separation of states into slow and fast ones and it can be applied not only to standard, but also to nonstandard nonlinear two-time scale systems. A numerical example is provided to illustrate the design developed in this paper.

MIBiClus: Mutual Information based Biclustering Algorithm

Most of the biclustering/projected clustering algorithms are based either on the Euclidean distance or correlation coefficient which capture only linear relationships. However, in many applications, like gene expression data and word-document data, non linear relationships may exist between the objects. Mutual Information between two variables provides a more general criterion to investigate dependencies amongst variables. In this paper, we improve upon our previous algorithm that uses mutual information for biclustering in terms of computation time and also the type of clusters identified. The algorithm is able to find biclusters with mixed relationships and is faster than the previous one. To the best of our knowledge, none of the other existing algorithms for biclustering have used mutual information as a similarity measure. We present the experimental results on synthetic data as well as on the yeast expression data. Biclusters on the yeast data were found to be biologically and statistically significant using GO Tool Box and FuncAssociate.

Customer Knowledge and Service Development, the Web 2.0 Role in Co-production
The paper is concerned with relationships between SSME and ICTs and focuses on the role of Web 2.0 tools in the service development process. The research presented aims at exploring how collaborative technologies can support and improve service processes, highlighting customer centrality and value coproduction. The core idea of the paper is the centrality of user participation and the collaborative technologies as enabling factors; Wikipedia is analyzed as an example. The result of such analysis is the identification and description of a pattern characterising specific services in which users collaborate by means of web tools with value co-producers during the service process. The pattern of collaborative co-production concerning several categories of services including knowledge based services is then discussed.
A New IT-Convergence Service Design Framework
In many countries, digital city or ubiquitous city (u-City) projects have been initiated to provide digitalized economic environments to cities. Recently in Korea, Kangwon Province has started the u-Kangwon project to boost local economy with digitalized tourism services. We analyze the limitations of the ubiquitous IT approach through the u-Kangwon case. We have found that travelers are more interested in quality over speed in access of information. For improved service quality, we are looking to develop an IT-convergence service design framework (ISDF). The ISDF is based on the service engineering technique and composed of three parts: Service Design, Service Simulation, and the Service Platform.
Construction of Intersection of Nondeterministic Finite Automata using Z Notation
Functionalities and control behavior are both primary requirements in design of a complex system. Automata theory plays an important role in modeling behavior of a system. Z is an ideal notation which is used for describing state space of a system and then defining operations over it. Consequently, an integration of automata and Z will be an effective tool for increasing modeling power for a complex system. Further, nondeterministic finite automata (NFA) may have different implementations and therefore it is needed to verify the transformation from diagrams to a code. If we describe formal specification of an NFA before implementing it, then confidence over transformation can be increased. In this paper, we have given a procedure for integrating NFA and Z. Complement of a special type of NFA is defined. Then union of two NFAs is formalized after defining their complements. Finally, formal construction of intersection of NFAs is described. The specification of this relationship is analyzed and validated using Z/EVES tool.
Environmental Efficiency of Electric Power Industry of the United States: A Data Envelopment Analysis Approach

Importance of environmental efficiency of electric power industry stems from high demand for energy combined with global warming concerns. It is especially essential for the world largest economies like that of the United States. The paper introduces a Data Envelopment Analysis (DEA) model of environmental efficiency using indicators of fossil fuels utilization, emissions rate, and electric power losses. Using DEA is advantageous in this situation over other approaches due to its nonparametric nature. The paper analyzes data for the period of 1990 - 2006 by comparing actual yearly levels in each dimension with the best values of partial indicators for the period. As positive factors of efficiency, tendency to the decline in emissions rates starting 2000, and in electric power losses starting 2004 may be mentioned together with increasing trend of fuel utilization starting 1999. As a result, dynamics of environmental efficiency is positive starting 2002. The main concern is the decline in fossil fuels utilization in 2006. This negative change should be reversed to comply with ecological and economic requirements.

The Experimental and Numerical Analysis of a Lightpipe using a Simulation Software
A lightpipe is an about 99 percent specular reflective mirror pipe or duct that is used for the transmission of the daylight from the outside into a building. The lightpipes are usually used in the daylighting buildings, in the residential, industrial and commercial sectors. This paper is about the performances of a lightpipe installed in a laboratory (3 m x 2.6 m x 3 m) without windows. The aim is to analyse the luminous intensity distribution for several sky/sun conditions. The lightpipe was monitored during the year 2006. The lightpipe is 1 m long and the diameter of the top collector and of the internal diffuser device is 0.25 m. In the laboratory there are seven illuminance sensors: one external is located on the roof of the laboratory and six internal sensors are connected to a data acquisition system. The internal sensors are positioned under the internal diffusive device at an height of 0.85 m from the floor to simulate a working plane. The numerical data are obtained through a simulation software. This paper shows the comparison between the experimental and numerical results concerning the behavior of the lightpipe.
Intelligent Mobile Search Oriented to Global e-Commerce
In this paper we propose a novel approach for searching eCommerce products using a mobile phone, illustrated by a prototype eCoMobile. This approach aims to globalize the mobile search by integrating the concept of user multilinguism into it. To show that, we particularly deal with English and Arabic languages. Indeed the mobile user can formulate his query on a commercial product in either language (English/Arabic). The description of his information need on commercial products relies on the ontology that represents the conceptualization of the product catalogue knowledge domain defined in both English and Arabic languages. A query expressed on a mobile device client defines the concept that corresponds to the name of the product followed by a set of pairs (property, value) specifying the characteristics of the product. Once a query is submitted it is then communicated to the server side which analyses it and in its turn performs an http request to an eCommerce application server (like Amazon). This latter responds by returning an XML file representing a set of elements where each element defines an item of the searched product with its specific characteristics. The XML file is analyzed on the server side and then items are displayed on the mobile device client along with its relevant characteristics in the chosen language.
The Application of Non-quantitative Modelling in the Analysis of a Network Warfare Environment

Network warfare is an emerging concept that focuses on the network and computer based forms through which information is attacked and defended. Various computer and network security concepts thus play a role in network warfare. Due the intricacy of the various interacting components, a model to better understand the complexity in a network warfare environment would be beneficial. Non-quantitative modeling is a useful method to better characterize the field due to the rich ideas that can be generated based on the use of secular associations, chronological origins, linked concepts, categorizations and context specifications. This paper proposes the use of non-quantitative methods through a morphological analysis to better explore and define the influential conditions in a network warfare environment.

E-health in Rural Areas: Case of Developing Countries

The Application of e-health solutions has brought superb advancements in the health care industry. E-health solutions have already been embraced in the industrialized countries. In an effort to catch up with the growth, the developing countries have strived to revolutionize the healthcare industry by use of Information technology in different ways. Based on a technology assessment carried out in Kenya – one of the developing countries – and using multiple case studies in Nyanza Province, this work focuses on an investigation on how five rural hospitals are adapting to the technology shift. The issues examined include the ICT infrastructure and e-health technologies in place, the knowledge of participants in terms of benefits gained through the use of ICT and the challenges posing barriers to the use of ICT technologies in these hospitals. The results reveal that the ICT infrastructure in place is inadequate for e-health implementations as a result to various challenges that exist. Consequently, suggestions on how to tackle the various challenges have been addressed in this paper.

Convergence of ICT and Education
Information and communication technology (ICT) has become, within a very short time, one of the basic building blocks of modern society. Many countries now understanding the importance of ICT and mastering the basic skills and concepts of it as part of the core of education. Organizations, experts and practitioners in the education sector increasingly recognizing the importance of ICT in supporting educational improvement and reform. This paper addresses the convergence of ICT and education. When two technologies are converging to each other, together they will generate some great opportunities and challenges. This paper focuses on these issues. In introduction section, it explains the ICT, education, and ICT-enhanced education. In next section it describes need of ICT in education, relationship between ICT skills and education, and stages of teaching learning process. The next two sections describe opportunities and challenges in integrating ICT in education. Finally the concluding section summaries the idea and its usefulness.
QoS Management in the Future Internet
The talks about technological convergence had been around for almost twenty years. Today Internet made it possible. And this is not only technical evolution. The way it changed our lives reflected in variety of applications, services and technologies used in day-to-day life. Such benefits imposed even more requirements on heterogeneous and unreliable IP networks. Current paper outlines QoS management system developed in the NetQoS [1] project. It describes an overall architecture of management system for heterogeneous networks and proposes automated multi-layer QoS management. Paper focuses on the structure of the most crucial modules of the system that enable autonomous and multi-layer provisioning and dynamic adaptation.
A Noble Flow Rate Control based on Leaky Bucket Method for Multi-Media OBS Networks
Optical burst switching (OBS) has been proposed to realize the next generation Internet based on the wavelength division multiplexing (WDM) network technologies. In the OBS, the burst contention is one of the major problems. The deflection routing has been designed for resolving the problem. However, the deflection routing becomes difficult to prevent from the burst contentions as the network load becomes high. In this paper, we introduce a flow rate control methods to reduce burst contentions. We propose new flow rate control methods based on the leaky bucket algorithm and deflection routing, i.e. separate leaky bucket deflection method, and dynamic leaky bucket deflection method. In proposed methods, edge nodes which generate data bursts carry out the flow rate control protocols. In order to verify the effectiveness of the flow rate control in OBS networks, we show that the proposed methods improve the network utilization and reduce the burst loss probability through computer simulations.
Remarks Regarding Queuing Model and Packet Loss Probability for the Traffic with Self-Similar Characteristics
Network management techniques have long been of interest to the networking research community. The queue size plays a critical role for the network performance. The adequate size of the queue maintains Quality of Service (QoS) requirements within limited network capacity for as many users as possible. The appropriate estimation of the queuing model parameters is crucial for both initial size estimation and during the process of resource allocation. The accurate resource allocation model for the management system increases the network utilization. The present paper demonstrates the results of empirical observation of memory allocation for packet-based services.
A Distributed Topology Control Algorithm to Conserve Energy in Heterogeneous Wireless Mesh Networks
A considerable amount of energy is consumed during transmission and reception of messages in a wireless mesh network (WMN). Reducing per-node transmission power would greatly increase the network lifetime via power conservation in addition to increasing the network capacity via better spatial bandwidth reuse. In this work, the problem of topology control in a hybrid WMN of heterogeneous wireless devices with varying maximum transmission ranges is considered. A localized distributed topology control algorithm is presented which calculates the optimal transmission power so that (1) network connectivity is maintained (2) node transmission power is reduced to cover only the nearest neighbours (3) networks lifetime is extended. Simulations and analysis of results are carried out in the NS-2 environment to demonstrate the correctness and effectiveness of the proposed algorithm.
An Efficient Data Mining Approach on Compressed Transactions
In an era of knowledge explosion, the growth of data increases rapidly day by day. Since data storage is a limited resource, how to reduce the data space in the process becomes a challenge issue. Data compression provides a good solution which can lower the required space. Data mining has many useful applications in recent years because it can help users discover interesting knowledge in large databases. However, existing compression algorithms are not appropriate for data mining. In [1, 2], two different approaches were proposed to compress databases and then perform the data mining process. However, they all lack the ability to decompress the data to their original state and improve the data mining performance. In this research a new approach called Mining Merged Transactions with the Quantification Table (M2TQT) was proposed to solve these problems. M2TQT uses the relationship of transactions to merge related transactions and builds a quantification table to prune the candidate itemsets which are impossible to become frequent in order to improve the performance of mining association rules. The experiments show that M2TQT performs better than existing approaches.
mCRM-s New Opportunities of Customer Satisfaction
This paper aims at a new challenge of customer satisfaction on mobile customer relationship management. In this paper presents a conceptualization of mCRM on its unique characteristics of customer satisfaction. Also, this paper develops an empirical framework in conception of customer satisfaction in mCRM. A single-case study is applied as the methodology. In order to gain an overall view of the empirical case, this paper accesses to invisible and important information of company in this investigation. Interview is the key data source form the main informants of the company through which the issues are identified and the proposed framework is built. It supports the development of customer satisfaction in mCRM; links this theoretical framework into practice; and provides the direction for future research. Therefore, this paper is very useful for the industries as it helps them to understand how customer satisfaction changes the mCRM structure and increase the business competitive advantage. Finally, this paper provides a contribution in practice by linking a theoretical framework in conception of customer satisfaction in mCRM for companies to a practical real case.
Casting Users- Perspectives on Foundries as Suppliers
Global competition is tightening and companies have to think how to remain competitive. The main aim of this paper is to discuss how Finnish foundries will remain competitive. To fulfil the aim, we conducted interviews in nine companies using castings and analysed buyer–supplier relationships, current competitive advantages of Finnish foundries and customer perspectives on how Finnish foundries remain competitive. We found that the customerfoundry relationship is still closer to traditional subcontracting than partnering and general image of foundries is negative. Current competitive advantages of Finnish foundries include designing cooperation, proximity and flexibility. Casting users state that Finnish foundries should sell their know-how and services instead of their capacity, concentrate on prototype, single and short series castings and supply ready-to-install cast components directly to customers- assembly lines.
Issues in Procurement of Castings
The aim of this paper is to present current and future procedures in castings procurement. Differences in procurement are highlighted. The supplier selection criteria used in practice is compared to literature findings. Different trends related to supply chains are presented and it is described how they are reflected in reality to castings procurement. To fulfil the aim, interviews were conducted in nine companies using castings. It was found that largest casting users have the most subcontractor foundries and it is more typical that they have multiple suppliers for the same parts. Currently only two companies out of nine purchase castings outside Europe, but the others are also progressing in the same direction. The main reason is the need to lower purchasing costs. Another trend is that all companies want to buy cast components or sub-assemblies instead of raw castings from foundries. It was found that price is a main supplier selection criterion. All companies use competitive bidding in supplier selection.
The Epistemological Crisis in the Theory of Vittorio Guidano

This work shows a basic philosophical difficulty in the constructivist foundations of the cognitive posracionalist psychology of Vittorio Guidano. This is a difficulty caused by the problem of the existential crisis. It will be analyzed how Guidano-s suggestions about this problem depend on felt experience. Then it will appear how Guidano-s philosophy and psychotherapy must turn towards a phenomenological approach. Finally, some references are given about Eugen Gendlin-s philosophy which could be considered as a radical way to confront these questions.

The Use of a Tactical Simulator as a Learning Resource at the Norwegian Military Academy
The Norwegian Military Academy (Army) has been using a tactical simulator for the last two years. During this time there has been some discussion concerning how to use the simulator most efficiently and what type of learning one achieves by using the simulator. The problem that is addressed in this paper is how simulators can be used as a learning resource for students concerned with developing their military profession. The aim of this article is to create a wider consciousness regarding the use of a simulator while educating officers in a military profession. The article discusses the use of simulators from two different perspectives. The first perspective deals with using the simulator as a computer game, and the second perspective looks at the simulator as a socio-cultural artefact. Furthermore the article discusses four different ways the simulator can be looked upon as a useful learning resource when educating students of a military profession.
Building an e-Learning System Model with Implications for Research and Instructional Use
This paper demonstrates a model of an e-Learning system based on nowadays learning theory and distant education practice. The relationships in the model are designed to be simple and functional and do not necessarily represent any particular e- Learning environments. It is meant to be a generic e-Learning system model with implications for any distant education course instructional design. It allows online instructors to move away from the discrepancy between the courses and body of knowledge. The interrelationships of four primary sectors that are at the e-Learning system are presented in this paper. This integrated model includes [1] pedagogy, [2] technology, [3] teaching, and [4] learning. There are interactions within each of these sectors depicted by system loop map.
Instruction Resource Recommendation Services for Elementary Schools in Taiwan

In the past, there were more researches of recommendation system in applied electronic commerce. However, because all circles promote information technology integrative instruction actively, the quantity of instruction resources website is more and more increasing on the Internet. But there are less website including recommendation service, especially for teachers. This study established an instruction resource recommendation website that analyzed teaching style of teachers, then provided appropriate instruction resources for teachers immediately. We used the questionnaire survey to realize teacher-s suggestions and satisfactions with the instruction resource contents and recommendation results. The study shows: (1)The website used “Transactional Ability Inventory" that realized teacher-s style and provided appropriate instruction resources for teachers in a short time, it reduced the step of data filter. (2)According to the content satisfaction of questionnaire survey, four styles teachers were almost satisfied with the contents of the instruction resources that the website recommended, thus, the conception of developing instruction resources with different teaching style is accepted. (3) According to the recommendation satisfaction of questionnaire survey, four styles teachers were almost satisfied with the recommendation service of the website, thus, the recommendation strategy that provide different results for teachers in different teaching styles is accepted.

Impact Assessment using Path Models of Microentrepreneurs developed by a Business Corporation in India
For scores of years now, several microfinance organizations, non governmental organizations and other welfare organizations have, with a view to aiding the progress of communities rooted in poverty have been focusing on creating microentrepreneurs, besides taking several other measures. In recent times, business corporations have joined forces to combat poverty by taking up microenterprise development. Hindustan Unilever Limited (HUL), the Indian subsidiary of Unilever Limited exemplifies this through its Project Shakti. The company through the Project creates rural women entrepreneurs by making them direct to home sales distributors of its products in villages that have thus far been ignored by multinational corporations. The members participating in Project Shakti are largely self help group members. The paper focuses on assessing the impact made by the company on the members engaged in Project Shakti. The analysis involves use of quantitative methods to study the effect of Project Shakti on those self help group members engaged in Project Shakti and those not engaged with Project Shakti. Path analysis has been used to study the impact made on those members engaged in Project Shakti. Significant differences were observed on fronts of entrepreneurial development, economic empowerment and social empowerment between members associated with Project Shakti and those not associated with Project Shakti. Path analysis demonstrated that involvement in Project Shakti led to entrepreneurial development resulting in economic empowerment that in turn led to social empowerment and that these three elements independently induced a feeling of privilege in the women for being associated with the Project.
Human Capital and Capability Approach in European Lifelong Learning Development: A Case Study of Macedonia in the Balkan

The paper discusses European Lifelong Learning policy in the European enlargement to the Balkan. The European Lifelong Learning policy with Human Capital approach is researched in the country case of Macedonia. The paper argues that Human Capital approach focusing on instrumental and economic importance of learning for employability and economic growth needs to be complemented with Capability Approach for intrinsic and noneconomic needs of learning among the ethnic minorities. The paper identifies two dimensions of importance – minority languages and civic education – that the Capability Approach may develop to guarantee equal opportunities to all to benefit from European educational and lifelong learning development and to build an inclusive and socially just democracy in Macedonia.

The Relationship between Burnout, Negative Affectivity and Organizational Citizenship Behavior for Human Services Employees
The purpose of this study was to explore the relationship between Burnout, Negative Affectivity, and Organizational Citizenship Behavior (OCB) for social service workers at two agencies serving homeless populations. Thirty two subjects completed surveys. Significant correlations between major variables and subscales were found.
Analyses of Socio-Cognitive Identity Styles by Slovak Adolescents
The contribution deals with analysis of identity style at adolescents (N=463) at the age from 16 to 19 (the average age is 17,7 years). We used the Identity Style Inventory by Berzonsky, distinguishing three basic, measured identity styles: informational, normative, diffuse-avoidant identity style and also commitment. The informational identity style influencing on personal adaptability, coping strategies, quality of life and the normative identity style, it means the style in which an individual takes on models of authorities at self-defining were found to have the highest representation in the studied group of adolescents by higher scores at girls in comparison with boys. The normative identity style positively correlates with the informational identity style. The diffuse-avoidant identity style was found to be positively associated with maladaptive decisional strategies, neuroticism and depressive reactions. There is the style, in which the individual shifts aside defining his personality. In our research sample the lowest score represents it and negatively correlates with commitment, it means with coping strategies, thrust in oneself and the surrounding world. The age of adolescents did not significantly differentiate representation of identity style. We were finding the model, in which informational and normative identity style had positive relationship and the informational and diffuseavoidant style had negative relationship, which were determinated with commitment. In the same time the commitment is influenced with other outside factors.
Influence of Social-Psychological Training on Selected Features of University Students
We presented results of research aimed on findings influence of social - psychological training (realized with students of Constantine the Philosopher University- future teachers within their undergraduate preparation) on the choice of intrapersonal and interpersonal features. After social- psychological training using Interpersonal Check List (ICL) we found out shift of behavior to more adaptive forms in categories, which are characterized by extroversive friendly behavior, willingness to cooperation, conformity regard to social situation, responsible and regardful behavior. Using State-Trait Anxiety Inventory (STAI) we found out the cut down of state anxiety and of trait anxiety. The report was processed within grants KEGA 3/5269/07 and VEGA 1/3675/06.
Regret, Choice, and Outcome
In two studies we challenged the well consolidated position in regret literature according to which the necessary condition for the emergence of regret is a bad outcome ensuing from free decisions. Without free choice, and, consequently, personal responsibility, other emotions, such as disappointment, but not regret, are supposed to be elicited. In our opinion, a main source of regret is being obliged by circumstance out of our control to chose an undesired option. We tested the hypothesis that regret resulting from a forced choice is more intense than regret derived from a free choice and that the outcome affects the latter, not the former. Besides, we investigated whether two other variables – the perception of the level of freedom of the choice and the choice justifiability – mediated the relationships between choice and regret, as well as the other four emotions we examined: satisfaction, anger toward oneself, disappointment, anger towards circumstances. The two studies were based on the scenario methodology and implied a 2 x 2 (choice x outcome) between design. In the first study the foreseen short-term effects of the choice were assessed; in the second study the experienced long-term effects of the choice were assessed. In each study 160 students of the Second University of Naples participated. Results largely corroborated our hypotheses. They were discussed in the light of the main theories on regret and decision making.
Cohabiting in Multiethnic Community: Forms, Representations and Images of the Diversity

Modern culture, based on disinhibition of cultural trends and on heterodirection, is promoting openmindedness attitudes towards ethnic diversity, but on the other hand also new forms of social representations of the foreigner. Social representation is situated between the psychic field and the social one; it is the representation of oneself and of the other one, hanging between social categories and individual inner world. We will produce the results of a research on the representation of the foreigner, built on the type of prejudice prevailing among middle-low or middle-high educational qualification subjects, in which prejudicial attitudes seem to descend from precise mental images of the foreigner.

Terrorism's Fear : Perceived Personal and National Threats

Terrorism represents an unexpected and unwanted change which challenges one-s social identity. We carried out a study to explore the demographic variables- role on the perception of personal and national threat, and to investigate the effects of perceived terrorist threat on people-s ways of life, moods, opinions and hopes. 313 residents of Palermo (Italy) were interviewed. The results pointed out that the fear of terrorism affects three areas: the cognitive, the emotional and the behavioural one.

Aggressive Driving in Young Motorists
Road rage is an increasingly prevalent expression of aggression in our society. Its dangers are apparent and understanding its causes may shed light on preventative measures. This study involved a fifteen-minute survey administered to 147 undergraduate students at a North Eastern suburban university. The survey consisted of a demographics section, questions regarding financial investment in respondents- vehicles, experience driving, habits of driving, experiences witnessing role models driving, and an evaluation of road rage behavior using the Driving Vengeance Questionnaire. The study found no significant differences in driving aggression between respondents who were financially invested in their vehicle compared to those who were not, or between respondents who drove in heavy traffic hours compared to those who did not, suggesting internal factors correlate with aggressive driving habits. The study also found significant differences in driving aggression between males versus females, those with more points on their license versus fewer points, and those who witnessed parents driving aggressively very often versus rarely or never. Additional studies can investigate how witnessing parents driving aggressively is related to future driving behaviors.
Explorations in the Role of Emotion in Moral Judgment
Recent theorizations on the cognitive process of moral judgment have focused on the role of intuitions and emotions, marking a departure from previous emphasis on conscious, step-by-step reasoning. My study investigated how being in a disgusted mood state affects moral judgment. Participants were induced to enter a disgusted mood state through listening to disgusting sounds and reading disgusting descriptions. Results shows that they, when compared to control who have not been induced to feel disgust, are more likely to endorse actions that are emotionally aversive but maximizes utilitarian return The result is analyzed using the 'emotion-as-information' approach to decision making. The result is consistent with the view that emotions play an important role in determining moral judgment.
Reform-Oriented Teaching of Introductory Statistics in the Health, Social and Behavioral Sciences – Historical Context and Rationale

There is widespread emphasis on reform in the teaching of introductory statistics at the college level. Underpinning this reform is a consensus among educators and practitioners that traditional curricular materials and pedagogical strategies have not been effective in promoting statistical literacy, a competency that is becoming increasingly necessary for effective decision-making and evidence-based practice. This paper explains the historical context of, and rationale for reform-oriented teaching of introductory statistics (at the college level) in the health, social and behavioral sciences (evidence-based disciplines). A firm understanding and appreciation of the basis for change in pedagogical approach is important, in order to facilitate commitment to reform, consensus building on appropriate strategies, and adoption and maintenance of best practices. In essence, reform-oriented pedagogy, in this context, is a function of the interaction among content, pedagogy, technology, and assessment. The challenge is to create an appropriate balance among these domains.

Sun, Salon, and Cosmetic Tanning: Predictors and Motives

The appearance management behavior of tanning by gay men is examined through the lens of Impression Formation. The study proposes that body image, self-esteem, and internalized homophobia are connected and affect the motives for engaging in sun, salon, and cosmetic tanning. Motives examined were: to look masculine, to look attractive to (potential) partners, to look attractive in general, to socialize, to meet a peer standard, and for personal satisfaction. Using regression analysis to examine data of 103 gay men who engage in at least one method of tanning, results reveal that components of body image and internalized homophobia–but not self-esteem–are linked to various motives and methods of tanning. These findings support and extend the literature of Impression Formation Theory and provide practitioners in the health and healthrelated fields new avenues to pursue when dealing with diseases related to tanning.

A Hybrid Neural Network and Traditional Approach for Forecasting Lumpy Demand
Accurate demand forecasting is one of the most key issues in inventory management of spare parts. The problem of modeling future consumption becomes especially difficult for lumpy patterns, which characterized by intervals in which there is no demand and, periods with actual demand occurrences with large variation in demand levels. However, many of the forecasting methods may perform poorly when demand for an item is lumpy. In this study based on the characteristic of lumpy demand patterns of spare parts a hybrid forecasting approach has been developed, which use a multi-layered perceptron neural network and a traditional recursive method for forecasting future demands. In the described approach the multi-layered perceptron are adapted to forecast occurrences of non-zero demands, and then a conventional recursive method is used to estimate the quantity of non-zero demands. In order to evaluate the performance of the proposed approach, their forecasts were compared to those obtained by using Syntetos & Boylan approximation, recently employed multi-layered perceptron neural network, generalized regression neural network and elman recurrent neural network in this area. The models were applied to forecast future demand of spare parts of Arak Petrochemical Company in Iran, using 30 types of real data sets. The results indicate that the forecasts obtained by using our proposed mode are superior to those obtained by using other methods.
The Risk and Value Engineering Structures and their Integration with Industrial Projects Management (A Case Study on I. K.Corporation)
Value engineering is an efficacious contraption for administrators to make up their minds. Value perusals proffer the gaffers a suitable instrument to decrease the expenditures of the life span, quality amelioration, structural improvement, curtailment of the construction schedule, longevity prolongation or a merging of the aforementioned cases. Subjecting organizers to pressures on one hand and their accountability towards their pertinent fields together with inherent risks and ambiguities of other options on the other hand set some comptrollers in a dilemma utilization of risk management and the value engineering in projects manipulation with regard to complexities of implementing projects can be wielded as a contraption to identify and efface each item which wreaks unnecessary expenses and time squandering sans inflicting any damages upon the essential project applications. Of course It should be noted that implementation of risk management and value engineering with regard to the betterment of efficiency and functions may lead to the project implementation timing elongation. Here time revamping does not refer to time diminishing in the whole cases. his article deals with risk and value engineering conceptualizations at first. The germane reverberations effectuated due to its execution in Iran Khodro Corporation are regarded together with the joint features and amalgamation of the aforesaid entia; hence the proposed blueprint is submitted to be taken advantage of in engineering and industrial projects including Iran Khodro Corporation.
Using Degree of Adaptive (DOA) Model for Partner Selection in Supply Chain
In order to reduce cost, increase quality, and for timely supplying production systems has considerably taken the advantages of supply chain management and these advantages are also competitive. Selection of appropriate supplier has an important role in improvement and efficiency of systems. The models of supplier selection which have already been used by researchers have considered selection one or more suppliers from potential suppliers but in this paper selecting one supplier as partner from one supplier that have minimum one period supplying to buyer is considered. This paper presents a conceptual model for partner selection and application of Degree of Adoptive (DOA) model for final selection. The attributes weight in this model is prepared through AHP model. After making the descriptive model, determining the attributes and measuring the parameters of the adaptive is examined in an auto industry of Iran(Zagross Khodro co.) and results are presented.
An Agent-Based Scheduling Framework for Flexible Manufacturing Systems
The concept of flexible manufacturing is highly appealing in gaining a competitive edge in the market by quickly adapting to the changing customer needs. Scheduling jobs on flexible manufacturing systems (FMSs) is a challenging task of managing the available flexibility on the shop floor to react to the dynamics of the environment in real-time. In this paper, an agent-oriented scheduling framework that can be integrated with a real or a simulated FMS is proposed. This framework works in stochastic environments with a dynamic model of job arrival. It supports a hierarchical cooperative scheduling that builds on the available flexibility of the shop floor. Testing the framework on a model of a real FMS showed the capability of the proposed approach to overcome the drawbacks of the conventional approaches and maintain a near optimal solution despite the dynamics of the operational environment.
Application of Association Rule Mining in Supplier Selection Criteria
In this paper the application of rule mining in order to review the effective factors on supplier selection is reviewed in the following three sections 1) criteria selecting and information gathering 2) performing association rule mining 3) validation and constituting rule base. Afterwards a few of applications of rule base is explained. Then, a numerical example is presented and analyzed by Clementine software. Some of extracted rules as well as the results are presented at the end.
Risk Evaluation of Information Technology Projects Based on Fuzzy Analytic Hierarchal Process
Information Technology (IT) projects are always accompanied by various risks and because of high rate of failure in such projects, managing risks in order to neutralize or at least decrease their effects on the success of the project is strongly essential. In this paper, fuzzy analytical hierarchy process (FAHP) is exploited as a means of risk evaluation methodology to prioritize and organize risk factors faced in IT projects. A real case of IT projects, a project of design and implementation of an integrated information system in a vehicle producing company in Iran is studied. Related risk factors are identified and then expert qualitative judgments about these factors are acquired. Translating these judgments to fuzzy numbers and using them as an input to FAHP, risk factors are then ranked and prioritized by FAHP in order to make project managers aware of more important risks and enable them to adopt suitable measures to deal with these highly devastative risks.
An Intelligent System Framework for Generating Activity List of a Project Using WBS Mind map and Semantic Network
Work Breakdown Structure (WBS) is one of the most vital planning processes of the project management since it is considered to be the fundamental of other processes like scheduling, controlling, assigning responsibilities, etc. In fact WBS or activity list is the heart of a project and omission of a simple task can lead to an irrecoverable result. There are some tools in order to generate a project WBS. One of the most powerful tools is mind mapping which is the basis of this article. Mind map is a method for thinking together and helps a project manager to stimulate the mind of project team members to generate project WBS. Here we try to generate a WBS of a sample project involving with the building construction using the aid of mind map and the artificial intelligence (AI) programming language. Since mind map structure can not represent data in a computerized way, we convert it to a semantic network which can be used by the computer and then extract the final WBS from the semantic network by the prolog programming language. This method will result a comprehensive WBS and decrease the probability of omitting project tasks.
A New Framework and a Model for Product Development with an Application in the Telecommunications Services Sector
This paper argues that a product development exercise involves in addition to the conventional stages, several decisions regarding other aspects. These aspects should be addressed simultaneously in order to develop a product that responds to the customer needs and that helps realize objectives of the stakeholders in terms of profitability, market share and the like. We present a framework that encompasses these different development dimensions. The framework shows that a product development methodology such as the Quality Function Deployment (QFD) is the basic tool which allows definition of the target specifications of a new product. Creativity is the first dimension that enables the development exercise to live and end successfully. A number of group processes need to be followed by the development team in order to ensure enough creativity and innovation. Secondly, packaging is considered to be an important extension of the product. Branding strategies, quality and standardization requirements, identification technologies, design technologies, production technologies and costing and pricing are also integral parts to the development exercise. These dimensions constitute the proposed framework. The paper also presents a mathematical model used to calculate the design targets based on the target costing principle. The framework is used to study a case of a new product development in the telecommunications services sector.
Developing Efficient Testing and Unloading Procedures for a Local Sewage Holding Pit
A local municipality has decided to build a sewage pit to receive residential sewage waste arriving by tank trucks. Daily accumulated waste are to be pumped to a nearby waste water treatment facility to be re-consumed for agricultural and construction projects. A discrete-event simulation model using Arena Software was constructed to assist in defining the capacity of the system in cubic meters, number of tank trucks to use the system, number of unload docks required, number of standby areas needed and manpower required for data collection at entrance checkpoint and truck tank load toxicity testing. The results of the model are statistically validated. Simulation turned out to be an excellent tool in the facility planning effort for the pit project, as it insured smooth flow lines of tank trucks load discharge and best utilization of facilities on site.
CFD Analysis of Two Phase Flow in a Horizontal Pipe – Prediction of Pressure Drop
In designing of condensers, the prediction of pressure drop is as important as the prediction of heat transfer coefficient. Modeling of two phase flow, particularly liquid – vapor flow under diabatic conditions inside a horizontal tube using CFD analysis is difficult with the available two phase models in FLUENT due to continuously changing flow patterns. In the present analysis, CFD analysis of two phase flow of refrigerants inside a horizontal tube of inner diameter, 0.0085 m and 1.2 m length is carried out using homogeneous model under adiabatic conditions. The refrigerants considered are R22, R134a and R407C. The analysis is performed at different saturation temperatures and at different flow rates to evaluate the local frictional pressure drop. Using Homogeneous model, average properties are obtained for each of the refrigerants that is considered as single phase pseudo fluid. The so obtained pressure drop data is compared with the separated flow models available in literature.
Experimental and Numerical Simulation of Fire in a Scaled Underground Station
The objective of this study is to investigate fire behaviors, experimentally and numerically, in a scaled version of an underground station. The effect of ventilation velocity on the fire is examined. Fire experiments are simulated by burning 10 ml isopropyl alcohol fuel in a fire pool with dimensions 5cm x 10cm x 4 mm at the center of 1/100 scaled underground station model. A commercial CFD program FLUENT was used in numerical simulations. For air flow simulations, k-ω SST turbulence model and for combustion simulation, non-premixed combustion model are used. This study showed that, the ventilation velocity is increased from 1 m/s to 3 m/s the maximum temperature in the station is found to be less for ventilation velocity of 1 m/s. The reason for these experimental result lies on the relative dominance of oxygen supply effect on cooling effect. Without piston effect, maximum temperature occurs above the fuel pool. However, when the ventilation velocity increased the flame was tilted in the direction of ventilation and the location of maximum temperature moves along the flow direction. The velocities measured experimentally in the station at different locations are well matched by the CFD simulation results. The prediction of general flow pattern is satisfactory with the smoke visualization tests. The backlayering in velocity is well predicted by CFD simulation. However, all over the station, the CFD simulations predicted higher temperatures compared to experimental measurements.
The Application of HLLC Numerical Solver to the Reduced Multiphase Model

The performance of high-resolution schemes is investigated for unsteady, inviscid and compressible multiphase flows. An Eulerian diffuse interface approach has been chosen for the simulation of multicomponent flow problems. The reduced fiveequation and seven equation models are used with HLL and HLLC approximation. The authors demonstrated the advantages and disadvantages of both seven equations and five equations models studying their performance with HLL and HLLC algorithms on simple test case. The seven equation model is based on two pressure, two velocity concept of Baer–Nunziato [10], while five equation model is based on the mixture velocity and pressure. The numerical evaluations of two variants of Riemann solvers have been conducted for the classical one-dimensional air-water shock tube and compared with analytical solution for error analysis.

The Comparison of Form Drag and Profile Dragof a Wind Turbine Blade Section in Pitching Oscillation
Extensive wind tunnel tests have been conducted to investigate the unsteady flow field over and behind a 2D model of a 660 kW wind turbine blade section in pitching motion. The surface pressure and wake dynamic pressure variation at a distance of 1.5 chord length from trailing edge were measured by pressure transducers during several oscillating cycles at 3 reduced frequencies and oscillating amplitudes. Moreover, form drag and linear momentum deficit are extracted and compared at various conditions. The results show that the wake velocity field and surface pressure of the model have similar behavior before and after the airfoil beyond the static stall angle of attack. In addition, the effects of reduced frequency and oscillation amplitudes are discussed.
The Effect of Confinement Shapes on Over-Reinforced HSC Beams
High strength concrete (HSC) provides high strength but lower ductility than normal strength concrete. This low ductility limits the benefit of using HSC in building safe structures. On the other hand, when designing reinforced concrete beams, designers have to limit the amount of tensile reinforcement to prevent the brittle failure of concrete. Therefore the full potential of the use of steel reinforcement can not be achieved. This paper presents the idea of confining concrete in the compression zone so that the HSC will be in a state of triaxial compression, which leads to improvements in strength and ductility. Five beams made of HSC were cast and tested. The cross section of the beams was 200×300 mm, with a length of 4 m and a clear span of 3.6 m subjected to four-point loading, with emphasis placed on the midspan deflection. The first beam served as a reference beam. The remaining beams had different tensile reinforcement and the confinement shapes were changed to gauge their effectiveness in improving the strength and ductility of the beams. The compressive strength of the concrete was 85 MPa and the tensile strength of the steel was 500 MPa and for the stirrups and helixes was 250 MPa. Results of testing the five beams proved that placing helixes with different diameters as a variable parameter in the compression zone of reinforced concrete beams improve their strength and ductility.
Magnetohydrodynamic Damping of Natural Convection Flows in a Rectangular Enclosure
We numerically study the three-dimensional magnetohydrodynamics (MHD) stability of oscillatory natural convection flow in a rectangular cavity, with free top surface, filled with a liquid metal, having an aspect ratio equal to A=L/H=5, and subjected to a transversal temperature gradient and a uniform magnetic field oriented in x and z directions. The finite volume method was used in order to solve the equations of continuity, momentum, energy, and potential. The stability diagram obtained in this study highlights the dependence of the critical value of the Grashof number Grcrit , with the increase of the Hartmann number Ha for two orientations of the magnetic field. This study confirms the possibility of stabilization of a liquid metal flow in natural convection by application of a magnetic field and shows that the flow stability is more important when the direction of magnetic field is longitudinal than when the direction is transversal.
Performance Improvement of a Supersonic External Compression Inlet by Heat Source Addition
Heat source addition to the axisymmetric supersonic inlet may improve the performance parameters, which will increase the inlet efficiency. In this investigation the heat has been added to the flow field at some distance ahead of an axisymmetric inlet by adding an imaginary thermal source upstream of cowl lip. The effect of heat addition on the drag coefficient, mass flow rate and the overall efficiency of the inlet have been investigated. The results show that heat addition causes flow separation, hence to prevent this phenomena, roughness has been added on the spike surface. However, heat addition reduces the drag coefficient and the inlet mass flow rate considerably. Furthermore, the effects of position, size, and shape on the inlet performance were studied. It is found that the thermal source deflects the flow streamlines. By improper location of the thermal source, the optimum condition has been obtained. For the optimum condition, the drag coefficient is considerably reduced and the inlet mass flow rate and its efficiency have been increased slightly. The optimum shape of the heat source is obtained too.
Control of Pressure Gradient in the Contraction of a Wind Tunnel
Subsonic wind tunnel experiments were conducted to study the effect of tripped boundary layer on the pressure distribution in the contraction region of the tunnel. Measurements were performed by installing trip strip at two different positions in the concave portion of the contraction. The results show that installation of the trip strips, have significant effects on both turbulence and pressure distribution. The reduction in the free stream turbulence and reduction of the wall static pressure distribution deferred signified with the location of the trip strip.
Electroviscous Effects in Low Reynolds Number Flow through a Microfluidic Contraction with Rectangular Cross-Section
The electrokinetic flow resistance (electroviscous effect) is predicted for steady state, pressure-driven liquid flow at low Reynolds number in a microfluidic contraction of rectangular cross-section. Calculations of the three dimensional flow are performed in parallel using a finite volume numerical method. The channel walls are assumed to carry a uniform charge density and the liquid is taken to be a symmetric 1:1 electrolyte. Predictions are presented for a single set of flow and electrokinetic parameters. It is shown that the magnitude of the streaming potential gradient and the charge density of counter-ions in the liquid is greater than that in corresponding two-dimensional slit-like contraction geometry. The apparent viscosity is found to be very close to the value for a rectangular channel of uniform cross-section at the chosen Reynolds number (Re = 0.1). It is speculated that the apparent viscosity for the contraction geometry will increase as the Reynolds number is reduced.
Fluid Flow Analysis and Design of a Flow Distributor in a Domestic Gas Boiler Using a Commercial CFD Software
The aim of the study was to investigate the possible use of commercial Computational Fluid Dynamics (CFD) software in the design process of a domestic gas boiler. Because of the limited computational resources some simplifications had to be made in order to contribute to the design in a reasonable timescale. The porous media model was used in order to simulate the influence of the pressure drop characteristic of particular elements of a heat transfer system on the water-flow distribution in the system. Further, a combination of CFD analyses and spread sheet calculations was used in order to solve the flow distribution problem.
On Discretization of Second-order Derivatives in Smoothed Particle Hydrodynamics
Discretization of spatial derivatives is an important issue in meshfree methods especially when the derivative terms contain non-linear coefficients. In this paper, various methods used for discretization of second-order spatial derivatives are investigated in the context of Smoothed Particle Hydrodynamics. Three popular forms (i.e. "double summation", "second-order kernel derivation", and "difference scheme") are studied using one-dimensional unsteady heat conduction equation. To assess these schemes, transient response to a step function initial condition is considered. Due to parabolic nature of the heat equation, one can expect smooth and monotone solutions. It is shown, however in this paper, that regardless of the type of kernel function used and the size of smoothing radius, the double summation discretization form leads to non-physical oscillations which persist in the solution. Also, results show that when a second-order kernel derivative is used, a high-order kernel function shall be employed in such a way that the distance of inflection point from origin in the kernel function be less than the nearest particle distance. Otherwise, solutions may exhibit oscillations near discontinuities unlike the "difference scheme" which unconditionally produces monotone results.
Application of Artificial Neural Network for the Prediction of Pressure Distribution of a Plunging Airfoil
Series of experimental tests were conducted on a section of a 660 kW wind turbine blade to measure the pressure distribution of this model oscillating in plunging motion. In order to minimize the amount of data required to predict aerodynamic loads of the airfoil, a General Regression Neural Network, GRNN, was trained using the measured experimental data. The network once proved to be accurate enough, was used to predict the flow behavior of the airfoil for the desired conditions. Results showed that with using a few of the acquired data, the trained neural network was able to predict accurate results with minimal errors when compared with the corresponding measured values. Therefore with employing this trained network the aerodynamic coefficients of the plunging airfoil, are predicted accurately at different oscillation frequencies, amplitudes, and angles of attack; hence reducing the cost of tests while achieving acceptable accuracy.
The Investigation of Motor Cooling Performance
This study experimentally and numerically investigates motor cooling performance. The motor consists of a centrifugal fan, two axial fans, a shaft, a stator, a rotor and a heat exchanger with 637 cooling tubes. The pressure rise-flow rate (P-Q) performance curves of the cooling fans at 1800 rpm are tested using a test apparatus complying with the Chinese National Standard (CNS) 2726. Compared with the experimental measurements, the numerical analysis results show that the P-Q performance curves of the axial fan and centrifugal fan can be estimated within about 2% and 6%, respectively. By using the simplified model, setting up the heat exchanger and stator as porous media, the flow field in the motor is calculated. By using the results of the flow field near the rotor and stator, and subjecting the heat generation rate as a boundary condition, the temperature distributions of the stator and rotor are also calculated. The simulation results show that the calculated temperature of the stator winding near the axial fans is lower by about 5% than the measured value, and the calculated temperature of the stator core located at the center of the stator is about 1% higher than the measured value. Besides, discussion is made to improve the motor cooling performance.
On the Flow of a Third Grade Viscoelastic Fluid in an Orthogonal Rheometer

The flow of a third grade fluid in an orthogonal rheometer is studied. We employ the admissible velocity field proposed in [5]. We solve the problem and obtain the velocity field as well as the components for the Cauchy tensor. We compare the results with those from [9]. Some diagrams concerning the velocity and Cauchy stress components profiles are presented for different values of material constants and compared with the corresponding values for a linear viscous fluid.

Effects of Annealing Treatment on Optical Properties of Anatase TiO2 Thin Films
In this investigation, anatase TiO2 thin films were grown by radio frequency magnetron sputtering on glass substrates at a high sputtering pressure and room temperature. The anatase films were then annealed at 300-600 °C in air for a period of 1 hour. To examine the structure and morphology of the films, X-ray diffraction (XRD) and atomic force microscopy (AFM) methods were used respectively. From X-ray diffraction patterns of the TiO2 films, it was found that the as-deposited film showed some differences compared with the annealed films and the intensities of the peaks of the crystalline phase increased with the increase of annealing temperature. From AFM images, the distinct variations in the morphology of the thin films were also observed. The optical constants were characterized using the transmission spectra of the films obtained by UV-VIS-IR spectrophotometer. Besides, optical thickness of the film deposited at room temperature was calculated and cross-checked by taking a cross-sectional image through SEM. The optical band gaps were evaluated through Tauc model. It was observed that TiO2 films produced at room temperatures exhibited high visible transmittance and transmittance decreased slightly with the increase of annealing temperatures. The films were found to be crystalline having anatase phase. The refractive index of the films was found from 2.31-2.35 in the visible range. The extinction coefficient was nearly zero in the visible range and was found to increase with annealing temperature. The allowed indirect optical band gap of the films was estimated to be in the range from 3.39 to 3.42 eV which showed a small variation. The allowed direct band gap was found to increase from 3.67 to 3.72 eV. The porosity was also found to decrease at a higher annealing temperature making the film compact and dense.
Experimental Validation of the Predicted Performance of a Wind Driven Venturi Ventilator
The paper presents the results of simple measurements conducted on a model of a wind-driven venturi-type room ventilator. The ventilator design is new and was developed employing mathematical modeling. However, the computational model was not validated experimentally for the particular application considered. The paper presents the performance of the ventilator model under laboratory conditions, for five different wind tunnel speeds. The results are used to both demonstrate the effectiveness of the new design and to validate the computational model employed to develop it.
The New Semi-Experimental Method for Simulation of Turbine Flow Meters Rotation in the Transitional Flow
The new semi-experimental method for simulation of the turbine flow meters rotation in the transitional flow has been developed. The method is based on the experimentally established exponential low of changing of dimensionless relative turbine gas meter rotation frequency and meter inertia time constant. For experimental evaluation of the meter time constant special facility has been developed. The facility ensures instant switching of turbine meter under test from one channel to the other channel with different flow rate and measuring the meter response. The developed method can be used for evaluation and predication of the turbine meters response and dynamic error in the transitional flow with any arbitrary law of flow rate changing. The examples of the method application are presented.
Low Air Velocity Measurement Characteristics- Variation Due to Flow Regime

The paper depicts air velocity values, reproduced by laser Doppler anemometer (LDA) and ultrasonic anemometer (UA), relations with calculated ones from flow rate measurements using the gas meter which calibration uncertainty is ± (0.15 – 0.30) %. Investigation had been performed in channel installed in aerodynamical facility used as a part of national standard of air velocity. Relations defined in a research let us confirm the LDA and UA for air velocity reproduction to be the most advantageous measures. The results affirm ultrasonic anemometer to be reliable and favourable instrument for measurement of mean velocity or control of velocity stability in the velocity range of 0.05 m/s – 10 (15) m/s when the LDA used. The main aim of this research is to investigate low velocity regularities, starting from 0.05 m/s, including region of turbulent, laminar and transitional air flows. Theoretical and experimental results and brief analysis of it are given in the paper. Maximum and mean velocity relations for transitional air flow having unique distribution are represented. Transitional flow having distinctive and different from laminar and turbulent flow characteristics experimentally have not yet been analysed.

Experimental Study of the Pressure Drop after Fractal-Shaped Orifices in a Turbulent Flow Pipe
The fractal-shaped orifices are assumed to have a significant effect on the pressure drop downstream pipe flow due to their edge self-similarity shape which enhances the mixing properties. Here, we investigate the pressure drop after these fractals using a digital micro-manometer at different stations downstream a turbulent flow pipe then a direct comparison has been made with the pressure drop measured from regular orifices with the same flow area. Our results showed that the fractal-shaped orifices have a significant effect on the pressure drop downstream the flow. Also the pressure drop measured across the fractal-shaped orifices is noticed to be lower that that from ordinary orifices of the same flow areas. This result could be important in designing piping systems from point of view of losses consideration with the same flow control area. This is promising to use the fractal-shaped orifices as flowmeters as they can sense the pressure drop across them accurately with minimum losses than the regular ones.
Optimal Design of Airfoil with High Aspect Ratio in Unmanned Aerial Vehicles
Shape optimization of the airfoil with high aspect ratio of long endurance unmanned aerial vehicle (UAV) is performed by the multi-objective optimization technology coupled with computational fluid dynamics (CFD). For predicting the aerodynamic characteristics around the airfoil the high-fidelity Navier-Stokes solver is employed and SMOGA (Simple Multi-Objective Genetic Algorithm), which is developed by authors, is used for solving the multi-objective optimization problem. To obtain the optimal solutions of the design variable (i.e., sectional airfoil profile, wing taper ratio and sweep) for high performance of UAVs, both the lift and lift-to-drag ratio are maximized whereas the pitching moment should be minimized, simultaneously. It is found that the lift force and lift-to-drag ratio are linearly dependent and a unique and dominant solution are existed. However, a trade-off phenomenon is observed between the lift-to-drag ratio and pitching moment. As the result of optimization, sixty-five (65) non-dominated Pareto individuals at the cutting edge of design spaces that is decided by airfoil shapes can be obtained.
Super Resolution Blind Reconstruction of Low Resolution Images using Wavelets based Fusion
Crucial information barely visible to the human eye is often embedded in a series of low resolution images taken of the same scene. Super resolution reconstruction is the process of combining several low resolution images into a single higher resolution image. The ideal algorithm should be fast, and should add sharpness and details, both at edges and in regions without adding artifacts. In this paper we propose a super resolution blind reconstruction technique for linearly degraded images. In our proposed technique the algorithm is divided into three parts an image registration, wavelets based fusion and an image restoration. In this paper three low resolution images are considered which may sub pixels shifted, rotated, blurred or noisy, the sub pixel shifted images are registered using affine transformation model; A wavelet based fusion is performed and the noise is removed using soft thresolding. Our proposed technique reduces blocking artifacts and also smoothens the edges and it is also able to restore high frequency details in an image. Our technique is efficient and computationally fast having clear perspective of real time implementation.
A 3D Virtual Navigation System Integrating User Positioning and Pre-Download Mechanism
This paper takes the actual scene of Aletheia University campus – the Class 2 national monument, the first educational institute in northern Taiwan as an example, to present a 3D virtual navigation system which supports user positioning and pre-download mechanism. The proposed system was designed based on the principle of Voronoi Diagra) to divide the virtual scenes and its multimedia information, which combining outdoor GPS positioning and the indoor RFID location detecting function. When users carry mobile equipments such as notebook computer, UMPC, EeePC...etc., walking around the actual scenes of indoor and outdoor areas of campus, this system can automatically detect the moving path of users and pre-download the needed data so that users will have a smooth and seamless navigation without waiting.
Performance Enhancement of Motion Estimation Using SSE2 Technology
Motion estimation is the most computationally intensive part in video processing. Many fast motion estimation algorithms have been proposed to decrease the computational complexity by reducing the number of candidate motion vectors. However, these studies are for fast search algorithms themselves while almost image and video compressions are operated with software based. Therefore, the timing constraints for running these motion estimation algorithms not only challenge for the video codec but also overwhelm for some of processors. In this paper, the performance of motion estimation is enhanced by using Intel's Streaming SIMD Extension 2 (SSE2) technology with Intel Pentium 4 processor.
Realtime Lip Contour Tracking For Audio-Visual Speech Recognition Applications
Detection and tracking of the lip contour is an important issue in speechreading. While there are solutions for lip tracking once a good contour initialization in the first frame is available, the problem of finding such a good initialization is not yet solved automatically, but done manually. We have developed a new tracking solution for lip contour detection using only few landmarks (15 to 25) and applying the well known Active Shape Models (ASM). The proposed method is a new LMS-like adaptive scheme based on an Auto regressive (AR) model that has been fit on the landmark variations in successive video frames. Moreover, we propose an extra motion compensation model to address more general cases in lip tracking. Computer simulations demonstrate a fair match between the true and the estimated spatial pixels. Significant improvements related to the well known LMS approach has been obtained via a defined Frobenius norm index.
Infrared Face Recognition Using Distance Transforms
In this work we present an efficient approach for face recognition in the infrared spectrum. In the proposed approach physiological features are extracted from thermal images in order to build a unique thermal faceprint. Then, a distance transform is used to get an invariant representation for face recognition. The obtained physiological features are related to the distribution of blood vessels under the face skin. This blood network is unique to each individual and can be used in infrared face recognition. The obtained results are promising and show the effectiveness of the proposed scheme.
Monitoring of Dielectric Losses and Use of Ferrofluids for Bushing Cooling
At present, the tendency to implement the conditionbased maintenance (CBM), which allows the optimization of the expenses for equipment monitoring, is more and more evident; also, the transformer substations with remote monitoring are increasingly used. This paper reviews all the advantages of the on-line monitoring and presents an equipment for on-line monitoring of bushings, which is the own contribution of specialists who are the authors of this paper. The paper presents a study of the temperature field, using the finite element method. For carrying out this study, the 3D modelling of the above mentioned bushing was performed. The analysis study is done taking into account the extreme thermal stresses, focusing at the level of the first cooling wing section of the ceramic insulator. This fact enables to justify the tanδ variation in time, depending on the transformer loading and the environmental conditions. With a view to reducing the variation of dielectric losses in bushing insulation, the use of ferrofuids instead of mineral oils is proposed.
Hybrid Method Using Wavelets and Predictive Method for Compression of Speech Signal
The development of the signal compression algorithms is having compressive progress. These algorithms are continuously improved by new tools and aim to reduce, an average, the number of bits necessary to the signal representation by means of minimizing the reconstruction error. The following article proposes the compression of Arabic speech signal by a hybrid method combining the wavelet transform and the linear prediction. The adopted approach rests, on one hand, on the original signal decomposition by ways of analysis filters, which is followed by the compression stage, and on the other hand, on the application of the order 5, as well as, the compression signal coefficients. The aim of this approach is the estimation of the predicted error, which will be coded and transmitted. The decoding operation is then used to reconstitute the original signal. Thus, the adequate choice of the bench of filters is useful to the transform in necessary to increase the compression rate and induce an impercevable distortion from an auditive point of view.
Electrical Impedance Imaging Using Eddy Current
Electric impedance imaging is a method of reconstructing spatial distribution of electrical conductivity inside a subject. In this paper, a new method of electrical impedance imaging using eddy current is proposed. The eddy current distribution in the body depends on the conductivity distribution and the magnetic field pattern. By changing the position of magnetic core, a set of voltage differences is measured with a pair of electrodes. This set of voltage differences is used in image reconstruction of conductivity distribution. The least square error minimization method is used as a reconstruction algorithm. The back projection algorithm is used to get two dimensional images. Based on this principle, a measurement system is developed and some model experiments were performed with a saline filled phantom. The shape of each model in the reconstructed image is similar to the corresponding model, respectively. From the results of these experiments, it is confirmed that the proposed method is applicable in the realization of electrical imaging.
Intelligent Off-Grid Photovoltaic Supply Systems
Off-grid Photovoltaic (PV) systems are empowering technology in underdeveloped countries like Ethiopia where many people live far away from the modern world. Where there is relatively low energy consumption, providing energy from grid systems is not commercially cost-effective. As a result, significant people groups worldwide stay without access to electricity. One remote village in northern Ethiopia was selected by the United Nations for a pilot project to improve its living conditions. As part of this comprehensive project, an intelligent charge controller circuit for Off-grid PV systems was designed for the clinic in that village. In this paper, design aspects of an intelligent charge controller unit and its load driver circuits are discussed for an efficient utilization of PVbased supply systems.
Sensitivity of Small Disturbance Angle Stability to the System Parameters of Future Power Networks

The incorporation of renewable energy sources for the sustainable electricity production is undertaking a more prominent role in electric power systems. Thus, it will be an indispensable incident that the characteristics of future power networks, their prospective stability for instance, get influenced by the imposed features of sustainable energy sources. One of the distinctive attributes of the sustainable energy sources is exhibiting the stochastic behavior. This paper investigates the impacts of this stochastic behavior on the small disturbance rotor angle stability in the upcoming electric power networks. Considering the various types of renewable energy sources and the vast variety of system configurations, the sensitivity analysis can be an efficient breakthrough towards generalizing the effects of new energy sources on the concept of stability. In this paper, the definition of small disturbance angle stability for future power systems and the iterative-stochastic way of its analysis are presented. Also, the effects of system parameters on this type of stability are described by performing a sensitivity analysis for an electric power test system.

Applying Wavelet Entropy Principle in Fault Classification
The ability to detect and classify the type of fault plays a great role in the protection of power system. This procedure is required to be precise with no time consumption. In this paper detection of fault type has been implemented using wavelet analysis together with wavelet entropy principle. The simulation of power system is carried out using PSCAD/EMTDC. Different types of faults were studied obtaining various current waveforms. These current waveforms were decomposed using wavelet analysis into different approximation and details. The wavelet entropy of such decompositions is analyzed reaching a successful methodology for fault classification. The suggested approach is tested using different fault types and proven successful identification for the type of fault.
The Influence of the Commons Structure Modification on the Allocation
The tracing methods determine the contribution the power system sources have in their supplying. The methods can be used to assess the transmission prices, but also to recover the transmission fixed cost. In this paper is presented the influence of the modification of commons structure has on the specific price of transfer. The operator must make use of a few basic principles about allocation. Most tracing methods are based on the proportional sharing principle. In this paper Kirschen method is used. In order to illustrate this method, the 25- bus test system is used, elaborated within the Electrical Power Engineering Department, from Timisoara, Romania.
Analysis of Electrical Installation of a Photovoltaic Power Park in Greece
The scope of this paper is to describe a real electrical installation of renewable energy using photovoltaic cells. The displayed power grid connected network was established in 2007 at area of Northern Greece. The photovoltaic park is composed of 6120 photovoltaic cells able to deliver a total power of 1.101.600 Wp. For the transformation of DC voltage to AC voltage have been used 25 stand alone three phases inverters and for the connection at the medium voltage network of Greek Power Authority have been installed two oil immersed transformer of 630 kVA each one. Due to the wide space area of installation a specific external lightning protection system has been designed. Additionally, due to the sensitive electronics of the control and protection systems of park, surge protection, equipotent bonding and shielding were also of major importance.
DC Link Floating for Grid Connected PV Converters
Nowadays there are several grid connected converter in the grid system. These grid connected converters are generally the converters of renewable energy sources, industrial four quadrant drives and other converters with DC link. These converters are connected to the grid through a three phase bridge. The standards prescribe the maximal harmonic emission which could be easily limited with high switching frequency. The increased switching losses can be reduced to the half with the utilization of the wellknown Flat-top modulation. The suggested control method is the expansion of the Flat-top modulation with which the losses could be also reduced to the half compared to the Flat-top modulation. Comparing to traditional control these requirements can be simultaneously satisfied much better with the DLF (DC Link Floating) method.
3 State Current Mode of a Grid Connected PV Converter
Nowadays in applications of renewable energy sources it is important to develop powerful and energy-saving photovoltaic converters and to keep the prescriptions of the standards. In grid connected PV converters the obvious solution to increase the efficiency is to reduce the switching losses. Our new developed control method reduces the switching losses and keeps the limitations of the harmonic distortion standards. The base idea of the method is the utilization of 3-state control causing discontinuous current mode at low input power. In the following sections the control theory, the realizations and the simulation results are presented.
Simplified Models to Determine Nodal Voltagesin Problems of Optimal Allocation of Capacitor Banks in Power Distribution Networks
This paper presents two simplified models to determine nodal voltages in power distribution networks. These models allow estimating the impact of the installation of reactive power compensations equipments like fixed or switched capacitor banks. The procedure used to develop the models is similar to the procedure used to develop linear power flow models of transmission lines, which have been widely used in optimization problems of operation planning and system expansion. The steady state non-linear load flow equations are approximated by linear equations relating the voltage amplitude and currents. The approximations of the linear equations are based on the high relationship between line resistance and line reactance (ratio R/X), which is valid for power distribution networks. The performance and accuracy of the models are evaluated through comparisons with the exact results obtained from the solution of the load flow using two test networks: a hypothetical network with 23 nodes and a real network with 217 nodes.
Closely Parametrical Model for an Electrical Arc Furnace
To maximise furnace production it-s necessary to optimise furnace control, with the objectives of achieving maximum power input into the melting process, minimum network distortion and power-off time, without compromise on quality and safety. This can be achieved with on the one hand by an appropriate electrode control and on the other hand by a minimum of AC transformer switching. Electrical arc is a stochastic process; witch is the principal cause of power quality problems, including voltages dips, harmonic distortion, unbalance loads and flicker. So it is difficult to make an appropriate model for an Electrical Arc Furnace (EAF). The factors that effect EAF operation are the melting or refining materials, melting stage, electrode position (arc length), electrode arm control and short circuit power of the feeder. So arc voltages, current and power are defined as a nonlinear function of the arc length. In this article we propose our own empirical function of the EAF and model, for the mean stages of the melting process, thanks to the measurements in the steel factory.
Stability of Electrical Drives Supplied by a Three Level Inverter
The development of the power electronics has allowed increasing the precision and reliability of the electrical devices, thanks to the adjustable inverters, as the Pulse Wide Modulation (PWM) applied to the three level inverters, which is the object of this study. The authors treat the relation between the law order adopted for a given system and the oscillations of the electrical and mechanical parameters of which the tolerance depends on the process with which they are integrated (paper factory, lifting of the heavy loads, etc.).Thus, the best choice of the regulation indexes allows us to achieve stability and safety training without investment (management of existing equipment). The optimal behavior of any electric device can be achieved by the minimization of the stored electrical and mechanical energy.
Project Selection by Using a Fuzzy TOPSIS Technique
Selection of a project among a set of possible alternatives is a difficult task that the decision maker (DM) has to face. In this paper, by using a fuzzy TOPSIS technique we propose a new method for a project selection problem. After reviewing four common methods of comparing investment alternatives (net present value, rate of return, benefit cost analysis and payback period) we use them as criteria in a TOPSIS technique. First we calculate the weight of each criterion by a pairwise comparison and then we utilize the improved TOPSIS assessment for the project selection.
Is Management Science doing Enough to Improve Healthcare?

Healthcare issues continue to pose huge problems and incur massive costs. As a result there are many challenging problems still unresolved. In this paper, we will carry out an extensive scientific survey of different areas of management and planning in an attempt to identify where there has already been a substantial contribution from management science methods to healthcare problems and where there is a clear potential for more work to be done. The focus will be on the read-across to the healthcare domain from such approaches applied generally to management and planning and how the methods can be used to improvement patient care. We conclude that, since the healthcare domain significantly differs from traditional areas of management and planning, in some cases there is a need to modify the approaches so as to incorporate the complexities of healthcare, and fully exploit the potential for improvement.

Parallel Direct Integration Variable Step Block Method for Solving Large System of Higher Order Ordinary Differential Equations
The aim of this paper is to investigate the performance of the developed two point block method designed for two processors for solving directly non stiff large systems of higher order ordinary differential equations (ODEs). The method calculates the numerical solution at two points simultaneously and produces two new equally spaced solution values within a block and it is possible to assign the computational tasks at each time step to a single processor. The algorithm of the method was developed in C language and the parallel computation was done on a parallel shared memory environment. Numerical results are given to compare the efficiency of the developed method to the sequential timing. For large problems, the parallel implementation produced 1.95 speed-up and 98% efficiency for the two processors.
The Riemann Barycenter Computation and Means of Several Matrices

An iterative definition of any n variable mean function is given in this article, which iteratively uses the two-variable form of the corresponding two-variable mean function. This extension method omits recursivity which is an important improvement compared with certain recursive formulas given before by Ando-Li-Mathias, Petz- Temesi. Furthermore it is conjectured here that this iterative algorithm coincides with the solution of the Riemann centroid minimization problem. Certain simulations are given here to compare the convergence rate of the different algorithms given in the literature. These algorithms will be the gradient and the Newton mehod for the Riemann centroid computation.

Direct Block Backward Differentiation Formulas for Solving Second Order Ordinary Differential Equations
In this paper, a direct method based on variable step size Block Backward Differentiation Formula which is referred as BBDF2 for solving second order Ordinary Differential Equations (ODEs) is developed. The advantages of the BBDF2 method over the corresponding sequential variable step variable order Backward Differentiation Formula (BDFVS) when used to solve the same problem as a first order system are pointed out. Numerical results are given to validate the method.
Parallel Block Backward Differentiation Formulas for Solving Ordinary Differential Equations
A parallel block method based on Backward Differentiation Formulas (BDF) is developed for the parallel solution of stiff Ordinary Differential Equations (ODEs). Most common methods for solving stiff systems of ODEs are based on implicit formulae and solved using Newton iteration which requires repeated solution of systems of linear equations with coefficient matrix, I - hβJ . Here, J is the Jacobian matrix of the problem. In this paper, the matrix operations is paralleled in order to reduce the cost of the iterations. Numerical results are given to compare the speedup and efficiency of parallel algorithm and that of sequential algorithm.
Group Similarity Transformation of a Time Dependent Chemical Convective Process

The time dependent progress of a chemical reaction over a flat horizontal plate is here considered. The problem is solved through the group similarity transformation method which reduces the number of independent by one and leads to a set of nonlinear ordinary differential equation. The problem shows a singularity at the chemical reaction order n=1 and is analytically solved through the perturbation method. The behavior of the process is then numerically investigated for n≠1 and different Schmidt numbers. Graphical results for the velocity and concentration of chemicals based on the analytical and numerical solutions are presented and discussed.

Reflection of Plane Waves at Free Surface of an Initially Stressed Dissipative Medium

The paper discuses the effect of initial stresses on the reflection coefficients of plane waves in a dissipative medium. Basic governing equations are formulated in context of Biot's incremental deformation theory. These governing equations are solved analytically to obtain the dimensional phase velocities of plane waves propagating in plane of symmetry. Closed-form expressions for the reflection coefficients of P and SV waves- incident at the free surface of an initially stressed dissipative medium are obtained. Numerical computations, using these expressions, are carried out for a particular model. Computations made with the results predicted in presence and absence of the initial stresses and the results have been shown graphically. The study shows that the presence of compressive initial stresses increases the velocity of longitudinal wave (P-wave) but diminishes that of transverse wave (SV-wave). Also the numerical results presented indicate that initial stresses and dissipation might affect the reflection coefficients significantly.

Error Propagation in the RK5GL3 Method
The RK5GL3 method is a numerical method for solving initial value problems in ordinary differential equations, and is based on a combination of a fifth-order Runge-Kutta method and 3-point Gauss-Legendre quadrature. In this paper we describe the propagation of local errors in this method, and show that the global order of RK5GL3 is expected to be six, one better than the underlying Runge- Kutta method.
Local Error Control in the RK5GL3 Method
The RK5GL3 method is a numerical method for solving initial value problems in ordinary differential equations, and is based on a combination of a fifth-order Runge-Kutta method and 3-point Gauss-Legendre quadrature. In this paper we describe an effective local error control algorithm for RK5GL3, which uses local extrapolation with an eighth-order Runge-Kutta method in tandem with RK5GL3, and a Hermite interpolating polynomial for solution estimation at the Gauss-Legendre quadrature nodes.
Comparison of Finite Difference Schemes for Water Flow in Unsaturated Soils
Flow movement in unsaturated soil can be expressed by a partial differential equation, named Richards equation. The objective of this study is the finding of an appropriate implicit numerical solution for head based Richards equation. Some of the well known finite difference schemes (fully implicit, Crank Nicolson and Runge-Kutta) have been utilized in this study. In addition, the effects of different approximations of moisture capacity function, convergence criteria and time stepping methods were evaluated. Two different infiltration problems were solved to investigate the performance of different schemes. These problems include of vertical water flow in a wet and very dry soils. The numerical solutions of two problems were compared using four evaluation criteria and the results of comparisons showed that fully implicit scheme is better than the other schemes. In addition, utilizing of standard chord slope method for approximation of moisture capacity function, automatic time stepping method and difference between two successive iterations as convergence criterion in the fully implicit scheme can lead to better and more reliable results for simulation of fluid movement in different unsaturated soils.
A Two-Species Model for a Fishing System with Marine Protected Areas
A model of a system concerning one species of demersal (inshore) fish and one of pelagic (offshore) fish undergoing fishing restricted by marine protected areas is proposed in this paper. This setup was based on the FISH-BE model applied to the Tabina fishery in Zamboanga del Sur, Philippines. The components of the model equations have been adapted from widely-accepted mechanisms in population dynamics. The model employs Gompertz-s law of growth and interaction on each type of protected and unprotected subpopulation. Exchange coefficients between protected and unprotected areas were assumed to be proportional to the relative area of the entry region. Fishing harvests were assumed to be proportional to both the number of fishers and the number of unprotected fish. An extra term was included for the pelagic population to allow for the exchange between the unprotected area and the outside environment. The systems were found to be bounded for all parameter values. The equations for the steady state were unsolvable analytically but the existence and uniqueness of non-zero steady states can be proven. Plots also show that an MPA size yielding the maximum steady state of the unprotected population can be found. All steady states were found to be globally asymptotically stable for the entire range of parameter values.
Relationship between Sums of Squares in Linear Regression and Semi-parametric Regression
In this paper, the sum of squares in linear regression is reduced to sum of squares in semi-parametric regression. We indicated that different sums of squares in the linear regression are similar to various deviance statements in semi-parametric regression. In addition to, coefficient of the determination derived in linear regression model is easily generalized to coefficient of the determination of the semi-parametric regression model. Then, it is made an application in order to support the theory of the linear regression and semi-parametric regression. In this way, study is supported with a simulated data example.
On Diffusion Approximation of Discrete Markov Dynamical Systems
The paper is devoted to stochastic analysis of finite dimensional difference equation with dependent on ergodic Markov chain increments, which are proportional to small parameter ". A point-form solution of this difference equation may be represented as vertexes of a time-dependent continuous broken line given on the segment [0,1] with "-dependent scaling of intervals between vertexes. Tending " to zero one may apply stochastic averaging and diffusion approximation procedures and construct continuous approximation of the initial stochastic iterations as an ordinary or stochastic Ito differential equation. The paper proves that for sufficiently small " these equations may be successfully applied not only to approximate finite number of iterations but also for asymptotic analysis of iterations, when number of iterations tends to infinity.
Vol:11 No:04 2017Vol:11 No:03 2017Vol:11 No:02 2017Vol:11 No:01 2017
Vol:10 No:12 2016Vol:10 No:11 2016Vol:10 No:10 2016Vol:10 No:09 2016Vol:10 No:08 2016Vol:10 No:07 2016Vol:10 No:06 2016Vol:10 No:05 2016Vol:10 No:04 2016Vol:10 No:03 2016Vol:10 No:02 2016Vol:10 No:01 2016
Vol:9 No:12 2015Vol:9 No:11 2015Vol:9 No:10 2015Vol:9 No:09 2015Vol:9 No:08 2015Vol:9 No:07 2015Vol:9 No:06 2015Vol:9 No:05 2015Vol:9 No:04 2015Vol:9 No:03 2015Vol:9 No:02 2015Vol:9 No:01 2015
Vol:8 No:12 2014Vol:8 No:11 2014Vol:8 No:10 2014Vol:8 No:09 2014Vol:8 No:08 2014Vol:8 No:07 2014Vol:8 No:06 2014Vol:8 No:05 2014Vol:8 No:04 2014Vol:8 No:03 2014Vol:8 No:02 2014Vol:8 No:01 2014
Vol:7 No:12 2013Vol:7 No:11 2013Vol:7 No:10 2013Vol:7 No:09 2013Vol:7 No:08 2013Vol:7 No:07 2013Vol:7 No:06 2013Vol:7 No:05 2013Vol:7 No:04 2013Vol:7 No:03 2013Vol:7 No:02 2013Vol:7 No:01 2013
Vol:6 No:12 2012Vol:6 No:11 2012Vol:6 No:10 2012Vol:6 No:09 2012Vol:6 No:08 2012Vol:6 No:07 2012Vol:6 No:06 2012Vol:6 No:05 2012Vol:6 No:04 2012Vol:6 No:03 2012Vol:6 No:02 2012Vol:6 No:01 2012
Vol:5 No:12 2011Vol:5 No:11 2011Vol:5 No:10 2011Vol:5 No:09 2011Vol:5 No:08 2011Vol:5 No:07 2011Vol:5 No:06 2011Vol:5 No:05 2011Vol:5 No:04 2011Vol:5 No:03 2011Vol:5 No:02 2011Vol:5 No:01 2011
Vol:4 No:12 2010Vol:4 No:11 2010Vol:4 No:10 2010Vol:4 No:09 2010Vol:4 No:08 2010Vol:4 No:07 2010Vol:4 No:06 2010Vol:4 No:05 2010Vol:4 No:04 2010Vol:4 No:03 2010Vol:4 No:02 2010Vol:4 No:01 2010
Vol:3 No:12 2009Vol:3 No:11 2009Vol:3 No:10 2009Vol:3 No:09 2009Vol:3 No:08 2009Vol:3 No:07 2009Vol:3 No:06 2009Vol:3 No:05 2009Vol:3 No:04 2009Vol:3 No:03 2009Vol:3 No:02 2009Vol:3 No:01 2009
Vol:2 No:12 2008Vol:2 No:11 2008Vol:2 No:10 2008Vol:2 No:09 2008Vol:2 No:08 2008Vol:2 No:07 2008Vol:2 No:06 2008Vol:2 No:05 2008Vol:2 No:04 2008Vol:2 No:03 2008Vol:2 No:02 2008Vol:2 No:01 2008
Vol:1 No:12 2007Vol:1 No:11 2007Vol:1 No:10 2007Vol:1 No:09 2007Vol:1 No:08 2007Vol:1 No:07 2007Vol:1 No:06 2007Vol:1 No:05 2007Vol:1 No:04 2007Vol:1 No:03 2007Vol:1 No:02 2007Vol:1 No:01 2007