Excellence in Research and Innovation for Humanity

International Science Index

Commenced in January 1999 Frequency: Monthly Edition: International Paper Count: 150

Modeling and FOS Feedback Based Control of SISO Intelligent Structures with Embedded Shear Sensors and Actuators
Active vibration control is an important problem in structures. The objective of active vibration control is to reduce the vibrations of a system by automatic modification of the system-s structural response. In this paper, the modeling and design of a fast output sampling feedback controller for a smart flexible beam system embedded with shear sensors and actuators for SISO system using Timoshenko beam theory is proposed. FEM theory, Timoshenko beam theory and the state space techniques are used to model the aluminum cantilever beam. For the SISO case, the beam is divided into 5 finite elements and the control actuator is placed at finite element position 1, whereas the sensor is varied from position 2 to 5, i.e., from the nearby fixed end to the free end. Controllers are designed using FOS method and the performance of the designed FOS controller is evaluated for vibration control for 4 SISO models of the same plant. The effect of placing the sensor at different locations on the beam is observed and the performance of the controller is evaluated for vibration control. Some of the limitations of the Euler-Bernoulli theory such as the neglection of shear and axial displacement are being considered here, thus giving rise to an accurate beam model. Embedded shear sensors and actuators have been considered in this paper instead of the surface mounted sensors and actuators for vibration suppression because of lot of advantages. In controlling the vibration modes, the first three dominant modes of vibration of the system are considered.
Self-Organization of Clusters having Locally Distributed Patterns for Synchronized Inputs
Many experimental results suggest that more precise spike timing is significant in neural information processing. We construct a self-organization model using the spatiotemporal patterns, where Spike-Timing Dependent Plasticity (STDP) tunes the conduction delays between neurons. We show that the fluctuation of conduction delays causes globally continuous and locally distributed firing patterns through the self-organization.
A New Quantile Based Fuzzy Time Series Forecasting Model

Time series models have been used to make predictions of academic enrollments, weather, road accident, casualties and stock prices, etc. Based on the concepts of quartile regression models, we have developed a simple time variant quantile based fuzzy time series forecasting method. The proposed method bases the forecast using prediction of future trend of the data. In place of actual quantiles of the data at each point, we have converted the statistical concept into fuzzy concept by using fuzzy quantiles using fuzzy membership function ensemble. We have given a fuzzy metric to use the trend forecast and calculate the future value. The proposed model is applied for TAIFEX forecasting. It is shown that proposed method work best as compared to other models when compared with respect to model complexity and forecasting accuracy.

Stochastic Simulation of Reaction-Diffusion Systems

Reactiondiffusion systems are mathematical models that describe how the concentration of one or more substances distributed in space changes under the influence of local chemical reactions in which the substances are converted into each other, and diffusion which causes the substances to spread out in space. The classical representation of a reaction-diffusion system is given by semi-linear parabolic partial differential equations, whose general form is ÔêétX(x, t) = DΔX(x, t), where X(x, t) is the state vector, D is the matrix of the diffusion coefficients and Δ is the Laplace operator. If the solute move in an homogeneous system in thermal equilibrium, the diffusion coefficients are constants that do not depend on the local concentration of solvent and of solutes and on local temperature of the medium. In this paper a new stochastic reaction-diffusion model in which the diffusion coefficients are function of the local concentration, viscosity and frictional forces of solvent and solute is presented. Such a model provides a more realistic description of the molecular kinetics in non-homogenoeus and highly structured media as the intra- and inter-cellular spaces. The movement of a molecule A from a region i to a region j of the space is described as a first order reaction Ai k- → Aj , where the rate constant k depends on the diffusion coefficient. Representing the diffusional motion as a chemical reaction allows to assimilate a reaction-diffusion system to a pure reaction system and to simulate it with Gillespie-inspired stochastic simulation algorithms. The stochastic time evolution of the system is given by the occurrence of diffusion events and chemical reaction events. At each time step an event (reaction or diffusion) is selected from a probability distribution of waiting times determined by the specific speed of reaction and diffusion events. Redi is the software tool, developed to implement the model of reaction-diffusion kinetics and dynamics. It is a free software, that can be downloaded from http://www.cosbi.eu. To demonstrate the validity of the new reaction-diffusion model, the simulation results of the chaperone-assisted protein folding in cytoplasm obtained with Redi are reported. This case study is redrawing the attention of the scientific community due to current interests on protein aggregation as a potential cause for neurodegenerative diseases.

An Analysis of Global Stability of Cohen-Grossberg Neural Networks with Multiple Time Delays
This paper presents a new sufficient condition for the existence, uniqueness and global asymptotic stability of the equilibrium point for Cohen-Grossberg neural networks with multiple time delays. The results establish a relationship between the network parameters of the neural system independently of the delay parameters. The results are also compared with the previously reported results in the literature.
Knowledge Management in Cross- Organizational Networks as Illustrated by One of the Largest European ICT Associations A Case Study of the “METORA

In networks, mainly small and medium-sized businesses benefit from the knowledge, experiences and solutions offered by experts from industry and science or from the exchange with practitioners. Associations which focus, among other things, on networking, information and knowledge transfer and which are interested in supporting such cooperations are especially well suited to provide such networks and the appropriate web platforms. Using METORA as an example – a project developed and run by the Federal Association for Information Economy, Telecommunications and New Media e.V. (BITKOM) for the Federal Ministry of Economics and Technology (BMWi) – This paper will discuss how associations and other network organizations can achieve this task and what conditions they have to consider.

A Novel Fuzzy Logic Based Controller to Adjust the Brightness of the Television Screen with Respect to Surrounding Light
One of the major cause of eye strain and other problems caused while watching television is the relative illumination between the screen and its surrounding. This can be overcome by adjusting the brightness of the screen with respect to the surrounding light. A controller based on fuzzy logic is proposed in this paper. The fuzzy controller takes in the intensity of light surrounding the screen and the present brightness of the screen as input. The output of the fuzzy controller is the grid voltage corresponding to the required brightness. This voltage is given to CRT and brightness is controller dynamically. For the given test system data, different de-fuzzifier methods have been implemented and the results are compared. In order to validate the effectiveness of the proposed approach, a fuzzy controller has been designed by obtaining a test data from a real time system. The simulations are performed in MATLAB and are verified with standard system data. The proposed approach can be implemented for real time applications.
An Intelligent Cascaded Fuzzy Logic Based Controller for Controlling the Room Temperature in Hydronic Heating System
Heating systems are a necessity for regions which brace extreme cold weather throughout the year. To maintain a comfortable temperature inside a given place, heating systems making use of- Hydronic boilers- are used. The principle of a single pipe system serves as a base for their working. It is mandatory for these heating systems to control the room temperature, thus maintaining a warm environment. In this paper, the concept of regulation of the room temperature over a wide range is established by using an Adaptive Fuzzy Controller (AFC). This fuzzy controller automatically detects the changes in the outside temperatures and correspondingly maintains the inside temperature to a palatial value. Two separate AFC's are put to use to carry out this function: one to determine the quantity of heat needed to reach the prospective temperature required and to set the desired temperature; the other to control the position of the valve, which is directly proportional to the error between the present room temperature and the user desired temperature. The fuzzy logic controls the position of the valve as per the requirement of the heat. The amount by which the valve opens or closes is controlled by 5 knob positions, which vary from minimum to maximum, thereby regulating the amount of heat flowing through the valve. For the given test system data, different de-fuzzifier methods have been implemented and the results are compared. In order to validate the effectiveness of the proposed approach, a fuzzy controller has been designed by obtaining a test data from a real time system. The simulations are performed in MATLAB and are verified with standard system data. The proposed approach can be implemented for real time applications.
Laplace Transformation on Ordered Linear Space of Generalized Functions
Aim. We have introduced the notion of order to multinormed spaces and countable union spaces and their duals. The topology of bounded convergence is assigned to the dual spaces. The aim of this paper is to develop the theory of ordered topological linear spaces La,b, L(w, z), the dual spaces of ordered multinormed spaces La,b, ordered countable union spaces L(w, z), with the topology of bounded convergence assigned to the dual spaces. We apply Laplace transformation to the ordered linear space of Laplace transformable generalized functions. We ultimately aim at finding solutions to nonhomogeneous nth order linear differential equations with constant coefficients in terms of generalized functions and comparing different solutions evolved out of different initial conditions. Method. The above aim is achieved by • Defining the spaces La,b, L(w, z). • Assigning an order relation on these spaces by identifying a positive cone on them and studying the properties of the cone. • Defining an order relation on the dual spaces La,b, L(w, z) of La,b, L(w, z) and assigning a topology to these dual spaces which makes the order dual and the topological dual the same. • Defining the adjoint of a continuous map on these spaces and studying its behaviour when the topology of bounded convergence is assigned to the dual spaces. • Applying the two-sided Laplace Transformation on the ordered linear space of generalized functions W and studying some properties of the transformation which are used in solving differential equations. Result. The above techniques are applied to solve non-homogeneous n-th order linear differential equations with constant coefficients in terms of generalized functions and to compare different solutions of the differential equation.
Real Time Force Sensing Mat for Human Gait Analysis
This paper presents a real time force sensing instrument that is designed for human gait analysis purposes. This instrument mainly consists of three main elements: the force sensing mat, signal conditioning and switching circuit and data acquisition device. In order to control and to process the incoming signals from the force sensing mat, Force-Logger and Force-Reloader program are developed using Labview 8.0. This paper describes the architecture of the force sensing mat, signal conditioning and switching circuit and the real time streaming of the incoming data from the force sensing mat.
Decision Support System for Flood Crisis Management using Artificial Neural Network
This paper presents an alternate approach that uses artificial neural network to simulate the flood level dynamics in a river basin. The algorithm was developed in a decision support system environment in order to enable users to process the data. The decision support system is found to be useful due to its interactive nature, flexibility in approach and evolving graphical feature and can be adopted for any similar situation to predict the flood level. The main data processing includes the gauging station selection, input generation, lead-time selection/generation, and length of prediction. This program enables users to process the flood level data, to train/test the model using various inputs and to visualize results. The program code consists of a set of files, which can as well be modified to match other purposes. This program may also serve as a tool for real-time flood monitoring and process control. The running results indicate that the decision support system applied to the flood level seems to have reached encouraging results for the river basin under examination. The comparison of the model predictions with the observed data was satisfactory, where the model is able to forecast the flood level up to 5 hours in advance with reasonable prediction accuracy. Finally, this program may also serve as a tool for real-time flood monitoring and process control.
Novel Rao-Blackwellized Particle Filter for Mobile Robot SLAM Using Monocular Vision
This paper presents the novel Rao-Blackwellised particle filter (RBPF) for mobile robot simultaneous localization and mapping (SLAM) using monocular vision. The particle filter is combined with unscented Kalman filter (UKF) to extending the path posterior by sampling new poses that integrate the current observation which drastically reduces the uncertainty about the robot pose. The landmark position estimation and update is also implemented through UKF. Furthermore, the number of resampling steps is determined adaptively, which seriously reduces the particle depletion problem, and introducing the evolution strategies (ES) for avoiding particle impoverishment. The 3D natural point landmarks are structured with matching Scale Invariant Feature Transform (SIFT) feature pairs. The matching for multi-dimension SIFT features is implemented with a KD-Tree in the time cost of O(log2 N). Experiment results on real robot in our indoor environment show the advantages of our methods over previous approaches.
Counterpropagation Neural Network for Solving Power Flow Problem
Power flow (PF) study, which is performed to determine the power system static states (voltage magnitudes and voltage angles) at each bus to find the steady state operating condition of a system, is very important and is the most frequently carried out study by power utilities for power system planning, operation and control. In this paper, a counterpropagation neural network (CPNN) is proposed to solve power flow problem under different loading/contingency conditions for computing bus voltage magnitudes and angles of the power system. The counterpropagation network uses a different mapping strategy namely counterpropagation and provides a practical approach for implementing a pattern mapping task, since learning is fast in this network. The composition of the input variables for the proposed neural network has been selected to emulate the solution process of a conventional power flow program. The effectiveness of the proposed CPNN based approach for solving power flow is demonstrated by computation of bus voltage magnitudes and voltage angles for different loading conditions and single line-outage contingencies in IEEE 14-bus system.
Target Detection with Improved Image Texture Feature Coding Method and Support Vector Machine

An image texture analysis and target recognition approach of using an improved image texture feature coding method (TFCM) and Support Vector Machine (SVM) for target detection is presented. With our proposed target detection framework, targets of interest can be detected accurately. Cascade-Sliding-Window technique was also developed for automated target localization. Application to mammogram showed that over 88% of normal mammograms and 80% of abnormal mammograms can be correctly identified. The approach was also successfully applied to Synthetic Aperture Radar (SAR) and Ground Penetrating Radar (GPR) images for target detection.

Decision Algorithm for Smart Airbag Deployment Safety Issues
Airbag deployment has been known to be responsible for huge death, incidental injuries and broken bones due to low crash severity and wrong deployment decisions. Therefore, the authorities and industries have been looking for more innovative and intelligent products to be realized for future enhancements in the vehicle safety systems (VSSs). Although the VSSs technologies have advanced considerably, they still face challenges such as how to avoid unnecessary and untimely airbag deployments that can be hazardous and fatal. Currently, most of the existing airbag systems deploy without regard to occupant size and position. As such, this paper will focus on the occupant and crash sensing performances due to frontal collisions for the new breed of so called smart airbag systems. It intends to provide a thorough discussion relating to the occupancy detection, occupant size classification, occupant off-position detection to determine safe distance zone for airbag deployment, crash-severity analysis and airbag decision algorithms via a computer modeling. The proposed system model consists of three main modules namely, occupant sensing, crash severity analysis and decision fusion. The occupant sensing system module utilizes the weight sensor to determine occupancy, classify the occupant size, and determine occupant off-position condition to compute safe distance for airbag deployment. The crash severity analysis module is used to generate relevant information pertinent to airbag deployment decision. Outputs from these two modules are fused to the decision module for correct and efficient airbag deployment action. Computer modeling work is carried out using Simulink, Stateflow, SimMechanics and Virtual Reality toolboxes.
Vibration Suppression of Timoshenko Beams with Embedded Piezoelectrics Using POF
This paper deals with the design of a periodic output feedback controller for a flexible beam structure modeled with Timoshenko beam theory, Finite Element Method, State space methods and embedded piezoelectrics concept. The first 3 modes are considered in modeling the beam. The main objective of this work is to control the vibrations of the beam when subjected to an external force. Shear piezoelectric sensors and actuators are embedded into the top and bottom layers of a flexible aluminum beam structure, thus making it intelligent and self-adaptive. The composite beam is divided into 5 finite elements and the control actuator is placed at finite element position 1, whereas the sensor is varied from position 2 to 5, i.e., from the nearby fixed end to the free end. 4 state space SISO models are thus developed. Periodic Output Feedback (POF) Controllers are designed for the 4 SISO models of the same plant to control the flexural vibrations. The effect of placing the sensor at different locations on the beam is observed and the performance of the controller is evaluated for vibration control. Conclusions are finally drawn.
Admission Control Approaches in the IMS Presence Service
In this research, we propose a weighted class based queuing (WCBQ) mechanism to provide class differentiation and to reduce the load for the IMS (IP Multimedia Subsystem) presence server (PS). The tasks of admission controller for the PS are demonstrated. Analysis and simulation models are developed to quantify the performance of WCBQ scheme. An optimized dropping time frame has been developed based on which some of the preexisting messages are dropped from the PS-buffer. Cost functions are developed and simulation comparison has been performed with FCFS (First Come First Served) scheme. The results show that the PS benefits significantly from the proposed queuing and dropping algorithm (WCBQ) during heavy traffic.
New Wavelet-Based Superresolution Algorithm for Speckle Reduction in SAR Images

This paper describes a novel projection algorithm, the Projection Onto Span Algorithm (POSA) for wavelet-based superresolution and removing speckle (in wavelet domain) of unknown variance from Synthetic Aperture Radar (SAR) images. Although the POSA is good as a new superresolution algorithm for image enhancement, image metrology and biometric identification, here one will use it like a tool of despeckling, being the first time that an algorithm of super-resolution is used for despeckling of SAR images. Specifically, the speckled SAR image is decomposed into wavelet subbands; POSA is applied to the high subbands, and reconstruct a SAR image from the modified detail coefficients. Experimental results demonstrate that the new method compares favorably to several other despeckling methods on test SAR images.

A Specification-Based Approach for Retrieval of Reusable Business Component for Software Reuse
Software reuse can be considered as the most realistic and promising way to improve software engineering productivity and quality. Automated assistance for software reuse involves the representation, classification, retrieval and adaptation of components. The representation and retrieval of components are important to software reuse in Component-Based on Software Development (CBSD). However, current industrial component models mainly focus on the implement techniques and ignore the semantic information about component, so it is difficult to retrieve the components that satisfy user-s requirements. This paper presents a method of business component retrieval based on specification matching to solve the software reuse of enterprise information system. First, a business component model oriented reuse is proposed. In our model, the business data type is represented as sign data type based on XML, which can express the variable business data type that can describe the variety of business operations. Based on this model, we propose specification match relationships in two levels: business operation level and business component level. In business operation level, we use input business data types, output business data types and the taxonomy of business operations evaluate the similarity between business operations. In the business component level, we propose five specification matches between business components. To retrieval reusable business components, we propose the measure of similarity degrees to calculate the similarities between business components. Finally, a business component retrieval command like SQL is proposed to help user to retrieve approximate business components from component repository.
Turbine Follower Control Strategy Design Based on Developed FFPP Model
In this paper a comprehensive model of a fossil fueled power plant (FFPP) is developed in order to evaluate the performance of a newly designed turbine follower controller. Considering the drawbacks of previous works, an overall model is developed to minimize the error between each subsystem model output and the experimental data obtained at the actual power plant. The developed model is organized in two main subsystems namely; Boiler and Turbine. Considering each FFPP subsystem characteristics, different modeling approaches are developed. For economizer, evaporator, superheater and reheater, first order models are determined based on principles of mass and energy conservation. Simulations verify the accuracy of the developed models. Due to the nonlinear characteristics of attemperator, a new model, based on a genetic-fuzzy systems utilizing Pittsburgh approach is developed showing a promising performance vis-à-vis those derived with other methods like ANFIS. The optimization constraints are handled utilizing penalty functions. The effect of increasing the number of rules and membership functions on the performance of the proposed model is also studied and evaluated. The turbine model is developed based on the equation of adiabatic expansion. Parameters of all evaluated models are tuned by means of evolutionary algorithms. Based on the developed model a fuzzy PI controller is developed. It is then successfully implemented in the turbine follower control strategy of the plant. In this control strategy instead of keeping control parameters constant, they are adjusted on-line with regard to the error and the error rate. It is shown that the response of the system improves significantly. It is also shown that fuel consumption decreases considerably.
Development of a Semantic Wiki-based Feature Library for the Extraction of Manufacturing Feature and Manufacturing Information
A manufacturing feature can be defined simply as a geometric shape and its manufacturing information to create the shape. In a feature-based process planning system, feature library that consists of pre-defined manufacturing features and the manufacturing information to create the shape of the features, plays an important role in the extraction of manufacturing features with their proper manufacturing information. However, to manage the manufacturing information flexibly, it is important to build a feature library that can be easily modified. In this paper, the implementation of Semantic Wiki for the development of the feature library is proposed.
A Reusability Evaluation Model for OO-Based Software Components
The requirement to improve software productivity has promoted the research on software metric technology. There are metrics for identifying the quality of reusable components but the function that makes use of these metrics to find reusability of software components is still not clear. These metrics if identified in the design phase or even in the coding phase can help us to reduce the rework by improving quality of reuse of the component and hence improve the productivity due to probabilistic increase in the reuse level. CK metric suit is most widely used metrics for the objectoriented (OO) software; we critically analyzed the CK metrics, tried to remove the inconsistencies and devised the framework of metrics to obtain the structural analysis of OO-based software components. Neural network can learn new relationships with new input data and can be used to refine fuzzy rules to create fuzzy adaptive system. Hence, Neuro-fuzzy inference engine can be used to evaluate the reusability of OO-based component using its structural attributes as inputs. In this paper, an algorithm has been proposed in which the inputs can be given to Neuro-fuzzy system in form of tuned WMC, DIT, NOC, CBO , LCOM values of the OO software component and output can be obtained in terms of reusability. The developed reusability model has produced high precision results as expected by the human experts.
A New Approach for Fingerprint Classification based on Minutiae Distribution
The paper describes a new approach for fingerprint classification, based on the distribution of local features (minute details or minutiae) of the fingerprints. The main advantage is that fingerprint classification provides an indexing scheme to facilitate efficient matching in a large fingerprint database. A set of rules based on heuristic approach has been proposed. The area around the core point is treated as the area of interest for extracting the minutiae features as there are substantial variations around the core point as compared to the areas away from the core point. The core point in a fingerprint has been located at a point where there is maximum curvature. The experimental results report an overall average accuracy of 86.57 % in fingerprint classification.
A Preliminary Study on the Suitability of Data Driven Approach for Continuous Water Level Modeling
Reliable water level forecasts are particularly important for warning against dangerous flood and inundation. The current study aims at investigating the suitability of the adaptive network based fuzzy inference system for continuous water level modeling. A hybrid learning algorithm, which combines the least square method and the back propagation algorithm, is used to identify the parameters of the network. For this study, water levels data are available for a hydrological year of 2002 with a sampling interval of 1-hour. The number of antecedent water level that should be included in the input variables is determined by two statistical methods, i.e. autocorrelation function and partial autocorrelation function between the variables. Forecasting was done for 1-hour until 12-hour ahead in order to compare the models generalization at higher horizons. The results demonstrate that the adaptive networkbased fuzzy inference system model can be applied successfully and provide high accuracy and reliability for river water level estimation. In general, the adaptive network-based fuzzy inference system provides accurate and reliable water level prediction for 1-hour ahead where the MAPE=1.15% and correlation=0.98 was achieved. Up to 12-hour ahead prediction, the model still shows relatively good performance where the error of prediction resulted was less than 9.65%. The information gathered from the preliminary results provide a useful guidance or reference for flood early warning system design in which the magnitude and the timing of a potential extreme flood are indicated.
Study of Compaction in Hot-Mix Asphalt Using Computer Simulations
During the process of compaction in Hot-Mix Asphalt (HMA) mixtures, the distance between aggregate particles decreases as they come together and eliminate air-voids. By measuring the inter-particle distances in a cut-section of a HMA sample the degree of compaction can be estimated. For this, a calibration curve is generated by computer simulation technique when the gradation and asphalt content of the HMA mixture are known. A two-dimensional cross section of HMA specimen was simulated using the mixture design information (gradation, asphalt content and air-void content). Nearest neighbor distance methods such as Delaunay triangulation were used to study the changes in inter-particle distance and area distribution during the process of compaction in HMA. Such computer simulations would enable making several hundreds of repetitions in a short period of time without the necessity to compact and analyze laboratory specimens in order to obtain good statistics on the parameters defined. The distributions for the statistical parameters based on computer simulations showed similar trends as those of laboratory specimens.
Modified Fuzzy ARTMAP and Supervised Fuzzy ART: Comparative Study with Multispectral Classification

In this article a modification of the algorithm of the fuzzy ART network, aiming at returning it supervised is carried out. It consists of the search for the comparison, training and vigilance parameters giving the minimum quadratic distances between the output of the training base and those obtained by the network. The same process is applied for the determination of the parameters of the fuzzy ARTMAP giving the most powerful network. The modification consist in making learn the fuzzy ARTMAP a base of examples not only once as it is of use, but as many time as its architecture is in evolution or than the objective error is not reached . In this way, we don-t worry about the values to impose on the eight (08) parameters of the network. To evaluate each one of these three networks modified, a comparison of their performances is carried out. As application we carried out a classification of the image of Algiers-s bay taken by SPOT XS. We use as criterion of evaluation the training duration, the mean square error (MSE) in step control and the rate of good classification per class. The results of this study presented as curves, tables and images show that modified fuzzy ARTMAP presents the best compromise quality/computing time.

ORank: An Ontology Based System for Ranking Documents
Increasing growth of information volume in the internet causes an increasing need to develop new (semi)automatic methods for retrieval of documents and ranking them according to their relevance to the user query. In this paper, after a brief review on ranking models, a new ontology based approach for ranking HTML documents is proposed and evaluated in various circumstances. Our approach is a combination of conceptual, statistical and linguistic methods. This combination reserves the precision of ranking without loosing the speed. Our approach exploits natural language processing techniques for extracting phrases and stemming words. Then an ontology based conceptual method will be used to annotate documents and expand the query. To expand a query the spread activation algorithm is improved so that the expansion can be done in various aspects. The annotated documents and the expanded query will be processed to compute the relevance degree exploiting statistical methods. The outstanding features of our approach are (1) combining conceptual, statistical and linguistic features of documents, (2) expanding the query with its related concepts before comparing to documents, (3) extracting and using both words and phrases to compute relevance degree, (4) improving the spread activation algorithm to do the expansion based on weighted combination of different conceptual relationships and (5) allowing variable document vector dimensions. A ranking system called ORank is developed to implement and test the proposed model. The test results will be included at the end of the paper.
A Pattern Language for Software Debugging
In spite of all advancement in software testing, debugging remains a labor-intensive, manual, time consuming, and error prone process. A candidate solution to enhance debugging process is to fuse it with testing process. To achieve this integration, a possible solution may be categorizing common software tests and errors followed by the effort on fixing the errors through general solutions for each test/error pair. Our approach to address this issue is based on Christopher Alexander-s pattern and pattern language concepts. The patterns in this language are grouped into three major sections and connect the three concepts of test, error, and debug. These patterns and their hierarchical relationship shape a pattern language that introduces a solution to solve software errors in a known testing context. Finally, we will introduce our developed framework ADE as a sample implementation to support a pattern of proposed language, which aims to automate the whole process of evolving software design via evolutionary methods.
Improving Image Segmentation Performance via Edge Preserving Regularization
This paper presents an improved image segmentation model with edge preserving regularization based on the piecewise-smooth Mumford-Shah functional. A level set formulation is considered for the Mumford-Shah functional minimization in segmentation, and the corresponding partial difference equations are solved by the backward Euler discretization. Aiming at encouraging edge preserving regularization, a new edge indicator function is introduced at level set frame. In which all the grid points which is used to locate the level set curve are considered to avoid blurring the edges and a nonlinear smooth constraint function as regularization term is applied to smooth the image in the isophote direction instead of the gradient direction. In implementation, some strategies such as a new scheme for extension of u+ and u- computation of the grid points and speedup of the convergence are studied to improve the efficacy of the algorithm. The resulting algorithm has been implemented and compared with the previous methods, and has been proved efficiently by several cases.
Analyzing and Formulation of Product Lead Time
Product Lead Time (PLT) is the period of time from receiving a customer's order to delivering the final product. PLT is an indicator of the manufacturing controllability, efficiency and performance. Due to the explosion in the rate of technological innovations and the rapid changes in the nature of manufacturing processes, manufacturing firms can bring the new products to market quicker only if they can reduce their PLT and speed up the rate at which they can design, plan, control, and manufacture. Although there is a substantial body of research on manufacturing relating to cost and quality issues, there is no much specific research conducted in relation to the formulation of PLT, despite its significance and importance. This paper analyzes and formulates PLT which can be used as a guideline for achieving the shorter PLT. Further more this paper identifies the causes of delay and factors that contributes to the increased product lead-time.
A New Automatic System of Cell Colony Counting
The counting process of cell colonies is always a long and laborious process that is dependent on the judgment and ability of the operator. The judgment of the operator in counting can vary in relation to fatigue. Moreover, since this activity is time consuming it can limit the usable number of dishes for each experiment. For these purposes, it is necessary that an automatic system of cell colony counting is used. This article introduces a new automatic system of counting based on the elaboration of the digital images of cellular colonies grown on petri dishes. This system is mainly based on the algorithms of region-growing for the recognition of the regions of interest (ROI) in the image and a Sanger neural net for the characterization of such regions. The better final classification is supplied from a Feed-Forward Neural Net (FF-NN) and confronted with the K-Nearest Neighbour (K-NN) and a Linear Discriminative Function (LDF). The preliminary results are shown.
Random Projections for Dimensionality Reduction in ICA
In this paper we present a technique to speed up ICA based on the idea of reducing the dimensionality of the data set preserving the quality of the results. In particular we refer to FastICA algorithm which uses the Kurtosis as statistical property to be maximized. By performing a particular Johnson-Lindenstrauss like projection of the data set, we find the minimum dimensionality reduction rate ¤ü, defined as the ratio between the size k of the reduced space and the original one d, which guarantees a narrow confidence interval of such estimator with high confidence level. The derived dimensionality reduction rate depends on a system control parameter β easily computed a priori on the basis of the observations only. Extensive simulations have been done on different sets of real world signals. They show that actually the dimensionality reduction is very high, it preserves the quality of the decomposition and impressively speeds up FastICA. On the other hand, a set of signals, on which the estimated reduction rate is greater than 1, exhibits bad decomposition results if reduced, thus validating the reliability of the parameter β. We are confident that our method will lead to a better approach to real time applications.
Decomposition Method for Neural Multiclass Classification Problem
In this article we are going to discuss the improvement of the multi classes- classification problem using multi layer Perceptron. The considered approach consists in breaking down the n-class problem into two-classes- subproblems. The training of each two-class subproblem is made independently; as for the phase of test, we are going to confront a vector that we want to classify to all two classes- models, the elected class will be the strongest one that won-t lose any competition with the other classes. Rates of recognition gotten with the multi class-s approach by two-class-s decomposition are clearly better that those gotten by the simple multi class-s approach.
Automatic Musical Genre Classification Using Divergence and Average Information Measures
Recently many research has been conducted to retrieve pertinent parameters and adequate models for automatic music genre classification. In this paper, two measures based upon information theory concepts are investigated for mapping the features space to decision space. A Gaussian Mixture Model (GMM) is used as a baseline and reference system. Various strategies are proposed for training and testing sessions with matched or mismatched conditions, long training and long testing, long training and short testing. For all experiments, the file sections used for testing are never been used during training. With matched conditions all examined measures yield the best and similar scores (almost 100%). With mismatched conditions, the proposed measures yield better scores than the GMM baseline system, especially for the short testing case. It is also observed that the average discrimination information measure is most appropriate for music category classifications and on the other hand the divergence measure is more suitable for music subcategory classifications.
Neuro-Fuzzy Algorithm for a Biped Robotic System
This paper summaries basic principles and concepts of intelligent controls, implemented in humanoid robotics as well as recent algorithms being devised for advanced control of humanoid robots. Secondly, this paper presents a new approach neuro-fuzzy system. We have included some simulating results from our computational intelligence technique that will be applied to our humanoid robot. Subsequently, we determine a relationship between joint trajectories and located forces on robot-s foot through a proposed neuro-fuzzy technique.
Automation of Heat Exchanger using Neural Network
In this paper the development of a heat exchanger as a pilot plant for educational purpose is discussed and the use of neural network for controlling the process is being presented. The aim of the study is to highlight the need of a specific Pseudo Random Binary Sequence (PRBS) to excite a process under control. As the neural network is a data driven technique, the method for data generation plays an important role. In light of this a careful experimentation procedure for data generation was crucial task. Heat exchange is a complex process, which has a capacity and a time lag as process elements. The proposed system is a typical pipe-in- pipe type heat exchanger. The complexity of the system demands careful selection, proper installation and commissioning. The temperature, flow, and pressure sensors play a vital role in the control performance. The final control element used is a pneumatically operated control valve. While carrying out the experimentation on heat exchanger a welldrafted procedure is followed giving utmost attention towards safety of the system. The results obtained are encouraging and revealing the fact that if the process details are known completely as far as process parameters are concerned and utilities are well stabilized then feedback systems are suitable, whereas neural network control paradigm is useful for the processes with nonlinearity and less knowledge about process. The implementation of NN control reinforces the concepts of process control and NN control paradigm. The result also underlined the importance of excitation signal typically for that process. Data acquisition, processing, and presentation in a typical format are the most important parameters while validating the results.
LSGENSYS - An Integrated System for Pattern Recognition and Summarisation
This paper presents a new system developed in Java® for pattern recognition and pattern summarisation in multi-band (RGB) satellite images. The system design is described in some detail. Results of testing the system to analyse and summarise patterns in SPOT MS images and LANDSAT images are also discussed.
Intelligent Agent Approach to the Control of Critical Infrastructure Networks
In this paper we propose an intelligent agent approach to control the electric power grid at a smaller granularity in order to give it self-healing capabilities. We develop a method using the influence model to transform transmission substations into information processing, analyzing and decision making (intelligent behavior) units. We also develop a wireless communication method to deliver real-time uncorrupted information to an intelligent controller in a power system environment. A combined networking and information theoretic approach is adopted in meeting both the delay and error probability requirements. We use a mobile agent approach in optimizing the achievable information rate vector and in the distribution of rates to users (sensors). We developed the concept and the quantitative tools require in the creation of cooperating semiautonomous subsystems which puts the electric grid on the path towards intelligent and self-healing system.
Source Direction Detection based on Stationary Electronic Nose System
Electronic nose (array of chemical sensors) are widely used in food industry and pollution control. Also it could be used to locate or detect the direction of the source of emission odors. Usually this task is performed by electronic nose (ENose) cooperated with mobile vehicles, but when a source is instantaneous or surrounding is hard for vehicles to reach, problem occurs. Thus a method for stationary ENose to detect the direction of the source and locate the source will be required. A novel method which uses the ratio between the responses of different sensors as a discriminant to determine the direction of source in natural wind surroundings is presented in this paper. The result shows that the method is accurate and easily to be implemented. This method could be also used in movably, as an optimized algorithm for robot tracking source location.
A New Self-Adaptive EP Approach for ANN Weights Training
Evolutionary Programming (EP) represents a methodology of Evolutionary Algorithms (EA) in which mutation is considered as a main reproduction operator. This paper presents a novel EP approach for Artificial Neural Networks (ANN) learning. The proposed strategy consists of two components: the self-adaptive, which contains phenotype information and the dynamic, which is described by genotype. Self-adaptation is achieved by the addition of a value, called the network weight, which depends on a total number of hidden layers and an average number of neurons in hidden layers. The dynamic component changes its value depending on the fitness of a chromosome, exposed to mutation. Thus, the mutation step size is controlled by two components, encapsulated in the algorithm, which adjust it according to the characteristics of a predefined ANN architecture and the fitness of a particular chromosome. The comparative analysis of the proposed approach and the classical EP (Gaussian mutation) showed, that that the significant acceleration of the evolution process is achieved by using both phenotype and genotype information in the mutation strategy.
Distortion Estimation in Digital Image Watermarking using Genetic Programming
This paper introduces a technique of distortion estimation in image watermarking using Genetic Programming (GP). The distortion is estimated by considering the problem of obtaining a distorted watermarked signal from the original watermarked signal as a function regression problem. This function regression problem is solved using GP, where the original watermarked signal is considered as an independent variable. GP-based distortion estimation scheme is checked for Gaussian attack and Jpeg compression attack. We have used Gaussian attacks of different strengths by changing the standard deviation. JPEG compression attack is also varied by adding various distortions. Experimental results demonstrate that the proposed technique is able to detect the watermark even in the case of strong distortions and is more robust against attacks.
Universal Method for Timetable Construction based on Evolutionary Approach
Timetabling problems are often hard and timeconsuming to solve. Most of the methods of solving them concern only one problem instance or class. This paper describes a universal method for solving large, highly constrained timetabling problems from different domains. The solution is based on evolutionary algorithm-s framework and operates on two levels – first-level evolutionary algorithm tries to find a solution basing on given set of operating parameters, second-level algorithm is used to establish those parameters. Tabu search is employed to speed up the solution finding process on first level. The method has been used to solve three different timetabling problems with promising results.
Designing a Novel General Sorting Network Constructor Using Artificial Evolution
A method is presented for the construction of arbitrary even-input sorting networks exhibiting better properties than the networks created using a conventional technique of the same type. The method was discovered by means of a genetic algorithm combined with an application-specific development. Similarly to human inventions in the area of theoretical computer science, the evolved invention was analyzed: its generality was proven and area and time complexities were determined.
Representing Shared Join Points with State Charts: A High Level Design Approach
Aspect Oriented Programming promises many advantages at programming level by incorporating the cross cutting concerns into separate units, called aspects. Join Points are distinguishing features of Aspect Oriented Programming as they define the points where core requirements and crosscutting concerns are (inter)connected. Currently, there is a problem of multiple aspects- composition at the same join point, which introduces the issues like ordering and controlling of these superimposed aspects. Dynamic strategies are required to handle these issues as early as possible. State chart is an effective modeling tool to capture dynamic behavior at high level design. This paper provides methodology to formulate the strategies for multiple aspect composition at high level, which helps to better implement these strategies at coding level. It also highlights the need of designing shared join point at high level, by providing the solutions of these issues using state chart diagrams in UML 2.0. High level design representation of shared join points also helps to implement the designed strategy in systematic way.
Social, Group and Individual Mind extracted from Rule Bases of Multiple Agents
This paper shows possibility of extraction Social, Group and Individual Mind from Multiple Agents Rule Bases. Types those Rule bases are selected as two fuzzy systems, namely Mambdani and Takagi-Sugeno fuzzy system. Their rule bases are describing (modeling) agent behavior. Modifying of agent behavior in the time varying environment will be provided by learning fuzzyneural networks and optimization of their parameters with using genetic algorithms in development system FUZNET. Finally, extraction Social, Group and Individual Mind from Multiple Agents Rule Bases are provided by Cognitive analysis and Matching criterion.
A Model Predictive Control and Time Series Forecasting Framework for Supply Chain Management
Model Predictive Control has been previously applied to supply chain problems with promising results; however hitherto proposed systems possessed no information on future demand. A forecasting methodology will surely promote the efficiency of control actions by providing insight on the future. A complete supply chain management framework that is based on Model Predictive Control (MPC) and Time Series Forecasting will be presented in this paper. The proposed framework will be tested on industrial data in order to assess the efficiency of the method and the impact of forecast accuracy on overall control performance of the supply chain. To this end, forecasting methodologies with different characteristics will be implemented on test data to generate forecasts that will serve as input to the Model Predictive Control module.
Design and Layout of a X-Band MMIC Power Amplifier in a Phemt Technology
The design of Class A and Class AB 2-stage X band Power Amplifier is described in this report. This power amplifier is part of a transceiver used in radar for monitoring iron characteristics in a blast furnace. The circuit was designed using foundry WIN Semiconductors. The specification requires 15dB gain in the linear region, VSWR nearly 1 at input as well as at the output, an output power of 10 dBm and good stable performance in the band 10.9-12.2 GHz. The design was implemented by using inter-stage configuration, the Class A amplifier was chosen for driver stage i.e. the first amplifier focusing on the gain and the output amplifier conducted at Class AB with more emphasis on output power.
Evolutionary Training of Hybrid Systems of Recurrent Neural Networks and Hidden Markov Models
We present a hybrid architecture of recurrent neural networks (RNNs) inspired by hidden Markov models (HMMs). We train the hybrid architecture using genetic algorithms to learn and represent dynamical systems. We train the hybrid architecture on a set of deterministic finite-state automata strings and observe the generalization performance of the hybrid architecture when presented with a new set of strings which were not present in the training data set. In this way, we show that the hybrid system of HMM and RNN can learn and represent deterministic finite-state automata. We ran experiments with different sets of population sizes in the genetic algorithm; we also ran experiments to find out which weight initializations were best for training the hybrid architecture. The results show that the hybrid architecture of recurrent neural networks inspired by hidden Markov models can train and represent dynamical systems. The best training and generalization performance is achieved when the hybrid architecture is initialized with random real weight values of range -15 to 15.
Online Collaborative Learning System Using Speech Technology
A Web-based learning tool, the Learn IN Context (LINC) system, designed and being used in some institution-s courses in mixed-mode learning, is presented in this paper. This mode combines face-to-face and distance approaches to education. LINC can achieve both collaborative and competitive learning. In order to provide both learners and tutors with a more natural way to interact with e-learning applications, a conversational interface has been included in LINC. Hence, the components and essential features of LINC+, the voice enhanced version of LINC, are described. We report evaluation experiments of LINC/LINC+ in a real use context of a computer programming course taught at the Université de Moncton (Canada). The findings show that when the learning material is delivered in the form of a collaborative and voice-enabled presentation, the majority of learners seem to be satisfied with this new media, and confirm that it does not negatively affect their cognitive load.
DIFFER: A Propositionalization approach for Learning from Structured Data
Logic based methods for learning from structured data is limited w.r.t. handling large search spaces, preventing large-sized substructures from being considered by the resulting classifiers. A novel approach to learning from structured data is introduced that employs a structure transformation method, called finger printing, for addressing these limitations. The method, which generates features corresponding to arbitrarily complex substructures, is implemented in a system, called DIFFER. The method is demonstrated to perform comparably to an existing state-of-art method on some benchmark data sets without requiring restrictions on the search space. Furthermore, learning from the union of features generated by finger printing and the previous method outperforms learning from each individual set of features on all benchmark data sets, demonstrating the benefit of developing complementary, rather than competing, methods for structure classification.
Extension and Evaluation of Interface “2D-RIB“ for Impression-Based Retrieval
Recently, lots of researchers are attracted to retrieving multimedia database by using some impression words and their values. Ikezoe-s research is one of the representatives and uses eight pairs of opposite impression words. We had modified its retrieval interface and proposed '2D-RIB'. In '2D-RIB', after a retrieval person selects a single basic music, the system visually shows some other music around the basic one along relative position. He/she can select one of them fitting to his/her intention, as a retrieval result. The purpose of this paper is to improve his/her satisfaction level to the retrieval result in 2D-RIB. One of our extensions is to define and introduce the following two measures: 'melody goodness' and 'general acceptance'. We implement them in different five combinations. According to an evaluation experiment, both of these two measures can contribute to the improvement. Another extension is three types of customization. We have implemented them and clarified which customization is effective.
Inferring the Dynamics of “Hidden“ Neurons from Electrophysiological Recordings
Statistical analysis of electrophysiological recordings obtained under, e.g. tactile, stimulation frequently suggests participation in the network dynamics of experimentally unobserved “hidden" neurons. Such interneurons making synapses to experimentally recorded neurons may strongly alter their dynamical responses to the stimuli. We propose a mathematical method that formalizes this possibility and provides an algorithm for inferring on the presence and dynamics of hidden neurons based on fitting of the experimental data to spike trains generated by the network model. The model makes use of Integrate and Fire neurons “chemically" coupled through exponentially decaying synaptic currents. We test the method on simulated data and also provide an example of its application to the experimental recording from the Dorsal Column Nuclei neurons of the rat under tactile stimulation of a hind limb.
Simple Agents Benefit Only from Simple Brains

In order to answer the general question: “What does a simple agent with a limited life-time require for constructing a useful representation of the environment?" we propose a robot platform including the simplest probabilistic sensory and motor layers. Then we use the platform as a test-bed for evaluation of the navigational capabilities of the robot with different “brains". We claim that a protocognitive behavior is not a consequence of highly sophisticated sensory–motor organs but instead emerges through an increment of the internal complexity and reutilization of the minimal sensory information. We show that the most fundamental robot element, the short-time memory, is essential in obstacle avoidance. However, in the simplest conditions of no obstacles the straightforward memoryless robot is usually superior. We also demonstrate how a low level action planning, involving essentially nonlinear dynamics, provides a considerable gain to the robot performance dynamically changing the robot strategy. Still, however, for very short life time the brainless robot is superior. Accordingly we suggest that small organisms (or agents) with short life-time does not require complex brains and even can benefit from simple brain-like (reflex) structures. To some extend this may mean that controlling blocks of modern robots are too complicated comparative to their life-time and mechanical abilities.

Central Pattern Generator Incorporating the Actuator Dynamics for a Hexapod Robot
We proposed the use of a Toda-Rayleigh ring as a central pattern generator (CPG) for controlling hexapodal robots. We show that the ring composed of six Toda-Rayleigh units coupled to the limb actuators reproduces the most common hexapodal gaits. We provide an electrical circuit implementation of the CPG and test our theoretical results obtaining fixed gaits. Then we propose a method of incorporation of the actuator (motor) dynamics in the CPG. With this approach we close the loop CPG – environment – CPG, thus obtaining a decentralized model for the leg control that does not require higher level intervention to the CPG during locomotion in a nonhomogeneous environments. The gaits generated by the novel CPG are not fixed, but adapt to the current robot bahvior.
On-Line Geometrical Identification of Reconfigurable Machine Tool using Virtual Machining
One of the main research directions in CAD/CAM machining area is the reducing of machining time. The feedrate scheduling is one of the advanced techniques that allows keeping constant the uncut chip area and as sequel to keep constant the main cutting force. They are two main ways for feedrate optimization. The first consists in the cutting force monitoring, which presumes to use complex equipment for the force measurement and after this, to set the feedrate regarding the cutting force variation. The second way is to optimize the feedrate by keeping constant the material removal rate regarding the cutting conditions. In this paper there is proposed a new approach using an extended database that replaces the system model. The feedrate scheduling is determined based on the identification of the reconfigurable machine tool, and the feed value determination regarding the uncut chip section area, the contact length between tool and blank and also regarding the geometrical roughness. The first stage consists in the blank and tool monitoring for the determination of actual profiles. The next stage is the determination of programmed tool path that allows obtaining the piece target profile. The graphic representation environment models the tool and blank regions and, after this, the tool model is positioned regarding the blank model according to the programmed tool path. For each of these positions the geometrical roughness value, the uncut chip area and the contact length between tool and blank are calculated. Each of these parameters are compared with the admissible values and according to the result the feed value is established. We can consider that this approach has the following advantages: in case of complex cutting processes the prediction of cutting force is possible; there is considered the real cutting profile which has deviations from the theoretical profile; the blank-tool contact length limitation is possible; it is possible to correct the programmed tool path so that the target profile can be obtained. Applying this method, there are obtained data sets which allow the feedrate scheduling so that the uncut chip area is constant and, as a result, the cutting force is constant, which allows to use more efficiently the machine tool and to obtain the reduction of machining time.
Using Multi-Thread Technology Realize Most Short-Path Parallel Algorithm
The shortest path question is in a graph theory model question, and it is applied in many fields. The most short-path question may divide into two kinds: Single sources most short-path, all apexes to most short-path. This article mainly introduces the problem of all apexes to most short-path, and gives a new parallel algorithm of all apexes to most short-path according to the Dijkstra algorithm. At last this paper realizes the parallel algorithms in the technology of C # multithreading.
The Application of Specialized Memory Manager in Interactive CAD Systems
Interactive CAD systems have to allocate and deallocate memory frequently. Frequent memory allocation and deallocation can play a significant role in degrading application performance. An application may use memory in a very specific way and pay a performance penalty for functionality it does not need. We could counter that by developing specialized memory managers.
Computer Aided Detection on Mammography
A typical definition of the Computer Aided Diagnosis (CAD), found in literature, can be: A diagnosis made by a radiologist using the output of a computerized scheme for automated image analysis as a diagnostic aid. Often it is possible to find the expression Computer Aided Detection (CAD or CADe): this definition emphasizes the intent of CAD to support rather than substitute the human observer in the analysis of radiographic images. In this article we will illustrate the application of CAD systems and the aim of these definitions. Commercially available CAD systems use computerized algorithms for identifying suspicious regions of interest. In this paper are described the general CAD systems as an expert system constituted of the following components: segmentation / detection, feature extraction, and classification / decision making. As example, in this work is shown the realization of a Computer- Aided Detection system that is able to assist the radiologist in identifying types of mammary tumor lesions. Furthermore this prototype of station uses a GRID configuration to work on a large distributed database of digitized mammographic images.
Quantification of Heart Rate Variability: A Measure based on Unique Heart Rates

It is established that the instantaneous heart rate (HR) of healthy humans keeps on changing. Analysis of heart rate variability (HRV) has become a popular non invasive tool for assessing the activities of autonomic nervous system. Depressed HRV has been found in several disorders, like diabetes mellitus (DM) and coronary artery disease, characterised by autonomic nervous dysfunction. A new technique, which searches for pattern repeatability in a time series, is proposed specifically for the analysis of heart rate data. These set of indices, which are termed as pattern repeatability measure and pattern repeatability ratio are compared with approximate entropy and sample entropy. In our analysis, based on the method developed, it is observed that heart rate variability is significantly different for DM patients, particularly for patients with diabetic foot ulcer.

Diagnosing Dangerous Arrhythmia of Patients by Automatic Detecting of QRS Complexes in ECG
In this paper, an automatic detecting algorithm for QRS complex detecting was applied for analyzing ECG recordings and five criteria for dangerous arrhythmia diagnosing are applied for a protocol type of automatic arrhythmia diagnosing system. The automatic detecting algorithm applied in this paper detected the distribution of QRS complexes in ECG recordings and related information, such as heart rate and RR interval. In this investigation, twenty sampled ECG recordings of patients with different pathologic conditions were collected for off-line analysis. A combinative application of four digital filters for bettering ECG signals and promoting detecting rate for QRS complex was proposed as pre-processing. Both of hardware filters and digital filters were applied to eliminate different types of noises mixed with ECG recordings. Then, an automatic detecting algorithm of QRS complex was applied for verifying the distribution of QRS complex. Finally, the quantitative clinic criteria for diagnosing arrhythmia were programmed in a practical application for automatic arrhythmia diagnosing as a post-processor. The results of diagnoses by automatic dangerous arrhythmia diagnosing were compared with the results of off-line diagnoses by experienced clinic physicians. The results of comparison showed the application of automatic dangerous arrhythmia diagnosis performed a matching rate of 95% compared with an experienced physician-s diagnoses.
Placebo Effect on Psychoemotional Status and Clinical Analyses of Armenian Crohn's Disease Patients

The Zung self-depression scale and Beck Anxiety Inventory were used both to study the depression and anxiety levels of Armenian Crohn’s disease patients, and to reveal the relation between emotional status and placebo effect of these patients. On the other hand, the blood cell analyses and gut bacteria investigations were used to assess the placebo effect on ESR, and haemoglobin-s and leukocyte-s levels as well as gut commensal E. coli quantities of these patients. Despite of registered high levels of depression and anxiety, the high placebo effect on psychoemotional status for investigated patients during the investigations was described. On the other hand, no positive effect of placebo on measurements of ESP and hemoglobin-s levels of Crohn’s disease patients was revealed. The importance of use of psychotherapies for optimal outcomes during treatments of Crohn’s disease is discussed.

Modeling HIV/AIDS Prevention by Defense
The functional response of an infective is the relationship between an infected individual-s infection rate and the abundance of the number of susceptibles that one can potentially be infected. In this paper, we consider defensive attitudes for HIV prevention (primary prevention) while at the same time emphasizing on offensive attitudes that reduce infection for those infected (secondary prevention). We look at how defenses can protect an uninfected individual in the case where high risk groups such as commercial sex workers and those who deliberately go out to look for partners. We propose an infection cycle that begins with a search, then an encounter, a proposal and contact. The infection cycle illustrates the various steps an infected individual goes through to successfully infect a susceptible. For heterogeneous transmission of HIV, there will be no infection unless there is contact. The ability to avoid an encounter, detection, proposal and contact constitute defense.
Automatic 3D Reconstruction of Coronary Artery Centerlines from Monoplane X-ray Angiogram Images
We present a new method for the fully automatic 3D reconstruction of the coronary artery centerlines, using two X-ray angiogram projection images from a single rotating monoplane acquisition system. During the first stage, the input images are smoothed using curve evolution techniques. Next, a simple yet efficient multiscale method, based on the information of the Hessian matrix, for the enhancement of the vascular structure is introduced. Hysteresis thresholding using different image quantiles, is used to threshold the arteries. This stage is followed by a thinning procedure to extract the centerlines. The resulting skeleton image is then pruned using morphological and pattern recognition techniques to remove non-vessel like structures. Finally, edge-based stereo correspondence is solved using a parallel evolutionary optimization method based on f symbiosis. The detected 2D centerlines combined with disparity map information allow the reconstruction of the 3D vessel centerlines. The proposed method has been evaluated on patient data sets for evaluation purposes.
Chaotic Properties of Hemodynamic Responsein Functional Near Infrared Spectroscopic Measurement of Brain Activity
Functional near infrared spectroscopy (fNIRS) is a practical non-invasive optical technique to detect characteristic of hemoglobin density dynamics response during functional activation of the cerebral cortex. In this paper, fNIRS measurements were made in the area of motor cortex from C4 position according to international 10-20 system. Three subjects, aged 23 - 30 years, were participated in the experiment. The aim of this paper was to evaluate the effects of different motor activation tasks of the hemoglobin density dynamics of fNIRS signal. The chaotic concept based on deterministic dynamics is an important feature in biological signal analysis. This paper employs the chaotic properties which is a novel method of nonlinear analysis, to analyze and to quantify the chaotic property in the time series of the hemoglobin dynamics of the various motor imagery tasks of fNIRS signal. Usually, hemoglobin density in the human brain cortex is found to change slowly in time. An inevitable noise caused by various factors is to be included in a signal. So, principle component analysis method (PCA) is utilized to remove high frequency component. The phase pace is reconstructed and evaluated the Lyapunov spectrum, and Lyapunov dimensions. From the experimental results, it can be conclude that the signals measured by fNIRS are chaotic.
Functional Near Infrared Spectroscope for Cognition Brain Tasks by Wavelets Analysis and Neural Networks
Brain Computer Interface (BCI) has been recently increased in research. Functional Near Infrared Spectroscope (fNIRs) is one the latest technologies which utilize light in the near-infrared range to determine brain activities. Because near infrared technology allows design of safe, portable, wearable, non-invasive and wireless qualities monitoring systems, fNIRs monitoring of brain hemodynamics can be value in helping to understand brain tasks. In this paper, we present results of fNIRs signal analysis indicating that there exist distinct patterns of hemodynamic responses which recognize brain tasks toward developing a BCI. We applied two different mathematics tools separately, Wavelets analysis for preprocessing as signal filters and feature extractions and Neural networks for cognition brain tasks as a classification module. We also discuss and compare with other methods while our proposals perform better with an average accuracy of 99.9% for classification.
Time and Frequency Domain Analysis of Heart Rate Variability and their Correlations in Diabetes Mellitus
Diabetes mellitus (DM) is frequently characterized by autonomic nervous dysfunction. Analysis of heart rate variability (HRV) has become a popular noninvasive tool for assessing the activities of autonomic nervous system (ANS). In this paper, changes in ANS activity are quantified by means of frequency and time domain analysis of R-R interval variability. Electrocardiograms (ECG) of 16 patients suffering from DM and of 16 healthy volunteers were recorded. Frequency domain analysis of extracted normal to normal interval (NN interval) data indicates significant difference in very low frequency (VLF) power, low frequency (LF) power and high frequency (HF) power, between the DM patients and control group. Time domain measures, standard deviation of NN interval (SDNN), root mean square of successive NN interval differences (RMSSD), successive NN intervals differing more than 50 ms (NN50 Count), percentage value of NN50 count (pNN50), HRV triangular index and triangular interpolation of NN intervals (TINN) also show significant difference between the DM patients and control group.
Modelling the Role of Prophylaxis in Malaria Prevention

Malaria is by far the world-s most persistent tropical parasitic disease and is endemic to tropical areas where the climatic and weather conditions allow continuous breeding of the mosquitoes that spread malaria. A mathematical model for the transmission of malaria with prophylaxis prevention is analyzed. The stability analysis of the equilibria is presented with the aim of finding threshold conditions under which malaria clears or persists in the human population. Our results suggest that eradication of mosquitoes and prophylaxis prevention can significantly reduce the malaria burden on the human population.

On The Analysis of a Compound Neural Network for Detecting Atrio Ventricular Heart Block (AVB) in an ECG Signal
Heart failure is the most common reason of death nowadays, but if the medical help is given directly, the patient-s life may be saved in many cases. Numerous heart diseases can be detected by means of analyzing electrocardiograms (ECG). Artificial Neural Networks (ANN) are computer-based expert systems that have proved to be useful in pattern recognition tasks. ANN can be used in different phases of the decision-making process, from classification to diagnostic procedures. This work concentrates on a review followed by a novel method. The purpose of the review is to assess the evidence of healthcare benefits involving the application of artificial neural networks to the clinical functions of diagnosis, prognosis and survival analysis, in ECG signals. The developed method is based on a compound neural network (CNN), to classify ECGs as normal or carrying an AtrioVentricular heart Block (AVB). This method uses three different feed forward multilayer neural networks. A single output unit encodes the probability of AVB occurrences. A value between 0 and 0.1 is the desired output for a normal ECG; a value between 0.1 and 1 would infer an occurrence of an AVB. The results show that this compound network has a good performance in detecting AVBs, with a sensitivity of 90.7% and a specificity of 86.05%. The accuracy value is 87.9%.
Speech Data Compression using Vector Quantization
Mostly transforms are used for speech data compressions which are lossy algorithms. Such algorithms are tolerable for speech data compression since the loss in quality is not perceived by the human ear. However the vector quantization (VQ) has a potential to give more data compression maintaining the same quality. In this paper we propose speech data compression algorithm using vector quantization technique. We have used VQ algorithms LBG, KPE and FCG. The results table shows computational complexity of these three algorithms. Here we have introduced a new performance parameter Average Fractional Change in Speech Sample (AFCSS). Our FCG algorithm gives far better performance considering mean absolute error, AFCSS and complexity as compared to others.
A Tool for Audio Quality Evaluation Under Hostile Environment
In this paper is to evaluate audio and speech quality with the help of Digital Audio Watermarking Technique under the different types of attacks (signal impairments) like Gaussian Noise, Compression Error and Jittering Effect. Further attacks are considered as Hostile Environment. Audio and Speech Quality Evaluation is an important research topic. The traditional way for speech quality evaluation is using subjective tests. They are reliable, but very expensive, time consuming, and cannot be used in certain applications such as online monitoring. Objective models, based on human perception, were developed to predict the results of subjective tests. The existing objective methods require either the original speech or complicated computation model, which makes some applications of quality evaluation impossible.
Centralized Monitoring and Self-protected against Fiber Fault in FTTH Access Network
This paper presented a new approach for centralized monitoring and self-protected against fiber fault in fiber-to-the-home (FTTH) access network by using Smart Access Network Testing, Analyzing and Database (SANTAD). SANTAD will be installed with optical line terminal (OLT) at central office (CO) for in-service transmission surveillance and fiber fault localization within FTTH with point-to-multipoint (P2MP) configuration downwardly from CO towards customer residential locations based on the graphical user interface (GUI) processing capabilities of MATLAB software. SANTAD is able to detect any fiber fault as well as identify the failure location in the network system. SANTAD enable the status of each optical network unit (ONU) connected line is displayed onto one screen with capability to configure the attenuation and detect the failure simultaneously. The analysis results and information will be delivered to the field engineer for promptly actions, meanwhile the failure line will be diverted to protection line to ensure the traffic flow continuously. This approach has a bright prospect to improve the survivability and reliability as well as increase the efficiency and monitoring capabilities in FTTH.
Improved Modulo 2n +1 Adder Design
Efficient modulo 2n+1 adders are important for several applications including residue number system, digital signal processors and cryptography algorithms. In this paper we present a novel modulo 2n+1 addition algorithm for a recently represented number system. The proposed approach is introduced for the reduction of the power dissipated. In a conventional modulo 2n+1 adder, all operands have (n+1)-bit length. To avoid using (n+1)-bit circuits, the diminished-1 and carry save diminished-1 number systems can be effectively used in applications. In the paper, we also derive two new architectures for designing modulo 2n+1 adder, based on n-bit ripple-carry adder. The first architecture is a faster design whereas the second one uses less hardware. In the proposed method, the special treatment required for zero operands in Diminished-1 number system is removed. In the fastest modulo 2n+1 adders in normal binary system, there are 3-operand adders. This problem is also resolved in this paper. The proposed architectures are compared with some efficient adders based on ripple-carry adder and highspeed adder. It is shown that the hardware overhead and power consumption will be reduced. As well as power reduction, in some cases, power-delay product will be also reduced.
Context Modeling and Context-Aware Service Adaptation for Pervasive Computing Systems

Devices in a pervasive computing system (PCS) are characterized by their context-awareness. It permits them to provide proactively adapted services to the user and applications. To do so, context must be well understood and modeled in an appropriate form which enhance its sharing between devices and provide a high level of abstraction. The most interesting methods for modeling context are those based on ontology however the majority of the proposed methods fail in proposing a generic ontology for context which limit their usability and keep them specific to a particular domain. The adaptation task must be done automatically and without an explicit intervention of the user. Devices of a PCS must acquire some intelligence which permits them to sense the current context and trigger the appropriate service or provide a service in a better suitable form. In this paper we will propose a generic service ontology for context modeling and a context-aware service adaptation based on a service oriented definition of context.

Auction Theory: Bidder-s Perspective in a Public Out-Cry English Auction
This paper provides an overview of auction theory literature. We present a general review on literature of various auctions and focus ourselves specifically on an English auction. We are interested in modelling bidder-s behavior in an English auction environment. And hence, we present an overview of the New Zealand wool auction followed by a model that would describe a bidder-s decision making behavior from the New Zealand wool auction. The mathematical assumptions in an English auction environment are demonstrated from the perspective of the New Zealand wool auction.
A Feasible Path Selection QoS Routing Algorithm with two Constraints in Packet Switched Networks
Over the past several years, there has been a considerable amount of research within the field of Quality of Service (QoS) support for distributed multimedia systems. One of the key issues in providing end-to-end QoS guarantees in packet networks is determining a feasible path that satisfies a number of QoS constraints. The problem of finding a feasible path is NPComplete if number of constraints is more than two and cannot be exactly solved in polynomial time. We proposed Feasible Path Selection Algorithm (FPSA) that addresses issues with pertain to finding a feasible path subject to delay and cost constraints and it offers higher success rate in finding feasible paths.
A Characterized and Optimized Approach for End-to-End Delay Constrained QoS Routing
QoS Routing aims to find paths between senders and receivers satisfying the QoS requirements of the application which efficiently using the network resources and underlying routing algorithm to be able to find low-cost paths that satisfy given QoS constraints. The problem of finding least-cost routing is known to be NP hard or complete and some algorithms have been proposed to find a near optimal solution. But these heuristics or algorithms either impose relationships among the link metrics to reduce the complexity of the problem which may limit the general applicability of the heuristic, or are too costly in terms of execution time to be applicable to large networks. In this paper, we analyzed two algorithms namely Characterized Delay Constrained Routing (CDCR) and Optimized Delay Constrained Routing (ODCR). The CDCR algorithm dealt an approach for delay constrained routing that captures the trade-off between cost minimization and risk level regarding the delay constraint. The ODCR which uses an adaptive path weight function together with an additional constraint imposed on the path cost, to restrict search space and hence ODCR finds near optimal solution in much quicker time.
Complexity of Mathematical Expressions in Adaptive Multimodal Multimedia System Ensuring Access to Mathematics for Visually Impaired Users
Our adaptive multimodal system aims at correctly presenting a mathematical expression to visually impaired users. Given an interaction context (i.e. combination of user, environment and system resources) as well as the complexity of the expression itself and the user-s preferences, the suitability scores of different presentation format are calculated. Unlike the current state-of-the art solutions, our approach takes into account the user-s situation and not imposes a solution that is not suitable to his context and capacity. In this wok, we present our methodology for calculating the mathematical expression complexity and the results of our experiment. Finally, this paper discusses the concepts and principles applied on our system as well as their validation through cases studies. This work is our original contribution to an ongoing research to make informatics more accessible to handicapped users.
Satellite Data Classification Accuracy Assessment Based from Reference Dataset
In order to develop forest management strategies in tropical forest in Malaysia, surveying the forest resources and monitoring the forest area affected by logging activities is essential. There are tremendous effort has been done in classification of land cover related to forest resource management in this country as it is a priority in all aspects of forest mapping using remote sensing and related technology such as GIS. In fact classification process is a compulsory step in any remote sensing research. Therefore, the main objective of this paper is to assess classification accuracy of classified forest map on Landsat TM data from difference number of reference data (200 and 388 reference data). This comparison was made through observation (200 reference data), and interpretation and observation approaches (388 reference data). Five land cover classes namely primary forest, logged over forest, water bodies, bare land and agricultural crop/mixed horticultural can be identified by the differences in spectral wavelength. Result showed that an overall accuracy from 200 reference data was 83.5 % (kappa value 0.7502459; kappa variance 0.002871), which was considered acceptable or good for optical data. However, when 200 reference data was increased to 388 in the confusion matrix, the accuracy slightly improved from 83.5% to 89.17%, with Kappa statistic increased from 0.7502459 to 0.8026135, respectively. The accuracy in this classification suggested that this strategy for the selection of training area, interpretation approaches and number of reference data used were importance to perform better classification result.
Representing Uncertainty in Computer-Generated Forces
The Integrated Performance Modelling Environment (IPME) is a powerful simulation engine for task simulation and performance analysis. However, it has no high level cognition such as memory and reasoning for complex simulation. This article introduces a knowledge representation and reasoning scheme that can accommodate uncertainty in simulations of military personnel with IPME. This approach demonstrates how advanced reasoning models that support similarity-based associative process, rule-based abstract process, multiple reasoning methods and real-time interaction can be integrated with conventional task network modelling to provide greater functionality and flexibility when modelling operator performance.
Automated Knowledge Engineering
This article outlines conceptualization and implementation of an intelligent system capable of extracting knowledge from databases. Use of hybridized features of both the Rough and Fuzzy Set theory render the developed system flexibility in dealing with discreet as well as continuous datasets. A raw data set provided to the system, is initially transformed in a computer legible format followed by pruning of the data set. The refined data set is then processed through various Rough Set operators which enable discovery of parameter relationships and interdependencies. The discovered knowledge is automatically transformed into a rule base expressed in Fuzzy terms. Two exemplary cancer repository datasets (for Breast and Lung Cancer) have been used to test and implement the proposed framework.
A New Group Key Management Protocol for Wireless Ad-Hoc Networks
Ad hoc networks are characterized by multi-hop wireless connectivity and frequently changing network topology. Forming security association among a group of nodes in ad-hoc networks is more challenging than in conventional networks due to the lack of central authority, i.e. fixed infrastructure. With that view in mind, group key management plays an important building block of any secure group communication. The main contribution of this paper is a low complexity key management scheme that is suitable for fully self-organized ad-hoc networks. The protocol is also password authenticated, making it resilient against active attacks. Unlike other existing key agreement protocols, ours make no assumption about the structure of the underlying wireless network, making it suitable for “truly ad-hoc" networks. Finally, we will analyze our protocol to show the computation and communication burden on individual nodes for key establishment.
OXADM Asymmetrical Optical Device: Extending the Application to FTTH System
With the drastically growth in optical communication technology, a lossless, low-crosstalk and multifunction optical switch is most desirable for large-scale photonic network. To realize such a switch, we have introduced the new architecture of optical switch that embedded many functions on single device. The asymmetrical architecture of OXADM consists of 3 parts; selective port, add/drop operation, and path routing. Selective port permits only the interest wavelength pass through and acts as a filter. While add and drop function can be implemented in second part of OXADM architecture. The signals can then be re-routed to any output port or/and perform an accumulation function which multiplex all signals onto single path and then exit to any interest output port. This will be done by path routing operation. The unique features offered by OXADM has extended its application to Fiber to-the Home Technology (FTTH), here the OXADM is used as a wavelength management element in Optical Line Terminal (OLT). Each port is assigned specifically with the operating wavelengths and with the dynamic routing management to ensure no traffic combustion occurs in OLT.
VFAST TCP: A delay-based enhanced version of FAST TCP
This paper is aimed at describing a delay-based endto- end (e2e) congestion control algorithm, called Very FAST TCP (VFAST), which is an enhanced version of FAST TCP. The main idea behind this enhancement is to smoothly estimate the Round-Trip Time (RTT) based on a nonlinear filter, which eliminates throughput and queue oscillation when RTT fluctuates. In this context, an evaluation of the suggested scheme through simulation is introduced, by comparing our VFAST prototype with FAST in terms of throughput, queue behavior, fairness, stability, RTT and adaptivity to changes in network. The achieved simulation results indicate that the suggested protocol offer better performance than FAST TCP in terms of RTT estimation and throughput.
A Formatting Method for Transforming XML Data into HTML

In this paper, we propose a fixed formatting method of PPX(Pretty Printer for XML). PPX is a query language for XML database which has extensive formatting capability that produces HTML as the result of a query. The fixed formatting method is to completely specify the combination of variables and layout specification operators within the layout expression of the GENERATE clause of PPX. In the experiment, a quick comparison shows that PPX requires far less description compared to XSLT or XQuery programs doing the same tasks.

Geometric Operators in the Selection of Human Resources
We study the possibility of using geometric operators in the selection of human resources. We develop three new methods that use the ordered weighted geometric (OWG) operator in different indexes used for the selection of human resources. The objective of these models is to manipulate the neutrality of the old methods so the decision maker is able to select human resources according to his particular attitude. In order to develop these models, first a short revision of the OWG operator is developed. Second, we briefly explain the general process for the selection of human resources. Then, we develop the three new indexes. They will use the OWG operator in the Hamming distance, in the adequacy coefficient and in the index of maximum and minimum level. Finally, an illustrative example about the new approach is given.
Structural Parsing of Natural Language Text in Tamil Using Phrase Structure Hybrid Language Model
Parsing is important in Linguistics and Natural Language Processing to understand the syntax and semantics of a natural language grammar. Parsing natural language text is challenging because of the problems like ambiguity and inefficiency. Also the interpretation of natural language text depends on context based techniques. A probabilistic component is essential to resolve ambiguity in both syntax and semantics thereby increasing accuracy and efficiency of the parser. Tamil language has some inherent features which are more challenging. In order to obtain the solutions, lexicalized and statistical approach is to be applied in the parsing with the aid of a language model. Statistical models mainly focus on semantics of the language which are suitable for large vocabulary tasks where as structural methods focus on syntax which models small vocabulary tasks. A statistical language model based on Trigram for Tamil language with medium vocabulary of 5000 words has been built. Though statistical parsing gives better performance through tri-gram probabilities and large vocabulary size, it has some disadvantages like focus on semantics rather than syntax, lack of support in free ordering of words and long term relationship. To overcome the disadvantages a structural component is to be incorporated in statistical language models which leads to the implementation of hybrid language models. This paper has attempted to build phrase structured hybrid language model which resolves above mentioned disadvantages. In the development of hybrid language model, new part of speech tag set for Tamil language has been developed with more than 500 tags which have the wider coverage. A phrase structured Treebank has been developed with 326 Tamil sentences which covers more than 5000 words. A hybrid language model has been trained with the phrase structured Treebank using immediate head parsing technique. Lexicalized and statistical parser which employs this hybrid language model and immediate head parsing technique gives better results than pure grammar and trigram based model.
A Simplified Approach for Load Flow Analysis of Radial Distribution Network
This paper presents a simple approach for load flow analysis of a radial distribution network. The proposed approach utilizes forward and backward sweep algorithm based on Kirchoff-s current law (KCL) and Kirchoff-s voltage law (KVL) for evaluating the node voltages iteratively. In this approach, computation of branch current depends only on the current injected at the neighbouring node and the current in the adjacent branch. This approach starts from the end nodes of sub lateral line, lateral line and main line and moves towards the root node during branch current computation. The node voltage evaluation begins from the root node and moves towards the nodes located at the far end of the main, lateral and sub lateral lines. The proposed approach has been tested using four radial distribution systems of different size and configuration and found to be computationally efficient.
Use of Novel Algorithms MAJE4 and MACJER-320 for Achieving Confidentiality and Message Authentication in SSL and TLS
Extensive use of the Internet coupled with the marvelous growth in e-commerce and m-commerce has created a huge demand for information security. The Secure Socket Layer (SSL) protocol is the most widely used security protocol in the Internet which meets this demand. It provides protection against eaves droppings, tampering and forgery. The cryptographic algorithms RC4 and HMAC have been in use for achieving security services like confidentiality and authentication in the SSL. But recent attacks against RC4 and HMAC have raised questions in the confidence on these algorithms. Hence two novel cryptographic algorithms MAJE4 and MACJER-320 have been proposed as substitutes for them. The focus of this work is to demonstrate the performance of these new algorithms and suggest them as dependable alternatives to satisfy the need of security services in SSL. The performance evaluation has been done by using practical implementation method.
Transmitter Design for LMS-MIMO-MCCDMA Systems with Pilot Channel Estimates and Zero Forcing Equalizer
We propose a downlink multiple-input multipleoutput (MIMO) multi-carrier code division multiple access (MCCDMA) system with adaptive beamforming algorithm for smart antennas. The algorithm used in this paper is based on the Least Mean Square (LMS), with pilot channel estimation (PCE) and the zero forcing equalizer (ZFE) in the receiver, requiring reference signal and no knowledge channel. MC-CDMA is studied in a multiple antenna context in order to efficiently exploit robustness against multipath effects and multi-user flexibility of MC-CDMA and channel diversity offered by MIMO systems for radio mobile channels. Computer simulations, considering multi-path Rayleigh Fading Channel, interference inter symbol and interference are presented to verify the performance. Simulation results show that the scheme achieves good performance in a multi-user system.
The Investigation of the Role of Institutions in the Process of Growth and Development of Economy
The new institutional Economics helps generalization and expansion of new classic by adding the institution theories to Economic. It is clear that the appropriate institution is among the factors that lead to success in Economic programs. If the institutional are appropriate, the society will save the source and when we make use of time to apply the program, there will be welfare and average revenue product will also increase. In Economy, one should not expect the real manifestation of Economic programs only with a model for estimating and predicting rather institutions of the same purpose and along with production are needed to form the process of growth and development costs. In this research, the institution role in transaction costs, financial markets, distribution of revenue and capital and its influence on the process of growth and development are investigated so that handicaps and problems of Iran Economic Institutions can be recognized. In other words, incapability, non productivity and ambiguity of the institution in Iran Economic are some of the factors that handicap Economic growth and development. For example, Iran government as an important institution while having 20 ministries,83 organizations and 60 years of programming could not go along the growth and development but why?
Clustering Protein Sequences with Tailored General Regression Model Technique
Cluster analysis divides data into groups that are meaningful, useful, or both. Analysis of biological data is creating a new generation of epidemiologic, prognostic, diagnostic and treatment modalities. Clustering of protein sequences is one of the current research topics in the field of computer science. Linear relation is valuable in rule discovery for a given data, such as if value X goes up 1, value Y will go down 3", etc. The classical linear regression models the linear relation of two sequences perfectly. However, if we need to cluster a large repository of protein sequences into groups where sequences have strong linear relationship with each other, it is prohibitively expensive to compare sequences one by one. In this paper, we propose a new technique named General Regression Model Technique Clustering Algorithm (GRMTCA) to benignly handle the problem of linear sequences clustering. GRMT gives a measure, GR*, to tell the degree of linearity of multiple sequences without having to compare each pair of them.
Simulation of Superconducting Nanowire Single-Photon Detector with Circuit Modeling
Single photon detectors have been fabricated NbN nano wire. These detectors are fabricated from high quality, ultra high vacuum sputtered NbN thin films on a sapphire substrate. In this work a typical schematic of the nanowire Single Photon Detector structure and then driving and measurement electronic circuit are shown. The response of superconducting nanowire single photon detectors during a photo detection event, is modeled by a special electrical circuits (two circuit). Finally, current through the wire is calculated by solving equations of models.
A Multiresolution Approach for Noised Texture Classification based on the Co-occurrence Matrix and First Order Statistics
Wavelet transform provides several important characteristics which can be used in a texture analysis and classification. In this work, an efficient texture classification method, which combines concepts from wavelet and co-occurrence matrices, is presented. An Euclidian distance classifier is used to evaluate the various methods of classification. A comparative study is essential to determine the ideal method. Using this conjecture, we developed a novel feature set for texture classification and demonstrate its effectiveness
Data Extraction of XML Files using Searching and Indexing Techniques

XML files contain data which is in well formatted manner. By studying the format or semantics of the grammar it will be helpful for fast retrieval of the data. There are many algorithms which describes about searching the data from XML files. There are no. of approaches which uses data structure or are related to the contents of the document. In these cases user must know about the structure of the document and information retrieval techniques using NLPs is related to content of the document. Hence the result may be irrelevant or not so successful and may take more time to search.. This paper presents fast XML retrieval techniques by using new indexing technique and the concept of RXML. When indexing an XML document, the system takes into account both the document content and the document structure and assigns the value to each tag from file. To query the system, a user is not constrained about fixed format of query.

Chattering-free Sliding Mode Control for an Active Magnetic Bearing System

In this paper, a few chattering-free Sliding Mode Controllers (SMC) are proposed to stabilize an Active Magnetic Bearing (AMB) system with gyroscopic effect that is proportional to the rotor speed. The improved switching terms of the controller inherited from the saturation-type function and boundary layer control technique is shown to be able to achieve bounded and asymptotic stability, respectively, while the chattering effect in the input is attenuated. This is proven to be advantageous for AMB system since minimization of chattering results in optimized control effort. The performance of each controller is demonstrated via result of simulation in which the measurement of the total consumed energy and maximum control magnitude of each controller illustrates the effectiveness of the proposed controllers.

Human Face Detection and Segmentation using Eigenvalues of Covariance Matrix, Hough Transform and Raster Scan Algorithms
In this paper we propose a novel method for human face segmentation using the elliptical structure of the human head. It makes use of the information present in the edge map of the image. In this approach we use the fact that the eigenvalues of covariance matrix represent the elliptical structure. The large and small eigenvalues of covariance matrix are associated with major and minor axial lengths of an ellipse. The other elliptical parameters are used to identify the centre and orientation of the face. Since an Elliptical Hough Transform requires 5D Hough Space, the Circular Hough Transform (CHT) is used to evaluate the elliptical parameters. Sparse matrix technique is used to perform CHT, as it squeeze zero elements, and have only a small number of non-zero elements, thereby having an advantage of less storage space and computational time. Neighborhood suppression scheme is used to identify the valid Hough peaks. The accurate position of the circumference pixels for occluded and distorted ellipses is identified using Bresenham-s Raster Scan Algorithm which uses the geometrical symmetry properties. This method does not require the evaluation of tangents for curvature contours, which are very sensitive to noise. The method has been evaluated on several images with different face orientations.
Performance Trade-Off of File System between Overwriting and Dynamic Relocation on a Solid State Drive
Most file systems overwrite modified file data and metadata in their original locations, while the Log-structured File System (LFS) dynamically relocates them to other locations. We design and implement the Evergreen file system that can select between overwriting or relocation for each block of a file or metadata. Therefore, the Evergreen file system can achieve superior write performance by sequentializing write requests (similar to LFS-style relocation) when space utilization is low and overwriting when utilization is high. Another challenging issue is identifying performance benefits of LFS-style relocation over overwriting on a newly introduced SSD (Solid State Drive) which has only Flash-memory chips and control circuits without mechanical parts. Our experimental results measured on a SSD show that relocation outperforms overwriting when space utilization is below 80% and vice versa.
Dynamic Metrics for Polymorphism in Object Oriented Systems
Metrics is the process by which numbers or symbols are assigned to attributes of entities in the real world in such a way as to describe them according to clearly defined rules. Software metrics are instruments or ways to measuring all the aspect of software product. These metrics are used throughout a software project to assist in estimation, quality control, productivity assessment, and project control. Object oriented software metrics focus on measurements that are applied to the class and other characteristics. These measurements convey the software engineer to the behavior of the software and how changes can be made that will reduce complexity and improve the continuing capability of the software. Object oriented software metric can be classified in two types static and dynamic. Static metrics are concerned with all the aspects of measuring by static analysis of software and dynamic metrics are concerned with all the measuring aspect of the software at run time. Major work done before, was focusing on static metric. Also some work has been done in the field of dynamic nature of the software measurements. But research in this area is demanding for more work. In this paper we give a set of dynamic metrics specifically for polymorphism in object oriented system.
Predicting Protein Function using Decision Tree
The drug discovery process starts with protein identification because proteins are responsible for many functions required for maintenance of life. Protein identification further needs determination of protein function. Proposed method develops a classifier for human protein function prediction. The model uses decision tree for classification process. The protein function is predicted on the basis of matched sequence derived features per each protein function. The research work includes the development of a tool which determines sequence derived features by analyzing different parameters. The other sequence derived features are determined using various web based tools.
Decision Support System for Farm Management
The emergence of information technology has resulted in an ever-increasing demand to use computers for the efficient management and dissemination of information. Keeping in view the strong need of farmers to collect important and updated information for interactive, flexible and quick decision-making, a model of Decision Support System for Farm Management is developed. The paper discusses the use of Internet technology for the farmers to take decisions. A model is developed for the farmers to access online interactive and flexible information for their farm management. The workflow of the model is presented highlighting the information transfer between different modules.
A Comparative Analysis of Fuzzy, Neuro-Fuzzy and Fuzzy-GA Based Approaches for Software Reusability Evaluation
Software Reusability is primary attribute of software quality. There are metrics for identifying the quality of reusable components but the function that makes use of these metrics to find reusability of software components is still not clear. These metrics if identified in the design phase or even in the coding phase can help us to reduce the rework by improving quality of reuse of the component and hence improve the productivity due to probabilistic increase in the reuse level. In this paper, we have devised the framework of metrics that uses McCabe-s Cyclometric Complexity Measure for Complexity measurement, Regularity Metric, Halstead Software Science Indicator for Volume indication, Reuse Frequency metric and Coupling Metric values of the software component as input attributes and calculated reusability of the software component. Here, comparative analysis of the fuzzy, Neuro-fuzzy and Fuzzy-GA approaches is performed to evaluate the reusability of software components and Fuzzy-GA results outperform the other used approaches. The developed reusability model has produced high precision results as expected by the human experts.
An Economic Analysis of Phu Kradueng National Park
The purposes of this study were as follows to evaluate the economic value of Phu Kradueng National Park by the travel cost method (TCM) and the contingent valuation method (CVM) and to estimate the demand for traveling and the willingness to pay. The data for this study were collected by conducting two large scale surveys on users and non-users. A total of 1,016 users and 1,034 non-users were interviewed. The data were analyzed using multiple linear regression analysis, logistic regression model and the consumer surplus (CS) was the integral of demand function for trips. The survey found, were as follows: 1)Using the travel cost method which provides an estimate of direct benefits to park users, we found that visitors- total willingness to pay per visit was 2,284.57 bath, of which 958.29 bath was travel cost, 1,129.82 bath was expenditure for accommodation, food, and services, and 166.66 bath was consumer surplus or the visitors -net gain or satisfaction from the visit (the integral of demand function for trips). 2) Thai visitors to Phu Kradueng National Park were further willing to pay an average of 646.84 bath per head per year to ensure the continued existence of Phu Kradueng National Park and to preserve their option to use it in the future. 3) Thai non-visitors, on the other hand, are willing to pay an average of 212.61 bath per head per year for the option and existence value provided by the Park. 4) The total economic value of Phu Kradueng National Park to Thai visitors and non-visitors taken together stands today at 9,249.55 million bath per year. 5) The users- average willingness to pay for access to Phu Kradueng National Park rises from 40 bath to 84.66 bath per head per trip for improved services such as road improvement, increased cleanliness, and upgraded information. This paper was needed to investigate of the potential market demand for bio prospecting in Phu Kradueng national Park and to investigate how a larger share of the economic benefits of tourism could be distributed income to the local residents.
Design of a Three Phase Active Power Filter with Sliding Mode Control and Energy Feedback
Nonlinear and unbalance loads in three phase networks create harmonics and losses. Active and passive filters are used for elimination or reduction of these effects. Passive filters have some limitations. For example, they are designed only for a specific frequency and they may cause to resonance in the network at the point of common coupling. The other drawback of a passive filter is that the sizes of required elements are normally large. The active filter can improve some of limitations of passive filter for example; they can eliminate more than one harmonic and don't cause resonance in the network. In this paper inverter analysis have been done simultaneously in three phase and the RL impedance of the line have been considered. A sliding mode control based on energy feedback of capacitors is employed in the design with this method, the dynamic speed of the filter is improved effectively and harmonics and load unbalance is compensating quickly.
Mathematical Modeling of Current Harmonics Caused by Personal Computers
Personal computers draw non-sinusoidal current with odd harmonics more significantly. Power Quality of distribution networks is severely affected due to the flow of these generated harmonics during the operation of electronic loads. In this paper, mathematical modeling of odd harmonics in current like 3rd, 5th, 7th and 9th influencing the power quality has been presented. Live signals have been captured with the help of power quality analyzer for analysis purpose. The interesting feature is that Total Harmonic Distortion (THD) in current decreases with the increase of nonlinear loads has been verified theoretically. The results obtained using mathematical expressions have been compared with the practical results and exciting results have been found.
The Implementation of Remote Automation Execution Agent over ACL on QOS POLICY Based System
This paper will present the implementation of QoS policy based system by utilizing rules on Access Control List (ACL) over Layer 3 (L3) switch. Also presented is the architecture on that implementation; the tools being used and the result were gathered. The system architecture has an ability to control ACL rules which are installed inside an external L3 switch. ACL rules used to instruct the way of access control being executed, in order to entertain all traffics through that particular switch. The main advantage of using this approach is that the single point of failure could be prevented when there are any changes on ACL rules inside L3 switches. Another advantage is that the agent could instruct ACL rules automatically straight away based on the changes occur on policy database without configuring them one by one. Other than that, when QoS policy based system was implemented in distributed environment, the monitoring process can be synchronized easily due to the automate process running by agent over external policy devices.
A Hybrid Machine Learning System for Stock Market Forecasting
In this paper, we propose a hybrid machine learning system based on Genetic Algorithm (GA) and Support Vector Machines (SVM) for stock market prediction. A variety of indicators from the technical analysis field of study are used as input features. We also make use of the correlation between stock prices of different companies to forecast the price of a stock, making use of technical indicators of highly correlated stocks, not only the stock to be predicted. The genetic algorithm is used to select the set of most informative input features from among all the technical indicators. The results show that the hybrid GA-SVM system outperforms the stand alone SVM system.
Analysis and Application of in Indirect MinimumJerk Method for Higher order Differential Equation in Dynamics Optimization Systems
Both the minimum energy consumption and smoothness, which is quantified as a function of jerk, are generally needed in many dynamic systems such as the automobile and the pick-and-place robot manipulator that handles fragile equipments. Nevertheless, many researchers come up with either solely concerning on the minimum energy consumption or minimum jerk trajectory. This research paper considers the indirect minimum Jerk method for higher order differential equation in dynamics optimization proposes a simple yet very interesting indirect jerks approaches in designing the time-dependent system yielding an alternative optimal solution. Extremal solutions for the cost functions of indirect jerks are found using the dynamic optimization methods together with the numerical approximation. This case considers the linear equation of a simple system, for instance, mass, spring and damping. The simple system uses two mass connected together by springs. The boundary initial is defined the fix end time and end point. The higher differential order is solved by Galerkin-s methods weight residual. As the result, the 6th higher differential order shows the faster solving time.
Precise Measurement of Displacement using Pixels
Manufacturing processes demand tight dimensional tolerances. The paper concerns a transducer for precise measurement of displacement, based on a camera containing a linescan chip. When tests were conducted using a track of black and white stripes with a 2mm pitch, errors in measuring on individual cycle amounted to 1.75%, suggesting that a precision of 35 microns is achievable.
Determination of Stress Concentration Factors of a Steam Turbine Rotor by FEA
Stress Concentration Factors are significant in machine design as it gives rise to localized stress when any change in the design of surface or abrupt change in the cross section occurs. Almost all machine components and structural members contain some form of geometrical or microstructural discontinuities. These discontinuities are very dangerous and lead to failure. So, it is very much essential to analyze the stress concentration factors for critical applications like Turbine Rotors. In this paper Finite Element Analysis (FEA) with extremely fine mesh in the vicinity of the blades of Steam Turbine Rotor is applied to determine stress concentration factors. A model of Steam Turbine Rotor is shown in Fig. 1.
Piezoelectric Transducer Modeling: with System Identification (SI) Method
System identification is the process of creating models of dynamic process from input- output signals. The aim of system identification can be identified as “ to find a model with adjustable parameters and then to adjust them so that the predicted output matches the measured output". This paper presents a method of modeling and simulating with system identification to achieve the maximum fitness for transformation function. First by using optimized KLM equivalent circuit for PVDF piezoelectric transducer and assuming different inputs including: sinuside, step and sum of sinusides, get the outputs, then by using system identification toolbox in MATLAB, we estimate the transformation function from inputs and outputs resulted in last program. Then compare the fitness of transformation function resulted from using ARX,OE(Output- Error) and BJ(Box-Jenkins) models in system identification toolbox and primary transformation function form KLM equivalent circuit.
Over-Height Vehicle Detection in Low Headroom Roads Using Digital Video Processing
In this paper we present a new method for over-height vehicle detection in low headroom streets and highways using digital video possessing. The accuracy and the lower price comparing to present detectors like laser radars and the capability of providing extra information like speed and height measurement make this method more reliable and efficient. In this algorithm the features are selected and tracked using KLT algorithm. A blob extraction algorithm is also applied using background estimation and subtraction. Then the world coordinates of features that are inside the blobs are estimated using a noble calibration method. As, the heights of the features are calculated, we apply a threshold to select overheight features and eliminate others. The over-height features are segmented using some association criteria and grouped using an undirected graph. Then they are tracked through sequential frames. The obtained groups refer to over-height vehicles in a scene.
Fusion of ETM+ Multispectral and Panchromatic Texture for Remote Sensing Classification
This paper proposes to use ETM+ multispectral data and panchromatic band as well as texture features derived from the panchromatic band for land cover classification. Four texture features including one 'internal texture' and three GLCM based textures namely correlation, entropy, and inverse different moment were used in combination with ETM+ multispectral data. Two data sets involving combination of multispectral, panchromatic band and its texture were used and results were compared with those obtained by using multispectral data alone. A decision tree classifier with and without boosting were used to classify different datasets. Results from this study suggest that the dataset consisting of panchromatic band, four of its texture features and multispectral data was able to increase the classification accuracy by about 2%. In comparison, a boosted decision tree was able to increase the classification accuracy by about 3% with the same dataset.
Artificial Neural Network based Modeling of Evaporation Losses in Reservoirs
An Artificial Neural Network based modeling technique has been used to study the influence of different combinations of meteorological parameters on evaporation from a reservoir. The data set used is taken from an earlier reported study. Several input combination were tried so as to find out the importance of different input parameters in predicting the evaporation. The prediction accuracy of Artificial Neural Network has also been compared with the accuracy of linear regression for predicting evaporation. The comparison demonstrated superior performance of Artificial Neural Network over linear regression approach. The findings of the study also revealed the requirement of all input parameters considered together, instead of individual parameters taken one at a time as reported in earlier studies, in predicting the evaporation. The highest correlation coefficient (0.960) along with lowest root mean square error (0.865) was obtained with the input combination of air temperature, wind speed, sunshine hours and mean relative humidity. A graph between the actual and predicted values of evaporation suggests that most of the values lie within a scatter of ±15% with all input parameters. The findings of this study suggest the usefulness of ANN technique in predicting the evaporation losses from reservoirs.
Oxygen Transfer by Multiple Inclined Plunging Water Jets
There has been a growing interest in the oxygenation by plunging water jets in the last few years due to their inherent advantages, like energy-efficient, low operation cost, etc. Though a lot of work has been reported on the oxygen-transfer by single plunging water jets but very few studies have been carried out using multiple plunging jets. In this paper, volumetric oxygen-transfer coefficient and oxygen-transfer efficiency has been studied experimentally for multiple inclined plunging jets (having jet plunge angle of 60 0 ) in a pool of water for different configurations, in terms of varying number of jets and jet diameters. This research suggests that the volumetric oxygen-transfer coefficient and oxygentransfer efficiency of the multiple inclined plunging jets for air-water system are significantly higher than those of a single vertical as well as inclined plunging jet for same flow area and other similar conditions. The study also reveals that the oxygen-transfer increase with increase in number of multiple jets under similar conditions, which will be most advantageous and energy-efficient in practical situations when large volumes of wastewaters are to be treated. A relationship between volumetric oxygen-transfer coefficient and jet parameters is also proposed. The suggested relationship predicts the volumetric oxygen-transfer coefficient for multiple inclined plunging jet(s) within a scatter of ±15 percent. The relationship will be quite useful in scale-up and in deciding optimum configuration of multiple inclined plunging jet aeration system.
An Example of Open Robot Controller Architecture - For Power Distribution Line Maintenance Robot System -
In this paper, we propose an architecture for easily constructing a robot controller. The architecture is a multi-agent system which has eight agents: the Man-machine interface, Task planner, Task teaching editor, Motion planner, Arm controller, Vehicle controller, Vision system and CG display. The controller has three databases: the Task knowledge database, the Robot database and the Environment database. Based on this controller architecture, we are constructing an experimental power distribution line maintenance robot system and are doing the experiment for the maintenance tasks, for example, “Bolt insertion task".
Decision Rule Induction in a Learning Content Management System
A learning content management system (LCMS) is an environment to support web-based learning content development. Primary function of the system is to manage the learning process as well as to generate content customized to meet a unique requirement of each learner. Among the available supporting tools offered by several vendors, we propose to enhance the LCMS functionality to individualize the presented content with the induction ability. Our induction technique is based on rough set theory. The induced rules are intended to be the supportive knowledge for guiding the content flow planning. They can also be used as decision rules to help content developers on managing content delivered to individual learner.
Knowledge Mining in Web-based Learning Environments
The state of the art in instructional design for computer-assisted learning has been strongly influenced by advances in information technology, Internet and Web-based systems. The emphasis of educational systems has shifted from training to learning. The course delivered has also been changed from large inflexible content to sequential small chunks of learning objects. The concepts of learning objects together with the advanced technologies of Web and communications support the reusability, interoperability, and accessibility design criteria currently exploited by most learning systems. These concepts enable just-in-time learning. We propose to extend theses design criteria further to include the learnability concept that will help adapting content to the needs of learners. The learnability concept offers a better personalization leading to the creation and delivery of course content more appropriate to performance and interest of each learner. In this paper we present a new framework of learning environments containing knowledge discovery as a tool to automatically learn patterns of learning behavior from learners' profiles and history.
Component Based Framework for Authoring and Multimedia Training in Mathematics
The new programming technologies allow for the creation of components which can be automatically or manually assembled to reach a new experience in knowledge understanding and mastering or in getting skills for a specific knowledge area. The project proposes an interactive framework that permits the creation, combination and utilization of components that are specific to mathematical training in high schools. The main framework-s objectives are: • authoring lessons by the teacher or the students; all they need are simple operating skills for Equation Editor (or something similar, or Latex); the rest are just drag & drop operations, inserting data into a grid, or navigating through menus • allowing sonorous presentations of mathematical texts and solving hints (easier understood by the students) • offering graphical representations of a mathematical function edited in Equation • storing of learning objects in a database • storing of predefined lessons (efficient for expressions and commands, the rest being calculations; allows a high compression) • viewing and/or modifying predefined lessons, according to the curricula The whole thing is focused on a mathematical expressions minicompiler, storing the code that will be later used for different purposes (tables, graphics, and optimisations). Programming technologies used. A Visual C# .NET implementation is proposed. New and innovative digital learning objects for mathematics will be developed; they are capable to interpret, contextualize and react depending on the architecture where they are assembled.
Optimizing of Fuzzy C-Means Clustering Algorithm Using GA
Fuzzy C-means Clustering algorithm (FCM) is a method that is frequently used in pattern recognition. It has the advantage of giving good modeling results in many cases, although, it is not capable of specifying the number of clusters by itself. In FCM algorithm most researchers fix weighting exponent (m) to a conventional value of 2 which might not be the appropriate for all applications. Consequently, the main objective of this paper is to use the subtractive clustering algorithm to provide the optimal number of clusters needed by FCM algorithm by optimizing the parameters of the subtractive clustering algorithm by an iterative search approach and then to find an optimal weighting exponent (m) for the FCM algorithm. In order to get an optimal number of clusters, the iterative search approach is used to find the optimal single-output Sugenotype Fuzzy Inference System (FIS) model by optimizing the parameters of the subtractive clustering algorithm that give minimum least square error between the actual data and the Sugeno fuzzy model. Once the number of clusters is optimized, then two approaches are proposed to optimize the weighting exponent (m) in the FCM algorithm, namely, the iterative search approach and the genetic algorithms. The above mentioned approach is tested on the generated data from the original function and optimal fuzzy models are obtained with minimum error between the real data and the obtained fuzzy models.
Analytical Solution for Free Vibration of Rectangular Kirchhoff Plate from Wave Approach
In this paper, an analytical approach for free vibration analysis of four edges simply supported rectangular Kirchhoff plates is presented. The method is based on wave approach. From wave standpoint vibration propagate, reflect and transmit in a structure. Firstly, the propagation and reflection matrices for plate with simply supported boundary condition are derived. Then, these matrices are combined to provide a concise and systematic approach to free vibration analysis of a simply supported rectangular Kirchhoff plate. Subsequently, the eigenvalue problem for free vibration of plates is formulated and the equation of plate natural frequencies is constructed. Finally, the effectiveness of the approach is shown by comparison of the results with existing classical solution.
Real-Time Control of a Two-Wheeled Inverted Pendulum Mobile Robot

The research on two-wheeled inverted pendulum (TWIP) mobile robots or commonly known as balancing robots have gained momentum over the last decade in a number of robotic laboratories around the world. This paper describes the hardware design of such a robot. The objective of the design is to develop a TWIP mobile robot as well as MATLAB interfacing configuration to be used as flexible platform comprises of embedded unstable linear plant intended for research and teaching purposes. Issues such as selection of actuators and sensors, signal processing units, MATLAB Real Time Workshop coding, modeling and control scheme will be addressed and discussed. The system is then tested using a wellknown state feedback controller to verify its functionality.

Direct Power Control Strategies for Multilevel Inverter Based Custom Power Devices
Custom power is a technology driven product and service solution which embraces a family devices such as Dynamic Voltage Restorer (DVR), Distributed Shunt Compensator (DSTATCOM), Solid State Breaker (SSB) etc which will provide power quality functions at distribution voltages. The rapid response of these devices enables them to operate in real time, providing continuous and dynamic control of the supply including voltage and reactive power regulation, harmonic reduction and elimination of voltage dips. This paper presents the benefits of multilevel inverters when they are used for DPC based custom power devices. Power flow control mechanism, salient features, advantages and disadvantages of direct power control (DPC) using lookup table, SVM, predictive voltage vector and hybrid DPC strategies are discussed in this paper. Simulation results of three level inverter based STATCOM, harmonic analysis of multi level inverters are presented at the end.
Investigation of Various PWM Techniques for Shunt Active Filter
Pulse width modulation (PWM) techniques have been the subject of intensive research for different industrial and power sector applications. A large variety of methods, different in concept and performance, have been newly developed and described. This paper analyzes the comparative merits of Sinusoidal Pulse Width Modulation (SPWM) and Space Vector Pulse Width Modulation (SVPWM) techniques and the suitability of these techniques in a Shunt Active Filter (SAF). The objective is to select the scheme that offers effective utilization of DC bus voltage and also harmonic reduction at the input side. The effectiveness of the PWM techniques is tested in the SAF configuration with a non linear load. The performance of the SAF with the SPWM and (SVPWM) techniques are compared with respect to the THD in source current. The study reveals that in the context of closed loop SAF control with the SVPWM technique there is only a minor improvement in THD. The utilization of the DC bus with SVPWM is also not significant compared to that with SPWM because of the non sinusoidal modulating signal from the controller in SAF configuration.
Classification of Discharges Initiated by Liquid Droplet on Insulation Material under AC Voltages Adopting UHF Technique
In the present work, an attempt has been made to understand the feasibility of using UHF technique for identification of any corona discharges/ arcing in insulating material due to water droplets. The sensors of broadband type are useful for identification of such discharges. It is realised that arcing initiated by liquid droplet radiates UHF signals in the entire bandwidth up to 2 GHz. The frequency content of the UHF signal generated due to corona/arcing is not much varied in epoxy nanocomposites with different weight percentage of clay content. The exfoliated/intercalated properties were analysed through TEM studies. It is realized that corona initiated discharges are of intermittent process. The hydrophobicity of the material characterized through contact angle measurement. It is realized that low Wt % of nanoclay content in epoxy resin reduces the surface carbonization due to arcing/corona discharges. The results of the study with gamma irradiated specimen indicates that contact angle, discharge inception time and evaporation time of the liquid are much lower than the virgin epoxy nanocomposite material.
Understanding the Discharge Activities in Transformer Oil under AC and DC Voltage Adopting UHF Technique
Design of Converter transformer insulation is a major challenge. The insulation of these transformers is stressed by both AC and DC voltages. Particle contamination is one of the major problems in insulation structures, as they generate partial discharges leading it to major failure of insulation. Similarly corona discharges occur in transformer insulation. This partial discharge due to particle movement / corona formation in insulation structure under different voltage wave shapes, are different. In the present study, UHF technique is adopted to understand the discharge activity and could be realized that the characteristics of UHF signal generated under low and high fields are different. In the case of corona generated signal, the frequency content of the UHF sensor output lies in the range 0.3-1.2 GHz and is not much varied except for its increase in magnitude of discharge with the increase in applied voltage. It is realized that the current signal injected due to partial discharges/corona is about 4ns duration measured for first one half cycle. Wavelet technique is adopted in the present study. It allows one to identify the frequency content present in the signal at different instant of time. The STD-MRA analysis helps one to identify the frequency band in which the energy content of the UHF signal is maximum.
Synchronization Technique for Random Switching Frequency Pulse-Width Modulation

This paper proposes a synchronized random switching frequency pulse width modulation (SRSFPWM). In this technique, the clock signal is used to control the random noise frequency which is produced by the feedback voltage of a hysteresis circuit. These make the triangular carrier frequency equaling to the random noise frequency in each switching period with the symmetrical positive and negative slopes of triangular carrier. Therefore, there is no error voltage in PWM signal. The PSpice simulated results shown the proposed technique improved the performance in case of low frequency harmonics of PWM signal comparing with conventional random switching frequency PWM.

Feed-Forward Control in Resonant DC Link Inverter
This paper proposes a feed-forward control in resonant dc link inverter. The feed-forward control configuration is based on synchronous sigma-delta modulation. The simulation results showing the proposed technique can reject non-ideal dc bus improving the total harmonic distortion.
Feed-Forward Control in Half-Bridge Resonant DC Link Inverter
This paper proposes a feed-forward control in a halfbridge resonant dc link inverter. The configuration of feed-forward control is based on synchronous sigma-delta modulation and the halfbridge resonant dc link inverter consists of two inductors, one capacitor and two power switches. The simulation results show the proposed technique can reject non-ideal dc bus improving the total harmonic distortion.
Impact of Loading Conditions on the Emission- Economic Dispatch
Environmental awareness and the recent environmental policies have forced many electric utilities to restructure their operational practices to account for their emission impacts. One way to accomplish this is by reformulating the traditional economic dispatch problem such that emission effects are included in the mathematical model. This paper presents a Particle Swarm Optimization (PSO) algorithm to solve the Economic- Emission Dispatch problem (EED) which gained recent attention due to the deregulation of the power industry and strict environmental regulations. The problem is formulated as a multi-objective one with two competing functions, namely economic cost and emission functions, subject to different constraints. The inequality constraints considered are the generating unit capacity limits while the equality constraint is generation-demand balance. A novel equality constraint handling mechanism is proposed in this paper. PSO algorithm is tested on a 30-bus standard test system. Results obtained show that PSO algorithm has a great potential in handling multi-objective optimization problems and is capable of capturing Pareto optimal solution set under different loading conditions.
Model and Control of Renewable Energy Systems
This paper presents a developed method for controlling multi-renewable energy generators. The control system depends basically on three sensors (wind anemometer, solar sensor, and voltage sensor). These sensors represent PLC-s analogue inputs. Controlling the output voltage supply can be achieved by an enhanced method of interlocking between the renewable energy generators, depending on those sensors and output contactors.
Sensitivity and Removed THD of a Phase- Cutting Dimmer
In this paper, we consider a designed and implemented phase-cutting dimmer. In fact, the dimmer is closed loop and a microcontroller calculates and then regulates the firing delay angles of each channel. Depending on the firing angle, the harmonic distortion in the input current will not comply with international standards, such as IEC 61000-3-2 (class C equipments). For solving this problem, eight harmonic compensators have been added to the dimmer. So, the proposed dimmer has a little harmonic distortion in the input current whereas conventional phase-cutting dimmers are not so. Sensitivity and removed THD of the proposed dimmer will be presented.
Implementation of Adder-Subtracter Design with VerilogHDL
According to the density of the chips, designers are trying to put so any facilities of computational and storage on single chips. Along with the complexity of computational and storage circuits, the designing, testing and debugging become more and more complex and expensive. So, hardware design will be built by using very high speed hardware description language, which is more efficient and cost effective. This paper will focus on the implementation of 32-bit ALU design based on Verilog hardware description language. Adder and subtracter operate correctly on both unsigned and positive numbers. In ALU, addition takes most of the time if it uses the ripple-carry adder. The general strategy for designing fast adders is to reduce the time required to form carry signals. Adders that use this principle are called carry look- ahead adder. The carry look-ahead adder is to be designed with combination of 4-bit adders. The syntax of Verilog HDL is similar to the C programming language. This paper proposes a unified approach to ALU design in which both simulation and formal verification can co-exist.
Numerical Simulation of a Conventional Heat Pipe

The steady incompressible flow has been solved in cylindrical coordinates in both vapour region and wick structure. The governing equations in vapour region are continuity, Navier-Stokes and energy equations. These equations have been solved using SIMPLE algorithm. For study of parameters variation on heat pipe operation, a benchmark has been chosen and the effect of changing one parameter has been analyzed when the others have been fixed.

Rock Textures Classification Based on Textural and Spectral Features
In this paper, we proposed a method to classify each type of natural rock texture. Our goal is to classify 26 classes of rock textures. First, we extract five features of each class by using principle component analysis combining with the use of applied spatial frequency measurement. Next, the effective node number of neural network was tested. We used the most effective neural network in classification process. The results from this system yield quite high in recognition rate. It is shown that high recognition rate can be achieved in separation of 26 stone classes.
Tsunami Modelling using the Well-Balanced Scheme
A well balanced numerical scheme based on stationary waves for shallow water flows with arbitrary topography has been introduced by Thanh et al. [18]. The scheme was constructed so that it maintains equilibrium states and tests indicate that it is stable and fast. Applying the well-balanced scheme for the one-dimensional shallow water equations, we study the early shock waves propagation towards the Phuket coast in Southern Thailand during a hypothetical tsunami. The initial tsunami wave is generated in the deep ocean with the strength that of Indonesian tsunami of 2004.
Accurate Time Domain Method for Simulation of Microstructured Electromagnetic and Photonic Structures
A time-domain numerical model within the framework of transmission line modeling (TLM) is developed to simulate electromagnetic pulse propagation inside multiple microcavities forming photonic crystal (PhC) structures. The model developed is quite general and is capable of simulating complex electromagnetic problems accurately. The field quantities can be mapped onto a passive electrical circuit equivalent what ensures that TLM is provably stable and conservative at a local level. Furthermore, the circuit representation allows a high level of hybridization of TLM with other techniques and lumped circuit models of components and devices. A photonic crystal structure formed by rods (or blocks) of high-permittivity dieletric material embedded in a low-dielectric background medium is simulated as an example. The model developed gives vital spatio-temporal information about the signal, and also gives spectral information over a wide frequency range in a single run. The model has wide applications in microwave communication systems, optical waveguides and electromagnetic materials simulations.
Next Generation Networks and Their Relation with Ad-hoc Networks
The communication networks development and advancement during two last decades has been toward a single goal and that is gradual change from circuit-switched networks to packed switched ones. Today a lot of networks operates are trying to transform the public telephone networks to multipurpose packed switch. This new achievement is generally called "next generation networks". In fact, the next generation networks enable the operators to transfer every kind of services (sound, data and video) on a network. First, in this report the definition, characteristics and next generation networks services and then ad-hoc networks role in the next generation networks are studied.
A Low Power SRAM Base on Novel Word-Line Decoding
This paper proposes a low power SRAM based on five transistor SRAM cell. Proposed SRAM uses novel word-line decoding such that, during read/write operation, only selected cell connected to bit-line whereas, in conventional SRAM (CV-SRAM), all cells in selected row connected to their bit-lines, which in turn develops differential voltages across all bit-lines, and this makes energy consumption on unselected bit-lines. In proposed SRAM memory array divided into two halves and this causes data-line capacitance to reduce. Also proposed SRAM uses one bit-line and thus has lower bit-line leakage compared to CV-SRAM. Furthermore, the proposed SRAM incurs no area overhead, and has comparable read/write performance versus the CV-SRAM. Simulation results in standard 0.25μm CMOS technology shows in worst case proposed SRAM has 80% smaller dynamic energy consumption in each cycle compared to CV-SRAM. Besides, energy consumption in each cycle of proposed SRAM and CV-SRAM investigated analytically, the results of which are in good agreement with the simulation results.
A Novel Four-Transistor SRAM Cell with Low Dynamic Power Consumption
This paper presents a novel CMOS four-transistor SRAM cell for very high density and low power embedded SRAM applications as well as for stand-alone SRAM applications. This cell retains its data with leakage current and positive feedback without refresh cycle. The new cell size is 20% smaller than a conventional six-transistor cell using same design rules. Also proposed cell uses two word-lines and one pair bit-line. Read operation perform from one side of cell, and write operation perform from another side of cell, and swing voltage reduced on word-lines thus dynamic power during read/write operation reduced. The fabrication process is fully compatible with high-performance CMOS logic technologies, because there is no need to integrate a poly-Si resistor or a TFT load. HSPICE simulation in standard 0.25μm CMOS technology confirms all results obtained from this paper.
Numerical Simulation of a Single Air Bubble Rising in Water with Various Models of Surface Tension Force

Different numerical methods are employed and developed for simulating interfacial flows. A large range of applications belong to this group, e.g. two-phase flows of air bubbles in water or water drops in air. In such problems surface tension effects often play a dominant role. In this paper, various models of surface tension force for interfacial flows, the CSF, CSS, PCIL and SGIP models have been applied to simulate the motion of small air bubbles in water and the results were compared and reviewed. It has been pointed out that by using SGIP or PCIL models, we are able to simulate bubble rise and obtain results in close agreement with the experimental data.

Generalization of SGIP Surface Tension Force Model in Three-Dimensional Flows and Compare to Other Models in Interfacial Flows
In this paper, the two-dimensional stagger grid interface pressure (SGIP) model has been generalized and presented into three-dimensional form. For this purpose, various models of surface tension force for interfacial flows have been investigated and compared with each other. The VOF method has been used for tracking the interface. To show the ability of the SGIP model for three-dimensional flows in comparison with other models, pressure contours, maximum spurious velocities, norm spurious flow velocities and pressure jump error for motionless drop of liquid and bubble of gas are calculated using different models. It has been pointed out that SGIP model in comparison with the CSF, CSS and PCIL models produces the least maximum and norm spurious velocities. Additionally, the new model produces more accurate results in calculating the pressure jumps across the interface for motionless drop of liquid and bubble of gas which is generated in surface tension force.
2D Human Motion Regeneration with Stick Figure Animation Using Accelerometers
This paper explores the opportunity of using tri-axial wireless accelerometers for supervised monitoring of sports movements. A motion analysis system for the upper extremities of lawn bowlers in particular is developed. Accelerometers are placed on parts of human body such as the chest to represent the shoulder movements, the back to capture the trunk motion, back of the hand, the wrist and one above the elbow, to capture arm movements. These sensors placement are carefully designed in order to avoid restricting bowler-s movements. Data is acquired from these sensors in soft-real time using virtual instrumentation; the acquired data is then conditioned and converted into required parameters for motion regeneration. A user interface was also created to facilitate in the acquisition of data, and broadcasting of commands to the wireless accelerometers. All motion regeneration in this paper deals with the motion of the human body segment in the X and Y direction, looking into the motion of the anterior/ posterior and lateral directions respectively.
Learning Monte Carlo Data for Circuit Path Length
This paper analyzes the patterns of the Monte Carlo data for a large number of variables and minterms, in order to characterize the circuit path length behavior. We propose models that are determined by training process of shortest path length derived from a wide range of binary decision diagram (BDD) simulations. The creation of the model was done use of feed forward neural network (NN) modeling methodology. Experimental results for ISCAS benchmark circuits show an RMS error of 0.102 for the shortest path length complexity estimation predicted by the NN model (NNM). Use of such a model can help reduce the time complexity of very large scale integrated (VLSI) circuitries and related computer-aided design (CAD) tools that use BDDs.
Tele-Operated Anthropomorphic Arm and Hand Design
In this project, a tele-operated anthropomorphic robotic arm and hand is designed and built as a versatile robotic arm system. The robot has the ability to manipulate objects such as pick and place operations. It is also able to function by itself, in standalone mode. Firstly, the robotic arm is built in order to interface with a personal computer via a serial servo controller circuit board. The circuit board enables user to completely control the robotic arm and moreover, enables feedbacks from user. The control circuit board uses a powerful integrated microcontroller, a PIC (Programmable Interface Controller). The PIC is firstly programmed using BASIC (Beginner-s All-purpose Symbolic Instruction Code) and it is used as the 'brain' of the robot. In addition a user friendly Graphical User Interface (GUI) is developed as the serial servo interface software using Microsoft-s Visual Basic 6. The second part of the project is to use speech recognition control on the robotic arm. A speech recognition circuit board is constructed with onboard components such as PIC and other integrated circuits. It replaces the computers- Graphical User Interface. The robotic arm is able to receive instructions as spoken commands through a microphone and perform operations with respect to the commands such as picking and placing operations.
A Fuzzy Implementation for Optimization of Storage Locations in an Industrial AS/RS
Warehousing is commonly used in factories for the storage of products until delivery of orders. As the amount of products stored increases it becomes tedious to be carried out manually. In recent years, the manual storing has converted into fully or partially computer controlled systems, also known as Automated Storage and Retrieval Systems (AS/RS). This paper discusses an ASRS system, which was designed such that the best storage location for the products is determined by utilizing a fuzzy control system. The design maintains the records of the products to be/already in store and the storage/retrieval times along with the availability status of the storage locations. This paper discusses on the maintenance of the above mentioned records and the utilization of the concept of fuzzy logic in order to determine the optimum storage location for the products. The paper will further discuss on the dynamic splitting and merging of the storage locations depending on the product sizes.
Handwritten Character Recognition Using Multiscale Neural Network Training Technique
Advancement in Artificial Intelligence has lead to the developments of various “smart" devices. Character recognition device is one of such smart devices that acquire partial human intelligence with the ability to capture and recognize various characters in different languages. Firstly multiscale neural training with modifications in the input training vectors is adopted in this paper to acquire its advantage in training higher resolution character images. Secondly selective thresholding using minimum distance technique is proposed to be used to increase the level of accuracy of character recognition. A simulator program (a GUI) is designed in such a way that the characters can be located on any spot on the blank paper in which the characters are written. The results show that such methods with moderate level of training epochs can produce accuracies of at least 85% and more for handwritten upper case English characters and numerals.
Evolutionary Search Techniques to Solve Set Covering Problems
Set covering problem is a classical problem in computer science and complexity theory. It has many applications, such as airline crew scheduling problem, facilities location problem, vehicle routing, assignment problem, etc. In this paper, three different techniques are applied to solve set covering problem. Firstly, a mathematical model of set covering problem is introduced and solved by using optimization solver, LINGO. Secondly, the Genetic Algorithm Toolbox available in MATLAB is used to solve set covering problem. And lastly, an ant colony optimization method is programmed in MATLAB programming language. Results obtained from these methods are presented in tables. In order to assess the performance of the techniques used in this project, the benchmark problems available in open literature are used.
Solving Machine Loading Problem in Flexible Manufacturing Systems Using Particle Swarm Optimization
In this paper, a particle swarm optimization (PSO) algorithm is proposed to solve machine loading problem in flexible manufacturing system (FMS), with bicriterion objectives of minimizing system unbalance and maximizing system throughput in the occurrence of technological constraints such as available machining time and tool slots. A mathematical model is used to select machines, assign operations and the required tools. The performance of the PSO is tested by using 10 sample dataset and the results are compared with the heuristics reported in the literature. The results support that the proposed PSO is comparable with the algorithms reported in the literature.
Kinematics and Control System Design of Manipulators for a Humanoid Robot
In this work, a new approach is proposed to control the manipulators for Humanoid robot. The kinematics of the manipulators in terms of joint positions, velocity, acceleration and torque of each joint is computed using the Denavit Hardenberg (D-H) notations. These variables are used to design the manipulator control system, which has been proposed in this work. In view of supporting the development of a controller, a simulation of the manipulator is designed for Humanoid robot. This simulation is developed through the use of the Virtual Reality Toolbox and Simulink in Matlab. The Virtual Reality Toolbox in Matlab provides the interfacing and controls to an environment which is developed based on the Virtual Reality Modeling Language (VRML). Chains of bones were used to represent the robot.
Bio-mechanical Analysis of Human Joints and Extension of the Study to Robot

In this paper, the bio-mechanical analysis of human joints is carried out and the study is extended to the robot manipulator. This study will first focus on the kinematics of human arm which include the movement of each joint in shoulder, wrist, elbow and finger complexes. Those analyses are then extended to the design of a human robot manipulator. A simulator is built for Direct Kinematics and Inverse Kinematics of human arm. In the simulation of Direct Kinematics, the human joint angles can be inserted, while the position and orientation of each finger tips (end-effector) are shown. Inverse Kinematics does the reverse of the Direct Kinematics. Based on previous materials obtained from kinematics analysis, the human manipulator joints can be designed to follow prescribed position trajectories.

Vol:11 No:04 2017Vol:11 No:03 2017Vol:11 No:02 2017Vol:11 No:01 2017
Vol:10 No:12 2016Vol:10 No:11 2016Vol:10 No:10 2016Vol:10 No:09 2016Vol:10 No:08 2016Vol:10 No:07 2016Vol:10 No:06 2016Vol:10 No:05 2016Vol:10 No:04 2016Vol:10 No:03 2016Vol:10 No:02 2016Vol:10 No:01 2016
Vol:9 No:12 2015Vol:9 No:11 2015Vol:9 No:10 2015Vol:9 No:09 2015Vol:9 No:08 2015Vol:9 No:07 2015Vol:9 No:06 2015Vol:9 No:05 2015Vol:9 No:04 2015Vol:9 No:03 2015Vol:9 No:02 2015Vol:9 No:01 2015
Vol:8 No:12 2014Vol:8 No:11 2014Vol:8 No:10 2014Vol:8 No:09 2014Vol:8 No:08 2014Vol:8 No:07 2014Vol:8 No:06 2014Vol:8 No:05 2014Vol:8 No:04 2014Vol:8 No:03 2014Vol:8 No:02 2014Vol:8 No:01 2014
Vol:7 No:12 2013Vol:7 No:11 2013Vol:7 No:10 2013Vol:7 No:09 2013Vol:7 No:08 2013Vol:7 No:07 2013Vol:7 No:06 2013Vol:7 No:05 2013Vol:7 No:04 2013Vol:7 No:03 2013Vol:7 No:02 2013Vol:7 No:01 2013
Vol:6 No:12 2012Vol:6 No:11 2012Vol:6 No:10 2012Vol:6 No:09 2012Vol:6 No:08 2012Vol:6 No:07 2012Vol:6 No:06 2012Vol:6 No:05 2012Vol:6 No:04 2012Vol:6 No:03 2012Vol:6 No:02 2012Vol:6 No:01 2012
Vol:5 No:12 2011Vol:5 No:11 2011Vol:5 No:10 2011Vol:5 No:09 2011Vol:5 No:08 2011Vol:5 No:07 2011Vol:5 No:06 2011Vol:5 No:05 2011Vol:5 No:04 2011Vol:5 No:03 2011Vol:5 No:02 2011Vol:5 No:01 2011
Vol:4 No:12 2010Vol:4 No:11 2010Vol:4 No:10 2010Vol:4 No:09 2010Vol:4 No:08 2010Vol:4 No:07 2010Vol:4 No:06 2010Vol:4 No:05 2010Vol:4 No:04 2010Vol:4 No:03 2010Vol:4 No:02 2010Vol:4 No:01 2010
Vol:3 No:12 2009Vol:3 No:11 2009Vol:3 No:10 2009Vol:3 No:09 2009Vol:3 No:08 2009Vol:3 No:07 2009Vol:3 No:06 2009Vol:3 No:05 2009Vol:3 No:04 2009Vol:3 No:03 2009Vol:3 No:02 2009Vol:3 No:01 2009
Vol:2 No:12 2008Vol:2 No:11 2008Vol:2 No:10 2008Vol:2 No:09 2008Vol:2 No:08 2008Vol:2 No:07 2008Vol:2 No:06 2008Vol:2 No:05 2008Vol:2 No:04 2008Vol:2 No:03 2008Vol:2 No:02 2008Vol:2 No:01 2008
Vol:1 No:12 2007Vol:1 No:11 2007Vol:1 No:10 2007Vol:1 No:09 2007Vol:1 No:08 2007Vol:1 No:07 2007Vol:1 No:06 2007Vol:1 No:05 2007Vol:1 No:04 2007Vol:1 No:03 2007Vol:1 No:02 2007Vol:1 No:01 2007