Excellence in Research and Innovation for Humanity

International Science Index

Commenced in January 1999 Frequency: Monthly Edition: International Paper Count: 155

A Modified Spiral Search Algorithm and Its Embedded System Architecture Design

One of the most growing areas in the embedded community is multimedia devices. Multimedia devices incorporate a number of complicated functions for their operation, like motion estimation. A multitude of different implementations have been proposed to reduce motion estimation complexity, such as spiral search. We have studied the implementations of spiral search and identified areas of improvement. We propose a modified spiral search algorithm, with lower computational complexity compared to the original spiral search. We have implemented our algorithm on an embedded ARM based architecture, with custom memory hierarchy. The resulting system yields energy consumption reduction up to 64% and performance increase up to 77%, with a small penalty of 2.3 dB, in average, of video quality compared with the original spiral search algorithm.

Using a Trust-Based Environment Key for Mobile Agent Code Protection
Human activities are increasingly based on the use of remote resources and services, and on the interaction between remotely located parties that may know little about each other. Mobile agents must be prepared to execute on different hosts with various environmental security conditions. The aim of this paper is to propose a trust based mechanism to improve the security of mobile agents and allow their execution in various environments. Thus, an adaptive trust mechanism is proposed. It is based on the dynamic interaction between the agent and the environment. Information collected during the interaction enables generation of an environment key. This key informs on the host-s trust degree and permits the mobile agent to adapt its execution. Trust estimation is based on concrete parameters values. Thus, in case of distrust, the source of problem can be located and a mobile agent appropriate behavior can be selected.
A Critical Review of the Adequacy of EIA Reports-Evidence from Pakistan

The preparation of good-quality Environmental Impact Assessment (EIA) reports contribute to enhancing overall effectiveness of EIA. This component of the EIA process becomes more important in situation where public participation is weak and there is lack of expertise on the part of the competent authority. In Pakistan, EIA became mandatory for every project likely to cause adverse environmental impacts from July 1994. The competent authority also formulated guidelines for preparation and review of EIA reports in 1997. However, EIA is yet to prove as a successful decision support tool to help in environmental protection. One of the several reasons of this ineffectiveness is the generally poor quality of EIA reports. This paper critically reviews EIA reports of some randomly selected projects. Interviews of EIA consultants, project proponents and concerned government officials have also been conducted to underpin the root causes of poor quality of EIA reports. The analysis reveals several inadequacies particularly in areas relating to identification, evaluation and mitigation of key impacts and consideration of alternatives. The paper identifies some opportunities and suggests measures for improving the quality of EIA reports and hence making EIA an effective tool to help in environmental protection.

Momentum and Heat Transfer in the Flow of a Viscoelastic Fluid Past a Porous Flat Plate Subject to Suction or Blowing

An analysis is made of the flow of an incompressible viscoelastic fluid (of small memory) over a porous plate subject to suction or blowing. It is found that velocity at a point increases with increase in the elasticity in the fluid. It is also shown that wall shear stress depends only on suction and is also independent of the material of fluids. No steady solution for velocity distribution exists when there is blowing at the plate. Temperature distribution in the boundary layer is determined and it is found that temperature at a point decreases with increase in the elasticity in the fluid.

Evolution, Tendencies and Impact of Standardization of Input/Output Platforms in Full Scale Simulators for Training Power Plant Operators

This article presents the evolution and technological changes implemented on the full scale simulators developed by the Simulation Department of the Instituto de Investigaciones Eléctricas1 (Mexican Electric Research Institute) and located at different training centers around the Mexican territory, and allows US to know the last updates, basically from the input/output view point, of the current simulators at some facilities of the electrical sector as well as the compatible industry of the electrical manufactures and industries such as Comision Federal de Electricidad (CFE*, The utility Mexican company). Tendencies of these developments and impact within the operators- scope are also presented.

Application of Load Transfer Technique for Distribution Power Flow Analysis
Installation of power compensation equipment in some cases places additional buses into the system. Therefore, a total number of power flow equations and voltage unknowns increase due to additional locations of installed devices. In this circumstance, power flow calculation is more complicated. It may result in a computational convergence problem. This paper presents a power flow calculation by using Newton-Raphson iterative method together with the proposed load transfer technique. This concept is to eliminate additional buses by transferring installed loads at the new buses to existing two adjacent buses. Thus, the total number of power flow equations is not changed. The overall computational speed is expectedly shorter than that of solving the problem without applying the load transfer technique. A 15-bus test system is employed for test to evaluate the effectiveness of the proposed load transfer technique. As a result, the total number of iteration required and execution time is significantly reduced.
Automation System for Optimization of Electrical and Thermal Energy Production in Cogenerative Gas Power Plants
The system is made with main distributed components: First Level: Industrial Computers placed in Control Room (monitors thermal and electrical processes based on the data provided by the second level); Second Level: PLCs which collects data from process and transmits information on the first level; also takes commands from this level which are further, passed to execution elements from third level; Third Level: field elements consisting in 3 categories: data collecting elements; data transfer elements from the third level to the second; execution elements which take commands from the second level PLCs and executes them after which transmits the confirmation of execution to them. The purpose of the automatic functioning is the optimization of the co-generative electrical energy commissioning in the national energy system and the commissioning of thermal energy to the consumers. The integrated system treats the functioning of all the equipments and devices as a whole: Gas Turbine Units (GTU); MT 20kV Medium Voltage Station (MVS); 0,4 kV Low Voltage Station (LVS); Main Hot Water Boilers (MHW); Auxiliary Hot Water Boilers (AHW); Gas Compressor Unit (GCU); Thermal Agent Circulation Pumping Unit (TPU); Water Treating Station (WTS).
Interactive Fuzzy Multi-objective Programming in Land Re-organisational Planning for Sustainable Rural Development
Sustainability in rural production system can only be achieved if it can suitably satisfy the local requirement as well as the outside demand with the changing time. With the increased pressure from the food sector in a globalised world, the agrarian economy needs to re-organise its cultivable land system to be compatible with new management practices as well as the multiple needs of various stakeholders and the changing resource scenario. An attempt has been made to transform this problem into a multi-objective decisionmaking problem considering various objectives, resource constraints and conditional constraints. An interactive fuzzy multi-objective programming approach has been used for such a purpose taking a case study in Indian context to demonstrate the validity of the method.
Adaptive Bidirectional Flow for Image Interpolation and Enhancement

Image interpolation is a common problem in imaging applications. However, most interpolation algorithms in existence suffer visually the effects of blurred edges and jagged artifacts in the image to some extent. This paper presents an adaptive feature preserving bidirectional flow process, where an inverse diffusion is performed to sharpen edges along the normal directions to the isophote lines (edges), while a normal diffusion is done to remove artifacts (“jaggies") along the tangent directions. In order to preserve image features such as edges, corners and textures, the nonlinear diffusion coefficients are locally adjusted according to the directional derivatives of the image. Experimental results on synthetic images and nature images demonstrate that our interpolation algorithm substantially improves the subjective quality of the interpolated images over conventional interpolations.

Adaptive Anisotropic Diffusion for Ultrasonic Image Denoising and Edge Enhancement

Utilizing echoic intension and distribution from different organs and local details of human body, ultrasonic image can catch important medical pathological changes, which unfortunately may be affected by ultrasonic speckle noise. A feature preserving ultrasonic image denoising and edge enhancement scheme is put forth, which includes two terms: anisotropic diffusion and edge enhancement, controlled by the optimum smoothing time. In this scheme, the anisotropic diffusion is governed by the local coordinate transformation and the first and the second order normal derivatives of the image, while the edge enhancement is done by the hyperbolic tangent function. Experiments on real ultrasonic images indicate effective preservation of edges, local details and ultrasonic echoic bright strips on denoising by our scheme.

Multiwavelet and Biological Signal Processing

In this paper we are to find the optimum multiwavelet for compression of electrocardiogram (ECG) signals and then, selecting it for using with SPIHT codec. At present, it is not well known which multiwavelet is the best choice for optimum compression of ECG. In this work, we examine different multiwavelets on 24 sets of ECG data with entirely different characteristics, selected from MIT-BIH database. For assessing the functionality of the different multiwavelets in compressing ECG signals, in addition to known factors such as Compression Ratio (CR), Percent Root Difference (PRD), Distortion (D), Root Mean Square Error (RMSE) in compression literature, we also employed the Cross Correlation (CC) criterion for studying the morphological relations between the reconstructed and the original ECG signal and Signal to reconstruction Noise Ratio (SNR). The simulation results show that the Cardinal Balanced Multiwavelet (cardbal2) by the means of identity (Id) prefiltering method to be the best effective transformation. After finding the most efficient multiwavelet, we apply SPIHT coding algorithm on the transformed signal by this multiwavelet.

A Web Oriented Spread Spectrum Watermarking Procedure for MPEG-2 Videos
In the last decade digital watermarking procedures have become increasingly applied to implement the copyright protection of multimedia digital contents distributed on the Internet. To this end, it is worth noting that a lot of watermarking procedures for images and videos proposed in literature are based on spread spectrum techniques. However, some scepticism about the robustness and security of such watermarking procedures has arisen because of some documented attacks which claim to render the inserted watermarks undetectable. On the other hand, web content providers wish to exploit watermarking procedures characterized by flexible and efficient implementations and which can be easily integrated in their existing web services frameworks or platforms. This paper presents how a simple spread spectrum watermarking procedure for MPEG-2 videos can be modified to be exploited in web contexts. To this end, the proposed procedure has been made secure and robust against some well-known and dangerous attacks. Furthermore, its basic scheme has been optimized by making the insertion procedure adaptive with respect to the terminals used to open the videos and the network transactions carried out to deliver them to buyers. Finally, two different implementations of the procedure have been developed: the former is a high performance parallel implementation, whereas the latter is a portable Java and XML based implementation. Thus, the paper demonstrates that a simple spread spectrum watermarking procedure, with limited and appropriate modifications to the embedding scheme, can still represent a valid alternative to many other well-known and more recent watermarking procedures proposed in literature.
Improved Automated Classification of Alcoholics and Non-alcoholics
In this paper, several improvements are proposed to previous work of automated classification of alcoholics and nonalcoholics. In the previous paper, multiplayer-perceptron neural network classifying energy of gamma band Visual Evoked Potential (VEP) signals gave the best classification performance using 800 VEP signals from 10 alcoholics and 10 non-alcoholics. Here, the dataset is extended to include 3560 VEP signals from 102 subjects: 62 alcoholics and 40 non-alcoholics. Three modifications are introduced to improve the classification performance: i) increasing the gamma band spectral range by increasing the pass-band width of the used filter ii) the use of Multiple Signal Classification algorithm to obtain the power of the dominant frequency in gamma band VEP signals as features and iii) the use of the simple but effective knearest neighbour classifier. To validate that these two modifications do give improved performance, a 10-fold cross validation classification (CVC) scheme is used. Repeat experiments of the previously used methodology for the extended dataset are performed here and improvement from 94.49% to 98.71% in maximum averaged CVC accuracy is obtained using the modifications. This latest results show that VEP based classification of alcoholics is worth exploring further for system development.
An ICA Algorithm for Separation of Convolutive Mixture of Speech Signals

This paper describes Independent Component Analysis (ICA) based fixed-point algorithm for the blind separation of the convolutive mixture of speech, picked-up by a linear microphone array. The proposed algorithm extracts independent sources by non- Gaussianizing the Time-Frequency Series of Speech (TFSS) in a deflationary way. The degree of non-Gaussianization is measured by negentropy. The relative performances of algorithm under random initialization and Null beamformer (NBF) based initialization are studied. It has been found that an NBF based initial value gives speedy convergence as well as better separation performance

A Survey of Business Component Identification Methods and Related Techniques
With deep development of software reuse, componentrelated technologies have been widely applied in the development of large-scale complex applications. Component identification (CI) is one of the primary research problems in software reuse, by analyzing domain business models to get a set of business components with high reuse value and good reuse performance to support effective reuse. Based on the concept and classification of CI, its technical stack is briefly discussed from four views, i.e., form of input business models, identification goals, identification strategies, and identification process. Then various CI methods presented in literatures are classified into four types, i.e., domain analysis based methods, cohesion-coupling based clustering methods, CRUD matrix based methods, and other methods, with the comparisons between these methods for their advantages and disadvantages. Additionally, some insufficiencies of study on CI are discussed, and the causes are explained subsequently. Finally, it is concluded with some significantly promising tendency about research on this problem.
An Adversarial Construction of Instability Bounds in LIS Networks
In this work, we study the impact of dynamically changing link slowdowns on the stability properties of packetswitched networks under the Adversarial Queueing Theory framework. Especially, we consider the Adversarial, Quasi-Static Slowdown Queueing Theory model, where each link slowdown may take on values in the two-valued set of integers {1, D} with D > 1 which remain fixed for a long time, under a (w, ¤ü)-adversary. In this framework, we present an innovative systematic construction for the estimation of adversarial injection rate lower bounds, which, if exceeded, cause instability in networks that use the LIS (Longest-in- System) protocol for contention-resolution. In addition, we show that a network that uses the LIS protocol for contention-resolution may result in dropping its instability bound at injection rates ¤ü > 0 when the network size and the high slowdown D take large values. This is the best ever known instability lower bound for LIS networks.
MONPAR - A Page Replacement Algorithm for a Spatiotemporal Database
For a spatiotemporal database management system, I/O cost of queries and other operations is an important performance criterion. In order to optimize this cost, an intense research on designing robust index structures has been done in the past decade. With these major considerations, there are still other design issues that deserve addressing due to their direct impact on the I/O cost. Having said this, an efficient buffer management strategy plays a key role on reducing redundant disk access. In this paper, we proposed an efficient buffer strategy for a spatiotemporal database index structure, specifically indexing objects moving over a network of roads. The proposed strategy, namely MONPAR, is based on the data type (i.e. spatiotemporal data) and the structure of the index structure. For the purpose of an experimental evaluation, we set up a simulation environment that counts the number of disk accesses while executing a number of spatiotemporal range-queries over the index. We reiterated simulations with query sets with different distributions, such as uniform query distribution and skewed query distribution. Based on the comparison of our strategy with wellknown page-replacement techniques, like LRU-based and Prioritybased buffers, we conclude that MONPAR behaves better than its competitors for small and medium size buffers under all used query-distributions.
A Metric-Set and Model Suggestion for Better Software Project Cost Estimation
Software project effort estimation is frequently seen as complex and expensive for individual software engineers. Software production is in a crisis. It suffers from excessive costs. Software production is often out of control. It has been suggested that software production is out of control because we do not measure. You cannot control what you cannot measure. During last decade, a number of researches on cost estimation have been conducted. The metric-set selection has a vital role in software cost estimation studies; its importance has been ignored especially in neural network based studies. In this study we have explored the reasons of those disappointing results and implemented different neural network models using augmented new metrics. The results obtained are compared with previous studies using traditional metrics. To be able to make comparisons, two types of data have been used. The first part of the data is taken from the Constructive Cost Model (COCOMO'81) which is commonly used in previous studies and the second part is collected according to new metrics in a leading international company in Turkey. The accuracy of the selected metrics and the data samples are verified using statistical techniques. The model presented here is based on Multi-Layer Perceptron (MLP). Another difficulty associated with the cost estimation studies is the fact that the data collection requires time and care. To make a more thorough use of the samples collected, k-fold, cross validation method is also implemented. It is concluded that, as long as an accurate and quantifiable set of metrics are defined and measured correctly, neural networks can be applied in software cost estimation studies with success
Lithofacies Classification from Well Log Data Using Neural Networks, Interval Neutrosophic Sets and Quantification of Uncertainty

This paper proposes a novel approach to the question of lithofacies classification based on an assessment of the uncertainty in the classification results. The proposed approach has multiple neural networks (NN), and interval neutrosophic sets (INS) are used to classify the input well log data into outputs of multiple classes of lithofacies. A pair of n-class neural networks are used to predict n-degree of truth memberships and n-degree of false memberships. Indeterminacy memberships or uncertainties in the predictions are estimated using a multidimensional interpolation method. These three memberships form the INS used to support the confidence in results of multiclass classification. Based on the experimental data, our approach improves the classification performance as compared to an existing technique applied only to the truth membership. In addition, our approach has the capability to provide a measure of uncertainty in the problem of multiclass classification.

Coalescing Data Marts
OLAP uses multidimensional structures, to provide access to data for analysis. Traditionally, OLAP operations are more focused on retrieving data from a single data mart. An exception is the drill across operator. This, however, is restricted to retrieving facts on common dimensions of the multiple data marts. Our concern is to define further operations while retrieving data from multiple data marts. Towards this, we have defined six operations which coalesce data marts. While doing so we consider the common as well as the non-common dimensions of the data marts.
A Security Analysis for Home Gateway Architectures
Providing Services at Home has become over the last few years a very dynamic and promising technological domain. It is likely to enable wide dissemination of secure and automated living environments. We propose a methodology for identifying threats to Services at Home Delivery systems, as well as a threat analysis of a multi-provider Home Gateway architecture. This methodology is based on a dichotomous positive/preventive study of the target system: it aims at identifying both what the system must do, and what it must not do. This approach completes existing methods with a synthetic view of potential security flaws, thus enabling suitable measures to be taken into account. Security implications of the evolution of a given system become easier to deal with. A prototype is built based on the conclusions of this analysis.
Performance Evaluation of Data Transfer Protocol GridFTP for Grid Computing
In Grid computing, a data transfer protocol called GridFTP has been widely used for efficiently transferring a large volume of data. Currently, two versions of GridFTP protocols, GridFTP version 1 (GridFTP v1) and GridFTP version 2 (GridFTP v2), have been proposed in the GGF. GridFTP v2 supports several advanced features such as data streaming, dynamic resource allocation, and checksum transfer, by defining a transfer mode called X-block mode. However, in the literature, effectiveness of GridFTP v2 has not been fully investigated. In this paper, we therefore quantitatively evaluate performance of GridFTP v1 and GridFTP v2 using mathematical analysis and simulation experiments. We reveal the performance limitation of GridFTP v1, and quantitatively show effectiveness of GridFTP v2. Through several numerical examples, we show that by utilizing the data streaming feature, the average file transfer time of GridFTP v2 is significantly smaller than that of GridFTP v1.
Expressive Modes and Species of Language
Computer languages are usually lumped together into broad -paradigms-, leaving us in want of a finer classification of kinds of language. Theories distinguishing between -genuine differences- in language has been called for, and we propose that such differences can be observed through a notion of expressive mode. We outline this concept, propose how it could be operationalized and indicate a possible context for the development of a corresponding theory. Finally we consider a possible application in connection with evaluation of language revision. We illustrate this with a case, investigating possible revisions of the relational algebra in order to overcome weaknesses of the division operator in connection with universal queries.
Agents Network on a Grid: An Approach with the Set of Circulant Operators
In this work we present some matrix operators named circulant operators and their action on square matrices. This study on square matrices provides new insights into the structure of the space of square matrices. Moreover it can be useful in various fields as in agents networking on Grid or large-scale distributed self-organizing grid systems.
Evaluation of Risk Attributes Driven by Periodically Changing System Functionality
Modeling of the distributed systems allows us to represent the whole its functionality. The working system instance rarely fulfils the whole functionality represented by model; usually some parts of this functionality should be accessible periodically. The reporting system based on the Data Warehouse concept seams to be an intuitive example of the system that some of its functionality is required only from time to time. Analyzing an enterprise risk associated with the periodical change of the system functionality, we should consider not only the inaccessibility of the components (object) but also their functions (methods), and the impact of such a situation on the system functionality from the business point of view. In the paper we suggest that the risk attributes should be estimated from risk attributes specified at the requirements level (Use Case in the UML model) on the base of the information about the structure of the model (presented at other levels of the UML model). We argue that it is desirable to consider the influence of periodical changes in requirements on the enterprise risk estimation. Finally, the proposition of such a solution basing on the UML system model is presented.
Interactive Model Based On an Extended CPN

The UML modeling of complex distributed systems often is a great challenge due to the large amount of parallel real-time operating components. In this paper the problems of verification of such systems are discussed. ECPN, an Extended Colored Petri Net is defined to formally describe state transitions of components and interactions among components. The relationship between sequence diagrams and Free Choice Petri Nets is investigated. Free Choice Petri Net theory helps verifying the liveness of sequence diagrams. By converting sequence diagrams to ECPNs and then comparing behaviors of sequence diagram ECPNs and statecharts, the consistency among models is analyzed. Finally, a verification process for an example model is demonstrated.

Extraction of Temporal Relation by the Creation of Historical Natural Disaster Archive
In historical science and social science, the influence of natural disaster upon society is a matter of great interest. In recent years, some archives are made through many hands for natural disasters, however it is inefficiency and waste. So, we suppose a computer system to create a historical natural disaster archive. As the target of this analysis, we consider newspaper articles. The news articles are considered to be typical examples that prescribe the temporal relations of affairs for natural disaster. In order to do this analysis, we identify the occurrences in newspaper articles by some index entries, considering the affairs which are specific to natural disasters, and show the temporal relation between natural disasters. We designed and implemented the automatic system of “extraction of the occurrences of natural disaster" and “temporal relation table for natural disaster."
Measuring the Comprehensibility of a UML-B Model and a B Model

Software maintenance, which involves making enhancements, modifications and corrections to existing software systems, consumes more than half of developer time. Specification comprehensibility plays an important role in software maintenance as it permits the understanding of the system properties more easily and quickly. The use of formal notation such as B increases a specification-s precision and consistency. However, the notation is regarded as being difficult to comprehend. Semi-formal notation such as the Unified Modelling Language (UML) is perceived as more accessible but it lacks formality. Perhaps by combining both notations could produce a specification that is not only accurate and consistent but also accessible to users. This paper presents an experiment conducted on a model that integrates the use of both UML and B notations, namely UML-B, versus a B model alone. The objective of the experiment was to evaluate the comprehensibility of a UML-B model compared to a traditional B model. The measurement used in the experiment focused on the efficiency in performing the comprehension tasks. The experiment employed a cross-over design and was conducted on forty-one subjects, including undergraduate and masters students. The results show that the notation used in the UML-B model is more comprehensible than the B model.

A New Method in Detection of Ceramic Tiles Color Defects Using Genetic C-Means Algorithm

In this paper an algorithm is used to detect the color defects of ceramic tiles. First the image of a normal tile is clustered using GCMA; Genetic C-means Clustering Algorithm; those results in best cluster centers. C-means is a common clustering algorithm which optimizes an objective function, based on a measure between data points and the cluster centers in the data space. Here the objective function describes the mean square error. After finding the best centers, each pixel of the image is assigned to the cluster with closest cluster center. Then, the maximum errors of clusters are computed. For each cluster, max error is the maximum distance between its center and all the pixels which belong to it. After computing errors all the pixels of defected tile image are clustered based on the centers obtained from normal tile image in previous stage. Pixels which their distance from their cluster center is more than the maximum error of that cluster are considered as defected pixels.

A Support System Applicable to Multiple APIs for Haptic VR Application Designers
This paper describes a proposed support system which enables applications designers to effectively create VR applications using multiple haptic APIs. When the VR designers create applications, it is often difficult to handle and understand many parameters and functions that have to be set in the application program using documentation manuals only. This complication may disrupt creative imagination and result in inefficient coding. So, we proposed the support application which improved the efficiency of VR applications development and provided the interactive components of confirmation of operations with haptic sense previously. In this paper, we describe improvements of our former proposed support application, which was applicable to multiple APIs and haptic devices, and evaluate the new application by having participants complete VR program. Results from a preliminary experiment suggest that our application facilitates creation of VR applications.
Shot Transition Detection with Minimal Decoding of MPEG Video Streams
Digital libraries become more and more necessary in order to support users with powerful and easy-to-use tools for searching, browsing and retrieving media information. The starting point for these tasks is the segmentation of video content into shots. To segment MPEG video streams into shots, a fully automatic procedure to detect both abrupt and gradual transitions (dissolve and fade-groups) with minimal decoding in real time is developed in this study. Each was explored through two phases: macro-block type's analysis in B-frames, and on-demand intensity information analysis. The experimental results show remarkable performance in detecting gradual transitions of some kinds of input data and comparable results of the rest of the examined video streams. Almost all abrupt transitions could be detected with very few false positive alarms.
Environmental Capacity and Sustainability of European Regional Airports: A Case Study
Airport capacity has always been perceived in the traditional sense as the number of aircraft operations during a specified time corresponding to a tolerable level of average delay and it mostly depends on the airside characteristics, on the fleet mix variability and on the ATM. The adoption of the Directive 2002/30/EC in the EU countries drives the stakeholders to conceive airport capacity in a different way though. Airport capacity in this sense is fundamentally driven by environmental criteria, and since acoustical externalities represent the most important factors, those are the ones that could pose a serious threat to the growth of airports and to aviation market itself in the short-medium term. The importance of the regional airports in the deregulated market grew fast during the last decade since they represent spokes for network carriers and a preferential destination for low-fares carriers. Not only regional airports have witnessed a fast and unexpected growth in traffic but also a fast growth in the complaints for the nuisance by the people living near those airports. In this paper the results of a study conducted in cooperation with the airport of Bologna G. Marconi are presented in order to investigate airport acoustical capacity as a defacto constraint of airport growth.
An Efficient Hardware Implementation of Extended and Fast Physical Addressing in Microprocessor-Based Systems Using Programmable Logic

This paper describes an efficient hardware implementation of a new technique for interfacing the data exchange between the microprocessor-based systems and the external devices. This technique, based on the use of software/hardware system and a reduced physical address, enlarges the interfacing capacity of the microprocessor-based systems, uses the Direct Memory Access (DMA) to increases the frequency of the new bus, and improves the speed of data exchange. While using this architecture in microprocessor-based system or in computer, the input of the hardware part of our system will be connected to the bus system, and the output, which is a new bus, will be connected to an external device. The new bus is composed of a data bus, a control bus and an address bus. A Xilinx Integrated Software Environment (ISE) 7.1i has been used for the programmable logic implementation.

Numerical Modeling of Gas Turbine Engines

In contrast to existing methods which do not take into account multiconnectivity in a broad sense of this term, we develop mathematical models and highly effective combination (BIEM and FDM) numerical methods of calculation of stationary and quasi-stationary temperature field of a profile part of a blade with convective cooling (from the point of view of realization on PC). The theoretical substantiation of these methods is proved by appropriate theorems. For it, converging quadrature processes have been developed and the estimations of errors in the terms of A.Ziqmound continuity modules have been received. For visualization of profiles are used: the method of the least squares with automatic conjecture, device spline, smooth replenishment and neural nets. Boundary conditions of heat exchange are determined from the solution of the corresponding integral equations and empirical relationships. The reliability of designed methods is proved by calculation and experimental investigations heat and hydraulic characteristics of the gas turbine first stage nozzle blade.

Specification of a Model of Honeypot Attack Based On Raised Data

The security of their network remains the priorities of almost all companies. Existing security systems have shown their limit; thus a new type of security systems was born: honeypots. Honeypots are defined as programs or intended servers which have to attract pirates to study theirs behaviours. It is in this context that the leurre.com project of gathering about twenty platforms was born. This article aims to specify a model of honeypots attack. Our model describes, on a given platform, the evolution of attacks according to theirs hours. Afterward, we show the most attacked services by the studies of attacks on the various ports. It is advisable to note that this article was elaborated within the framework of the research projects on honeyspots within the LABTIC (Laboratory of Information Technologies and Communication).

Analysis of a PWM Boost Inverter for Solar Home Application

Solar Cells are destined to supply electric energy beginning from primary resources. It can charge a battery up to 12V dc. For residential use an inverter for 12V dc to 220Vac conversion is desired. For this a static DC-AC converter is necessarily inserted between the solar cells and the distribution network. This paper describes a new P.W.M. strategy for a voltage source inverter. This modulation strategy reduces the energy losses and harmonics in the P.W.M. voltage source inverter. This technique allows the P.W.M. voltage source inverter to become a new feasible solution for solar home application.

A Secure Semi-Fragile Watermarking Scheme for Authentication and Recovery of Images Based On Wavelet Transform

Authentication of multimedia contents has gained much attention in recent times. In this paper, we propose a secure semi-fragile watermarking, with a choice of two watermarks to be embedded. This technique operates in integer wavelet domain and makes use of semi fragile watermarks for achieving better robustness. A self-recovering algorithm is employed, that hides the image digest into some Wavelet subbands to detect possible malevolent object manipulation undergone by the image (object replacing and/or deletion). The Semi-fragility makes the scheme tolerant for JPEG lossy compression as low as quality of 70%, and locate the tempered area accurately. In addition, the system ensures more security because the embedded watermarks are protected with private keys. The computational complexity is reduced using parameterized integer wavelet transform. Experimental results show that the proposed scheme guarantees the safety of watermark, image recovery and location of the tempered area accurately.

Dynamic Adaptability Using Reflexivity for Mobile Agent Protection

The paradigm of mobile agent provides a promising technology for the development of distributed and open applications. However, one of the main obstacles to widespread adoption of the mobile agent paradigm seems to be security. This paper treats the security of the mobile agent against malicious host attacks. It describes generic mobile agent protection architecture. The proposed approach is based on the dynamic adaptability and adopts the reflexivity as a model of conception and implantation. In order to protect it against behaviour analysis attempts, the suggested approach supplies the mobile agent with a flexibility faculty allowing it to present an unexpected behaviour. Furthermore, some classical protective mechanisms are used to reinforce the level of security.

Some Computational Results on MPI Parallel Implementation of Dense Simplex Method
There are two major variants of the Simplex Algorithm: the revised method and the standard, or tableau method. Today, all serious implementations are based on the revised method because it is more efficient for sparse linear programming problems. Moreover, there are a number of applications that lead to dense linear problems so our aim in this paper is to present some computational results on parallel implementation of dense Simplex Method. Our implementation is implemented on a SMP cluster using C programming language and the Message Passing Interface MPI. Preliminary computational results on randomly generated dense linear programs support our results.
A Hybrid Particle Swarm Optimization Solution to Ramping Rate Constrained Dynamic Economic Dispatch
This paper presents the application of an enhanced Particle Swarm Optimization (EPSO) combined with Gaussian Mutation (GM) for solving the Dynamic Economic Dispatch (DED) problem considering the operating constraints of generators. The EPSO consists of the standard PSO and a modified heuristic search approaches. Namely, the ability of the traditional PSO is enhanced by applying the modified heuristic search approach to prevent the solutions from violating the constraints. In addition, Gaussian Mutation is aimed at increasing the diversity of global search, whilst it also prevents being trapped in suboptimal points during search. To illustrate its efficiency and effectiveness, the developed EPSO-GM approach is tested on the 3-unit and 10-unit 24-hour systems considering valve-point effect. From the experimental results, it can be concluded that the proposed EPSO-GM provides, the accurate solution, the efficiency, and the feature of robust computation compared with other algorithms under consideration.
A New Similarity Measure Based On Edge Counting

In the field of concepts, the measure of Wu and Palmer [1] has the advantage of being simple to implement and have good performances compared to the other similarity measures [2]. Nevertheless, the Wu and Palmer measure present the following disadvantage: in some situations, the similarity of two elements of an IS-A ontology contained in the neighborhood exceeds the similarity value of two elements contained in the same hierarchy. This situation is inadequate within the information retrieval framework. To overcome this problem, we propose a new similarity measure based on the Wu and Palmer measure. Our objective is to obtain realistic results for concepts not located in the same way. The obtained results show that compared to the Wu and Palmer approach, our measure presents a profit in terms of relevance and execution time.

Architectural Stratification and Woody Species Diversity of a Subtropical Forest Grown in a Limestone Habitat in Okinawa Island, Japan
The forest stand consisted of four layers. The species composition between the third and the bottom layers was almost similar, whereas it was almost exclusive between the top and the lower three layers. The values of Shannon-s index H' and Pielou-s index J ' tended to increase from the bottom layer upward, except for H' -value of the top layer. The values of H' and J ' were 4.21 bit and 0.73, respectively, for the total stand. High woody species diversity of the forest depended on large trees in the upper layers, which trend was different from a subtropical evergreen broadleaf forest grown in silicate habitat in the northern part of Okinawa Island. The spatial distribution of trees was overlapped between the third and the bottom layers, whereas it was independent or slightly exclusive between the top and the lower three layers. Mean tree weight of each layer decreased from the top toward the bottom layer, whereas the corresponding tree density increased from the top downward. This relationship was analogous to the process of self-thinning plant populations.
Induction Motor Speed Control Using Fuzzy Logic Controller

Because of the low maintenance and robustness induction motors have many applications in the industries. The speed control of induction motor is more important to achieve maximum torque and efficiency. Various speed control techniques like, Direct Torque Control, Sensorless Vector Control and Field Oriented Control are discussed in this paper. Soft computing technique – Fuzzy logic is applied in this paper for the speed control of induction motor to achieve maximum torque with minimum loss. The fuzzy logic controller is implemented using the Field Oriented Control technique as it provides better control of motor torque with high dynamic performance. The motor model is designed and membership functions are chosen according to the parameters of the motor model. The simulated design is tested using various tool boxes in MATLAB. The result concludes that the efficiency and reliability of the proposed speed controller is good.

iDENTM Phones Automated Stress Testing
System testing is actually done to the entire system against the Functional Requirement Specification and/or the System Requirement Specification. Moreover, it is an investigatory testing phase, where the focus is to have almost a destructive attitude and test not only the design, but also the behavior and even the believed expectations of the customer. It is also intended to test up to and beyond the bounds defined in the software/hardware requirements specifications. In Motorola®, Automated Testing is one of the testing methodologies uses by GSG-iSGT (Global Software Group - iDEN TM Subcriber Group-Test) to increase the testing volume, productivity and reduce test cycle-time in iDEN TM phones testing. Testing is able to produce more robust products before release to the market. In this paper, iHopper is proposed as a tool to perform stress test on iDEN TM phonse. We will discuss the value that automation has brought to iDEN TM Phone testing such as improving software quality in the iDEN TM phone together with some metrics. We will also look into the advantages of the proposed system and some discussion of the future work as well.
Optimal Preventive Maintenance of the Reserve Source in the Industrial Electric Network
The great majority of the electric installations belong to the first and second category. In order to ensure a high level of reliability of their electric system feeder, two power supply sources are envisaged, one principal, the other of reserve, generally a cold reserve (electric diesel group). The principal source being under operation, its control can be ideal and sure, however for the reserve source being in stop, a preventive maintenance-s which proceeds on time intervals (periodicity) and for well defined lengths of time are envisaged, so that this source will always available in case of the principal source failure. The choice of the periodicity of preventive maintenance of the source of reserve influences directly the reliability of the electric feeder system. On the basis of the semi-markovians processes, the influence of the periodicity of the preventive maintenance of the source of reserve is studied and is given the optimal periodicity.
Optimal Data Compression and Filtering: The Case of Infinite Signal Sets

We present a theory for optimal filtering of infinite sets of random signals. There are several new distinctive features of the proposed approach. First, we provide a single optimal filter for processing any signal from a given infinite signal set. Second, the filter is presented in the special form of a sum with p terms where each term is represented as a combination of three operations. Each operation is a special stage of the filtering aimed at facilitating the associated numerical work. Third, an iterative scheme is implemented into the filter structure to provide an improvement in the filter performance at each step of the scheme. The final step of the concerns signal compression and decompression. This step is based on the solution of a new rank-constrained matrix approximation problem. The solution to the matrix problem is described in this paper. A rigorous error analysis is given for the new filter.

Interest of the Sequences Pseudo Noises Codes of Different Lengths for the Reduction from the Interference between Users of CDMA Network
The third generation (3G) of cellular system adopted the spread spectrum as solution for the transmission of the data in the physical layer. Contrary to systems IS-95 or CDMAOne (systems with spread spectrum of the preceding generation), the new standard, called Universal Mobil Telecommunications System (UMTS), uses long codes in the down link. The system is conceived for the vocal communication and the transmission of the data. In particular, the down link is very important, because of the asymmetrical request of the data, i.e., more remote loading towards the mobiles than towards the basic station. Moreover, the UMTS uses for the down link an orthogonal spreading out with a variable factor of spreading out (OVSF for Orthogonal Variable Spreading Factor). This characteristic makes it possible to increase the flow of data of one or more users by reducing their factor of spreading out without changing the factor of spreading out of other users. In the current standard of the UMTS, two techniques to increase the performances of the down link were proposed, the diversity of sending antenna and the codes space-time. These two techniques fight only fainding. The receiver proposed for the mobil station is the RAKE, but one can imagine a receiver more sophisticated, able to reduce the interference between users and the impact of the coloured noise and interferences to narrow band. In this context, where the users have long codes synchronized with variable factor of spreading out and ignorance by the mobile of the other active codes/users, the use of the sequences of code pseudo-noises different lengths is presented in the form of one of the most appropriate solutions.
Use of Fuzzy Edge Image in Block Truncation Coding for Image Compression
An image compression method has been developed using fuzzy edge image utilizing the basic Block Truncation Coding (BTC) algorithm. The fuzzy edge image has been validated with classical edge detectors on the basis of the results of the well-known Canny edge detector prior to applying to the proposed method. The bit plane generated by the conventional BTC method is replaced with the fuzzy bit plane generated by the logical OR operation between the fuzzy edge image and the corresponding conventional BTC bit plane. The input image is encoded with the block mean and standard deviation and the fuzzy bit plane. The proposed method has been tested with test images of 8 bits/pixel and size 512×512 and found to be superior with better Peak Signal to Noise Ratio (PSNR) when compared to the conventional BTC, and adaptive bit plane selection BTC (ABTC) methods. The raggedness and jagged appearance, and the ringing artifacts at sharp edges are greatly reduced in reconstructed images by the proposed method with the fuzzy bit plane.
A Unified Robust Algorithm for Detection of Human and Non-human Object in Intelligent Safety Application
This paper presents a general trainable framework for fast and robust upright human face and non-human object detection and verification in static images. To enhance the performance of the detection process, the technique we develop is based on the combination of fast neural network (FNN) and classical neural network (CNN). In FNN, a useful correlation is exploited to sustain high level of detection accuracy between input image and the weight of the hidden neurons. This is to enable the use of Fourier transform that significantly speed up the time detection. The combination of CNN is responsible to verify the face region. A bootstrap algorithm is used to collect non human object, which adds the false detection to the training process of the human and non-human object. Experimental results on test images with both simple and complex background demonstrate that the proposed method has obtained high detection rate and low false positive rate in detecting both human face and non-human object.
Mean-Square Performance of Adaptive Filter Algorithms in Nonstationary Environments
Employing a recently introduced unified adaptive filter theory, we show how the performance of a large number of important adaptive filter algorithms can be predicted within a general framework in nonstationary environment. This approach is based on energy conservation arguments and does not need to assume a Gaussian or white distribution for the regressors. This general performance analysis can be used to evaluate the mean square performance of the Least Mean Square (LMS) algorithm, its normalized version (NLMS), the family of Affine Projection Algorithms (APA), the Recursive Least Squares (RLS), the Data-Reusing LMS (DR-LMS), its normalized version (NDR-LMS), the Block Least Mean Squares (BLMS), the Block Normalized LMS (BNLMS), the Transform Domain Adaptive Filters (TDAF) and the Subband Adaptive Filters (SAF) in nonstationary environment. Also, we establish the general expressions for the steady-state excess mean square in this environment for all these adaptive algorithms. Finally, we demonstrate through simulations that these results are useful in predicting the adaptive filter performance.
New Nonlinear Filtering Strategies for Eliminating Short and Long Tailed Noise in Images with Edge Preservation Properties
Midpoint filter is quite effective in recovering the images confounded by the short-tailed (uniform) noise. It, however, performs poorly in the presence of additive long-tailed (impulse) noise and it does not preserve the edge structures of the image signals. Median smoother discards outliers (impulses) effectively, but it fails to provide adequate smoothing for images corrupted with nonimpulse noise. In this paper, two nonlinear techniques for image filtering, namely, New Filter I and New Filter II are proposed based on a nonlinear high-pass filter algorithm. New Filter I is constructed using a midpoint filter, a highpass filter and a combiner. It suppresses uniform noise quite well. New Filter II is configured using an alpha trimmed midpoint filter, a median smoother of window size 3x3, the high pass filter and the combiner. It is robust against impulse noise and attenuates uniform noise satisfactorily. Both the filters are shown to exhibit good response at the image boundaries (edges). The proposed filters are evaluated for their performance on a test image and the results obtained are included.
Analysis of Blind Decision Feedback Equalizer Convergence: Interest of a Soft Decision
In this paper the behavior of the decision feedback equalizers (DFEs) adapted by the decision-directed or the constant modulus blind algorithms is presented. An analysis of the error surface of the corresponding criterion cost functions is first developed. With the intention of avoiding the ill-convergence of the algorithm, the paper proposes to modify the shape of the cost function error surface by using a soft decision instead of the hard one. This was shown to reduce the influence of false decisions and to smooth the undesirable minima. Modified algorithms using the soft decision during a pseudo-training phase with an automatic switch to the properly tracking phase are then derived. Computer simulations show that these modified algorithms present better ability to avoid local minima than conventional ones.
Blind Non-Minimum Phase Channel Identification Using 3rd and 4th Order Cumulants
In this paper we propose a family of algorithms based on 3rd and 4th order cumulants for blind single-input single-output (SISO) Non-Minimum Phase (NMP) Finite Impulse Response (FIR) channel estimation driven by non-Gaussian signal. The input signal represents the signal used in 10GBASE-T (or IEEE 802.3an-2006) as a Tomlinson-Harashima Precoded (THP) version of random Pulse-Amplitude Modulation with 16 discrete levels (PAM-16). The proposed algorithms are tested using three non-minimum phase channel for different Signal-to-Noise Ratios (SNR) and for different data input length. Numerical simulation results are presented to illustrate the performance of the proposed algorithms.
Template-Based Object Detection through Partial Shape Matching and Boundary Verification
This paper presents a novel template-based method to detect objects of interest from real images by shape matching. To locate a target object that has a similar shape to a given template boundary, the proposed method integrates three components: contour grouping, partial shape matching, and boundary verification. In the first component, low-level image features, including edges and corners, are grouped into a set of perceptually salient closed contours using an extended ratio-contour algorithm. In the second component, we develop a partial shape matching algorithm to identify the fractions of detected contours that partly match given template boundaries. Specifically, we represent template boundaries and detected contours using landmarks, and apply a greedy algorithm to search the matched landmark subsequences. For each matched fraction between a template and a detected contour, we estimate an affine transform that transforms the whole template into a hypothetic boundary. In the third component, we provide an efficient algorithm based on oriented edge lists to determine the target boundary from the hypothetic boundaries by checking each of them against image edges. We evaluate the proposed method on recognizing and localizing 12 template leaves in a data set of real images with clutter back-grounds, illumination variations, occlusions, and image noises. The experiments demonstrate the high performance of our proposed method1.
Computer Generated Hologram for SemiFragile Watermarking with Encrypted Images
The protection of the contents of digital products is referred to as content authentication. In some applications, to be able to authenticate a digital product could be extremely essential. For example, if a digital product is used as a piece of evidence in the court, its integrity could mean life or death of the accused. Generally, the problem of content authentication can be solved using semifragile digital watermarking techniques. Recently many authors have proposed Computer Generated Hologram Watermarking (CGHWatermarking) techniques. Starting from these studies, in this paper a semi-fragile Computer Generated Hologram coding technique is proposed, which is able to detect malicious tampering while tolerating some incidental distortions. The proposed technique uses as watermark an encrypted image, and it is well suitable for digital image authentication.
Frame Texture Classification Method (FTCM) Applied on Mammograms for Detection of Abnormalities
Texture classification is an important image processing task with a broad application range. Many different techniques for texture classification have been explored. Using sparse approximation as a feature extraction method for texture classification is a relatively new approach, and Skretting et al. recently presented the Frame Texture Classification Method (FTCM), showing very good results on classical texture images. As an extension of that work the FTCM is here tested on a real world application as detection of abnormalities in mammograms. Some extensions to the original FTCM that are useful in some applications are implemented; two different smoothing techniques and a vector augmentation technique. Both detection of microcalcifications (as a primary detection technique and as a last stage of a detection scheme), and soft tissue lesions in mammograms are explored. All the results are interesting, and especially the results using FTCM on regions of interest as the last stage in a detection scheme for microcalcifications are promising.
Improved Closed Set Text-Independent Speaker Identification by Combining MFCC with Evidence from Flipped Filter Banks

A state of the art Speaker Identification (SI) system requires a robust feature extraction unit followed by a speaker modeling scheme for generalized representation of these features. Over the years, Mel-Frequency Cepstral Coefficients (MFCC) modeled on the human auditory system has been used as a standard acoustic feature set for SI applications. However, due to the structure of its filter bank, it captures vocal tract characteristics more effectively in the lower frequency regions. This paper proposes a new set of features using a complementary filter bank structure which improves distinguishability of speaker specific cues present in the higher frequency zone. Unlike high level features that are difficult to extract, the proposed feature set involves little computational burden during the extraction process. When combined with MFCC via a parallel implementation of speaker models, the proposed feature set outperforms baseline MFCC significantly. This proposition is validated by experiments conducted on two different kinds of public databases namely YOHO (microphone speech) and POLYCOST (telephone speech) with Gaussian Mixture Models (GMM) as a Classifier for various model orders.

Speech Enhancement by Marginal Statistical Characterization in the Log Gabor Wavelet Domain
This work presents a fusion of Log Gabor Wavelet (LGW) and Maximum a Posteriori (MAP) estimator as a speech enhancement tool for acoustical background noise reduction. The probability density function (pdf) of the speech spectral amplitude is approximated by a Generalized Laplacian Distribution (GLD). Compared to earlier estimators the proposed method estimates the underlying statistical model more accurately by appropriately choosing the model parameters of GLD. Experimental results show that the proposed estimator yields a higher improvement in Segmental Signal-to-Noise Ratio (S-SNR) and lower Log-Spectral Distortion (LSD) in two different noisy environments compared to other estimators.
Embedding a Large Amount of Information Using High Secure Neural Based Steganography Algorithm
In this paper, we construct and implement a new Steganography algorithm based on learning system to hide a large amount of information into color BMP image. We have used adaptive image filtering and adaptive non-uniform image segmentation with bits replacement on the appropriate pixels. These pixels are selected randomly rather than sequentially by using new concept defined by main cases with sub cases for each byte in one pixel. According to the steps of design, we have been concluded 16 main cases with their sub cases that covere all aspects of the input information into color bitmap image. High security layers have been proposed through four layers of security to make it difficult to break the encryption of the input information and confuse steganalysis too. Learning system has been introduces at the fourth layer of security through neural network. This layer is used to increase the difficulties of the statistical attacks. Our results against statistical and visual attacks are discussed before and after using the learning system and we make comparison with the previous Steganography algorithm. We show that our algorithm can embed efficiently a large amount of information that has been reached to 75% of the image size (replace 18 bits for each pixel as a maximum) with high quality of the output.
On the EM Algorithm and Bootstrap Approach Combination for Improving Satellite Image Fusion
This paper discusses EM algorithm and Bootstrap approach combination applied for the improvement of the satellite image fusion process. This novel satellite image fusion method based on estimation theory EM algorithm and reinforced by Bootstrap approach was successfully implemented and tested. The sensor images are firstly split by a Bayesian segmentation method to determine a joint region map for the fused image. Then, we use the EM algorithm in conjunction with the Bootstrap approach to develop the bootstrap EM fusion algorithm, hence producing the fused targeted image. We proposed in this research to estimate the statistical parameters from some iterative equations of the EM algorithm relying on a reference of representative Bootstrap samples of images. Sizes of those samples are determined from a new criterion called 'hybrid criterion'. Consequently, the obtained results of our work show that using the Bootstrap EM (BEM) in image fusion improve performances of estimated parameters which involve amelioration of the fused image quality; and reduce the computing time during the fusion process.
Optimal Channel Equalization for MIMO Time-Varying Channels
We consider optimal channel equalization for MIMO (multi-input/multi-output) time-varying channels in the sense of MMSE (minimum mean-squared-error), where the observation noise can be non-stationary. We show that all ZF (zero-forcing) receivers can be parameterized in an affine form which eliminates completely the ISI (inter-symbol-interference), and optimal channel equalizers can be designed through minimization of the MSE (mean-squarederror) between the detected signals and the transmitted signals, among all ZF receivers. We demonstrate that the optimal channel equalizer is a modified Kalman filter, and show that under the AWGN (additive white Gaussian noise) assumption, the proposed optimal channel equalizer minimizes the BER (bit error rate) among all possible ZF receivers. Our results are applicable to optimal channel equalization for DWMT (discrete wavelet multitone), multirate transmultiplexers, OFDM (orthogonal frequency division multiplexing), and DS (direct sequence) CDMA (code division multiple access) wireless data communication systems. A design algorithm for optimal channel equalization is developed, and several simulation examples are worked out to illustrate the proposed design algorithm.
Motion Area Estimated Motion Estimation with Triplet Search Patterns for H.264/AVC
In this paper a fast motion estimation method for H.264/AVC named Triplet Search Motion Estimation (TS-ME) is proposed. Similar to some of the traditional fast motion estimation methods and their improved proposals which restrict the search points only to some selected candidates to decrease the computation complexity, proposed algorithm separate the motion search process to several steps but with some new features. First, proposed algorithm try to search the real motion area using proposed triplet patterns instead of some selected search points to avoid dropping into the local minimum. Then, in the localized motion area a novel 3-step motion search algorithm is performed. Proposed search patterns are categorized into three rings on the basis of the distance from the search center. These three rings are adaptively selected by referencing the surrounding motion vectors to early terminate the motion search process. On the other hand, computation reduction for sub pixel motion search is also discussed considering the appearance probability of the sub pixel motion vector. From the simulation results, motion estimation speed improved by a factor of up to 38 when using proposed algorithm than that of the reference software of H.264/AVC with ignorable picture quality loss.
Image Compression Using Hybrid Vector Quantization
In this paper, image compression using hybrid vector quantization scheme such as Multistage Vector Quantization (MSVQ) and Pyramid Vector Quantization (PVQ) are introduced. A combined MSVQ and PVQ are utilized to take advantages provided by both of them. In the wavelet decomposition of the image, most of the information often resides in the lowest frequency subband. MSVQ is applied to significant low frequency coefficients. PVQ is utilized to quantize the coefficients of other high frequency subbands. The wavelet coefficients are derived using lifting scheme. The main aim of the proposed scheme is to achieve high compression ratio without much compromise in the image quality. The results are compared with the existing image compression scheme using MSVQ.
Effect of Scene Changing on Image Sequences Compression Using Zero Tree Coding
We study in this paper the effect of the scene changing on image sequences coding system using Embedded Zerotree Wavelet (EZW). The scene changing considered here is the full motion which may occurs. A special image sequence is generated where the scene changing occurs randomly. Two scenarios are considered: In the first scenario, the system must provide the reconstruction quality as best as possible by the management of the bit rate (BR) while the scene changing occurs. In the second scenario, the system must keep the bit rate as constant as possible by the management of the reconstruction quality. The first scenario may be motivated by the availability of a large band pass transmission channel where an increase of the bit rate may be possible to keep the reconstruction quality up to a given threshold. The second scenario may be concerned by the narrow band pass transmission channel where an increase of the bit rate is not possible. In this last case, applications for which the reconstruction quality is not a constraint may be considered. The simulations are performed with five scales wavelet decomposition using the 9/7-tap filter bank biorthogonal wavelet. The entropy coding is performed using a specific defined binary code book and EZW algorithm. Experimental results are presented and compared to LEAD H263 EVAL. It is shown that if the reconstruction quality is the constraint, the system increases the bit rate to obtain the required quality. In the case where the bit rate must be constant, the system is unable to provide the required quality if the scene change occurs; however, the system is able to improve the quality while the scene changing disappears.
Hybrid Genetic-Simulated Annealing Approach for Fractal Image Compression
In this paper a hybrid technique of Genetic Algorithm and Simulated Annealing (HGASA) is applied for Fractal Image Compression (FIC). With the help of this hybrid evolutionary algorithm effort is made to reduce the search complexity of matching between range block and domain block. The concept of Simulated Annealing (SA) is incorporated into Genetic Algorithm (GA) in order to avoid pre-mature convergence of the strings. One of the image compression techniques in the spatial domain is Fractal Image Compression but the main drawback of FIC is that it involves more computational time due to global search. In order to improve the computational time along with acceptable quality of the decoded image, HGASA technique has been proposed. Experimental results show that the proposed HGASA is a better method than GA in terms of PSNR for Fractal image Compression.
Using Different Aspects of the Signings for Appearance-based Sign Language Recognition

Sign language is used by the deaf and hard of hearing people for communication. Automatic sign language recognition is a challenging research area since sign language often is the only way of communication for the deaf people. Sign language includes different components of visual actions made by the signer using the hands, the face, and the torso, to convey his/her meaning. To use different aspects of signs, we combine the different groups of features which have been extracted from the image frames recorded directly by a stationary camera. We combine the features in two levels by employing three techniques. At the feature level, an early feature combination can be performed by concatenating and weighting different feature groups, or by concatenating feature groups over time and using LDA to choose the most discriminant elements. At the model level, a late fusion of differently trained models can be carried out by a log-linear model combination. In this paper, we investigate these three combination techniques in an automatic sign language recognition system and show that the recognition rate can be significantly improved.

A Type-2 Fuzzy Adaptive Controller of a Class of Nonlinear System
In this paper we propose a robust adaptive fuzzy controller for a class of nonlinear system with unknown dynamic. The method is based on type-2 fuzzy logic system to approximate unknown non-linear function. The design of the on-line adaptive scheme of the proposed controller is based on Lyapunov technique. Simulation results are given to illustrate the effectiveness of the proposed approach.
Discovery of Sequential Patterns Based On Constraint Patterns

This paper proposes a method that discovers sequential patterns corresponding to user-s interests from sequential data. This method expresses the interests as constraint patterns. The constraint patterns can define relationships among attributes of the items composing the data. The method recursively decomposes the constraint patterns into constraint subpatterns. The method evaluates the constraint subpatterns in order to efficiently discover sequential patterns satisfying the constraint patterns. Also, this paper applies the method to the sequential data composed of stock price indexes and verifies its effectiveness through comparing it with a method without using the constraint patterns.

Analysis of Textual Data Based On Multiple 2-Class Classification Models

This paper proposes a new method for analyzing textual data. The method deals with items of textual data, where each item is described based on various viewpoints. The method acquires 2- class classification models of the viewpoints by applying an inductive learning method to items with multiple viewpoints. The method infers whether the viewpoints are assigned to the new items or not by using the models. The method extracts expressions from the new items classified into the viewpoints and extracts characteristic expressions corresponding to the viewpoints by comparing the frequency of expressions among the viewpoints. This paper also applies the method to questionnaire data given by guests at a hotel and verifies its effect through numerical experiments.

Decision Making with Dempster-Shafer Theory of Evidence Using Geometric Operators

We study the problem of decision making with Dempster-Shafer belief structure. We analyze the previous work developed by Yager about using the ordered weighted averaging (OWA) operator in the aggregation of the Dempster-Shafer decision process. We discuss the possibility of aggregating with an ascending order in the OWA operator for the cases where the smallest value is the best result. We suggest the introduction of the ordered weighted geometric (OWG) operator in the Dempster-Shafer framework. In this case, we also discuss the possibility of aggregating with an ascending order and we find that it is completely necessary as the OWG operator cannot aggregate negative numbers. Finally, we give an illustrative example where we can see the different results obtained by using the OWA, the Ascending OWA (AOWA), the OWG and the Ascending OWG (AOWG) operator.

A Sequential Pattern Mining Method Based On Sequential Interestingness

Sequential mining methods efficiently discover all frequent sequential patterns included in sequential data. These methods use the support, which is the previous criterion that satisfies the Apriori property, to evaluate the frequency. However, the discovered patterns do not always correspond to the interests of analysts, because the patterns are common and the analysts cannot get new knowledge from the patterns. The paper proposes a new criterion, namely, the sequential interestingness, to discover sequential patterns that are more attractive for the analysts. The paper shows that the criterion satisfies the Apriori property and how the criterion is related to the support. Also, the paper proposes an efficient sequential mining method based on the proposed criterion. Lastly, the paper shows the effectiveness of the proposed method by applying the method to two kinds of sequential data.

Network Intrusion Detection Design Using Feature Selection of Soft Computing Paradigms
The network traffic data provided for the design of intrusion detection always are large with ineffective information and enclose limited and ambiguous information about users- activities. We study the problems and propose a two phases approach in our intrusion detection design. In the first phase, we develop a correlation-based feature selection algorithm to remove the worthless information from the original high dimensional database. Next, we design an intrusion detection method to solve the problems of uncertainty caused by limited and ambiguous information. In the experiments, we choose six UCI databases and DARPA KDD99 intrusion detection data set as our evaluation tools. Empirical studies indicate that our feature selection algorithm is capable of reducing the size of data set. Our intrusion detection method achieves a better performance than those of participating intrusion detectors.
Modified Diffie-Hellman Protocol By Extend The Theory of The Congruence
This paper is introduced a modification to Diffie- Hellman protocol to be applicable on the decimal numbers, which they are the numbers between zero and one. For this purpose we extend the theory of the congruence. The new congruence is over the set of the real numbers and it is called the “real congruence" or the “real modulus". We will refer to the existing congruence by the “integer congruence" or the “integer modulus". This extension will define new terms and redefine the existing terms. As the properties and the theorems of the integer modulus are extended as well. Modified Diffie-Hellman key exchange protocol is produced a sharing, secure and decimal secret key for the the cryptosystems that depend on decimal numbers.
A New Velocity Expression for Open Channel and its Application to Lyari River
In this communication an expression for mean velocity of waste flow via an open channel is proposed which is an improvement over Manning formula. The discharges, storages and depths are computed at all locations of the Lyari river by utilizing proposed expression. The results attained through proposed expression are in good agreement with the observed data and better than those acquired using Manning formula.
Cubic Splines and Fourier Series Approach to Study Temperature Variation in Dermal Layers of Elliptical Shaped Human Limbs
An attempt has been made to develop a seminumerical model to study temperature variations in dermal layers of human limbs. The model has been developed for two dimensional steady state case. The human limb has been assumed to have elliptical cross section. The dermal region has been divided into three natural layers namely epidermis, dermis and subdermal tissues. The model incorporates the effect of important physiological parameters like blood mass flow rate, metabolic heat generation, and thermal conductivity of the tissues. The outer surface of the limb is exposed to the environment and it is assumed that heat loss takes place at the outer surface by conduction, convection, radiation, and evaporation. The temperature of inner core of the limb also varies at the lower atmospheric temperature. Appropriate boundary conditions have been framed based on the physical conditions of the problem. Cubic splines approach has been employed along radial direction and Fourier series along angular direction to obtain the solution. The numerical results have been computed for different values of eccentricity resembling with the elliptic cross section of the human limbs. The numerical results have been used to obtain the temperature profile and to study the relationships among the various physiological parameters.
Isomorphism on Fuzzy Graphs
In this paper, the order, size and degree of the nodes of the isomorphic fuzzy graphs are discussed. Isomorphism between fuzzy graphs is proved to be an equivalence relation. Some properties of self complementary and self weak complementary fuzzy graphs are discussed.
Modeling and Simulation for Physical Vapor Deposition: Multiscale Model
In this paper we present modeling and simulation for physical vapor deposition for metallic bipolar plates. In the models we discuss the application of different models to simulate the transport of chemical reactions of the gas species in the gas chamber. The so called sputter process is an extremely sensitive process to deposit thin layers to metallic plates. We have taken into account lower order models to obtain first results with respect to the gas fluxes and the kinetics in the chamber. The model equations can be treated analytically in some circumstances and complicated multi-dimensional models are solved numerically with a software-package (UG unstructed grids, see [1]). Because of multi-scaling and multi-physical behavior of the models, we discuss adapted schemes to solve more accurate in the different domains and scales. The results are discussed with physical experiments to give a valid model for the assumed growth of thin layers.
Factoring a Polynomial with Multiple-Roots
A given polynomial, possibly with multiple roots, is factored into several lower-degree distinct-root polynomials with natural-order-integer powers. All the roots, including multiplicities, of the original polynomial may be obtained by solving these lowerdegree distinct-root polynomials, instead of the original high-degree multiple-root polynomial directly. The approach requires polynomial Greatest Common Divisor (GCD) computation. The very simple and effective process, “Monic polynomial subtractions" converted trickily from “Longhand polynomial divisions" of Euclidean algorithm is employed. It requires only simple elementary arithmetic operations without any advanced mathematics. Amazingly, the derived routine gives the expected results for the test polynomials of very high degree, such as p( x) =(x+1)1000.
The Structure of Weakly Left C-wrpp Semigroups
In this paper, the class of weakly left C-wrpp semigroups which includes the class of weakly left C-rpp semigroups as a subclass is introduced. To particularly show that the spined product of a left C-wrpp semigroup and a right normal band which is a weakly left C-wrpp semifroup by virtue of left C-full Ehremann cyber groups recently obtained by authors Li-Shum, results obtained by Tang and Du-Shum are extended and strengthened.
Numerical Solving of General Fuzzy Linear Systems by Huang's Method

In this paper the Huang-s method for solving a m×n fuzzy linear system when, m≤ n, is considered. The method in detail is discussed and illustrated by solving some numerical examples.

Process-based Business Transformation through Services Computing
Business transformation initiatives are required by any organization to jump from its normal mode of operation to the one that is suitable for the change in the environment such as competitive pressures, regulatory requirements, changes in labor market, etc., or internal such as changes in strategy/vision, changes in the capability, change in the management, etc. Recent advances in information technology in automating the business processes have the potential to transform an organization to provide it with a sustained competitive advantage. Process constitutes the skeleton of a business. Thus, for a business to exist and compete well, it is essential for the skeleton to be robust and agile. This paper details “transformation" from a business perspective, methodologies to bring about an effective transformation, process-based transformation, and the role of services computing in this. Further, it details the benefits that could be achieved through services computing.
The Economic Lot Scheduling Problem in Flow Lines with Sequence-Dependent Setups
The problem of lot sizing, sequencing and scheduling multiple products in flow line production systems has been studied by several authors. Almost all of the researches in this area assumed that setup times and costs are sequence –independent even though sequence dependent setups are common in practice. In this paper we present a new mixed integer non linear program (MINLP) and a heuristic method to solve the problem in sequence dependent case. Furthermore, a genetic algorithm has been developed which applies this constructive heuristic to generate initial population. These two proposed solution methods are compared on randomly generated problems. Computational results show a clear superiority of our proposed GA for majority of the test problems.
A Unity Gain Fully-Differential 10bit and 40MSps Sample-And-Hold Amplifier in 0.18um CMOS

A 10bit, 40 MSps, sample and hold, implemented in 0.18-μm CMOS technology with 3.3V supply, is presented for application in the front-end stage of an analog-to-digital converter. Topology selection, biasing, compensation and common mode feedback are discussed. Cascode technique has been used to increase the dc gain. The proposed opamp provides 149MHz unity-gain bandwidth (wu), 80 degree phase margin and a differential peak to peak output swing more than 2.5v. The circuit has 55db Total Harmonic Distortion (THD), using the improved fully differential two stage operational amplifier of 91.7dB gain. The power dissipation of the designed sample and hold is 4.7mw. The designed system demonstrates relatively suitable response in different process, temperature and supply corners (PVT corners).

Learning through Shared Procedures -A Case of Using Technology to Bridge the Gap between Theory and Practice in Officer Education

In this article we explore how computer assisted exercises may allow for bridging the traditional gap between theory and practice in professional education. To educate officers able to master the complexity of the battlefield the Norwegian Military Academy needs to develop a learning environment that allows for creating viable connections between the educational environment and the field of practice. In response to this challenge we explore the conditions necessary to make computer assisted training systems (CATS) a useful tool to create structural similarities between an educational context and the field of military practice. Although, CATS may facilitate work procedures close to real life situations, this case do demonstrate how professional competence also must build on viable learning theories and environments. This paper explores the conditions that allow for using simulators to facilitate professional competence from within an educational setting. We develop a generic didactic model that ascribes learning to participation in iterative cycles of action and reflection. The development of this model is motivated by the need to develop an interdisciplinary professional education rooted in the pattern of military practice.

Optimal Manufacturing Scheduling for Dependent Details Processing
The increasing competitiveness in manufacturing industry is forcing manufacturers to seek effective processing schedules. The paper presents an optimization manufacture scheduling approach for dependent details processing with given processing sequences and times on multiple machines. By defining decision variables as start and end moments of details processing it is possible to use straightforward variables restrictions to satisfy different technological requirements and to formulate easy to understand and solve optimization tasks for multiple numbers of details and machines. A case study example is solved for seven base moldings for CNC metalworking machines processed on five different machines with given processing order among details and machines and known processing time-s duration. As a result of linear optimization task solution the optimal manufacturing schedule minimizing the overall processing time is obtained. The manufacturing schedule defines the moments of moldings delivery thus minimizing storage costs and provides mounting due-time satisfaction. The proposed optimization approach is based on real manufacturing plant problem. Different processing schedules variants for different technological restrictions were defined and implemented in the practice of Bulgarian company RAIS Ltd. The proposed approach could be generalized for other job shop scheduling problems for different applications.
A New Shock Model for Systems Subject to Random Threshold Failure
This paper generalizes Yeh Lam-s shock model for renewal shock arrivals and random threshold. Several interesting statistical measures are explicitly obtained. A few special cases and an optimal replacement problem are also discussed.
Privacy in New Mobile Payment Protocol
The increasing development of wireless networks and the widespread popularity of handheld devices such as Personal Digital Assistants (PDAs), mobile phones and wireless tablets represents an incredible opportunity to enable mobile devices as a universal payment method, involving daily financial transactions. Unfortunately, some issues hampering the widespread acceptance of mobile payment such as accountability properties, privacy protection, limitation of wireless network and mobile device. Recently, many public-key cryptography based mobile payment protocol have been proposed. However, limited capabilities of mobile devices and wireless networks make these protocols are unsuitable for mobile network. Moreover, these protocols were designed to preserve traditional flow of payment data, which is vulnerable to attack and increase the user-s risk. In this paper, we propose a private mobile payment protocol which based on client centric model and by employing symmetric key operations. The proposed mobile payment protocol not only minimizes the computational operations and communication passes between the engaging parties, but also achieves a completely privacy protection for the payer. The future work will concentrate on improving the verification solution to support mobile user authentication and authorization for mobile payment transactions.
A Multiple-Objective Environmental Rationalization and Optimization for Material Substitution in the Production of Stone-Washed Jeans- Garments
As the Textile Industry is the second largest industry in Egypt and as small and medium-sized enterprises (SMEs) make up a great portion of this industry therein it is essential to apply the concept of Cleaner Production for the purpose of reducing pollution. In order to achieve this goal, a case study concerned with ecofriendly stone-washing of jeans-garments was investigated. A raw material-substitution option was adopted whereby the toxic potassium permanganate and sodium sulfide were replaced by the environmentally compatible hydrogen peroxide and glucose respectively where the concentrations of both replaced chemicals together with the operating time were optimized. In addition, a process-rationalization option involving four additional processes was investigated. By means of criteria such as product quality, effluent analysis, mass and heat balance; and cost analysis with the aid of a statistical model, a process optimization treatment revealed that the superior process optima were 50%, 0.15% and 50min for H2O2 concentration, glucose concentration and time, respectively. With these values the superior process ought to reduce the annual cost by about EGP 105 relative to the currently used conventional method.
Thermo-Sensitive Hydrogel: Control of Hydrophilic-Hydrophobic Transition
The study investigated the hydrophilic to hydrophobic transition of modified polyacrylamide hydrogel with the inclusion of N-isopropylacrylamide (NIAM). The modification was done by mimicking micellar polymerization, which resulted in better arrangement of NIAM chains in the polyacrylamide network. The degree of NIAM arrangement is described by NH number. The hydrophilic to hydrophobic transition was measured through the partition coefficient, K, of Orange II and Methylene Blue in hydrogel and in water. These dyes were chosen as a model for solutes with different degree of hydrophobicity. The study showed that the hydrogel with higher NH values resulted in better solubility of both dyes. Moreover, in temperature above the lower critical solution temperature (LCST) of Poly(N-isopropylacrylamide) (PNIAM)also caused the collapse of NIPAM chains which results in a more hydrophobic environment that increases the solubility of Methylene Blue and decreases the solubility of Orange II in the hydrogels with NIPAM present.
Effect of One-Handed Pushing and Puling Strength at Different Handle Heights in Vertical Direction
The purpose of this study was to measure the maximal isometric strength and to investigate the effects of different handleheights and elbow angles with respect to Mid. sagittal plane on the pushing and pulling strength in vertical direction. Eight male subjects performed a series of static strength measurement for each subject. The highest isometric strength was found in pulling at shoulder height (S.H.) (Mean = 60.29 lb., SD = 16.78 lb.) and the lowest isometric strength was found also in pulling at elbow height (E.H.) (Mean = 33.06 lb., SD = 6.56 lb.). Although the isometric strengths were higher at S.H than at E.H. for both activities, the maximal isometric strengths were compared statistically. ANOVA was performed. The results of the experiment revealed that there was a significant different between handle heights. However, there were no significant different between angles and activities, also no correlation between grip strength and activities.
Large-Deflection Analysis of Automotive Vehicle's Door Wiring Harness System Using Finite Element Method

A Vehicle-s door wireing harness arrangement structure is provided. In vehicle-s door wiring harness(W/H) system is more toward to arrange a passenger compartment than a hinge and a weatherstrip. This article gives some insight into the dimensioning process, with special focus on large deflection analysis of wiring harness(W/H) in vehicle-s door structures for durability problem. An Finite elements analysis for door wiring harness(W/H) are used for residual stresses and dimensional stability with bending flexible. Durability test data for slim test specimens were compared with the numerical predicted fatigue life for verification. The final lifing of the component combines the effects of these microstructural features with the complex stress state arising from the combined service loading and residual stresses.

Encrypted Audio Transmission Using Synchronized Nd: YAG Lasers

Encoded information based on synchronization of coupled chaotic Nd:YAG lasers in master-slave configuration is numerically studied. Encoding, transmission, and decoding of information in optical chaotic communication with a single channel is presented. We analyze the robustness of the encrypted audio transmission in a channel noise. In order to illustrate this synchronization robustness, we present two cases of study: synchronization and transmission with a single channel without and with noise in the channel.

Physical Modeling of Oil Well Fire Extinguishing Using a Turbojet on a Barge

There are reports of gas and oil wells fire due to different accidents. Many different methods are used for fire fighting in gas and oil industry. Traditional fire extinguishing techniques are mostly faced with many problems and are usually time consuming and needs lots of equipments. Besides, they cause damages to facilities, and create health and environmental problems. This article proposes innovative approach in fire extinguishing techniques in oil and gas industry, especially applicable for burning oil wells located offshore. Fire extinguishment employing a turbojet is a novel approach which can help to extinguishment the fire in short period of time. Divergent and convergent turbojets modeled in laboratory scale along with a high pressure flame were used. Different experiments were conducted to determine the relationship between output discharges of trumpet and oil wells. The results were corrected and the relationship between dimensionless parameters of flame and fire extinguishment distances and also the output discharge of turbojet and oil wells in specified distances are demonstrated by specific curves.

Predicting Radiative Heat Transfer in Arbitrary Two and Three-Dimensional Participating Media
The radiative exchange method is introduced as a numerical method for the simulation of radiative heat transfer in an absorbing, emitting and isotropically scattering media. In this method, the integro-differential radiative balance equation is solved by using a new introduced concept for the exchange factor. Even though the radiative source term is calculated in a mesh structure that is coarser than the structure used in computational fluid dynamics, calculating the exchange factor between different coarse elements by using differential integration elements makes the result of the method close to that of integro-differential radiative equation. A set of equations for calculating exchange factors in two and threedimensional Cartesian coordinate system is presented, and the method is used in the simulation of radiative heat transfer in twodimensional rectangular case and a three-dimensional simple cube. The result of using this method in simulating different cases is verified by comparing them with those of using other numerical radiative models.
Recycling Poultry Feathers for Pb Removal from Wastewater: Kinetic and Equilibrium Studies
Chicken feathers were used as biosorbent for Pb removal from aqueous solution. In this paper, the kinetics and equilibrium studies at several pH, temperature, and metal concentration values are reported. For tested conditions, the Pb sorption capacity of this poultry waste ranged from 0.8 to 8.3 mg/g. Optimal conditions for Pb removal by chicken feathers have been identified. Pseudo-first order and pseudo-second order equations were used to analyze the experimental data. In addition, the sorption isotherms were fitted to classical Langmuir and Freundlich models. Finally, thermodynamic parameters for the sorption process have been determined. In summary, the results showed that chicken feathers are an alternative and promising sorbent for the treatment of effluents polluted by Pb ions.
Arsenic Mobility from Mining Tailings of Monte San Nicolas to Presa de Mata in Guanajuato, Mexico

Mining tailings represent a generating source of rich heavy metal material with a potential danger the public health and the environment, since these metals, under certain conditions, can leach and contaminate aqueous systems that serve like supplying potable water sources. The strategy for this work is based on the observation, experimentation and the simulation that can be obtained by binding real answers of the hydrodynamic behavior of metals leached from mining tailings, and the applied mathematics that provides the logical structure to decipher the individual effects of the general physicochemical phenomenon. The case of study presented herein focuses on mining tailings deposits located in Monte San Nicolas, Guanajuato, Mexico, an abandoned mine. This was considered the contamination source that under certain physicochemical conditions can favor the metal leaching, and its transport towards aqueous systems. In addition, the cartography, meteorology, geology and the hydrodynamics and hydrological characteristics of the place, will be helpful in determining the way and the time in which these systems can interact. Preliminary results demonstrated that arsenic presents a great mobility, since this one was identified in several superficial aqueous systems of the micro watershed, as well as in sediments in concentrations that exceed the established maximum limits in the official norms. Also variations in pH and potential oxide-reduction were registered, conditions that favor the presence of different species from this element its solubility and therefore its mobility.

Auto Tuning PID Controller based on Improved Genetic Algorithm for Reverse Osmosis Plant
An optimal control of Reverse Osmosis (RO) plant is studied in this paper utilizing the auto tuning concept in conjunction with PID controller. A control scheme composing an auto tuning stochastic technique based on an improved Genetic Algorithm (GA) is proposed. For better evaluation of the process in GA, objective function defined newly in sense of root mean square error has been used. Also in order to achieve better performance of GA, more pureness and longer period of random number generation in operation are sought. The main improvement is made by replacing the uniform distribution random number generator in conventional GA technique to newly designed hybrid random generator composed of Cauchy distribution and linear congruential generator, which provides independent and different random numbers at each individual steps in Genetic operation. The performance of newly proposed GA tuned controller is compared with those of conventional ones via simulation.
Numerical Investigation of the Chilling of Food Products by Air-Mist Spray
Spray chilling using air-mist nozzles has received much attention in the food processing industry because of the benefits it has shown over forced air convection. These benefits include an increase in the heat transfer coefficient and a reduction in the water loss by the product during cooling. However, few studies have simulated the heat transfer and aerodynamics phenomena of the air-mist chilling process for optimal operating conditions. The study provides insight into the optimal conditions for spray impaction, heat transfer efficiency and control of surface flooding. A computational fluid dynamics model using a two-phase flow composed of water droplets injected with air is developed to simulate the air-mist chilling of food products. The model takes into consideration droplet-to-surface interaction, water-film accumulation and surface runoff. The results of this study lead to a better understanding of the heat transfer enhancement, water conservation, and to a clear direction for the optimal design of air-mist chilling systems that can be used in commercial applications in the food and meat processing industries.
Some Aspects Regarding I. R. Absorbing Materials Based On Thin Alumina Films for Solar-Thermal Energy Conversion, Using X-Ray Diffraction Technique

Solar energy is the most “available", ecological and clean energy. This energy can be used in active or passive mode. The active mode implies the transformation of solar energy into a useful energy. The solar energy can be transformed into thermal energy, using solar collectors. In these collectors, the active and the most important element is the absorber, material which performs the absorption of solar radiation and, in at the same time, limits its reflection. The paper presents some aspects regarding the IR absorbing material – a type of cermets, used as absorber in the solar collectors, by X Ray Diffraction Technique (XRD) characterization.

A Parallel Architecture for the Real Time Correction of Stereoscopic Images
In this paper, we will present an architecture for the implementation of a real time stereoscopic images correction's approach. This architecture is parallel and makes use of several memory blocs in which are memorized pre calculated data relating to the cameras used for the acquisition of images. The use of reduced images proves to be essential in the proposed approach; the suggested architecture must so be able to carry out the real time reduction of original images.
Power Flow Control with UPFC in Power Transmission System
In this paper the performance of unified power flow controller is investigated in controlling the flow of po wer over the transmission line. Voltage sources model is utilized to study the behaviour of the UPFC in regulating the active, reactive power and voltage profile. This model is incorporated in Newton Raphson algorithm for load flow studies. Simultaneous method is employed in which equations of UPFC and the power balance equations of network are combined in to one set of non-linear algebraic equations. It is solved according to the Newton raphson algorithm. Case studies are carried on standard 5 bus network. Simulation is done in Matlab. The result of network with and without using UPFC are compared in terms of active and reactive power flows in the line and active and reactive power flows at the bus to analyze the performance of UPFC.
Numerical Solution of Infinite Boundary Integral Equation by Using Galerkin Method with Laguerre Polynomials

In this paper the exact solution of infinite boundary integral equation (IBIE) of the second kind with degenerate kernel is presented. Moreover Galerkin method with Laguerre polynomial is applied to get the approximate solution of IBIE. Numerical examples are given to show the validity of the method presented.

Determining Optimal Production Plan by Revised Surrogate Worth Trade-off Method

The authors of this work indicate by means of a concrete example that it is possible to apply efficaciously the method of multiple criteria programming in dealing with the problem of determining the optimal production plan for a certain period of time. The work presents: (1) the selection of optimization criteria, (2) the setting of the problem of determining an optimal production plan, (3) the setting of the model of multiple criteria programming in finding a solution to a given problem, (4) the revised surrogate trade-off method, (5) generalized multicriteria model for solving production planning problem and problem of choosing technological variants in the metal manufacturing industry. In the final part of this work the authors reflect on the application of the method of multiple criteria programming while determining the optimal production plan in manufacturing enterprises.

On Preprocessing of Speech Signals

Preprocessing of speech signals is considered a crucial step in the development of a robust and efficient speech or speaker recognition system. In this paper, we present some popular statistical outlier-detection based strategies to segregate the silence/unvoiced part of the speech signal from the voiced portion. The proposed methods are based on the utilization of the 3 σ edit rule, and the Hampel Identifier which are compared with the conventional techniques: (i) short-time energy (STE) based methods, and (ii) distribution based methods. The results obtained after applying the proposed strategies on some test voice signals are encouraging.

A New Approach for the Fingerprint Classification Based On Gray-Level Co- Occurrence Matrix

In this paper, we propose an approach for the classification of fingerprint databases. It is based on the fact that a fingerprint image is composed of regular texture regions that can be successfully represented by co-occurrence matrices. So, we first extract the features based on certain characteristics of the cooccurrence matrix and then we use these features to train a neural network for classifying fingerprints into four common classes. The obtained results compared with the existing approaches demonstrate the superior performance of our proposed approach.

Single-Camera EKF-vSLAM
This paper presents an Extended Kaman Filter implementation of a single-camera Visual Simultaneous Localization and Mapping algorithm, a novel algorithm for simultaneous localization and mapping problem widely studied in mobile robotics field. The algorithm is vision and odometry-based, The odometry data is incremental, and therefore it will accumulate error over time, since the robot may slip or may be lifted, consequently if the odometry is used alone we can not accurately estimate the robot position, in this paper we show that a combination of odometry and visual landmark via the extended Kalman filter can improve the robot position estimate. We use a Pioneer II robot and motorized pan tilt camera models to implement the algorithm.
TRS: System for Recommending Semantic Web Service Composition Approaches
A large number of semantic web service composition approaches are developed by the research community and one is more efficient than the other one depending on the particular situation of use. So a close look at the requirements of ones particular situation is necessary to find a suitable approach to use. In this paper, we present a Technique Recommendation System (TRS) which using a classification of state-of-art semantic web service composition approaches, can provide the user of the system with the recommendations regarding the use of service composition approach based on some parameters regarding situation of use. TRS has modular architecture and uses the production-rules for knowledge representation.
Effect of Endplate Shape on Performance and Stability of Wings-in Ground (WIG) Craft
Numerical analysis for the aerodynamic characteristics of the WIG (wing-in ground effect) craft with highly cambered and aspect ratio of one is performed to predict the ground effect for the case of with- and without- lower-extension endplate. The analysis is included varying angles of attack from 0 to10 deg. and ground clearances from 5% of chord to 50%. Due to the ground effect, the lift by rising in pressure on the lower surface is increased and the influence of wing-tip vortices is decreased. These two significant effects improve the lift-drag ratio. On the other hand, the endplate prevents the high-pressure air escaping from the air cushion at the wing tip and causes to increase the lift and lift-drag ratio further. It is found from the visualization of computation results that two wing-tip vortices are generated from each surface of the wing tip and their strength are weak and diminished rapidly. Irodov-s criteria are also evaluated to investigate the static height stability. The comparison of Irodov-s criteria shows that the endplate improves the deviation of the static height stability with respect to pitch angles and heights. As the results, the endplate can improve the aerodynamic characteristics and static height stability of wings in ground effect, simultaneously.
Application of Boost Converter for Ride-through Capability of Adjustable Speed Drives during Sag and Swell Conditions
Process control and energy conservation are the two primary reasons for using an adjustable speed drive. However, voltage sags are the most important power quality problems facing many commercial and industrial customers. The development of boost converters has raised much excitement and speculation throughout the electric industry. Now utilities are looking to these devices for performance improvement and reliability in a variety of areas. Examples of these include sags, spikes, or transients in supply voltage as well as unbalanced voltages, poor electrical system grounding, and harmonics. In this paper, simulations results are presented for the verification of the proposed boost converter topology. Boost converter provides ride through capability during sag and swell. Further, input currents are near sinusoidal. This eliminates the need of braking resistor also.
An Approach in the Improvement of the Reliability of Impedance Relay

The distance protection mainly the impedance relay which is considered as the main protection for transmission lines can be subjected to impedance measurement error which is, mainly, due to the fault resistance and to the power fluctuation. Thus, the impedance relay may not operate for a short circuit at the far end of the protected line (case of the under reach) or operates for a fault beyond its protected zone (case of overreach). In this paper, an approach to fault detection by a distance protection, which distinguishes between the faulty conditions and the effect of overload operation mode, has been developed. This approach is based on the symmetrical components; mainly the negative sequence, and it is taking into account both the effect of fault resistance and the overload situation which both have an effect upon the reliability of the protection in terms of dependability for the former and security for the latter.

Simulation and Configuration of Hydrogen Assisted Renewable Energy Power System
A renewable energy system discussed in this paper is a stand-alone wind-hydrogen system for a remote island in Australia. The analysis of an existing wind-diesel power system was performed. Simulation technique was used to model the power system currently employed on the island, and simulated different configurations of additional hydrogen energy system. This study aims to determine the suitable hydrogen integrated configuration to setting up the prototype system for the island, which helps to reduce the diesel consumption on the island. A set of configurations for the hydrogen system and associated parameters that consists of wind turbines, electrolysers, hydrogen internal combustion engines, and storage tanks has been purposed. The simulation analyses various configurations that perfectly balances the system to meet the demand on the island.
Modeling and Simulation of a Hybrid Scooter

This paper presents a hybrid electric scooter model developed and simulated using Matlab/Simulink. This hybrid scooter modeled has a parallel hybrid structure. The main propulsion units consist of a two stroke internal combustion engine and a hub motor attached to the front wheel of the scooter. The methodology used to optimize the energy and fuel consumption of the hybrid electric scooter is the multi-mode approach. Various case studies were presented to check the model and were compared to the literatures. Results shown that the model developed was feasible and valuable.

Micro-Controller Based Oxy-Fuel Profile Cutting System

In today-s era of plasma and laser cutting, machines using oxy-acetylene flame are also meritorious due to their simplicity and cost effectiveness. The objective to devise a Computer controlled Oxy-Fuel profile cutting machine arose from the increasing demand for metal cutting with respect to edge quality, circularity and lesser formation of redeposit material. The System has an 8 bit micro controller based embedded system, which assures stipulated time response. A new window based Application software was devised which takes a standard CAD file .DXF as input and converts it into numerical data required for the controller. It uses VB6 as a front end whereas MS-ACCESS and AutoCAD as back end. The system is designed around AT89C51RD2, powerful 8 bit, ISP micro controller from Atmel and is optimized to achieve cost effectiveness and also maintains the required accuracy and reliability for complex shapes. The backbone of the system is a cleverly designed mechanical assembly along with the embedded system resulting in an accuracy of about 10 microns while maintaining perfect linearity in the cut. This results in substantial increase in productivity. The observed results also indicate reduced inter laminar spacing of pearlite with an increase in the hardness of the edge region.

Combining Minimum Energy and Minimum Direct Jerk of Linear Dynamic Systems
Both the minimum energy consumption and smoothness, which is quantified as a function of jerk, are generally needed in many dynamic systems such as the automobile and the pick-and-place robot manipulator that handles fragile equipments. Nevertheless, many researchers come up with either solely concerning on the minimum energy consumption or minimum jerk trajectory. This research paper proposes a simple yet very interesting when combining the minimum energy and jerk of indirect jerks approaches in designing the time-dependent system yielding an alternative optimal solution. Extremal solutions for the cost functions of the minimum energy, the minimum jerk and combining them together are found using the dynamic optimization methods together with the numerical approximation. This is to allow us to simulate and compare visually and statistically the time history of state inputs employed by combining minimum energy and jerk designs. The numerical solution of minimum direct jerk and energy problem are exactly the same solution; however, the solutions from problem of minimum energy yield the similar solution especially in term of tendency.
The Optimal Design for Grip Force of Material Handling
Applied a mouse-s roller with a gripper to increase the efficiency for a gripper can learn to a material handling without slipping. To apply a gripper, we use the optimize principle to develop material handling by use a signal for checking a roller mouse that rotate or not. In case of the roller rotates means that the material slips. A gripper will slide to material handling until the roller will not rotate. As this experiment has test material handling for comparing a grip force that uses to material handling of the 10-human with the applied gripper. We can summarize that human exert the material handling more than the applied gripper. Because of the gripper can exert more befit to material handling than human and may be a minimum force to lift a material without slipping.
Distribution Voltage Regulation Under Three- Phase Fault by Using D-STATCOM
This paper presents the voltage regulation scheme of D-STATCOM under three-phase faults. It consists of the voltage detection and voltage regulation schemes in the 0dq reference. The proposed control strategy uses the proportional controller in which the proportional gain, kp, is appropriately adjusted by using genetic algorithms. To verify its use, a simplified 4-bus test system is situated by assuming a three-phase fault at bus 4. As a result, the DSTATCOM can resume the load voltage to the desired level within 1.8 ms. This confirms that the proposed voltage regulation scheme performs well under three-phase fault events.
Speed Regulation of a Small BLDC Motor Using Genetic-Based Proportional Control

This paper presents the speed regulation scheme of a small brushless dc motor (BLDC motor) with trapezoidal back-emf consideration. The proposed control strategy uses the proportional controller in which the proportional gain, kp, is appropriately adjusted by using genetic algorithms. As a result, the proportional control can perform well in order to compensate the BLDC motor with load disturbance. This confirms that the proposed speed regulation scheme gives satisfactory results.

Simulating Action Potential as a Linear Combination of Gating Dynamics

In this research we show that the dynamics of an action potential in a cell can be modeled with a linear combination of the dynamics of the gating state variables. It is shown that the modeling error is negligible. Our findings can be used for simplifying cell models and reduction of computational burden i.e. it is useful for simulating action potential propagation in large scale computations like tissue modeling. We have verified our finding with the use of several cell models.

Representation of Power System for Electromagnetic Transient Calculation
The new idea of analyze of power system failure with use of artificial neural network is proposed. An analysis of the possibility of simulating phenomena accompanying system faults and restitution is described. It was indicated that the universal model for the simulation of phenomena in whole analyzed range does not exist. The main classic method of search of optimal structure and parameter identification are described shortly. The example with results of calculation is shown.
A Finite Element Model for Estimating Young-s Modulus of Carbon Nanotube Reinforced Composites Incorporating Elastic Cross-Links

The presence of chemical bonding between functionalized carbon nanotubes and matrix in carbon nanotube reinforced composites is modeled by elastic beam elements representing covalent bonding characteristics. Neglecting other reinforcing mechanisms in the composite such as relatively weak interatomic Van der Waals forces, this model shows close results to the Rule of Mixtures model-s prediction for effective Young-s modulus of a Representative Volume Element of composite for small volume fractions (~1%) and high aspect ratios (L/D>200) of CNTs.

Automatic Fingerprint Classification Using Graph Theory

Using efficient classification methods is necessary for automatic fingerprint recognition system. This paper introduces a new structural approach to fingerprint classification by using the directional image of fingerprints to increase the number of subclasses. In this method, the directional image of fingerprints is segmented into regions consisting of pixels with the same direction. Afterwards the relational graph to the segmented image is constructed and according to it, the super graph including prominent information of this graph is formed. Ultimately we apply a matching technique to compare obtained graph with the model graphs in order to classify fingerprints by using cost function. Increasing the number of subclasses with acceptable accuracy in classification and faster processing in fingerprints recognition, makes this system superior.

Design of the Mathematical Model of the Respiratory System Using Electro-acoustic Analogy
The article deals with development, design and implementation of a mathematical model of the human respiratory system. The model is designed in order to simulate distribution of important intrapulmonary parameters along the bronchial tree such as pressure amplitude, tidal volume and effect of regional mechanical lung properties upon the efficiency of various ventilatory techniques. Therefore exact agreement of the model structure with the lung anatomical structure is required. The model is based on the lung morphology and electro-acoustic analogy is used to design the model.
The First Integral Approach in Stability Problem of Large Scale Nonlinear Dynamical Systems
In analyzing large scale nonlinear dynamical systems, it is often desirable to treat the overall system as a collection of interconnected subsystems. Solutions properties of the large scale system are then deduced from the solution properties of the individual subsystems and the nature of the interconnections. In this paper a new approach is proposed for the stability analysis of large scale systems, which is based upon the concept of vector Lyapunov functions and the decomposition methods. The present results make use of graph theoretic decomposition techniques in which the overall system is partitioned into a hierarchy of strongly connected components. We show then, that under very reasonable assumptions, the overall system is stable once the strongly connected subsystems are stables. Finally an example is given to illustrate the constructive methodology proposed.
Design of Smith-like Predictive Controller with Communication Delay Adaptation
This paper addresses the design of predictive networked controller with adaptation of a communication delay. The networked control system contains random delays from sensor to controller and from controller to actuator. The proposed predictive controller includes an adaptation loop which decreases the influence of communication delay on the control performance. Also, the predictive controller contains a filter which improves the robustness of the control system. The performance of the proposed adaptive predictive controller is demonstrated by simulation results in comparison with PI controller and predictive controller with constant delay.
Neural Network Controller for Mobile Robot Motion Control
In this paper the neural network-based controller is designed for motion control of a mobile robot. This paper treats the problems of trajectory following and posture stabilization of the mobile robot with nonholonomic constraints. For this purpose the recurrent neural network with one hidden layer is used. It learns relationship between linear velocities and error positions of the mobile robot. This neural network is trained on-line using the backpropagation optimization algorithm with an adaptive learning rate. The optimization algorithm is performed at each sample time to compute the optimal control inputs. The performance of the proposed system is investigated using a kinematic model of the mobile robot.
Optimization of Kinematics for Birds and UAVs Using Evolutionary Algorithms

The aim of this work is to present a multi-objective optimization method to find maximum efficiency kinematics for a flapping wing unmanned aerial vehicle. We restrained our study to rectangular wings with the same profile along the span and to harmonic dihedral motion. It is assumed that the birdlike aerial vehicle (whose span and surface area were fixed respectively to 1m and 0.15m2) is in horizontal mechanically balanced motion at fixed speed. We used two flight physics models to describe the vehicle aerodynamic performances, namely DeLaurier-s model, which has been used in many studies dealing with flapping wings, and the model proposed by Dae-Kwan et al. Then, a constrained multi-objective optimization of the propulsive efficiency is performed using a recent evolutionary multi-objective algorithm called є-MOEA. Firstly, we show that feasible solutions (i.e. solutions that fulfil the imposed constraints) can be obtained using Dae-Kwan et al.-s model. Secondly, we highlight that a single objective optimization approach (weighted sum method for example) can also give optimal solutions as good as the multi-objective one which nevertheless offers the advantage of directly generating the set of the best trade-offs. Finally, we show that the DeLaurier-s model does not yield feasible solutions.

Graphical Environment for Modeling Control Systems in Full Scope Training Simulators
This paper describes the development of a control system model using a graphical software tool. This control system is part of an operator training simulator developed for the National Training Center for Operators of Ixtapantongo (CNCAOI, acronym according to its name in Spanish language) of the Mexico-s Federal Commission of Electricity, CFE). The Department of Simulation of the Electrical Research Institute (IIE) developed this simulator using as reference the Unit I of the Combined Cycle Power Plant El Sauz, located at the centre of Mexico. The first step in the project was the developing of the Gas Turbine System and its control system simulator. The Turbo Gas simulator was finished and delivered to CNCAOI in March 2007 for commercial operation. This simulator is a high-fidelity real time dynamic simulator built and tested for accurate operation over the entire load range. The simulator was used primarily for operator training although it has been used for procedure development and evaluation of plant transients.
110 MW Geothermal Power Plant Multiple Simulator, Using Wireless Technology
A geothermal power plant multiple simulator for operators training is presented. The simulator is designed to be installed in a wireless local area network and has a capacity to train one to six operators simultaneously, each one with an independent simulation session. The sessions must be supervised only by one instructor. The main parts of this multiple simulator are: instructor and operator-s stations. On the instructor station, the instructor controls the simulation sessions, establishes training exercises and supervises each power plant operator in individual way. This station is hosted in a Main Personal Computer (NS) and its main functions are: to set initial conditions, snapshots, malfunctions or faults, monitoring trends, and process and soft-panel diagrams. On the other hand the operators carry out their actions over the power plant simulated on the operator-s stations; each one is also hosted in a PC. The main software of instructor and operator-s stations are executed on the same NS and displayed in PCs through graphical Interactive Process Diagrams (IDP). The geothermal multiple simulator has been installed in the Geothermal Simulation Training Center (GSTC) of the Comisi├│n Federal de Electricidad, (Federal Commission of Electricity, CFE), Mexico, and is being utilized as a part of the training courses for geothermal power plant operators.
Compression and Filtering of Random Signals under Constraint of Variable Memory
We study a new technique for optimal data compression subject to conditions of causality and different types of memory. The technique is based on the assumption that some information about compressed data can be obtained from a solution of the associated problem without constraints of causality and memory. This allows us to consider two separate problem related to compression and decompression subject to those constraints. Their solutions are given and the analysis of the associated errors is provided.
Performance of Random Diagonal Codes for Spectral Amplitude Coding Optical CDMA Systems
In this paper we study the use of a new code called Random Diagonal (RD) code for Spectral Amplitude Coding (SAC) optical Code Division Multiple Access (CDMA) networks, using Fiber Bragg-Grating (FBG), FBG consists of a fiber segment whose index of reflection varies periodically along its length. RD code is constructed using code level and data level, one of the important properties of this code is that the cross correlation at data level is always zero, which means that Phase intensity Induced Phase (PIIN) is reduced. We find that the performance of the RD code will be better than Modified Frequency Hopping (MFH) and Hadamard code It has been observed through experimental and theoretical simulation that BER for RD code perform significantly better than other codes. Proof –of-principle simulations of encoding with 3 channels, and 10 Gbps data transmission have been successfully demonstrated together with FBG decoding scheme for canceling the code level from SAC-signal.
An Algorithm for Computing the Analytic Singular Value Decomposition

A proof of convergence of a new continuation algorithm for computing the Analytic SVD for a large sparse parameter– dependent matrix is given. The algorithm itself was developed and numerically tested in [5].

Modeling of Dielectric Heating in Radio- Frequency Applicator Optimized for Uniform Temperature by Means of Genetic Algorithms
The paper presents an optimization study based on genetic algorithms (GA-s) for a radio-frequency applicator used in heating dielectric band products. The weakly coupled electro-thermal problem is analyzed using 2D-FEM. The design variables in the optimization process are: the voltage of a supplementary “guard" electrode and six geometric parameters of the applicator. Two objective functions are used: temperature uniformity and total active power absorbed by the dielectric. Both mono-objective and multiobjective formulations are implemented in GA optimization.
An Autonomous Collaborative Forecasting System Implementation – The First Step towards Successful CPFR System
In the past decade, artificial neural networks (ANNs) have been regarded as an instrument for problem-solving and decision-making; indeed, they have already done with a substantial efficiency and effectiveness improvement in industries and businesses. In this paper, the Back-Propagation neural Networks (BPNs) will be modulated to demonstrate the performance of the collaborative forecasting (CF) function of a Collaborative Planning, Forecasting and Replenishment (CPFR®) system. CPFR functions the balance between the sufficient product supply and the necessary customer demand in a Supply and Demand Chain (SDC). Several classical standard BPN will be grouped, collaborated and exploited for the easy implementation of the proposed modular ANN framework based on the topology of a SDC. Each individual BPN is applied as a modular tool to perform the task of forecasting SKUs (Stock-Keeping Units) levels that are managed and supervised at a POS (point of sale), a wholesaler, and a manufacturer in an SDC. The proposed modular BPN-based CF system will be exemplified and experimentally verified using lots of datasets of the simulated SDC. The experimental results showed that a complex CF problem can be divided into a group of simpler sub-problems based on the single independent trading partners distributed over SDC, and its SKU forecasting accuracy was satisfied when the system forecasted values compared to the original simulated SDC data. The primary task of implementing an autonomous CF involves the study of supervised ANN learning methodology which aims at making “knowledgeable" decision for the best SKU sales plan and stocks management.
Study of a Four-Bed Pressure Swing Adsorption for Oxygen Separation from Air
This article is presented an experimental and modeling study of a four-bed pressure swing adsorption process using zeolite13X to provide oxygen-enriched air. The binary mixture N2/O2 (79/21 vol %) was used as a feed stream. The effects of purge/feed ratio (P/F), adsorption pressure, cyclic time and product flow rate on product purity and recovery under nonisothermal condition were studied. The adsorption dynamics of process were determined using a mathematical model incorporated mass and energy balances. A Mathlab code using finite difference method was developed to solve the set of coupled differential-algebraic equations, and the simulation results are agreed well with experimental results.
Removal of Boron from Waste Waters by Ion- Exchange in a Batch System
Boron minerals are very useful for various industrial activities, such as glass industry and detergent industry, due to its mechanical and chemical properties. During the production of boron compounds, many of these are introduced into the environment in the form of waste. Boron is also an important micro nutrient for the plants to vegetate but if it exists in high concentrations, it could have toxic effects. The maximum boron level in drinking water for human health is given as 0.3 mg/L in World Health Organization (WHO) standards. The toxic effects of boron should be noted especially for dry regions, thus, in recent years, increasing attention has been paid to remove the boron from waste waters. In this study, boron removal is implemented by ion exchange process using Amberlite IRA-743 resin. Amberlite IRA-743 resin is a boron specific resin and it belongs to the polymerizate sorbent group within the aminopolyol functional group. Batch studies were performed to investigate the effects of various experimental parameters, such as adsorbent dose, initial concentration and pH, on the removal of boron. It is found that, when the adsorbent dose increases removal of boron from the liquid phase increases. However, an increase in the initial concentration decreases the removal of boron. The effective pH values for removal of boron are determined between 8.5 and 9. Equilibrium isotherms were also analyzed by Langmuir and Freundlich isotherm models. The Langmuir isotherm is obeyed better than the Freundlich isotherm.
New SUZ-4 Zeolite Membrane from Sol-Gel Technique
A new SUZ-4 zeolite membrane with tetraethlyammonium hydroxide as the template was fabricated on mullite tube via hydrothermal sol-gel synthesis in a rotating autoclave reactor. The suitable synthesis condition was SiO2:Al2O3 ratio of 21.2 for 4 days at 155 °C crystallization under autogenous pressure. The obtained SUZ-4 possessed a high BET surface area of 396.4 m2/g, total pore volume at 2.611 cm3/g, and narrow pore size distribution with 97 nm mean diameter and 760 nm long of needle crystal shape. The SUZ-4 layer obtained from seeding crystallization was thicker than that of without seeds or in situ crystallization.
QCM-D Study on Relationship of PEG Coated Stainless Steel Surfaces to Protein Resistance
Nonspecific protein adsorption generally occurs on any solid surfaces and usually has adverse consequences. Adsorption of proteins onto a solid surface is believed to be the initial and controlling step in biofouling. Surfaces modified with end-tethered poly(ethylene glycol) (PEG) have been shown to be protein-resistant to some degree. In this study, the adsorption of β-casein and lysozyme was performed on 6 different types of surfaces where PEG was tethered onto stainless steel by polyethylene imine (PEI) through either OH or NHS end groups. Protein adsorption was also performed on the bare stainless steel surface as a control. The adsorption was conducted at 23 °C and pH 7.2. In situ QCM-D was used to determine PEG adsorption kinetics, plateau PEG chain densities, protein adsorption kinetics and plateau protein adsorbed quantities. PEG grafting density was the highest for a NHS coupled chain, around 0.5 chains / nm2. Interestingly, lysozyme which has smaller size than β-casein, appeared to adsorb much less mass than that of β- casein. Overall, the surface with high PEG grafting density exhibited a good protein rejection.
The Game of Col on Complete K-ary Trees
Col is a classic combinatorial game played on graphs and to solve a general instance is a PSPACE-complete problem. However, winning strategies can be found for some specific graph instances. In this paper, the solution of Col on complete k-ary trees is presented.
Immobilized Liquid Membrane for Propylene- Propane Separation
Separation of propylene-propane mixture using immobilized liquid membrane was investigated. The effect of transmembrane pressure and carrier concentration on membrane separation performance was studied. It was observed that for 30:70 (vol. %) propylene-propane mixture, at pressure of 120kPa and carrier concentration of 20wt. %, a separation factor of 474 was obtained.
Effect of Flowrate and Coolant Temperature on the Efficiency of Progressive Freeze Concentration on Simulated Wastewater
Freeze concentration freezes or crystallises the water molecules out as ice crystals and leaves behind a highly concentrated solution. In conventional suspension freeze concentration where ice crystals formed as a suspension in the mother liquor, separation of ice is difficult. The size of the ice crystals is still very limited which will require usage of scraped surface heat exchangers, which is very expensive and accounted for approximately 30% of the capital cost. This research is conducted using a newer method of freeze concentration, which is progressive freeze concentration. Ice crystals were formed as a layer on the designed heat exchanger surface. In this particular research, a helical structured copper crystallisation chamber was designed and fabricated. The effect of two operating conditions on the performance of the newly designed crystallisation chamber was investigated, which are circulation flowrate and coolant temperature. The performance of the design was evaluated by the effective partition constant, K, calculated from the volume and concentration of the solid and liquid phase. The system was also monitored by a data acquisition tool in order to see the temperature profile throughout the process. On completing the experimental work, it was found that higher flowrate resulted in a lower K, which translated into high efficiency. The efficiency is the highest at 1000 ml/min. It was also found that the process gives the highest efficiency at a coolant temperature of -6 °C.
Photo-Fenton Treatment of 1,3-dichloro-2- Propanol Aqueous Solutions Using UV Radiation and H2O2 – A Kinetic Study

The photochemical and photo-Fenton oxidation of 1,3-dichloro-2-propanol was performed in a batch reactor, at room temperature, using UV radiation, H2O2 as oxidant, and Fenton-s reagent. The effect of the oxidative agent-s initial concentration was investigated as well as the effect of the initial concentration of Fe(II) by following the target compound degradation, the total organic carbon removal and the chloride ion production. Also, from the kinetic analysis conducted and proposed reaction scheme it was deduced that the addition of Fe(II) significantly increases the production and the further oxidation of the chlorinated intermediates.

Artificial Neural Networks for Identification and Control of a Lab-Scale Distillation Column Using LABVIEW

LABVIEW is a graphical programming language that has its roots in automation control and data acquisition. In this paper we have utilized this platform to provide a powerful toolset for process identification and control of nonlinear systems based on artificial neural networks (ANN). This tool has been applied to the monitoring and control of a lab-scale distillation column DELTALAB DC-SP. The proposed control scheme offers high speed of response for changes in set points and null stationary error for dual composition control and shows robustness in presence of externally imposed disturbance.

Software Tools for System Identification and Control using Neural Networks in Process Engineering
Neural networks offer an alternative approach both for identification and control of nonlinear processes in process engineering. The lack of software tools for the design of controllers based on neural network models is particularly pronounced in this field. SIMULINK is properly a widely used graphical code development environment which allows system-level developers to perform rapid prototyping and testing. Such graphical based programming environment involves block-based code development and offers a more intuitive approach to modeling and control task in a great variety of engineering disciplines. In this paper a SIMULINK based Neural Tool has been developed for analysis and design of multivariable neural based control systems. This tool has been applied to the control of a high purity distillation column including non linear hydrodynamic effects. The proposed control scheme offers an optimal response for both theoretical and practical challenges posed in process control task, in particular when both, the quality improvement of distillation products and the operation efficiency in economical terms are considered.
An Experimental Study on Autoignition of Wood
Experiments were conducted to characterize fire properties of wood exposed to the certain external heat flux and under variety of wood moisture content. Six kinds of Indonesian wood: keruing, sono, cemara, kamper, pinus, and mahoni were exposed to radiant heat from a conical heater, result in appearance of a stable flame on the wood surface caused by spontaneous ignition. A thermocouple K-type was used to measure the wood surface temperature. Temperature histories were recorded throughout each experiment at 1 s intervals using a TC-08. Data of first ignition time and temperature, end ignition time and temperature, and charring rate have been successfully collected. It was found that the ignition temperature and charring rate depend on moisture content of wood.
Dynamic Modeling and Simulation of Heavy Paraffin Dehydrogenation Reactor for Selective Olefin Production in Linear Alkyl Benzene Production Plant
Modeling of a heterogeneous industrial fixed bed reactor for selective dehydrogenation of heavy paraffin with Pt-Sn- Al2O3 catalyst has been the subject of current study. By applying mass balance, momentum balance for appropriate element of reactor and using pressure drop, rate and deactivation equations, a detailed model of the reactor has been obtained. Mass balance equations have been written for five different components. In order to estimate reactor production by the passage of time, the reactor model which is a set of partial differential equations, ordinary differential equations and algebraic equations has been solved numerically. Paraffins, olefins, dienes, aromatics and hydrogen mole percent as a function of time and reactor radius have been found by numerical solution of the model. Results of model have been compared with industrial reactor data at different operation times. The comparison successfully confirms validity of proposed model.
A Study on Polymer Coated Colour Pigments for Water-Based Ink
The pigments covered by film-forming polymers have opened a prospect to improve the quality of water-based printing inks. In this study such pigments were prepared by the initiated polymerization of styrene and methacrylate derivative monomers in the aqueous pigment dispersions. The formation of polymer films covering pigment cores depends on the polymerization time and the ratio of pigment to monomers. At the time of 4 hours and the ratio of 1/10 almost pigment particles are coated by the polymer. The formed polymer covers of pigments have the average thickness of 5.95 nm. The size increasing percentage of the coated particles after a week is 4.5 %, about fourteen-fold lower than of the original ones. The obtained results indicate that the coated pigments are improved dispersion stability in water medium along with a guarantee for the optical colour.
Isobaric Vapor-Liquid Equilibrium Data for Binary Mixture of 2-Methyltetrahydrofuran and Cumene
Isobaric vapor-liquid equilibrium measurements are reported for binary mixture of 2-Methyltetrahydrofuran and Cumene at 97.3 kPa. The data were obtained using a vapor recirculating type (modified Othmer's) equilibrium still. The mixture shows slight negative deviation from ideality. The system does not form an azeotrope. The experimental data obtained in this study are thermodynamically consistent according to the Herington test. The activity coefficients have been satisfactorily correlated by means of the Margules, and NRTL equations. Excess Gibbs free energy has been calculated from the experimental data. The values of activity coefficients have also been obtained by the UNIFAC group contribution method.
Isobaric Vapor-Liquid Equilibrium of Binary Mixture of Methyl Acetate with Isopropylbenzene at 97.3 kPa
Isobaric vapor-liquid equilibrium measurements are reported for the binary mixture of Methyl acetate and Isopropylbenzene at 97.3 kPa. The measurements have been performed using a vapor recirculating type (modified Othmer's) equilibrium still. The mixture shows positive deviation from ideality and does not form an azeotrope. The activity coefficients have been calculated taking into consideration the vapor phase nonideality. The data satisfy the thermodynamic consistency tests of Herington and Black. The activity coefficients have been satisfactorily correlated by means of the Margules, NRTL, and Black equations. A comparison of the values of activity coefficients obtained by experimental data with the UNIFAC model has been made.
Density Functional Calculations of 27Al, 11B,and 14N and NQR Parameters in the (6, 0) BN_AlN Nanotube Junction

Density functional theory (DFT) calculations were performed to calculate aluminum-27, boron-11, and nitrogen-14 quadrupole coupling constant (CQ) in the representative considered model of (6, 0) boron nitride-aluminum nitride nanotube junction (BN-AlNNT) for the first time. To this aim, 1.3 nm length of BNAlN consisting of 18 Al, 18 B, and 36 N atoms was selected where the end atoms capped by hydrogen atoms. The calculated CQ values for optimized BN-AlNNT system reveal different electrostatic environment in the mentioned system. The calculations were performed using the Gaussian 98 package of program.

Structural Study of Boron - Nitride Nanotube with Magnetic Resonance (NMR) Parameters Calculation via Density Functional Theory Method (DFT)

A model of (4, 4) single-walled boron-nitride nanotube as a representative of armchair boron-nitride nanotubes studied. At first the structure optimization performed and then Nuclear Magnetic Resonance parameters (NMR) by Density Functional Theory (DFT) method at 11B and 15N nuclei calculated. Resulted parameters evaluation presents electrostatic environment heterogeneity along the nanotube and especially at the ends but the nuclei in a layer feel the same electrostatic environment. All of calculations carried out using Gaussian 98 Software package.

Density Functional Calculations of N-14 andB-11 NQR Parameters in the H-capped (5, 5)Single-Wall BN Nanotube
Density functional theory (DFT) calculations were performed to compute nitrogen-14 and boron-11 nuclear quadrupole resonance (NQR) spectroscopy parameters in the representative model of armchair boron nitride nanotube (BNNT) for the first time. The considered model consisting of 1 nm length of H-capped (5, 5) single-wall BNNT were first allowed to fully relax and then the NQR calculations were carried out on the geometrically optimized model. The evaluated nuclear quadrupole coupling constants and asymmetry parameters for the mentioned nuclei reveal that the model can be divided into seven layers of nuclei with an equivalent electrostatic environment where those nuclei at the ends of tubes have a very strong electrostatic environment compared to the other nuclei along the length of tubes. The calculations were performed via Gaussian 98 package of program.
A Review of Heat Pipe Heat Exchangers Activity in Asia
Heat pipes are two-phase heat transfer devices with high effective thermal conductivity. Due to the high heat transport capacity, heat exchanger with heat pipes has become much smaller than traditional heat exchangers in handling high heat fluxes. With the working fluid in a heat pipe, heat can be absorbed on the evaporator region and transported to the condenser region where the vapour condenses releasing the heat to the cooling media. Heat pipe technology has found increasing applications in enhancing the thermal performance of heat exchangers in microelectranics, energy saving in HVAC systems for operating rooms,surgery centers, hotels, cleanrooms etc, temperature regulation systems for the human body and other industrial sectors. Development activity in heat pipe and thermosyphon technology in asia in recent years is surveyed. Some new results obtained in Australia and other countries are also included.
Simulation of Thin Film Relaxation by Buried Misfit Networks
The present work is motivated by the idea that the layer deformation in anisotropic elasticity can be estimated from the theory of interfacial dislocations. In effect, this work which is an extension of a previous approach given by one of the authors determines the anisotropic displacement fields and the critical thickness due to a complex biperiodic network of MDs lying just below the free surface in view of the arrangement of dislocations. The elastic fields of such arrangements observed along interfaces play a crucial part in the improvement of the physical properties of epitaxial systems. New results are proposed in anisotropic elasticity for hexagonal networks of MDs which contain intrinsic and extrinsic stacking faults. We developed, using a previous approach based on the relative interfacial displacement and a Fourier series formulation of the displacement fields, the expressions of elastic fields when there is a possible dissociation of MDs. The numerical investigations in the case of the observed system Si/(111)Si with low twist angles show clearly the effect of the anisotropy and thickness when the misfit networks are dissociated.
Heat Transfer at Convective Solid Melting in Fixed Bed
A method to determine experimentally the melting rate, rm, and the heat transfer coefficients, αv (W/(m3K)), at convective melting in a fixed bed of particles under adiabatic regime is established in this paper. The method lies in the determining of the melting rate by measuring the fixed bed height in time. Experimental values of rm, α and α v were determined using cylindrical particles of ice (d = 6.8 mm, h = 5.5 mm) and, as a melting agent, aqueous NaCl solution with a temperature of 283 K at different values of the liquid flow rate (11.63·10-6, 28.83·10-6, 38.83·10-6 m3/s). Our experimental results were compared with those existing in literature being noticed a good agreement for Re values higher than 50.
Experimental Study on Gas-Viscous Liquid Mixture Flow Regimes and Transitions Criteria in Vertical Narrow Rectangular Channels

In the study the influence of the physical-chemical properties of a liquid, the width of a channel gap and the superficial liquid and gas velocities on the patterns formed during two phase flows in vertical, narrow mini-channels was investigated. The research was performed in the channels of rectangular cross-section and of dimensions: 15 x 0.65 mm and 7.5 x 0.73 mm. The experimental data were compared with the published criteria of the transitions between the patterns of two-phase flows.

Vol:11 No:04 2017Vol:11 No:03 2017Vol:11 No:02 2017Vol:11 No:01 2017
Vol:10 No:12 2016Vol:10 No:11 2016Vol:10 No:10 2016Vol:10 No:09 2016Vol:10 No:08 2016Vol:10 No:07 2016Vol:10 No:06 2016Vol:10 No:05 2016Vol:10 No:04 2016Vol:10 No:03 2016Vol:10 No:02 2016Vol:10 No:01 2016
Vol:9 No:12 2015Vol:9 No:11 2015Vol:9 No:10 2015Vol:9 No:09 2015Vol:9 No:08 2015Vol:9 No:07 2015Vol:9 No:06 2015Vol:9 No:05 2015Vol:9 No:04 2015Vol:9 No:03 2015Vol:9 No:02 2015Vol:9 No:01 2015
Vol:8 No:12 2014Vol:8 No:11 2014Vol:8 No:10 2014Vol:8 No:09 2014Vol:8 No:08 2014Vol:8 No:07 2014Vol:8 No:06 2014Vol:8 No:05 2014Vol:8 No:04 2014Vol:8 No:03 2014Vol:8 No:02 2014Vol:8 No:01 2014
Vol:7 No:12 2013Vol:7 No:11 2013Vol:7 No:10 2013Vol:7 No:09 2013Vol:7 No:08 2013Vol:7 No:07 2013Vol:7 No:06 2013Vol:7 No:05 2013Vol:7 No:04 2013Vol:7 No:03 2013Vol:7 No:02 2013Vol:7 No:01 2013
Vol:6 No:12 2012Vol:6 No:11 2012Vol:6 No:10 2012Vol:6 No:09 2012Vol:6 No:08 2012Vol:6 No:07 2012Vol:6 No:06 2012Vol:6 No:05 2012Vol:6 No:04 2012Vol:6 No:03 2012Vol:6 No:02 2012Vol:6 No:01 2012
Vol:5 No:12 2011Vol:5 No:11 2011Vol:5 No:10 2011Vol:5 No:09 2011Vol:5 No:08 2011Vol:5 No:07 2011Vol:5 No:06 2011Vol:5 No:05 2011Vol:5 No:04 2011Vol:5 No:03 2011Vol:5 No:02 2011Vol:5 No:01 2011
Vol:4 No:12 2010Vol:4 No:11 2010Vol:4 No:10 2010Vol:4 No:09 2010Vol:4 No:08 2010Vol:4 No:07 2010Vol:4 No:06 2010Vol:4 No:05 2010Vol:4 No:04 2010Vol:4 No:03 2010Vol:4 No:02 2010Vol:4 No:01 2010
Vol:3 No:12 2009Vol:3 No:11 2009Vol:3 No:10 2009Vol:3 No:09 2009Vol:3 No:08 2009Vol:3 No:07 2009Vol:3 No:06 2009Vol:3 No:05 2009Vol:3 No:04 2009Vol:3 No:03 2009Vol:3 No:02 2009Vol:3 No:01 2009
Vol:2 No:12 2008Vol:2 No:11 2008Vol:2 No:10 2008Vol:2 No:09 2008Vol:2 No:08 2008Vol:2 No:07 2008Vol:2 No:06 2008Vol:2 No:05 2008Vol:2 No:04 2008Vol:2 No:03 2008Vol:2 No:02 2008Vol:2 No:01 2008
Vol:1 No:12 2007Vol:1 No:11 2007Vol:1 No:10 2007Vol:1 No:09 2007Vol:1 No:08 2007Vol:1 No:07 2007Vol:1 No:06 2007Vol:1 No:05 2007Vol:1 No:04 2007Vol:1 No:03 2007Vol:1 No:02 2007Vol:1 No:01 2007