Prof. Irad E. Ben-Gal- List of Selected Abstracts

 

1.     PROBABILISTIC SEQUENTIAL METHODOLOGY FOR DESIGNING A FACTORIAL SYSTEM WITH MULTIPLE RESPONSES. (pdf File)

Ben-Gal I., Braha D., and Maimon, O. (1999)

International Journal of Production Research, Vol. 37(12) 2703-2724.

 

Abstract

This paper addresses the problem of optimizing a factorial system with multiple responses. A heuristic termed probabilistic sequential methodology (PSM) is proposed. The PSM identifies those designs that maximize the likelihood of satisfying a given set of functional requirements. It is based on sequential experimentation, statistical inference and a probabilistic local search. The PSM comprises three main steps: (1) screening and estimating the main location and dispersion effects by applying fractional factorial experiments (FFE) techniques; (2) based on these effects, establishing probabilistic measures for different combinations of factor-levels; and (3) constructing a set of candidate designs from which the best solution is selected by applying a heuristic local search. The PSM is attractive when the exact analytic relationship between factor-level combinations and the system’s responses is unknown; when the system involves qualitative factors; and when the number of experiments is limited. The PSM is illustrated by a detailed case study of a Flexible Manufacturing Cell (FMC) design.

 

Keywords:  probability search, design of experiments, optimization, ranking and selection,  statistical inference, location effects, dispersion effects, Flexible Manufacturing Systems.

 

2.     ON THE UNCERTAINTIES OF DECENTRALIZED CONTROLLERS IN A TRANSFER PRODUCTION LINE. (pdf File)

Ben-Gal I., Khmelnitsky E., (2000).

IIE Transactions on Design and Manufacturing, 32, 953-961,

 

Abstract

In this paper, an information theoretic approach is applied to analyze the performance of a decentralized control system. The control system plays the role of a correcting device which decreases the uncertainties associated with state variables of a production line by applying an appropriate ``correcting signal'' for each deviation from the target. In particular, a distributed feedback control policy is considered to govern a transfer production line, which consists of machines and buffers and processes a single part type in response to a stochastic demand. It is shown how the uncertainty of the demand propagates dynamically into the production system, causing uncertainties associated with buffer levels and machine production rates. The paper proposes upper estimates for these uncertainties as functions of the demand variance, parameters of the distributed controllers and some physical properties of the production line. The bounds are based on dynamic entropy measures of the system state and the control variables. Some practical implications into the area of decentralized controller design are proposed, an information-economical analysis is presented and a numerical study is performed.

 

Keywords:  decentralize and centralize controllers, entropy measure, control complexity, stochastic demands, feedback control policy, transfer lines, information theory.

 

3.     AN APPLICATION OF INFORMATION THEORY AND ERROR-CORRECTING CODES TO FRACTIONAL FACTORIAL EXPERIMENTS. (pdf File)
Ben-Gal I., Levitin L. (2000)

Journal of Statistical Planning and Inference. 92/1-2, 267-282.

 

Abstract

The objective of design of experiments (DOE) is addressed by introducing an information optimality criterion, which is based on concepts adopted from information theory. In particular, experiments are specified to maximize the information in the system responses about estimators of the system parameters. It is shown that one has to maintain certain resolution of the design matrix to maximize the information, obtainable by a design, about a system described by a linear model with interactions. The correspondence between error-correcting codes and fractional factorial experiments provides a method to attain the required resolution with a smaller fractional factorial experiment by increasing the number of levels associated with each factor – a result that in the context of experimental design seems counterintuitive. In particular, the Gilbert-Varshamov and the Singleton bounds are employed to obtain bounds on the size of the fractional experiment. Analytical approximations and numerical results are given and illustrated by examples.

 

Keywords:  Design of Experiments, information in experiments, H-optimal designs, coding theory, design matrix, fractional factorial experiments, error correcting codes, design resolution, alphabetic optimality. 

 

4.     DESIGN OF CONTROL AND MONITORING RULES FOR STATE DEPENDENT PROCESSES (pdf File)

Ben-Gal I., Shmilovici A., Morag G. (2000),

The International Journal for Manufacturing Science and Production, 3, NOS. 2-4, pp. 85-93.

 

Abstract

Most statistical process control (SPC) methods assume that the faulty products are the result of independent and identically distributed error modes in the production system. This paper proposes a general framework for extending SPC methods to autocorrelated and state dependent processes. The suggested methodology does not require an a-priory knowledge of the process distribution. A context identification model is used to estimate the probabilities of different process outputs based on the contexts in which they appear. The Kullback-Leibler measure for the discrimination between distributions is adapted for dealing with distribution that can be described with context models. It is demonstrated that the Kullback-Leibler statistic is approximately Chi-square distributed with the number of degrees of freedom depending on the number of contexts and symbol types. Control limits are developed for the dependent processes, and numerical examples are presented.

Keywords: Process control, control charts, context algorithm, variable order Markov models, decision trees, Kullback-Leibler, SPC, auto-correlated data.

 

5.     SEQUENTIAL DOE VIA DYNAMIC PROGRAMMING (pdf File)

Ben-Gal I., Caramanis M. (2002)

IIE Transactions on Quality and Reliability, (2002) 34, 1087-1100.

 

The paper considers a sequential design of experiments (DOE) scheme. Our objective is to maximize both information and economic measures over a feasible set of experiments. Optimal DOE strategies are developed by introducing information criteria based on measures adopted from information theory. The evolution of acquired information along various stages of experimentation is analyzed for linear models with a Gaussian noise term. We show that for particular cases, although the amount of information is unbounded, the desired rate of acquiring information decreases with the number of experiments. This observation implies that at a certain point in time it is no longer efficient to continue experimenting. Accordingly, we investigate methods of stochastic dynamic programming under imperfect state information as appropriate means to obtain optimal experimentation policies. We propose cost-to-go functions that model the trade-off between the cost of additional experiments and the benefit of incremental information. We formulate a general stochastic dynamic programming framework for design of experiments and illustrate it by analytic and numerical implementation examples.

 

Keywords:  adaptive design of experiment, dynamic experimentation, cost of experiments, measure for uncertainty,  stochastic dynamic programming, dynamic programming for imperfect state information, alphabetic optimality criteria, value of information. 

 

6.     SELF-CORRECTING INSPECTION PROCEDURE UNDER INSPECTION ERRORS. (pdf File)

Ben-Gal I., Herer Y., Raz T. (2002)

IIE Transactions on Quality and Reliability, 2002, 34(6), pp. 529-540.

 

Abstract

In this paper we present a novel treatment of the inspection-system design problem when inspection is unreliable and subject to classification errors. Our approach, based on the theory of Error-Correcting Codes (ECC), leads to the development of a Self- Correcting Inspection (SCI) decision rule that does not require complete knowledge of inspection error probabilities. We show that the proposed rule assures correct classification, if the number of inspection errors is less than a certain number. We analyze the performance of the SCI decision rule under different inspection situations, including some situations that are uncommon in the field of error-correcting codes. Then, we show how the underlying mathematical structure can be applied to determine the number of inspections and the level of inspection reliability in order to minimize the sum of inspection-related costs. The practical contribution of this work lies in that it expands the ability of the designer of inspection systems to deal with cases where there is very little or no information regarding the reliability of the inspection operations.

 

Keywords: unreliable inspection, error-correcting codes, classification of errors,  cost of experiments, value of information, reliability, sensors.

 

 

7.     THE ERGONOMIC DESIGN OF WORKSTATION USING RAPID PROTOTYPING AND RESPONSE SURFACE METHODOLOGY. (pdf File)

Ben-Gal I., Bukchin J. (2001)

IIE Transactions on Design and Manufacturing, 2001; 34 (4) : 375-391

 

Abstract

The increasing use of computerized tools for virtual manufacturing in workstation design has two main advantages over traditional methods; first, it enables the designer to examine a large number of design solutions; and second, simulation of the work task may be performed in order to obtain the values of various performance measures. In this paper a new structural methodology for the workstation design is presented. Factorial experiments and the response surface methodology are integrated in order to reduce the number of examined design solutions and obtain an estimate for the best design configuration with respect to multi-objective requirements.  

 

Keywords: work stations, ergonomic design, rapid prototyping, design of experiments, computer simulations, multi-objective criteria, bio-mechanic design, fractional factorial experiments, and statistical inference.

 

 

 

8.     THROUGHPUT OF MULTIPLE PART-TYPE SYSTEMS: A SUBJECTIVE LINEAR MEASURE (pdf File)

Hevron A., Khmelnitsky E., Ben-Gal I., (2005)

The International Journal for Manufacturing Science and Production (IJMSP), Vol. 6, no. 3, pp. 143-151.

 

Abstract

The term throughput, which is commonly used as a performance measure of various production systems, has not been uniquely defined for multiple part-type systems. In some cases, the procedures that were developed to maximize throughput of multiple part-type systems are difficult to evaluate and to justify. In other cases the inaccurate definition of this term motivated incorrect operational concepts. This paper discusses some of the problems in the traditional definition of throughput and suggests a new, more suitable definition for a multiple-product manufacturing system.

 

Keywords: FMS, performance measure, productivity, workload, make span, cycle time, multiple part-types, multiple models, production line, optimization, throughput measure, production systems, design.

 


 

9.     CSPC: A MONITORING PROCEDURE FOR STATE DEPENDENT PROCESSES

Ben-Gal I., Morag G., Shmilovici A. (2003) (pdf File)

Technometrics, Vol. 45, no. 4, pp. 293-311.

 

Abstract

Most Statistical Process Control (SPC) methods are not suitable for monitoring non-linear and state-dependent processes. This paper introduces the Context-based SPC (CSPC) methodology for state-dependent data generated by a finite-memory source. The key idea of the CSPC is to monitor the statistical attributes of a process by comparing two context trees at any monitoring period of time. The first is a reference tree that represents the 'in control' reference behavior of the process; the second is a monitored tree, generated periodically from a sample of sequenced observations,  that represents the behavior of the process at that period. The Kullback-Leibler (KL) statistic is used to measure the relative 'distance' between these two trees, and an analytic distribution of this statistic is derived. Monitoring the KL statistic indicates whether there has been any significant change in the process that requires intervention. An example of buffer-level monitoring in a production system demonstrates the viability of the new method with respect to conventional methods.

 

Keywords: Nonparametric Statistical process control (SPC), distribution free SPC, Context Tree, Kullback-Leibler statistic, control charts, context algorithm, variable order Markov models, decision trees, auto-correlated data, dependent data, stochastic processes, data analysis, production systems, queueing systems.

 

 

 

 

10.  STATISTICAL PROCESS CONTROL VIA CONTEXT MODELING OF FINITE STATE PROCESSES: AN APPLICATION TO PRODUCTION MONITORING (pdf File)

Ben-Gal I., Singer G. (2004)

IIE Transactions on Quality and Reliability, vol. 36, no. 5, pp.401-415.

 

Abstract

Conventional Statistical Process Control (SPC) schemes fail to monitor nonlinear and finite-state processes that often result from feedback-controlled processes. SPC methods that are designed to monitor autocorrelated processes usually assume a known model (often an ARIMA) that might poorly describe the real process. In this paper, we present a novel SPC methodology based on context modeling of finite-state processes. The method utilizes a series of context-tree models to estimate the conditional distribution of the process output given the context of previous observations. The Kullback-Leibler divergence statistic is derived to indicate significant changes in the trees along the process. The method is implemented in a simulated flexible manufacturing system in order to detect significant changes in its production mix ratio output.

 

Keywords: Nonparametric Statistical process control (SPC), distribution free SPC, engineering process control, feedback control, Context Tree, Kullback-Leibler statistic, control charts, context algorithm, variable order Markov models, decision trees, auto-correlated data, dependent data, stochastic processes, data analysis, production systems,

 

11.  AN UPPER BOUND OF THE WEIGHT-BALANCED TESTING PROCEDURE WITH MULTIPLE TESTERS (pdf File)

Ben-Gal I. (2004)

IIE Transactions on Quality and Reliability, vol. 36, no. 5, 481-491.

 

 

Abstract

This paper presents the performance of the Weight-Balanced Testing (WBT) algorithm with multiple testers. The WBT algorithm aims to minimize the expected number of (round of) tests and has been proposed for coding, memory storage, search and testing applications. It often provides reasonable results if used with a single tester. Yet, the performance of the WBT algorithm with multiple testers and particularly its upper bound have not been previously analyzed, despite the large body of literature that exists on the WBT algorithm, and the recent papers that suggest its use in various testing applications. Here we demonstrate that WBT algorithm with multiple testers is far from being the optimal search procedure. The main result of this paper is the generalization of the upper bound on the expected number of tests previously obtained for a single-tester WBT algorithm. For this purpose, we first draw an analogy between the WBT algorithm and alphabetic codes; both being represented by the same Q-ary search tree. The upper bound is then obtained on the expected path length of a Q-ary tree, which is constructed by the WBT algorithm. Applications to the field of testing and some numerical examples are presented for illustrative purposes.

 

Keywords:  decision trees, dynamic testing, adaptive search, weight balance algorithm, queueing systems, multiple sensors, search trees, alphabetic codes, Huffman codes, Hu-Tucker codes, Vitter algorithm, data compression, entropy, information theory.  

 

12.  USING A PSEUDO-STOCHASTIC APPROACH FOR MULTIPLE-PARTS SCHEDULING ON AN UNRELIABLE MACHINE. (pdf File)

Herbon A., Khmelnitsky E., Ben-Gal I. (2005)

IIE Transactions on Operations Engineering, vol. 37, no. 3, pp. 189 - 199.

 

Abstract

In this paper we follow previous “pseudo-stochastic” approaches that solve stochastic control problems by using deterministic optimal control methods. Similarly to the certainty equivalence principle, the suggested model maximizes a given profit function of the expected system outcome. However, unlike the certainty equivalence principle, we model the expected influences of all future events, including those that are expected beyond the planning horizon, as encapsulated by their density functions and not only by their mean values. The model is applied to the optimal scheduling of multiple part-types on a single machine that is subjected to random failures and repairs. The objective of the scheduler is to maximize the profit function of the produced multiple-part mix. A numerical study is performed to evaluate the suggested pseudo-stochastic solutions under various conditions. These solutions are compared to a profit upper bound of the stochastic optimal control solutions.

 

Keywords: pseudo-stochastic, multiple part type, scheduling, unreliable machine, stochastic control, uncertainty equivalence principle, stochastic optimization, numerical methods, dynamic programming.

 

13.  IDENTIFICATION OF TRANSCRIPTION FACTOR BINDING SITES WITH VARIABLE-ORDER BAYESIAN NETWORKS (pdf File)

Ben-Gal I., Shani A., Gohr A., Grau J., Arviv S., Shmilovici A., Posch S. and Grosse I., (2005)

Bioinformatics, vol. 21, no. 11, 2657-2666.

 

Abstract

Motivation: We propose a new class of variable-order Bayesian network (VOBN) models for the identification of transcription factor binding sites (TFBSs). The proposed models generalize the widely used position weight matrix (PWM) models, Markov models and Bayesian network models. In contrast to these models, where for each position a fixed subset of the remaining positions is used to model dependencies, in VOBN models, these subsets may vary based on the specific nucleotides observed, which are called the context. This flexibility turns out to be of advantage for the classification and analysis of TFBSs, as statistical dependencies between nucleotides in different TFBS positions (not necessarily adjacent) may be taken into account

efficiently—in a position-specific and context-specific manner.

Results: We apply the VOBN model to a set of 238 experimentally verified sigma-70 binding sites in Escherichia coli. We find that the VOBN model can distinguish these 238 sites from a set of 472 intergenicnonpromoter’ sequences with a higher accuracy than fixed-order Markov models or Bayesian trees. We use a replicated stratified-holdout experiment having a fixed true-negative rate of 99.9%. We find that for a foreground inhomogeneous VOBN model of order 1 and a background homogeneous variable-order Markov (VOM) model of order 5, the obtained mean true-positive (TP) rate is 47.56%. In comparison, the best TP rate for the conventional models is 44.39%, obtained from a foreground PWM model and a background 2nd-order Markov model. As the standard deviation of the estimated TP rate is ~0.01%, this improvement is highly significant.

Availability: All datasets are available upon request from the authors. A web server for the VOBN and VOM models is available http://www.eng.tau.ac.il/~bengal/

 

Keywords: context tree, variable order Markov models, probability weight matrix, hidden and interpolated Markov models, Bayesian networks, context specific, promotes, non-promoters, DNA modeling, transcription factor binding sites (TFBS), bioinformatics, statistical dependence, nucleotides, bases, sigma 70, E. coli, inhomogeneous models.

 


14.  ECONOMIC OPTIMIZATION OF OFF-LINE INSPECTION IN A PROCESS SUBJECT TO FAILURE AND RECOVERY. (pdf File)

Finkelshtein A., Herer Y., Raz T., Ben-Gal I. (2005)

IIE Transactions on Quality and Reliability, 37, 995–1009.

 

Abstract

In certain types of processes, verification of the quality of the output units is possible only after the entire batch has been processed. We develop a model that prescribes which units should be inspected and how the units that were not inspected should be disposed of, in order to minimize the expected sum of inspection costs and disposition error costs, for processes that are subject to random failure and recovery. The model is based on a dynamic programming algorithm that has a low computational complexity. The study also includes a sensitivity analysis under a variety of cost and probability scenarios, supplemented by an analysis of the smallest batch that requires inspection, the expected number of inspections, and the performance of an easy to implement heuristic.

Keywords: Offline / online quality control, offline online inspection, unreliable inspection, adaptive testing, dynamic programming, value of information, production process, transfer line

 

 

15.  USING A COMPRESSIBILITY MEASURE TO DISTINGUISH CODING AND NONCODING DNA  (pdf File)

Shmilovici A., Ben-Gal I.,(2004)

Far East Journal of Theoretical Statistics (FEJT).Vol. 13(2), 215-235.

 

Abstract

DNA sequences consist of protein coding and noncoding regions. Recognition of coding regions is an important phase in gene-finding procedures. This paper presents a new method for distinguishing coding and noncoding DNA regions. The proposed method implements compressibility measures that results from Variable Order Markov (VOM) models. In contrast to fixed-order Markov models, where the model order is identical for all positions and for all contexts, in VOM models the order may vary – based on a nucleotide position and its contexts. As a result, VOM models are more flexible with respect to model parameterization. Preliminary experimental results on benchmark datasets demonstrate that the proposed methodology classifies coding and noncoding DNA more accurately than traditional coding measures presented in the literature.

 

Keywords: DNA compression, variable order Markov model, coding and noncoding DNA, context tree, sequence analysis, gene prediction, gene annotation, sequencing errors, detection and correction.

 

 


16.  ON THE USE OF DATA COMPRESSION MEASURES TO ASSESS ROBUST DESIGNS (pdf File)

Ben-Gal I. (2005)

IEEE Trans. on Reliability, Vol. 54, no. 3, 381-388.

 

Abstract

In this paper we suggest a potential use of data compression measures, such as the Entropy and the Huffman Coding, to assess the effects of noise factors on the reliability of tested systems. In particular, we extend the Taguchi method for robust design by computing the entropy of the percent contribution values of the noise factors. The new measures are computed already at the parameter-design stage and together with the traditional S/N ratios enable the specification of a robust design. Assuming that (some of) the noise factors should be naturalized, the entropy of a design reflects the potential efforts that will be required in the tolerance-design stage to reach a more reliable system. Using a small example, we illustrate the contribution of the new measure that might alter the designer decision in comparison with the traditional Taguchi method and ultimately obtain a system with a lower quality loss.

Assuming that the percent contribution values can reflect the probability of a noise factor to trigger a disturbance in the system response, a series of probabilistic algorithms can be applied to the robust design problem. We focus on the Huffman coding algorithm and show how to implement this algorithm such that the designer obtains the minimal expected number of tests in order to find the disturbing noise factor. The entropy measure, in this case, provides the lower bound on the algorithm's performance.

Keywords: Control & noise factors, experimentation, compression rate, robust designs, Taguchi method, S/N ratio, entropy, performance measure, Information theory, Huffman coding.

 

 

17.  ASSESSING THE EFFICIENT MARKET HYPOTHESIS VIA A UNIVERSAL PREDICTION STATISTIC

Shmilovici A. and Ben-Gal I. (after review), 2005.

 

Abstract

In an efficient market, it is assumed that all available information is fully reflected in the security prices. In recent years weak-form tests are primarily concerned with the time-series predictability of stock returns. Non of which was able to reject the weak-form of market efficiency. The purpose of this research, is to offer another test of the weak form of the market efficiency hypothesis.  The test is based on the stochastic complexity of a time series as a measure for the number of bits needed to represent the information in the time series. The length of the data after compression is a practical measure for the stochastic complexity of the data. The key idea is that in an efficient market, compression of the time series is not possible (that is the stochastic complexity is high). In this research, Rissanen’s context tree algorithm is used to identify recurring patterns in the data, and use them for compression. If the series are compressible, this indicates potential market inefficiency.  The weak market efficiency hypothesis was tested for 13 international stock indices, and for all the stocks composing the Tel-Aviv 25 index. Using sliding windows of 50, 75 and 100 consecutive trading days, it was found that for 10 international stock indices, and for 60% to 84% of the stocks tested, compressibility beyond a random is possible. These findings imply that the market is inefficient in some of the trading days that may be used to generate excess returns.

 

Keywords: Stochastic Complexity, the efficient market hypothesis, context tree, stocks prediction, stocks behavior, revenues, variable-order Markov models, portfolio analysis..

 

 

18.  USING A VOM MODEL FOR RECONSTRUCTING POTENTIAL CODING REGIONS IN EST SEQUENCES

Shmilovici A. and Ben-Gal I. (in review), 2005.

 

Abstract

This paper presents a new method for annotating coding and noncoding DNA regions using Variable Order Markov (VOM) models. In contrast to fixed-order Markov models, where the model order is identical for all positions and for all contexts, in VOM models the order may vary. As a result, VOM models are more flexible with respect to model parameterization and can be trained on relatively short sequences, and on low-quality datasets, such as Expressed Sequence Tags (ESTs) that are considered in this paper. Furthermore, the proposed method is found to be robust to random substitution errors in the genetic sequences. Accordingly, the paper presents a modified VOM model for detecting and correcting insertion and deletion sequencing errors.

 

Keywords: variable order Markov model, coding and noncoding DNA, context tree, gene annotation, sequencing error, detection and correction, gene finding, sequence analysis, gene prediction

 

 

19.  DESIGNING EXPERIMENTS FOR ROBUST OPTIMIZATION PROBLEMS: THE VS-OPTIMALITY CRITERION.  (pdf File)

Ginsburg H. and Ben-Gal I. (2006)

IIE Transactions on Quality and Reliability, vol 38, 445 – 461.

 

Abstract

We suggest an experimentation strategy for robust design of empirically fitted models. The suggested approach is used to design experiments that minimize the variance of the optimal robust solution. The new design of experiment (DOE) optimality criterion, termed Vs-optimal, prioritizes the estimation of the model's coefficients, such that the variance of the optimal solution is minimized by the performed experiments. It is discussed how the new criterion is related to known optimality criteria. We present an analytical formulation of the suggested approach for linear models and a numerical procedure for higher-order or non-polynomial models. In comparison to conventional robust-design methods, our approach provides more information on the robust solution by numerically generating its multidimensional distribution. Moreover, in a case study, the proposed approach results in a better robust solution in comparison with these standard methods.

 

Keywords: Robust design, robust optimization, design of experiments, DOE alphabetic optimality criteria, design matrix, computerized experiments, Taguchi, Box, signal to noise ratio,D-optimality, experimental design.

 

20.  THE FUNNEL EXPERIMENT: A MARKOV-BASED SPC APPROACH

Singer G. and Ben-Gal I. (2005).

 

Abstract

This paper extends the classical funnel experiment that was used by Deming to promote Statistical Process Control (SPC). The popular example illustrate that feedback control rules violate the independence assumption, but it did not indicate how to implement SPC in such cases. We extend the funnel example by introducing a simple feedback-control rule. We analyze the resulting non-linear process to which traditional SPC methods cannot be applied, then we use a simple Markov-based SPC approach to model and monitor the controlled process.

 

Keywords: funnel experiment, Deming, Quality control, feedback control, simulation, simulated experiments, statistical process control (spc), engineering process control (epc), integration, FMS.

 

 

21.  GENE-FINDING WITH THE VOM MODEL (pdf File)

Shohat-Zaidenraise K.O., Shmilovici A., Ben-Gal I. (2006)

Journal of Computational Methods in Sciences and Engineering.

 

Abstract

We present the architecture of an elementary gene-finding algorithm that is based on a Variable Order Markov model (VOM). The VOM model is a generalization of the traditional Markov model that can cope with varying memory dependencies. The VOM model is more efficient in terms of its parameterization and therefore can be trained on relatively short sequences. Experiments with the proposed gene-finder on three prokaryotic genomes indicate its potential advantage on the detection of short genes.

 

Keywords: Gene finding, variable order Markov models, context tree, sequence analysis, gene prediction, coding noncoding DNA, gene annotation, sequencing errors,  detection and correction

 

22.  ROBUST ECO-DESIGN: A NEW APPLICATION FOR QUALITY ENGINEERING

Katz R., Ben-Gal I., Bukchin J. (2005)

 

Abstract

The method of robust design has long been used for the design of systems that are insensitive to noises. In this paper we demonstrate how this approach can be used to obtain a robust eco-design (ecological design). In our case study, we implement the robust design principles to the design of a factory smoke-stack, relying on the Gaussian Plume model (GPM). The GPM is a known model for describing pollutant dispersal from a point source, subject to various atmospheric conditions. The mean-square-errors (MSE) of the accumulated pollution values around a given target is defined as the performance measure and used to adjust the design parameters. Both analytical and numerical approaches are used to evaluate the MSE measure over the design space. We demonstrate how the non-linearity in the GPM can be exploited to reach a respectively low MSE value by a cheaper design configuration. We focus and analyze the differences between the manufacturer viewpoint and the environmentalist viewpoint with respect to the considered eco-design problem.

 

Keywords: Eco design, sustainable design, Taguchi method, green manufacturing, Ecology and production, air pollution, cleaner production, design for the environment, end-of -pipe

 

23.  BACKUP STRATEGY FOR ROBOTS' FAILURES IN AN AUTOMOTIVE ASSEMBLY SYSTEM

Kahan T., Bukchin Y., Menassa R., Ben-Gal I.

 

Abstract

Robotic assembly lines are often characterized by robots' failures that reduce the throughput rate of the line. A common scenario under such failures is a full stoppage of the line during the robots' repair period. This paper presents a backup strategy in which working robots backup the failed ones by performing the latter's tasks. The strategy aims at minimizing the throughput loss by utilizing the robots' flexibility and effectively managing the redundancies in the system. In case no backup is available, the tasks of the failed robot are performed manually, resulting in an impaired quality and a reduced throughput rate.  The need for an efficient robotic backup strategy is common to automotive welding assembly lines. The paper considers such lines and formulates a Mixed Integer Linear Programming (MILP) model for minimizing the cycle time when a failure occurs. Experiments illustrate the effect of the problem parameters on the cycle time. The performance of the proposed approach is then compared with heuristic rules in a stochastic environment, where the failure rates and the repair times are randomly distributed. A quality factor, which is affected by the amount of manual work, is also evaluated. The proposed algorithm is found superior to these rules with regard to both the throughput and the quality factor.

 

Keywords: Assembly lines, Cycle-time reduction, Throughput reduction, Optimization of production systems, optimization by simulation, Robots  reliability, Automotive industry. 

 

24.  ROBUST GENE EXPRESSION CLUSTERING VIA MUTUAL INFORMATION DISTANCE MEASURE

      Priness I., Maimon O., Ben-Gal I.

 

Abstract

Motivation: The definition of a distance measure plays a key role for a successful clustering of gene expression pro-files. In this paper we compare the robustness of the Mutual Information (MI) measure to that of the Euclidean distance and the Pearson correlation coefficient.

Results: It is found that the MI measure yields a more significant differentiation among erroneous clustering solutions. The proposed measure is also used to analyze the performance of several known clustering algorithms.

 

Keywords: Gene expression data, clustering, genetics profiles, distance measures, distance metrics, homogeneity and separation. 

 

25.   VOMBAT: PREDICTION OF TRANSCRIPTION FACTOR BINDING SITES USING VARIABLE ORDER BAYESIAN TREES (pdf File)

           

      Grau J.,  Ben-Gal I., Posch S., Grosse I.

     

      Nucleic Acids Research, Vol. 34, Web Server issue, W529–W533

 

Abstract

Variable order Markov models and variable order Bayesian trees have been proposed for the recognition of transcription factor binding sites, and it could be demonstrated that they outperform traditional models such as position weight matrices, Markov models, and Bayesian trees. We develop a web-server for the recognition of DNA binding sites based on variable order Markov models and variable order Bayesian

trees offering the following functionality: (i) given data sets with annotated binding sites and genomic background sequences, variable order Markov models and variable order Bayesian trees can be trained; (ii) given a set of trained models, putative DNA binding sites can be predicted in a given set of genomic sequences; and (iii) given a data set with annotated binding sites and a data set with genomic background sequences, cross-validation experiments for different model combinations with different parameter settings can be performed. Several of the offered services are computationally demanding, such as genome-wide predictions of DNA binding sites in mammalian genomes or sets of 104-fold cross-validation experiments for different model combinations based on problem-specific data sets. In order to execute

these jobs, and in order to serve multiple users at the same time, the web-server is attached to a Linux cluster with 150 processors.

 

VOMBAT is available at http://pdw-24.ipk-gatersleben.de:8080/VOMBAT/.

 

Keywords: context tree, variable order Markov models, probability weight matrix, hidden and interpolated Markov models, Bayesian networks, context specific, promotes, non-promoters, DNA modeling, transcription factor binding sites (TFBS), bioinformatics, statistical dependence, nucleotides, bases, sigma 70, E. coli, inhomogeneous models.

 

26.  STATISTICAL PROCESS CONTROL OF THE STOCHASTIC COMPLEXITY OF DISCRETE PROCESSES

      Shmilovici A. and Ben-Gal I.,

 

Communications on Dependability and Quality Management in Engineering, 8(3), 55-61.

 

Abstract

Changes in stochastic processes often affect their description length, and reflected by their stochastic complexity measures. Monitoring the stochastic complexity of a sequence (or, equivalently, its code length) can detect process changes that may be undetectable by traditional SPC methods. The context tree is proposed here as a universal compression algorithm for measuring the stochastic complexity of a state-dependent discrete process. The advantage of the proposed method is in the reduced number of samples that are needed for reliable monitoring.

 

Keywords: Process control; Control charts; Stochastic complexity; Context tree algorithm. Variable order Markov.

 

 

line

Return to:
[Irad E. Ben-Gal's - Home Page | Department of Industrial Engineering | Faculty of Engineering ]

line

Last modified: December. 2010