Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Faramarz Khosravi is active.

Publication


Featured researches published by Faramarz Khosravi.


design automation conference | 2014

Multi-Objective Local-Search Optimization using Reliability Importance Measuring

Faramarz Khosravi; Felix Reimann; Michael Glaß; Jürgen Teich

In recent years, reliability has become a major issue and objective during the design of embedded systems. Here, different techniques to increase reliability like hardware-/software-based redundancy or component hardening are applied systematically during Design Space Exploration (DSE), aiming at achieving highest reliability at lowest possible cost. Existing approaches typically solely provide reliability measures, e. g. failure rate or Mean-Time-To-Failure (MTTF), to the optimization engine, poorly guiding the search which parts of the implementation to change. As a remedy, this work proposes an efficient approach that (a) determines the importance of resources with respect to the systems reliability and (b) employs this knowledge as part of a local search to guide the optimization engine which components/design decisions to investigate. First, we propose a novel approach to derive Importance Measures (IMs) using a structural evaluation of Success Trees (STs). Since ST-based reliability analysis is already used for MTTF calculation, our approach comes at almost no overhead. Second, we enrich the global DSE with a local search. Here, we propose strategies guided by the IMs that directly change and enhance the implemen- tation. In our experimental setup, the available measures to enhance reliability are the selection of hardening levels during resource allocation and software-based redundancy during task binding; exemplarily, the proposed local search considers the selected hardening levels. The results show that the proposed method outperforms a state-of-the-art approach regarding optimization quality, particularly in the search for highly-reliable yet affordable implementations - at negligible runtime overhead.


adaptive hardware and systems | 2015

A co-design approach for fault-tolerant loop execution on Coarse-Grained Reconfigurable Arrays

Vahid Lari; Alexandru Tanase; Jürgen Teich; Michael Witterauf; Faramarz Khosravi; Frank Hannig; Brett H. Meyer

We present a co-design approach to establish redundancy schemes such as Dual Modular Redundancy (DMR) and Triple Modular Redundancy (TMR) to a whole region of a processor array for a class of Coarse-Grained Reconfigurable Arrays (CGRAs). The approach is applied to applications with mixed-criticality properties and experiencing varying Soft Error Rates (SERs) due to environmental reasons, e. g., changing altitude. The core idea is to adapt the degree of fault protection for loop programs executing in parallel on a CGRA to the level of reliability required as well as SER profiles. This is realized through claiming neighbor regions of processing elements for the execution of replicated loop nests. First, at the source code level, a compiler transformation is proposed that realizes these replication schemes in two steps: (1) replicate given parallel loop program two or three times for DMR or TMR, respectively, and (2) add appropriate error handling functions (voting or comparison) in order to detect respectively correct any single errors. Then, using the opportunities of hardware/software co-design, we propose optimized implementations of the error handling functions in software as well as in hardware. Finally, experimental results are given for the analysis of reliability gains for each proposed scheme of array replication in dependence of different SERs.


design, automation, and test in europe | 2015

Uncertainty-aware reliability analysis and optimization

Faramarz Khosravi; Malte Müller; Michael Glaß; Jürgen Teich

Due to manufacturing tolerances and aging effects, future embedded systems have to cope with unreliable components. The intensity of such effects depends on uncertain aspects like environmental or usage conditions such that highly safety-critical systems are pessimistically designed for worst-case mission profiles. In this work, we propose to explicitly model the uncertain characteristics of system components, i. e. we model components using reliability functions with parameters distributed between a best and worst case. Since destructive effects like temperature may affect several components simultaneously (e. g. those in the same package), a correlation between uncertainties of components exists. The proposed uncertainty-aware method combines a formal analysis approach and a Monte Carlo simulation to consider uncertain characteristics and their different correlations. It delivers a holistic view on the systems reliability with best/worst/average-case behavior and also insights on variance and quantiles. But, existing optimization approaches typically assume design objectives to be single values or to follow a predefined distribution. As a remedy, we propose a dominance criterion for meta-heuristic optimization approaches like evolutionary algorithms that enables the comparison of system implementations with arbitrarily distributed characteristics. Our presented experimental results show that (a) the proposed analysis comes at low overhead while capturing existing uncertainties with sufficient accuracy, and (b) the optimization process is significantly enhanced when guiding the search process by additional aspects like variance and the 95% quantile, delivering better system implementations as found by an uncertainty-oblivious optimization approach.


Information Technology | 2016

Providing fault tolerance through invasive computing

Vahid Lari; Andreas Weichslgartner; Alexandru Tanase; Michael Witterauf; Faramarz Khosravi; Jürgen Teich; Jan Heißwolf; Stephanie Friederich; Jürgen Becker

Abstract As a consequence of technology scaling, todays complex multi-processor systems have become more and more susceptible to errors. In order to satisfy reliability requirements, such systems require methods to detect and tolerate errors. This entails two major challenges: (a) providing a comprehensive approach that ensures fault-tolerant execution of parallel applications across different types of resources, and (b) optimizing resource usage in the face of dynamic fault probabilities or with varying fault tolerance needs of different applications. In this paper, we present a holistic and adaptive approach to provide fault tolerance on Multi-Processor System-on-a-Chip (MPSoC) on demand of an application or environmental needs based on invasive computing. We show how invasive computing may provide adaptive fault tolerance on a heterogeneous MPSoC including hardware accelerators and communication infrastructure such as a Network-on-Chip (NoC). In addition, we present (a) compile-time transformations to automatically adopt well-known redundancy schemes such as Dual Modular Redundancy (DMR) and Triple Modular Redundancy (TMR) for fault-tolerant loop execution on a class of massively parallel arrays of processors called as Tightly Coupled Processor Arrays (). Based on timing characteristics derived from our compilation flow, we further develop (b) a reliability analysis guiding the selection of a suitable degree of fault tolerance. Finally, we present (c) a methodology to detect and adaptively mitigate faults in invasive NoCs.


Information Technology | 2015

Application-aware cross-layer reliability analysis and optimization

Michael Glaß; Hananeh Aliee; Liang Chen; Mojtaba Ebrahimi; Faramarz Khosravi; Veit B. Kleeberger; Alexandra Listl; Daniel Müller-Gritschneder; Fabian Oboril; Ulf Schlichtmann; Mehdi Baradaran Tahoori; Jürgen Teich; Norbert Wehn; Christian Weis

Abstract The increasing error susceptibility of semiconductor devices has put reliability in the focus of modern design methodologies. Low-level techniques alone cannot economically tackle this problem. Instead, counter measures on all system layers from device and circuit up to the application are required. As these counter measures are not for free, orchestrating them across different layers to achieve optimum trade-offs for the application wrt. reliability but also cost, timeliness, or energy consumption becomes a challenge. This typically requires a combination of analysis techniques to quantify the achieved reliability and optimization techniques that search for the best combination of counter measures. This work presents five recent approaches for application-aware cross-layer reliability optimization from within the embedded domain. Moreover, the Resilience Articulation Point (RAP) as a concept cooperatively developed to model errors across different layers is discussed. The developed approaches are showcased via applications, ranging from MIMO systems to distributed embedded control applications.


international conference on hardware/software codesign and system synthesis | 2014

An efficient technique for computing importance measures in automatic design of dependable embedded systems

Hananeh Aliee; Michael Glaß; Faramarz Khosravi; Jürgen Teich

Importance measure analysis judges the relative importance of components in a system and reveals how each component contributes to the system reliability. In the design of large and complex systems, importance measure analysis can therefore be employed to guide an optimization tool which design decisions to investigate to gain higher reliability. While previous research has mainly concentrated on developing analytical importance measure techniques, the automatic and frequent computing of importance measures as required in the context of design space exploration has got very few, if any attention. This paper presents a highly efficient technique to compute the reliability and structural importance measures of components of a system. The proposed technique considers the reliability of a system implementation and subsequently analyzes the importance measures of its components based on a state-of-the-art Monte Carlo simulation. The technique can therefore estimate the importance measures of all components concurrently, highly improving the performance of the computation compared, e. g., to the well-known Birnbaum approach by the factor of 2n with n being the number of components. Moreover, we show how this algorithm can be extended to support importance measure analysis in the existence of transient faults which is essential since in future systems, transient faults are expected to cause relatively more failures than permanent faults. We integrated the proposed analysis approach in an existing multi-objective local-search algorithm that is part of an automatic system-level design space exploration which seeks for system implementations with highest reliability at lowest possible cost. Experimental results show that the proposed algorithm performs efficiently with negligible imprecision, even for large realworld examples.


embedded systems for real time multimedia | 2017

System-level reliability analysis considering imperfect fault coverage

Faramarz Khosravi; Hananeh Aliee; Jürgen Teich

Safety-critical systems rely on redundancy schemes such as k-out-of-n structures which enable tolerance against multiple faults. These techniques are subject to Imperfect Fault Coverage (IFC) as error detection and recovery might be prone to errors or even impossible for certain fault models. As a result, these techniques may act as single points of failure in the system where uncovered faults might be overlooked and lead to wrong system outputs. Neglecting IFC in reliability analysis may lead to fatal overestimations in case of safety-critical applications. Yet, existing techniques that do consider IFC are overly pessimistic in assuming that the occurrence of an uncovered fault always results in a system failure. But often, in particular in complex systems with nested redundant structures, a fault that is not noticed by an inner redundancy scheme might be caught by an outer redundancy scheme. This paper proposes to automatically incorporate IFC into reliability models, i. e. Binary Decision Diagrams (BDDs), to enable an accurate reliability analysis for complex system structures including nested redundancies and repeated components. It also shows that IFC does not equally affect different redundancy schemes. Experimental results presented for applications in multimedia and automotive confirm that the proposed approach can analyze system reliability more accurately at an acceptable execution time and memory overhead compared to the underlying IFC-unaware technique.


Journal of Systems Architecture | 2015

Techniques for on-demand structural redundancy for massively parallel processor arrays

Vahid Lari; Jürgen Teich; Alexandru Tanase; Michael Witterauf; Faramarz Khosravi; Brett H. Meyer


IEEE Transactions on Reliability | 2017

Automatic Reliability Analysis in the Presence of Probabilistic Common Cause Failures

Faramarz Khosravi; Michael Glaß; Jürgen Teich


arXiv: Distributed, Parallel, and Cluster Computing | 2014

Towards Cross-layer Reliability Analysis of Transient and Permanent Faults

Hananeh Aliee; Liang Chen; Mojtaba Ebrahimi; Michael Glaß; Faramarz Khosravi; Mehdi Baradaran Tahoori

Collaboration


Dive into the Faramarz Khosravi's collaboration.

Top Co-Authors

Avatar

Jürgen Teich

University of Erlangen-Nuremberg

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Hananeh Aliee

University of Erlangen-Nuremberg

View shared research outputs
Top Co-Authors

Avatar

Alexandru Tanase

University of Erlangen-Nuremberg

View shared research outputs
Top Co-Authors

Avatar

Michael Witterauf

University of Erlangen-Nuremberg

View shared research outputs
Top Co-Authors

Avatar

Vahid Lari

University of Erlangen-Nuremberg

View shared research outputs
Top Co-Authors

Avatar

Liang Chen

Karlsruhe Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Malte Müller

University of Erlangen-Nuremberg

View shared research outputs
Top Co-Authors

Avatar

Mehdi Baradaran Tahoori

Karlsruhe Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Mojtaba Ebrahimi

Karlsruhe Institute of Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge