๐”– Bobbio Scriptorium
โœฆ   LIBER   โœฆ

Identification algorithms for fuzzy relational matrices, Part 2: Optimizing algorithms

โœ Scribed by Mary M. Bourke; D. Grant Fisher


Publisher
Elsevier Science
Year
2000
Tongue
English
Weight
207 KB
Volume
109
Category
Article
ISSN
0165-0114

No coin nor oath required. For personal study only.

โœฆ Synopsis


This paper, Part 2 of a two part series, reviews and evaluates four (4) algorithms that identify fuzzy relational matrices by optimizing a user-speciรฟed performance index [6,8,29,34]. The performance of the Recursive Parameter method [34] was unsatisfactory but the Probabilistic Descent [6], Neural Learning [29] and Quasi-Newton [8] methods all gave comparable results that, in general, were better than the non-optimizing algorithms [3,9,23,32] reviewed in Part 1 [1]. However, the tuning and iteration required for these optimizing algorithms makes them less desirable for most on-line applications than the non-optimizing techniques of Shaw et al. [23] and Pedrycz [9]. It was also noted that results expressed in terms of fuzzy indices, Q q, were very poorly correlated (in fact tended to an inverse correlation) with results based on the non-fuzzy discrete indices. Jq preferred in many practical applications.


๐Ÿ“œ SIMILAR VOLUMES


Identification algorithms for fuzzy rela
โœ Mary M. Bourke; D. Grant Fisher ๐Ÿ“‚ Article ๐Ÿ“… 2000 ๐Ÿ› Elsevier Science ๐ŸŒ English โš– 182 KB

This paper is the รฟrst of a two part series that reviews and critiques several identiรฟcation algorithms for fuzzy relational matrices. Part 1 reviews and evaluates algorithms that do not optimize or minimize a speciรฟed performance criteria [3,9,20,24]. It compliments and extends a recent comparative

Novel neural algorithms based on fuzzy ฮด
โœ Xiaozhong Li; Da Ruan ๐Ÿ“‚ Article ๐Ÿ“… 1999 ๐Ÿ› Elsevier Science ๐ŸŒ English โš– 709 KB

In this paper, we first design a fuzzy neuron which possesses some generality. This fuzzy neuron is founded by replacing the operators of the traditional neuron with a pair of abstract fuzzy operators as (~, ~ ) which we call fuzzy neuron operators. For example, it may be (+, o), (A, "), (V, o), or

Novel neural algorithms based on fuzzy ฮด
โœ Xiaozhong Li; Da Ruan ๐Ÿ“‚ Article ๐Ÿ“… 2000 ๐Ÿ› Elsevier Science ๐ŸŒ English โš– 112 KB

In our previous work (Li and Ruan, 1997) we proposed a max-min operator network and a series of training algorithms, called fuzzy rules, which could be used to solve fuzzy relation equations. The most basic and important result is the convergence theorem of fuzzy perceptron based on max-min operator

Two new algorithms for solving optimizat
โœ Ali Abbasi Molai ๐Ÿ“‚ Article ๐Ÿ“… 2010 ๐Ÿ› Elsevier Science ๐ŸŒ English โš– 877 KB

This paper studies the optimization model of a linear objective function subject to a system of fuzzy relation inequalities (FRI) with the max-Einstein composition operator. If its feasible domain is non-empty, then we show that its feasible solution set is completely determined by a maximum solutio

Amer. inst. chem. eng. j.: Pho, T. K. an
๐Ÿ“‚ Article ๐Ÿ“… 1974 ๐Ÿ› Elsevier Science ๐ŸŒ English โš– 116 KB

The SAR70 program has been written as a tool within the SAR design method for supports. A support is a building in which a variety of floorplans for dwellings can hc made. The design method allows evaluation of such asupport. SAR7O can be used to generate all variations possible within a support str