Sie sind auf Seite 1von 302

SPECIAL FUZZY MATRICES FOR

SOCIAL SCIENTISTS

W. B. Vasantha Kandasamy
e-mail: vasanthakandasamy@gmail.com
web: http://mat.iitm.ac.in/~wbv
www.vasantha.net

Florentin Smarandache
e-mail: smarand@unm.edu

K. Ilanthenral
e-mail: ilanthenral@gmail.com

INFOLEARNQUEST
Ann Arbor
2007
This book can be ordered in a paper bound reprint from:

Books on Demand
ProQuest Information & Learning
(University of Microfilm International)
300 N. Zeeb Road
P.O. Box 1346, Ann Arbor
MI 48106-1346, USA
Tel.: 1-800-521-0600 (Customer Service)
http://wwwlib.umi.com/bod/

Peer reviewers:

Prof. Mihaly Bencze, University of Brasov, Romania

Prof. Valentin Boju, Ph.D.,


Officer of the Order “Cultural Merit”
Category — “Scientific Research”
MontrealTech — Institut de Techologie de Montreal
Director, MontrealTech Press, P.O. Box 78574
Station Wilderton, Montreal, Quebec, H3S2W9, Canada

Prof. Mircea Eugen Selariu,


Polytech University of Timisoara, Romania.

Copyright 2007 by InfoLearnQuest and authors


Cover Design and Layout by Kama Kandasamy

Many books can be downloaded from the following


Digital Library of Science:
http://www.gallup.unm.edu/~smarandache/eBooks-otherformats.htm

ISBN-10: 1-59973-030-8
ISBN-13: 978-1-59973-030-1
EAN: 9781599730301

Standard Address Number: 297-5092


Printed in the United States of America

2
CONTENTS

Preface 5

Chapter One
A NEW CLASS OF SPECIAL FUZZY
MATRICES AND SPECIAL NEUTROSOPHIC
MATRICES 7

1.1 Introduction to Fuzzy Matrices 7


1.2 Special Classes of Fuzzy Matrices 26
1.3 Special Neutrosophic matrices and fuzzy
neutrosophic matrices and some essential
operators using them 98

Chapter Two
SPECIAL FUZZY MODELS AND SPECIAL
NEUTROSOPHIC MODELS AND THEIR
GENERALIZATIONS 171

2.1 Basic Description of Fuzzy Models 172


2.1.1 Definition of Fuzzy Cognitive Maps 172
2.1.2 Definition and Illustration of Fuzzy
Relational Maps (FRMs) 176
2.1.3 Properties of Fuzzy Relations and FREs 180
2.2 Neutrosophy and Neutrosophic models 187
2.2.1 An Introduction to Neutrosophy 187

3
2.2.2 Some Basic Neutrosophic Structures 190
2.2.3 On Neutrosophic Cognitive Maps 196

2.2.4 Neutrosophic Relational Maps 201


2.2.5 Binary Neutrosophic Relation and
their Properties 205
2.3 Special Fuzzy Cognitive Models and
their Neutrosophic Analogue 216
2.4 Special FRMs and NRMs and their generalizations 237
2.5 Special FRE models and NRE models 258
2.6 Some Programming Problems for Computer Experts 263

FURTHER READING 265

INDEX 293

ABOUT THE AUTHORS 301

4
PREFACE

This book is a continuation of the book, "Elementary fuzzy


matrix and fuzzy models for socio-scientists" by the same
authors. This book is a little advanced because we introduce a
multi-expert fuzzy and neutrosophic models. It mainly tries to
help social scientists to analyze any problem in which they need
multi-expert systems with multi-models.

To cater to this need, we have introduced new classes of fuzzy


and neutrosophic special matrices. The first chapter is
essentially spent on introducing the new notion of different
types of special fuzzy and neutrosophic matrices, and the simple
operations on them which are needed in the working of these
multi expert models.

In the second chapter, new set of multi expert models are


introduced; these special fuzzy models and special fuzzy
neutrosophic models that can cater to adopt any number of
experts. The working of the model is also explained by
illustrative examples.

However, these special fuzzy models can also be used by


applied mathematicians to study social and psychological
problems. These models can also be used by doctors, engineers,
scientists and statisticians. The SFCM, SMFCM, SNCM,
SMNCM, SFRM, SNRM, SMFRM, SMNRM, SFNCMs,
SFNRMs, etc. can give the special hidden pattern for any given
special input vector.

5
The working of these SFREs, SMFREs and their neutrosophic
analogues depends heavily upon the problems and the experts'
expectation. The authors have given a long list for further
reading which may help the socio scientists to know more about
SFRE and SMFREs.

We thank Dr. K. Kandasamy and Meena, without their


unflinching support, this book would have never been possible.

W.B.VASANTHA KANDASAMY
FLORENTIN SMARANDACHE
ILANTHENRAL. K

6
Chapter One

A NEW CLASS OF
SPECIAL FUZZY MATRICES AND SPECIAL
NEUTROSOPHIC MATRICES

In this chapter for the first time we introduce some new classes
of special fuzzy matrices and illustrate them with examples.
Also we give the main type of operations carried out on them.
All these special fuzzy matrices will be used in the special fuzzy
models which will be introduced in chapter two of this book.
This chapter has three sections. In sections one we introduce the
notion of fuzzy matrices and give the operations used on them
like min max operations or max min operations. In section two
we introduce the new classes of special fuzzy matrices define
special operations on them and illustrate them with examples. In
section three neutrosophic matrices, special operations on them
are introduced and described. Several illustrative examples are
given to make the operations explicit.

1.1 Introduction to Fuzzy Matrices

Here we just recall the definition of fuzzy matrices for more


about these concepts please refer [106]. Throughout this book
the unit interval [0, 1] denotes the fuzzy interval. However in
certain fuzzy matrices we also include the interval [–1, 1] to be

7
the fuzzy interval. So any element aij ∈ [–1, 1] can be positive or
negative. If aij is positive then 0 < aij ≤ 1, if aij is negative then
–1 ≤ aij ≤ 0; aij = 0 can also occur. So [0, 1] or [–1, 1] will be
known as fuzzy interval.
Thus if A = (aij) is a matrix and if in particular aij ∈ [0, 1]
(or [–1, 1]) we call A to be a fuzzy matrix. So all fuzzy matrices
are matrices but every matrix in general need not be a fuzzy
matrix. Hence fuzzy matrices forms a subclass of matrices.
Now we give some examples of fuzzy matrices.

Example 1.1.1: Let


⎡ 0.3 0.1 0.4 1 ⎤
A = ⎢⎢ 0.2 1 0.7 0 ⎥⎥
⎢⎣ 0.9 0.8 0.5 0.8⎥⎦

be a matrix. Every element in A is in the unit interval [0, 1].


Thus A is a fuzzy matrix. We can say A is a 3 × 4 rectangular
fuzzy matrix.

Example 1.1.2: Let

⎡ 0.1 1 0 0.3 0.6 0.2 ⎤


⎢ 1 0.5 1 0.8 0.9 1 ⎥
⎢ ⎥
B=⎢ 0 1 0.3 0.9 0.7 0.5⎥
⎢ ⎥
⎢0.4 0 1 0.6 0.3 1 ⎥
⎢⎣ 1 0.8 0.9 0.3 0.8 0 ⎥⎦

be a fuzzy matrix B is a 5 × 5 square fuzzy matrix. It is clear all


the entries of B are from the unit interval [0, 1].

Example 1.1.3: Consider the matrix

⎡1 0 1⎤
C = ⎢⎢ −1 1 −1⎥⎥
⎢⎣ 0 −1 1 ⎥⎦

8
C is a matrix, its entries are from the set {–1, 0,1}. C is a 3 × 3
square fuzzy matrix.

Example 1.1.4: Let A = [0 0.3 0.1 0.5 1 0.8 0.9 1 0]. A is a 1 ×


9 fuzzy matrix will also be known as the fuzzy row vector or
row fuzzy matrix.

Example 1.1.5: Let

⎡ 1 ⎤
⎢ 0.3⎥
⎢ ⎥
⎢0.7 ⎥
T= ⎢ ⎥
⎢ 0.5 ⎥
⎢ 0.1⎥
⎢ ⎥
⎣⎢ 0.9 ⎦⎥

be 6 × 1 fuzzy matrix. T is also know as the fuzzy column


vector or fuzzy column matrix. Thus if A = [a1 a2 … an] where ai
∈ [0, 1]; 1 ≤ i ≤ n, A will be known as the fuzzy row matrix or
the fuzzy row vector.

Let
⎡ b1 ⎤
⎢b ⎥
⎢ 2⎥
B = ⎢ b3 ⎥
⎢ ⎥
⎢ ⎥
⎢⎣ b m ⎥⎦

where bj ∈ [0, 1], 1 ≤ j ≤ m, B will be known as the fuzzy


column vector or the fuzzy column matrix.
Let A = (aij) with aij ∈ [0, 1], 1 ≤ i ≤ n and 1 ≤ j ≤ n. A will
be known as the n × n fuzzy square matrix. Suppose C = (cij)
with cij ∈ [0, 1]; 1 ≤ i ≤ n and 1 ≤ j ≤ m then C is known as the n
× m rectangular fuzzy matrix. We have seen examples of these
types of fuzzy matrices. A = [0 0 … 0] will be know as the zero

9
⎡0 ⎤
⎢0 ⎥
fuzzy row vector, B = ⎢ ⎥ will be known as the zero fuzzy
⎢ ⎥
⎢ ⎥
⎣0 ⎦
column vector and X = [1 1 … 1] is the fuzzy row unit vector
⎡1⎤
⎢1⎥
and Y = ⎢ ⎥ will be known as the fuzzy column unit vector.
⎢⎥
⎢⎥
⎣1⎦
Thus the unit fuzzy row vector and unit row vector are one and
the same so is the zero fuzzy row vector and zero fuzzy column
vector they are identical with the zero row vector and the zero
column vector respectively.

Now we see the usual matrix addition of fuzzy matrices in


general does not give a fuzzy matrix.

This is clearly evident from the following example.

Example 1.1.6: Consider the 3 × 3 fuzzy matrices A and B,


where
⎡ 0.3 0.8 1 ⎤
A = ⎢⎢ 1 0.8 0 ⎥⎥
⎢⎣ 0.9 0.5 0.7 ⎥⎦
and
⎡ 0.5 0.7 0.3⎤
B = ⎢⎢0.2 0.1 0.6 ⎥⎥ .
⎢⎣ 0.3 0.8 0.9 ⎥⎦

Now under usual matrix addition

⎡ 0.3 0.8 1 ⎤ ⎡ 0.5 0.7 0.3⎤


A + B = ⎢⎢ 1 0.8 0 ⎥⎥ + ⎢⎢0.2 0.1 0.6 ⎥⎥
⎢⎣ 0.9 0.5 0.7 ⎥⎦ ⎢⎣ 0.3 0.8 0.9 ⎥⎦

10
⎡0.8 1.5 1.3 ⎤
= ⎢⎢1.2 0.9 0.6 ⎥⎥ .
⎢⎣1.2 1.3 1.6 ⎥⎦

Clearly all entries in A + B are not in [0, 1]. Thus A + B is only


a 3 × 3 matrix and is not a 3 × 3 fuzzy matrix.

On similar lines we see the product of two fuzzy matrices under


usual matrix multiplication in general does not lead to a fuzzy
matrix. This is evident from the following example.

Example 1.1.7: Let


⎡ 0.8 0.9 ⎤
A=⎢ ⎥
⎣ 1 0.3⎦
and
⎡1 0.8⎤
B=⎢ ⎥
⎣ 0 0.9 ⎦

be two 2 × 2 fuzzy matrices.


Now the product of matrices A with B is given by

⎡ 0.8 0.9 ⎤ ⎡ 1 0.8⎤


A×B= ⎢ ⎥⎢ ⎥
⎣ 1 0.3⎦ ⎣ 0 0.9 ⎦

⎡ 0.8 + 0 0.64 + 0.81⎤


= ⎢
⎣ 1+ 0 0.8 + 0.27 ⎥⎦

⎡ 0.8 1.45 ⎤
= ⎢ ⎥.
⎣ 1 1.07 ⎦

We see all the entries in A × B which will also be denoted by


AB are not in [0, 1] so AB is not a fuzzy matrix. Thus under the
usual multiplication the product of two fuzzy matrices in
general need not yield a fuzzy matrix. So we are forced to find

11
different operations on fuzzy matrices so that under those
operations we get the resultant to be also a fuzzy matrix.

Let A = (aij) and B = (bij) be any two m × n fuzzy matrices;


define Max (A, B) = (Max (aij, bij)), aij ∈ A and bij ∈ B, 1 ≤ i ≤
m and 1 ≤ j ≤ n. Then Max (A, B) is a fuzzy matrix. This
operation will be known as Max operation.

Example 1.1.8: Let


⎡ 0.8 1 0 0.3⎤
A = ⎢ 0.3 0.2 0.4 1 ⎥⎥

⎢⎣ 0.1 0 0.7 0.8⎥⎦
and
⎡ 0.9 0.8 0.7 0 ⎤
B = ⎢⎢ 0.1 1 0 0.3⎥⎥
⎢⎣ 0.2 0.5 0.5 0.8⎥⎦

be any two 3 × 4 fuzzy matrices Max {A, B} =

⎡ Max(0.8,0.9) Max(1,0.8) Max(0,0.7) Max(0.3,0) ⎤


⎢ Max(0.3,0.1) Max(0.2,1) Max(0.4,0) Max(1,0.3) ⎥⎥

⎢⎣ Max(0.1,0.2) Max(0,0.5) Max(0.7,0.5) Max(0.8,0.8) ⎥⎦

⎡0.9 1 0.7 0.3⎤


= ⎢⎢ 0.3 1 0.4 1 ⎥⎥ ;
⎢⎣0.2 0.5 0.7 0.8⎥⎦

clearly Max (A, B) is again a fuzzy matrix as every entry given


by Max (A, B) belongs to the interval [0, 1].

It is interesting to note Max (A, A) = A and Max ((0), A) =


A where (0) is the zero matrix of the same order as that of A.

Now we proceed on to define yet another operation on


fuzzy matrices. Let A and B be any two m × n fuzzy matrices

12
Min (A, B) = (Min (aij, bij)) where A = (aij) and B = (bij), 1 ≤ i <
m and 1 ≤ j ≤ n.

We illustrate this Min operation on two fuzzy matrices by the


following example.

Example 1.1.9: Let


⎡ 0.3 1 0.8⎤
⎢ 1 0.3 0.9 ⎥
⎢ ⎥
A = ⎢ 0 0.8 0.3⎥
⎢ ⎥
⎢ 0.7 0.2 1 ⎥
⎢⎣ 1 0 0.8⎥⎦

⎡ 1 0.8 0 ⎤
⎢ 0.3 0.2 0.5⎥
⎢ ⎥
B = ⎢ 0.1 1 0.5⎥
⎢ ⎥
⎢ 1 0.3 0 ⎥
⎢⎣ 0.5 0 1 ⎥⎦

be any two 5 × 3 fuzzy matrices.


Now
⎡ Min(0.3,1) Min(1,0.8) Min(0.8,0) ⎤
⎢ Min(1,0.3) Min(0.3,0.2) Min(0.9,0.5) ⎥
⎢ ⎥
Min (A, B) = ⎢ Min(0,0.1) Min(0.8,1) Min(0.3,0.5) ⎥
⎢ ⎥
⎢ Min(0.7,1) Min(0.2,0.3) Min(1,0) ⎥
⎢⎣ Min(1,0.5) Min(0,0) Min(0.8,1) ⎥⎦

⎡ 0.3 0.8 0 ⎤
⎢ 0.3 0.2 0.5⎥⎥

= ⎢ 0 0.8 0.3⎥ .
⎢ ⎥
⎢ 0.7 0.2 0 ⎥
⎢⎣ 0.5 0 0.8⎥⎦

13
Clearly Min(A, B) is a fuzzy matrix as all entries in Min(A, B)
belong to the unit interval [0, 1].
Now it is interesting to note that Min(A, A) =A where as
Min (A, (0)) = (0).
Further we see Min (A, B) = Min(B, A) and Max(A, B) =
Max(B, A).
We can have other types of operations on the matrices A
and B called max min operation or min max operations. Let P
and Q be two fuzzy matrices where P = (pik) be a m × n matrix;
1 ≤ i ≤ m and 1 ≤ k ≤ n, Q = (qkj) be a n × t matrix where 1 ≤ k ≤
n and 1 ≤ j ≤ t; then max min operations of P and Q is given by
R = (rij) = max min (pik, qkj), where 1 ≤ i ≤ m and 1 ≤ j ≤ t and R
is a m × t matrix. Clearly R = (rij) is a fuzzy matrix.

We illustrate this by the following example.

Example 1.1.10: Let


⎡ 0.3 0.1 0.6 ⎤
A = ⎢⎢ 0 0.7 1 ⎥⎥
⎢⎣ 0.4 0.2 0.3⎥⎦
and
⎡ 0.6 0.2 0 0.7 ⎤
B = ⎢⎢ 0.3 0.8 0.2 0 ⎥⎥
⎢⎣ 1 0.1 0.4 1 ⎥⎦

be any two fuzzy matrices where A is a 3 × 3 fuzzy matrix and


B is a 3 × 4 fuzzy matrix, max min (A, B) is given by R:

⎧ ⎡ 0.3 0.1 0.6 ⎤ ⎡0.6 0.2 0 0.7 ⎤ ⎫


⎪ ⎢ 0.3 0.8 0.2 0 ⎥ ⎪
R = max min ⎨ ⎢⎢ 0 0.7 1 ⎥⎥ , ⎢ ⎥⎬
⎪ ⎢0.4 0.2 0.3⎥ ⎢⎣ 1 0.1 0.4 1 ⎥⎦ ⎪⎭
⎩⎣ ⎦

⎡ r11 r12 r13 r14 ⎤


R = ⎢⎢ r21 r22 r23 r24 ⎥⎥ ;
⎣⎢ r31 r32 r33 r34 ⎦⎥

14
where

r11 = max {min(0.3, 0.6), min (0.1, 0.3), min (0.6, 1)}
= max {0.3, 0.1, 0.6}
= 0.6.
r12 = max {min (0.3, 0.2), min (0.1, 0.8), min (0.6, 0.1)}
= max {0.2, 0.1, 0.1}
= 0.2.
r13 = max {min (0.3, 0), min (0.1, 0.2), min (0.6, 0.4)}
= max {0, 0.1, 0.4}
= 0.4.
r14 = max {min (0.3, 0.7), min (0.1, 0), min (0.6, 1)}
= max {0.3, 0, 0.6}
= 0.6.
r21 = max {min (0, 0.6), min (0.7, 0.3), min (1, 1)}
= max {0, 0.3, 1}
= 1;

and so on and

r34 = max {min (0.4, 0.7), min (0.2, 0), min (0.3, 1)}
= max {0.4, 0, 0.3}
= 0.4.

⎡ 0.6 0.2 0.4 0.6 ⎤


R = ⎢ 1 0.7 0.4 1 ⎥
⎢ ⎥
⎢⎣ 0.4 0.2 0.3 0.4 ⎥⎦

= (rij).

Likewise we can also define the notion of min max operator on


A, B as follows; if A = (aik) be a m × n matrix with 1 ≤ i ≤ m
and 1 ≤ k ≤ n and B = (bkj) be a n × t matrix where 1 ≤ k ≤ n and
1 ≤ j ≤ t, then C = (cij) = min max {aik, bkj}, where 1 ≤ i ≤ m and
1 ≤ j ≤ t,

We illustrate this by the following example. Further it is


pertinent to mention here that in general

15
max min {a ik , b kj } ≠ min max {a ik , b kj } .
k k

We find the min max (A, B) for the same A and B given in
example 1.1.10.

Example 1.1.11: Let


⎡ 0.3 0.1 0.6 ⎤
A = ⎢⎢ 0 0.7 1 ⎥⎥
⎢⎣ 0.4 0.2 0.3⎥⎦
and
⎡0.6 0.2 0 0.7 ⎤
B = ⎢⎢ 0.3 0.8 0.2 0 ⎥⎥
⎢⎣ 1 0.1 0.4 1 ⎥⎦

be the fuzzy matrices given in example 1.1.10.


Now

min max (A, B)

⎧ ⎡ 0.3 0.1 0.6 ⎤ ⎡ 0.6 0.2 0 0.7 ⎤ ⎫


⎪ ⎢ 0.3 0.8 0.2 0 ⎥ ⎪
= min {max ⎨ ⎢⎢ 0 0.7 1 ⎥⎥ , ⎢ ⎥⎬
⎪ ⎢0.4 0.2 0.3⎥ ⎢⎣ 1 0.1 0.4 1 ⎥⎦ ⎪⎭
⎩⎣ ⎦

⎡ p11 p12 p13 p14 ⎤


= ⎢p p 22 p 23 p 24 ⎥⎥ = P
⎢ 21
⎢⎣ p31 p32 p33 p34 ⎥⎦

p11 = min {max (0.3, 0.6), max (0.1, 0.3), max (0.6, 1)}
= min {0.6, 0.3, 1}
= 0.3.
p12 = min {max (0.3, 0.2), max (0.1, 0.8), max (0.6, 1)}
= min {0.3, 0.8, 1}
= 0.3.
p13 = min {max (0.3, 0), max (0.1, 0.2), max (0.6, 0.4)}

16
= min {0.3, 0.2, 0.6}
= 0.2.
p14 = min {max (0.3, 0.7), max (0.1, 0), max (0.6, 1)}
= min {0.7, 0.1, 1}
= 0.1.
p21 = min {max (0, 0.6), max (0.7, 0.3), max (1, 1)}
= min {0.6, 0.7, 1}
= 0.6
and so on and
p34 = min {max (0.4, 0.7), max (0.2, 0), max (0.3, 1)}
= min {0.7, 0.2, 1}
= 0.2.

Thus we get
⎡ 0.3 0.3 0.2 0.1⎤
P = ⎢⎢ 0.6 0.2 0 0.7 ⎥⎥ .
⎢⎣ 0.3 0.3 0.2 0.2 ⎥⎦

We see max {min(a ik , b kj )} ≠ min {max (a ik , b kj )} from the


k k
examples 1.1.10 and 1.1.11.

Note: It is important to note that the expert who works with


the fuzzy models may choose to have the max min operator or
the min max operator. For we see in general the resultant given
by max min operator will be always greater than or equal to the
min max operator. That is in other words we can say that min
max operator in general will yield a value which will always be
less than or equal to the max min operator.
Now we have to answer the question will max min operator
be always defined for any two fuzzy matrices. We see the max
min operator or the min max operator will not in general be
defined for any fuzzy matrices A and B. The max min operator
{A, B} or the min max operator {A, B} will be defined if and
only if the number of columns in A equal to the number of rows
of B otherwise max min or min max of A, B will not be defined.
Further max min operator of A, B in general will not be equal to
the max min operator of B, A. In general even if max min
operator of A, B be defined than max min operator of B, A may

17
not be even defined. This is true of min max operator also we
see max min of A, B is defined in the example 1.1.10 but max
min {B, A} cannot be found at all i.e., max min {A B} is not
defined.
Similarly in example 1.1.11 we see min max of A, B is
defined where as min max {B, A} is not defined. Thus we can
say that for any two fuzzy matrices A and B, in general max
min (A, B) ≠ max min (B, A) (may not be even compatible).
Like wise min max {A, B} ≠ min max (B, A) (may not be
defined or compatible at all).
Now we see even if both max min (A, B) and max min (B,
A) are defined they may not be equal in general.

For we will give an example to this effect.

Example 1.1.12: Let


⎡ 0.3 0.1 1 ⎤
A = ⎢⎢ 0.6 0.3 0.8⎥⎥
⎢⎣ 0 0.4 0.5⎥⎦
and
⎡ 1 0.4 0.9 ⎤
B = ⎢⎢ 0.8 0.6 0.2 ⎥⎥
⎢⎣ 0.7 0.4 1 ⎥⎦

be any two 3 × 3 fuzzy matrices. Now

max min (A, B)

⎧ ⎧ ⎡ 0.3 0.1 1 ⎤ ⎡ 1 0.4 0.9 ⎤ ⎫ ⎫


⎪ ⎪ ⎪⎪
= max ⎨min ⎨ ⎢⎢0.6 0.3 0.8⎥⎥ , ⎢⎢ 0.8 0.6 0.2 ⎥⎥ ⎬⎬
⎪ ⎪ ⎢ 0 0.4 0.5⎥ ⎢0.7 0.4 1 ⎥ ⎪⎪
⎩ ⎩⎣ ⎦ ⎣ ⎦ ⎭⎭

⎡ 0.7 0.4 1 ⎤
= ⎢⎢ 0.7 0.4 0.8⎥⎥ .
⎣⎢ 0.5 0.4 0.5⎥⎦

18
max min (B, A)

⎧ ⎧ ⎡ 1 0.4 0.9 ⎤ ⎡ 0.3 0.1 1 ⎤ ⎫ ⎫


⎪ ⎪ ⎪⎪
= max ⎨ min ⎨ ⎢⎢ 0.8 0.6 0.2 ⎥⎥ , ⎢⎢0.6 0.3 0.8⎥⎥ ⎬⎬
⎪ ⎪ ⎢0.7 0.4 1 ⎥ ⎢ 0 0.4 0.5⎥ ⎪⎪
⎩ ⎩⎣ ⎦ ⎣ ⎦ ⎭⎭

⎡ 0.4 0.4 1 ⎤
= ⎢⎢ 0.6 0.3 0.8 ⎥⎥ .
⎢⎣ 0.4 0.4 0.7 ⎥⎦

We see max min (A, B) ≠ max (min) (B, A).

⎡ 0.7 0.4 1 ⎤ ⎡ 0.4 0.4 1 ⎤


⎢ 0.7 0.4 0.8⎥ ≠ ⎢0.6 0.3 0.8 ⎥ .
⎢ ⎥ ⎢ ⎥
⎢⎣ 0.5 0.4 0.5⎥⎦ ⎢⎣ 0.4 0.4 0.7 ⎥⎦

We see in general min max (A, B) ≠ min max (A, B). We


illustrate this by the following.

Example 1.1.13: Let


⎡ 1 0.3 0.2 ⎤
A = ⎢⎢ 0.4 1 0.5⎥⎥
⎢⎣ 0.7 0.3 1 ⎥⎦
and
⎡ 0.3 1 0.8⎤
B = ⎢⎢0.7 0.7 1 ⎥⎥
⎣⎢ 1 0.6 0.3⎦⎥
be any two fuzzy matrices.
min max (A, B)
⎧ ⎧ ⎡ 1 0.3 0.2 ⎤ ⎡ 0.3 1 0.8⎤ ⎫ ⎫
⎪ ⎪ ⎪⎪
= min ⎨ max ⎨ ⎢⎢ 0.4 1 0.5⎥⎥ , ⎢⎢ 0.7 0.7 1 ⎥⎥ ⎬⎬
⎪ ⎪ ⎢0.7 0.3 1 ⎥ ⎢ 1 0.6 0.3⎥ ⎪⎪
⎩ ⎩⎣ ⎦ ⎣ ⎦ ⎭⎭

19
⎡ 0.7 0.6 0.3⎤
= ⎢⎢ 0.4 0.6 0.5⎥⎥ .
⎢⎣ 0.7 0.7 0.8⎥⎦
Now we find
min max (B, A)

⎧ ⎧ ⎡ 0.3 1 0.8⎤ ⎡ 1 0.3 0.2 ⎤ ⎫ ⎫


⎪ ⎪⎢ ⎪⎪
= min ⎨ max ⎨ ⎢0.7 0.7 1 ⎥⎥ , ⎢⎢ 0.4 1 0.5⎥⎥ ⎬⎬
⎪ ⎪ ⎢ 1 0.6 0.3⎥ ⎢ 0.7 0.3 1 ⎥ ⎪⎪
⎩ ⎩⎣ ⎦ ⎣ ⎦ ⎭⎭

⎡ 0.8 0.3 0.3⎤


= ⎢⎢ 0.7 0.7 0.7 ⎥⎥ .
⎢⎣ 0.6 0.3 0.6 ⎥⎦

We see clearly min max (A, B) ≠ min max (B, A).

⎡ 0.7 0.6 0.3⎤ ⎡ 0.8 0.3 0.3⎤


i.e. ⎢ 0.4 0.6 0.5⎥ ≠ ⎢⎢ 0.7 0.7 0.7 ⎥⎥ .
⎢ ⎥
⎢⎣ 0.7 0.7 0.8⎥⎦ ⎢⎣ 0.6 0.3 0.6 ⎥⎦

Now we also work with special type of products using fuzzy


matrices. Suppose we have a n × n square fuzzy matrix A = (aij);
aij ∈ {–1, 0, 1} and 1 ≤ i, j ≤ n.
Let X = [x1 x2 … xn] be a fuzzy row vector where xi ∈ {0,
1}, 1 ≤ i ≤ n. Now we find the product X o A = [y1 y2 … yn]
where the product X o A is the usual matrix product. Now yi’s 1
≤ i ≤ n need not in general belong to {0, 1}, we make the
following operations so that yi’s belong to {0, 1} which will be
known as the updating operation and thresholding operation.
We define if yi > 0 replace them by 1, 1 ≤ i ≤ n and if yi’s
are ≤ 0 then replace yi by 0. Now this operation is known as the
thresholding operation. It is important to mention here that in
general we can replace yi > 0 by a value say a > 0, a ∈ [0, 1] and
yi ≤ 0 by 0. It is left to the choice of the expert who works with

20
it. Now the updating operation is dependent on the X = [x1 …
xn] with which we started to work with. If xk in X was 1 in 1 ≤ k
≤ n; then we demand in the resultant Y = X o A = [y1 y2 … yn]
the yk must be 1, 1 ≤ k ≤ n, this operation on the resultant Y is
known as the updating operation.
Now using Y' which is the updated and thresholded fuzzy
row vector of Y = X o A we find Y' o A = Z (say) Z = [z1 z2 …
zn]. We now threshold and update Z to get Z' this process is
repeated till we arrive at a fixed point or a limit cycle i.e. if X o
A = Y, Y' after updating and thresholding Y we find Y' o A = Z,
if Z' is got after updating and thresholding Z we proceed on till
some R' o A = S and S' is Y' or Z' and this process will repeat,
then S' will be known as the limit cycle given by the fuzzy row
vector X using the fuzzy matrix A. [108, 112]

Now we illustrate this by the following example.

Example 1.1.14: Let A= (aij) be a 5 × 5 fuzzy matrix with aij ∈


{–1, 0, 1},
⎡0 1 0 −1 0⎤
⎢ −1 0 −1 0 1 ⎥⎥

i.e., A = ⎢ 0 −1 0 1 −1⎥ .
⎢ ⎥
⎢0 0 0 0 1⎥
⎢⎣ 1 −1 1 0 0 ⎥⎦

Let X = [0 1 0 0 1] be the fuzzy row vector (given), to find the


resultant of X using A.
⎡0 1 0 −1 0⎤
⎢ −1 0 −1 0 1 ⎥⎥

X o A = [ 0 1 0 0 1] o ⎢ 0 −1 0 1 −1⎥
⎢ ⎥
⎢0 0 0 0 1⎥
⎢⎣ 1 −1 1 0 0 ⎥⎦

= [0 –1 0 0 1]
= Y.

21
Now

Y' = [0 1 0 0 1]

which is got after updating and thresholding Y. Now

Y' o A = [0 –1 0 0 1]
= Z.

Let Z' be the vector got after thresholding and updating Z and
we see Z' =Y'. Thus it gives us the fixed point. Suppose X = [1
0 1 0 1] to find the effect of X on A.

⎡ 0 1 0 −1 0 ⎤
⎢ −1 0 −1 0 1 ⎥
⎢ ⎥
X o A = [1 0 1 0 1] o ⎢ 0 −1 0 1 −1⎥
⎢ ⎥
⎢0 0 0 0 1⎥
⎢⎣ 1 −1 1 0 0 ⎥⎦
= [1 –1 1 0 –1]
= Y.

Y' got by updating and thresholding Y is got as [1 0 1 0 1],


which shows X remains unaffected by this product.

Let X = [0 0 1 0 0] to find the effect of X on A.

⎡ 0 1 0 −1 0 ⎤
⎢ −1 0 −1 0 1 ⎥
⎢ ⎥
X o A = [ 0 0 1 0 0] o ⎢ 0 −1 0 1 −1⎥
⎢ ⎥
⎢0 0 0 0 1⎥
⎢⎣ 1 −1 1 0 0 ⎥⎦

= [0 –1 0 1 –1]
= Y;

after updating and thresholding Y we get Y' = [0 0 1 1 0].

22
Now
⎡ 0 1 0 −1 0 ⎤
⎢ −1 0 −1 0 1 ⎥
⎢ ⎥
Y' o A = [ 0 0 1 1 0] o ⎢ 0 −1 0 1 −1⎥
⎢ ⎥
⎢0 0 0 0 1⎥
⎢⎣ 1 −1 1 0 0 ⎥⎦
= [0 –1 0 1 0]
= Z.
After updating and thresholding Z we get Z' = [0 0 1 1 0].
Thus the special hidden pattern is a special fixed point given
by [0 0 1 1 0].
It is very usual we get back the same X after a sequence of
operations such cases one says the special hidden pattern is a
special limit cycle. Now we use after set of operations when the
given fuzzy matrix is not a square matrix. Let B = (bij) where bij
∈ {–1, 0, 1} and B is a n × m matrix m ≠ n with 1 ≤ i ≤ n and 1
≤ j ≤ m. Suppose X = [x1 … xn] be a fuzzy row vector, xi ∈ {0,
1}; 1 ≤ i ≤ n. Then X o B = Y = [y1 … ym] where yi ∉ {0, 1} we
update and threshold Y to Y'; Y' is a 1 × m fuzzy row vector.
We find Y' o BT = Z where Z = [z1 z2 … zn]. Let Z' be the
fuzzy row vector after thresholding and updating Z. Now we
calculate Z' o A = P where P is a 1 × m row matrix and P' is the
updated and thresholded resultant of P. Find P' o BT and so on is
continued till one arrives at a fixed binary pair or a limit bicycle.

We illustrate this by the following example.

Example 1.1.15: Let B = (bij) be a 6 × 4 fuzzy matrix, bij ∈ {–1,


0, 1} where

⎡ 1 −1 0 1 ⎤
⎢0 1 0 0⎥
⎢ ⎥
⎢ −1 0 1 0 ⎥
B=⎢ ⎥.
⎢ 0 0 0 −1⎥
⎢0 1 1 0⎥
⎢ ⎥
⎢⎣ 1 1 −1 1 ⎥⎦

23
Suppose X = [1 0 1 0 1 1] be given to find the resultant of X and
B.

⎡ 1 −1 0 1 ⎤
⎢0 1 0 0⎥
⎢ ⎥
⎢ −1 0 1 0 ⎥
X o B = [1 0 1 0 1 1] o ⎢ ⎥
⎢ 0 0 0 −1⎥
⎢0 1 1 0⎥
⎢ ⎥
⎣⎢ 1 1 −1 1 ⎦⎥
= [1 1 1 2]
= Y;

after thresholding Y we get Y' = [1 1 1 1]. Clearly as Y is not a


given fuzzy vector we need not update Y. Now we find

⎡1 0 −1 0 0 1⎤
⎢ −1 1 0 0 1 1 ⎥⎥
Y o BT = [1 1 1 1] o ⎢
⎢0 0 1 0 1 −1⎥
⎢ ⎥
⎣1 0 0 −1 0 1⎦

= [1 1 0 –1 2 2]
= Z.

Now Z' got after updating and thresholding Z is given by Z' = [1


1 1 0 1 1].
We find
⎡ 1 −1 0 1 ⎤
⎢0 1 0 0⎥
⎢ ⎥
⎢ −1 0 1 0 ⎥
Z' o B = [1 1 1 0 1 1] o ⎢ ⎥
⎢ 0 0 0 −1⎥
⎢0 1 1 0⎥
⎢ ⎥
⎣⎢ 1 1 −1 1 ⎦⎥
= [1 2 1 2]
= P.

24
P after thresholding we get P' = [1 1 1 1]. We see P' o BT gives
Q and Q' after thresholding and updating is [1 1 1 0 1 1] = Z'.
Thus we get the resultant as a fixed binary pair given by {[1 1 1
0 1 1], [1 1 1 1]}.
Now we study the effect of a 1 × 4 row vector T = [1 0 0 1]
on B
⎡ 1 0 −1 0 0 1 ⎤
⎢ −1 1 0 0 1 1 ⎥
T o BT = [1 0 0 1] o ⎢ ⎥
⎢ 0 0 1 0 1 −1⎥
⎢ ⎥
⎣ 1 0 0 −1 0 1 ⎦
= [2 0 –1 –1 0 2]
= X.

X’ after thresholding X is given by X' = [1 0 0 0 0 1].


Now effect of X' on B is given by
⎡ 1 −1 0 1 ⎤
⎢0 1 0 0⎥
⎢ ⎥
⎢ −1 0 1 0 ⎥
X' o B = [1 0 0 0 0 1] ⎢ ⎥
⎢ 0 0 0 −1⎥
⎢0 1 1 0⎥
⎢ ⎥
⎣⎢ 1 1 −1 1 ⎦⎥
= [2 0 –1 2]
= Y.

Y' = [1 0 0 1] the updated and thresholded vector which is T.


Thus we get the fixed binary pair in case of T = [1 0 0 1] is
given by {[1 0 0 1], [1 0 0 0 0 1]}.
We now proceed on to give yet another operation viz max
min operation on fuzzy matrices.
Let A = (aij) be a m × n fuzzy matrix, m = [m1 m2 … mn]
where aij ∈ [0, 1] and mk ∈ {0, 1} 1 ≤ i ≤ m, 1 ≤ j ≤ n and 1 ≤ k
≤ n, we calculate X = A o M that is xi = max min (aij, mj) where
1 ≤ j ≤ n and i = 1, 2, …, m.
X o A = Y is calculated using this X.

We explicitly show this by the following example.

25
Example 1.1.16: Let us consider the fuzzy matrix A = (aij)
where A is a 5 × 7 matrix with aij ∈ [0, 1].

⎡ 0.1 0.3 1 0.2 1 0 0.8⎤


⎢ 0.5 1 0.5 0.6 1 0.7 0.2 ⎥
⎢ ⎥
i.e. A = ⎢ 1 0.4 0.5 1 0.7 0.7 1 ⎥
⎢ ⎥
⎢ 0.7 0 1 0.2 0.6 1 0 ⎥
⎢⎣ 0.6 0.8 0.6 0.3 1 0.2 0.3⎥⎦

Let M = [1 0 0 1 0 1 1] be the given fuzzy vector.

AoM = max min {aij, mj}


= max ({0.1 0 0 0.2 0 0 0.8}, {0.5 0 0 0.6 0 0.7 0.2},
{1 0 0 1 0 0.7 1}, {0.7 0 0 0.2 0 1 0}, {0.6 0 0 0.3 0
0.2 0.3})
= [0.8 0.7 1 1 0.6]
= B
BoA = max min {bj aji}
= [1 0.7 1 1 0.8 1 1]
= M1.

Now we can find M1 o A and so on.

This type of operator will be used in the fuzzy models.

1.2 Special Classes of Fuzzy Matrices

In this section we introduce a special classes of fuzzy matrices.


We illustrate them with examples. Certain special type of
operations are defined on them which will be used in the fuzzy
special models which will be constructed in chapter two of this
book.

DEFINITION 1.2.1: Let M = M1 ∪ M2 ∪ … ∪ Mn; (n ≥ 2) where


each Mi is a t × t fuzzy matrix. We call M to be a special fuzzy t

26
× t square matrix i.e. M = M1 ∪ M2 ∪ … ∪ Mn = (m1ij) ∪ (m2ij)
∪ … ∪ (mnij), 1 ≤ i ≤ t and 1 ≤ j ≤ t.

We illustrate this by the following example.

Example 1.2.1: Let M = M1 ∪ M2 ∪ M3 ∪ M4 be a 4 × 4 square


fuzzy matrix where

⎡ 1 0.2 0.7 0.9 ⎤


⎢ 0.3 1 0.2 0.5⎥
M1 = ⎢ ⎥
⎢ 0.8 0.9 1 1 ⎥
⎢ ⎥
⎣ 0 0.6 0.4 0 ⎦

⎡ 0.3 0.4 0.5 0.6 ⎤


⎢ 1 0.7 0.9 0.5⎥
M2 = ⎢ ⎥
⎢ 0.2 0.3 0.4 1 ⎥
⎢ ⎥
⎣ 0.5 1 0 0.8⎦

⎡ 0.8 1 0 0.9 ⎤
⎢ 0.3 0.2 1 0.2 ⎥
M3 = ⎢ ⎥
⎢0.6 0.7 0.5 1 ⎥
⎢ ⎥
⎣1 0 0.8 0 ⎦
and
⎡1 0 1 0 ⎤
⎢ −1 0.7 0.2 0.5⎥
M4 = ⎢ ⎥.
⎢ 0.3 0.8 1 0.9 ⎥
⎢ ⎥
⎣ 0.1 0.6 0.7 1 ⎦

M is a special fuzzy 4 × 4 square matrix.

Now we proceed on to define the notion of special fuzzy row


matrix / vector and special fuzzy column vector / matrix.

27
DEFINITION 1.2.2: Let X = X1 ∪ X2 ∪ … ∪ XM (M ≥ 2) where
each Xi is a 1 × s fuzzy row vector / matrix then we define X to
be a special fuzzy row vector / matrix (i = 1, 2, …, M). If in
particular X = X1 ∪ X2 ∪ … ∪ XM (M ≥ 2) where each Xi is a 1
× si fuzzy row vector/matrix where for atleast one si ≠ sj with i ≠
j, 1 ≤ i, j ≤ M then we define X to be a special fuzzy mixed row
vector / matrix.

We illustrate first these two concepts by some examples.

Example 1.2.2: Let

X = X1 ∪ X2 ∪ X3 ∪ X4 ∪ X5
= [0.1 0.2 0.1 0.5 0.7] ∪ [1 0 1 0.9 1] ∪ [0 0.1 0.8 0.9 0.4]
∪ [0.8 0.6 0.4 0.5 0.7] ∪ [0.3 0.1 0.5 0.3 0.2]

where each Xi is a 1 × 6 fuzzy row vector; i = 1, 2, 3, 4, 5. X is a


special fuzzy row vector / matrix.

Now we proceed on to give an example of a special fuzzy


mixed row vector / matrix.

Example 1.2.3: Let

Y = Y1 ∪ Y2 ∪ Y3 ∪ Y4 ∪ Y5
= [1 0.3 0.2 0.9 1 0.3] ∪ [1 1 1 0.3] ∪ [0.3 1 0 0.2 0.5] ∪
[1 0.3 0.4] ∪ [1 0.8 0.9 0.7 0.6 0.5]

where Y1 is a 1 × 6 fuzzy row vector, Y2 is a 1 × 4 fuzzy row


vector, Y3 is a 1 × 5 fuzzy row vector, Y4 is a 1 × 3 fuzzy row
vector and Y5 is a l × 6 fuzzy row vector. We see Y is a special
fuzzy mixed row vector/matrix.

Now we proceed on to define the notion of special fuzzy


column vector/matrix and the notion of special fuzzy mixed
column vector/matrix.

DEFINITION 1.2.3: Let Y = Y1 ∪ Y2 ∪ … ∪ Ym (m ≥ 2) we have


each Yi to be a t × 1 fuzzy column vector/ matrix then we define

28
Y to be a special fuzzy column vector / matrix. If in particular in
Y = Y1 ∪ Y2 ∪ … ∪ Ym (m ≥ 2) we have each Yi to be ti × 1 fuzzy
column vector where atleast for one or some ti ≠ tj for i ≠ j, 1 ≤
i, j ≤ m then we define Y to be a special fuzzy mixed column
vector/matrix.

Now we proceed on to describe these two concepts with


examples.

Example 1.2.4: Let us consider

Z = Z1 ∪ Z2 ∪ Z3 ∪ Z4 ∪ Z5

⎡ 0.3⎤ ⎡ 1 ⎤ ⎡ 0.5 ⎤ ⎡ 0.8⎤ ⎡ 0.5⎤


⎢ 0.8 ⎥ ⎢ 0 ⎥ ⎢0.7 ⎥ ⎢ 0.1⎥ ⎢ 0.7 ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢ 0.1⎥ ⎢ 1 ⎥ ⎢ 0.6 ⎥ ⎢ 0.3⎥ ⎢ 0.8⎥
= ⎢ ⎥∪⎢ ⎥∪⎢ ⎥∪⎢ ⎥∪⎢ ⎥
⎢ 0.5 ⎥ ⎢ 1 ⎥ ⎢ 0.9 ⎥ ⎢ 0.1⎥ ⎢ 0.6 ⎥
⎢ 1 ⎥ ⎢0.8⎥ ⎢ 1 ⎥ ⎢ 0.8⎥ ⎢ 0.3⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢⎣ 0.7 ⎥⎦ ⎢⎣ 0.1⎥⎦ ⎢⎣ 0 ⎥⎦ ⎢⎣ 0.4 ⎥⎦ ⎢⎣ 0.1⎥⎦

where each Zi is a 6 × 1 fuzzy column vector. We see Z is a


special fuzzy column vector/matrix, 1≤ i ≤ 5.

Example 1.2.5: Let


T = T1 ∪ T2 ∪ T3 ∪ T4 ∪ T5 ∪ T6 ∪ T7

⎡ 0.8 ⎤
⎢ 0.71 ⎥ ⎡ 0.91⎤
⎢ ⎥ ⎢ ⎥
⎡ 0.3⎤ ⎢ 0.7 ⎥
⎢ 1 ⎥ ⎢ 0.512 ⎥
⎢ ⎥ ⎢ 0.11⎥ ⎡ 1 ⎤
⎢ ⎥ ⎢ ⎥
∪ ⎢ 0.44 ⎥ ∪ ⎢⎢ 0 ⎥⎥
⎢ 0.3 ⎥
= ⎢ 0 ⎥ ∪
⎢ ⎥ ⎢ 0.031 ⎥
⎢ ⎥ ⎢ 0.13⎥ ⎢⎣0.31⎥⎦
⎢ 0.8 ⎥ ⎢ ⎥
⎢⎣ 0.7 ⎥⎦ ⎢ 0.54 ⎥
⎢ 0.111⎥ ⎢ 0.71⎥
⎢ ⎥ ⎢ 0.8 ⎥
⎣ ⎦
⎣⎢ 0.14 ⎦⎥

29
⎡ 0.9 ⎤
⎢ 0.3 ⎥
⎡ 1 ⎤ ⎢ ⎥
⎢ 0.81⎥ ⎢ 0.81 ⎥
⎢ ⎥ ⎢ ⎥ ⎡ 0.9 ⎤
∪ ⎢⎢ 0.8 ⎥⎥ ;
⎢ 0.116 ⎥
∪ ⎢ 0.6 ⎥ ∪
⎢ ⎥ ⎢ 0.08 ⎥
⎢ 1 ⎥ ⎢ ⎥ ⎢⎣0.15⎥⎦
⎢⎣ 0.9 ⎥⎦ ⎢ 1 ⎥
⎢ 0 ⎥
⎢ ⎥
⎢⎣ 0.4 ⎥⎦

clearly T is a special fuzzy mixed column vector/matrix.

The following facts are important and is to be noted. We see in


a special fuzzy mixed column (row) matrix we can have two or
more fuzzy column (row) vector to have same number of rows.
But in the special fuzzy column (row) matrix we must have each
fuzzy column (row) vector should have the same number of
rows. Further in case of special fuzzy row (column)
vector/matrix we can also have the same fuzzy row (column)
vector to repeat itself.
For instance if
X = X1 ∪ X2 ∪ X3 ∪ X4

⎡1 ⎤ ⎡1 ⎤ ⎡1 ⎤ ⎡0⎤
⎢0⎥ ⎢0 ⎥ ⎢0⎥ ⎢1 ⎥
= ⎢ ⎥ ∪ ⎢ ⎥ ∪ ⎢ ⎥ ∪ ⎢ ⎥
⎢1 ⎥ ⎢1 ⎥ ⎢1 ⎥ ⎢1 ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎣1 ⎦ ⎣1 ⎦ ⎣1 ⎦ ⎣1 ⎦

then also X is a special fuzzy column matrix.


Like wise if

Y = Y1 ∪ Y2 ∪ Y3 ∪ Y4 ∪ X5
= [1 0 0 0 1 0.7 0.5] ∪ [1 0 0 0 1 0.7 0.5] ∪ [1 0 0 0 1 0.7
0.5] ∪ [1 0 0 0 1 0.7 0.5] ∪ [1 0 0 0 1 0.7 0.5] ∪ [1 0 0
0 1 0.7 0.5] ∪ [1 0 0 0 1 0.7 0.5],

30
we say Y is a special fuzzy row matrix, each Yi is 1 × 7 fuzzy
row vector and we have

Y1 = Y2 = Y3 = Y4 = Y5 = [1 0 0 0 1 0.7 0.5].

Now we proceed on to define the notion of special fuzzy mixed


square matrix.

DEFINITION 1.2.4: Let V = V1 ∪ V2 ∪ … ∪ Vn (n ≥ 2) where


each Vi is a ni × ni square fuzzy matrix where atleast one ni ≠ nj ,
i ≠ j, (1 ≤ i, j ≤ n). Then we define V to be a special fuzzy mixed
square matrix.

We illustrate this by the following example.

Example 1.2.6: Let


V = V1 ∪ V2 ∪ V3 ∪ V4 ∪ V5 ∪ V6

⎡ 0.9 1 0 0.9 ⎤
⎡ 0.3 0.8 1 ⎤ ⎢ 1 0.7 1 0.5⎥
= ⎢⎢ 0.9 1 0.5⎥⎥ ∪ ⎢ ⎥
⎢ 0.8 0.1 0.5 1 ⎥
⎢⎣ 0.2 0.4 0 ⎥⎦ ⎢ ⎥
⎣ 0.6 1 0.7 0 ⎦

⎡0.91 0 1 0.8 0.7 ⎤


⎢ 0 1 0 1 0 ⎥⎥
⎡ 0.91 0.82 ⎤ ⎢
∪ ⎢ ⎥ ∪ ⎢ 1 0.7 0.7 0.6 0.2 ⎥
⎣ 0.45 1 ⎦ ⎢ ⎥
⎢ 0.8 0.6 0.3 0.1 1 ⎥
⎢⎣ 0.3 0.1 0.11 0 0.3⎥⎦

⎡ 0.8 0.1 0.4 0.9 1 ⎤


⎢ 0.3 1 0.2 0.8 0.3⎥
⎡1 0 0.3⎤ ⎢ ⎥
∪ ⎢⎢0.9 1 0.8⎥⎥ ∪ ⎢ 0.4 0.2 1 0.2 0.5⎥ .
⎢ ⎥
⎣⎢ 1 0.4 1 ⎦⎥ ⎢ 0.3 1 0.3 0.6 1 ⎥
⎢⎣ 1 0.7 1 0.2 0.6 ⎥⎦

31
Clearly V is a special fuzzy square mixed matrix.

Now we proceed on to define the notion of special fuzzy


rectangular matrix.

DEFINITION 1.2.5: Let S = S1 ∪ S2 ∪ … ∪ Sm (m ≥ 2) where


each Si is a t × s rectangular fuzzy matrix t ≠ s; 1 ≤ i ≤ m then
we define S to be a special fuzzy rectangular matrix.

We illustrate this notion by an example.

Example 1.2.7: Let S = S1 ∪ S2 ∪ … ∪ S5

⎡0.2 1 0 ⎤ ⎡ 0.3 1 0.31⎤ ⎡ 1 0.5 1 ⎤


⎢ 0.1 0.1 1 ⎥ ⎢ 0.2 0 1 ⎥⎥ ⎢ 0 1 0.2 ⎥⎥
=⎢ ⎥ ∪ ⎢ ∪ ⎢
⎢ 0.3 1 0 ⎥ ⎢ 0.7 0.3 0.9 ⎥ ⎢ 0.1 0 0.3⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎣0.4 0 0.7 ⎦ ⎣ 0.2 1 0.1 ⎦ ⎣ 0.4 1 1 ⎦

⎡ 1 0.3 0.9 ⎤ ⎡ 0.3 0.2 0.9 ⎤


⎢ 0.6 0.2 0.1⎥ ⎢ ⎥
∪⎢ ⎥ ∪ ⎢ 0.8 1 0.4 ⎥ .
⎢0.7 0.8 0.5⎥ ⎢ 1 0.5 1 ⎥
⎢ ⎥ ⎢ ⎥
⎣ 0.1 1 0.4 ⎦ ⎣ 0.7 0.4 0.3⎦

We see each Si is a 4 × 3 rectangular fuzzy matrix, 1 ≤ i ≤ 5;


hence S is a special fuzzy rectangular matrix.

Now we proceed on to define the notion of special fuzzy mixed


rectangular matrix.

DEFINITION 1.2.6: Let P = P1 ∪ P2 ∪ … ∪ Pm where each Pi is


a si × ti rectangular fuzzy matrix (si ≠ ti), 1 ≤ i ≤ m and atleast
one Pi ≠ Pj for i ≠ j i.e., si ≠ sj or ti ≠ tj, 1 ≤ i, j≤ m, then we
define P to be a special fuzzy mixed rectangular matrix.

We now illustrate this by the following example.

32
Example 1.2.8: Let

S = S1 ∪ S2 ∪ S3 ∪ S4 ∪ S5 ∪ S6

⎡ 1 0.3⎤
⎢ 0.8 1 ⎥⎥
⎢ ⎡ 0.3 0.8 1 0.7 0.5 0.3 0.2 ⎤
= ⎢ 0.7 0.4 ⎥ ∪ ⎢
⎢ ⎥ ⎣ 0.1 1 0.9 0 1 0.9 0.14 ⎥⎦
⎢ 0.5 0.8⎥
⎢⎣ 1 0.9 ⎥⎦

⎡ 1 0.1⎤
⎡ 0.3 1 ⎤ ⎢0.7 1 ⎥
⎢ 0.8 0.9 ⎥ ⎢ ⎥
∪ ⎢ ⎥ ∪ ⎢ 0.8 0.7 ⎥
⎢1 0 ⎥ ⎢ ⎥
⎢ ⎥ ⎢ 0.4 0 ⎥
⎣ 0.9 0.4 ⎦ ⎢⎣ 1 0.2 ⎥⎦

⎡ 0.3 1 0.4 1 0.5 ⎤


⎡ 0.3 0.2 1 ⎤ ⎢ 0.9 0.4 1 0.8
⎢ 0.1 1 0.3⎥ ⎢ 1 ⎥⎥
∪ ⎢ ⎥ ∪ ⎢ 1 0.6 0.7 0.6 0.3⎥ .
⎢ 0.9 0 0.9 ⎥ ⎢ ⎥
⎢ ⎥ ⎢ 0.5 1 0.1 1 0.7 ⎥
⎣ 0.1 1 0.4 ⎦ ⎢⎣ 0.8 0.7 0 0.8 0.4 ⎥⎦

We see S is a special fuzzy mixed rectangular matrix.

Example 1.2.9: Let


X = X1 ∪ X2 ∪ X3 ∪ X4

⎡ 0.1 0.8⎤ ⎡ 0.1 0.8⎤ ⎡1 0⎤ ⎡ 0.9 0.3⎤


⎢ 0.9 1 ⎥ ⎢ 0.9 1 ⎥ ⎢ ⎥ ⎢ 0.8 1 ⎥
= ⎢ ⎥ ∪ ⎢ ⎥ ∪ ⎢0 1⎥
∪ ⎢ ⎥
⎢ 0.3 0.4 ⎥ ⎢ 0.3 0.4⎥ ⎢1 1⎥ ⎢ 0 0.1⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎣ 0.7 0 ⎦ ⎣ 0.7 0 ⎦ ⎣0 0⎦ ⎣1 0⎦

is the special fuzzy rectangular matrix.

33
We see
⎡ 0.1 0.8⎤
⎢ 0.9 1 ⎥
X1 = X2 = ⎢ ⎥
⎢ 0.3 0.4 ⎥
⎢ ⎥
⎣ 0.7 0 ⎦

a 4 × 2 fuzzy matrix.

Now we proceed on to define the notion of special fuzzy mixed


matrix.

DEFINITION 1.2.7: Let X = X1 ∪ X2 ∪ X3 ∪ …∪ Xn (n > 2)


where Xi is a ti × ti fuzzy square matrix and some Xj is a pj × qj
(pj ≠ qj) fuzzy rectangular matrix. Then we define X to be a
special fuzzy mixed matrix.

We now illustrate this by the following examples.

Example 1.2.10: Let


T = T1 ∪ T2 ∪ T3 ∪ T4 ∪ T5 =

⎡ 0.8 1 ⎤
⎢ 0.7 0.6 ⎥⎥
⎡ 0.3 1 0.8⎤ ⎢ ⎡ 1 0.3 0.4 0.7 0.5 ⎤
⎢ ⎥ ⎢ 0.5 0.4 ⎥ ⎢ ⎥
⎢ 1 0.9 0.5⎥ ∪ ⎢ 0.6 0.3⎥
⎥ ∪ ⎢ 0.3 1 0.3 0.5 1 ⎥
⎢⎣ 0.6 0.7 1 ⎥⎦ ⎢ ⎢⎣ 1 0 1 0 0.7 ⎥⎦
⎢ 0.2 0.1⎥
⎢ ⎥
⎣⎢ 1 0 ⎦⎥

⎡ 0.4 1 0.3 0.9 ⎤


⎢ 0.8 0.6 1 0.5⎥
∪ ⎢ ⎥ ∪ ⎡ 0.8 1 ⎤ ,
⎢ 0.1 1 0.8 1 ⎥ ⎢0.9 0.2 ⎥
⎣ ⎦
⎢ ⎥
⎣ 0.7 0.5 0.7 0.3 ⎦

T is a special fuzzy mixed matrix.

34
Example 1.2.11: Let
S = S1 ∪ S2 ∪ S3 ∪ S4

⎡ 0.1 0.2 ⎤
⎢ 0.4 0.5⎥⎥
⎡ 0.3 0.2 1 ⎤ ⎢
= ⎢⎢ 0.9 1 0.3⎥⎥ ∪ ⎢ 1 0.9 ⎥ ∪
⎢ ⎥
⎣⎢ 0.2 0.7 0.5⎦⎥ ⎢ 0.7 0.4 ⎥
⎢⎣ 1 0.8⎥⎦

⎡ 1 0 0.8⎤
⎢ 0 1 0.3⎥ ∪ ⎡0.8 1 ⎤ ;
⎢ ⎥ ⎢ 1 0.3⎥
⎢⎣0.7 0 1 ⎥⎦ ⎣ ⎦

S is a special fuzzy mixed matrix.

Example 1.2.12: Let


T = T1 ∪ T2 ∪ T3 ∪ T4 ∪ T5 =

⎡1 ⎤
⎢ 0 ⎥
⎢ ⎥
[1 0 0.3 1 0.5] ∪ ⎢ 0.8⎥ ∪ [ 0.3 1 0.8 0.5 0.31 0 1]
⎢ ⎥
⎢ 0.9 ⎥
⎢⎣ 0.6 ⎥⎦

⎡1 ⎤
⎢ 0 ⎥
⎢ ⎥
⎢ 0.8⎥
⎢ ⎥
∪ ⎢0.9 ⎥ ∪ [ 0.1 0.3 0.2 1 0.7 0.8 0.9 1]
⎢1 ⎥
⎢ ⎥
⎢ 0 ⎥
⎢ 0.1⎥
⎣ ⎦

35
T is a special fuzzy rectangular mixed matrix. We see T is not a
special fuzzy mixed matrix.

Now we proceed on to define how in the special fuzzy matrices


we get its transpose.

DEFINITION 1.2.8: Let P = P1 ∪ P2 ∪ … ∪ Pn (n > 2) be a


special fuzzy square (rectangular) matrix. Now the transpose of
P denoted by PT = ( P1 ∪ P2 ∪ … ∪ Pn ) = P1T ∪ P2T ∪ … ∪ PnT ,
T

PT is also a special fuzzy square (rectangular) matrix.

DEFINITION 1.2.9: Let X = X1 ∪ X2 ∪ … ∪ Xn be a special


fuzzy column matrix, then XT the transpose of X is
( X 1 ∪ X 2 ∪ … ∪ X n ) = X 1T ∪ X 2T ∪ … ∪ X nT = X T . Clearly XT
T

is a special fuzzy row matrix. Thus we see if Y = Y1 ∪ Y2 ∪ …


∪ Ym is a special fuzzy row matrix then YT =
(Y1 ∪ Y2 ∪ … ∪ Ym ) = Y1T ∪ Y2T ∪ … ∪ YmT is a special fuzzy
T

column matrix. Like wise if P = P1 ∪ P2 ∪ … ∪ Pm is a special


fuzzy mixed column matrix then the transpose of P, PT = PT1 ∪
PT2 ∪ … ∪ PTm is a special fuzzy mixed row matrix and vice
versa.

We just illustrate these by the following examples.

Example 1.2.13: Let T = T1 ∪ T2 ∪ T3 ∪ T4 = [1 0 0.3 0.7 1


0.8] ∪ [1 1 1 0 0.8 1] ∪ [0.8 0.1 1 0 0.7 0.9] ∪ [1 1 1 0 1 1] be a
special fuzzy mixed row matrix.
TT= ( T1 ∪ T2 ∪ T3 ∪ T4 ) = T1T ∪ T2T ∪ T3T ∪ T4T =
T

⎡ 1 ⎤ ⎡ 1 ⎤ ⎡ 0.8 ⎤ ⎡1 ⎤
⎢ 0 ⎥ ⎢ 1 ⎥ ⎢ 0.1⎥ ⎢1 ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢ 0.3⎥ ⎢ 1 ⎥ ⎢ 1 ⎥ ⎢1 ⎥
⎢ ⎥ ∪ ⎢ ⎥ ∪⎢ ⎥ ∪ ⎢ ⎥ .
⎢ 0.7 ⎥ ⎢ 0 ⎥ ⎢ 0 ⎥ ⎢0⎥
⎢ 1 ⎥ ⎢0.8⎥ ⎢0.7 ⎥ ⎢1 ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢⎣ 0.8 ⎥⎦ ⎢⎣ 1 ⎥⎦ ⎢⎣ 0.9 ⎥⎦ ⎢⎣1 ⎥⎦

36
Clearly TT is a special fuzzy mixed column matrix.

Example 1.2.14: Consider the special fuzzy mixed row matrix


P = P1 ∪ P2 ∪ P3 ∪ P4 ∪ P5
= [0 1 0.3 0.9] ∪ [1 0 1 1 0.9 0.3 0.7] ∪ [0 1 0.8] ∪ [0.9
0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1] ∪ [1 0.9 0.6].

Now the transpose of P is


⎡ 0.9 ⎤
⎢ 0.8 ⎥
⎡ 1 ⎤ ⎢ ⎥
⎢ 0 ⎥ ⎢ 0.7 ⎥
⎡ 0 ⎤ ⎢ ⎥ ⎢ ⎥
⎢ 1 ⎥ ⎢ 1 ⎥ ⎡ 0 ⎤ ⎢ 0.6 ⎥ ⎡ 1 ⎤
⎢ ⎥
P = ⎢ ⎥ ∪ ⎢ 1 ⎥ ∪ ⎢⎢ 1 ⎥⎥ ∪ ⎢ 0.5 ⎥ ∪ ⎢⎢0.9 ⎥⎥ .
T
⎢ 0.3⎥ ⎢ ⎥
⎢ ⎥ ⎢ 0.9 ⎥ ⎢⎣ 0.8⎥⎦ ⎢ 0.4 ⎥ ⎢⎣0.6 ⎥⎦
⎣ 0.9 ⎦ ⎢ ⎥ ⎢ 0.3⎥
⎢ 0.3⎥ ⎢ ⎥
⎢0.7 ⎥ ⎢ 0.2 ⎥
⎣ ⎦
⎢ 0.1⎥
⎣ ⎦

We see PT is a special fuzzy mixed column matrix.

It can be easily verified that the transpose of a special fuzzy


square (mixed) matrix will once again be a special fuzzy square
(mixed) matrix. Likewise the transpose of a special fuzzy
rectangular (mixed) matrix will once again will be a special
fuzzy rectangular (mixed).

We are forced to discuss about this because we will be


using these concepts in the special fuzzy models which we will
be constructing in chapter two.

Now we give an example of how the transpose of a special


fuzzy mixed matrix looks like.

Example 1.2.15: Let


S = S1 ∪ S2 ∪ S3 ∪ S4 ∪ S5

37
⎡0.9 ⎤ ⎡ 0.3 0.8 1 0.7 ⎤
⎢1 ⎥ ⎢ 0.1 0.3 0 1 ⎥⎥
= [1 0 0.9 0.2 1 0.7 0.8] ∪ ⎢ ⎥ ∪ ⎢
⎢0.2 ⎥ ⎢ 1 0.4 0.3 0.7 ⎥
⎢ ⎥ ⎢ ⎥
⎣ 0.1⎦ ⎣ 0.9 1 0.4 1 ⎦

⎡ 0.3 0.2 ⎤
⎢1 0.7 ⎥⎥
⎢ ⎡ 0.3 0.2 0.7 0.9 0.2 0.5 ⎤
⎢ 0.9 0.5 ⎥ ⎢ ⎥
∪ ⎢ ⎥ ∪ ⎢ 1 0.5 0.8 1 0.3 0.6 ⎥
⎢ 0.8 0.3⎥
⎢⎣ 0.9 0.6 0 0.1 0.4 0.7 ⎥⎦
⎢ 0.6 0.1⎥
⎢ ⎥
⎣⎢ 1 0 ⎦⎥

be a special fuzzy mixed matrix.

Now the transpose of S denoted by

ST = S1T ∪ ST2 ∪ S3T ∪ ST4 ∪ S5T


⎡ 1 ⎤
⎢ 0 ⎥
⎢ ⎥ ⎡ 0.3 0.1 1 0.9 ⎤
⎢ 0.9 ⎥ ⎢ 0.8 0.3 0.4 1 ⎥
⎢ ⎥
= ⎢ 0.2 ⎥ ∪ [ 0.9 1 0.2 0.1] ∪ ⎢ ⎥
⎢ 1 0 0.3 0.4 ⎥
⎢ 1 ⎥ ⎢ ⎥
⎢ ⎥ ⎣ 0.7 1 0.7 1 ⎦
⎢ 0.7 ⎥
⎢ 0.8 ⎥
⎣ ⎦
⎡ 0.3 1 0.9 ⎤
⎢ 0.2 0.5 0.6 ⎥⎥

⎡ 0.3 1 0.9 0.8 0.6 1 ⎤ ⎢ 0.7 0.8 0 ⎥
∪⎢ ⎥ ∪ ⎢ ⎥.
⎣0.2 0.7 0.5 0.3 0.1 0 ⎦ ⎢ 0.9 1 0.1⎥
⎢ 0.2 0.3 0.4 ⎥
⎢ ⎥
⎢⎣ 0.5 0.6 0.7 ⎥⎦

38
We see ST is also a special fuzzy mixed matrix.

Now we proceed on to introduce some special fuzzy operations


on special fuzzy square matrix which will be used on these
special fuzzy models. Thus at this juncture we do not promise to
give all types of operations that can be carried out on these class
of matrices we give here only those relevant operations on these
new class of special fuzzy matrices which will be described in
chapter two of this book. Operations on special fuzzy square
matrices with a special fuzzy operator 'o' which yields a fixed
point or a limit cycle.
Let
T = T1 ∪ T2 ∪ … ∪ Tm
be a special fuzzy square matrix where each Ti is a n × n
fuzzy matrix with entries from the set {–1, 0, 1}, 1 ≤ i ≤ m.
Now suppose we have a special fuzzy row vector say
X = X1 ∪ X2 ∪ … ∪ Xm

where X i = ⎡⎣ x1i x i2 … x in ⎤⎦ , 1 ≤ i ≤ m. To find the effect of X


on the special fuzzy square matrix T. It is to be noted that the
Xi’s need not all be distinct, we can have Xi = Xj without i = j.
Now the entries in each X i = ⎡⎣ x1i x i2 … x in ⎤⎦ is such that xik ∈
{0, 1}; i = 1, 2, …, m; k = 1, 2, …, n.

Now the operation

XoT = (X1 ∪ X2 ∪ … ∪ Xm) o (T1 ∪ T2 ∪ … ∪ Tn)


= X1 o T1 ∪ X2 o T2 ∪ … ∪ Xm o Tm.

The value of Xi o Ti is calculated as mentioned in pages 20-1 of


this book, for i = 1, 2, …, m.

We illustrate this by the following example.

Example 1.2.16: Let

T = T1 ∪ T2 ∪ T3 ∪ T4 ∪ T5

39
⎡0 1 0 −1⎤
0
⎢1 0 1 1 0 ⎥⎥

= ⎢0 1 0 −1 0 ⎥ ∪
⎢ ⎥
⎢0 0 0 0 1⎥
⎢⎣1 0 −1 0 0 ⎥⎦

⎡0 0 1 0 1⎤ ⎡0 1 0 1 0⎤
⎢ −1 0 0 ⎥
−1 0 ⎥ ⎢0 0 −1 0 1 ⎥⎥
⎢ ⎢
⎢0 0 0 1 0 ⎥ ∪ ⎢1 0 0 1 0⎥ ∪
⎢ ⎥ ⎢ ⎥
⎢1 0 0 0 1⎥ ⎢0 1 0 0 1⎥
⎢⎣ 0 1 0 0 0 ⎥⎦ ⎢⎣1 0 0 1 0 ⎥⎦

⎡0 0 0 1 1⎤ ⎡0 1 0 −1 0 ⎤
⎢ −1 0 1 0 0⎥ ⎥ ⎢1 0 −1 0 0 ⎥⎥
⎢ ⎢
⎢0 0 0 1 0⎥ ∪⎢0 1 0 −1 0 ⎥
⎢ ⎥ ⎢ ⎥
⎢0 1 0 0 −1⎥ ⎢0 0 −1 0 1 ⎥
⎢⎣ 1 −1 0 1 0 ⎥⎦ ⎢⎣ −1 0 0 1 0 ⎥⎦

be a special fuzzy square matrix. Let

X = X1 ∪ X2 ∪ X3 ∪ X4 ∪ X5
= [0 1 0 0 0] ∪ [1 0 0 0 1] ∪ [0 0 1 0 0] ∪ [0 1 0 0 0]
∪ [0 0 1 0 0]

be a special fuzzy row matrix.

To find X o T, to be in more technical terms the resultant of X


on T.
Now

XoT = {[X1 ∪ X2 ∪ X3 ∪ X4 ∪ X5] o [T1 ∪ T2 ∪ T3 ∪ T4


∪ T5]}
= X1 o T1 ∪ X2 o T2 ∪ X3 o T3 ∪ X4 o T4 ∪ X5 o T5

40
⎡0 1 0 −1⎤
0
⎢1 0 1 1 0 ⎥⎥

= [ 0 1 0 0 0] o ⎢ 0 1 0 −1 0 ⎥ ∪
⎢ ⎥
⎢0 0 0 0 1⎥
⎢⎣1 0 −1 0 0 ⎥⎦

⎡0 0 1 0 1⎤
⎢ −1 0 0 −1 0 ⎥⎥

[1 0 0 0 1] o ⎢ 0 0 0 1 0⎥ ∪
⎢ ⎥
⎢1 0 0 0 1⎥
⎢⎣ 0 1 0 0 0 ⎥⎦

⎡0 1 0 1 0⎤
⎢0 0 −1 0 1 ⎥⎥

[ 0 0 1 0 0] o ⎢1 0 0 1 0⎥ ∪
⎢ ⎥
⎢0 1 0 0 1⎥
⎢⎣1 0 0 1 0 ⎥⎦

⎡0 0 0 1 1⎤
⎢ −1 0 1 0 0 ⎥⎥

[ 0 1 0 0 0] o ⎢ 0 0 0 1 0⎥ ∪
⎢ ⎥
⎢0 1 0 0 −1⎥
⎢⎣ 1 −1 0 1 0 ⎥⎦

⎡0 1 0 −1 0 ⎤
⎢1 0 −1 0 0 ⎥⎥

[ 0 0 1 0 0] o ⎢ 0 1 0 −1 0 ⎥
⎢ ⎥
⎢0 0 −1 0 1 ⎥
⎢⎣ −1 0 0 1 0 ⎥⎦

41
= [1 0 1 1 0] ∪ [0 1 1 0 1] ∪ [1 0 0 1 0] ∪ [–1 0 1 0
0] ∪ [0 1 0 –1 0]
= Y'.

After updating and thresholding mentioned in pages 20-1 we get

Y = [1 1 1 1 0] ∪ [1 1 1 0 1] ∪ [1 0 1 1 0] ∪ [0 1 1 0 0]
∪ [0 1 1 0 0]
= Y1 ∪ Y2 ∪ Y3 ∪ Y4 ∪ Y5.

Now we find the effect of Yon T.

YoT = [Y1 ∪ Y2 ∪ Y3 ∪ Y4 ∪ Y5] o [T1 ∪ T2 ∪ T3 ∪ T4 ∪


T5 ]
= Y1 o T1 ∪ Y2 o T2 ∪ Y3 o T3 ∪ Y4 o T4 ∪ Y5 o T5

⎡0 1 0 −1⎤
0
⎢1 0 1 1 0 ⎥⎥

= [1 1 1 1 0] o ⎢0 1 0 −1 0 ⎥ ∪
⎢ ⎥
⎢0 0 0 0 1⎥
⎢⎣1 0 −1 0 0 ⎥⎦

⎡0 0 1 0 1⎤
⎢ −1 0 0 −1 0 ⎥⎥

[1 1 1 0 1] o ⎢ 0 0 0 1 0⎥ ∪
⎢ ⎥
⎢1 0 0 0 1⎥
⎢⎣ 0 1 0 0 0 ⎥⎦

⎡0 1 0 1 0⎤
⎢0 0 −1 0 1 ⎥⎥

[1 0 1 1 0] o ⎢1 0 0 1 0⎥ ∪
⎢ ⎥
⎢0 1 0 0 1⎥
⎢⎣1 0 0 1 0 ⎥⎦

42
⎡0 0 0 1 1⎤
⎢ −1 0 1 0 0 ⎥⎥

[ 0 1 1 0 0] o ⎢ 0 0 0 1 0⎥ ∪
⎢ ⎥
⎢0 1 0 0 −1⎥
⎢⎣ 1 −1 0 1 0 ⎥⎦

⎡0 1 0 −1 0 ⎤
⎢1 0 −1 0 0 ⎥⎥

[ 0 1 1 0 0] o ⎢ 0 1 0 −1 0 ⎥ .
⎢ ⎥
⎢0 0 −1 0 1 ⎥
⎢⎣ −1 0 0 1 0 ⎥⎦

= [1 2 1 0 0] ∪ [–1 1 1 0 1] ∪ [1 2 0 2 1] ∪ [–1 0 1 1
0] ∪ [1 1 –1 –1 0]
= Z',

after updating and thresholding Z' we get

Z = [1 1 1 0 0] ∪ [1 1 1 0 1] ∪ [1 1 1 1 1] ∪ [0 1 1 1 0]
∪ [1 1 1 0 0]
= Z1 ∪ Z2 ∪ Z3 ∪ Z4 ∪ Z5.

Now we find

ZoT = [Z1 ∪ Z2 ∪ Z3 ∪ Z4 ∪ Z5] o [T1 ∪ T2 ∪ T3 ∪ T4 ∪


T5 ]
= Z1 o T1 ∪ Z2 o T2 ∪ Z3 o T3 ∪ Z4 o T4 ∪ Z5 o T5

⎡0 1 0 −1⎤
0
⎢1 0 1 1 0 ⎥⎥

= [1 1 1 0 0] o ⎢0 1 0 −1 0 ⎥ ∪
⎢ ⎥
⎢0 0 0 0 1⎥
⎢⎣1 0 −1 0 0 ⎥⎦

43
⎡0 0 1 0 1⎤
⎢ −1 0 0 −1 0 ⎥⎥

[1 1 1 0 1] o ⎢ 0 0 0 1 0⎥ ∪
⎢ ⎥
⎢1 0 0 0 1⎥
⎢⎣ 0 1 0 0 0 ⎥⎦

⎡0 1 0 1 0⎤
⎢0 0 −1 0 1 ⎥⎥

[1 1 1 1 1] o ⎢1 0 0 1 0⎥ ∪
⎢ ⎥
⎢0 1 0 0 1⎥
⎢⎣1 0 0 1 0 ⎥⎦

⎡0 0 0 1 1⎤
⎢ −1 0 1 0 0 ⎥⎥

[ 0 1 1 1 0] o ⎢ 0 0 0 1 0⎥ ∪
⎢ ⎥
⎢0 1 0 0 −1⎥
⎢⎣ 1 −1 0 1 0 ⎥⎦

⎡0 1 −1 0 ⎤
0
⎢1 0 −1 0 0 ⎥⎥

[1 1 1 0 0] o ⎢ 0 1 0 −1 0 ⎥
⎢ ⎥
⎢0 0 −1 0 1 ⎥
⎢⎣ −1 0 0 1 0 ⎥⎦

= [1 2 1 0 –1] ∪ [–1 1 1 0 1] ∪ [2 2 –1 3 2] ∪ [–1 1 1


1 –1] ∪ [1 2 –1 –2 0]
= P' .

Let P be the special fuzzy row vector got by thresholding and


updating P'.

P = [1 1 1 0 0] ∪ [1 1 1 0 1] ∪ [1 1 1 1 1] ∪ [0 1 1 1 0]
∪ [1 1 1 0 0].

44
Thus we get the resultant of the special fuzzy row vector X to be
a special fixed point which is a special fuzzy row vector. This
sort of operations will be used in the special fuzzy cognitive
models which will be introduced in chapter two of this book.

Next we proceed on to give some special fuzzy operations in


case of special fuzzy mixed square matrix.
Let
P = P1 ∪ P2 ∪ … ∪ Pm
be a special fuzzy mixed square matrix where Pk = (pij), k = 1, 2,
…, m and Pk is a tk × tk square fuzzy matrix with pij ∈ {–1, 0,
1}, 1 ≤ i, j ≤ tk.
Let X = X1 ∪ X2 ∪ … ∪ Xm be a special fuzzy mixed row
vector such that each Xk is a 1 × tk fuzzy row vector k = 1, 2, …,
m with
X k = ⎡⎣ x1k x k2 … x kt k ⎤⎦
and x ik ∈ {0, 1}, 1 ≤ i ≤ tk; k = 1, 2, …, m.

Now
XoP = [X1 ∪ X2 ∪ … ∪ Xm] o [P1 ∪ P2 ∪ … ∪ Pm]
= X1 o P1 ∪ X2 o P2 ∪ … ∪ Xm o Pm.

The operation Xk o Pk is carried out and described in page 22.

Now we illustrate this explicitly by an example.

Example 1.2.17: Let


P = P1 ∪ P2 ∪ P3 ∪ P4

⎡0 1 0 0 1⎤
⎡0 1⎤ ⎢1 0 1 0 0 ⎥⎥
1 0
⎢1 0 −1 0 ⎥⎥ ⎢
= ⎢ ∪ ⎢0 0 0 1 1⎥ ∪
⎢ −1 1 0 0⎥ ⎢ ⎥
⎢ ⎥ ⎢1 1 0 0 0⎥
⎣0 0 1 0⎦
⎢⎣ 0 0 1 0 1 ⎥⎦

45
⎡0 1 1 0 0 0⎤
⎢1 0 0 0 1 0 ⎥⎥
⎢ ⎡0 0 0 1⎤
⎢0 0 0 1 0 1⎥ ⎢0 1 0 0 ⎥⎥
⎢ ⎥ ∪ ⎢
⎢1 0 0 0 1 0⎥ ⎢0 1 0 1⎥
⎢0 ⎥ ⎢ ⎥
1 0 1 0 0 ⎣1 0 1 0⎦
⎢ ⎥
⎢⎣ 0 0 1 0 1 0 ⎥⎦

be a special fuzzy mixed square matrix. Suppose


X = [1 0 0 0] ∪ [0 1 0 0 1] ∪ [0 0 1 1 0 0] ∪ [0 0 0 1]
= X1 ∪ X2 ∪ X3 ∪ X4,
be the special fuzzy mixed row vector.
To find
X o P = [X1 ∪ X2 ∪ X3 ∪ X4] o [P1 ∪ P2 ∪ P3 ∪ P4]
= X1 o P1 ∪ X2 o P2 ∪ X3 o P3 ∪ X4 o P4

⎡0 1 1⎤
0
⎢1 0 −1 0 ⎥⎥
= [1 0 0 0] o ⎢⎢ −1 1 0 0⎥

⎢ ⎥
⎣0 0 1 0⎦

⎡0 1 0 0 1⎤
⎢1 0 1 0 0 ⎥⎥

[0 1 0 0 1] o ⎢0 0 0 1 1⎥ ∪
⎢ ⎥
⎢1 1 0 0 0⎥
⎢⎣0 0 1 0 1 ⎥⎦

⎡0 1 1 0 0 0⎤
⎢1 0 0 0 1 0 ⎥⎥

⎢0 0 0 1 0 1⎥
[0 0 1 1 0 0] o ⎢ ⎥ ∪
⎢1 0 0 0 1 0⎥
⎢0 1 0 1 0 0⎥
⎢ ⎥
⎣⎢ 0 0 1 0 1 0 ⎦⎥

46
⎡0 0 0 1⎤
⎢0 1 0 0 ⎥⎥
[ 0 0 0 1] ⎢⎢0
o
1 0 1⎥
⎢ ⎥
⎣1 0 1 0⎦

= [0 1 0 1] ∪ [1 0 2 0 1] ∪ [1 0 0 1 1 1] ∪ [1 0 1 0]
= Y',

after updating and thresholding Y' we get

Y = Y1 ∪ Y2 ∪ Y3 ∪ Y4
= [1 1 0 1] ∪ [1 1 1 0 1] ∪ [1 0 1 1 1 1] ∪ [1 0 1 1].
Now we find
Y o P = (Y1 ∪ Y2 ∪ Y3 ∪ Y4) o (P1 ∪ P2 ∪ P3 ∪ P4)
= Y1 o P1 ∪ Y2 o P2 ∪ Y3 o P3 ∪ Y4 o P4

⎡0 1 1⎤
0
⎢1 0 −1 0 ⎥⎥
= [1 1 0 1] o ⎢⎢ −1 1 0 0⎥

⎢ ⎥
⎣0 0 1 0⎦

⎡0 1 0 0 1⎤
⎢1 0 1 0 0 ⎥⎥

[1 1 1 0 1] o ⎢0 0 0 1 1⎥ ∪
⎢ ⎥
⎢1 1 0 0 0⎥
⎢⎣ 0 0 1 0 1 ⎥⎦

⎡0 1 1 0 0 0⎤
⎢1 0 0 0 1 0 ⎥⎥

⎢0 0 0 1 0 1⎥
[1 0 1 1 1 1] o ⎢ ⎥ ∪
⎢1 0 0 0 1 0⎥
⎢0 1 0 1 0 0⎥
⎢ ⎥
⎢⎣0 0 1 0 1 0 ⎥⎦

47
⎡0 0 0 1⎤
⎢0 1 0 0 ⎥⎥
[1 0 1 1] ⎢⎢0
o
1 0 1⎥
⎢ ⎥
⎣1 0 1 0⎦

= [1 1 0 1] ∪ [1 1 2 1 3] ∪ [1 2 2 2 2 1] ∪ [1 1 1 2]
= R'.

Let R be the special fuzzy state vector obtained after


thresholding and updating R'

R = [1 1 0 1] ∪ [1 1 1 1 1] ∪ [1 1 1 1 1 1] ∪ [1 1 1 1].

We see R o P = R. It is left as an exercise for the reader to


calculate R o P. Thus we arrive at a special fixed point which is
also a special fuzzy mixed row vector.

Next we proceed on to define special fuzzy operations on


special fuzzy rectangular matrix R = R1 ∪ R2 ∪ … ∪ Rt (t ≥ 2)
where Ri’s are m × n rectangular fuzzy matrices i = 1, 2, …, t
(m ≠ n); Let X = X1 ∪ X2 ∪ … ∪ Xt where each Xi is a 1 × m
fuzzy row vector with entries of Xi from the set {0,1}. X is
clearly a special fuzzy row vector, 1 ≤ i ≤ t.
Now how to study the effect of X on R i.e., how does X
operate on R.
X o R = (X1 ∪ X2 ∪ … ∪ Xt) o (R1 ∪ R2 ∪ … ∪ Rt)
= X1 o R1 ∪ X2 o R2 ∪ … ∪ Xt o Rt

where Xi o Ri operation is described in pages 23-4 where i = 1,


2, …, t.
Now let

X o R = Y'
= Y'1 ∪ Y'2 ∪ … ∪ Y't,
clearly each Yi is a 1 × n row vector but now entries of Y'i
need not even belong to the unit fuzzy interval [0, 1]. So we
threshold and update Y'i as described in pages 20-1 for i = 1, 2,
3, …, t.

48
Now let the updated and thresholded Y’i be denoted by Yi =
[Yi1, Yi2, …, Yit] we see Yij ∈ {0, 1} for j = 1, 2, …, t, 1 ≤ i ≤ t.
Let Y = Y1 ∪ Y2 ∪ … ∪ Yt, Y is a special fuzzy row vector.
Now using Y we can work only on the transpose of R i.e.,
RT, otherwise we do not have the compatibility of ‘o’ operation
with R.
We find Y o RT = (Y1 ∪ Y2 ∪ … ∪ Yt) o (R1 ∪ R2 ∪ … ∪
Rt)T = Y1 o RT1 ∪ Y2 o RT2 ∪ … ∪ Yt o RTt = Z'1 ∪ Z'2 ∪ … ∪
Z't = Z' we see Z' in general need not be a special fuzzy row
vector. We update and threshold each Z'i to Zi (1 ≤ i ≤ t) to get Z
= Z1 ∪ Z2 ∪ … ∪ Zt. Z is a special fuzzy row vector. We can
work with Z as Z o R and so on until we arrive at an equilibrium
i.e., a special fixed point or a limit cycle.

We illustrate this by the following example.

Example 1.2.18: Let

T = T1 ∪ T2 ∪ T3 ∪ T4 ∪ T5

be a special fuzzy rectangular matrix where

⎡ 1 0 −1⎤ ⎡0 1 1⎤ ⎡0 1 0⎤
⎢0 1 0⎥ ⎢1 0 1 ⎥ ⎢1 0 0⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢ 0 −1 1 ⎥ ⎢1 1 0 ⎥ ⎢0 0 1⎥
T=⎢ ⎥ ∪ ⎢ ⎥ ∪ ⎢ ⎥
⎢1 0 1⎥ ⎢0 1 0⎥ ⎢ 1 −1 0 ⎥
⎢ −1 1 0 ⎥ ⎢1 0 0 ⎥ ⎢ −1 0 1 ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢⎣ 0 0 0 ⎥⎦ ⎢⎣ 0 0 1 ⎥⎦ ⎢⎣ 1 1 −1⎥⎦

⎡1 0 0⎤ ⎡ 1 1 −1⎤
⎢0 1 0⎥ ⎢0 0 1⎥
⎢ ⎥ ⎢ ⎥
⎢0 0 1⎥ ⎢1 0 0⎥
∪ ⎢ ⎥ ∪ ⎢ ⎥;
⎢ 1 −1 0 ⎥ ⎢0 1 0⎥
⎢ 0 1 −1⎥ ⎢ −1 0 1 ⎥
⎢ ⎥ ⎢ ⎥
⎢⎣ −1 0 1 ⎥⎦ ⎢⎣ 1 −1 0 ⎥⎦

49
T is clearly a 6 × 3 rectangular special fuzzy matrix with
entries from {–1, 0, 1}.
Let X = [1 0 0 1 0 0] ∪ [0 0 0 1 0 0] ∪ [0 0 0 0 0 1] ∪ [1 0
0 0 0 0] ∪ [0 1 0 0 0 0] = X1 ∪ X2 ∪ X3 ∪ X4 ∪ X5 be a special
fuzzy row vector with entries from the set {0, 1}.

The effect of X on T is given by

XoT = (X1 ∪ X2 ∪ X3 ∪ X4 ∪ X5) o (T1 ∪ T2 ∪ T3 ∪ T4 ∪


T5 )
= X1 o T1 ∪ X2 o T2 ∪ X3 o T3 ∪ X4 o T4 ∪ X5 o T5

⎡ 1 0 −1⎤
⎢0 1 0⎥
⎢ ⎥
⎢ 0 −1 1 ⎥
= [1 0 0 1 0 0] o ⎢ ⎥ ∪
⎢1 0 1⎥
⎢ −1 1 0 ⎥
⎢ ⎥
⎢⎣ 0 0 0 ⎥⎦

⎡0 1 1⎤
⎢1 0 1 ⎥⎥

⎢1 1 0⎥
[ 0 0 0 1 0 0] o ⎢ ⎥ ∪
⎢0 1 0⎥
⎢1 0 0⎥
⎢ ⎥
⎢⎣0 0 1 ⎥⎦

⎡0 1 0⎤
⎢1 0 0⎥
⎢ ⎥
⎢0 0 1⎥
[0 0 0 0 0 1] o ⎢ ⎥ ∪
⎢ 1 −1 0 ⎥
⎢ −1 0 1 ⎥
⎢ ⎥
⎣⎢ 1 1 −1⎥⎦

50
⎡1 0 0⎤
⎢0 1 0⎥
⎢ ⎥
⎢0 0 1⎥
[1 0 0 0 0 0] o ⎢ ⎥ ∪
⎢ 1 −1 0 ⎥
⎢ 0 1 −1⎥
⎢ ⎥
⎢⎣ −1 0 1 ⎥⎦

⎡ 1 1 −1⎤
⎢0 0 1⎥
⎢ ⎥
⎢1 0 0⎥
[ 0 1 0 0 0 0] o ⎢ 0 1 0 ⎥
⎢ ⎥
⎢ −1 0 1 ⎥
⎢ ⎥
⎢⎣ 1 −1 0 ⎥⎦

= [2 0 0] ∪ [0 1 0] ∪ [1 1 –1] ∪ [1 0 0] ∪ [0 0 1]
= Z'
= Z'1 ∪ Z'2 ∪ Z'3 ∪ Z'4 ∪ Z'5

clearly Z’ is not a special fuzzy row vector so we threshold it


and obtain
Z = Z1 ∪ Z2 ∪ … ∪ Z5
= [1 0 0] ∪ [0 1 0] ∪ [1 1 0] ∪ [1 0 0] ∪ [0 0 1].
Now
Z o TT = (Z1 ∪ Z2 ∪ Z3 ∪ Z4 ∪ Z5) o (TT1 ∪ TT2 ∪ … ∪ TT5)
= Z1 o TT1 ∪ Z2 o TT2 ∪ … ∪ Z5 o TT5

⎡ 1 0 0 1 −1 0 ⎤
= [1 0 0] o ⎢⎢ 0 1 −1 0 1 0⎥⎥ ∪
⎢⎣ −1 0 1 1 0 0 ⎥⎦

⎡ 0 1 1 0 1 0⎤
[0 1 0] o ⎢⎢ 1 0 1 1 0 0⎥⎥ ∪
⎢⎣ 1 1 0 0 0 1 ⎥⎦

51
⎡ 0 1 0 1 −1 1 ⎤
[1 1 0] o ⎢⎢1 0 0 −1 0 1 ⎥⎥ ∪
⎢⎣ 0 0 1 0 1 −1⎥⎦

⎡1 0 0 1 0 −1⎤
[1 0 0] o ⎢⎢0 1 0 −1 1 0 ⎥⎥ ∪
⎢⎣ 0 0 1 0 −1 1 ⎥⎦

⎡ 1 0 1 0 −1 1 ⎤
[0 0 1] o ⎢⎢ 1 0 0 1 0 −1⎥⎥
⎢⎣ −1 1 0 0 1 0 ⎥⎦

= [1 0 0 1 –1 0] ∪ [1 0 1 1 0 0] ∪ [1 1 0 0 –1 2] ∪ [1
0 0 1 0 –1] ∪ [–1 1 0 0 1 0]
= P’
= P’1 ∪ P’2 ∪ … ∪ P’5

clearly P’ is a not a special fuzzy row matrix.

We update and threshold P’ to obtain

P = P1 ∪ P2 ∪ P3 ∪ P4 ∪ P5
= [1 0 0 1 0 0] ∪ [1 0 1 1 0 0] ∪ [1 1 0 0 0 1] ∪ [1 0 0
1 0 0] ∪ [0 1 0 0 1 0].

P is a special fuzzy row vector. Now we can find P o T and so


on.

We now define special fuzzy operator on special fuzzy


mixed rectangular matrix
Suppose
P = P1 ∪ P2 ∪ … ∪ Pn

(n ≥ 2) be a special fuzzy mixed rectangular matrix i.e., each Pi


is a fuzzy rectangular si × ti (si ≠ ti) matrix and i = 1, 2, …, n.

52
Let X = X1 ∪ X2 ∪ … ∪ Xn be a special fuzzy mixed row
vector were each Xi is a l × si fuzzy row vector with entries from
the set {0, 1}; i = 1, 2, …, n.

Now we find

XoP = [X1 ∪ X2 ∪ … ∪ Xn] o [P1 ∪ P2 ∪ … ∪ Pn]


= X1 o P1 ∪ X2 o P2 ∪ … ∪ Xn o Pn.

Each Xi o Pi is obtained as described in pages 23-4; i = 1, 2, …,


n now let

XoP = Y'1 ∪ Y'2 ∪ … ∪ Y'n


= Y'.

Clearly Y' is not a special fuzzy row vector. We threshold Y'


and obtain

Y = Y1 ∪ Y2 ∪ … ∪ Yn;

Y is a special fuzzy mixed row vector with each Yi a 1 × ti fuzzy


row vector with entries from {0, 1}; i = 1, 2, …, n.

We find
Y o PT = (Y1 ∪ Y2 ∪ … ∪ Yn) o (PT1 ∪ PT2 ∪ … ∪ PTn)
= Y1 o PT1 ∪ Y2 P2T ∪ … ∪ Yn PnT .

Let
Y o PT = Z'1 ∪ Z'2 ∪ … ∪ Z'n
= Z';

Z' need not be a special fuzzy row vector we update and


threshold Z' and obtain Z = Z1 ∪ Z2 ∪ … ∪ Zn which is a
special fuzzy row vector and each Zi is a 1 × si fuzzy row vector
for i = 1, 2, …, n. Thus Z = Z1 ∪ Z2 ∪ … ∪ Zn is a special
fuzzy mixed row vector. Using Z we can find Z o P and so on.

Now we illustrate this by the following example.

53
Example 1.2.19: Let T be a special fuzzy mixed rectangular
matrix, where

⎡1 0 1 −1⎤
⎢0 1 0 0 ⎥⎥
⎢ ⎡ 0 1 0 1 0 −1 0 ⎤
⎢0 0 1 0⎥ ⎢ ⎥
T= ⎢ ⎥ ∪ ⎢1 0 −1 0 0 0 0 ⎥ ∪
⎢1 0 0 −1⎥
⎢ 0 0 1 0 1 0 −1⎥⎦
⎢0 1 0 1⎥ ⎣
⎢ ⎥
⎣⎢ −1 0 1 0 ⎦⎥

⎡0 1 0 0 1⎤
0
⎢1 0 −1 0 0 0 ⎥⎥ ⎡ −1 0 0 ⎤

⎢0 0 0 1 −1 0 ⎥ ⎢⎢ 0 1 0 ⎥⎥
⎢ ⎥
⎢0 1 0 0 0 −1⎥ ∪ ⎢ 1 −1 1 ⎥
⎢1 ⎢ ⎥
0 −1 0 1 0 ⎥ ⎢ 0 0 0 ⎥
⎢ ⎥
⎢ −1 1 −1 0 0 0 ⎥ ⎢⎣ 1 0 −1⎥⎦
⎢0 0 −1 1 1 ⎥⎦
⎣ 0

⎡ 1 0 0 0 −1⎤
⎢ 0 1 −1 0 0 ⎥
⎢ ⎥
⎢0 1 0 1 0⎥
⎢ ⎥
⎢ 1 0 0 0 −1⎥
∪ ⎢ 0 −1 0 1 0 ⎥
⎢ ⎥
⎢0 0 1 0 1⎥
⎢ −1 0 1 0 0 ⎥
⎢ ⎥
⎢ 0 0 1 −1 1 ⎥
⎢0 1 0 0 1⎥
⎣ ⎦

we see T1 is a 6 × 4 matrix fuzzy matrix, T2 a 3 × 7 matrix, T3 a


7 × 6 fuzzy matrix, T4 a 5 × 3 fuzzy matrix and a T5 a 9 × 5

54
fuzzy matrix. Thus T is a special fuzzy mixed rectangular
matrix.
Suppose

X = X1 ∪ X2 ∪ X3 ∪ X4 ∪ X5
= [1 0 0 1 0 0] ∪ [0 1 0] ∪ [0 0 1 1 0 0 0] ∪ [1 0 0 0
0] ∪ [1 0 0 0 0 0 0 0 0]

be a special fuzzy mixed row vector to find the special product


of X with T.

XoT = [X1 ∪ X2 ∪ … ∪ X5] o [T1 ∪ T2 ∪ … ∪ T5]


= X1 o T1 ∪ X2 o T2 ∪ X3 o T3 ∪ X4 o T4 ∪ X5 o T5

⎡1 0 1 −1⎤
⎢0 1 0 0 ⎥⎥

⎢0 0 1 0⎥
= [1 0 0 1 0 0] o ⎢ ⎥ ∪
⎢1 0 0 −1⎥
⎢0 1 0 1⎥
⎢ ⎥
⎣⎢ −1 0 1 0 ⎦⎥

⎡ 0 1 0 1 0 −1 0 ⎤
[0 1 0] o ⎢⎢1 0 −1 0 0 0 0 ⎥⎥ ∪
⎢⎣ 0 0 1 0 1 0 −1⎥⎦

⎡0 1 0 0 0 1⎤
⎢1 0 −1 0 0 0 ⎥⎥

⎢0 0 0 1 −1 0 ⎥
[0 0 1 1 0 0 0] o ⎢⎢ 0 ⎥
1 0 0 0 −1⎥ ∪
⎢1 0 −1 0 1 0 ⎥
⎢ ⎥
⎢ −1 1 −1 0 0 0 ⎥
⎢0 0 0 −1 1 1 ⎥⎦

55
⎡ −1 0 0 ⎤
⎢0 1 0⎥
⎢ ⎥
[1 0 0 0 0] o ⎢ 1 −1 1 ⎥ ∪
⎢ ⎥
⎢0 0 0⎥
⎢⎣ 1 0 −1⎥⎦

⎡ 1 0 0 0 −1⎤
⎢ 0 1 −1 0 0 ⎥
⎢ ⎥
⎢0 1 0 1 0⎥
⎢ ⎥
⎢ 1 0 0 0 −1⎥
[1 0 0 0 0 0 0 0 0 0] o ⎢⎢ 0 −1 0 1 0 ⎥⎥
⎢0 0 1 0 1⎥
⎢ −1 0 1 0 0 ⎥
⎢ ⎥
⎢ 0 0 1 −1 1 ⎥
⎢0 1 0 0 1⎥
⎣ ⎦

= [2 0 1 –2] ∪ [1 0 –1 0 0 0 0] ∪ [0 1 0 1 –1 –1] ∪
[–1 0 0] ∪ [1 0 0 0 –1]
= Y'1 ∪ Y'2 ∪ Y'3 ∪ Y'4 ∪ Y'5
= Y'.

Clearly Y' is not a special fuzzy mixed row vector. After


thresholding Y' we get

Y = [1 0 1 0] ∪ [1 0 0 0 0 0 0] ∪ [0 1 0 1 0 0] ∪ [0 0 0]
∪ [1 0 0 0 0].

Y is a special fuzzy mixed row vector. We find

Y o TT = [Y1 ∪ Y2 ∪ … ∪ Y5] o [T1T ∪ T2T ∪ T3T ∪ T4T ∪ T5T ]


= Y1 o T1T ∪ Y2 o T2T ∪ Y3 o T3T ∪ Y4 o T4T ∪ Y5 o T5T

56
⎡1 0 0 0 −1⎤
1
⎢0 1 0 0 1 0 ⎥⎥
= [1 0 1 0 ] ⎢⎢ 1
o
0 1 0 0 1⎥

⎢ ⎥
⎣ −1 0 0 −1 1 0 ⎦

⎡0 1 0⎤
⎢1 0 0⎥
⎢ ⎥
⎢ 0 −1 1 ⎥
[1 0 0 0 0 0 0] o ⎢⎢ 1 0 0 ⎥⎥ ∪
⎢0 0 1⎥
⎢ ⎥
⎢ −1 0 0 ⎥
⎢ 0 0 −1⎥
⎣ ⎦

⎡ 0 1 0 0 1 −1 0 ⎤
⎢1 0 0 1 0 1 0 ⎥
⎢ ⎥
⎢ 0 −1 0 0 −1 −1 0 ⎥
[0 1 0 1 0 0] o ⎢ ⎥ ∪
⎢ 0 0 1 0 0 0 −1⎥
⎢ 0 0 −1 0 1 0 1 ⎥
⎢ ⎥
⎣⎢1 0 0 −1 0 0 1 ⎥⎦

⎡ −1 0 1 0 1 ⎤
[0 0 0] o ⎢⎢ 0 1 −1 0 0 ⎥⎥ ∪
⎢⎣ 0 0 1 0 −1⎥⎦

⎡1 0 0 1 0 0 −1 0 0⎤
⎢0 1 1 0 −1 0 0 0 1 ⎥⎥

[1 0 0 0 0] o ⎢ 0 −1 0 0 0 1 1 1 0⎥ .
⎢ ⎥
⎢0 0 1 0 1 0 0 −1 0 ⎥
⎢⎣ −1 0 0 −1 0 1 0 1 1 ⎥⎦

57
= [2 0 1 1 0 0] ∪ [0 1 0] ∪ [1 0 1 1 0 1 –1] ∪ [0 0 0 0
0] ∪ [1 0 0 1 0 0 –1 0 0]
= Z'
= Z'1 ∪ Z'2 ∪ Z'3 ∪ Z'4 ∪ Z'5.

We see Z' is not a special fuzzy mixed row vector. Now we


update and threshold Z' to

Z = Z1 ∪ Z2 ∪ Z3 ∪ Z4 ∪ Z5
= [1 0 1 1 0 0] ∪ [0 1 0] ∪ [1 0 1 1 0 1 0] ∪ [1 0 0 0
0] ∪ [1 0 0 1 0 0 0 0 0]

which is a special fuzzy mixed row vector.

Now we can work with

ZoT = [Z1 ∪ Z2 ∪ … ∪ Z5] o [T1 ∪ T2 ∪ … ∪ T5]


= Z1 o T1 ∪ Z2 o T2 ∪ … ∪ Z5 o T5

and so on.
Thus we have seen a special type of operation on the special
fuzzy mixed row vector with special fuzzy mixed rectangular
matrix.

Now we proceed on to find yet another special type of


operations on the four types of special fuzzy matrices viz. the
special min max operator on the special fuzzy square matrix.
Suppose W = (W1 ∪ W2 ∪ … ∪ Wm) (m ≥ 2) be a special
fuzzy square matrix with entries from [0, 1]. Further let us
assume each Wi is a n × n square fuzzy matrix, 1 ≤ i ≤ m.
Let

X = X1 ∪ X2 ∪ … ∪ Xm

with Xj = ⎡⎣ x1j x 2j … x mj ⎤⎦ be a special fuzzy row vector, where


x tj ∈ [0, 1] for j = 1, 2, …, m and 1 ≤ t ≤ n. To use special min
max operator, to find the special product X and W.

58
Now

min max {X, W}

= min max {(X1 ∪ X2 ∪ … ∪ Xm), (W1 ∪ W2 ∪ … ∪ Wm)}


= min max{X1, W1} ∪ min max{X2, W2} ∪ … ∪ min
max{Xm, Wm}.

The min max operator has been defined between a fuzzy row
vector and a fuzzy square matrix in pages 14-21.

Now we illustrate this by the following example.

Example 1.2.20: Let W = W1 ∪ W2 ∪ W3 ∪ W4 ∪ W5 be a


special fuzzy square matrix where W

⎡ 0 0.3 0.5 0.1 0.2 ⎤


⎢ 1 0.4 0.6 0.8 0 ⎥⎥

= ⎢ 0.2 0.2 0 0.9 0.4 ⎥
⎢ ⎥
⎢ 0.7 0 0.5 0 0.8⎥
⎢⎣ 0.9 0.1 0 0.7 0 ⎥⎦

⎡ 1 0.2 0.4 0.6 0.8 ⎤


⎢ 0 0.3 0.5 0.7 0.9 ⎥⎥

∪ ⎢1 0 0.2 1 0.7 ⎥
⎢ ⎥
⎢ 0 0.5 0.4 0 0.3⎥
⎢⎣ 0.1 1 0.7 0.8 1 ⎥⎦

⎡ 0.3 1 0.2 1 0.4 ⎤


⎢ 1 0.2 0 0.5 0.7 ⎥
⎢ ⎥
∪ ⎢ 0 0.8 1 0.4 0.9 ⎥
⎢ ⎥
⎢ 0.8 1 0 0.7 0 ⎥
⎢⎣ 0.4 0 0.3 1 0.1⎥⎦

59
⎡ 1 0.2 0.4 0.7 ⎤1
⎢ 0 0.2 1 0.7 0.5 ⎥⎥

∪ ⎢ 0.7 1 0.5 1 0.4 ⎥
⎢ ⎥
⎢ 0.9 0.7 0.3 0 1 ⎥
⎢⎣ 1 0.2 0.4 0.7 0.4 ⎥⎦

⎡1 0 0.3 0.5 0.7 ⎤


⎢ 0 0.7 1 0.9 0.2 ⎥⎥

∪ ⎢ 0.4 0.6 0.8 1 0 ⎥
⎢ ⎥
⎢ 0.8 0.3 1 0.3 0.7 ⎥
⎢⎣ 1 0.4 0 0.2 1 ⎥⎦

Clearly entries of each Wi are in [0, 1], i = 1, 2, 3, 4, 5.

Let
X = X1 ∪ X2 ∪ X3 ∪ X4 ∪ X5
= [1 0 0 0 1] ∪ [0 1 0 1 0] ∪ [0 0 1 1 0] ∪ [1 0 1 0 0]
∪ [0 1 0 1 0]

be a special fuzzy row vector with entries form the set {0, 1}.
To find the special product of X with W using the min max
operator

min{max(X, W)}

= min{max(X1 ∪ X2 ∪ … ∪ X5), (W1 ∪ W2 ∪ … ∪ W5)}


= min max{X1, W1} ∪ min max{X2, W2} ∪ min max{X3,
W3} ∪ min max{X4, W4} ∪ min max{X5, W5}

⎧ ⎡ 0 0.3 0.5 0.1 0.2 ⎤ ⎫


⎪ ⎢ 1 ⎪
⎪ ⎢ 0.4 0.6 0.8 0 ⎥⎥ ⎪
⎪ ⎪
= min max ⎨[1 0 0 0 1] , ⎢ 0.2 0.2 0 0.9 0.4 ⎥ ⎬
⎪ ⎢ ⎥
⎢ 0.7 0 0.5 0 0.8⎥ ⎪
⎪ ⎪
⎪⎩ ⎢⎣ 0.9 0.1 0 0.7 0 ⎥⎦ ⎪⎭

60
⎧ ⎡ 1 0.2 0.4 0.6 0.8 ⎤ ⎫
⎪ ⎢ 0 0.3 0.5 0.7 ⎪
⎪ ⎢ 0.9 ⎥⎥ ⎪
⎪ ⎪
∪ min max ⎨[ 0 1 0 1 0] , ⎢1 0 0.2 1 0.7 ⎥ ⎬
⎪ ⎢ ⎥
⎢ 0 0.5 0.4 0 0.3⎥ ⎪
⎪ ⎪
⎪⎩ ⎢⎣0.1 1 0.7 0.8 1 ⎥⎦ ⎪⎭

⎧ ⎡ 0.3 1 0.2 1 0.4 ⎤ ⎫


⎪ ⎢ 1 0.2 0 0.5 0.7 ⎥ ⎪
⎪ ⎢ ⎥ ⎪⎪

∪ min max ⎨[ 0 0 1 1 0] , ⎢ 0 0.8 1 0.4 0.9 ⎥ ⎬
⎪ ⎢ ⎥
⎢ 0.8 1 0 0.7 0 ⎥ ⎪
⎪ ⎪
⎢⎣ 0.4 0 0.3 1 0.1⎥⎦ ⎪
⎩⎪ ⎭

⎧ ⎡ 1 0.2 0.4 1 0.7 ⎤ ⎫


⎪ ⎢ 0 ⎪
⎪ ⎢ 0.2 1 0.7 0.5 ⎥⎥ ⎪
⎪ ⎪
∪ min max ⎨[1 0 1 0 0] , ⎢0.7 1 0.5 1 0.4 ⎥ ⎬
⎪ ⎢ ⎥
⎢ 0.9 0.7 0.3 0 1 ⎥⎪
⎪ ⎪
⎪⎩ ⎢⎣ 1 0.2 0.4 0.7 0.4 ⎥⎦ ⎪⎭

⎧ ⎡1 0 0.3 0.5 0.7 ⎤ ⎫


⎪ ⎢ 0 ⎪
⎪ ⎢ 0.7 1 0.9 0.2 ⎥⎥ ⎪
⎪ ⎪
∪ min max ⎨[ 0 1 0 1 0] , ⎢ 0.4 0.6 0.8 1 0 ⎥⎬
⎪ ⎢ ⎥
⎢ 0.8 0.3 1 0.3 0.7 ⎥ ⎪
⎪ ⎪
⎪⎩ ⎢⎣ 1 0.4 0 0.2 1 ⎥⎦ ⎪⎭

= [0.2 0 0 0 0] ∪ [0.1 0 0.2 0.6 0.7] ∪ [0.3 0 0 0.5


0.4] ∪ [0 0.2 0.3 0 0.4] ∪ [0.4 0 0 0.2 0]
= Y
= Y1 ∪ Y2 ∪ Y3 ∪ Y4 ∪ Y5.

Now using the same min max operation, we can find min max
{Y o W} = Z and so on.

61
Here we define a special max min operator of special fuzzy
mixed square matrix. Let M = M1 ∪ M2 ∪ … ∪ Mn be a special
fuzzy mixed square matrix i.e., each Mi is a ti × ti fuzzy matrix; i
= 1, 2, …, n. Suppose X = X1 ∪ X2 ∪ … ∪ Xn be a special
fuzzy mixed row vector i.e., each Xi is a ⎡⎣ x1i , … , x iti ⎤⎦ fuzzy row
vector x ik ∈ {0, 1}, i = 1, 2, …, n; 1 ≤ k ≤ ti.

max min {X , M}
= max min{X1, M1} ∪ max min{X2, M2} ∪ … ∪
max min{Xn, Mn}

where max min {Xi, Mi} is found in pages 25 and 26 of this


book, i = 1, 2, …, n.

Now we illustrate this situation by the following example.

Example 1.2.21: Let M = M1 ∪ M2 ∪ M3 ∪ M4 ∪ M5 ∪ M6 be


a special fuzzy mixed square matrix; that is entries in each Mi
are from the unit interval [0, 1]; i = 1, 2, …, 6. Suppose X = X1
∪ X2 ∪ … ∪ X6 be a special fuzzy mixed row vector. We will
illustrate how max min {X, M} is obtained. Given

M = M1 ∪ M2 ∪ M3 ∪ M4 ∪ M5 ∪ M6

⎡ 0.4 0.6 1 0 0.5 ⎤


⎢ 0.2 0 0.6 1 0.7 ⎥⎥
⎢ ⎡ 0.3 1 0.8 ⎤
⎢ 1 0.3 1 0 0.5 ⎥ ⎢ ⎥
= ⎢ ⎥ ∪ ⎢ 1 0.4 0.6 ⎥
⎢ 0.6 1 0.3 1 0.2 ⎥
⎢⎣ 0.5 0 0.7 ⎥⎦
⎢ 0 0.4 0.2 0 1 ⎥
⎢ ⎥
⎢⎣ 0.5 1 0.6 0.7 0.5 ⎥⎦

62
⎡ 1 0.6 0.3 1 0.8⎤
⎡ 0.3 1 0.2 0.9 ⎤ ⎢ 0.3 1 0.6 0.7 1 ⎥
⎢ 1 0.8 1 0.6 ⎥ ⎢ ⎥
∪ ⎢ ⎥ ∪ ⎢0.6 0.5 0.4 0.5 0.5⎥
⎢ 0.9 0.6 0.1 0.5⎥ ⎢ ⎥
⎢ ⎥ ⎢ 0.8 0.2 0.1 0.2 0.1⎥
⎣ 0.7 0.8 0.7 0.3⎦ ⎢⎣ 0.5 0 0.3 0 0.4 ⎥⎦

⎡ 0.3 1 0.8⎤
⎡ 1 0.3⎤
∪ ⎢⎢ 1 0.3 0.4 ⎥⎥ ∪ ⎢
⎣ 0.7 0.2 ⎥⎦
⎢⎣ 0.7 0.6 0.2 ⎥⎦

is a special fuzzy mixed square matrix.


Given
X = X1 ∪ X2 ∪ X3 ∪ X4 ∪ X5 ∪ X6
= [1 0 0 0 0] ∪ [0 1 0] ∪ [0 0 1 0] ∪ [0 0 0 0 1] ∪
[0 0 1] ∪ [1 0]

be the special fuzzy mixed row vector. To find X o M using the


max min operation.

max min {X o M}
= max min {(X1 ∪ X2 ∪ X3 ∪ X4 ∪ X5 ∪ X6), (M1 ∪
M2 ∪ M3 ∪ M4 ∪ M5 ∪ M6)}

= max min(X1, M1) ∪ max min(X2, M2) ∪ … ∪ max


min(X6, M6)

⎧ ⎡0.4 0.6 1 0 0.5 ⎤ ⎫


⎪ ⎢0.2 0 0.6 1 ⎪
⎪ ⎢ 0.7 ⎥⎥ ⎪
⎪⎪ ⎢ 1 0.3 1 0 0.5 ⎥ ⎪⎪
= max min ⎨ [1 0 0 0 0] , ⎢ ⎥⎬
⎪ ⎢0.6 1 0.3 1 0.2 ⎥ ⎪
⎪ ⎢ 0 0.4 0.2 0 1 ⎥⎪
⎪ ⎢ ⎥⎪
⎩⎪ ⎢⎣ 0.5 1 0.6 0.7 0.5 ⎥⎦ ⎭⎪

63
⎧ ⎡ 0.3 1 0.8 ⎤ ⎫
⎪ ⎢ 1 0.4 0.6 ⎥ ⎪
∪ max min ⎨ [ 0 1 0] , ⎢ ⎥⎬
⎪ ⎢⎣ 0.5 0 0.7 ⎥⎦ ⎪⎭

⎧ ⎡ 0.3 1 0.2 0.9 ⎤ ⎫


⎪ ⎢ 1 0.8 1 0.6 ⎥ ⎪
⎪ ⎥ ⎪⎬
∪ max min ⎨[ 0 0 1 0] , ⎢
⎪ ⎢ 0.9 0.6 0.1 0.5⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣ 0.7 0.8 0.7 0.3⎦ ⎪⎭

⎧ ⎡ 1 0.6 0.3 1 0.8⎤ ⎫


⎪ ⎢ 0.3 1 0.6 0.7 1 ⎥ ⎪
⎪ ⎢ ⎥ ⎪⎪

∪ max min ⎨ [ 0 0 0 0 1] , ⎢ 0.6 0.5 0.4 0.5 0.5 ⎥ ⎬
⎪ ⎢ ⎥⎪
⎪ ⎢ 0.8 0.2 0.1 0.2 0.1⎥ ⎪
⎪⎩ ⎢⎣ 0.5 0 0.3 0 0.4 ⎥⎦ ⎪

⎧ ⎡ 0.3 1 0.8⎤ ⎫
⎪ ⎢ 1 0.3 0.4 ⎥ ⎪
= max min ⎨[ 0 0 1] , ⎢ ⎥⎬
⎪ ⎢⎣0.7 0.6 0.2 ⎥⎦ ⎭⎪

⎧⎪ ⎡ 1 0.3⎤ ⎫⎪
∪ max min ⎨[1 0] , ⎢ ⎥⎬
⎩⎪ ⎣ 0.7 0.2 ⎦ ⎭⎪

= [0.4 0.6 1 0 0.5] ∪ [1 0.4 0.6] ∪ [0.9 0.6 0.1 0.5] ∪


[0.5 0 0.3 0 0.4] ∪ [0.7 0.6 0.2] ∪ [1 0.3]
= Y
= Y1 ∪ Y2 ∪ Y3 ∪ Y4 ∪ Y5 ∪ Y6;

which is a special fuzzy mixed row vector. Now one can work
with max min of Y o M.
This work of finding max min Y o M is left as an exercise
for the reader.

64
Now we proceed on to show how we use the special max min
operator and work with the special fuzzy rectangular matrix.
Let P = P1 ∪ P2 ∪ … ∪ Pn (n ≥ 2) be a special fuzzy
rectangular matrix where each Pi is a t × s rectangular matrix t
≠ s, i = 1, 2, …, n.
Suppose
X = X1 ∪ X2 ∪ … ∪ Xn
be a special fuzzy row matrix where each Xi is a 1 × t fuzzy row
vector i = 1, 2, …, n. To find using the special max min operator

max min{X o P}
= max min (X, P)
= max min{(X1 ∪ X2 ∪ … ∪ Xn ), (P1 ∪ P2 ∪ … ∪
Pn)}

(say) now Y is again a special fuzzy row vector with each Yj a 1


× s fuzzy row vector. j = 1, 2, …, n.
Now using Y one can find

max min {Y, PT}


= max min {(Y1 ∪ Y2 ∪ … ∪ Yn) , (P1T ∪ P2T ∪ …
∪ PnT)}
= max min {Y1, P1T} ∪ max min {Y2, P2T} ∪ … ∪
max min {Yn, PnT}
= (Z1 ∪ Z2 ∪ … ∪ Zn)
= Z (say).

Now once again Z is a special fuzzy row vector and each Zk is a


1 × t fuzzy row vector for k = 1, 2, …, n. Using Z one can find
max min Z, P and so on.

Now we illustrate the situation by an example.

Example 1.2.22: Let P = P1 ∪ P2 ∪ P3 ∪ P4 ∪ P5 be a special


fuzzy rectangular matrix. Suppose X = X1 ∪ X2 ∪ X3 ∪ X4 ∪
X5 be a special fuzzy row vector. Let each Pi be a 6 × 5
rectangular fuzzy matrix i = 1, 2, …, 5. Clearly each Xi is a 1 ×
6 fuzzy row vector i = 1, 2, …, 5.

65
Given

P = P1 ∪ P2 ∪ P3 ∪ P4 ∪ P5

⎡ 0.3 1 0.2 1 0.8⎤


⎢ 1 0.7 0.4 0 0.1⎥
⎢ ⎥
⎢ 0.5 1 0.8 0.9 1 ⎥
= ⎢ ⎥ ∪
⎢ 0.7 0.9 1 0.6 0 ⎥
⎢ 0 1 0.7 0 0.9 ⎥
⎢ ⎥
⎢⎣ 1 0.8 1 0.9 0 ⎥⎦

⎡1 0.8 0.6 0.4 0.2 ⎤


⎢ 0.3 0.5 1 0.7 0.9 ⎥⎥

⎢ 0 0.1 0.3 0 0.5⎥
⎢ ⎥ ∪
⎢1 0 0.4 0.6 1 ⎥
⎢ 0 0.7 0 0.9 0.1⎥
⎢ ⎥
⎢⎣ 0.2 0.8 1 0.4 0 ⎥⎦

⎡ 0 1 0.9 0.3 1 ⎤
⎢ 0.1 0.2 1 0.5 0.8⎥
⎢ ⎥
⎢ 0.7 1 0.7 1 0.1⎥
⎢ ⎥ ∪
⎢ 1 0.6 0.4 0.9 0.5⎥
⎢ 0.5 1 0 1 0.4 ⎥
⎢ ⎥
⎢⎣ 0 0.7 0.8 0 1 ⎥⎦

⎡ 0 0.8 0 10.7 ⎤
⎢ 0.9 1 0.2 0 0.4 ⎥⎥

⎢ 0.1 0.2 1 0.3 0 ⎥
⎢ ⎥ ∪
⎢ 0.5 0.6 0.7 1 0.8 ⎥
⎢1 0 0.9 0.4 1 ⎥
⎢ ⎥
⎣⎢ 0.2 0.4 1 0 0.6 ⎦⎥

66
⎡ 0.5 0.7 1 0.9 0 ⎤
⎢ 1 0.9 0 0.1 0.2 ⎥⎥

⎢ 0 1 0.2 0.4 0.6 ⎥
⎢ ⎥.
⎢ 0.8 0 1 0.5 0.4 ⎥
⎢ 1 0.9 0 0.4 1 ⎥
⎢ ⎥
⎢⎣ 0.7 0.6 0.4 1 0 ⎥⎦

We see that each Pi is a 6 × 5 fuzzy matrix. Given

X = [0 0 0 0 0 1] ∪ [1 0 0 0 0 0] ∪ [0 1 0 0 0 0] ∪ [0 0 1
0 0 0] ∪ [0 0 0 0 1 0]

to be a 1 × 6 special fuzzy row vector. Using special max min


operator we find
max min {X, P}
= max min {(X1 ∪ X2 ∪ X3 ∪ X4 ∪ X5) , (P1 ∪ P2 ∪
P3 ∪ P4 ∪ P5)}
= max min(X1, P1) ∪ max min(X2, P2) ∪ … ∪ max
min(X5, P5)

⎧ ⎡ 0.3 1 0.2 1 0.8⎤ ⎫


⎪ ⎢ 1 0.7 0.4 0 0.1⎥ ⎪
⎪ ⎢ ⎥⎪
⎪⎪ ⎢ 0.5 1 0.8 0.9 1 ⎥ ⎪⎪
= max min ⎨[ 0 0 0 0 0 1] , ⎢ ⎥⎬
⎪ ⎢ 0.7 0.9 1 0.6 0 ⎥ ⎪
⎪ ⎢ 0 1 0.7 0 0.9 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪⎩ ⎣⎢ 1 0.8 1 0.9 0 ⎥⎦ ⎪⎭

⎧ ⎡1 0.8 0.6 0.4 0.2 ⎤ ⎫


⎪ ⎢ 0.3 ⎪
⎪ ⎢ 0.5 1 0.7 0.9 ⎥⎥ ⎪
⎪⎪ ⎢ 0 0.1 0.3 0 0.5⎥ ⎪⎪
∪ max min ⎨[1 0 0 0 0 0] , ⎢ ⎥⎬
⎪ ⎢1 0 0.4 0.6 1 ⎥ ⎪
⎪ ⎢ 0 0.7 0 0.9 0.1⎥ ⎪
⎪ ⎢ ⎥⎪
⎩⎪ ⎣⎢ 0.2 0.8 1 0.4 0 ⎦⎥ ⎭⎪

67
⎧ ⎡ 0 1 0.9 0.3 1 ⎤ ⎫
⎪ ⎢ 0.1 0.2 1 0.5 0.8⎥ ⎪
⎪ ⎢ ⎥⎪
⎪⎪ ⎢ 0.7 1 0.7 1 0.1⎥ ⎪⎪
∪ max min ⎨[ 0 1 0 0 0 0] , ⎢ ⎥⎬
⎪ ⎢ 1 0.6 0.4 0.9 0.5⎥ ⎪
⎪ ⎢ 0.5 1 0 1 0.4 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎩⎪ ⎢⎣ 0 0.7 0.8 0 1 ⎦⎥ ⎪⎭

⎧ ⎡ 0 0.8 0 10.7 ⎤ ⎫
⎪ ⎢ 0.9 ⎪
⎪ ⎢ 1 0.2 00.4 ⎥⎥ ⎪
⎪⎪ ⎢ 0.1 0.2 1 0.3 0 ⎥ ⎪⎪
∪ max min ⎨[ 0 0 1 0 0 0] , ⎢ ⎥⎬
⎪ ⎢ 0.5 0.6 0.7 1 0.8 ⎥ ⎪
⎪ ⎢1 0 0.9 0.4 1 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎩⎪ ⎢⎣ 0.2 0.4 1 0 0.6 ⎥⎦ ⎭⎪

⎧ ⎡ 0.5 0.7 1 0.9 0 ⎤ ⎫


⎪ ⎢ 1 0.9 0 ⎪
⎪ ⎢ 0.1 0.2 ⎥⎥ ⎪
⎪⎪ ⎢ 0 1 0.2 0.4 0.6 ⎥ ⎪⎪
∪ max min ⎨[ 0 0 0 0 1 0] , ⎢ ⎥⎬
⎪ ⎢ 0.8 0 1 0.5 0.4 ⎥ ⎪
⎪ ⎢ 1 0.9 0 0.4 1 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪⎩ ⎣⎢ 0.7 0.6 0.4 1 0 ⎦⎥ ⎪⎭

= [1 0.8 1 0.9 0] ∪ [1 0.8 0.6 0.4 0.2] ∪ [0.1 0.2 1 0.5


0.8] ∪ [0.1 0.2 1 0.3 0] ∪ [1 0.9 0 0.4 1]
= Y1 ∪ Y2 ∪ Y3 ∪ Y4 ∪ Y5
= Y.

Y is again a special fuzzy row vector, each Yi is 1 × 5 fuzzy


vector; i = 1, 2, …, 5.

Now using the special max min operator we find

68
Y o PT = max min {(Y1 ∪ Y2 ∪ … ∪ Y5),
(P1
T
∪ P2T ∪ … ∪ P5T ) }
= max min {Y1, P } ∪ max min { Y2, P2T } ∪ … ∪
1
T

T
max min {Y5, P } 5

=
⎧ ⎡ 0.3 1 0.5 0.7 0 1 ⎤⎫
⎪ ⎢ 1 0.7 1 0.9 1 0.8⎥ ⎪
⎪ ⎢ ⎥ ⎪⎪

max min ⎨[1 0.8 1 0.9 0] , ⎢0.2 0.4 0.8 1 0.7 1 ⎥ ⎬
⎪ ⎢ ⎥
⎢1 0 0.9 0.6 0 0.9 ⎥ ⎪
⎪ ⎪
⎪⎩ ⎢⎣ 0.8 0.1 1 0 0.9 0 ⎥⎦ ⎪⎭

⎧ ⎡ 1 0.3 0 1 0 0.2 ⎤ ⎫
⎪ ⎢ 0.8 0.5 0.1 0 0.7 0.8⎥ ⎪
⎪ ⎢ ⎥ ⎪⎪

max min ⎨[1 0.8 0.6 0.4 0.2] , ⎢0.6 1 0.3 0.4 0 1 ⎥⎬
⎪ ⎢ ⎥⎪
⎪ ⎢0.4 0.7 0 0.6 0.9 0.4 ⎥ ⎪
⎢⎣0.2 0.9 0.5 1 0.1 0 ⎥⎦ ⎪
⎩⎪ ⎭

⎧ ⎡ 0 0.1 0.7 1 0.5 0 ⎤ ⎫
⎪ ⎢ 1 0.2 1 0.6 1 0.7 ⎥ ⎪
⎪ ⎢ ⎥ ⎪⎪

max min ⎨[ 0.1 0.2 1 0.5 0.8] , ⎢0.9 1 0.7 0.4 0 0.8 ⎥ ⎬
⎪ ⎢ ⎥
⎢ 0.3 0.5 1 0.9 1 0 ⎥⎪
⎪ ⎪
⎪⎩ ⎢⎣ 1 0.8 0.1 0.5 0.4 1 ⎥⎦ ⎪


⎧ ⎡ 0 0.9 0.1 0.5 1 0.2 ⎤ ⎫
⎪ ⎢ 0.8 1 0.2 0.6 0 0.4 ⎥ ⎪
⎪ ⎢ ⎥ ⎪⎪

max min ⎨[ 0.1 0.2 1 0.3 0] , ⎢ 0 0.2 1 0.7 0.9 1 ⎥ ⎬
⎪ ⎢ ⎥
⎢ 1 0 0.3 1 0.4 0 ⎥ ⎪
⎪ ⎪
⎪⎩ ⎢⎣ 0.7 0.4 0 0.8 1 0.6 ⎥⎦ ⎪

69
⎧ ⎡ 0.5 1 0 0.8 1 0.7 ⎤ ⎫
⎪ ⎢ 0.7 0.9 1 ⎪
⎪ ⎢ 0 0.9 0.6 ⎥⎥ ⎪
⎪ ⎪
max min ⎨[1 0.9 0 0.4 1] , ⎢ 1 0 0.2 1 0 0.4 ⎥ ⎬
⎪ ⎢ ⎥
⎢ 0.9 0.1 0.4 0.5 0.4 1 ⎥ ⎪
⎪ ⎪
⎪⎩ ⎢⎣ 0 0.2 0.6 0.4 1 0 ⎥⎦ ⎪⎭

= [0.9 1 0.9 1 0.8 1] ∪ [1 0.6 0.3 1 0.7 0.8] ∪ [0.9 1


0.7 0.5 0.5 0.8] ∪ [0.3 0.2 1 0.7 0.9 1] ∪ [0.7 1 0.9
0.8 1 0.7]
= Z1 ∪ Z2 ∪ Z3 ∪ Z4 ∪ Z5
= Z;

is once again a special fuzzy row vector and each Zi in this case
is a 1 × 6 fuzzy vector. Now using Z we can calculate Z o P
using the special max min operator.

Now we define special max min operator when we have the


special fuzzy matrix to be a special fuzzy mixed rectangular
matrix and the special fuzzy row vector which works on it is a
special fuzzy mixed row vector.
Let V = V1 ∪ V2 ∪ V3 ∪ V4 ∪ … ∪ Vn where V is a special
fuzzy mixed rectangular matrix where each Vi is a si × ti fuzzy
matrix si ≠ ti, i = 1, 2, …, n. Let X = X1 ∪ X2 ∪ X3 ∪ X4 ∪ …
∪ Xn be a special fuzzy mixed row vector where each Xi is a 1 ×
si fuzzy row vector. To use the special max min operator and
find

max min {X, V}


= max min {X, V}
= max min {(X1 ∪ X2 ∪ … ∪ Xn), (V1 ∪ V2 ∪ … ∪
Vn)}
= max min {X1, V1} ∪ max min {X2, V2} ∪ … ∪
max min {Xn, Vn}
= Y1 ∪ Y2 ∪ … ∪ Yn
= Y

70
where Y is a special fuzzy mixed row vector with each Yj a 1 ×
ti fuzzy vector for j = 1, 2, …, n.

We now find max min {Y, VT} and so on, which will once
again give a special fuzzy mixed row vector and so on.

Now we illustrate this situation by the following example.

Example 1.2.23: Let us consider the special fuzzy mixed


rectangular matrix

V = V1 ∪ V2 ∪ V3 ∪ V4 ∪ V5

⎡ 0.3 0 1 0.8 ⎤
⎢ ⎥
⎡ 0.4 1 0.3⎤ ⎢ 1 0.9 0 0.1⎥
⎢ 1 0.2 0.5⎥ ⎢0.2 1 0.3 0.7 ⎥
⎢ ⎥ ⎢ ⎥
= ⎢ 0.6 1 0.2 ⎥ ∪ ⎢ 0.5 0 0.1 0 ⎥
⎢ ⎥ ⎢ 1 0.5 0.8 0 ⎥
⎢ 0 0.3 1 ⎥ ⎢ ⎥
⎢⎣ 0.7 1 0.3⎥⎦ ⎢ 0.3 0 0.9 1 ⎥
⎢0.2 0.5 1 0.7 ⎥
⎣ ⎦

⎡ 1 0.2 0.4 0.6 0.8 0 ⎤


⎢ 0 1 0.6 1 0.2 1 ⎥⎥
∪ ⎢
⎢ 0.2 0.3 0.5 0 0.9 0.3⎥
⎢ ⎥
⎣ 0.9 0.8 1 0.5 0 0.2 ⎦

⎡ 0.3 0.8 1 0.5 0.3⎤ ⎡ 0.2 0.3 0.1⎤


⎢ 0.1 1 0 1 0.7 ⎥⎥ ⎢ 1 0 1 ⎥⎥
⎢ ⎢
⎢ 0.4 0 0.2 0.2 1 ⎥ ⎢ 0.2 1 0.3⎥
∪ ⎢ ⎥ ∪ ⎢ ⎥.
⎢ 0.8 0.2 1 0 1 ⎥ ⎢ 1 0.2 1 ⎥
⎢ 0.9 0 0.3 1 0.4 ⎥ ⎢ 0.6 0.7 0 ⎥
⎢ ⎥ ⎢ ⎥
⎣⎢ 0.7 1 0 0.2 0.9 ⎦⎥ ⎣⎢ 0.7 1 0.4 ⎦⎥

71
Let

X = [1 0 0 0 1] ∪ [0 0 0 0 0 0 1] ∪ [0 1 0 0] ∪ [0 0 0 1 0
0] ∪ [0 0 0 0 1 0]

be the given special fuzzy mixed row vector. To find using the
special max min operator the value of X, V

max min (X, V)

= max min {(X1 ∪ X2 ∪ X3 ∪ X4 ∪ X5), (V1 ∪ V2 ∪


V3 ∪ V4 ∪ V5)}

= max min(X1, V1) ∪ max min(X2, V2) ∪ … ∪ max


min(X5, V5)

⎧ ⎡ 0.4 1 0.3⎤ ⎫
⎪ ⎢ 1 0.2 0.5⎥ ⎪
⎪ ⎢ ⎥ ⎪⎪

= max min ⎨[1 0 0 0 1] , ⎢ 0.6 1 0.2 ⎥ ⎬
⎪ ⎢ ⎥⎪
⎪ ⎢ 0 0.3 1 ⎥ ⎪
⎪⎩ ⎢⎣ 0.7 1 0.3⎥⎦ ⎪

⎧ ⎡ 0.3 0 1 0.8 ⎤ ⎫
⎪ ⎢ 1 0.9 0 0.1⎥ ⎪
⎪ ⎢ ⎥⎪
⎪ ⎢0.2 1 0.3 0.7 ⎥ ⎪
⎪ ⎢ ⎥⎪
∪ max min ⎨[ 0 0 0 0 0 0 1] , ⎢ 0.5 0 0.1 0 ⎥ ⎬
⎪ ⎢ 1 0.5 0.8 0 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪ ⎢ 0.3 0 0.9 1 ⎥ ⎪
⎪ ⎢0.2 0.5 1 0.7 ⎥ ⎪
⎩ ⎣ ⎦⎭

⎧ ⎡ 1 0.2 0.4 0.6 0.8 0 ⎤ ⎫


⎪ ⎢ 0 ⎪
⎪ 1 0.6 1 0.2 1 ⎥⎥ ⎪
∪ max min ⎨[ 0 1 0 0] , ⎢ ⎬
⎪ ⎢ 0.2 0.3 0.5 0 0.9 0.3⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣ 0.9 0.8 1 0.5 0 0.2 ⎦ ⎪⎭

72
⎧ ⎡ 0.3 0.8 1 0.5 0.3⎤ ⎫
⎪ ⎢ 0.1 1 ⎪
⎪ ⎢ 0 1 0.7 ⎥⎥ ⎪
⎪⎪ ⎢ 0.4 0 0.2 0.2 1 ⎥ ⎪⎪
∪ max min ⎨[ 0 0 0 1 0 0] , ⎢ ⎥⎬
⎪ ⎢ 0.8 0.2 1 0 1 ⎥⎪
⎪ ⎢ 0.9 0 0.3 1 0.4 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎩⎪ ⎣⎢ 0.7 1 0 0.2 0.9 ⎦⎥ ⎭⎪

⎧ ⎡ 0.2 0.3 0.1⎤ ⎫


⎪ ⎢ 1 ⎪
⎪ ⎢ 0 1 ⎥⎥ ⎪
⎪⎪ ⎢ 0.2 1 0.3⎥ ⎪⎪
∪ max min ⎨[ 0 0 0 0 1 0] , ⎢ ⎥⎬
⎪ ⎢ 1 0.2 1 ⎥ ⎪
⎪ ⎢ 0.6 0.7 0 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪⎩ ⎢⎣ 0.7 1 0.4 ⎥⎦ ⎪⎭

= [0.7 1 0.3] ∪ [0.2 0.5 1 0.7] ∪ [0 1 0.6 1 0.2 1] ∪


[0.8 0.2 1 0 1] ∪ [0.6 0.7 0]
= Y1 ∪ Y2 ∪ Y3 ∪ Y4 ∪ Y5
= Y.

We see Y is once again a special fuzzy mixed row vector. Now


we see Y o V is not defined, only max min{Y, VT} is defined so
we find Y o VT using the special max min operator.

max min{Y, VT}

= max min {(Y1 ∪ Y2 ∪ Y3 ∪ Y4 ∪ Y5), (V1T ∪ V2T


∪ V3T ∪ V4T ∪ V5T)}
= max min {Y1, V1T} ∪ max min {Y2, V2T} ∪ … ∪
max min {Y5, V5T}

⎧ ⎡0.4 1 0.6 0 0.7 ⎤ ⎫


⎪ ⎢ 1 0.2 1 0.3 1 ⎥ ⎪
= max min ⎨[ 0.7 1 0.3] , ⎢ ⎥⎬
⎪ ⎢⎣ 0.3 0.5 0.2 1 0.3⎥⎦ ⎭⎪

73
∪ max min
⎧ ⎡ 0.3 1 0.2 0.5 1 0.3 0.2 ⎤ ⎫
⎪ ⎢ 0 0.9 1 ⎪
⎪ 0 0.5 0 0.5 ⎥⎥ ⎪
⎨[ 0.2 0.5 1 0.7 ] , ⎢ ⎬
⎪ ⎢1 0 0.3 0.1 0.8 0.9 1 ⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣ 0.8 0.1 0.7 0 0 1 0.7 ⎦ ⎪⎭

⎧ ⎡1 0 0.2 0.9 ⎤ ⎫
⎪ ⎢0.2 1 0.3 0.8⎥ ⎪
⎪ ⎢ ⎥⎪
⎪⎪ ⎢0.4 0.6 0.5 1 ⎥ ⎪⎪
∪ max min ⎨[ 0 1 0.6 1 0.2 1] , ⎢ ⎥⎬
⎪ ⎢0.6 1 0 0.5⎥ ⎪
⎪ ⎢ 0.8 0.2 0.9 0 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎩⎪ ⎢⎣ 0 1 0.3 0.2 ⎥⎦ ⎭⎪

∪ max min
⎧ ⎡ 0.3 0.1 0.4 0.8 0.9 0.7 ⎤ ⎫
⎪ ⎢0.8 1 ⎪
⎪ ⎢ 0 0.2 0 1 ⎥⎥ ⎪
⎪ ⎪
⎨[ 0.8 0.2 1 0 1] , ⎢1 0 0.2 1 0.3 0 ⎥ ⎬
⎪ ⎢ ⎥
⎢0.5 1 0.2 0 1 0.2 ⎥ ⎪
⎪ ⎪
⎪⎩ ⎢⎣ 0.3 0.7 1 1 0.4 0.9 ⎥⎦ ⎪⎭

⎧ ⎡ 0.2 1 0.2 1 0.6 0.7 ⎤ ⎫


⎪ ⎢ 0.3 0 1 0.2 0.7 1 ⎥ ⎪
∪ max min ⎨[ 0.6 0.7 0] , ⎢ ⎥⎬
⎪ ⎢⎣ 0.1 1 0.3 1 0 0.4 ⎥⎦ ⎪⎭

= [1 0.7 1 0.3 1] ∪ [1 0.5 0.7 0.2 0.8 0.9 1] ∪ [0.6 1


0.5 0.8] ∪ [1 0.7 1 1 0.8 0.9] ∪ [0.3 0.6 0.7 0.6 0.7
0.7]
= T1 ∪ T2 ∪ T3 ∪ T4 ∪ T5
= T;

is a special fuzzy mixed row vector. Now using this T and V


one can calculate the special max min value of {T, V} and so
on.

74
Now we proceed on to define the special min max operator
on the special class of fuzzy rectangular matrices and the special
class of fuzzy mixed rectangular matrices.

Let S = S1 ∪ S2 ∪ … ∪ Sn be the set of special fuzzy


rectangular matrices with each Si a t × s rectangular matrix,

X = X1 ∪ X2 ∪ … ∪ Xn;

where Xi is a 1 × t fuzzy row vector i = 1, 2, …, n, the special


fuzzy row vector. To define the special min max operator using
X and S.

min max (X, S)

= min max {(X1 ∪ X2 ∪ … ∪ Xn), (S1 ∪ S2 ∪ … ∪


Sn)}
= min max {X1, S1} ∪ min max {X2, S2} ∪ … ∪ min
max {Xn, Sn}.

Each min max {Xi, Si} is calculated in pages 14-21 of this book.

Now
min max (X, S)
= Y1 ∪ Y2 ∪ … ∪ Yn.
= Y

is a special fuzzy row vector; each Yi is a 1 × s fuzzy row vector


i = 1, 2, …, n. Special min max operator is obtained for {Y, ST}
and so on.

Now we illustrate this by the following example.

Example 1.2.24: Let W = W1 ∪ W2 ∪ … ∪ W6 be a special


fuzzy rectangular matrix where

75
⎡ 0.1 1 0 0.5 0.6 1 ⎤
⎢ 0.5 0 0.7 1 1 0.8⎥⎥
W = ⎢ ∪
⎢ 0.4 1 0.2 0.3 0.4 1 ⎥
⎢ ⎥
⎣ 1 0.6 1 0.7 1 0 ⎦

⎡ 1 0.6 0.9 0 0 0.9 ⎤


⎢ 0 0.8 1 0.6 0.9 0.2 ⎥
⎢ ⎥ ∪
⎢ 0.2 1 0 1 0 0.3⎥
⎢ ⎥
⎣ 0.4 0 0.5 0.2 0.4 1 ⎦

⎡ 1 0.2 0.6 0.4 0.8 0.9 ⎤


⎢ 0 1 0.7 0.9 0.5 0.3⎥⎥
⎢ ∪
⎢ 0.5 0 0.2 0 1 0.1⎥
⎢ ⎥
⎣ 0.6 0.8 1 0.5 0.3 0 ⎦

⎡ 0.8 0.1 0 1 0.9 0.5⎤


⎢ 0 1 0.7 0.5 0.7 0.3⎥⎥
⎢ ∪
⎢ 0.6 0.5 1 0 0.9 0.4 ⎥
⎢ ⎥
⎣ 0.7 0 0.6 0.3 1 0 ⎦

⎡ 1 0.8 0 1 0 0.8 ⎤
⎢0 0 1 1 0.1 1 ⎥⎥
⎢ ∪
⎢0 1 0.1 0.3 0.5 0.7 ⎥
⎢ ⎥
⎣ 0.3 0.2 0.4 0.2 1 0 ⎦

⎡ 0.3 1 0.2 0 0.1 1⎤


⎢ 0.5 0 0.4 1 0.2 0 ⎥⎥

⎢ 0.7 1 0.6 0 0.3 0⎥
⎢ ⎥
⎣ 0.9 0 0.8 1 0.4 1⎦

76
Given
X = X1 ∪ X2 ∪ … ∪ X6
= [1 0 0 0] ∪ [0 1 0 0] ∪ [0 0 1 0] ∪ [0 0 0 1] ∪
[0 1 1 0] ∪ [1 0 0 1]

a special fuzzy row vector. To find the special min max value of
X o W.
min max (X, W)
= min max {(X1 ∪ X2 ∪ … ∪ X6), (W1 ∪ W2 ∪ … ∪
W6)}
= min max {X1, W1} ∪ min max {X2, W2} ∪ … ∪
min max {X6, W6}

⎧ ⎡ 0.1 1 0 0.5 0.6 1 ⎤ ⎫


⎪ ⎢ 0.5 0 0.7 1 ⎪
⎪ 1 0.8⎥⎥ ⎪
= min max ⎨[1 0 0 0] , ⎢ ⎬
⎪ ⎢ 0.4 1 0.2 0.3 0.4 1 ⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣ 1 0.6 1 0.7 1 0 ⎦ ⎪⎭

⎧ ⎡ 1 0.6 0.9 0 0 0.9 ⎤ ⎫


⎪ ⎢ 0 0.8 1 0.6 0.9 ⎪
⎪ 0.2 ⎥⎥ ⎪
∪ min max ⎨[ 0 1 0 0] , ⎢ ⎬
⎪ ⎢ 0.2 1 0 1 0 0.3⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣ 0.4 0 0.5 0.2 0.4 0.1⎦ ⎪⎭

⎧ ⎡ 1 0.2 0.6 0.4 0.8 0.9 ⎤ ⎫


⎪ ⎢ 0 ⎪
⎪ 1 0.7 0.9 0.5 0.3⎥⎥ ⎪
∪ min max ⎨[ 0 0 1 0] , ⎢ ⎬
⎪ ⎢ 0.5 0 0.2 0 1 0.1⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣ 0.6 0.8 1 0.5 0.3 0 ⎦ ⎪⎭

⎧ ⎡ 0.8 0.1 0 1 0.9 0.5⎤ ⎫


⎪ ⎢ 0 ⎪
⎪ 1 0.7 0.5 0.7 0.3⎥⎥ ⎪
∪ min max ⎨[ 0 0 0 1] , ⎢ ⎬
⎪ ⎢ 0.6 0.5 1 0 0.9 0.4 ⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣0.7 0 0.6 0.3 1 0 ⎦ ⎪⎭

77
⎧ ⎡ 1 0.8 0 1 0 0.8 ⎤ ⎫
⎪ ⎢0 ⎪
⎪ 0 1 1 0.1 1 ⎥⎥ ⎪
∪ min max ⎨[ 0 1 1 0] , ⎢ ⎬
⎪ ⎢0 1 0.1 0.3 0.5 0.7 ⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣0.3 0.2 0.4 0.2 1 0 ⎦ ⎪⎭

⎧ ⎡ 0.3 1 0.2 0 0.1 1⎤ ⎫


⎪ ⎢ 0.5 ⎪
⎪ 0 0.4 1 0.2 0 ⎥⎥ ⎪
∪ min max ⎨[1 0 0 1] , ⎢ ⎬
⎪ ⎢ 0.7 1 0.6 0 0.3 0⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣ 0.9 0 0.8 1 0.4 1 ⎦ ⎪⎭

= [0.4 0 0.2 0.3 0.4 0] ∪ [0.2 0 0 0 0 0.1] ∪ [0 0.2 0.6


0.4 0.3 0] ∪ [0 0.1 0 0 0.7 0.3] ∪ [0.3 0.2 0 0.2 0 0]
∪ [0.5 0 0.4 0 0.2 0]
= Z1 ∪ Z2 ∪ Z3 ∪ Z4 ∪ Z5 ∪ Z6
= Z;

is a special fuzzy row vector and each Zi is a 1 × 6 fuzzy row


vector; i = 1, 2, …, 6.
Now using Z we find min max of
min max {Z, WT}
= min max {(Z1 ∪ Z2 ∪ … ∪ Z6), (W1T ∪ W2T ∪ …
∪ W6T)
= min max {Z1, W1T} ∪ max mix {Z2, W2T} ∪ … ∪
max mix {Z6, W6T}

=
⎧ ⎡ 0.1 0.5 0.4 1 ⎤ ⎫
⎪ ⎢1 ⎪
⎪ ⎢ 0 1 0.6 ⎥⎥ ⎪
⎪⎪ ⎢ 0 0.7 0.2 1 ⎥ ⎪⎪
min max ⎨[ 0.4 0 0.2 0.3 0.4 0] , ⎢ ⎥⎬
⎪ ⎢ 0.5 1 0.3 0.7 ⎥ ⎪
⎪ ⎢0.6 1 0.4 1 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎩⎪ ⎢⎣ 1 0.8 1 0 ⎥⎦ ⎭⎪

78
⎧ ⎡1 0 0.2 0.4 ⎤ ⎫
⎪ ⎢ 0.6 ⎪
⎪ ⎢ 0.8 1 0 ⎥⎥ ⎪
⎪⎪ ⎢ 0.9 1 0 0.5 ⎥ ⎪⎪
∪ min max ⎨[ 0.2 0 0 0 0 0.1] , ⎢ ⎥⎬
⎪ ⎢ 0 0.6 1 0.2 ⎥ ⎪
⎪ ⎢ 0 0.9 0 0.4 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎩⎪ ⎣⎢ 0.9 0.2 0.3 0.1⎦⎥ ⎭⎪

⎧ ⎡1 0 0.5 0.6 ⎤ ⎫
⎪ ⎢ 0.2 1 ⎪
⎪ ⎢ 0 0.8⎥⎥ ⎪
⎪⎪ ⎢ 0.6 0.7 0.2 1 ⎥ ⎪⎪
∪ min max ⎨[ 0 0.2 0.6 0.4 0.3 0] , ⎢ ⎥⎬
⎪ ⎢ 0.4 0.9 0 0.5⎥ ⎪
⎪ ⎢ 0.8 0.5 1 0.3⎥ ⎪
⎪ ⎢ ⎥⎪
⎩⎪ ⎢⎣ 0.9 0.3 0.1 0 ⎥⎦ ⎪⎭

⎧ ⎡ 0.8 0 0.6 0.7 ⎤ ⎫


⎪ ⎢ 0.1 ⎪
⎪ ⎢ 1 0.5 0 ⎥⎥ ⎪
⎪⎪ ⎢ 0 0.7 1 0.6 ⎥ ⎪⎪
∪ min max ⎨[ 0 0.1 0 0 0.7 0.3] , ⎢ ⎥⎬
⎪ ⎢1 0.5 0 0.3⎥ ⎪
⎪ ⎢ 0.9 0.7 0.9 1 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪⎩ ⎣⎢ 0.5 0.3 0.4 0 ⎦⎥ ⎪⎭

⎧ ⎡1 0 0 0.3⎤ ⎫
⎪ ⎢ 0.8 0 ⎪
⎪ ⎢ 1 0.2 ⎥⎥ ⎪
⎪⎪ ⎢0 1 0.1 0.4 ⎥ ⎪⎪
∪ min max ⎨[ 0.3 0.2 0 0.2 0 0] , ⎢ ⎥⎬
⎪ ⎢1 1 0.3 0.2 ⎥ ⎪
⎪ ⎢ 0 0.1 0.5 1 ⎥⎪
⎪ ⎢ ⎥⎪
⎩⎪ ⎢⎣ 0.8 1 0.7 0 ⎦⎥ ⎭⎪

79
⎧ ⎡ 0.3 0.5 0.7 0.9 ⎤ ⎫
⎪ ⎢1 ⎪
⎪ ⎢ 0 1 0 ⎥⎥ ⎪
⎪⎪ ⎢ 0.2 0.4 0.6 0.8⎥ ⎪⎪
∪ min max ⎨[ 0.5 0 0.4 0 0.2 0] , ⎢ ⎥⎬
⎪ ⎢ 0 1 0 1 ⎥⎪
⎪ ⎢ 0.1 0.2 0.3 0.4 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎩⎪ ⎢⎣ 1 0 0 1 ⎦⎥ ⎪⎭

= [0.2 0 0.2 0] ∪ [0 0.2 0 0] ∪ [0.2 0 0.1 0] ∪ [0 0 0


0.1] ∪ [0 0.2 0.1 0] ∪ [0 0 0 0]
= T1 ∪ T2 ∪ T3 ∪ T4 ∪ T5 ∪ T6
= T;

where T is again a special fuzzy row vector and one can find
min max of {T, W} and so on.

Now we proceed on to work with special min max product


on a special fuzzy mixed rectangular matrix N = N1 ∪ N2 ∪ …
∪ Nt (t ≥ 2) where each Ni is a pi × qi rectangular fuzzy matrix i
= 1, 2, …, t. (pi ≠ qi).

Let X = X1 ∪ X2 ∪ … ∪ Xt be a special mixed row vector


where each Xi is 1 × pi fuzzy row vector i = 1, 2, …, t.

To find the special min max product of X with N ,

min max (X, N)

= min max {(X1 ∪ X2 ∪ … ∪ Xt), (N1 ∪ N2 ∪ … ∪


Nt)}
= min max {X1, N1} ∪ min max {X2, N2} ∪ … ∪
min max {Xt, Nt}
= Y1 ∪ Y2 ∪ … ∪ Yt.
= Y

where Y is a special fuzzy mixed row vector where each Yi is a


1 × qi fuzzy row vector i = 1, 2, …, t.

80
Now we will find special min max value of

min max {Y, NT}


= min max {(Y1 ∪ Y2 ∪ … ∪ Yt), (N1T ∪ N2T ∪ …
∪ NtT)}
= min max {Y1, N1T} ∪ min max {Y2, N2T} ∪ … ∪
min max {Yt, NtT}
= Z1 ∪ Z2 ∪ … ∪ Zt
= Z,

which is again a special fuzzy mixed row vector where each Zi


is a 1 × pi fuzzy row vector; i = 1, 2, …, t.
Now using Z and N we can find min max (Z, N) and so on.

We shall illustrate this by the following example.

Example 1.2.25: Let N = N1 ∪ N2 ∪ … ∪ N5 be a special fuzzy


mixed rectangular matrix where

⎡ 0.3 0.2 0.1 0.5 0.2 0 1 ⎤


N = ⎢ 0.1 0.7 0.8 1 0 1 0.6 ⎥⎥ ∪

⎣⎢ 0 1 0.9 0.2 0.7 0.9 0.8⎥⎦

⎡ 0.9 0.3 0.5 0.9 0.7 ⎤


⎢ 0.6 0.7 1 0.1 0 ⎥⎥

⎢ 1 0.8 0 1 0.2 ⎥
⎢ ⎥ ∪
⎢ 0.7 1 0.5 0.4 0.5 ⎥
⎢ 0.4 0 1 0.8 0.1⎥
⎢ ⎥
⎢⎣ 0 0.7 0.2 1 0.9 ⎥⎦

⎡ 0.2 0.4 ⎤
⎢ 1 0 ⎥⎥
⎢ ∪
⎢ 0.7 0.3⎥
⎢ ⎥
⎣ 0.2 1 ⎦

81
⎡ 0.3 0.2 0.8 1 ⎤
⎢ 0.2 1 0 0 ⎥⎥

⎢ 0.7 0.9 1 1 ⎥ ∪
⎢ ⎥
⎢ 1 0.8 0.4 0 ⎥
⎢⎣ 0.5 1 0.2 0.7 ⎥⎦

⎡ 0.2 0.3 0.5 ⎤


⎢1 0.1 0.8 ⎥⎥

⎢ 0.4 1 0 ⎥
⎢ ⎥.
⎢ 0.5 0.7 1 ⎥
⎢ 0.8 0 0.5 ⎥
⎢ ⎥
⎣⎢ 1 0.6 0.7 ⎦⎥

Suppose X = X1 ∪ X2 ∪ … ∪ X5 be a special fuzzy mixed row


vector where

X = [0 1 0] ∪ [0 0 0 0 0 1] ∪ [0 0 1 0] ∪ [0 0 1 0 0] ∪
[1 0 0 0 0 0].

Now using the special min max operator we calculate

min max (X, N)

= min max {(X1 ∪ X2 ∪ … ∪ X5), (N1 ∪ N2 ∪ … ∪


N5)}
= min max {X1, N1} ∪ min max {X2, N2} ∪ … ∪
min max {X5, N5}

⎧ ⎡0.3 0.2 0.1 0.5 0.2 0 1 ⎤⎫


⎪ ⎢ ⎪
= min max ⎨[ 0 1 0] , ⎢ 0.1 0.7 0.8 1 0 1 0.6 ⎥⎥ ⎬
⎪ ⎢⎣ 0 1 0.9 0.2 0.7 0.9 0.8⎥⎦ ⎪⎭

82
⎧ ⎡ 0.9 0.3 0.5 0.9 0.7 ⎤ ⎫
⎪ ⎢ 0.6 0.7 1 0.1 0 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪⎪ ⎢ 1 0.8 0 1 0.2 ⎥ ⎪⎪
∪ min max ⎨[ 0 0 0 0 0 1] , ⎢ ⎥⎬
⎪ ⎢0.7 1 0.5 0.4 0.5 ⎥ ⎪
⎪ ⎢ 0.4 0 1 0.8 0.1⎥ ⎪
⎪ ⎢ ⎥⎪
⎩⎪ ⎢⎣ 0 0.7 0.2 1 0.9 ⎦⎥ ⎪⎭

⎧ ⎡ 0.2 0.4 ⎤ ⎫
⎪ ⎢ 1 ⎪
⎪ 0 ⎥⎥ ⎪
∪ min max ⎨[ 0 0 1 0] , ⎢ ⎬
⎪ ⎢0.7 0.3⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣ 0.2 1 ⎦ ⎪⎭

⎧ ⎡ 0.3 0.2 0.8 1 ⎤ ⎫


⎪ ⎢ 0.2 1 ⎪
⎪ ⎢ 0 0 ⎥⎥ ⎪
⎪ ⎪
∪ min max ⎨[ 0 0 1 0 0] , ⎢ 0.7 0.9 1 1 ⎥⎬
⎪ ⎢ ⎥⎪
⎪ ⎢ 1 0.8 0.4 0 ⎥ ⎪
⎪⎩ ⎢⎣ 0.5 1 0.2 0.7 ⎥⎦ ⎪

⎧ ⎡0.2 0.3 0.5 ⎤ ⎫


⎪ ⎢1 ⎪
⎪ ⎢ 0.1 0.8 ⎥⎥ ⎪
⎪⎪ ⎢0.4 1 0 ⎥ ⎪⎪
∪ min max ⎨[1 0 0 0 0 0] , ⎢ ⎥⎬
⎪ ⎢ 0.5 0.7 1 ⎥⎪
⎪ ⎢ 0.8 0 0.5 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪⎩ ⎣⎢ 1 0.6 0.7 ⎦⎥ ⎪⎭

= [0 0.2 0.1 0.2 0.2 0 0.8] ∪ [0.4 0 0 0.1 0] ∪ [0.2 0]


∪ [0.2 0.2 0 0] ∪ [0.2 0 0]
= Z1 ∪ Z2 ∪ … ∪ Z5
= Z;

83
is the special fuzzy mixed row vector. Clearly min max {Z, N}
is not defined so we have to define only min max {Z, NT} and
find the special min max of {Z, NT}.

min max {(Z, NT)}


= min max {(Z1 ∪ Z2 ∪ … ∪ Z5), (N1T ∪ N2T ∪ … ∪
N5T)}
= min max {Z1, N1T} ∪ min max {Z2, N2T} ∪ … ∪
min max {Z5, N5T}

=
⎧ ⎡ 0.3 0.1 0 ⎤⎫
⎪ ⎢0.2 0.7 ⎪
⎪ ⎢ 1 ⎥⎥ ⎪
⎪ ⎢ 0.1 0.8 0.9 ⎥ ⎪
⎪ ⎢ ⎥⎪
min max ⎨[ 0 0.2 0.1 0.2 0.2 0 0.8] , ⎢ 0.5 1 0.2 ⎥ ⎬ ∪
⎪ ⎢0.2 0 0.7 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪ ⎢ 0 1 0.9 ⎥ ⎪
⎪ ⎢ 1 0.6 ⎪
⎩ ⎣ 0.8 ⎥⎦ ⎭

⎧ ⎡ 0.9 0.6 1 0.7 0.4 0 ⎤ ⎫


⎪ ⎢ 0.3 0.7 0.8 1 ⎪
⎪ ⎢ 0 0.7 ⎥⎥ ⎪
⎪ ⎪
min max ⎨[ 0.4 0 0 0.1 0] , ⎢ 0.5 1 0 0.5 1 0.2 ⎥ ⎬
⎪ ⎢ ⎥⎪
⎪ ⎢ 0.9 0.1 1 0.4 0.8 1 ⎥ ⎪
⎪⎩ ⎢⎣0.7 0 0.2 0.5 0.1 0.9 ⎥⎦ ⎪

⎧⎪ ⎡ 0.2 1 0.7 0.2 ⎤ ⎫⎪


∪ min max ⎨[ 0.2 0] , ⎢ ⎥⎬
⎩⎪ ⎣ 0.4 0 0.3 1 ⎦ ⎪⎭

⎧ ⎡ 0.3 0.2 0.7 1 0.5 ⎤ ⎫


⎪ ⎢0.2 1 0.9 0.8 1 ⎥ ⎪
⎪ ⎥ ⎪⎬
∪ min max ⎨[ 0.2 0.2 0 0] , ⎢
⎪ ⎢ 0.8 0 1 0.4 0.2 ⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣1 0 1 0 0.7 ⎦ ⎪⎭

84
⎧ ⎡0.2 1 0.4 0.5 0.8 1 ⎤ ⎫
⎪ ⎢ 0.3 0.1 1 0.7 0 0.6 ⎥ ⎪
∪ min max ⎨[ 0.2 0 0] , ⎢ ⎥⎬
⎪ ⎢⎣ 0.5 0.8 0 1 0.5 0.7 ⎥⎦ ⎪⎭

= [0 0.1 0] ∪ [0.3 0 0 0.4 0 0.2] ∪ [0.2 0 0.3 0.2] ∪


[0.2 0 0.7 0 0.2] ∪ [0.2 0.1 0 0.5 0 0.6]
= Z1 ∪ Z2 ∪ … ∪ Z5
= Z.

We see Z is a special fuzzy mixed row vector. Now we can find


min max {Z, N} and so on.

Now we proceed on to work with the special min max


operator on special fuzzy mixed matrix. Let M = M1 ∪ M2 ∪ …
∪ Ms (s ≥ 2) be the special fuzzy mixed matrix, i.e., some Mi
are ni × ni fuzzy square matrices and some of the Mj’ s are mj ×
tj fuzzy rectangular matrices; 1 ≤ i, j ≤ s (mj ≠ tj).
Suppose

X = X1 ∪ X2 ∪ … ∪ Xs

be a special fuzzy mixed row vector some Xi’ s are 1 × ni fuzzy


row vectors and other Xj’s are 1 × mj fuzzy row vectors. 1 ≤ i, j
≤ s. To use the special min max operator and find min max {X,
M}; i.e.,

min max (X, M)

= min max {(X1 ∪ X2 ∪ … ∪ Xs), (M1 ∪ M2 ∪ … ∪


Ms)}
= min max {X1, M1} ∪ min max {X2, M2} ∪ … ∪
min max {Xs, Ms}
= Y1 ∪ Y2 ∪ … ∪ Ys.
= Y

is once again a special fuzzy mixed row vector. Now we see


min max {Y, M} is not compatible under special min max

85
operator so we define special transpose of M since M is a very
peculiar fuzzy mixed matrix containing both square and
rectangular matrices.

Suppose M1 ∪ M2 ∪ … ∪ Ms; be a special mixed matrix (s


≥ 2) and if some of the Mi’s are ni × ni square fuzzy matrices
and some of the Mj’s are mj × tj (mj ≠ tj) rectangular fuzzy
matrices 1 ≤ i, j ≤ s then the special transpose of M is defined
and denoted by MST;

MST = M1T ∪ … ∪ Mi ∪ … ∪ MjT ∪ … ∪ Ms

where if Mi is a square fuzzy matrix we do not take its transpose


in MST only for the rectangular fuzzy matrices we take the
transpose. This transpose is defined in a special way is called
the special transpose.
We just for the sake of better understanding exhibit the
notion of special fuzzy transpose by the following example.

Suppose

M = M1 ∪ M2 ∪ … ∪ M5

⎡ 0.7 0.8 ⎤
⎢ 1 0 ⎥⎥

⎡ 0.3 1 0 ⎤ ⎢ 0 0.9 ⎥
⎢ ⎥ ⎢ ⎥
= ⎢ 0.8 0.9 1 ⎥ ∪ ⎢ 0.2 0.7 ⎥
⎢⎣ 0.7 0.4 0.5⎥⎦ ⎢ 0.3 0.2 ⎥
⎢ ⎥
⎢ 0.4 0.1⎥
⎢ 0.9 1 ⎥⎦

⎡ 0.3 0.5 0.8 0.9 ⎤


⎢1 ⎡ 0.9 1 0.2 0.7 0.4 0.5⎤
0 0.2 1 ⎥⎥
∪ ⎢ ∪ ⎢⎢ 1 0 0.9 1 0.7 0 ⎥⎥
⎢ 0.4 0.7 1 0.4 ⎥
⎢ ⎥ ⎢⎣ 0.7 1 0.2 0 0.6 1 ⎥⎦
⎣ 0.9 0.8 0.5 1 ⎦

86
⎡ 0.8 1 0.7 0.5 0.6 ⎤
⎢ 1 0 0.6 1 0.9 ⎥⎥

⎢ 0.9 1 0.2 0.5 1 ⎥
∪ ⎢ ⎥
⎢ 0.6 0.3 1 0.7 0.4 ⎥
⎢ 0.4 0.2 0.5 1 0.8⎥
⎢ ⎥
⎢⎣ 0.7 1 1 0 0.1⎥⎦

be a special fuzzy mixed matrix. The special transpose of M


denoted by

MST = (M1 ∪ M2 ∪ … ∪ M5)ST


= M1 ∪ MT2 ∪ M3 ∪ MT4 ∪ MT5.

⎡ 0.3 1 0 ⎤
= ⎢ 0.8 0.9 1 ⎥ ∪
⎢ ⎥
⎢⎣ 0.7 0.4 0.5⎥⎦

⎡ 0.7 1 0 0.2 0.3 0.4 0.9 ⎤


⎢ 0.8 0 0.9 0.7 0.2 0.1 1 ⎥ ∪
⎣ ⎦

⎡ 0.3 0.5 0.8 0.9 ⎤


⎢1 0 0.2 1 ⎥⎥
⎢ ∪
⎢ 0.4 0.7 1 0.4 ⎥
⎢ ⎥
⎣ 0.9 0.8 0.5 1 ⎦

⎡ 0.9 1 0.7 ⎤
⎢ 1 0 1 ⎥⎥

⎢ 0.2 0.9 0.2 ⎥
⎢ ⎥ ∪
⎢ 0.7 1 0 ⎥
⎢ 0.4 0.7 0.6 ⎥
⎢ ⎥
⎢⎣ 0.5 0 1 ⎥⎦

87
⎡ 0.8 1 0.9 0.6 0.4 0.7 ⎤
⎢ 1 0 1 0.3 0.2 1 ⎥⎥

⎢ 0.7 0.6 0.2 1 0.5 1 ⎥
⎢ ⎥
⎢ 0.5 1 0.5 0.7 1 0 ⎥
⎢⎣ 0.6 0.9 1 0.4 0.8 0.1⎥⎦

= M1 ∪ M2T ∪ M3 ∪ M4T ∪ M5T .

Thus one can see that special transpose of a special fuzzy mixed
matrix is different form the usual transpose of any special mixed
square or rectangular matrix.
Now we calculate the special min max operator X on with
M where M = M1 ∪ M2 ∪ … ∪ Ms is a special fuzzy mixed
matrix and X1 ∪ X2 ∪ … ∪ Xs is the special fuzzy mixed row
vector with conditions described just above in the example.
If min max {X, M} = Y1 ∪ Y2 ∪ … ∪ YS = Y be the
special fuzzy mixed row vector then we find the special value of
Y on MST using the special min max operator as follows.

min max {Y, MST}

= min max {(Y1 ∪ Y2 ∪ … ∪ Ys), (M1 ∪ M2T ∪ …


∪ Mi ∪ … ∪ MjT ∪ … ∪ Ms)}
= min max {Y1, M1} ∪ min max {Y2, M2T} ∪ … ∪
min max {Yi, Mi} ∪ … ∪ min max {Yj, MjT} ∪ …
∪ min max {Ys, Ms}
= Z1 ∪ Z2 ∪ … ∪ Zs
= Z;

where Z is a special fuzzy mixed row vector. Now we find min


max {Z, M} and so on.

Now we shall illustrate this situation by the following example.

Example 1.2.26: Let M = M1 ∪ M2 ∪ … ∪ M6 be a special


fuzzy mixed matrix where

88
⎡ 0.8 1 0.7 ⎤
⎢ 1 0.5 0 ⎥
⎡ 0.4 1 0 0.2 ⎤ ⎢ ⎥
⎢ 1 0.7 1 0.3⎥ ⎢ 0.9 1 0.9 ⎥
M = ⎢ ⎥ ∪ ⎢ 0.1 0.6 1 ⎥ ∪
⎢ 0.8 1 0.4 1 ⎥ ⎢ ⎥
⎢ ⎥ ⎢ 0.5 1 0.2 ⎥
⎣ 1 0.7 0 0.2 ⎦ ⎢ ⎥
⎢ 0.3 0.1 0.7 ⎥
⎢ 0.5 1 0.4 ⎥
⎣ ⎦

⎡ 0.7 0.6 1 0.7 0.9 0.3 0.8⎤


⎢ 0 1 0 0.5 1 0.6 1 ⎥⎥
⎢ ∪
⎢ 1 0.7 0.9 0.2 0.7 1 0.9 ⎥
⎢ ⎥
⎣ 0.8 0.5 1 0.3 1 0.8 0.4 ⎦

⎡ 0.8 1 0.9 1 ⎤
⎢ 1 0.7 0.1 0.4 ⎥
⎢ ⎥
⎢ 0.5 1 0.5 1 ⎥
⎡ 0.6 1 0.8⎤ ⎢ ⎥
⎢ 1 0.7 1 ⎥ ∪ ⎢ 0.9 0 1 0.7 ⎥

⎢ ⎥ ⎢ 0.2 1 0.8 0.3⎥
⎢⎣ 0.2 0.8 0 ⎥⎦ ⎢ ⎥
⎢ 0.7 0 1 0.2 ⎥
⎢ 1 0.7 0.6 0.5 ⎥
⎢ ⎥
⎣⎢ 0.3 1 0.4 1 ⎦⎥

⎡ 0.3 1 0.4 0.9 ⎤


⎢1 0 0.2 1 ⎥⎥

⎢ 0.5 0.7 1 0 ⎥.
⎢ ⎥
⎢ 1 0.9 0 1 ⎥
⎢⎣ 0.8 1 0.7 0.4 ⎥⎦

Suppose the special fuzzy mixed row vector

X = X1 ∪ X2 ∪ … ∪ X6

89
= [1 0 0 0] ∪ [0 1 0 0 0 0 0] ∪ [0 0 1 0] ∪ [0 0 1] ∪
[0 0 0 0 1 0 0 0] ∪ [1 0 0 0 0]

operates on M under special min max product to find

min max {X, M}


= min max {(X1 ∪ X2 ∪ … ∪ X6), (M1 ∪ M2 ∪ … ∪
M6)}
= min max {X1, M1} ∪ min max {X2, M2} ∪ … ∪
min max {X6, M6}

⎧ ⎡ 0.4 1 0 0.2 ⎤ ⎫
⎪ ⎢ 1 0.7 1 0.3⎥ ⎪
⎪ ⎥ ⎪⎬
= min max ⎨[1 0 0 0] , ⎢
⎪ ⎢ 0.8 1 0.4 1 ⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣ 1 0.7 0 0.2 ⎦ ⎪⎭

⎧ ⎡ 0.8 1 0.7 ⎤ ⎫
⎪ ⎢ 1 0.5 0 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪ ⎢0.9 1 0.9 ⎥ ⎪
⎪ ⎢ ⎥⎪
∪ min max ⎨[ 0 1 0 0 0 0 0] , ⎢ 0.1 0.6 1 ⎥ ⎬
⎪ ⎢ 0.5 1 0.2 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪ ⎢ 0.3 0.1 0.7 ⎥ ⎪
⎪ ⎢ 0.5 1 0.4 ⎥ ⎪
⎩ ⎣ ⎦⎭

⎧ ⎡0.7 0.6 1 0.7 0.9 0.3 0.8⎤ ⎫


⎪ ⎢ 0 ⎪
⎪ 1 0 0.5 1 0.6 1 ⎥⎥ ⎪
∪ min max ⎨[ 0 0 1 0] , ⎢ ⎬
⎪ ⎢ 1 0.7 0.9 0.2 0.7 1 0.9 ⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣ 0.8 0.5 1 0.3 1 0.8 0.4 ⎦ ⎪⎭

⎧ ⎡0.6 1 0.8⎤ ⎫
⎪ ⎢ 1 0.7 1 ⎥ ⎪
∪ min max ⎨[ 0 0 1] , ⎢ ⎥⎬
⎪ ⎢⎣0.2 0.8 0 ⎥⎦ ⎪⎭

90
⎧ ⎡ 0.8 1 0.9 1 ⎤ ⎫
⎪ ⎢ 1 0.7 0.1 0.4 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪ ⎢ 0.5 1 0.5 1 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪ ⎢ 0.9 0 1 0.7 ⎥ ⎪
∪ min max ⎨[ 0 0 0 0 1 0 0 0] ,
⎢ 0.2 1 0.8 0.3⎥ ⎬
⎪ ⎢ ⎥⎪
⎪ ⎢ 0.7 0 1 0.2 ⎥⎪
⎪ ⎢ 1 0.7 0.6 0.5 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪ ⎪
⎩ ⎣⎢ 0.3 1 0.4 1 ⎦⎥ ⎭

⎧ ⎡ 0.3 1 0.4 0.9 ⎤ ⎫


⎪ ⎢1 ⎪
⎪ ⎢ 0 0.2 1 ⎥⎥ ⎪
⎪ ⎪
∪ min max ⎨[1 0 0 0 0] , ⎢0.5 0.7 1 0 ⎥⎬
⎪ ⎢ ⎥
⎢ 1 0.9 0 1 ⎥⎪
⎪ ⎪
⎪⎩ ⎢⎣0.8 1 0.7 0.4 ⎥⎦ ⎪

= {[0.8 0.7 0 0.2] ∪ [0.1 0.1 0.2] ∪ [0 0.5 0 0.3 0.9


0.3 0.4] ∪ [0.6 0.7 0.8] ∪ [0.3 0 0.1 0.2] ∪ [0.5 0 0
0]}
= Y1 ∪ Y2 ∪ … ∪ Y6.
= Y

is again a special fuzzy mixed row vector.


Now we use special min max function to find the value of

Y on MST

= min max (Y, MST).


= min max {(Y1 ∪ Y2 ∪ … ∪ Y6), (M1 ∪ M2T ∪ M3T
∪ M4 ∪ M5T ∪ M6T).
= min max (Y1, M1) ∪ min max (Y2, M2T) ∪ min max
(Y3, M3T) ∪ min max (Y4, M4) ∪ min max (Y5,
M5T) ∪ min max (Y6, M6T).

91
⎧ ⎡ 0.4 1 0 0.2 ⎤ ⎫
⎪ ⎢ 1 0.7 1 0.3⎥ ⎪
⎪ ⎥ ⎪⎬
= min max ⎨[ 0.8 0.7 0 0.2] , ⎢
⎪ ⎢ 0.8 1 0.4 1 ⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣ 1 0.7 0 0.2 ⎦ ⎪⎭

⎧ ⎡ 0.8 1 0.9 0.1 0.5 0.3 0.5⎤ ⎫
⎪ ⎪
min max ⎨[ 0.1 0.1 0.2] , ⎢⎢ 1 0.5 1 0.6 1 0.1 1 ⎥⎥ ⎬
⎪ ⎢⎣0.7 0 0.9 1 0.2 0.7 0.4 ⎥⎦ ⎪⎭


⎧ ⎡ 0.7 0 1 0.8⎤ ⎫
⎪ ⎢ 0.6 1 0.7 0.5⎥ ⎪
⎪ ⎢ ⎥⎪
⎪ ⎢ 1 0 0.9 1 ⎥ ⎪
⎪ ⎢ ⎥⎪
min max ⎨[ 0 0.5 0 0.3 0.9 0.3 0.4] , ⎢ 0.7 0.5 0.2 0.3⎥ ⎬
⎪ ⎢ 0.9 1 0.7 1 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪ ⎢ 0.3 0.6 1 0.8⎥ ⎪
⎪ ⎢ 0.8 1 0.9 0.4 ⎥ ⎪
⎩ ⎣ ⎦⎭

⎧ ⎡ 0.6 1 0.8⎤ ⎫
⎪ ⎢ 1 0.7 1 ⎥ ⎪
∪ min max ⎨[ 0.6 0.7 0.8] , ⎢ ⎥⎬
⎪ ⎪
⎩ ⎣⎢ 0.2 0.8 0 ⎥⎦ ⎭
∪ min max
⎧ ⎡ 0.8 1 0.5 0.9 0.2 0.7 0.3⎤ ⎫
1
⎪ ⎢ 1 0.7 ⎪
⎪ 1 0 1 0 0.7 1 ⎥⎥ ⎪
⎨[ 0.3 0 0.1 0.2 ] , ⎢
⎢0.9 0.1

⎪ 0.5 1 0.8 1 0.6 0.4 ⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣ 1 0.4 1 0.7 0.3 0.2 0.5 1 ⎦ ⎪⎭

⎧ ⎡ 0.3 1 0.5 1 0.8 ⎤ ⎫


⎪ ⎢1 ⎪
⎪ 0 0.7 0.9 1 ⎥⎥ ⎪
∪ min max ⎨[ 0.5 0 0 0] , ⎢ ⎬
⎪ ⎢ 0.4 0.2 1 0 0.7 ⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣ 0.9 1 0 1 0.4 ⎦ ⎪⎭

92
= [0.8 0.7 0.2 0.2] ∪ [0.7 0.2 0.9 0.1 0.2 0.1 0.4] ∪
[0.3 0 0.3 0.3] ∪ [0.6 0.7 0.8] ∪ [0.8 0.1 0.5 0 0.3 0
0.5 0.3] ∪ [0.4 0 0 0 0.4]
= P1 ∪ P2 ∪ P3 ∪ P4 ∪ P5 ∪ P6
= P

is a special fuzzy mixed row vector. Now we can find min max
{P, M} and so on.

On similar lines we can work with special max min operator


also. Now having defined some of the major operations we now
proceed on to define mixed operations on these special fuzzy
matrices.

DEFINITION 1.2.10: Let S = S1 ∪ S2 ∪ … ∪ Sn (n ≥ 2) be a


special fuzzy mixed matrix X = X1 ∪ X2 ∪ … ∪ Xn (n ≥ 2) be a
special fuzzy mixed row vector. We define a new operation of X
on S called the special mixed operation denoted by om.

X om S = (X1 ∪ X2 ∪ … ∪ Xn) om (S1 ∪ S2 ∪ … ∪ Sn)


= X1 om S1 ∪ X2 om S2 ∪ … ∪ Xn om Sn.

is defined as follows.

If Si is a square fuzzy matrix with entries from the set {–1, 0, 1}


then Xi om Si is the operation described in pages 20-1 of this
book. If Sj is a rectangular fuzzy matrix with entries from the set
{–1, 0, 1} then Xj om Sj is the operation described in page 23 of
this book. Suppose Sk is a square fuzzy matrix with entries from
[0, 1] then Xk om Sk is the min max or max min operation
defined in pages 14-5 of this book.
If St is a rectangular fuzzy matrix with entries from [0, 1]
then Xt om St is the min max operation described in page 15 max
min operation defined in page 14 of this book for 1 ≤ i, j ≤ k, t ≤
n. Thus ‘om’ defined between X and S will be known as the
special mixed operation and denoted by om.

We will illustrate this by the following example.

93
Example 1.2.27: Let S = S1 ∪ S2 ∪ … ∪ S5 be a special fuzzy
mixed matrix where
⎡ 0 1 −1 0 0⎤
⎢1 0 0 0 1 ⎥⎥

S = ⎢ −1 0 0 1 0⎥ ∪
⎢ ⎥
⎢0 0 1 0 0⎥
⎢⎣ 0 −1 0 1 0 ⎥⎦

⎡ 0.3 0.8 1 0.5 0 0.7 ⎤


⎢ 0.5 0.7 0 1 0.6 1 ⎥⎥ ∪

⎢⎣ 1 0.5 0.2 0.7 0.2 0 ⎥⎦

⎡1 −1 0 1⎤
⎢0 0 ⎡ 0.9 0.2 1 0 0.5 ⎤
⎢ 1 0 ⎥⎥ ⎢ 0.3 0 0.5 1 0.7 ⎥
⎢0 1 0 1⎥ ⎢ ⎥
⎢ ⎥ ∪ ⎢ 1 0.2 1 0 0.3⎥ ∪
⎢1 0 1 −1⎥ ⎢ ⎥
⎢ 0 1 0.3 0.2 1 ⎥
⎢0 1 0 0 ⎥
⎢ ⎥ ⎢⎣ 0.9 0.7 0.6 0.7 0 ⎥⎦
⎣⎢ 0 1 1 0 ⎦⎥

⎡1 0 1 1 1 0 0 1 −1⎤
⎢0 1 0 0 0 1 0 0 0 ⎥⎥

⎢ 1 −1 0 0 0 0 0 −1 1 ⎥
⎢ ⎥.
⎢0 0 1 0 1 0 1 0 0⎥
⎢ −1 0 0 −1 1 0 0 0 0⎥
⎢ ⎥
⎣⎢ 0 0 −1 0 0 1 1 0 0 ⎦⎥

Let

X = X1 ∪ X2 ∪ X3 ∪ X4 ∪ X5
= [1 0 0 0 0] ∪ [0 0 1] ∪ [0 0 0 0 0 1] ∪ [0 1 0 0 1] ∪
[0 0 0 1 0 0]

94
be special fuzzy mixed row vector. We find X om S using the
special mixed operation.

X om S = (X1 ∪ X2 ∪ X3 ∪ X4 ∪ X5) om (S1 ∪ S2 ∪ … ∪ S5)


= X1 o1m S1 ∪ X 2 o 2m S2 ∪ X 3 o3m S3 ∪ X 4 o 4m S4
∪ X 5 o5m S5

where o1m is the thresholding and updating resultant row vectors


after usual matrix multiplication. o 2m is the min max operator
between X2 and S2. o3m and o 4m are usual matrix multiplication
which are updated and thresholded and are found sequentially
by finding Xi o Si and next Yi o SiT and so on. o5m is the max min
operation.

Now we explicitly show how the special mixed operation om


function works:

X o m S = X1 o1m S1 ∪ … ∪ X 5 o5m S5

⎡ 0 1 −1 0 0⎤
⎢1 0 0 0 1 ⎥⎥

= [1 0 0 0 0] o1m ⎢ −1 0 0 1 0 ⎥ ( o1m = 'o')
⎢ ⎥
⎢0 0 1 0 0⎥
⎢⎣ 0 −1 0 1 0 ⎥⎦

⎧ ⎡ 0.3 0.8 1 0.5 0 0.7 ⎤ ⎫


⎪ ⎢ 0.5 0.7 0 ⎪
min max ⎨[ 0 0 1] o 2m ⎢ 1 0.6 1 ⎥⎥ ⎬
⎪ ⎢⎣ 1 0.5 0.2 0.7 0.2 0.1⎥⎦ ⎪⎭

2
( o m = ',' )

95
⎧ ⎡1 −1 0 1 ⎤⎫
⎪ ⎢0 0 ⎪
⎪ ⎢ 1 0 ⎥⎥ ⎪
⎪⎪ ⎢0 1 0 1 ⎥ ⎪⎪ 3
⎨[ 0 0 0 0 0 1] o m ⎢ ⎥ ⎬ ( o m = 'o'
3

⎪ ⎢1 0 1 −1⎥ ⎪
⎪ ⎢0 1 0 0 ⎥⎪
⎪ ⎢ ⎥⎪
⎩⎪ ⎣⎢0 1 1 0 ⎦⎥ ⎭⎪
operations)

∪ max min
⎧ ⎡ 0.9 0.2 1 0 0.5 ⎤ ⎫
⎪ ⎢ 0.3 0 0.5 1 0.7 ⎥ ⎪
⎪ ⎢ ⎥ ⎪⎪

⎨[ 0 1 0 0 1] o m ⎢ 1 0.2 1 0 0.3⎥ ⎬ ( o 4m = ',')
4

⎪ ⎢ ⎥
⎢ 0 1 0.3 0.2 1 ⎥ ⎪
⎪ ⎪
⎪⎩ ⎢⎣ 0.9 0.7 0.6 0.7 0 ⎥⎦ ⎪

⎧ ⎡1 0 1 1 1 0 0 1 −1⎤ ⎫
⎪ ⎢0 1 0 0 ⎪
⎪ ⎢ 0 1 0 0 0 ⎥⎥ ⎪
⎪⎪ ⎢ 1 −1 0 0 0 0 0 −1 1 ⎥ ⎪⎪
⎨[ 0 0 0 1 0 0] o m ⎢ ⎥⎬
5

⎪ ⎢0 0 1 0 1 0 1 0 0 ⎥⎪
⎪ ⎢ −1 0 0 −1 1 0 0 0 0 ⎥⎪
⎪ ⎢ ⎥⎪
⎪⎩ ⎢⎣ 0 0 −1 0 0 1 1 0 0 ⎥⎦ ⎪⎭
( o5m = 'o')

Y = [0 1 –1 0 0] ∪ [0.3 0.7 0 0.5 0 0.7] ∪ [0 1 1 0] ∪


[0.9 0.7 0.6 1 0.7] ∪ [0 0 1 0 1 0 1 0 0]
= Y'1 ∪ Y'2 ∪ Y'3 ∪ Y'4 ∪ Y'5
= Y'.

We update and threshold the resultant wherever applicable since


Y' = Y'1 ∪ Y'2 ∪ Y'3 ∪ Y'4 ∪ Y'5 is not a special fuzzy mixed
row vector and find

Y = Y1 ∪ Y2 ∪ Y3 ∪ Y4 ∪ Y5.

96
= [1 1 0 0 0] ∪ [0.3 0.7 0 0.5 0 0.7] ∪ [0 1 1 0] ∪ [0.9
0.7 0.6 1 0.7] ∪ [0 0 1 0 1 0 1 0 0].

Now we find

Y o m SST = Y1 o1m S1 ∪ Y2 o 2m ST2 ∪ Y3 o3m S3T ∪ Y4 o 4m S4 ∪ Y5 o5m S5T

⎡ 0 1 −1 0 0⎤
⎢1 0 0 0 1 ⎥⎥

= [1 1 0 0 0] ⎢ −1 0 0 1 0⎥
⎢ ⎥
⎢0 0 1 0 0⎥
⎢⎣ 0 −1 0 1 0 ⎥⎦

⎧ ⎡ 0.3 0.5 1 ⎤ ⎫
⎪ ⎢ 0.8 0.7 0.5 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪⎪ ⎢ 1 0 0.2 ⎥ ⎪⎪
∪ min max ⎨[ 0.3 0.7 0 0.5 0 0.7 ] , ⎢ ⎥⎬
⎪ ⎢ 0.5 1 0.7 ⎥ ⎪
⎪ ⎢ 0 0.6 0.2 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎩⎪ ⎢⎣0.7 1 0.1⎥⎦ ⎭⎪

⎡1 0 0 1 0 0⎤
⎢ −1 0 1 0 1 1 ⎥⎥
∪ [ 0 1 1 0] o ⎢
⎢0 1 0 1 0 1⎥
⎢ ⎥
⎣1 0 1 −1 0 0⎦

∪ max min
⎧ ⎡ 0.9 0.2 1 0 0.5 ⎤ ⎫
⎪ ⎢ 0.3 0 0.5 1 0.7 ⎥ ⎪
⎪ ⎢ ⎥ ⎪⎪

⎨[ 0.9 0.7 0.6 1 0.7 ] , ⎢ 1 0.2 1 0 0.3⎥ ⎬
⎪ ⎢ ⎥
⎢ 0 1 0.3 0.2 1 ⎥ ⎪
⎪ ⎪
⎪⎩ ⎢⎣ 0.9 0.7 0.6 0.7 0 ⎥⎦ ⎪

97
⎡1 0 1 0 −1 0 ⎤
⎢0 1 −1 0 0 0 ⎥⎥

⎢1 0 0 1 0 −1⎥
⎢ ⎥
⎢1 0 0 0 −1 0 ⎥
[0 0 1 0 1 0 1 0 0] o ⎢⎢ 1 0 0 1 1 0⎥

⎢0 1 0 0 0 1⎥
⎢0 0 0 1 0 1⎥
⎢ ⎥
⎢1 0 −1 0 0 0⎥
⎢ −1 0 0 0 ⎥⎦
⎣ 0 1

= [1 1 –1 0 1] ∪ [0 0 0.2] ∪ [–1 1 1 1 1 2] ∪ [0.9 1


0.9 0.7 1] ∪ [2 0 0 3 1 0]
= Z'1 ∪ Z'2 ∪ Z'3 ∪ Z'4 ∪ Z'5
= Z';

after updating and thresholding Z' we get Z

= Z1 ∪ Z2 ∪ Z3 ∪ Z4 ∪ Z5
= [1 1 0 0 1] ∪ [0 0 0.2] ∪ [0 1 1 1 1 1] ∪ [0.9 1 0.9
0.7 1] ∪ [1 0 0 1 1 0].

Using Z and S we find Z om S using the special mixed operator


and so on. This sort of special mixed operator will be used when
we use special mixed fuzzy models which will be described in
Chapter two.

1.3 Special Neutrosophic matrices and fuzzy neutrosophic


matrices and some essential operators using them

In this section we just introduce the special neutrosophic


matrices and special fuzzy neutrosophic matrices and give some
of the essential operators using them. We just give some
information about neutrosophy. For more about these concepts
please refer [187-200].
We denote the indeterminate by I and I2 = I; further

98
I + I + … + I = nI , n ≥ 2.
n − times

We call [0 1] ∪ [0 I] to be the fuzzy neutrosophic interval. If we


take the set generated by 〈Z ∪ I〉 we call this as the neutrosophic
ring of integers. Likewise 〈Q ∪ I〉 denotes the neutrosophic ring
of rationals. 〈R ∪ I〉 the set generated by R and I is called as the
neutrosophic ring of reals.

For more about these concepts please refer [187-190].

Thus a pure neutrosophic number is nI; nI ∈ R and a mixed


neutrosophic integer is m + nI; n, m ∈ R. Now a matrix is called
a neutrosophic matrix if it takes its entries from 〈Z ∪ I〉 or 〈Q ∪
I〉 or 〈R ∪ I〉.

We now illustrate different types of neutrosophic matrices.

Example 1.3.1: Consider the neutrosophic matrix

⎡ I 0 2I + 1 0.8 ⎤

M=⎢ 7 8I 3I − 5 −2I ⎥⎥ .
⎢⎣9 − 3I 6I + 3 3 + 0.9I 12I ⎥⎦

We call M to be 3 × 4 rectangular neutrosophic matrix with


entries from the neutrosophic ring of reals.

Example 1.3.2: Let N be a neutrosophic matrix where

⎡ 3I 2 − 5I 8 ⎤

N=⎢ 0 1 9 + I ⎥⎥ .
⎢⎣ 2 + 7I 8I + 1 0 ⎥⎦

N is a neutrosophic square matrix.

99
Example 1.3.3: Consider
⎡ 0.9I ⎤
⎢8 + 2I ⎥
⎢ ⎥
⎢5I − 7 ⎥
⎢ ⎥
⎢ I ⎥
X= .
⎢ 0 ⎥
⎢ ⎥
⎢ 4I ⎥
⎢ −8I ⎥
⎢ ⎥
⎣⎢ 1 ⎦⎥

X is a neutrosophic matrix which is known as the neutrosophic


column vector/matrix.

Example 1.3.4: Consider the neutrosophic matrix

Y = [ 0.9 I + 8 7I 8I − 1 0.2 I + 0.3 0.1]


is called the neutrosophic row vector/matrix.

Example 1.3.5: Let

T = [ 0.2I 0.3 0.4I 0.2I − 0.7 1 0 0.9I]


be a neutrosophic matrix, T will be called the fuzzy
neutrosophic row vector/matrix.

Example 1.3.6: Let


⎡ 0.9 ⎤
⎢ 0.2 + I ⎥
⎢ ⎥
⎢ 0.3I − 1⎥
⎢ ⎥
1 ⎥
V=⎢
⎢ 0.8I ⎥
⎢ ⎥
⎢ I − 0.7 ⎥
⎢ I ⎥
⎢ ⎥
⎢⎣ 0 ⎥⎦

100
neutrosophic matrix. V is known as the fuzzy neutrosophic
column vector/matrix.

Example 1.3.7: Let us consider the neutrosophic matrix

⎡4 0 0 0 0⎤
⎢ 0 2I 0 0 0 ⎥⎥

S = ⎢ 0 0 1 + 9I 0 0⎥
⎢ ⎥
⎢0 0 0 I +1 0 ⎥
⎢⎣ 0 0 0 0 0.1⎥⎦

S is called the neutrosophic diagonal matrix.


Now we will recall the definition of neutrosophic matrix
addition and multiplication.

The neutrosophic zero matrix is the usual zero matrix i.e., a


matrix in which all elements in it are zero and denoted by

⎡0 0 … 0⎤
⎢0 0 … 0⎥
(0) = ⎢ ⎥.
⎢ ⎥
⎢ ⎥
⎣0 0 … 0⎦

Example 1.3.8: Let S and T be two 3 × 5 rectangular matrices


where
⎡ 0 4 7I − 1 2 −I ⎤

S = ⎢ 2I 0 4I + 1 0 9I ⎥⎥
⎣⎢ 7 + 3I 0.9 − 2I 3 9I 2 − I ⎦⎥
and

⎡ 5+I 7+I 2−I 9+I 0 ⎤



T = ⎢ 7I 2 − I 3 + 8I 1 7 ⎥⎥
⎣⎢3 − 0.8I 0.9 9 0 4 + I ⎦⎥

101
The neutrosophic matrix addition of S and T

⎡ 5+I 11 + I 6I + 1 11 + I −I ⎤

S + T = ⎢ 9I 2−I 4 + 12I 1 7 + 9I ⎥⎥ .
⎢⎣10 + 2.2I 1.8 − 2I 12 9I 6 ⎥⎦

We see S + T is also a neutrosophic matrix. We will state in


general sum of two neutrosophic matrices need not yield back a
neutrosophic matrix. This is shown by the following example.

Example 1.3.9: Let


⎡ 3I + 1 2I ⎤
P=⎢ ⎥
⎣17 − I 4 ⎦
and
⎡7 − 3I 8 − 2I ⎤
Q=⎢
⎣ 3+ I 2 ⎥⎦

be any two 2 × 2 square matrices. Now

⎡ 8 8⎤
P+Q=⎢ ⎥.
⎣ 20 6 ⎦

Clearly P + Q is not a neutrosophic matrix. Likewise in general


the product of two neutrosophic matrices under matrix product
need not yield a neutrosophic matrix.

Example 1.3.10: Let


⎡7 + I I ⎤ ⎡7 − I 0⎤
A=⎢ ⎥ and B = ⎢
⎣ I −6I ⎦ ⎣ I 0 ⎥⎦

be any two neutrosophic matrices AB the matrix product of


neutrosophic matrices

⎡7 + I I ⎤ ⎡7 − I 0⎤
AB = ⎢
⎣ I −6I ⎥⎦ ⎢⎣ I 0 ⎥⎦

102
⎡ 49 0 ⎤
= ⎢ ⎥
⎣ 0 0⎦

which is not a neutrosophic matrix.

However we illustrate by the following example that even if the


neutrosophic product AB is defined the product BA need not be
defined. If A is a m × t matrix and B is a t × s matrix then AB is
defined but BA is not defined (m ≠ s).

Example 1.3.11: Let

⎡ 0 I 2 − I⎤

A = ⎢4 − I 0 7 ⎥⎥
⎢⎣ 8I −1 0 ⎥⎦
and
⎡ 7I − 1 2 + I 3 − I 5 − I 0 ⎤
B = ⎢⎢ 0 7I 2 0 3⎥⎥
⎢⎣ 8 + I 3I −I 1 0 ⎥⎦
Now
⎡ 16 − 7I 10I I 2−I 3I ⎤
AB = ⎢ 29I + 52 8 + 22I 12 − 13I 27 − 8I 0 ⎥⎥

⎢⎣ 48I 17I 16I − 2 32I −3⎥⎦

which is a neutrosophic matrix. But BA is not defined as B is a


3 × 5 matrix and A is a 3 × 3 matrix, so BA is not defined.

Now we proceed on to define a new class of special


neutrosophic matrices.

DEFINITION 1.3.1: Let M = M1 ∪ M2 ∪ … ∪ Mn be a collection


of neutrosophic matrices where each Mi is a t × t neutrosophic
matrix; i = 1, 2, …, n. We call M to be a special neutrosophic
square matrix.

103
We illustrate this by the following example.

Example 1.3.12: Let M = M1 ∪ M2 ∪ … ∪ M5

⎡ 3I I+7 8−I 0 ⎤
⎢ 1 7I 5I − 1 8I ⎥⎥
= ⎢ ∪
⎢ 2I + 4 11I − 1 1 0 ⎥
⎢ ⎥
⎣ 0 1 9 − 5I 2 − I ⎦

⎡ 0 7−I 5+I 2I ⎤
⎢ 7I 0 8 9I − 1⎥⎥
⎢ ∪
⎢ 2I − 1 8I 1 21I ⎥
⎢ ⎥
⎣9 + 3I 3 2−I 5I ⎦

⎡ 3I − 1 0 1 3 + 2I ⎤
⎢ 2 + 4I 3I 4I − 1 0 ⎥⎥
⎢ ∪
⎢ 3− I 4 3I 1 ⎥
⎢ ⎥
⎣ 5I 9I 6 − 3I 0 ⎦

⎡ 0 3I 7I − 1 2 − 3I ⎤
⎢ 4I 0 6 + 5I 9I ⎥⎥
⎢ ∪
⎢ 2 − I 4 − 2I 1 17I ⎥
⎢ ⎥
⎣ 6I + 1 12I 5−I 3−I ⎦

⎡ 1 0 I −2 + I ⎤
⎢ I 7−I 8I 5I − 1 ⎥⎥
⎢ .
⎢ 6I + 1 2I 7 − 5I 8I ⎥
⎢ ⎥
⎣ 9I − 1 6I − 1 0 1 ⎦

be a special neutrosophic square matrix.

Note: Even if Mi = Mj (i ≠ j) still M is a special neutrosophic


square matrix.

104
Example 1.3.13: Let G = G1 ∪ G2 ∪ G3

⎡ 2I 6 − 4I ⎤ ⎡ 0 I ⎤ ⎡ 0 I ⎤
= ⎢ ⎥ ∪ ⎢ ⎥ ∪ ⎢ ⎥
⎣ 5I 7 + 3I ⎦ ⎣5I − 1 2I + 1⎦ ⎣5I − 1 2I + 1⎦

G is a special neutrosophic square matrix.

DEFINITION 1.3.2: Let T = T1 ∪ T2 ∪ … ∪ Ts if the Ti’s are ni ×


ni square neutrosophic matrices then we call T to be a special
neutrosophic mixed square matrix (if s ≥ 2) ni ≠ nj (i ≠ j at least
for some i and j) 1 ≤ i, j ≤ s.

Example 1.3.14: Let T = T1 ∪ T2 ∪ T3 ∪ T4

⎡ 0.3I 5I − 1 3I + 2 4I ⎤
⎢ 0 3I + 1 3 − 8I 9 ⎥⎥
= ⎢ ∪
⎢ I 2I − 1 4I − 1 21 ⎥
⎢ ⎥
⎣ 2I + 7 9 − 3I 9I − 1 9 − 7I ⎦

⎡ 1 0.3 + I 7 − 2I ⎤
⎢ 0 ⎡ 0.I 2 ⎤
⎢ 4I − 1 12I − 1⎥⎥ ∪ ⎢ 7I 4 − I ⎥ ∪
⎢⎣ 6I + 1 0 34 ⎥⎦ ⎣ ⎦

⎡ 0.1 I − 7 8 + I 5I − 1 8I ⎤
⎢ 9 12 + I 0 17I − 1 6I ⎥⎥

⎢8 + 2I 0 I 2 − 5I 7 ⎥ .
⎢ ⎥
⎢ I 9I − 1 9 + 3I 8I + 1 1 − 5I ⎥
⎢⎣ 5I − 1 12 − I 1 3 9 + I ⎥⎦

T is a special neutrosophic mixed square matrix.

DEFINITION 1.3.3: Let W = W1 ∪ W2 ∪ … ∪ Wn (n ≥ 2) be the


collection of p × q neutrosophic (p ≠ q) rectangular matrices.
We define W to be a special neutrosophic rectangular matrix.

105
We illustrate this by the following example.

Example 1.3.15: Let W = W1 ∪ W2 ∪ W2 ∪ W4 =

⎡ 0.3I 7I − 1 8I + 1 ⎤
⎢ 8 5I + 1 9I ⎥⎥

⎢ 21I − 4 9 + 6I 4 + 6I ⎥ ∪
⎢ ⎥
⎢ 9 − 5I 7 0.5 ⎥
⎢⎣ 2 + 6 5I 21 ⎥⎦

⎡ 9I 2I − 1 5 ⎤
⎢ 0 7 + 2I 6 + I ⎥⎥

⎢ 21 6I − 1 12I − 1⎥ ∪
⎢ ⎥
⎢ 7I + 3 15 + I 8I − 1 ⎥
⎢⎣9I − 5I 2 3I ⎥⎦

⎡ 6 8I − 1 9I ⎤
⎢5I − 1 0 8 ⎥⎥

⎢ 7I 9I + 8 2I + 5⎥ ∪
⎢ ⎥
⎢ 3− I 21 0 ⎥
⎢⎣ 1 15I 9 − 6I ⎥⎦

⎡ 21 − I 2 + 5I 8 ⎤
⎢ 9 0 20I ⎥⎥

⎢ 4 + 5I 7I 5 + 3I ⎥ ,
⎢ ⎥
⎢12 + I 5 0 ⎥
⎣⎢ 9I 7 − 5I 12 + I ⎦⎥

W is a special neutrosophic rectangular matrix.

Next we define the notion of special neutrosophic mixed


rectangular matrix.

106
DEFINITION 1.3.4: Let V = V1 ∪ V2 ∪ … ∪ Vn (n ≥ 2) be a
collection of neutrosophic rectangular matrices. Each Vi is a ti
× si (ti ≠ si) neutrosophic rectangular matrix; i = 1, 2, …, n. We
define V to be a special neutrosophic mixed rectangular matrix.

We illustrate this by the following example.

Example 1.3.16: Let V = V1 ∪ V2 ∪ V3 ∪ V4 ∪ V5 =

⎡ 0.3I I + 7 ⎤
⎢ 2I − 8I ⎥⎥

⎢ 5 9I − 4 ⎥
⎢ ⎥ ∪
⎢ 0 21I + 1⎥
⎢ 12I 3 ⎥
⎢ ⎥
⎣⎢ 17 6 − I ⎥⎦

⎡ 0.5I 7 − 5I 15I + 4 8I + 7 21 ⎤
⎢ 7I 0 21 − 7I 6I + 1 8I + 4 ⎥⎥ ∪

⎢⎣ 1 15I + 6 9I 12 0 ⎥⎦

⎡ 0.8 9I 5I + 1⎤
⎢ 3I + 1 7I − 1 16I ⎥
⎢ ⎥
⎢ 7 8I 1 ⎥
⎢ ⎥ ∪
⎢ 6I − 7 2 + I 0 ⎥
⎢ 10I 8 + 5I 9I − 1⎥
⎢ ⎥
⎣⎢ 7 − 8I 5 3 + 5I ⎥⎦

⎡ 2I 9 5 − I 5I − 1 9 − 2I ⎤
⎢ 0 5I + 1 25I 4 − I 2I ⎥⎥
⎢ ∪
⎢ 7 21 − 7I 4 9 + 3I 0 ⎥
⎢ ⎥
⎣15I − 0 1+ I 2II 19 ⎦

107
⎡ 8I −9 25 − I 7 + 8I 0 ⎤
⎢ 0 I +1 4 + 2I 8 5I ⎥⎥

⎢17 − I 1 0 9I 12 ⎥
⎢ ⎥.
⎢ 21 12 + 5I 8I + 1 0 7I − 5⎥
⎢ 4 + 5I 25I 16I − 1 6 − 7I 8 + I ⎥
⎢ ⎥
⎢⎣ 9I − 1 8I − 1 25I + 1 6I 75 ⎥⎦

V is a special neutrosophic mixed rectangular matrix.

DEFINITION 1.3.5: Let X = X1 ∪ X2 ∪ … ∪ Xn (n ≥ 2) be a


collection of neutrosophic row matrices. Here each Xi is a 1 × t
row neutrosophic matrix. We call X to be a special neutrosophic
row vector / matrix. If some Xj is a 1× tj neutrosophic row
matrix and some Xk is a 1 × tk neutrosophic row vector and tj ≠
tk for some k ≠ j. Then we define X to be a special neutrosophic
mixed row vector.

We illustrate them by the following example.

Example 1.3.17: Let

X = X1 ∪ X2 ∪ X3 ∪ X4 ∪ X5 ∪ X6
= [I 0 0.8 0.8 I 0.9 I 0.2] ∪ [0 I I 6 1 I 8I 0.7] ∪ [0.5
0.6 I 0.7 4 0.9I 0.6 9] ∪ [9 1 4 1 I 0 0 I] ∪ [0.2 I 0.2
0.2 0.3I 0.4I 8 0] ∪ [0 I 0 I 0 0 I 8],

we see X is a special neutrosophic row vector for we see each


Xi is a 1 × 8 neutrosophic row vector / matrix, 1 ≤ i ≤ 6.

Next we give an example of a special neutrosophic mixed


neutrosophic row vector.

Example 1.3.18: Let

X = X1 ∪ X2 ∪ X3 ∪ X4
= [0 I 0.7 0.8] ∪ [I 1 I 0 0 0 1 0 0 5] ∪ [0.2 0.9 I 0.6
0.8 1 0 0] ∪ [0 I 0 0 7 1 I 0],

108
X is a special neutrosophic mixed row vector / matrix.

We define the notion of special neutrosophic column vector


and special neutrosophic mixed column vector / matrix.

DEFINITION 1.3.6: Let Y = Y1 ∪ Y2 ∪ … ∪ Ym (m ≥ 2) where Yi


is a t×1 column neutrosophic vector/matrix for each i = 1, 2, …,
m. Then we call Y to be a special neutrosophic column
vector/matrix. Suppose if Y = Y1 ∪ Y2 ∪ … ∪ Ym be such that
some Yj is a tj × 1 neutrosophic column vector/matrix for j = 1,
2, …, m and tj ≠ ti for at least one i ≠ j, 1 ≤ i, j ≤ m. Then we call
Y to be a special neutrosophic mixed column vector / matrix.

Now we illustrate these concepts by the following examples.

Example 1.3.19: Let Y = Y1 ∪ Y2 ∪ … ∪ Y6 =

⎡ 0.2I ⎤ ⎡ I ⎤ ⎡ 6 ⎤
⎢ 1 ⎥ ⎢ 9I ⎥ ⎢ 9I ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢ 0.7 ⎥ ⎢ 0.8 ⎥ ⎢ 0.8I ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢ I ⎥ ∪ ⎢ I ⎥ ∪ ⎢ 0.7 ⎥
⎢ 9.3 ⎥ ⎢0.6I ⎥ ⎢ I ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢ 0 ⎥ ⎢ 1 ⎥ ⎢ 9 ⎥
⎢ 0.6I ⎥ ⎢ I ⎥ ⎢ 0 ⎥
⎣ ⎦ ⎣ ⎦ ⎣ ⎦

⎡I⎤ ⎡ 8 ⎤ ⎡1 ⎤
⎢14 ⎥ ⎢ 0 ⎥ ⎢ 6.5⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢ 4I ⎥ ⎢ I ⎥ ⎢ I ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
∪ ⎢I⎥ ∪ ⎢ 0 ⎥ ∪ ⎢0 ⎥ .
⎢1⎥ ⎢ 6 ⎥ ⎢ I ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢6⎥ ⎢ 6.8I ⎥ ⎢ 6I ⎥
⎢I⎥ ⎢ 9.9 ⎥ ⎢ I ⎥
⎣ ⎦ ⎣ ⎦ ⎣ ⎦

109
Y is a special neutrosophic column vector / matrix and each of
the Yi is a 7 × 1 fuzzy column vector, 1 ≤ i ≤ 6.

Example 1.3.20: Let Y = Y1 ∪ Y2 ∪ Y3 ∪ Y4 ∪ Y5 =

⎡ 6.16I ⎤
⎡ 0.8I ⎤ ⎢ 8.02 ⎥
⎢ 6.9 ⎥ ⎢ ⎥
⎡ 0.4I ⎤ ⎢ ⎥ ⎢ 6.7I ⎥
⎡ I ⎤ ⎢ I ⎥ ⎢ 4I ⎥ ⎢ ⎥
⎢ 6 ⎥ ⎡ 0.9I ⎤ ⎢ ⎥ ⎢ ⎥ ⎢ 6I ⎥
⎢ ⎥ ∪ ⎢ 0.6I ⎥ ∪ ⎢ 1 ⎥ ∪ ⎢ 9.6I ⎥ ∪ ⎢ 9 ⎥
⎢ 0.7I ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ 0.I ⎥ ⎢ ⎥
⎢ ⎥ ⎢
⎣ 6.3 ⎥⎦ ⎢ I ⎥ ⎢ ⎥ ⎢ 0.7I ⎥
⎣ 0.8 ⎦ ⎢⎣ 6.7 ⎥⎦ ⎢ 12 ⎥ ⎢ 8.9 ⎥
⎢ 8I ⎥ ⎢ ⎥
⎢ ⎥ ⎢ 0.8 ⎥
⎣⎢ 0.15⎦⎥ ⎢ 7I ⎥
⎣ ⎦

be the special neutrosophic mixed column vector / matrix. We


see each neutrosophic column vector/matrix has different order.
Now we proceed onto define some basic operations called
the special max min operator on special neutrosophic square
matrix and illustrate them with examples.
Let S = S1 ∪ S2 ∪ … ∪ Sn be a special neutrosophic square
matrix where each Si is a m × m neutrosophic matrix, i = 1, 2,
…, n. Let X = X1 ∪ X2 ∪ … ∪ Xn be a special neutrosophic
row vector where each Xi is a 1 × m neutrosophic row vector.
We define the special neutrosophic operation using X and S.
Suppose we define a max min operator, to find

max min (X o S)
= max min {(X1 ∪ X2 ∪ … ∪ Xn), (S1 ∪ S2 ∪ … ∪
Sn)}
= max min{(X1, S1)} ∪ max min {(X2, S2)} ∪ … ∪
max min {(Xn, Sn)}
= Y'1 ∪ Y'2 ∪ … ∪ Y'n
= Y'

110
now Y' may be a special neutrosophic row vector. Now we
calculate using Y' on S we find

max min (Y', S)


= max min {(Y1 ∪ Y2 ∪ … ∪ Yn), (S1 ∪ S2 ∪ … ∪
Sn)}
= max min {Y1, S1} ∪ max min (Y2, S2) ∪ … ∪ max
min {Yn, Sn}
= T1 ∪ T2 ∪ … ∪ Tn
and now we find max min {T, S} and so on.

We illustrate this by the following example.

Example 1.3.21: Let S = S1 ∪ S2 ∪ S3 ∪ S4 ∪ S5 be a special


neutrosophic square matrix and X = X1 ∪ X2 ∪ … ∪ X5 be a
special neutrosophic row vector; where

⎡ 0.7 I 3 2I ⎤ ⎡ I 2I 0 1 ⎤
⎢ 4I 0 I 0 ⎥ ⎢ ⎥
S = ⎢ ⎥ ∪ ⎢0 7 I 0 ⎥ ∪
⎢ I 1 0 0⎥ ⎢1 0 0 I ⎥
⎢ ⎥ ⎢ ⎥
⎣ 7 2I 0 5I ⎦ ⎣0 I 0 2⎦

⎡3 I 0 0 ⎤ ⎡0 2 0 5I ⎤ ⎡ 0 2I 1 0 ⎤
⎢ 0 4 2I 0 ⎥ ⎢ ⎥ ⎢ 2I 0 0 1 ⎥
⎢ ⎥ ∪ ⎢I 0 7 0⎥
∪ ⎢ ⎥
⎢ 5 0 3I 0 ⎥ ⎢3I 0 0 2⎥ ⎢ I 0 1 0⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎣ 0 2 0 4I ⎦ ⎣0 4 I 0⎦ ⎣ 0 4I 0 I ⎦

is a special neutrosophic square matrix with each Si a 4 × 4


neutrosophic square matrix for i = 1, 2, 3, 4, 5 and
X = X1 ∪ X2 ∪ … ∪ X5
= [I 0 0 0] ∪ [0 I 0 0] ∪ [0 0 I 0] ∪ [0 0 0 I] ∪ [0 1 0 I 0]

be the special neutrosophic row vector each Xi is a 1 × 4


neutrosophic row vector. To find special max min of X, S, i.e.
max min {X, S}

111
= max min {(X1 ∪ X2 ∪ X3 ∪ X4 ∪ X5), (S1 ∪ S2 ∪
… ∪ S5)}
= max min {(X1, S1)} ∪ max min {(X2, S2)} ∪ max
min {(X3, S3)} ∪ … ∪ max min {(X5, S5)}

=
⎧ ⎡0.7 I 3 2I ⎤ ⎫
⎪ ⎢ 4I 0 I 0 ⎥ ⎪
⎪ ⎥ ⎪⎬ ∪
max min ⎨[ I 0 0 0] , ⎢
⎪ ⎢ I 1 0 0 ⎥⎪
⎪ ⎢ ⎥
⎩ ⎣0.7 2I 0 5I ⎦ ⎪⎭

⎧ ⎡ I 2I 0 1 ⎤ ⎫
⎪ ⎢0 7 I 0 ⎥ ⎪
⎪ ⎥ ⎪⎬ ∪
max min ⎨[ 0 I 0 0] , ⎢
⎪ ⎢1 0 0 I ⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣0 I 0 2 ⎦ ⎪⎭

⎧ ⎡3 I 0 0 ⎤ ⎫
⎪ ⎢0 4 2I 0 ⎥ ⎪
⎪ ⎥ ⎪⎬ ∪
max min ⎨[ 0 0 I 0] , ⎢
⎪ ⎢5 0 3I 0 ⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣0 2 0 4I ⎦ ⎪⎭

⎧ ⎡0 2 0 5I ⎤ ⎫
⎪ ⎢I ⎪
⎪ 0 7 0 ⎥⎥ ⎪
max min ⎨[ 0 0 0 I ] , ⎢ ⎬ ∪
⎪ ⎢3I 0 0 2 ⎥⎪
⎪ ⎢ ⎥
⎩ ⎣0 4 I 0 ⎦ ⎪⎭

⎧ ⎡ 0 2I 1 0 ⎤ ⎫
⎪ ⎢ 2I 0 0 1 ⎥ ⎪
⎪ ⎥ ⎪⎬
max min ⎨[ 0 1 0 I ] , ⎢
⎪ ⎢ I 0 1 0⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣ 0 4I 0 I ⎦ ⎪⎭

112
How do we define min(m, n), for m < n and n < m cannot be
defined in a natural way as comparison of a neutrosophic
number and the usual reals is not possible. So we have defined
depending on the wishes of the expert according as the expert
wants to give what type of weightage to the indeterminacy.
Suppose n is a neutrosophic number like n = 5I and m = 8 then
min (5I, 8) = 5I and max (5I, 8) is 8; min(2, 7I) is 2 and max(2,
7I) is 7I and like wise when m and n are of the form xI and yI, x
and y real numbers max (xI, yI) = xI or yI according as x > y or
y > x and min (xI, yI) = xI or yI according as x < y or y < x and
max (nI, n) is nI and min (nI, n) = nI.
It is pertinent and important to mention here that if the
expert feels the presence indeterminacy is important he/she
defines min (n, nI) = nI and max (n, nI) = nI this does not affect
any logic for our definition of min max is different. Clearly one
can not have the usual max or min using reals and
indeterminacy.
Now using this mode of calculation we find

max min (S, T)


= [0.7 I I I] ∪ [0 I I 0] ∪ [5 0 I 0] ∪ [0 I I 0] ∪ [1 I 0 I]
= Y1 ∪ Y2 ∪ Y3 ∪ Y4 ∪ Y5
= Y,

Y is again a special neutrosophic row vector/matrix. Now we


can find max min (Y, S) and so on.

Now we proceed on to define special min max operator on


special neutrosophic mixed square matrix.

Now suppose
T = T1 ∪ T2 ∪ T3 ∪ T4 ∪ T5 ∪ … ∪ Tp

(p ≥ 2) be a special neutrosophic mixed square matrix, where Ti


is a ni × ni square neutrosophic matrix i = 1, 2, …, p with ni ≠ nj
i ≠ j for atleast some i and j. 1 ≤ i, j ≤ p. Suppose X = X1 ∪ X2
∪ … ∪ Xp be a special neutrosophic mixed row vector where
each Xi is a 1 × pi neutrosophic row vector i = 1, 2, …, p. To
find

113
max min (X, T)

= max min {(X1 ∪ X2 ∪ … ∪ Xp), (T1 ∪ T2 ∪ … ∪


Tp)}
= max min (X1, T1) ∪ max min (X2, T2) ∪ … ∪ max
min (Xp, Tp)
= Y1 ∪ Y2 ∪ … ∪ Yp
= Y

is once again a special neutrosophic mixed row vector.

We can find max min (Y, S) and so on.

We illustrate this by the following example.

Example 1.3.22: Let V = V1 ∪ V2 ∪ … ∪ V5

⎡ I 0 2 3I 1 ⎤ ⎡7 0 0 I 0⎤
⎢ 0 3I 0 0 2 ⎥ ⎢ 0 4I 0 0 1 ⎥⎥
⎢ ⎥ ⎢
= ⎢ 1 0 1 5 0 ⎥ ∪ ⎢3I 0 1 0 0⎥ ∪
⎢ ⎥ ⎢ ⎥
⎢ 0 0 2I 0 1 ⎥ ⎢0 0 0 4 3I ⎥
⎢⎣ 7 3 0 0 I ⎥⎦ ⎢⎣ 0 0 5I 0 0 ⎥⎦

⎡ 3 0 4I 0 0 ⎤ ⎡I 0 0 0 1 ⎤
⎢ 0 0 0 8 I⎥ ⎢ 2 0 0 0 5I ⎥
⎢ ⎥ ⎢ ⎥
⎢3I 0 0 1 0 ⎥ ∪ ⎢ 0 I 3 0 0 ⎥ ∪
⎢ ⎥ ⎢ ⎥
⎢ 0 2 0 0 0⎥ ⎢ 0 0 2I 0 0 ⎥
⎢⎣ 0 5I 0 1 0 ⎥⎦ ⎢⎣ 0 5 0 7I 0 ⎥⎦

⎡0 0 I 7 0⎤
⎢I 0 0 0 8 ⎥⎥

⎢0 8 0 I 0⎥
⎢ ⎥
⎢ 2I 0 4 0 I⎥
⎢⎣ 0 0 0 0 6 ⎥⎦

114
be a special neutrosophic square matrix and

X = X1 ∪ X2 ∪ X3 ∪ X4 ∪ X5
= [1 0 0 0 I] ∪ [0 I 0 0 0] ∪ [0 1 I 0 0] ∪ [0 0 0 I 1] ∪
[0 0 I 0 1]

be a special neutrosophic row vector. Now using the special


operator viz. min max (X, V) we get the following resultant
vector

min max (X, V)


= min max {(X1 ∪ X2 ∪ … ∪ X5), (V1 ∪ V2 ∪ … ∪
V5)}
= min max (X1, V1) ∪ … min max (X5, V5)

⎧ ⎡ I 0 2 3I 1 ⎤ ⎫
⎪ ⎢ 0 3I 0 0 2 ⎥ ⎪
⎪ ⎢ ⎥ ⎪⎪

= min max ⎨[1 0 0 0 I ] , ⎢1 0 1 5 0 ⎥ ⎬
⎪ ⎢ ⎥⎪
⎪ ⎢ 0 0 2I 0 1 ⎥ ⎪
⎪⎩ ⎢⎣ 7 3 0 0 I ⎥⎦ ⎪

⎧ ⎡7 0 0 I 0⎤⎫
⎪ ⎢ 0 4I 0 ⎪
⎪ ⎢ 0 I ⎥⎥ ⎪
⎪ ⎪
∪ min max ⎨[ 0 I 0 0 0] , ⎢3I 0 1 0 0⎥⎬
⎪ ⎢ ⎥
⎢0 0 0 4 3I ⎥ ⎪
⎪ ⎪
⎪⎩ ⎢⎣ 0 0 5I 0 0 ⎥⎦ ⎪⎭

⎧ ⎡ 3 0 4I 0 0 ⎤ ⎫
⎪ ⎢ 0 0 0 8 I⎥ ⎪
⎪ ⎢ ⎥ ⎪⎪

∪ min max ⎨[ 0 1 I 0 0] , ⎢3I 0 0 1 0 ⎥ ⎬
⎪ ⎢ ⎥ ⎪
⎪ ⎢ 0 2 0 0 0⎥ ⎪
⎢⎣ 0 5I 0 1 0 ⎥⎦ ⎪
⎩⎪ ⎭

115
⎧ ⎡I 0 01⎤ 0 ⎫
⎪ ⎢2 ⎪
⎪ ⎢ 0 05I ⎥⎥
0 ⎪
⎪ ⎪
∪ min max ⎨ [ 0 0 0 I 1] , ⎢0 I 3 0 0⎥ ⎬
⎪ ⎢ ⎥ ⎪
⎪ ⎢0 0 2I 0 0 ⎥

⎪⎩ ⎢⎣ 0 5 0 7I 0 ⎥⎦ ⎪⎭

⎧ ⎡0 0 I 7 0⎤ ⎫
⎪ ⎢I ⎪
⎪ ⎢ 0 0 0 8 ⎥⎥ ⎪
⎪ ⎪
∪ min max ⎨ [ 0 0 I 0 1] , ⎢0 8 0 I 0⎥ ⎬
⎪ ⎢ ⎥
⎢ 2I 0 4 0 I⎥ ⎪
⎪ ⎪
⎢⎣ 0 0 0 0 6 ⎥⎦ ⎪⎭
⎩⎪

= [0 0 0 0 0] ∪ [0 0 0 0 0] ∪ [0 0 0 0 0] ∪ [0 0 0 0 0]
∪ [0 0 0 0 0].

Thus the resultant of X on V is zero.


Now we proceed on to describe by an example the special
min max and max min operations using on special neutrosophic
mixed square matrices.

Example 1.3.23: Let X = X1 ∪ X2 ∪ X3 ∪ X4 ∪ X5

⎡0 3 I 0 1⎤
⎢1 0 0 I 0 ⎥⎥

= ⎢0 I 0 0 1⎥ ∪
⎢ ⎥
⎢I 0 0 0 0⎥
⎢⎣ 0 0 1 I I ⎥⎦

⎡0 4 I 0⎤
⎢1 0 0 I ⎥
⎢ ⎥ ∪
⎢ I 0 4 0⎥
⎢ ⎥
⎣0 I 0 0⎦

116
⎡0 2 0 I 1⎤
0
⎢1 0 0 0 2I 0 ⎥⎥

⎢I 0 3 0 0 0⎥
⎢ ⎥ ∪
⎢0 0 0 0 0 0⎥
⎢ 0 3I 0 4 0 3I ⎥
⎢ ⎥
⎢⎣ 0 0 2I 0 7 0 ⎥⎦

⎡0 1 I 0⎤
⎡0 7 I ⎤ ⎢
⎢ I 0 1 ⎥ ∪ ⎢ 2I 0 0 4 ⎥⎥
⎢ ⎥ ⎢ 0 3I 0 1⎥
⎢⎣ 0 I 0 ⎥⎦ ⎢ ⎥
⎣5 0 6 0⎦

be the special neutrosophic mixed square matrix. Suppose

T = T1 ∪ T2 ∪ … ∪ T5
= [I 0 0 0 1] ∪ [0 1 0 0] ∪ [I 0 1 0 0 1] ∪ [0 1 0] ∪ [I
0 1 0]

be the mixed special neutrosophic mixed row matrix. We find


using the min max operator on T and X and obtain the resultant.

min max (T, X)


= min max {(T1 ∪ T2 … ∪ T5), (X1 ∪ X2 ∪ … ∪
X5)}
= min max {T1, X1} ∪ min max {(T2, X2)} ∪ minmax
{T3, X3} ∪ min max{T4, X4} ∪ min max{T5, X5}

⎧ ⎡0 3 I 0 1⎤ ⎫
⎪ ⎢1 ⎪
⎪ ⎢ 0 0 I 0 ⎥⎥ ⎪
⎪ ⎪
= min max ⎨ [ I 0 0 0 1] , ⎢0 I 0 0 1⎥ ⎬ ∪
⎪ ⎢ ⎥
⎢I 0 0 0 0⎥ ⎪
⎪ ⎪
⎪⎩ ⎢⎣0 0 1 I I ⎥⎦ ⎭⎪

117
⎧ ⎡0 4 I 0⎤ ⎫
⎪ ⎢1 0 0 I ⎥ ⎪
⎪ ⎥ ⎪⎬ ∪
min max ⎨ [ 0 1 0 0] , ⎢
⎪ ⎢ I 0 4 0⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣ 0 I 0 0 ⎦ ⎭⎪

⎧ ⎡0 2 0 I 0 1 ⎤⎫
⎪ ⎢1 0 0 ⎪
⎪ ⎢ 0 2I 0 ⎥⎥ ⎪
⎪⎪ ⎢I 0 3 0 0 0 ⎥ ⎪⎪
min max ⎨ [ I 0 1 0 0 1] , ⎢ ⎥⎬ ∪
⎪ ⎢0 0 0 0 0 0 ⎥⎪
⎪ ⎢0 3I 0 4 0 3I ⎥ ⎪
⎪ ⎢ ⎥⎪
⎩⎪ ⎣⎢0 0 2I 0 7 0 ⎦⎥ ⎪⎭

⎧ ⎡0 7 I ⎤ ⎫
⎪ ⎢ I 0 1⎥ ⎪ ∪
min max ⎨ [ 0 1 0] , ⎢ ⎥ ⎬
⎪ ⎢⎣ 0 I 0 ⎥⎦ ⎭⎪

⎧ ⎡0 1 I 0⎤ ⎫
⎪ ⎢ 2I 0 0 ⎪
⎪ 4 ⎥⎥ ⎪
min max ⎨ [ I 0 1 0] , ⎢ ⎬
⎪ ⎢ 0 3I 0 1⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎩ ⎣5 0 6 0⎦ ⎭

= [0 0 0 0 0] ∪ [0 0 0 0] ∪ [0 0 0 0 0 0] ∪ [0 I 0] ∪
[I 0 0 0]
= S1 ∪ S2 ∪ S3 ∪ S4 ∪ S5
= S

is again a special neutrosophic mixed row vector.

Now using S we can find min max {S, X} and so on.

118
Next we proceed on to find using the special min max function
the value of a special neutrosophic row vector and the special
neutrosophic rectangular matrix.

Let W = W1 ∪ W2 ∪ … ∪ Wn (n ≥ 2) be a special neutrosophic


rectangular matrix where each Wi is a t × s (t ≠ s) rectangular
neutrosophic matrix (i = 1, 2, …, n).
Let X = X1 ∪ X2 ∪ … ∪ Xn (n ≥ 2) be the special neutrosophic
row vector/matrix where each Xi is a 1 × t neutrosophic vector i
= 1, 2, …, n.
To find

max min ({X, W})

= max min {(X1 ∪ X2 ∪ … ∪ Xn), (W1 ∪ W2 ∪ … ∪


Wn)}
= max min {X1, W1} ∪ max min {X2, W2} ∪ … ∪
max min {Xn, Wn}
= Y1 ∪ Y2 ∪ … ∪ Yn
= Y

now Y is a special neutrosophic row vector where each Yi is a 1


× s neutrosophic row vector for i = 1, 2, …, n.
Now we see max min {Y, W} is not defined so we find the
value of

max min {Y, WT}

= max min {(Y1 ∪ Y2 ∪ … ∪ Yn), ( W1T ∪ W2T ∪ …


∪ WnT )}
= max min (Y1, W1T ) ∪ max min (Y2, W2T ) ∪ … ∪
max min (Yn, WnT )
= T1 ∪ T2 ∪ … ∪ Tn
= T

119
where T is a special neutrosophic row vector with each Ti a 1 × t
neutrosophic row vector for i = 1, 2, …, n. We now find out
max min {T, W} and so on.

Now we illustrate this situation by the following example.

Example 1.3.24: Let W = W1 ∪ W2 ∪ … ∪ W5

⎡ 0 4 I 2 4I 0 0 7 1⎤
= ⎢ 0 0 0 0 0 8I 5 0 1⎥ ∪
⎢ ⎥
⎢⎣1 0 7I 0 0 0 3 0 1⎥⎦

⎡1 I 4 5 7 0 0 0 0 ⎤
⎢0 1 0 I 2 1 0 0 8 ⎥ ∪
⎢ ⎥
⎢⎣1 0 0 0 1 0 I 0 4I ⎥⎦

⎡9I 0 2 0 0 0 1 4 I ⎤
⎢ 0 I 3 0 0 7 0 5 0⎥ ∪
⎢ ⎥
⎢⎣ 2 0 1 0 8 0 I 0 1 ⎥⎦

⎡ 0 3I 0 0 0 2 7 1 0 ⎤
⎢0 0 5 7 I 3 0 1 0⎥ ∪
⎢ ⎥
⎢⎣ 7 0 0 0 2 0 8 0 8 ⎥⎦

⎡0 0 7 0 0 0 2 5 I ⎤
⎢ I 3 0 0 0 4 0 0 2⎥
⎢ ⎥
⎢⎣ 0 0 I 7I 8 0 0 0 1 ⎥⎦

be a special neutrosophic rectangular matrix where each Wi is a


3 × 9 neutrosophic rectangular matrix; i = 1, 2, …, 5. Suppose

X = X1 ∪ X2 ∪ … ∪ X5
= [0 0 I] ∪ [I 3I 0] ∪ [5 0 2I] ∪ [0 7I 0] ∪ [2 I 6]

120
be the special neutrosophic row vector where each Xj is a 1 × 3
row vector, j = 1, 2, 3, …, 5. To find the value of

max min (X, W)


= max min {(X1 ∪ X2 ∪ … ∪ X5), (W1 ∪ W2 ∪ … ∪
W5)}
= max min (X1, W1) ∪ max min (X2, W2) ∪ … ∪
max min (X5, W5)
=
⎧ ⎡ 0 4 I 2 4I 0 0 7 1⎤ ⎫
⎪ ⎪
max min ⎨[ 0 0 I ] , ⎢⎢ 0 0 0 0 0 8I 5 0 1⎥⎥ ⎬ ∪
⎪ ⎢⎣1 0 7I 0 0 0 3 0 1⎥⎦ ⎪⎭

⎧ ⎡1 I 4 5 7 0 0 0 0 ⎤ ⎫
⎪ ⎢0 1 0 I 2 1 0 0 8 ⎥ ⎪
max min ⎨[ I 3I 0] , ⎢ ⎥ ⎬ ∪
⎪ ⎢⎣1 0 0 0 1 0 I 0 4I ⎥⎦ ⎪
⎩ ⎭

⎧ ⎡9I 0 2 0 0 0 1 4 I ⎤ ⎫
⎪ ⎢ 0 I 3 0 0 7 0 5 0⎥ ⎪
max min ⎨[5 0 2I ] , ⎢ ⎥ ⎬ ∪
⎪ ⎢⎣ 2 0 1 0 8 0 I 0 1 ⎥⎦ ⎪
⎩ ⎭

⎧ ⎡ 0 3I 0 0 0 2 7 1 0 ⎤ ⎫
⎪ ⎢0 0 5 7 I 3 0 1 0⎥ ⎪ ∪
max min ⎨[ 0 7I 0] , ⎢ ⎥ ⎬
⎪ ⎢⎣ 7 0 0 0 2 0 0 0 8 ⎥⎦ ⎭⎪

⎧ ⎡0 0 7 0 0 0 2 5 I ⎤ ⎫
⎪ ⎢ I 3 0 0 0 4 0 0 2⎥ ⎪
max min ⎨[ 2 I 6] , ⎢ ⎥ ⎬
⎪ ⎢⎣0 0 I 7I 8 0 0 0 1 ⎥⎦ ⎪⎭

= [I 0 I 0 0 0 I 0 I] ∪ [I 1 I I 2 1 0 0 3I] ∪ [5 0 2 0 2I
0 I 4 I] ∪ [0 0 5 7I I 3 0 1 0] ∪ [I I 2 6 6 I 2 2 I]
= Y1 ∪ Y2 ∪ Y3 ∪ Y4 ∪ Y5
= Y;

121
now Y is a special neutrosophic row vector and each Yi is a 1 ×
9 neutrosophic row vector, i = 1, 2, 3, 4, 5.
Now we calculate the using special max min find the value
of {Y, WT} i.e.

max min (Y, WT)

= max min {[Y1 ∪ Y2 ∪ … ∪ Y5], [ W1T ∪ W2T ∪ …


∪ W5T ]}
= max min (Y1, W1T ) ∪ max min (Y2, W2T ) ∪ … ∪
max min (Y5, W5T )

⎧ ⎡0 0 1⎤ ⎫
⎪ ⎢4 0 0⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢ I 0 7I ⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢2 0 0⎥ ⎪
⎪ ⎪
= max min ⎨[ I 0 I 0 0 0 I 0 I ] , ⎢ 4I 0 0 ⎥ ⎬ ∪
⎪ ⎢ ⎥ ⎪
⎪ ⎢ 0 8I 0 ⎥ ⎪
⎪ ⎢0 5 3⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢7 0 0⎥ ⎪
⎪ ⎢1 1 1⎥ ⎪
⎩ ⎣ ⎦ ⎭

⎧ ⎡1 0 1⎤ ⎫
⎪ ⎢I ⎪
⎪ ⎢ 1 0 ⎥⎥ ⎪
⎪ ⎢4 0 0⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢5 I 0⎥ ⎪
⎪ ⎪
max min ⎨[ I 1 I I 2 1 0 0 3I ] , ⎢7 2 1⎥ ⎬ ∪
⎪ ⎢ ⎥ ⎪
⎪ ⎢0 1 0⎥

⎪ ⎢0 0 I⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢0 0 0⎥ ⎪
⎪ ⎢0 4I ⎥⎦ ⎪⎭
⎩ ⎣ 8

122
⎧ ⎡9I 0 2⎤ ⎫
⎪ ⎢0 ⎪
⎪ ⎢ I 0 ⎥⎥ ⎪
⎪ ⎢2 3 1⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢0 0 0⎥ ⎪
⎪ ⎪
max min ⎨[5 0 2 0 2I 0 I 4 I ] , ⎢0 0 8⎥ ⎬ ∪
⎪ ⎢ ⎥ ⎪
⎪ ⎢0 7 0⎥

⎪ ⎢1 0 I⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢4 5 0⎥ ⎪
⎪ ⎢I 0 1 ⎥⎦ ⎪⎭
⎩ ⎣

⎧ ⎡0 0 7⎤ ⎫
⎪ ⎢3I ⎪
⎪ ⎢ 0 0 ⎥⎥ ⎪
⎪ ⎢0 5 0⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢0 7 0⎥ ⎪
⎪ ⎪
max min ⎨[ 0 0 5 7I I 3 0 1 0] , ⎢0 I 2⎥ ⎬ ∪
⎪ ⎢ ⎥
⎢2 3 0⎥ ⎪
⎪ ⎪
⎪ ⎢7 0 0⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢1 1 0⎥ ⎪
⎪ ⎢0 0 8 ⎥⎦ ⎪⎭
⎩ ⎣

⎧ ⎡0 I 0⎤ ⎫
⎪ ⎢0 ⎪
⎪ ⎢ 3 0 ⎥⎥ ⎪
⎪ ⎢7 0 I⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢0 0 7I ⎥ ⎪
⎪ ⎪
max min ⎨[ I I I 6 6 I 2 2 I ] , ⎢0 0 8⎥ ⎬
⎪ ⎢ ⎥ ⎪
⎪ ⎢0 4 0⎥

⎪ ⎢2 0 0⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢5 0 0⎥ ⎪
⎪ ⎢I 1 ⎥⎦ ⎪⎭
⎩ ⎣ 2

123
= [I I I] ∪ [2 3I 3I] ∪ [5 4 2] ∪ [2 5 I] ∪ [2 I 6]
= P1 ∪ P2 ∪ P3 ∪ P4 ∪ P5
= P,

P is a special neutrosophic row vector, where each Pi is a 1 × 3


neutrosophic row vector i = 1, 2, …, 5. Now we can as before
calculate max min (P, W) and so on.

Next we proceed on to show how max min function works


when we have a special neutrosophic mixed rectangular matrix.
Suppose V = V1 ∪ V2 ∪ … ∪ Vn (n ≥ 2) be a special
neutrosophic mixed rectangular matrix where each Vi is a ti × si
(ti ≠ si) rectangular neutrosophic matrix i = 1, 2, …, n. We have
atleast for one pair i and j; i ≠ j, ti ≠ tj (or si ≠ sj), 1 ≤ j, i ≤ n.
Let X = X1 ∪ X2 ∪ … ∪ Xn (n ≥ 2) special neutrosophic
mixed row vector where each Xi is a 1 × ti neutrosophic row
vector, i = 1, 2, …, n. Now special max min of {(X, V)}

= max min {(X1 ∪ X2 ∪ … ∪ Xn), (V1 ∪ V2 ∪ … ∪


Vn)}
= max min {X1, V1} ∪ max min {X2, V2} ∪ … ∪
max min (Xn, Vn}
= Y1 ∪ Y2 ∪ … ∪ Yn
= Y

where Y is a special neutrosophic mixed row vector. Now we


find max min of {Y, VT}

= max min {(Y1 ∪ Y2 ∪ … ∪ Yn), ( V1T ∪ V2T ∪ …


∪ VnT )}
= max min {Y1, V1T } ∪ max min {Y2, V2T } ∪ … ∪
max min {Yn, VnT }
= Z1 ∪ Z2 ∪ … ∪ Zn
= Z,

where Z is a special neutrosophic mixed row vector. We can if


need be find max min {Z, V} and so on.

124
Now we illustrate this situation by the following example.

Example 1.3.25: Let V = V1 ∪ V2 ∪ V3 ∪ V4 ∪ V5 be a special


neutrosophic mixed rectangular matrix given by

⎡ 0 7 0 I 5 3I 1 ⎤
V = ⎢⎢ I 0 0 0 6 5 2 ⎥⎥ ∪
⎢⎣ 6 2I 1 0 7 0 3 ⎥⎦

⎡5 3 I 7⎤
⎢ 0 2I 0 0 ⎥
⎢ ⎥
⎢8I 0 0 6 ⎥
⎢ ⎥
⎢1 7 0 0⎥ ∪
⎢ 0 0 8 6I ⎥
⎢ ⎥
⎢1 0 0 2⎥
⎢ 0 0 6I 1 ⎥
⎣ ⎦

⎡ 0 3 1 0 9I ⎤
⎢ 9 0 0 3I 2 ⎥
⎢ ⎥
⎢1 4 6 0 7 ⎥
⎢ ⎥ ∪
⎢0 0 0 8 0 ⎥
⎢ 0 2I 0 0 0 ⎥
⎢ ⎥
⎣⎢ 7 0 2I 1 9 ⎦⎥

⎡5 0 I 0 1 0 6 0⎤
⎢ 0 9 0 7 2 4I 7 2 ⎥
⎢ ⎥ ∪
⎢ 0 2I 6 0 3 5 8I 3 ⎥
⎢ ⎥
⎣ 2 0 0 I 0 0 9 4⎦

125
⎡8 0 0 7I 0 0 9⎤0
⎢4 0 6 0 0 2I 0 1 ⎥⎥

⎢0 0 I 0 0 0 I 3⎥ .
⎢ ⎥
⎢ 2I 3 0 2 7 0 0 2I ⎥
⎢⎣ 1 2 3 4 5 6 7 8 ⎥⎦

Suppose

X = X1 ∪ X2 ∪ … ∪ X5
= [6 0 I] ∪ [4 0 0 0 0 I 0] ∪ [0 0 2I 0 9 0] ∪ [0 0 0 6]
∪ [I 0 0 7 0]

where X is a special neutrosophic mixed row matrix.

max min (X, V)


= max min {(X1 ∪ X2 ∪ … ∪ X5), (V1 ∪ V2 ∪ … ∪
V5)}
= max min {X1, V1} ∪ max min {X2, V2} ∪ … ∪
max min {X5, V5}

⎧ ⎡0 7 0 I 5 3I 1 ⎤ ⎫
⎪ ⎢ I 0 0 0 6 5 2⎥ ⎪ ∪
= max min ⎨[ 6 0 I] , ⎢ ⎥⎬
⎪ ⎢⎣6 2I 1 0 7 0 3 ⎥⎦ ⎪⎭

⎧ ⎡5 3 I 7⎤ ⎫
⎪ ⎢ 0 2I 0 0 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪ ⎢8I 0 0 6 ⎥ ⎪
⎪ ⎢ ⎥⎪
max min ⎨[ 4 0 0 0 0 I 0] , ⎢ 1 7 0 0 ⎥ ⎬ ∪
⎪ ⎢ 0 0 8 6I ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪ ⎢1 0 0 2⎥ ⎪
⎪ ⎢ 0 0 6I 1 ⎥ ⎪
⎩ ⎣ ⎦⎭

126
⎧ ⎡ 0 3 1 0 9I ⎤ ⎫
⎪ ⎢ 9 0 0 3I 2 ⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪⎪ ⎢1 4 6 0 7 ⎥ ⎪⎪
max min ⎨[ 0 0 2I 0 9 0] , ⎢ ⎥ ⎬∪
⎪ ⎢0 0 0 8 0 ⎥ ⎪
⎪ ⎢ 0 2I 0 0 0 ⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎩⎪ ⎣⎢7 0 2I 1 9 ⎦⎥ ⎭⎪

⎧ ⎡5 0 I 0 1 0 6 0⎤ ⎫
⎪ ⎢ 0 9 0 7 2 4I 7 2 ⎥ ⎪
⎪ ⎥ ⎪⎬ ∪
max min ⎨[ 0 0 0 6] , ⎢
⎪ ⎢ 0 2I 6 0 3 5 8I 3 ⎥⎪
⎪ ⎢ ⎥
⎩ ⎣ 2 0 0 I 0 0 9 4 ⎦ ⎪⎭

⎧ ⎡8 0 0 4I 0 0 0 9 ⎤⎫
⎪ ⎢4 0 ⎪
⎪ ⎢ 6 0 0 2I 0 1 ⎥⎥ ⎪
⎪ ⎪
max min ⎨[ I 0 0 7 0] , ⎢ 0 0 I 0 0 0 I 3 ⎥⎬
⎪ ⎢ ⎥
⎢ 2I 3 0 2 7 0 0 2I ⎥ ⎪
⎪ ⎪
⎪⎩ ⎢⎣ 1 2 3 4 5 6 7 8 ⎥⎦ ⎪⎭

= [I 6 I I 5 3I I] ∪ [4 3 I 4] ∪ [1 2I 2I 0 2I] ∪ [2 0 0 I
0 0 6 4] ∪ [2I 3 0 2 7 0 0 2I]
= Y1 ∪ Y2 ∪ Y3 ∪ Y4 ∪ Y5
= Y

where Y is a special neutrosophic mixed row vector. Now we


use Y and VT and find the value of

max min (Y, VT)


= max min {(Y1 ∪ Y2 ∪ … ∪ Y5), ( V1T ∪ V2T ∪ …
∪ V5T )}
= max min {Y1, V1T } ∪ max min {Y2, V2T } ∪ max
min {Y3, V3T } ∪ max min {Y4, V4T } ∪ max min
{Y5, V5T }

127
⎧ ⎡0 I 6⎤ ⎫
⎪ ⎢7 ⎪
⎪ ⎢ 0 2I ⎥⎥ ⎪
⎪ ⎢0 0 1⎥ ⎪
⎪ ⎢ ⎥ ⎪
= max min ⎨[ I 6 I I 5 3I I ] , ⎢I 0 0⎥ ⎬ ∪
⎪ ⎢5 6 7⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢3I 5 0⎥ ⎪
⎪ ⎢1 ⎪
⎩ ⎣ 2 3 ⎥⎦ ⎭

⎧ ⎡ 5 0 8I 1 0 1 0⎤ ⎫
⎪ ⎢ 3 2I 0 7 0 ⎪
⎪ 0 0 ⎥⎥ ⎪
max min ⎨[ 4 3 I 4] , ⎢ ⎬ ∪
⎪ ⎢I 0 0 0 8 0 6I ⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎩ ⎣ 7 0 6 0 6I 2 1⎦ ⎭

⎧ ⎡0 9 1 0 0 7⎤ ⎫
⎪ ⎢3 0 4 ⎪
⎪ ⎢ 0 2I 0 ⎥⎥ ⎪
⎪ ⎪
max min ⎨[1 2I 2I 0 2I] , ⎢1 0 6 0 0 2I ⎥ ⎬ ∪
⎪ ⎢ ⎥ ⎪
⎪ ⎢ 0 3I 0 8 0 1⎥

⎢⎣9I 2 7 9 ⎥⎦
⎩⎪ 0 0 ⎪⎭

⎧ ⎡5 0 0 2⎤ ⎫
⎪ ⎢ 0 9 2I 0 ⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢ I 0 6 0⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢0 7 0 I ⎥ ⎪
max min ⎨[ 2 0 0 I 0 0 6 4] , ⎬ ∪
⎢1 2 3 0 ⎥
⎪ ⎢ ⎥ ⎪
⎪ ⎢ 0 4I 5 0 ⎥ ⎪
⎪ ⎢ 6 7 8I 9 ⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢⎣ 0 2 3 4 ⎥⎦ ⎪
⎩ ⎭

128
⎧ ⎡ 8 4 0 2I 1 ⎤ ⎫
⎪ ⎢ 0 0 0 3 2⎥ ⎪
⎪ ⎢ ⎥⎪
⎪ ⎢ 0 6 I 0 3⎥ ⎪
⎪ ⎢ ⎥⎪
⎪ ⎢ 4I 0 0 2 4 ⎥ ⎪
max min ⎨[ 2I 3 0 2 7 0 0 2I] ,
⎢ 0 0 0 7 5⎥ ⎬
.
⎪ ⎢ ⎥ ⎪
⎪ ⎢ 0 2I 0 0 6 ⎥ ⎪
⎪ ⎢ 0 0 I 0 7⎥ ⎪
⎪ ⎢ ⎥⎪
⎪ ⎪
⎩ ⎣⎢ 9 1 3 2I 8 ⎦⎥ ⎭

= [6 5 5] ∪ [4 2I 4 3 4 2 I] ∪ [2I 2I 2I 0 2I 2I] ∪ [6 6
6 6] ∪ [2I 2I I 3 2]
= T1 ∪ T2 ∪ T3 ∪ T4 ∪ T5
= T

where T is a special neutrosophic mixed row vector. Now we


can find max min (T, V) and so on.

Now we define special max min function for special


neutrosophic mixed matrix. Let P = P1 ∪ P2 ∪ … ∪ Pn (n ≥ 2)
where Pi is a mi × ti (mi ≠ ti) are neutrosophic rectangular matrix
and Pj’s are pj × pj neutrosophic matrix i ≠ j; 1 ≤ i, j ≤ n.
Suppose
X = X1 ∪ X2 ∪ … ∪ Xn

(n ≥ 2) where Xi’s are 1 × mi neutrosophic mixed row vector


and Xj’s are 1 × pj neutrosophic row vector 1 ≤ i, j ≤ n.
We define

max min (X, P)


= max min {(X1 ∪ X2 ∪ … ∪ Xn), (P1 ∪ P2 ∪ … ∪
Pn)}
= max min {X1, P1} ∪ max min {X2, P2} ∪ … ∪ max
min {Xn, Pn}
= Y1 ∪ Y2 ∪ … ∪ Yn
= Y

129
where Y is a neutrosophic mixed row vector. We find at the
next step max min {Y, PST} (we have defined in pages 85-6 the
notion of special transpose).
Now

max min {Y, PST}


= max min {(Y1 ∪ Y2 ∪ … ∪ Yn),
( P1ST ∪ P2ST ∪ … ∪ PnST ) }
= max min (Y1, P ) ∪ max min (Y2, P2ST ) ∪ … ∪
1
ST

max min (Yn, PnST )


= Z1 ∪ Z2 ∪ … ∪ Zn
= Z

where Z is again a neutrosophic mixed row vector. We can find


max min (Z, P) and so on.

We illustrate this situation by the following example.

Example 1.3.26: Let P = P1 ∪ P2 ∪ P3 ∪ P4 ∪ P5 ∪ P6

⎡3 0 I⎤ ⎡3 0 4 2 ⎤
⎢1 2 3⎥ ⎢I 0 0 8 ⎥
⎢ ⎥ ⎡5 0 1 2 I ⎤ ⎢ ⎥
⎢ 0 4I 0 ⎥ ⎢ 0 3I 0 0 4 ⎥ ⎢0 9 7 0 ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
= ⎢ 5 6 7 ⎥ ∪ ⎢0 0 7 I 6 ⎥ ∪ ⎢6 I 2 0 ⎥ ∪
⎢ 0 0 8I ⎥ ⎢ ⎥ ⎢ 0 0 5I 9 ⎥
⎢ I 8 0 0 0⎥
⎢ ⎥ ⎢ ⎥
⎢9 8 7⎥ ⎢⎣ 0 0 0 4 0 ⎥⎦ ⎢I 0 0 0 ⎥
⎢ 4I 0 0 ⎥ ⎢ 5 3 0 5I ⎥
⎣ ⎦ ⎣ ⎦

⎡ 0 0 6I 2 0 1 0⎤
⎢2 0 0 0 8 2 9 ⎥⎥
⎢ ∪
⎢ 0 I 0 0 3I 3 0⎥
⎢ ⎥
⎣0 0 8 I 0 7 0⎦

130
⎡0 9 0 ⎤
⎢8 0 0 ⎥
⎡ 0 4 2I 0 ⎤ ⎢ ⎥
⎢8 0 0 I ⎥ ⎢0 0 7 ⎥
⎢ ⎥ ∪ ⎢ ⎥
⎢3 I 0 8⎥ ⎢I 0 0 ⎥
⎢ ⎥ ⎢ 0 2I 0 ⎥
⎣ 0 0 5I 0 ⎦
⎢ ⎥
⎢⎣ 0 0 3I ⎥⎦

be a neutrosophic mixed matrix.


Suppose
X = X1 ∪ X2 ∪ … ∪ X6
= [1 0 0 0 0 4 I] ∪ [0 I 7 0 0] ∪ [0 2 0 0 0 I 0] ∪ [0 3
0 I] ∪ [0 3 4I 0] ∪ [9 0 0 0 1 0]

be the neutrosophic mixed row vector. To find


max min (X, P)
= max min {(X1 ∪ X2 ∪ … ∪ X6), (V1 ∪ V2 ∪ … ∪
V6)}
= max min (X1, V1) ∪ max min (X2, V2) ∪ … ∪ max
min (X6, V6)

⎧ ⎡3 0 I⎤ ⎫
⎪ ⎢1 2 3⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢ 0 4I 0 ⎥ ⎪
⎪ ⎢ ⎥ ⎪
= max min ⎨[1 0 0 0 0 4 I ] , ⎢5 6 7⎥ ⎬ ∪
⎪ ⎢ 0 0 8I ⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢9 8 7⎥ ⎪
⎪ ⎢ 4I 0 0 ⎥ ⎪
⎩ ⎣ ⎦ ⎭

⎧ ⎡5 0 1 2 I ⎤ ⎫
⎪ ⎢0 3I 0 0 4 ⎥ ⎪
⎪ ⎢ ⎥ ⎪⎪

max min ⎨[ 0 I 7 0 0] , ⎢0 0 7 I 6 ⎥ ⎬ ∪
⎪ ⎢ ⎥ ⎪
⎪ ⎢ I 8 0 0 0⎥ ⎪
⎢⎣0 0 0 4 0 ⎥⎦ ⎪
⎩⎪ ⎭

131
⎧ ⎡3 0 4 2⎤ ⎫
⎪ ⎢I ⎪
⎪ ⎢ 0 0 8 ⎥⎥ ⎪
⎪ ⎢0 9 7 0⎥ ⎪
⎪ ⎢ ⎥ ⎪
max min ⎨[ 0 2 0 0 0 I 0] , ⎢6 I 2 0⎥ ⎬ ∪
⎪ ⎢0 0 5I 9 ⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢I 0 0 0⎥ ⎪
⎪ ⎢5 ⎪
⎩ ⎣ 3 0 5I ⎥⎦ ⎭

⎧ ⎡ 0 0 6I 2 0 1 0⎤ ⎫
⎪ ⎢2 0 0 0 8 2 ⎪
⎪ 9 ⎥⎥ ⎪
max min ⎨[ 0 3 0 I ] , ⎢ ⎬∪
⎪ ⎢ 0 I 0 0 3I 3 0⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎩ ⎣0 0 8 I 0 7 0⎦ ⎭

⎧ ⎡0 4 2I 0 ⎤ ⎫
⎪ ⎢8 0 0 I ⎥ ⎪
⎪ ⎥ ⎪⎬ ∪
max min ⎨[ 0 3 4I 0] , ⎢
⎪ ⎢3 I 0 8⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣0 0 5I 0 ⎦ ⎪⎭

⎧ ⎡0 9 0 ⎤ ⎫
⎪ ⎢8 0 0 ⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪⎪ ⎢0 0 7 ⎥ ⎪⎪
max min ⎨[9 0 0 0 1 0] , ⎢ ⎥ ⎬
⎪ ⎢I 0 0 ⎥ ⎪
⎪ ⎢0 2I 0 ⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪⎩ ⎣⎢0 0 3I ⎦⎥ ⎭⎪

= [4 4 4] ∪ [0 I 7 I 6] ∪ [I 0 0 2] ∪ [2 0 I I 3 2 3] ∪
[3 I 0 4I] ∪ [0 9 0]
= Y1 ∪ Y2 ∪ … ∪ Y6
= Y
is again a special neutrosophic mixed row vector. Using Y we
find out

132
max min {Y, PST}
= max min {(Y1 ∪ Y2 ∪ … ∪ Y6) ∪ (( P1T ∪ P2 ∪ P3T
∪ P4T ∪ P5 ∪ P6T )}
= max min {Y1, P1T } ∪ max min {Y2, P2} ∪ max min
{Y3, P3T } ∪ max min {Y4, P4T } ∪ max min {Y5,
P5} ∪ max min {Y6, P6T }
=
⎧ ⎡ 3 1 0 5 0 9 4I ⎤ ⎫
⎪ ⎢ 0 2 4I 6 0 8 0 ⎥ ⎪ ∪
max min ⎨[ 4 4 4] , ⎢ ⎥ ⎬
⎪ ⎢⎣ I 3 0 7 8I 7 0 ⎥⎦ ⎪⎭

⎧ ⎡5 0 1 2 I ⎤ ⎫
⎪ ⎢0 3I 0 0 4 ⎥ ⎪
⎪ ⎢ ⎥ ⎪⎪

max min ⎨[ 0 I 7 I 6] , ⎢0 0 7 I 6 ⎥ ⎬ ∪
⎪ ⎢ ⎥ ⎪
⎪ ⎢ I 8 0 0 0⎥ ⎪
⎢⎣0 0 0 4 0 ⎥⎦ ⎪
⎩⎪ ⎭

⎧ ⎡3 I 0 6 0 I 5 ⎤ ⎫
⎪ ⎢0 0 9 I 0 0 3 ⎥ ⎪
⎪ ⎪
max min ⎨[ I 0 0 2] , ⎢ ⎥ ⎬ ∪
⎪ ⎢ 4 0 7 2 5I 0 0 ⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎩ ⎣ 2 8 0 0 9 0 5I ⎦ ⎭

⎧ ⎡0 2 0 0⎤ ⎫
⎪ ⎢0 ⎪
⎪ ⎢ 0 I 0 ⎥⎥ ⎪
⎪ ⎢6I 0 0 8⎥ ⎪
⎪ ⎢ ⎥ ⎪
max min ⎨[ 2 0 I I 3 2 3] , ⎢2 0 0 I⎥ ⎬ ∪
⎪ ⎢0 8 3I 0 ⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢1 2 3 7⎥ ⎪
⎪ ⎢0 ⎪
⎩ ⎣ 9 0 0 ⎥⎦ ⎭

133
⎧ ⎡0 4 2I 0 ⎤ ⎫
⎪ ⎢8 0 0 I ⎥ ⎪
⎪ ⎥ ⎪⎬ ∪
max min ⎨[3 I 0 4I ] , ⎢
⎪ ⎢3 I 0 8⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣0 0 5I 0 ⎦ ⎪⎭

⎧ ⎡0 8 0 I 0 0 ⎤ ⎫
⎪ ⎢9 0 0 0 2I 0 ⎥ ⎪
max min ⎨[ 0 9 0] , ⎢ ⎥ ⎬
⎪ ⎢⎣ 0 0 7 0 0 3I ⎥⎦ ⎪
⎩ ⎭

= [3 3 4I 4 4 4 4I] ∪ [I I 7 4 6] ∪ [2 2 0 I 2 I 2] ∪ [I 3
3I 2] ∪ [I 3 4I I] ∪ [9 0 0 0 2I 0]
= Z1 ∪ Z2 ∪ … ∪ Z6
= Z.

We see Z is a special neutrosophic mixed vector / matrix. Now


using Z and P we can find max min {Z, P} and so on.
Now using a special neutrosophic mixed matrix we can find
using the min max operator for any special neutrosophic mixed
row vector.

We illustrate this by the following example.

Example 1.3.27: Let T = T1 ∪ T2 ∪ T3 ∪ T4 ∪ T5 be a special


neutrosophic mixed matrix, where

⎡0 5 0⎤
⎢ 2I 0 0 ⎥⎥
⎢ ⎡0 7 9 0⎤
⎢0 0 7⎥ ⎢ 4 0 0 2I ⎥
⎢ ⎥
T= ⎢4 I 0⎥ ∪ ⎢ ⎥ ∪
⎢ 0 8I 0 8 ⎥
⎢8 0 6I ⎥ ⎢ ⎥
⎢ ⎥ ⎣3I 0 5I 0 ⎦
⎢0 7 3I ⎥
⎢2 9 7 ⎥⎦

134
⎡0 0 4 8 0 2 0⎤
⎢1 7 0 0 9 9 4 ⎥⎥

⎢2 I 0 I 0 0 0⎥
⎢ ⎥
⎣ 0 0 I 0 3I 6 2I ⎦

⎡9 0 2I 0 1 6 0 0 ⎤
⎢ 0 8 0 I 0 0 7I 1 ⎥
⎡8 0 1 ⎤ ⎢ ⎥
∪ ⎢⎢ 4 5I 0 ⎥⎥ ∪ ⎢ 0 0 6 0 2I 0 0 0 ⎥
⎢ ⎥
⎣⎢ 0 0 6I ⎦⎥ ⎢ 0 6I 0 0 0 8I 0 0 ⎥
⎢⎣ 0 0 0 3 9 0 8 2I ⎥⎦

be a special neutrosophic mixed matrix.


Suppose

I = I1 ∪ I2 ∪ I3 ∪ I4 ∪ I5
= [0 9 0 0 1 0 1] ∪ [0 2 0 0] ∪ [0 0 0 6] ∪ [6 0 0] ∪
[8 0 0 0 0]

be a special neutrosophic mixed row vector/matrix. To find the


effect of I on T using max min operator.

max min (I, T)


= max min {(I1 ∪ I2 ∪ … ∪ I5), (T1 ∪ T2 ∪ … ∪ T5)}
= max min (I1, T1) ∪ max min (I2, T2) ∪ … ∪ max
min (I5, T5)

⎧ ⎡0 5 0⎤ ⎫
⎪ ⎢ 2I ⎪
⎪ ⎢ 0 0 ⎥⎥ ⎪
⎪ ⎢0 0 7⎥ ⎪
⎪ ⎢ ⎥ ⎪
= max min ⎨[ 0 9 0 0 1 0 1] , ⎢4 I 0⎥ ⎬ ∪
⎪ ⎢8 0 6I ⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢0 7 3I ⎥ ⎪
⎪ ⎢2 ⎪
⎩ ⎣ 9 7 ⎥⎦ ⎭

135
⎧ ⎡0 7 9 0⎤ ⎫
⎪ ⎢ 4 0 0 2I ⎥ ⎪
⎪ ⎥ ⎪⎬
max min ⎨[ 0 2 0 0] , ⎢
⎪ ⎢ 0 8I 0 8 ⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣3I 0 5I 0 ⎦ ⎪⎭

⎧ ⎡0 0 4 8 0 2 0⎤ ⎫
⎪ ⎢1 7 0 0 9 ⎪
⎪ 9 4 ⎥⎥ ⎪
max min ⎨[ 0 0 0 6] , ⎢ ⎬
⎪ ⎢2 I 0 I 0 0 0⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎩ ⎣ 0 0 I 0 3I 6 2I ⎦ ⎭

⎧ ⎡8 0 1 ⎤ ⎫
⎪ ⎢ 4 5I 0 ⎥ ⎪
max min ⎨[ 6 0 0] , ⎢ ⎥ ⎬ ∪
⎪ ⎢⎣ 0 0 6I ⎥⎦ ⎪
⎩ ⎭

⎧ ⎡9 0 2I 0 1 6 0 0 ⎤ ⎫
⎪ ⎢ 0 8 0 I 0 0 7I 1 ⎥ ⎪
⎪ ⎢ ⎥ ⎪⎪

max min ⎨[8 0 0 0 0] , ⎢ 0 0 6 0 2I 0 0 0 ⎥ ⎬
⎪ ⎢ ⎥⎪
⎪ ⎢ 0 6I 0 0 0 8I 0 0 ⎥ ⎪
⎪⎩ ⎢⎣ 0 0 0 3 9 0 8 2I ⎥⎦ ⎪

= [2I 1 1] ∪ [2 0 0 2I] ∪ [0 0 I 0 3I 6 2I] ∪ [6 0 1] ∪


[8 0 2I 0 I 6 0 0]
= P1 ∪ P2 ∪ P3 ∪ P4 ∪ P5
= P

where P is a special neutrosophic mixed row vector. We now


calculate the resultant of P
max min {P, TST)
= max min {(P1 ∪ P2 ∪ … ∪ P5), ( ( T1ST ∪ T2ST ∪ … ∪ T5ST ) }
= max min (P1, T ) ∪ max min (P2, T ) ∪ … ∪
1
ST ST
2
ST
max min (P5, T )
5

136
⎧ ⎡ 0 2I 0 4 8 0 2 ⎤ ⎫
⎪ ⎪
= max min ⎨[ 2I 1 1] , ⎢⎢ 5 0 0 I 0 7 9 ⎥⎥ ⎬ ∪
⎪ ⎢⎣ 0 0 7 0 6I 3I 7 ⎥⎦ ⎪⎭

⎧ ⎡0 7 9 0⎤ ⎫
⎪ ⎢ 4 0 0 2I ⎥ ⎪
⎪ ⎥ ⎪⎬ ∪
max min ⎨[ 2 0 0 2I ] , ⎢
⎪ ⎢ 0 8I 0 8 ⎥ ⎪
⎪ ⎢ ⎥
⎩ ⎣3I 0 5I 0 ⎦ ⎪⎭

⎧ ⎡0 1 2 0⎤ ⎫
⎪ ⎢0 ⎪
⎪ ⎢ 7 I 0 ⎥⎥ ⎪
⎪ ⎢4 0 0 I⎥ ⎪
⎪ ⎢ ⎥ ⎪
max min ⎨[ 0 0 I 0 3I 6 2I] , ⎢8 0 I 0⎥ ⎬ ∪
⎪ ⎢0 9 0 3I ⎥ ⎪
⎪ ⎢ ⎥ ⎪
⎪ ⎢2 9 0 6⎥ ⎪
⎪ ⎢0 ⎪
⎩ ⎣ 4 0 2I ⎥⎦ ⎭

⎧ ⎡8 0 1 ⎤ ⎫
⎪ ⎪
max min ⎨[ 6 0 1] , ⎢⎢ 4 5I 0 ⎥⎥ ⎬ ∪
⎪ ⎪
⎩ ⎣⎢ 0 0 6I ⎦⎥ ⎭

⎧ ⎡ 9 0 0 0 0 ⎤⎫
⎪ ⎢ 0 8 0 6I 0 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪ ⎢ 2I 0 6 0 0 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪ ⎢ 0 I 0 0 3 ⎥⎪
max min ⎨[8 0 2I 0 I 6 0 0] ,
⎢ 1 0 2I 0 9 ⎥ ⎬
⎪ ⎢ ⎥⎪
⎪ ⎢ 6 0 0 8I 0 ⎥ ⎪
⎪ ⎢ 0 7I 0 0 8 ⎥ ⎪
⎪ ⎢ ⎥⎪
⎪ ⎢⎣ 0 1 0 0 2I ⎥⎦ ⎪⎭

137
= [1 2I 1 2I 2I 1 2I] ∪ [2I 2 2I 0] ∪ [2 6 0 6] ∪ [6 0
1] ∪ [8 0 I 6 I]
= Z1 ∪ Z2 ∪ … ∪ Z5
= Z;

where Z is a special neutrosophic mixed row vector.

Z can now work on T as follows.

max min {Z, T}

= max min {(Z1 ∪ Z2 ∪ … ∪ Z5), (T1 ∪ T2 ∪ … ∪


T5)}
= max min {Z1, T} ∪ max min {Z2, T2} ∪ … ∪ max
min {Z5, T5}
= R1 ∪ R2 ∪ … ∪ R5
= R,

where R is a special neutrosophic mixed row vector.


We can proceed on to find max min {R, TST} and so on.

Other types of operation on the special fuzzy neutrosophic


matrices will be described in the following.

Recall a special neutrosophic matrix will be known as the


special fuzzy neutrosophic matrix if its entries are from Nf = {[0
I] ∪ [0 1]} where Nf is the fuzzy neutrosophic interval.
Any element from Nf will be of the form y = a + bI where a
and b ∈ [0, 1]. Thus any typical element will be 0.7 + 0.5I. If x
∈ Nf is of the form x = 0.7 I or I then we call x to be an absolute
or pure neutrosophic fuzzy number or pure fuzzy neutrosophic
number. y will be known as a fuzzy neutrosophic number or
neutrosophic fuzzy number if and only if a ≠ 0 and b ≠ 0. Thus
in a pure fuzzy neutrosophic number is one in which a = 0. If in
y = a + bI, b = 0 then we call y to be a fuzzy number. Thus all
special fuzzy neutrosophic matrices are special neutrosophic
matrices but we see all special neutrosophic matrices need not
be special fuzzy neutrosophic matrices.

138
Example 1.3.28: Let
⎡ 9 ⎤
⎡ 0.8I 7 I ⎤ ⎢ 0 ⎥
⎢ 0 0 8 ⎥ ⎢ ⎥
P = [ 0.8 7 5I 2 0.7I ] ∪ ⎢ ⎥ ∪ ⎢ 0.8I ⎥
⎢ I 0 0.3 ⎥ ⎢ ⎥
⎢ ⎥ ⎢ 0.6 ⎥
⎣ 4 18 0.4I ⎦ ⎢⎣ 6 ⎥⎦

is a special neutrosophic mixed rectangular matrix. Clearly P is


not a special neutrosophic fuzzy mixed rectangular matrix.

Now as in case of special neutrosophic matrices we can define


operations on special fuzzy neutrosophic matrices also. We can
apply threshold and updating operation at each stage and obtain
the resultant. This sort of operation will be used in special fuzzy
models which will be defined in chapter one of this book.

Now we define the special operation on the special fuzzy row


vector with special square matrix T.

Let T = T1 ∪ T2 ∪ … ∪ Tn be a special fuzzy neutrosophic


square matrix. Let each Ti be a m × n fuzzy neutrosophic
matrix. X = X1 ∪ X2 ∪ … ∪ Xn be a special fuzzy neutrosophic
row vector where each Xi is a fuzzy neutrosophic row vector
taking its value from the set {0, 1, I}; i = 1, 2, …, n.

Now we define a special operation using X and T.

XoT = (X1 ∪ … Xn) o (T1 ∪ … ∪ Tn)


= X1 o T1 ∪ X2 o T2 ∪ … ∪ Xn o Tn

where the operation Xi o Ti 1 ≤ i ≤ is described in the following.

Let
Xi = [0 I 0 0 0 1]
and

139
⎡0 0 0 0 0 1⎤
⎢1 0 0 0 I 0 ⎥⎥

⎢0 0 0 1 0 I⎥
Ti = ⎢ ⎥
⎢I 0 0 0 1 0⎥
⎢0 0 0 0 0 1⎥
⎢ ⎥
⎢⎣ 0 0 0 1 0 0 ⎥⎦

be the fuzzy neutrosophic matrix. We find

⎡0 0 0 0 0 1⎤
⎢1 0 0 0 I 0 ⎥⎥

⎢0 0 0 1 0 I⎥
Xi o Ti = [ 0 I 0 0 0 1] o ⎢ ⎥
⎢I 0 0 0 1 0⎥
⎢0 0 0 0 0 1⎥
⎢ ⎥
⎣⎢0 0 0 1 0 0 ⎦⎥

= [I 0 0 1 I 0] (we threshold any resultant vector in the


following way. Suppose

Xj o Tj = [a1 a2 … an]

then if ai is a any real number than


ai = 0 if ai ≤ 0
ai = 1 if ai > 0
if ai is a pure neutrosophic number of the form say miI and
if mi ≤ 0 then ai = 0
if mi > 0 then ai = I.

If ai is not a pure neutrosophic number and is of the form ai = ti


+ siI and if ti > si and if ti ≤ 0 then ai = 0. If ti > 0 then ai = 1.
Suppose si > ti and if si ≤ 0 than ai = 0. If si > 0 then ai = 1. If si =
ti then ai = I. We using this type of thresholding technique work
with the resultant vector. The updating is done as in case of
fuzzy vectors if the tth coordinate is 1 in the starting and if in the

140
resultant it becomes 0 or I we replace the tth coordinate once
again by 1).
(So
X o Ti = [I 0 0 1 I 0] = Y'i

after updating and thresholding Y'i becomes equal to Yi = [I I 0


1 I 1]. Now we find

⎡0 0 0 0 0 1⎤
⎢1 0 0 0 I 0 ⎥⎥

⎢0 0 0 1 0 I⎥
Yi o Ti = [ I I 0 1 I 1] o ⎢ ⎥
⎢I 0 0 0 1 0⎥
⎢0 0 0 0 0 1⎥
⎢ ⎥
⎣⎢ 0 0 0 1 0 0 ⎦⎥

= [2I 0 0 1 1 + I 2I]

after updating and thresholding we get Z = [I I 0 1 I 1].

⎡0 0 0 0 0 1⎤
⎢1 0 0 0 I 0 ⎥⎥

⎢0 0 0 1 0 I⎥
Z o Ti = [ I I 0 1 I 1] o ⎢ ⎥
⎢I 0 0 0 1 0⎥
⎢0 0 0 0 0 1⎥
⎢ ⎥
⎢⎣0 0 0 1 0 0 ⎥⎦

= [2I 0 0 1 I + 1 2I]

after updating and thresholding we get S = [I I 0 1 I 1] which is


a fixed point).
Now we find

XoT
= (X1 ∪ X2 ∪ … ∪ Xn) o (T1 ∪ T2 ∪ … ∪ Tn)
= X1 o T1 ∪ X2 ∪ T2 ∪ … ∪ Xn o Tn

141
(where the operations are carried out as
described above)
= Y'1 ∪ Y'2 ∪ … ∪ Y'n
= Y'.

Now Y' is updated and thresholded to Y = Y1 ∪ Y2 ∪ … ∪ Yn.


We find

YoT
= (Y1 ∪ Y2 ∪ … ∪ Yn) o (T1 ∪ T2 ∪ … ∪ Tn)
= Y1 o T1 ∪ Y2 o T2 ∪ … ∪ Yn o Tn
= Z'1 ∪ Z'2 ∪ … ∪ Z'n
= Z'.

Now Z' is updated and thresholded so that we obtain Z and we


can find Z o T and so on.

Now we illustrate the situation by the following example.

Example 1.3.29: Let T = T1 ∪ T2 ∪ T3 ∪ T4 ∪ T5

⎡0 1 0 0 I⎤ ⎡0 1 0 0 0⎤
⎢1 0 0 1 0⎥⎥ ⎢1 0 0 0 1 ⎥⎥
⎢ ⎢
= ⎢0 0 0 I 1 ⎥ ∪ ⎢0 I 0 0 −1⎥ ∪
⎢ ⎥ ⎢ ⎥
⎢I 1 0 0 0⎥ ⎢0 0 I 0 0⎥
⎢⎣ 0 0 1 0 0⎦⎥ ⎢⎣ 0 0 1 0 0 ⎥⎦

⎡0 I 0 0 0⎤ ⎡0 1 0 0 1⎤ ⎡0 I 0 0 0⎤
⎢1 0 0 0 I ⎥ ⎢0 0 1 I ⎥
0⎥ ⎢1 0 1 0 0 ⎥⎥
⎢ ⎥ ⎢ ⎢
⎢ 0 1 0 I 0 ⎥ ∪ ⎢1 I 0 0 0⎥ ∪ ⎢0 0 0 1 −1⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢0 0 0 0 1 ⎥ ⎢0 1 I 0 0⎥ ⎢I 0 0 0 0⎥
⎢⎣1 I 0 0 0 ⎥⎦ ⎢⎣ 0 1 0 0 ⎥
0⎦ ⎢⎣ 0 0 0 1 0 ⎥⎦

be a special fuzzy neutrosophic square matrix. Each Ti is a 5 × 5


square fuzzy neutrosophic matrix i = 1, 2, …, 5.
Let

142
X = X1 ∪ X2 ∪ … ∪ X5
= [0 1 0 0 0] ∪ [0 0 0 I 0] ∪ [1 0 0 0 0] ∪ [0 0 0 0 1]
∪ [0 I 1 0 0]

be a special neutrosophic fuzzy row vector. To find the effect of


X on T, i.e. to find
X o T = (X1 ∪ X2 ∪ … ∪ X5) o (T1 ∪ T2 ∪ … ∪ T5)
= X1 o T1 ∪ X2 o T2 ∪ … ∪ X5 o T5

⎡0 1 0 0 I⎤
⎢1 0 0 1 0 ⎥⎥

= [ 0 1 0 0 0] o ⎢0 0 0 I 1⎥ ∪
⎢ ⎥
⎢I 1 0 0 0⎥
⎢⎣0 0 1 0 0 ⎥⎦

⎡0 1 0 0 0⎤
⎢1 0 0 0 1 ⎥⎥

[ 0 0 0 I 0] o ⎢ 0 I 0 0 −1⎥ ∪
⎢ ⎥
⎢0 0 I 0 0⎥
⎢⎣ 0 0 1 0 0 ⎥⎦

⎡0 I 0 0 0 ⎤
⎢1 0 0 0 I ⎥
⎢ ⎥
[1 0 0 0 0] o ⎢0 1 0 I 0⎥ ∪
⎢ ⎥
⎢0 0 0 0 1 ⎥
⎢⎣1 I 0 0 0 ⎥⎦

⎡0 1 0 0 1⎤
⎢0 0 1 I 0 ⎥⎥

[0 0 0 0 1] o ⎢1 I 0 0 0⎥ ∪
⎢ ⎥
⎢0 1 I 0 0⎥
⎢⎣0 1 0 0 0 ⎥⎦

143
⎡0 I 0 0 0⎤
⎢1 0 1 0 0 ⎥⎥

[ 0 I 1 0 0] o ⎢ 0 0 0 1 −1⎥
⎢ ⎥
⎢I 0 0 0 0⎥
⎢⎣ 0 0 0 1 0 ⎥⎦

= [1 0 0 1 0] ∪ [0 0 I 0 0] ∪ [0 I 0 0 0] ∪ [0 1 0 0 0]
∪ [I 0 I 1 –1]
= Y'1 ∪ Y'2 ∪ Y'3 ∪ Y'4 ∪ Y'5
= Y';

Y = Y1 ∪ Y2 ∪ Y3 ∪ Y4 ∪ Y5
= [1 1 0 1 0] ∪ [0 0 I I 0] ∪ [1 I 0 0 0] ∪ [0 1 0 0 1]
∪ [I I 1 1 0]

is the special neutrosophic fuzzy row vector obtained after


updating Y′. Now we find

YoT = (Y1 ∪ Y2 ∪ … ∪ Y5) o (T1 ∪ T2 ∪ … ∪ T5)


= Y1 o T1 ∪ Y2 o T2 ∪ … ∪ Y5 o T5

⎡0 1 0 0 I⎤
⎢1 0 0 1 0 ⎥⎥

= [1 1 0 1 0] o ⎢0 0 0 I 1⎥ ∪
⎢ ⎥
⎢I 1 0 0 0⎥
⎢⎣0 0 1 0 0 ⎥⎦

⎡0 1 0 0 0⎤
⎢1 0 0 0 1 ⎥⎥

[ 0 0 I I 0] o ⎢ 0 I 0 0 −1⎥ ∪
⎢ ⎥
⎢0 0 I 0 0⎥
⎢⎣ 0 0 1 0 0 ⎥⎦

144
⎡0 I 0 0 0 ⎤
⎢1 0 0 0 I ⎥
⎢ ⎥
[1 I 0 0 0] o ⎢0 1 0 I 0⎥ ∪
⎢ ⎥
⎢0 0 0 0 1 ⎥
⎢⎣1 I 0 0 0 ⎥⎦

⎡0 1 0 0 1⎤
⎢0 0 1 I 0 ⎥⎥

[0 I 0 0 1] o ⎢1 I 0 0 0⎥ ∪
⎢ ⎥
⎢0 1 I 0 0⎥
⎢⎣0 1 0 0 0 ⎥⎦

⎡0 I 0 0 0⎤
⎢1 0 1 0 0 ⎥⎥

[ I I 1 1 0] o ⎢ 0 0 0 1 −1⎥
⎢ ⎥
⎢I 0 0 0 0⎥
⎢⎣0 0 0 1 0 ⎥⎦

= [1+I 2 0 1 I] ∪ [0 I I 0 –I] ∪ [I I 0 0 I] ∪ [0 1 I I 0]
∪ [2I I I 1 –1]
= Z'1 ∪ Z'2 ∪ Z'3 ∪ … ∪ Z'5
= Z'.

Z = Z1 ∪ Z2 ∪ Z3 ∪ Z4 ∪ Z5
= [I 1 0 1 I] ∪ [0 I I I 0] ∪ [1 I 0 0 I] ∪ [0 1 1 I 1] ∪
[I I 1 1 0]

is the special fuzzy neutrosophic row vector obtained by


updating and thresholding Z′.

Now we can find Z o T and so on till we arrive at a fixed


point or a limit cycle.

145
Now we illustrate the same circle operator ‘o’ using special
fuzzy neutrosophic mixed square matrix. Let S = S1 ∪ S2 ∪ …
∪ Sn (n ≥ 2) be a special fuzzy neutrosophic mixed square
matrix; where Si is a ti × ti neutrosophic fuzzy matrix i = 1, 2,
…, n (ti ≠ tj, i ≠ j, 1 ≤ i, j ≤ n). Let X = X1 ∪ X2 ∪ … ∪ Xn be a
special fuzzy neutrosophic mixed row vector to find

XoS = (X1 ∪ X2 ∪ … ∪ Xn) o (S1 ∪ S2 ∪ … ∪ Sn)


= X1 o S1 ∪ X2 o S2 ∪ … ∪ Xn o Sn
= Y'1 ∪ Y'2 ∪ … ∪ Y'n
= Y'.

Now we update and threshold Y' to Y as Y' may not in general


be a special neutrosophic fuzzy mixed row vector. Now let Y =
Y1 ∪ Y2 ∪ … ∪ Yn be the special neutrosophic fuzzy mixed
row vector. We find Y o S = Z'1 ∪ Z'2 ∪ … ∪ Z'n = Z'. We
update and threshold Z' to obtain Z a special fuzzy neutrosophic
mixed row vector. We find Z o S = T', if T the thresholded and
updated special neutrosophic fuzzy mixed row vector of T' is a
fixed point or a limit cycle we stop the ‘o’ operation otherwise
we proceed on to find T o S and so on.

Now we illustrate this by the following example.

Example 1.3.30: Let S = S1 ∪ S2 ∪ S3 ∪ S4 ∪ S5 =

⎡0 1 0 0 I 0⎤
⎢1 0 ⎥⎥
⎡0 0 1 0⎤ ⎢ 0 I 0 0
⎢I ⎡ ⎤
0 ⎥⎥
0 1 I
⎢0 0⎥
∪ ⎢⎢ 0 0 1 ⎥⎥ ∪ ⎢
⎢ 0 0 0 0 1 I
⎥ ∪
⎢0 1 0 I⎥ ⎢0 0 0 0 1 I⎥
⎢ ⎥ ⎢⎣1 I 0 ⎥⎦
⎣1 1 I 0⎦ ⎢1 I 0 0 0 0⎥
⎢ ⎥
⎣⎢ 0 0 1 I 0 0 ⎦⎥

146
⎡0 0 0 1 1 0⎤
⎡0 0⎤ ⎢I 0 0 ⎥⎥
1 0 0
⎢I 0 1 0 0⎥ ⎥ ⎢ 0 0 1
⎢ ⎢0 0 0 I 0 0⎥
⎢0 I 0 1 0⎥ ∪ ⎢ ⎥
⎢ ⎥ ⎢1 0 0 0 I 1⎥
⎢0 0 I 0 1⎥
⎢0 0 0 1 0 I⎥
⎢⎣1 1 0 0 0 ⎥⎦ ⎢ ⎥
⎢⎣ 0 1 I 0 0 0 ⎥⎦

be the given special fuzzy neutrosophic mixed square matrix.


Suppose
X = X1 ∪ X2 ∪ … ∪ X5
= [1 0 0 0] ∪ [0 I 0] ∪ [1 0 0 0 0 I] ∪ [0 1 0 0 0] ∪
[1 I 0 0 0 0]

be the special fuzzy neutrosophic mixed row vector. Now we


find
X o S = [X1 ∪ X2 ∪ … ∪ X5] o [S1 ∪ S2 ∪ … ∪ S5]
= X1 o S1 ∪ X2 o S2 ∪ … ∪ X5 o S5

⎡0 0 1 0⎤
⎢I 0 0 0 ⎥⎥
= [1 0 0 0 ] ⎢⎢
o ∪
0 1 0 I⎥
⎢ ⎥
⎣1 1 I 0⎦

⎡0 1 I ⎤
[0 I 0] o ⎢⎢0 0 1 ⎥⎥ ∪
⎢⎣1 I 0 ⎥⎦

⎡0 1 0 0 I 0⎤
⎢1 0 I 0 0 0 ⎥⎥

⎢0 0 0 1 I 0⎥
[1 0 0 0 0 I] o ⎢ ⎥ ∪
⎢0 0 0 0 1 I⎥
⎢1 I 0 0 0 0⎥
⎢ ⎥
⎣⎢0 0 1 I 0 0 ⎦⎥

147
⎡0 1 0 0 0⎤
⎢I 0 1 0 0 ⎥⎥

[ 0 1 0 0 0] o ⎢ 0 I 0 1 0⎥ ∪
⎢ ⎥
⎢0 0 I 0 1⎥
⎢⎣1 1 0 0 0 ⎥⎦

⎡0 0 0 1 1 0⎤
⎢I 0 0 0 1 0 ⎥⎥

⎢0 0 0 I 0 0⎥
[1 I 0 0 0 0] o ⎢1 0 0 0 I 1⎥


⎢0 0 0 1 0 I⎥
⎢ ⎥
⎣⎢0 1 I 0 0 0 ⎦⎥

= [0 0 1 0] ∪ [0 0 I] ∪ [0 1 I I I 0] ∪ [I 0 1 0 0] ∪
[I 0 0 1 I+1 0]
= Y'1 ∪ Y'2 ∪ Y'3 ∪ Y'4 ∪ Y'5
= Y'.

After updating Y' to

Y = Y1 ∪ Y2 ∪ Y3 ∪ Y4 ∪ Y5
= [1 0 1 0] ∪ [0 I I] ∪ [1 1 I I I I] ∪ [I 1 1 0 0] ∪
[1 I 0 1 I 0]
which is again special neutrosophic fuzzy mixed row vector.

YoS = (Y1 ∪ Y2 ∪ … ∪ Y5) o (S1 ∪ S2 ∪ … ∪ S5)


= Y1 o S1 ∪ Y2 o S2 ∪ … ∪ Y5 o S5

⎡0 0 1 0⎤
⎢I 0 0 0 ⎥⎥
= [1 0 1 0] o ⎢⎢0 1 0 I⎥

⎢ ⎥
⎣1 1 I 0⎦

148
⎡0 1 I ⎤
[0 I I] o ⎢⎢0 0 1 ⎥⎥ ∪
⎢⎣1 I 0 ⎥⎦

⎡0 1 0 0 I 0⎤
⎢1 0 I 0 0 0 ⎥⎥

⎢0 0 0 1 I 0⎥
[1 1 I I I I] o ⎢0 0 0 0 1 I⎥
⎥ ∪

⎢1 I 0 0 0 0⎥
⎢ ⎥
⎢⎣0 0 1 I 0 0 ⎥⎦

⎡0 1 0 0 0⎤
⎢I 0 1 0 0 ⎥⎥

[ I 1 1 0 0] o ⎢ 0 I 0 1 0⎥ ∪
⎢ ⎥
⎢0 0 I 0 1⎥
⎢⎣1 1 0 0 0 ⎥⎦

⎡0 0 0 1 1 0⎤
⎢I 0 0 0 1 0 ⎥⎥

⎢0 0 0 I 0 0⎥
[1 I 0 1 I 0] o ⎢ ⎥
⎢1 0 0 0 I 1⎥
⎢0 0 0 1 0 I⎥
⎢ ⎥
⎣⎢0 1 0 0 0 0 ⎦⎥

= [0 1 1 I] ∪ [I I I] ∪ [1+I 1+I 2I 2I 3I I] ∪ [I 2I 1 1
0] ∪ [I+1 0 0 1 1+2I 1+I]
= Z'1 ∪ Z'2 ∪ Z'3 ∪ Z'4 ∪ Z'5
= Z'.

Now we update and threshold Z' to get

149
Z = Z1 ∪ Z2 ∪ Z3 ∪ Z4 ∪ Z5
= [1 1 I I] ∪ [I I I] ∪ [1 I I I I I] ∪ [I 1 1 1 0] ∪
[1 I 0 1 I I];

Z is the special neutrosophic fuzzy mixed row vector. Now

ZoS = (Z1 ∪ Z2 ∪ … ∪ Z5) o (S1 ∪ S2 ∪ … ∪ S5)


= Z1 o S1 ∪ Z2 o S2 ∪ … ∪ Z5 o S5

⎡0 0 1 0⎤
⎢I 0 0 0 ⎥⎥
= [1 1 I I] o ⎢⎢ ∪
0 1 0 I⎥
⎢ ⎥
⎣1 1 I 0⎦

⎡0 1 I ⎤
[ I I I] o ⎢⎢0 0 1 ⎥⎥ ∪
⎢⎣1 I 0 ⎥⎦

⎡0 1 0 0 I 0⎤
⎢1 0 I 0 0 0 ⎥⎥

⎢0 0 0 1 I 0⎥
[1 I I I I I] o ⎢ ⎥ ∪
⎢0 0 0 0 1 I⎥
⎢1 I 0 0 0 0⎥
⎢ ⎥
⎢⎣ 0 0 1 I 0 0 ⎥⎦

⎡0 1 0 0 0⎤
⎢I 0 1 0 0 ⎥⎥

[ I 1 1 1 0] o ⎢ 0 I 0 1 0⎥ ∪
⎢ ⎥
⎢0 0 I 0 1⎥
⎢⎣1 1 0 0 0 ⎥⎦

150
⎡0 0 0 1 1 0⎤
⎢I 0 0 0 1 0 ⎥⎥

⎢0 0 0 I 0 0⎥
[1 I 0 1 I I] o ⎢ ⎥
⎢1 0 0 0 I 1⎥
⎢0 0 0 1 0 I⎥
⎢ ⎥
⎢⎣0 1 I 0 0 0 ⎥⎦

= [2I 2I 1+I I] ∪ [I 2I 2I] ∪ [2I 1+I 2I 2I, 3I I] ∪ [I


2I 1+I 1 1] ∪ [I+1 I I 1+2I 1+2I 1+I]

= P'1 ∪ P'2 ∪ P'3 ∪ P'4 ∪ P'5


= P'.

Now P' is not a special fuzzy neutrosophic mixed row vector we


now update and threshold P' to P where

P = P1 ∪ P2 ∪ P3 ∪ P4 ∪ P5
= [1 I I I] ∪ [I I I] ∪ [1 I I I I I] ∪ [I 1 I 1 1] ∪
[1 I I I I I]

where P is a special fuzzy neutrosophic mixed row vector.

We can find P o S and continue till we arrive at a fixed point or


a limit cycle.

Now we proceed on to show how the operations are performed


in case of special fuzzy neutrosophic rectangular matrix.
Let

T = T1 ∪ T2 ∪ … ∪ Tm,
(m ≥ 2) be a special fuzzy neutrosophic rectangular matrix
where Ti is a s × t (s ≠ t) neutrosophic fuzzy rectangular matrix
for i = 1, 2, …, m.
Suppose

X = X1 ∪ X2 ∪ … ∪ Xm

151
be a special fuzzy neutrosophic row vector where each Xi is a 1
× s neutrosophic fuzzy row vector i = 1, 2, …, m.

Now we find

XoT = (X1 ∪ X2 ∪ … ∪ Xm) o (T1 ∪ T2 ∪ … ∪ Tm)


= X1 o T1 ∪ X2 o T2 ∪ … ∪ Xm o Tm
= Y'1 ∪ Y'2 ∪ … ∪ Y'm
= Y',

Y' may not be even a special fuzzy neutrosophic row vector we


update and threshold Y' to

Y = Y1 ∪ Y2 ∪ … ∪ Ym

where each Yi is a 1 × t fuzzy neutrosophic row vector for i = 1,


2, …, m.

Y o TT = (Y1 ∪ Y2 ∪ … ∪ Ym) o ( T1t ∪ T2t ∪ … ∪ Tmt )


= Y1 o T1t ∪ Y2 o T2t ∪ … ∪ Ym o Tmt
= Z'1 ∪ Z'2 ∪ … ∪ Z'm
= Z'.

Z' is updated and thresholded to get Z = Z1 ∪ Z2 ∪ … Zm. Now


using Z we find Z o T and so on until we arrive at a special limit
cycle or a special fixed point.

In case of special fuzzy neutrosophic rectangular matrices we


get the resultant as a pair of fuzzy neutrosophic row vector be it
a fixed point or a limit cycle which we shall call as special
binary pair.

Now we illustrate this by the following example.

Example 1.3.31: Let T = T1 ∪ T2 ∪ … ∪ T6 be a special


neutrosophic fuzzy rectangular matrix where each Ti is a 7 × 3
special neutrosophic fuzzy rectangular matrix; i = 1, 2, …, 6.

152
Suppose

X = X1 ∪ X2 ∪ … ∪ X6
= [0 1 0 0 0 0 1] ∪ [0 0 0 0 1 0 0] ∪ [0 0 0 0 0 0 I] ∪
[1 0 0 0 0 0 0] ∪ [0 0 0 1 I 0 0] ∪ [0 0 0 0 1 0 0].

be a 1 × 7 fuzzy neutrosophic row vector. Let

⎡0 1 0⎤ ⎡0 0 1 ⎤
⎢1 0 0⎥ ⎥ ⎢ I 0 0⎥
⎢ ⎢ ⎥
⎢0 1 1⎥ ⎢0 1 0⎥
⎢ ⎥ ⎢ ⎥
T = ⎢1 1 0 ⎥ ∪ ⎢1 0 0 ⎥ ∪
⎢0 I 0⎥ ⎢1 1 0 ⎥
⎢ ⎥ ⎢ ⎥
⎢0 0 1⎥ ⎢1 0 1 ⎥
⎢0 1 ⎥⎦ ⎢0 1 1 ⎥
⎣ 1 ⎣ ⎦

⎡I 0 0⎤ ⎡1 0 0 ⎤
⎢0 1 0⎥ ⎥ ⎢0 1 0⎥
⎢ ⎢ ⎥
⎢0 0 1⎥ ⎢0 0 1 ⎥
⎢ ⎥ ⎢ ⎥
⎢1 0 0 ⎥ ∪ ⎢1 1 0 ⎥ ∪
⎢1 0 1⎥ ⎢0 0 I ⎥
⎢ ⎥ ⎢ ⎥
⎢0 1 1⎥ ⎢1 0 1 ⎥
⎢1 0 ⎥⎦ ⎢0 1 1 ⎥
⎣ 1 ⎣ ⎦

⎡1 1 0⎤ ⎡1 0 1 ⎤
⎢0 1 1⎥ ⎥ ⎢0 I 0⎥
⎢ ⎢ ⎥
⎢0 0 1⎥ ⎢0 1 1 ⎥
⎢ ⎥ ⎢ ⎥
⎢0 1 0 ⎥ ∪ ⎢1 1 0 ⎥ .
⎢1 0 0⎥ ⎢1 0 0 ⎥
⎢ ⎥ ⎢ ⎥
⎢0 0 1⎥ ⎢0 0 1 ⎥
⎢0 I ⎥⎦ ⎢1 0 0 ⎥
⎣ 0 ⎣ ⎦

153
be the given special neutrosophic fuzzy rectangular matrix.

We find

XoT = (X1 ∪ X2 ∪ … ∪ X6) o (T1 ∪ T2 ∪ … ∪ T6)


= X1 o T1 ∪ X2 o T2 ∪ … ∪ X6 o T6

⎡0 1 0⎤
⎢1 0 0 ⎥⎥

⎢0 1 1⎥
= [0 1 0 0 0 0 1] o ⎢⎢1 1

0⎥ ∪
⎢0 I 0⎥
⎢ ⎥
⎢0 0 1⎥
⎢0 1 ⎥⎦
⎣ 1

⎡0 0 1⎤
⎢I 0 0 ⎥⎥

⎢0 1 0⎥
[0 0 0 0 1 0 0] o ⎢⎢1 0

0⎥ ∪
⎢1 1 0⎥
⎢ ⎥
⎢1 0 1⎥
⎢0 1 ⎥⎦
⎣ 1

⎡I 0 0⎤
⎢0 1 0 ⎥⎥

⎢0 0 1⎥
[0 0 0 0 0 0 I] o ⎢⎢1 0

0⎥ ∪
⎢1 0 1⎥
⎢ ⎥
⎢0 1 1⎥
⎢1 0 ⎥⎦
⎣ 1

154
⎡1 0 0⎤
⎢0 1 0 ⎥⎥

⎢0 0 1⎥
[1 0 0 0 0 0 0] o ⎢⎢1 1

0⎥ ∪
⎢0 0 I⎥
⎢ ⎥
⎢1 0 1⎥
⎢0 1 ⎥⎦
⎣ 1

⎡1 1 0⎤
⎢0 1 1 ⎥⎥

⎢0 0 1⎥
[0 0 0 1 I 0 0] o ⎢⎢0 1

0⎥ ∪
⎢1 0 0⎥
⎢ ⎥
⎢0 0 1⎥
⎢0 I ⎥⎦
⎣ 0

⎡1 0 1⎤
⎢0 I 0 ⎥⎥

⎢0 1 1⎥
[0 0 0 0 1 0 0] o ⎢⎢1 1 0⎥

⎢1 0 0⎥
⎢ ⎥
⎢0 0 1⎥
⎢1 0 ⎥⎦
⎣ 0

= [1 1 1] ∪ [1 1 0] ∪ [I I 0] ∪ [1 0 0] ∪ [I 1 0] ∪ [1 0 0]
= Y1 ∪ Y2 ∪ … ∪ Y6
= Y.

Now calculate
YoTt = (Y1 ∪ Y2 ∪ … ∪ Y6) o ( T1t ∪ T2t ∪ … ∪ T6t )
= Y1 o T1t ∪ Y2 o T2t ∪ … ∪ Y6 o T6t

155
⎡0 1 0 1 0 0 0⎤
= [1 1 1] o ⎢⎢1 0 1 1 I 0 1 ⎥⎥ ∪
⎢⎣ 0 0 1 0 0 1 1 ⎥⎦

⎡0 I 0 1 1 1 0 ⎤
[1 1 0] o ⎢⎢0 0 1 0 1 0 1 ⎥⎥ ∪
⎢⎣1 0 0 0 0 1 1 ⎥⎦

⎡ I 0 0 1 1 0 1⎤
[ I I 0] o ⎢⎢0 1 0 0 0 1 1 ⎥⎥ ∪
⎢⎣ 0 0 1 0 1 1 0 ⎥⎦

⎡1 0 0 1 0 1 0 ⎤
[1 0 0] o ⎢⎢0 1 0 1 0 0 1 ⎥⎥ ∪
⎢⎣0 0 1 0 I 1 1 ⎥⎦

⎡1 0 0 0 1 0 0 ⎤
[ I 1 0] o ⎢⎢1 1 0 1 0 0 0⎥⎥ ∪
⎢⎣0 1 1 0 0 1 I ⎥⎦

⎡1 0 0 1 1 0 1 ⎤
[1 0 0] o ⎢⎢0 I 1 1 0 0 0⎥⎥
⎢⎣1 0 1 0 0 1 0 ⎥⎦

= [1 1 2 2 I 1 2] ∪ [0 I 1 1 2 1 1] ∪ [I I 0 I I I 2I] ∪ [1
0 0 1 0 1 0] ∪ [1+I 1 0 1 I 0 0] ∪ [1 0 0 1 1 0 1]
= S'1 ∪ S'2 ∪ … ∪ S'6
= S',

we update and threshold S' to obtain S as follows:

S = [1 1 1 1 I 1 1] ∪ [0 I 1 1 1 1 1] ∪ [I I 0 I I I I] ∪ [1
0 0 1 0 1 0] ∪ [I 1 0 1 I 0 0] ∪ [1 0 0 1 1 0 1].

156
Now we can find

SoT = (S1 ∪ S2 ∪ … ∪ S6) o (T1 ∪ T2 ∪ … ∪ T6)


= S1 o T1 ∪ … ∪ S6 o T6

and so on. Thus we proceed on till we obtain a special pair of


fixed points or a special pair of limit cycle or a fixed point and a
limit cycle.

Now we illustrate how the operation works on special fuzzy


neutrosophic mixed rectangular matrix. Let W = W1 ∪ W2 ∪ …
∪ Wn (n ≥ 2) be the given special fuzzy neutrosophic mixed
rectangular matrix where Wi's are si × ti (si ≠ ti) neutrosophic
fuzzy rectangular matrix and Wj (i ≠ j) are pj × qj (pj ≠ qj and pj
≠ si or qj ≠ ti) rectangular neutrosophic fuzzy matrix, 1 ≤ i, j ≤ n.
Let X = X1 ∪ X2 ∪ … ∪ Xn where Xi are 1 × si fuzzy
neutrosophic row vector and Xj are 1 × pj fuzzy neutrosophic
row vectors be the special fuzzy neutrosophic mixed row vector
1 ≤ i, j ≤ n.
Now we calculate

XoW = (X1 ∪ X2 ∪ … ∪ Xn) o (W1 ∪ W2 ∪ … ∪ Wn)


= X1 o W1 ∪ X2 o W2 ∪ … ∪ Xn o Wn
= Y'1 ∪ Y'2 ∪ … ∪ Y'n
= Y'.

Y' may or may not be a special fuzzy neutrosophic mixed row


vector so we threshold Y' to Y and obtain the special fuzzy
neutrosophic mixed row vector i.e. Y = Y1 ∪ Y2 ∪ … ∪ Yn is
the special fuzzy neutrosophic mixed row vector. We now find

Y o WT = (Y1 ∪ Y2 ∪ … ∪ Yn) o ( W1T ∪ W2T ∪ … ∪ WnT )


= Y1 o W1T ∪ Y2 o W2T ∪ … ∪ Yn o WnT
= P'1 ∪ P'2 ∪ … ∪ P'n
= P';

157
P' may or may not be a special fuzzy neutrosophic mixed row
vector, we update and threshold P' to P and obtain P = P1 ∪ P2
∪ … ∪ Pn to be the special fuzzy neutrosophic mixed row
vector. Now using P we find

PoW = (P1 ∪ P2 ∪ … ∪ Pn) o (W1 ∪ W2 ∪ … ∪ Wn)


= P1 o W1 ∪ P2 o W2 ∪ … ∪ Pn o Wn
= R'1 ∪ R'2 ∪ … ∪ R'n
= R'.

R' may or may not be a special fuzzy neutrosophic mixed row


vector. We threshold R' to R and obtain the special fuzzy
neutrosophic mixed row vector. Now we can find R o WT and
so on until we arrive at a special pair of fixed point or a special
pair of limit cycle or a limit cycle and a fixed point.

We illustrate this situation by the following example.

Example 1.3.32: Let W = W1 ∪ W2 ∪ W3 ∪ W4 ∪ W5 =

⎡0 1 0⎤
⎢I 0 1 ⎥⎥
⎢ ⎡1 0 0 0 I 0 0 0⎤
⎢0 1⎥ ⎢0 0 ⎥⎥
0
⎢ ⎥ 1 0 0 0 1 0
⎢1 1 0⎥ ∪ ⎢ ∪
⎢0 0 1 0 0 0 1 0⎥
⎢0 I 0⎥ ⎢ ⎥
⎢ ⎥ ⎣1 0 0 I 1 0 1 1⎦
⎢1 1 1⎥
⎢1 1 ⎥⎦
⎣ 0

⎡1 0 0 0 −1 1 0⎤
⎢0 0 1 0 0 0 0 ⎥⎥

⎢0 0 0 I 0 0 0⎥
⎢ ⎥ ∪
⎢0 0 0 0 1 0 1⎥
⎢1 0 0 1 0 0 1⎥
⎢ ⎥
⎢⎣ 0 0 1 0 0 1 0 ⎥⎦

158
⎡0 I 0 1 0 0 1 0 1⎤
⎢1 0 0 0 I 0 0 0 1 ⎥⎥

⎢0 0 0 1 1 1 0 0 0⎥
⎢ ⎥
⎢1 0 1 0 0 0 0 1 0⎥
⎢⎣ 0 0 1 1 0 0 0 1 0 ⎥⎦

⎡0 1 0 0 1 0⎤
⎢0 0 I 0 0 1 ⎥⎥

∪ ⎢I 0 0 0 1 1⎥
⎢ ⎥
⎢0 0 0 1 0 0⎥
⎢⎣ 0 1 1 1 0 0 ⎥⎦

be a special fuzzy neutrosophic mixed rectangular matrix.


Let

X = X1 ∪ X2 ∪ … ∪ X5
= [0 0 0 0 0 0 1] ∪ [I 0 0 0] ∪ [0 1 0 0 1 0] ∪ [1 0 0 0
1] ∪ [0 0 1 0 0]

be a special fuzzy neutrosophic mixed row vector.


To find

XoW = [X1 ∪ X2 ∪ … ∪ X5] o [W1 ∪ W2 ∪ … ∪ W5]


= X1 o W1 ∪ X2 o W2 ∪ … ∪ X5 o W5

⎡0 1 0⎤
⎢I 0 1 ⎥⎥

⎢0 0 1⎥
= [0 0 0 0 0 0 1] o ⎢⎢1 1

0⎥ ∪
⎢0 I 0⎥
⎢ ⎥
⎢1 1 1⎥
⎢1 1 ⎥⎦
⎣ 0

159
⎡1 0 0 0 I 0 0 0⎤
⎢0 1 0 0 0 1 0 0 ⎥⎥
[ I 0 0 0] o ⎢⎢0 0 1 0 0 0 1 0⎥

⎢ ⎥
⎣1 0 0 I 1 0 1 1⎦

⎡1 0 0 0 −1 1 0⎤
⎢0 0 1 0 0 0 0 ⎥⎥

⎢0 0 0 I 0 0 0⎥
[ 0 1 0 0 1 0] o ⎢ ⎥ ∪
⎢0 0 0 0 1 0 1⎥
⎢1 0 0 1 0 0 1⎥
⎢ ⎥
⎢⎣0 0 1 0 0 1 0 ⎥⎦

⎡0 I 0 1 0 0 1 0 1⎤
⎢1 0 0 0 I 0 0 0 1 ⎥⎥

[1 0 0 0 1] o ⎢0 0 0 1 1 1 0 0 0⎥ ∪
⎢ ⎥
⎢1 0 1 0 0 0 0 1 0⎥
⎢⎣ 0 0 1 1 0 0 0 1 0 ⎥⎦

⎡0 1 0 0 1 0⎤
⎢0 0 I 0 0 1 ⎥⎥

[ 0 0 1 0 0] o ⎢ I 0 0 0 1 1⎥
⎢ ⎥
⎢0 0 0 1 0 0⎥
⎢⎣ 0 1 1 1 0 0 ⎥⎦

= [1 0 1] ∪ [I 0 0 0 I 0 0 0] ∪ [1 0 1 1 0 0 1] ∪ [0 I 1
2 0 0 1 1 1] ∪ [I 0 0 0 1 1]
= Y'1 ∪ Y'2 ∪ … ∪ Y'5
= Y'.

Now by thresholding Y' we get

160
Y = Y1 ∪ Y2 ∪ … ∪ Y5
= [1 0 1] ∪ [I 0 0 0 I 0 0 0] ∪ [1 0 1 1 0 0 1] ∪ [0 I 1
1 0 0 1 1 1] ∪ [I 0 0 0 1 1].

To find

Y o WT = [Y1 ∪ … ∪ Y5] o [ W1T ∪ W2T ∪ … ∪ W5T ]


= Y1 o W1T ∪ Y2 o W2T ∪ … ∪ Y5 o W5T

⎡0 I 0 1 0 1 1 ⎤
= [1 0 1] o ⎢⎢1 0 0 1 I 1 0⎥⎥ ∪
⎢⎣ 0 1 1 0 0 1 1 ⎥⎦

⎡1 0 0 1⎤
⎢0 1 0 0 ⎥⎥

⎢0 0 1 0⎥
⎢ ⎥
0 0 0 I⎥
[ I 0 0 0 I 0 0 0] o ⎢⎢ ∪
I 0 0 1⎥
⎢ ⎥
⎢0 1 0 0⎥
⎢0 0 1 1⎥
⎢ ⎥
⎣⎢ 0 0 0 1 ⎦⎥

⎡1 0 0 0 1 0⎤
⎢0 0 0 0 0 0 ⎥⎥

⎢0 1 0 0 0 1⎥
⎢ ⎥
[1 0 1 1 0 0 1] o ⎢0 0 I 0 1 0⎥ ∪
⎢ −1 0 0 1 0 0⎥
⎢ ⎥
⎢1 0 0 0 0 1⎥
⎢0 0 ⎥⎦
⎣ 0 0 1 1

161
⎡0 1 0 1 0⎤
⎢I 0 0 0 0 ⎥⎥

⎢0 0 0 1 1⎥
⎢ ⎥
⎢1 0 1 0 1⎥
[0 I 1 1 0 0 1 1 1] o ⎢⎢0 I 1 0 0⎥ ∪

⎢0 0 1 0 0⎥
⎢1 0 0 0 0⎥
⎢ ⎥
⎢0 0 0 1 1⎥
⎢1 0 ⎥⎦
⎣ 1 0 0

⎡0 0 I 0 0⎤
⎢1 0 0 0 1 ⎥⎥

⎢0 I 0 0 1⎥
[ I 0 0 0 1 1] o ⎢ ⎥
⎢0 0 0 1 1⎥
⎢1 0 1 0 0⎥
⎢ ⎥
⎣⎢ 0 1 1 0 0 ⎦⎥

= [0 1+I 1 1 0 2 2] ∪ [2I 0 0 2I] ∪ [1 1 I 1 3 1] ∪ [3+I


1 1 2 3] ∪ [1 1 2+I 0 0]
= R'1 ∪ R'2 ∪ … ∪ R'5
= R';

we now update and threshold this R′ to obtain R as R′ is not


a special fuzzy neutrosophic mixed row vector.

R = [0 I 1 1 0 1 1] ∪ [I 0 0 I] ∪ [1 1 I 1 1 1] ∪ [1 1 1 1
1] ∪ [1 1 1 0 0].

We now proceed on to find

RoW = (R1 ∪ R2 ∪ … ∪ R5) o (W1 ∪ W2 ∪ … ∪ W5)


= R1 o W1 ∪ R2 o W2 ∪ … ∪ R5 o W5

162
⎡0 1 0⎤
⎢I 0 1 ⎥⎥

⎢0 0 1⎥
= [0 I 1 1 0 1 1] o ⎢⎢1 1

0⎥ ∪
⎢0 I 0⎥
⎢ ⎥
⎢1 1 1⎥
⎢1 1 ⎥⎦
⎣ 0

⎡1 0 0 0 I 0 0 0⎤
⎢0 1 0 0 0 1 0 0 ⎥⎥
[ I 0 0 I] o ⎢⎢0 0 1 0 0 0 1 0⎥

⎢ ⎥
⎣1 0 0 I 1 0 1 1⎦

⎡1 0 0 0 −1 1 0⎤
⎢0 0 1 0 0 0 0 ⎥⎥

⎢0 0 0 I 0 0 0⎥
[1 1 I 1 1 1] o ⎢ ⎥ ∪
⎢0 0 0 0 1 0 1⎥
⎢1 0 0 1 0 0 1⎥
⎢ ⎥
⎣⎢0 0 1 0 0 1 0 ⎦⎥

⎡0 I 0 1 0 0 1 0 1⎤
⎢1 0 0 0 I 0 0 0 1 ⎥⎥

[1 1 1 1 1] o ⎢0 0 0 1 1 1 0 0 0⎥ ∪
⎢ ⎥
⎢1 0 1 0 0 0 0 1 0⎥
⎢⎣0 0 1 1 0 0 0 1 0 ⎥⎦

⎡0 1 0 0 1 0⎤
⎢0 0 I 0 0 1 ⎥⎥

[1 1 1 0 0] o ⎢ I 0 0 0 1 1⎥
⎢ ⎥
⎢0 0 0 1 0 0⎥
⎢⎣ 0 1 1 1 0 0 ⎥⎦

163
= [3+I 2 3+I] ∪ [2I 0 0 I 2I 0 I I] ∪ [2 0 2 1+I 0 2 2]
∪ [2 I 2 3 1+I 1 1 2 2] ∪ [I 1 I 0 2 2]
= S'1 ∪ S'2 ∪ … ∪ S'5
= S'.

We threshold S′ and obtain

= [1 1 1] ∪ [I 0 0 I I 0 I I] ∪ [1 0 1 I 1 1 1] ∪ [1 I 1 1
I 1 1 1 1] ∪ [I 1 I 0 1 1]
= S1 ∪ S2 ∪ S3 ∪ S4 ∪ S5
= S.

We can calculate S o WT and so on until we find the special


fixed point and a special limit cycle or a special pair of fixed
point or a special pair of limit cycle.
Now we use the same operation on M and find the value of
the resultant vector where M is a special fuzzy neutrosophic
mixed matrix. Let M = M1 ∪ M2 ∪ … ∪ Ms (s ≥ 2) where Mi
are ni × ni square fuzzy neutrosophic matrix and mj’s are tj × pj
(tj ≠ pj) fuzzy neutrosophic rectangular matrices 1 ≤ i, j ≤ s.
Suppose X = X1 ∪ X2 ∪ … ∪ Xs where Xi’s are 1× ni
neutrosophic fuzzy row vector and Xj’s are 1 × tj neutrosophic
fuzzy row vector 1 ≤ i, j ≤ s. To find

XoM = (X1 ∪ X2 ∪ … ∪ Xs) o (M1 ∪ M2 ∪ … ∪ Ms)


= X1 o M1 ∪ X2 o M2 ∪ … ∪ Xs o Ms.
= Y'1 ∪ Y'2 ∪ … ∪ Y's
= Y';

Y' may or may not be a special neutrosophic fuzzy mixed row


vector. We threshold Y' to Y = Y1 ∪ Y2 ∪ … ∪ Ys to become a
special fuzzy neutrosophic mixed row vector. Now each Yi is a
1 × ni fuzzy neutrosophic row vector are Yj’s are 1 × pj fuzzy
neutrosophic row vector. Thus we find now Y o MST where MST
is the special transpose of M. Let

Y o MST = Z'1 ∪ Z'2 ∪ … ∪ Z's

164
= Z',

we threshold and update Z' to Z = Z1 ∪ Z2 ∪ … ∪ Zs and find Z


o M and so on until we arrive at a special fixed binary pair or a
special pair of limit cycle or a fixed point and a limit cycle.

We illustrate this by the following example.

Example 1.3.33: Let M = M1 ∪ M2 ∪ M3 ∪ M4 ∪ M5 =

⎡0 1 0 0 0⎤
⎢0 ⎡0 1 0 0 0 1 0 0 0⎤
0 ⎥⎥ ⎢I 0 1 0 0 0 0 1 1 ⎥⎥
⎢ 0 I 1
⎢1 I 0 1 0⎥ ∪ ⎢ ∪
⎢ ⎥ ⎢0 0 0 0 1 0 I 0 1⎥
⎢0 0 0 1 I⎥ ⎢ ⎥
⎢⎣ I ⎣1 I 0 I 0 0 0 0 1⎦
0 1 0 1 ⎥⎦

⎡0 1 0 0 0⎤
⎢I 1 ⎥⎥
⎢ 0 0 0 ⎡0 1 0⎤
⎡0 0⎤ ⎢1 0 ⎥⎥
1 0
⎢0 0⎥ ⎢I 0 ⎥⎥
0 1 I ⎢ 0
⎢ ⎥ 0 0
⎢1 0 I 1 0⎥ ∪ ⎢ ∪ ⎢0 0 I⎥
⎢0 0 1 0⎥ ⎢ ⎥
⎢0 I 0 0 I⎥ ⎢ ⎥
⎢ ⎥ ⎢1 1 0⎥
⎣0 0 0 1⎦
⎢⎣ 0
⎢1 1 0 1 0⎥ 0 1 ⎥⎦
⎢0 1 ⎥⎦
⎣ 1 1 0

be the given special fuzzy neutrosophic mixed matrix.


Suppose

X = [1 0 0 0 0] ∪ [0 0 0 1] ∪ [0 0 I 0 0 1 0] ∪ [1 0 0 1]
∪ [0 1 0 0 0]

be the special fuzzy neutrosophic mixed row vector. To find

XoM = (X1 ∪ X2 ∪ … ∪ X5) o (M1 ∪ M2 ∪ … ∪ M5)


= X1 o M1 ∪ X2 o M2 ∪ … ∪ X5 o M5

165
⎡0 1 0 0 0⎤
⎢0 0 I 1 0 ⎥⎥

= [1 0 0 0 0] o ⎢1 I 0 1 0⎥ ∪
⎢ ⎥
⎢0 0 0 1 I⎥
⎢⎣ I 0 1 0 1 ⎥⎦

⎡0 1 0 0 0 1 0 0 0⎤
⎢I 0 1 0 0 0 0 1 1 ⎥⎥
[0 0 0 1] o ⎢⎢0 0 0 0 1 0 I 0 1⎥

⎢ ⎥
⎣1 I 0 I 0 0 0 0 1⎦

⎡0 1 0 0 0⎤
⎢I 0 0 0 1 ⎥⎥

⎢0 0 1 I 0⎥
⎢ ⎥
[0 0 I 0 0 1 0] o ⎢1 0 I 1 0⎥ ∪
⎢0 I 0 0 I⎥
⎢ ⎥
⎢1 1 0 1 0⎥
⎢0 1 ⎥⎦
⎣ 1 1 0

⎡0 1 0 0⎤
⎢I 0 0 0 ⎥⎥
[1 0 0 1] o ⎢⎢0 0 1 0⎥

⎢ ⎥
⎣0 0 0 1⎦

⎡0 1 0⎤
⎢1 0 0 ⎥⎥

[ 0 1 0 0 0] o ⎢ 0 0 I⎥
⎢ ⎥
⎢1 1 0⎥
⎢⎣ 0 0 1 ⎥⎦

166
= [0 1 0 0 0] ∪ [1 I 0 I 0 0 0 0 1] ∪ [1 1 I 1+I 0] ∪
[0 1 0 1] ∪ [1 0 0]
= Z'1 ∪ Z'2 ∪ Z'3 ∪ Z'4 ∪ Z'5
= Z'.

We update and threshold Z' to Z and


Z = Z1 ∪ Z2 ∪ Z3 ∪ Z4 ∪ Z5
= [1 1 0 0 0] ∪ [1 I 0 I 0 0 0 0 1] ∪ [1 1 I I 0] ∪ [1 1
0 1] ∪ [1 0 0].

Clearly Z is again a special fuzzy neutrosophic mixed row


vector. Now we find

Z o MST = [Z1 ∪ Z2 ∪ … ∪ Z5] o [M1 ∪ M ST


2 ∪ M3
ST
∪ M4 ∪
M ST
5 ]

= Z1 o M1 ∪ Z2 o M 2t ∪ Z3 o M 3t ∪ Z4 o M4 ∪ Z5 o
M 5t
⎡0 1 0 0 0⎤
⎢0 0 I 1 0 ⎥⎥

= [1 1 0 0 0] o ⎢1 I 0 1 0⎥ ∪
⎢ ⎥
⎢0 0 0 1 I⎥
⎢⎣ I 0 1 0 1 ⎥⎦

⎡0 I 0 1⎤
⎢1 0 0 I ⎥⎥

⎢0 1 0 0⎥
⎢ ⎥
⎢0 0 0 I⎥
[1 I 0 I 0 0 0 0 1] o ⎢⎢0 0 1 0⎥ ∪

⎢1 0 0 0⎥
⎢0 0 I 0⎥
⎢ ⎥
⎢0 1 0 0⎥
⎢0 1 ⎥⎦
⎣ 1 1

167
⎡0 I 0 1 0 1 0⎤
⎢1 0 0 0 I 1 1 ⎥⎥

[1 1 I I 0] o ⎢0 0 1 I 0 0 1⎥ ∪
⎢ ⎥
⎢0 0 I 1 0 1 0⎥
⎢⎣ 0 1 0 0 I 0 1 ⎥⎦

⎡0 1 0 0⎤
⎢I 0 0 0 ⎥⎥
[1 1 0 1] o ⎢⎢0 0 1 0⎥

⎢ ⎥
⎣0 0 0 1⎦

⎡0 1 0 1 0 ⎤
[1 0 0] o ⎢⎢1 0 0 1 0⎥⎥
⎣⎢0 0 I 0 1 ⎥⎦

= [0 1 I 1 0] ∪ [I I+1 1 2 + 2I] ∪ [1 I 2I 1+2I I 2+I


1+I] ∪ [I 1 0 1] ∪ [0 1 0 1 0]
= P'1 ∪ P'2 ∪ P'3 ∪ P'4 ∪ P'5
= P';

we see P' is not even a special fuzzy neutrosophic mixed row


vector so we update and threshold P' to

P = P1 ∪ P2 ∪ … ∪ P5
= [1 1 I 1 0] ∪ [I I 1 1] ∪ [1 I I I I 1 I] ∪ [1 1 0 1] ∪
[1 1 0 1 0].

P is clearly a special fuzzy neutrosophic mixed row vector.

Now we calculate

PoM = (P1 ∪ P2 ∪ … ∪ P5) o (M1 ∪ M2 ∪ … ∪ M5)


= P1 o M1 ∪ P2 o M2 ∪ … ∪ P5 o M5

168
⎡0 1 0 0 0⎤
⎢0 0 I 1 0 ⎥⎥

= [1 1 I 1 0] o ⎢1 I 0 1 0⎥ ∪
⎢ ⎥
⎢0 0 0 1 I⎥
⎢⎣ I 0 1 0 1 ⎥⎦

⎡0 1 0 0 0 1 0 0 0⎤
⎢I 0 1 0 0 0 0 1 1 ⎥⎥
[ I I 1 1] o ⎢⎢0 0 0 0 1 0 I 0 1⎥

⎢ ⎥
⎣1 I 0 I 0 0 0 0 1⎦

⎡0 1 0 0 0⎤
⎢I 0 0 0 1 ⎥⎥

⎢0 0 1 I 0⎥
[1 I I I I 1 I] o ⎢⎢1 0 I 1

0⎥ ∪
⎢0 I 0 0 I⎥
⎢ ⎥
⎢1 1 0 1 0⎥
⎢0 1 ⎥⎦
⎣ 1 1 0

⎡0 1 0 0⎤
⎢I 0 0 0 ⎥⎥
[1 1 0 1] o ⎢⎢0 0 1 0⎥

⎢ ⎥
⎣0 0 0 1⎦

⎡0 1 0⎤
⎢1 0 0 ⎥⎥

[1 1 0 1 0] o ⎢0 0 I⎥
⎢ ⎥
⎢1 1 0⎥
⎢⎣ 0 0 1 ⎥⎦

169
= [I 1+I I 2+I I] ∪ [1+I 2I I I 1 I I I 2 + I] ∪ [2I+1
2+2I 3I 2I+1 3I] ∪ [I 1 0 1] ∪ [2 2 0]

After updating and thresholding we get

T = [1 I I I I] ∪ [I I I I 1 I I I I] ∪ [I I I I I] ∪ [1 1 0 1] ∪
[1 1 0]
= T1 ∪ T2 ∪ T3 ∪ T4 ∪ T5;

T is a special fuzzy neutrosophic mixed row vector. We


now calculate T o MST and so on.

170
Chapter Two

SPECIAL FUZZY MODELS AND


SPECIAL NEUTROSOPHIC MODELS
AND THEIR GENERALIZATIONS

In this chapter for the first time authors introduce seventeen new
types of special fuzzy models and special neutrosophic models.
This chapter has six sections. In the first section we just
recall the notion of three fuzzy models; Fuzzy Cognitive Maps
(FCMs), Fuzzy Relational Maps (FRMs) and Fuzzy Relational
Equations (FREs). For more information refer [232, 239].
Second section introduces the basic notion of neutrosophy
from [187-190] and three neutrosophic models analogous to
FCMs, FRMs and FREs namely NCMs, NRMs and NREs.
Please refer [231-2]. The third section introduces five types of
special fuzzy cognitive models and special Neutrosophic
cognitive models.
The forth section gives yet another eight new special fuzzy
and neutrosophic models which are multi expert models. Yet
another set of five new special fuzzy and neutrosophic models
are introduced in section five using FRE and NRE. A special
super model using all these six fuzzy and neutrosophic multi
expert model is also introduced in this section. The final section

171
proposes some simple programming problems for these new
special models.

2.1 Basic Description of Fuzzy Models

This section has three subsections. In the first subsection the


fuzzy cognitive maps model is described. In the second
subsection the notion of fuzzy relational maps which are a
particular generalization of FCMs that too when the number of
attributes are very large and can be divided into two classes are
recalled for more about these concepts refer [108, 112]. In the
final subsection we briefly recall the definition of FRE. Several
types of FRE are introduced in [106, 232].

2.1.1 Definition of Fuzzy Cognitive Maps

In this section we recall the notion of Fuzzy Cognitive Maps


(FCMs), which was introduced by Bart Kosko [108, 112] in the
year 1986. We also give several of its interrelated definitions.
FCMs have a major role to play mainly when the data
concerned is an unsupervised one. Further this method is most
simple and an effective one as it can analyse the data by
directed graphs and connection matrices.

DEFINITION 2.1.1.1: An FCM is a directed graph with concepts


like policies, events etc. as nodes and causalities as edges. It
represents causal relationship between concepts.

Example 2.1.1.1: In Tamil Nadu (a southern state in India) in


the last decade several new engineering colleges have been
approved and started. The resultant increase in the production of
engineering graduates in these years is disproportionate with the
need of engineering graduates. This has resulted in thousands of
unemployed and underemployed graduate engineers. Using an
expert's opinion we study the effect of such unemployed people
on the society. An expert spells out the five major concepts
relating to the unemployed graduated engineers as

172
E1 – Frustration
E2 – Unemployment
E3 – Increase of educated criminals
E4 – Under employment
E5 – Taking up drugs etc.

The directed graph where E1, …, E5 are taken as the nodes and
causalities as edges as given by an expert is given in the
following Figure 2.1.1.1:

E1 E2

E4 E3

E5

FIGURE: 2.1.1.1

According to this expert, increase in unemployment increases


frustration. Increase in unemployment, increases the educated
criminals. Frustration increases the graduates to take up to evils
like drugs etc. Unemployment also leads to the increase in
number of persons who take up to drugs, drinks etc. to forget
their worries and unoccupied time. Under-employment forces
then to do criminal acts like theft (leading to murder) for want
of more money and so on. Thus one cannot actually get data for
this but can use the expert's opinion for this unsupervised data
to obtain some idea about the real plight of the situation. This is
just an illustration to show how FCM is described by a directed
graph.

173
{If increase (or decrease) in one concept leads to increase (or
decrease) in another, then we give the value 1. If there exists no
relation between two concepts the value 0 is given. If increase
(or decrease) in one concept decreases (or increases) another,
then we give the value –1. Thus FCMs are described in this
way.}

DEFINITION 2.1.1.2: When the nodes of the FCM are fuzzy sets
then they are called as fuzzy nodes.

DEFINITION 2.1.1.3: FCMs with edge weights or causalities


from the set {–1, 0, 1} are called simple FCMs.

DEFINITION 2.1.1.4: Consider the nodes / concepts C1, …, Cn of


the FCM. Suppose the directed graph is drawn using edge
weight eij ∈ {0, 1, –1}. The matrix E be defined by E = (eij)
where eij is the weight of the directed edge Ci Cj . E is called the
adjacency matrix of the FCM, also known as the connection
matrix of the FCM.

It is important to note that all matrices associated with an FCM


are always square matrices with diagonal entries as zero.

DEFINITION 2.1.1.5: Let C1, C2, … , Cn be the nodes of an FCM.


A = (a1, a2, … , an) where ai ∈ {0, 1}. A is called the
instantaneous state vector and it denotes the on-off position of
the node at an instant;

ai = 0 if ai is off and
ai = 1 if ai is on
for i = 1, 2, …, n.

DEFINITION 2.1.1.6: Let C1, C2, … , Cn be the nodes of an


FCM. Let C1C 2 , C2C3 , C3C4 , … , Ci C j be the edges of the
FCM (i ≠ j). Then the edges form a directed cycle. An FCM is
said to be cyclic if it possesses a directed cycle. An FCM is said
to be acyclic if it does not possess any directed cycle.

174
DEFINITION 2.1.1.7: An FCM with cycles is said to have a
feedback.

DEFINITION 2.1.1.8: When there is a feedback in an FCM, i.e.,


when the causal relations flow through a cycle in a
revolutionary way, the FCM is called a dynamical system.

DEFINITION 2.1.1.9: Let C1C2 , C2C3 , … ,Cn−1Cn be a


cycle. When Ci is switched on and if the causality flows through
the edges of a cycle and if it again causes Ci , we say that the
dynamical system goes round and round. This is true for any
node Ci , for i = 1, 2, … , n. The equilibrium state for this
dynamical system is called the hidden pattern.

DEFINITION 2.1.1.10: If the equilibrium state of a dynamical


system is a unique state vector, then it is called a fixed point.

Example 2.1.1.2: Consider a FCM with C1, C2, …, Cn as nodes.


For example let us start the dynamical system by switching on
C1. Let us assume that the FCM settles down with C1 and Cn on
i.e. the state vector remains as (1, 0, 0, …, 0, 1) this state vector
(1, 0, 0, …, 0, 1) is called the fixed point. (It is denoted also as
[1 0 0 0 … 0 1] as the resultant is a row matrix).

DEFINITION 2.1.1.11: If the FCM settles down with a state


vector repeating in the form

A1 → A2 → … → Ai → A1

then this equilibrium is called a limit cycle.

DEFINITION 2.1.1.12: Finite number of FCMs can be combined


together to produce the joint effect of all the FCMs. Let E1, E2,
… , Ep be the adjacency matrices of the FCMs with nodes C1,
C2, …, Cn then the combined FCM is got by adding all the
adjacency matrices E1, E2, …, Ep .
We denote the combined FCM adjacency matrix by E = E1
+ E2 + …+ Ep .

175
NOTATION: Suppose A = (a1, … , an) is a vector which is
passed into a dynamical system E. Then AE = (a'1, … , a'n) after
thresholding and updating the vector suppose we get (b1, … , bn)
we denote that by
(a'1, a'2, … , a'n) → (b1, b2, … , bn).

Thus the symbol '→' means the resultant vector has been
thresholded and updated.
FCMs have several advantages as well as some
disadvantages. The main advantage of this method it is simple.
It functions on expert's opinion. When the data happens to be an
unsupervised one the FCM comes handy. This is the only
known fuzzy technique that gives the hidden pattern of the
situation. As we have a very well known theory, which states
that the strength of the data depends on, the number of experts'
opinion we can use combined FCMs with several experts'
opinions.
At the same time the disadvantage of the combined FCM is
when the weightages are 1 and –1 for the same Ci Cj, we have
the sum adding to zero thus at all times the connection matrices
E1, … , Ek may not be conformable for addition.
Combined conflicting opinions tend to cancel out and
assisted by the strong law of large numbers, a consensus
emerges as the sample opinion approximates the underlying
population opinion. This problem will be easily overcome if the
FCM entries are only 0 and 1.
We have just briefly recalled the definitions. For more about
FCMs please refer Kosko [108, 112].

2.1.2 Definition and Illustration of Fuzzy Relational Maps


(FRMs)

In this section, we introduce the notion of Fuzzy Relational


Maps (FRMs); they are constructed analogous to FCMs
described and discussed in the earlier sections. In FCMs we
promote the correlations between causal associations among
concurrently active units. But in FRMs we divide the very
causal associations into two disjoint units, for example, the

176
relation between a teacher and a student or relation between an
employee or employer or a relation between doctor and patient
and so on. Thus for us to define a FRM we need a domain space
and a range space which are disjoint in the sense of concepts.
We further assume no intermediate relation exists within the
domain elements or node and the range spaces elements. The
number of elements in the range space need not in general be
equal to the number of elements in the domain space.
Thus throughout this section we assume the elements of the
domain space are taken from the real vector space of dimension
n and that of the range space are real vectors from the vector
space of dimension m (m in general need not be equal to n). We
denote by R the set of nodes R1,…, Rm of the range space,
where R = {(x1,…, xm) ⏐xj = 0 or 1 } for j = 1, 2,…, m. If xi = 1
it means that the node Ri is in the on state and if xi = 0 it means
that the node Ri is in the off state. Similarly D denotes the nodes
D1, D2,…, Dn of the domain space where D = {(x1,…, xn) ⏐ xj =
0 or 1} for i = 1, 2,…, n. If xi = 1 it means that the node Di is in
the on state and if xi = 0 it means that the node Di is in the off
state.

Now we proceed on to define a FRM.

DEFINITION 2.1.2.1: A FRM is a directed graph or a map from


D to R with concepts like policies or events etc, as nodes and
causalities as edges. It represents causal relations between
spaces D and R .

Let Di and Rj denote that the two nodes of an FRM. The directed
edge from Di to Rj denotes the causality of Di on Rj called
relations. Every edge in the FRM is weighted with a number in
the set {0, ±1}. Let eij be the weight of the edge DiRj, eij ∈ {0,
±1}. The weight of the edge Di Rj is positive if increase in Di
implies increase in Rj or decrease in Di implies decrease in Rj ie
causality of Di on Rj is 1. If eij = 0, then Di does not have any
effect on Rj . We do not discuss the cases when increase in Di
implies decrease in Rj or decrease in Di implies increase in Rj .

177
DEFINITION 2.1.2.2: When the nodes of the FRM are fuzzy sets
then they are called fuzzy nodes. FRMs with edge weights {0,
±1} are called simple FRMs.

DEFINITION 2.1.2.3: Let D1, …, Dn be the nodes of the domain


space D of an FRM and R1, …, Rm be the nodes of the range
space R of an FRM. Let the matrix E be defined as E = (eij)
where eij is the weight of the directed edge DiRj (or RjDi), E is
called the relational matrix of the FRM.

Note: It is pertinent to mention here that unlike the FCMs the


FRMs can be a rectangular matrix with rows corresponding to
the domain space and columns corresponding to the range
space. This is one of the marked difference between FRMs and
FCMs. For more about FRMs refer [241, 250]

DEFINITION 2.1.2.4: Let D1, ..., Dn and R1,…, Rm denote the


nodes of the FRM. Let A = (a1,…,an), ai ∈ {0, 1}. A is called the
instantaneous state vector of the domain space and it denotes
the on-off position of the nodes at any instant. Similarly let B =
(b1,…, bm) bi ∈ {0, 1}. B is called instantaneous state vector of
the range space and it denotes the on-off position of the nodes
at any instant ai = 0 if ai is off and ai = 1 if ai is on for i= 1,
2,…, n Similarly, bi = 0 if bi is off and bi = 1 if bi is on, for i= 1,
2,…, m.

DEFINITION 2.1.2.5: Let D1, …, Dn and R1,…, Rm be the nodes


of an FRM. Let DiRj (or Rj Di) be the edges of an FRM, j = 1,
2,…, m and i= 1, 2,…, n. Let the edges form a directed cycle. An
FRM is said to be a cycle if it posses a directed cycle. An FRM
is said to be acyclic if it does not posses any directed cycle.

DEFINITION 2.1.2.6: An FRM with cycles is said to be an FRM


with feedback.

DEFINITION 2.1.2.7: When there is a feedback in the FRM, i.e.


when the causal relations flow through a cycle in a
revolutionary manner, the FRM is called a dynamical system.

178
DEFINITION 2.1.2.8: Let Di Rj (or Rj Di), 1 ≤ j ≤ m, 1 ≤ i ≤ n.
When Ri (or Dj) is switched on and if causality flows through
edges of the cycle and if it again causes Ri (orDj), we say that
the dynamical system goes round and round. This is true for any
node Rj (or Di) for 1 ≤ i ≤ n, (or 1 ≤ j ≤ m). The equilibrium
state of this dynamical system is called the hidden pattern.

DEFINITION 2.1.2.9: If the equilibrium state of a dynamical


system is a unique state vector, then it is called a fixed point.
Consider an FRM with R1, R2,…, Rm and D1, D2,…, Dn as nodes.
For example, let us start the dynamical system by switching on
R1 (or D1). Let us assume that the FRM settles down with R1 and
Rm (or D1 and Dn) on, i.e. the state vector remains as (1, 0, …,
0, 1) in R (or 1, 0, 0, … , 0, 1) in D), This state vector is called
the fixed point.

DEFINITION 2.1.2.10: If the FRM settles down with a state


vector repeating in the form

A1 → A2 → A3 → … → Ai → A1 (or B1 → B2 → …→ Bi → B1)

then this equilibrium is called a limit cycle.

METHODS OF DETERMINING THE HIDDEN PATTERN

Let R1, R2,…, Rm and D1, D2,…, Dn be the nodes of a FRM with
feedback. Let E be the relational matrix. Let us find a hidden
pattern when D1 is switched on i.e. when an input is given as
vector A1 = (1, 0, …, 0) in D1, the data should pass through the
relational matrix E. This is done by multiplying A1 with the
relational matrix E. Let A1E = (r1, r2,…, rm), after thresholding
and updating the resultant vector we get A1 E ∈ R. Now let B =
A1E we pass on B into ET and obtain BET. We update and
threshold the vector BET so that BET ∈ D. This procedure is
repeated till we get a limit cycle or a fixed point.

DEFINITION 2.1.2.11: Finite number of FRMs can be combined


together to produce the joint effect of all the FRMs. Let E1,…,
Ep be the relational matrices of the FRMs with nodes R1, R2,…,

179
Rm and D1, D2,…, Dn, then the combined FRM is represented by
the relational matrix E = E1+…+ Ep.

2.1.3 Properties of Fuzzy Relations and FREs

In this section we just recollect the properties of fuzzy relations


like, fuzzy equivalence relation, fuzzy compatibility relations,
fuzzy ordering relations, fuzzy morphisms and sup-i-
compositions of fuzzy relation. For more about these concepts
please refer [231, 240].

Now we proceed on to define fuzzy equivalence relation. A


crisp binary relation R(X, X) that is reflexive, symmetric and
transitive is called an equivalence relation. For each element x
in X, we can define a crisp set Ax, which contains all the
elements of X that are related to x, by the equivalence relation.

Ax = {y ⏐(x, y) ∈ R (X, X)}

Ax is clearly a subset of X. The element x is itself contained in


Ax due to the reflexivity of R, because R is transitive and
symmetric each member of Ax, is related to all the other
members of Ax. Further no member of Ax, is related to any
element of X not included in Ax. This set Ax is referred to an as
equivalence class of R (X, X) with respect to x. The members of
each equivalence class can be considered equivalent to each
other and only to each other under the relation R. The family of
all such equivalence classes defined by the relation which is
usually denoted by X / R, forms a partition on X.
A fuzzy binary relation that is reflexive, symmetric and
transitive is known as a fuzzy equivalence relation or similarity
relation. In the rest of this section let us use the latter term.
While the max-min form of transitivity is assumed, in the
following discussion on concepts; can be generalized to the
alternative definition of fuzzy transitivity.
While an equivalence relation clearly groups elements that
are equivalent under the relation into disjoint classes, the
interpretation of a similarity relation can be approached in two

180
different ways. First it can be considered to effectively group
elements into crisp sets whose members are similar to each
other to some specified degree. Obviously when this degree is
equal to 1, the grouping is an equivalence class. Alternatively
however we may wish to consider the degree of similarity that
the elements of X have to some specified element x ∈ X. Thus
for each x ∈ X, a similarity class can be defined as a fuzzy set
in which the membership grade of any particular element
represents the similarity of that element to the element x. If all
the elements in the class are similar to x to the degree of 1 and
similar to all elements outside the set to the degree of 0 then the
grouping again becomes an equivalence class. We know every
fuzzy relation R can be uniquely represented in terms of its α-
cuts by the formula

R = ∪α. α R .
α∈(0,1]

It is easily verified that if R is a similarity relation then each α-


cut, αR is a crisp equivalence relation. Thus we may use any
similarity relation R and by taking an α - cut αR for any value α
∈ (0, 1], create a crisp equivalence relation that represents the
presence of similarity between the elements to the degree α.
Each of these equivalence relations form a partition of X. Let π
(αR) denote the partition corresponding to the equivalence
relation αR. Clearly any two elements x and y belong to the
same block of this partition if and only if R (x, y) ≥ α. Each
similarity relation is associated with the set π (R) = {π (αR) ⏐α
∈ (0,1]} of partition of X. These partitions are nested in the
sense that π (αR) is a refinement of π ( βR) if and only α ≥ β.
The equivalence classes formed by the levels of refinement
of a similarity relation can be interpreted as grouping elements
that are similar to each other and only to each other to a degree
not less than α.
Just as equivalences classes are defined by an equivalence
relation, similarity classes are defined by a similarity relation.
For a given similarity relation R(X, X) the similarity class for
each x ∈ X is a fuzzy set in which the membership grade of

181
each element y ∈ X is simply the strength of that elements
relation to x or R(x, y). Thus the similarity class for an element
x represents the degree to which all the other members of X are
similar to x. Expect in the restricted case of equivalence classes
themselves, similarity classes are fuzzy and therefore not
generally disjoint.
Similarity relations are conveniently represented by
membership matrices. Given a similarity relation R, the
similarity class for each element is defined by the row of the
membership matrix of R that corresponds to that element.
Fuzzy equivalence is a cutworthy property of binary relation
R(X, X) since it is preserved in the classical sense in each α-cut
of R. This implies that the properties of fuzzy reflexivity,
symmetry and max-min transitivity are also cutworthy. Binary
relations are symmetric and transitive but not reflexive are
usually referred to as quasi equivalence relations.
The notion of fuzzy equations is associated with the concept
of compositions of binary relations. The composition of two
fuzzy binary relations P (X, Y) and Q (Y, Z) can be defined, in
general in terms of an operation on the membership matrices of
P and Q that resembles matrix multiplication. This operation
involves exactly the same combinations of matrix entries as in
the regular matrix multiplication. However the multiplication
and addition that are applied to these combinations in the matrix
multiplication are replaced with other operations, these
alternative operations represent in each given context the
appropriate operations of fuzzy set intersections and union
respectively. In the max-min composition for example, the
multiplication and addition are replaced with the min and max
operations respectively.
We shall give the notational conventions. Consider three
fuzzy binary relations P (X, Y), Q (Y, Z) and R (X, Z) which
are defined on the sets

X = {xi | i ∈ I}
Y = {yj | j ∈ J} and
Z = {zk | k ∈ K}

182
where we assume that I = Nn J = Nm and K = Ns. Let the
membership matrices of P, Q and R be denoted by P = [pij], Q =
[qij], R = [rik] respectively, where pij = P (xi, yj), qjk = Q (yj, zk),
rij = R (xi, zk) for all i ∈ I (=Nn), j ∈ J = (Nm) and k ∈ K (= Ns).
This clearly implies that all entries in the matrices P, Q, and R
are real numbers from the unit interval [0, 1]. Assume now that
the three relations constrain each other in such a way that P ° Q
= R where ° denotes max-min composition. This means that
max min (pij, qjk) = rik for all i ∈ I and k ∈- K. That is the
j∈ J

matrix equation P° Q = R encompasses n × s simultaneous


equations of the form max min (pij, qjk ) = rik. When two of the
j∈ J

components in each of the equations are given and one is


unknown these equations are referred to as fuzzy relation
equations.
When matrices P and Q are given the matrix R is to
determined using P ° Q = R. The problem is trivial. It is solved
simply by performing the max-min multiplication – like
operation on P and Q as defined by max min (pij, qjk ) = rik.
j∈ J

Clearly the solution in this case exists and is unique. The


problem becomes far from trivial when one of the two matrices
on the left hand side of P ° Q = R is unknown. In this case the
solution is guaranteed neither to exist nor to be unique.
Since R in P ° Q = R is obtained by composing P and Q it is
suggestive to view the problem of determining P (or
alternatively Q ) from R to Q (or alternatively R and P) as a
decomposition of R with respect to Q (or alternatively with
respect to P). Since many problems in various contexts can be
formulated as problems of decomposition, the utility of any
method for solving P ° Q = R is quite high. The use of fuzzy
relation equations in some applications is illustrated. Assume
that we have a method for solving P ° Q = R only for the first
decomposition problem (given Q and R).
Then we can directly utilize this method for solving the
second decomposition problem as well. We simply write P ° Q
= R in the form Q-1 o P-1 = R-1 employing transposed matrices.

183
We can solve Q-1 o P-1 = R-1 for Q-1 by our method and then
obtain the solution of P ° Q = R by (Q-1)-1 = Q.
We study the problem of partitioning the equations P ° Q =
R. We assume that a specific pair of matrices R and Q in the
equations P ° Q = R is given. Let each particular matrix P that
satisfies P ° Q = R is called its solution and let S (Q, R) = {P | P
° Q = R} denote the set of all solutions (the solution set).
It is easy to see this problem can be partitioned, without loss
of generality into a set of simpler problems expressed by the
matrix equations pi o Q = ri for all i∈I where
Pi = [pij | j ∈ J] and
ri = [rik | k ∈ K].

Indeed each of the equation in max min (pijqjk) = rik


j∈ J

contains unknown pij identified only by one particular value of


the index i, that is, the unknown pij distinguished by different
values of i do not appear together in any of the individual
equations. Observe that pi, Q, and ri in pi ° Q = ri represent
respectively, a fuzzy set on Y, a fuzzy relation on Y × Z and a
fuzzy set on Z. Let Si (Q, ri) = [pi | pi o Q = ri] denote, for each i
∈ I, the solution set of one of the simpler problem expressed by
pi ° Q = ri.
Thus the matrices P in S (Q, R) = [P | P ° Q = R ] can be
viewed as one column matrix

⎡ p1 ⎤
⎢p ⎥
P = ⎢ 2⎥
⎢ ⎥
⎢ ⎥
⎣pn ⎦

where pi ∈ Si (Q, ri) for all i ∈ I = (=Nn). It follows immediately


from max min (pij, qjk ) = rik. That if max qjk < rik for some i ∈
j∈ J j∈ J

I and some k ∈ K, then no values pij ∈ [0, 1] exists (j ∈ J) that


satisfy P ° Q = R, therefore no matrix P exists that satisfies the
matrix equation.

184
This proposition can be stated more concisely as follows if

max q jk < max rik


j∈J j∈J

for some k ∈ K then S (Q, R) = φ. This proposition allows us in


certain cases to determine quickly that P ° Q = R has no
solutions its negation however is only a necessary not sufficient
condition for the existence of a solution of P ° Q = R that is for
S (Q, R) ≠ φ. Since P ° Q = R can be partitioned without loss of
generality into a set of equations of the form pi ° Q = ri we need
only methods for solving equations of the later form in order to
arrive at a solution.
We may therefore restrict our further discussion of matrix
equations of the form P ° Q = R to matrix equation of the
simpler form P ° Q = r, where p = [pj | j ∈ J], Q = [qjk | j ∈ J, k ∈
K] and r = {rk | k ∈ K].
We just recall the solution method as discussed by [43]. For
the sake of consistency with our previous discussion, let us
again assume that p, Q and r represent respectively a fuzzy set
on Y, a fuzzy relation on Y × Z and a fuzzy set on Z. Moreover
let J = Nm and K = Ns and let S (Q, r) = {p | p ° Q = r} denote
the solution set of
p ° Q = r.

In order to describe a method of solving p ° Q = r we need


to introduce some additional concepts and convenient notation.
First let ℘ denote the set of all possible vectors.

p = {pj | j ∈ J}

such that pj ∈ [0, 1] for all j ∈ J, and let a partial ordering


on ℘ be defined as follows for any pair p1, p2 ∈ ℘ p1 ≤ p2 if
and only if pi2 ≤ p 2j for all j ∈J. Given an arbitrary pair p1, p2 ∈
℘ such that p1 ≤ p2 let [p1 , p2] = {p ∈ ℘ | p1 ≤ p < p2}. For any
pair p1, p2 ∈ ℘ ({p1, p2} ≤ } is a lattice.

185
Now we recall some of the properties of the solution set S
(Q, r). Employing the partial ordering on ℘, let an element p̂ of
S (Q, r) be called a maximal solution of p ° Q = r if for all p ∈ S
(Q, r), p ≥ p̂ implies p = p̂ if for all p ∈ S (Q, r) p < p then
that is the maximum solution. Similar discussion can be made
on the minimal solution of p ° Q = r. The minimal solution is
unique if p ≥ p̂ (i.e. p̂ is unique).
It is well known when ever the solution set S (Q, r) is not
empty it always contains a unique maximum solution p̂ and it
may contain several minimal solution. Let S (Q, r) denote the
set of all minimal solutions.
It is known that the solution set S (Q, r) is fully
characterized by the maximum and minimal solution in the
sense that it consists exactly of the maximum solution p̂ all the
minimal solutions and all elements of ℘ that are between p̂
and the numeral solution.
Thus S (Q, r) = ∪ [ p, pˆ ] where the union is taken for all
p

p ∈ S (Q, r). When S (Q, r) ≠ φ, the maximum solution.


p̂ = [ p̂ j | j ∈ J] of p ° Q = r
is determined as follows:
⎧r if q jk > rk
pˆ j = min σ (qik, rk) where σ (qjk, rk) = ⎨ k
k ∈K
⎩ 1 otherwise

when p̂ determined in this way does not satisfy p ° Q = r then


S(Q, r) = φ. That is the existence of the maximum solution p̂ as
determined by pˆ j = min σ (qik, rk) is a necessary and sufficient
k ∈K

condition for S (Q, r) ≠ φ. Once p̂ is determined by


pˆ j = min σ (qik, rk),
k ∈K

we must check to see if it satisfies the given matrix equations p


° Q = r. If it does not then the equation has no solution (S (Q, r)
= φ), otherwise p̂ in the maximum solution of the equation and
we next determine the set S (Q, r) of its minimal solutions.

186
2.2 Neutrosophy and Neutrosophic models

This section has five subsections. In the first subsection a very


brief introduction to neutrosophy is given for more refer [187-
190]. In the second subsection some basic neutrosophic
structures needed to make the book a self contained one are
introduced. Sub section three briefly describes Neutrosophic
Cognitive Maps (NCMs). Neutrosophic Relational Maps
(NRMs) are recollected in the subsection four. The final
subsection gives a brief description of binary neutrosophic
relation and their properties.

2.2.1 An Introduction to Neutrosophy

In this section we introduce the notion of neutrosophic logic


created by Florentine Smarandache [187-190], which is an
extension / combination of the fuzzy logic in which
indeterminacy is included. It has become very essential that the
notion of neutrosophic logic play a vital role in several of the
real world problems like law, medicine, industry, finance, IT,
stocks and share etc. Use of neutrosophic notions will be
illustrated/ applied in the later sections of this chapter. Fuzzy
theory only measures the grade of membership or the non-
existence of a membership in the revolutionary way but fuzzy
theory has failed to attribute the concept when the relations
between notions or nodes or concepts in problems are
indeterminate. In fact one can say the inclusion of the concept
of indeterminate situation with fuzzy concepts will form the
neutrosophic logic. As in this book the concept of only fuzzy
cognitive maps are dealt which mainly deals with the relation /
non-relation between two nodes or concepts but it fails to deal
the relation between two conceptual nodes when the relation is
an indeterminate one. Neutrosophic logic is the only tool known
to us, which deals with the notions of indeterminacy, and here
we give a brief description of it. For more about Neutrosophic
logic please refer Smarandache [187-190].

187
DEFINITION 2.2.1.1: In the neutrosophic logic every logical
variable x is described by an ordered triple x = (T, I, F) where
T is the degree of truth, F is the degree of false and I the level of
indeterminacy.

(A). To maintain consistency with the classical and fuzzy


logics and with probability there is the special case
where T + I + F = 1.

(B). But to refer to intuitionistic logic, which means


incomplete information on a variable proposition or
event one has T + I + F < 1.

(C). Analogically referring to Paraconsistent logic, which


means contradictory sources of information about a
same logical variable, proposition or event one has T +
I + F > 1.

Thus the advantage of using Neutrosophic logic is that this logic


distinguishes between relative truth that is a truth is one or a few
worlds only noted by 1 and absolute truth denoted by 1+.
Likewise neutrosophic logic distinguishes between relative
falsehood, noted by 0 and absolute falsehood noted by – 0.

It has several applications. One such given by [187-190] is as


follows:

Example 2.2.1.1: From a pool of refugees, waiting in a political


refugee camp in Turkey to get the American visa, a% have the
chance to be accepted – where a varies in the set A, r% to be
rejected – where r varies in the set R, and p% to be in pending
(not yet decided) – where p varies in P.

Say, for example, that the chance of someone Popescu in the


pool to emigrate to USA is (between) 40-60% (considering
different criteria of emigration one gets different percentages,
we have to take care of all of them), the chance of being
rejected is 20-25% or 30-35%, and the chance of being in

188
pending is 10% or 20% or 30%. Then the neutrosophic
probability that Popescu emigrates to the Unites States is

NP (Popescu) = ((40-60) (20-25) ∪ (30-35), {10,20,30}), closer


to the life.

This is a better approach than the classical probability, where 40


P(Popescu) 60, because from the pending chance – which will
be converted to acceptance or rejection – Popescu might get
extra percentage in his will to emigrating and also the superior
limit of the subsets sum
60 + 35 + 30 > 100

and in other cases one may have the inferior sum < 0, while in
the classical fuzzy set theory the superior sum should be 100
and the inferior sum µ 0. In a similar way, we could say about
the element Popescu that Popescu ((40-60), (20-25) ∪ (30-35),
{10, 20, 30}) belongs to the set of accepted refugees.

Example 2.2.1.2: The probability that candidate C will win an


election is say 25-30% true (percent of people voting for him),
35% false (percent of people voting against him), and 40% or
41% indeterminate (percent of people not coming to the ballot
box, or giving a blank vote – not selecting any one or giving a
negative vote cutting all candidate on the list). Dialectic and
dualism don’t work in this case anymore.

Example 2.2.1.3: Another example, the probability that


tomorrow it will rain is say 50-54% true according to
meteorologists who have investigated the past years weather, 30
or 34-35% false according to today’s very sunny and droughty
summer, and 10 or 20% undecided (indeterminate).

Example 2.2.1.4: The probability that Yankees will win


tomorrow versus Cowboys is 60% true (according to their
confrontation’s history giving Yankees’ satisfaction), 30-32%
false (supposing Cowboys are actually up to the mark, while
Yankees are declining), and 10 or 11 or 12% indeterminate (left
to the hazard: sickness of players, referee’s mistakes,

189
atmospheric conditions during the game). These parameters act
on players’ psychology.
As in this book we use mainly the notion of neutrosophic
logic with regard to the indeterminacy of any relation in
cognitive maps we are restraining ourselves from dealing with
several interesting concepts about neutrosophic logic. As FCMs
deals with unsupervised data and the existence or non-existence
of cognitive relation, we do not in this book elaborately describe
the notion of neutrosophic concepts.
However we just state, suppose in a legal issue the jury or
the judge cannot always prove the evidence in a case, in several
places we may not be able to derive any conclusions from the
existing facts because of which we cannot make a conclusion
that no relation exists or otherwise. But existing relation is an
indeterminate. So in the case when the concept of indeterminacy
exists the judgment ought to be very carefully analyzed be it a
civil case or a criminal case. FCMs are deployed only where the
existence or non-existence is dealt with but however in our
Neutrosophic Cognitive Maps we will deal with the notion of
indeterminacy of the evidence also. Thus legal side has lot of
Neutrosophic (NCMs) applications. Also we will show how
NCMs can be used to study factors as varied as stock markets,
medical diagnosis, etc.

2.2.2 Some Basic Neutrosophic Structures

In this section we define some new neutrosophic algebraic


structures like neutrosophic fields, neutrosophic spaces and
neutrosophic matrices and illustrate them with examples. For
these notions are used in the definition of neutrosophic
cognitive maps which is dealt in the later sections of this
chapter.
Throughout this book by 'I' we denote the indeterminacy of
any notion/ concept/ relation. That is when we are not in a
position to associate a relation between any two concepts then
we denote it as indeterminacy.
Further in this book we assume all fields to be real fields of
characteristic 0 and all vector spaces are taken as real spaces

190
over reals and we denote the indeterminacy by 'I' as i will make
a confusion as i denotes the imaginary value viz i2 = –1 that is
−1 = i.

DEFINITION 2.2.2.1: Let K be the field of reals. We call the field


generated by K ∪ I to be the neutrosophic field for it involves
the indeterminacy factor in it. We define I 2 = I, I + I = 2I i.e., I
+…+ I = nI, and if k ∈ K then k.I = kI, 0I = 0. We denote the
neutrosophic field by K(I) which is generated by K ∪ I that is K
(I) = 〈K ∪ I〉.

Example 2.2.2.1: Let R be the field of reals. The neutrosophic


field is generated by 〈R ∪ I〉 i.e. R(I) clearly R ⊂ 〈R ∪ I〉.

Example 2.2.2.2: Let Q be the field of rationals. The


neutrosophic field is generated by Q and I i.e. 〈Q ∪ I〉 denoted
by Q(I).

DEFINITION 2.2.2.2: Let K(I) be a neutrosophic field we say


K(I) is a prime neutrosophic field if K(I) has no proper subfield
which is a neutrosophic field.

Example 2.2.2.3: Q(I) is a prime neutrosophic field where as


R(I) is not a prime neutrosophic field for Q(I) ⊂ R (I).

It is very important to note that all neutrosophic fields are of


characteristic zero. Likewise we can define neutrosophic
subfield.

DEFINITION 2.2.2.3: Let K(I) be a neutrosophic field, P ⊂ K(I)


is a neutrosophic subfield of P if P itself is a neutrosophic field.
K(I) will also be called as the extension neutrosophic field of
the neutrosophic field P.

Now we proceed on to define neutrosophic vector spaces, which


are only defined over neutrosophic fields. We can define two
types of neutrosophic vector spaces one when it is a
neutrosophic vector space over ordinary field other being

191
neutrosophic vector space over neutrosophic fields. To this end
we have to define neutrosophic group under addition.

DEFINITION 2.2.2.4: We know Z is the abelian group under


addition. Z(I) denote the additive abelian group generated by
the set Z and I, Z(I) is called the neutrosophic abelian group
under ‘+’.

Thus to define basically a neutrosophic group under addition we


need a group under addition. So we proceed on to define
neutrosophic abelian group under addition. Suppose G is an
additive abelian group under ‘+’. G(I) = 〈G ∪ I〉, additive
group generated by G and I, G(I) is called the neutrosophic
abelian group under ‘+’.

Example 2.2.2.4: Let Q be the group under ‘+’; Q (I) = 〈Q ∪ I〉


is the neutrosophic abelian group under addition; ‘+’.

Example 2.2.2.5: R be the additive group of reals, R(I) = 〈R ∪


I〉 is the neutrosophic group under addition.

Example 2.2.2.6: Mn×m(I) = {(aij) ⏐ aij ∈ Z(I)} be the collection


of all n × m matrices under ‘+’ Mn×m(I) is a neutrosophic group
under ‘+’.

Now we proceed on to define neutrosophic subgroup.

DEFINITION 2.2.2.5: Let G(I) be the neutrosophic group under


addition. P ⊂ G(I) be a proper subset of G(I). P is said to be
neutrosophic subgroup of G(I) if P itself is a neutrosophic
group i.e. P = 〈P1 ∪ I〉 where P1 is an additive subgroup of G.

Example 2.2.2.7: Let Z(I) = 〈Z ∪ I〉 be a neutrosophic group


under ‘+’. 〈2Z ∪ I〉 = 2Z(I) is the neutrosophic subgroup of Z
(I).

In fact Z(I) has several neutrosophic subgroups.

192
Now we proceed on to define the notion of neutrosophic
quotient group.

DEFINITION 2.2.2.6: Let G (I) = 〈G ∪ I〉 be a neutrosophic


group under ‘+’, suppose P (I) be a neutrosophic subgroup of G
(I) then the neutrosophic quotient group

G(I)
= {a + P(I) a ∈ G (I)} .
P(I)

Example 2.2.2.8: Let Z (I) be a neutrosophic group under


addition, Z the group of integers under addition P = 2Z(I) is a
neutrosophic subgroup of Z(I) the neutrosophic subgroup of
Z(I), the neutrosophic quotient group

Z(I)
= {a + 2Z(I) a ∈ Z(I)} = {(2n+1) + (2n+1) I ⏐ n ∈ Z}.
P

Z(I)
Clearly is a group. For P = 2Z (I) serves as the additive
P
Z(I)
identity. Take a, b ∈ . If a, b ∈ Z(I) \ P then two
P
possibilities occur.

a + b is odd times I or a + b is odd or a + b is even times I or


even if a + b is even or even times I then a + b ∈ P. if a + b is
Z(I)
odd or odd times I a + b ∈ .
P = 2Z(I)
It is easily verified that P acts as the identity and every element
in

Z(I)
a + 2Z (I) ∈
2Z(I)

has inverse. Hence the claim.

193
Now we proceed on to define the notion of neutrosophic vector
spaces over fields and then we define neutrosophic vector
spaces over neutrosophic fields.

DEFINITION 2.2.2.7: Let G(I) by an additive abelian


neutrosophic group. K any field. If G(I) is a vector space over K
then we call G(I) a neutrosophic vector space over K.

Now we give the notion of strong neutrosophic vector space.

DEFINITION 2.2.2.8: Let G(I) be a neutrosophic abelian group.


K(I) be a neutrosophic field. If G(I) is a vector space over K(I)
then we call G(I) the strong neutrosophic vector space.

THEOREM 2.2.2.1: All strong neutrosophic vector space over


K(I) are a neutrosophic vector space over K; as K ⊂ K(I).

Proof: Follows directly by the very definitions.

Thus when we speak of neutrosophic spaces we mean either a


neutrosophic vector space over K or a strong neutrosophic
vector space over the neutrosophic field K(I). By basis we mean
a linearly independent set which spans the neutrosophic space.

Now we illustrate with an example.

Example 2.2.2.9: Let R(I) × R(I) = V be an additive abelian


neutrosophic group over the neutrosophic field R(I). Clearly V
is a strong neutrosophic vector space over R(I). The basis of V
are {(0,1), (1,0)}.

Example 2.2.2.10: Let V = R(I) × R(I) be a neutrosophic


abelian group under addition. V is a neutrosophic vector space
over R. The neutrosophic basis of V are {(1,0), (0,1), (I,0),
(0,I)}, which is a basis of the vector space V over R.

A study of these basis and its relations happens to be an


interesting form of research.

194
DEFINITION 2.2.2.9: Let G(I) be a neutrosophic vector space
over the field K. The number of elements in the neutrosophic
basis is called the neutrosophic dimension of G(I).

DEFINITION 2.2.2.10: Let G(I) be a strong neutrosophic vector


space over the neutrosophic field K(I). The number of elements
in the strong neutrosophic basis is called the strong
neutrosophic dimension of G(I).

We denote the neutrosophic dimension of G(I) over K by Nk


(dim) of G (I) and that the strong neutrosophic dimension of G
(I) by SNK(I) (dim) of G(I).
Now we define the notion of neutrosophic matrices.

DEFINITION 2.2.2.11: Let Mn×m = {(aij) ⏐ aij ∈ K(I)}, where K


(I), is a neutrosophic field. We call Mn×m to be the neutrosophic
matrix.

Example 2.2.2.11: Let Q(I) = 〈Q ∪ I〉 be the neutrosophic field.

⎡ 0 1 I⎤
⎢ −2 4I 0 ⎥
M4×3 = ⎢ ⎥
⎢ 1 −I 2⎥
⎢ ⎥
⎣ 3I 1 0 ⎦

is the neutrosophic matrix, with entries from rationals and the


indeterminacy I. We define product of two neutrosophic
matrices whenever the production is defined as follows:

Let
⎡ −1 2 − I ⎤
A= ⎢ ⎥
⎣ 3 I 0 ⎦ 2×3
and
⎡ −I 1 2 4 ⎤
B = ⎢⎢ 1 I 0 2 ⎥⎥
⎢⎣ 5 −2 3I − I ⎥⎦ 3× 4

195
⎡ −6I + 2 −1 + 4I −2 − 3I I ⎤
AB = ⎢
⎣ −2I 3+ I 6 12 + 2I ⎥⎦ 2×4

(we use the fact I2 = I).

To define neutrosophic cognitive maps we direly need the


notion of neutrosophic matrices. We use square neutrosophic
matrices for Neutrosophic Cognitive Maps (NCMs) and use
rectangular neutrosophic matrices for Neutrosophic Relational
Maps (NRMs).

2.2.3 On Neutrosophic Cognitive Maps

The notion of Fuzzy Cognitive Maps (FCMs) which are fuzzy


signed directed graphs with feedback are discussed and
described in section 2.1.1 of this book. The directed edge eij
from causal concept Ci to concept Cj measures how much Ci
causes Cj. The time varying concept function Ci(t) measures the
non negative occurrence of some fuzzy event, perhaps the
strength of a political sentiment, historical trend or opinion
about some topics like child labor or school dropouts etc. FCMs
model the world as a collection of classes and causal relations
between them.
The edge eij takes values in the fuzzy causal interval [–1, 1]
(eij = 0 indicates no causality, eij > 0 indicates causal increase;
that Cj increases as Ci increases and Cj decreases as Ci
decreases, eij < 0 indicates causal decrease or negative causality
Cj decreases as Ci increases or Cj, increases as Ci decreases.
Simple FCMs have edge value in {-1, 0, 1}. Thus if causality
occurs it occurs to maximal positive or negative degree.
It is important to note that eij measures only absence or
presence of influence of the node Ci on Cj but till now any
researcher has not contemplated the indeterminacy of any
relation between two nodes Ci and Cj. When we deal with
unsupervised data, there are situations when no relation can be
determined between some two nodes. So in this section we try
to introduce the indeterminacy in FCMs, and we choose to call
this generalized structure as Neutrosophic Cognitive Maps

196
(NCMs). In our view this will certainly give a more appropriate
result and also caution the user about the risk of indeterminacy.

Now we proceed on to define the concepts about NCMs.

DEFINITION 2.2.3.1: A Neutrosophic Cognitive Map (NCM) is a


neutrosophic directed graph with concepts like policies, events
etc. as nodes and causalities or indeterminates as edges. It
represents the causal relationship between concepts.

Let C1, C2, …, Cn denote n nodes, further we assume each node


is a neutrosophic vector from neutrosophic vector space V. So a
node Ci will be represented by (x1, …, xn) where xk’s are zero or
one or I (I is the indeterminate introduced in Sections 2.2 and
2.3 of the chapter 2) and xk = 1 means that the node Ck is in the
on state and xk = 0 means the node is in the off state and xk = I
means the nodes state is an indeterminate at that time or in that
situation.
Let Ci and Cj denote the two nodes of the NCM. The
directed edge from Ci to Cj denotes the causality of Ci on Cj
called connections. Every edge in the NCM is weighted with a
number in the set {-1, 0, 1, I}. Let eij be the weight of the
directed edge CiCj, eij ∈ {–1, 0, 1, I}. eij = 0 if Ci does not have
any effect on Cj, eij = 1 if increase (or decrease) in Ci causes
increase (or decreases) in Cj, eij = –1 if increase (or decrease) in
Ci causes decrease (or increase) in Cj . eij = I if the relation or
effect of Ci on Cj is an indeterminate.

DEFINITION 2.2.3.2: NCMs with edge weight from {-1, 0, 1, I}


are called simple NCMs.

DEFINITION 2.2.3.3: Let C1, C2, …, Cn be nodes of a NCM. Let


the neutrosophic matrix N(E) be defined as N(E) = (eij) where eij
is the weight of the directed edge Ci Cj, where eij ∈ {0, 1, -1, I}.
N(E) is called the neutrosophic adjacency matrix of the NCM.

DEFINITION 2.2.3.4: Let C1, C2, …, Cn be the nodes of the


NCM. Let A = (a1, a2,…, an) where ai ∈ {0, 1, I}. A is called the

197
instantaneous state neutrosophic vector and it denotes the on –
off – indeterminate state position of the node at an instant

ai = 0 if ai is off (no effect)


ai = 1 if ai is on (has effect)
ai = I if ai is indeterminate(effect cannot be determined)

for i = 1, 2,…, n.

DEFINITION 2.2.3.5: Let C1, C2, …, Cn be the nodes of the FCM.


Let C1C 2 , C 2 C 3 , C3C4 , … , Ci C j be the edges of the NCM.
Then the edges form a directed cycle. An NCM is said to be
cyclic if it possesses a directed cyclic. An NCM is said to be
acyclic if it does not possess any directed cycle.

DEFINITION 2.2.3.6: An NCM with cycles is said to have a


feedback. When there is a feedback in the NCM i.e. when the
causal relations flow through a cycle in a revolutionary manner
the NCM is called a dynamical system.

DEFINITION 2.2.3.7: Let C1C2 , C2 C3 , , Cn −1Cn be cycle, when


Ci is switched on and if the causality flow through the edges of a
cycle and if it again causes Ci, we say that the dynamical system
goes round and round. This is true for any node Ci, for i = 1,
2,…, n. the equilibrium state for this dynamical system is called
the hidden pattern.

DEFINITION 2.2.3.8: If the equilibrium state of a dynamical


system is a unique state vector, then it is called a fixed point.
Consider the NCM with C1, C2, …, Cn as nodes. For example let
us start the dynamical system by switching on C1. Let us assume
that the NCM settles down with C1 and Cn on, i.e. the state
vector remain as (1, 0, …, 1) this neutrosophic state vector (1,0,
…, 0, 1) is called the fixed point.

DEFINITION 2.2.3.9: If the NCM settles with a neutrosophic


state vector repeating in the form

198
A1 → A2 → … → Ai → A1,

then this equilibrium is called a limit cycle of the NCM.

METHODS OF DETERMINING THE HIDDEN PATTERN:

Let C1, C2, …, Cn be the nodes of an NCM, with feedback. Let


E be the associated adjacency matrix. Let us find the hidden
pattern when C1 is switched on when an input is given as the
vector A1 = (1, 0, 0, …, 0), the data should pass through the
neutrosophic matrix N(E), this is done by multiplying A1 by the
matrix N(E). Let A1N(E) = (a1, a2, …, an) with the threshold
operation that is by replacing ai by 1 if ai > k and ai by 0 if ai < k
(k – a suitable positive integer) and ai by I if ai is not a integer.
We update the resulting concept, the concept C1 is included in
the updated vector by making the first coordinate as 1 in the
resulting vector. Suppose A1N(E) → A2 then consider A2N(E)
and repeat the same procedure. This procedure is repeated till
we get a limit cycle or a fixed point.

DEFINITION 2.2.3.10: Finite number of NCMs can be combined


together to produce the joint effect of all NCMs. If N(E1), N(E2),
…, N(Ep) be the neutrosophic adjacency matrices of a NCM
with nodes C1, C2, …, Cn then the combined NCM is got by
adding all the neutrosophic adjacency matrices N(E1), …,
N(Ep). We denote the combined NCMs adjacency neutrosophic
matrix by N(E) = N(E1) + N(E2)+ …+ N(Ep).

NOTATION: Let (a1, a2, …, an) and (a'1, a'2, … , a'n) be two
neutrosophic vectors. We say (a1, a2, … , an) is equivalent to
(a'1, a'2, … , a'n) denoted by ((a1, a2, … , an) ~ (a'1, a'2, …, a'n) if
(a'1, a'2, … , a'n) is got after thresholding and updating the vector
(a1, a2, … , an) after passing through the neutrosophic adjacency
matrix N(E).

The following are very important:

199
Note 1: The nodes C1, C2, …, Cn are nodes are not
indeterminate nodes because they indicate the concepts which
are well known. But the edges connecting Ci and Cj may be
indeterminate i.e. an expert may not be in a position to say that
Ci has some causality on Cj either will he be in a position to
state that Ci has no relation with Cj in such cases the relation
between Ci and Cj which is indeterminate is denoted by I.

Note 2: The nodes when sent will have only ones and zeros i.e.
on and off states, but after the state vector passes through the
neutrosophic adjacency matrix the resultant vector will have
entries from {0, 1, I} i.e. they can be neutrosophic vectors.

The presence of I in any of the coordinate implies the expert


cannot say the presence of that node i.e. on state of it after
passing through N(E) nor can we say the absence of the node
i.e. off state of it the effect on the node after passing through the
dynamical system is indeterminate so only it is represented by I.
Thus only in case of NCMs we can say the effect of any node
on other nodes can also be indeterminates. Such possibilities
and analysis is totally absent in the case of FCMs.

Note 3: In the neutrosophic matrix N(E), the presence of I in the


aij the place shows, that the causality between the two nodes i.e.
the effect of Ci on Cj is indeterminate. Such chances of being
indeterminate is very possible in case of unsupervised data and
that too in the study of FCMs which are derived from the
directed graphs.

Thus only NCMs helps in such analysis.

Now we shall represent a few examples to show how in this set


up NCMs is preferred to FCMs. At the outset before we proceed
to give examples we make it clear that all unsupervised data
need not have NCMs to be applied to it. Only data which have
the relation between two nodes to be an indeterminate need to
be modeled with NCMs if the data has no indeterminacy factor
between any pair of nodes one need not go for NCMs; FCMs
will do the best job.

200
2.2.4 Neutrosophic Relational Maps

Neutrosophic Cognitive Maps (NCMs) promote the causal


relationships between concurrently active units or decides the
absence of any relation between two units or the indeterminacy
of any relation between any two units. But in Neutrosophic
Relational Maps (NRMs) we divide the very causal nodes into
two disjoint units.
Thus for the modeling of a NRM we need a domain space
and a range space which are disjoint in the sense of concepts.
We further assume no intermediate relations exist within the
domain and the range spaces. The number of elements or nodes
in the range space need not be equal to the number of elements
or nodes in the domain space.
Throughout this section we assume the elements of a
domain space are taken from the neutrosophic vector space of
dimension n and that of the range space are neutrosophic vector
space of dimension m. (m in general need not be equal to n). We
denote by R the set of nodes R1, …, Rm of the range space,
where R = {(x1, …, xm) ⏐xj = 0 or 1 for j = 1, 2, …, m}.
If xi = 1 it means that node Ri is in the on state and if xi = 0
it means that the node Ri is in the off state and if xi = I in the
resultant vector it means the effect of the node xi is
indeterminate or whether it will be off or on cannot be predicted
by the neutrosophic dynamical system.
It is very important to note that when we send the state
vectors they are always taken as the real state vectors for we
know the node or the concept is in the on state or in the off state
but when the state vector passes through the Neutrosophic
dynamical system some other node may become indeterminate
i.e. due to the presence of a node we may not be able to predict
the presence or the absence of the other node i.e., it is
indeterminate, denoted by the symbol I, thus the resultant vector
can be a neutrosophic vector.

DEFINITION 2.2.4.1: A Neutrosophic Relational Map (NRM) is


a Neutrosophic directed graph or a map from D to R with

201
concepts like policies or events etc. as nodes and causalities as
edges. (Here by causalities we mean or include the
indeterminate causalities also). It represents Neutrosophic
Relations and Causal Relations between spaces D and R .
Let Di and Rj denote the nodes of an NRM. The directed
edge from Di to Rj denotes the causality of Di on Rj called
relations. Every edge in the NRM is weighted with a number in
the set {0, +1, –1, I}.
Let eij be the weight of the edge Di Rj, eij ∈ {0, 1, –1, I}. The
weight of the edge Di Rj is positive if increase in Di implies
increase in Rj or decrease in Di implies decrease in Rj i.e.
causality of Di on Rj is 1. If eij = –1 then increase (or decrease)
in Di implies decrease (or increase) in Rj. If eij = 0 then Di does
not have any effect on Rj. If eij = I it implies we are not in a
position to determine the effect of Di on Rj i.e. the effect of Di on
Rj is an indeterminate so we denote it by I.

DEFINITION 2.2.4.2: When the nodes of the NRM take edge


values from {0, 1, –1, I} we say the NRMs are simple NRMs.

DEFINITION 2.2.4.3: Let D1, …, Dn be the nodes of the domain


space D of an NRM and let R1, R2, …, Rm be the nodes of the
range space R of the same NRM. Let the matrix N(E) be defined
as N(E) = (eij ) where eij is the weight of the directed edge Di Rj
(or Rj Di ) and eij ∈ {0, 1, –1, I}. N(E) is called the Neutrosophic
Relational Matrix of the NRM.

The following remark is important and interesting to find its


mention in this book.

Remark: Unlike NCMs, NRMs can also be rectangular


matrices with rows corresponding to the domain space and
columns corresponding to the range space. This is one of the
marked difference between NRMs and NCMs. Further the
number of entries for a particular model which can be treated as
disjoint sets when dealt as a NRM has very much less entries
than when the same model is treated as a NCM.

202
Thus in many cases when the unsupervised data under study
or consideration can be spilt as disjoint sets of nodes or
concepts; certainly NRMs are a better tool than the NCMs.

DEFINITION 2.2.4.4: Let D1, …, Dn and R1,…, Rm denote the


nodes of a NRM. Let A = (a1,…, an ), ai ∈ {0, 1} is called the
neutrosophic instantaneous state vector of the domain space
and it denotes the on-off position of the nodes at any instant.
Similarly let B = (b1,…, bn) bi ∈ {0, 1}, B is called
instantaneous state vector of the range space and it denotes the
on-off position of the nodes at any instant, ai = 0 if ai is off and
ai = 1 if ai is on for i = 1, 2, …, n. Similarly, bi = 0 if bi is off
and bi = 1 if bi is on for i = 1, 2,…, m.

DEFINITION 2.2.4.5: Let D1,…, Dn and R1, R2,…, Rm be the


nodes of a NRM. Let Di Rj (or Rj Di ) be the edges of an NRM, j
= 1, 2,…, m and i = 1, 2,…, n. The edges form a directed cycle.
An NRM is said to be a cycle if it possess a directed cycle. An
NRM is said to be acyclic if it does not possess any directed
cycle.

DEFINITION 2.2.4.6: A NRM with cycles is said to be a NRM


with feedback.

DEFINITION 2.2.4.7: When there is a feedback in the NRM i.e.


when the causal relations flow through a cycle in a
revolutionary manner the NRM is called a Neutrosophic
dynamical system.

DEFINITION 2.2.4.8: Let Di Rj (or Rj Di) 1 ≤ j ≤ m, 1 ≤ i ≤ n,


when Rj (or Di ) is switched on and if causality flows through
edges of a cycle and if it again causes Rj (or Di ) we say that the
Neutrosophical dynamical system goes round and round. This is
true for any node Rj ( or Di ) for 1 ≤ j ≤ m (or 1 ≤ i ≤ n). The
equilibrium state of this Neutrosophical dynamical system is
called the Neutrosophic hidden pattern.

DEFINITION 2.2.4.9: If the equilibrium state of a


Neutrosophical dynamical system is a unique Neutrosophic

203
state vector, then it is called the fixed point. Consider an NRM
with R1, R2, …, Rm and D1, D2,…, Dn as nodes. For example let
us start the dynamical system by switching on R1 (or D1). Let us
assume that the NRM settles down with R1 and Rm (or D1 and
Dn) on, or indeterminate on, i.e. the Neutrosophic state vector
remains as (1, 0, 0,…, 1) or (1, 0, 0,…I) (or (1, 0, 0,…1) or (1,
0, 0,…I) in D), this state vector is called the fixed point.

DEFINITION 2.2.4.10: If the NRM settles down with a state


vector repeating in the form A1 → A2 → A3 → …→ Ai → A1 (or
B1 → B2 → …→ Bi → B1) then this equilibrium is called a limit
cycle.

METHODS OF DETERMINING THE HIDDEN PATTERN IN A


NRM

Let R1, R2,…, Rm and D1, D2,…, Dn be the nodes of a NRM


with feedback. Let N(E) be the Neutrosophic Relational Matrix.
Let us find the hidden pattern when D1 is switched on i.e. when
an input is given as a vector; A1 = (1, 0, …, 0) in D; the data
should pass through the relational matrix N(E).
This is done by multiplying A1 with the Neutrosophic
relational matrix N(E). Let A1N(E) = (r1, r2,…, rm) after
thresholding and updating the resultant vector we get A1E ∈ R,
Now let B = A1E we pass on B into the system (N(E))T and
obtain B(N(E))T. We update and threshold the vector B(N(E))T
so that B(N(E))T ∈ D.

This procedure is repeated till we get a limit cycle or a fixed


point.

DEFINITION 2.2.4.11: Finite number of NRMs can be combined


together to produce the joint effect of all NRMs. Let N(E1),
N(E2),…, N(Er) be the Neutrosophic relational matrices of the
NRMs with nodes R1,…, Rm and D1,…,Dn, then the combined
NRM is represented by the neutrosophic relational matrix N(E)
= N(E1) + N(E2) +…+ N(Er).

204
2.2.5 Binary Neutrosophic Relation and their Properties

In this section we introduce the notion of neutrosophic relational


equations and fuzzy neutrosophic relational equations and
analyze and apply them to real-world problems, which are
abundant with the concept of indeterminacy. We also mention
that most of the unsupervised data also involve at least to certain
degrees the notion of indeterminacy.
Throughout this section by a neutrosophic matrix we mean
a matrix whose entries are from the set N = [0, 1] ∪ I and by a
fuzzy neutrosophic matrix we mean a matrix whose entries are
from N’ = [0, 1] ∪ {nI / n ∈ (0, 1]}.
Now we proceed on to define binary neutrosophic relations
and binary neutrosophic fuzzy relation.
A binary neutrosophic relation RN(x, y) may assign to each
element of X two or more elements of Y or the indeterminate I.
Some basic operations on functions such as the inverse and
composition are applicable to binary relations as well. Given a
neutrosophic relation RN(X, Y) its domain is a neutrosophic set
on X ∪ I domain R whose membership function is defined by
domR(x) = max R N ( x , y) for each x ∈ X ∪ I.
y∈X ∪ I

That is each element of set X ∪ I belongs to the domain of


R to the degree equal to the strength of its strongest relation to
any member of set Y ∪ I. The degree may be an indeterminate I
also. Thus this is one of the marked difference between the
binary fuzzy relation and the binary neutrosophic relation. The
range of RN(X, Y) is a neutrosophic relation on Y, ran R whose
membership is defined by ran R(y) = max R N ( x, y) for each y ∈
x∈X
Y, that is the strength of the strongest relation that each element
of Y has to an element of X is equal to the degree of that
element’s membership in the range of R or it can be an
indeterminate I.
The height of a neutrosophic relation RN(x, y) is a number
h(R) or an indeterminate I defined by hN(R) = max max RN(x,
y∈Y ∪ I x∈X ∪ I

y). That is hN(R) is the largest membership grade attained by


any pair (x, y) in R or the indeterminate I.

205
A convenient representation of the neutrosophic binary
relation RN(X, Y) are membership matrices R = [γxy] where γxy
∈ RN(x, y). Another useful representation of a binary
neutrosophic relation is a neutrosophic sagittal diagram. Each of
the sets X, Y represented by a set of nodes in the diagram, nodes
corresponding to one set are clearly distinguished from nodes
representing the other set. Elements of X’ × Y’ with non-zero
membership grades in RN(X, Y) are represented in the diagram
by lines connecting the respective nodes. These lines are labeled
with the values of the membership grades.
An example of the neutrosophic sagittal diagram is a binary
neutrosophic relation RN(X, Y) together with the membership

x1
.5 y1
x2 .3
.4 y2
x3 1
.2 y3
x4
.5 y4
x5 .7
FIGURE: 2.2.5.1

neutrosophic matrix is given below.

y1 y2 y3 y4
x1 ⎡ I 0 0.5⎤
0
x2 ⎢0.3 0 0.4 0 ⎥⎥

x3 ⎢ 1 0 0 0.2⎥ .
⎢ ⎥
x4 ⎢ 0 I 0 0⎥
x 5 ⎢⎣ 0 0 0.5 0.7⎥⎦

The inverse of a neutrosophic relation RN(X, Y) = R(x, y)


for all x ∈ X and all y ∈ Y. A neutrosophic membership matrix
−1
R–1 = [ ryx ] representing R −N1 (Y, X) is the transpose of the
matrix R for RN(X, Y) which means that the rows of R-1 equal

206
the columns of R and the columns of R-1 equal rows of R.
Clearly (R-1)-1 = R for any binary neutrosophic relation.
Consider any two binary neutrosophic relation PN(X, Y) and
QN(Y, Z) with a common set Y. The standard composition of
these relations which is denoted by PN(X, Y) • QN(Y, Z)
produces a binary neutrosophic relation RN(X, Z) on X × Z
defined by RN(x, z) = [P • Q]N(x, z) = max min[PN(x, y), QN(x,
y∈Y

y)] for all x∈ X and all z ∈ Z.


This composition which is based on the standard tN-norm
and tN-co-norm, is often referred to as the max-min
composition. It can be easily verified that even in the case of
binary neutrosophic relations [PN(X, Y) • QN(Y, Z)]-1 = Q −N1 (Z,
Y) • PN−1 (Y, X). [PN(X, Y) • QN(Y, Z)] • RN(Z, W) = PN(X, Y) •
[QN(Y, Z) • RN(Z, W)], that is, the standard (or max-min)
composition is associative and its inverse is equal to the reverse
composition of the inverse relation. However, the standard
composition is not commutative, because QN(Y, Z) • PN(X, Y)
is not well defined when X ≠ Z. Even if X = Z and QN (Y, Z) °
PN (X, Y) are well defined still we can have PN (X, Y) ° Q (Y,
Z) ≠ Q (Y, Z) ° P (X, Y).
Compositions of binary neutrosophic relation can the
performed conveniently in terms of membership matrices of the
relations. Let P = [pik], Q = [qkj ] and R = [rij] be membership
matrices of binary relations such that R = P ° Q. We write this
using matrix notation
[rij] = [pik] o [qkj]
where rij = max min (pik, qkj).
k
A similar operation on two binary relations, which differs
from the composition in that it yields triples instead of pairs, is
known as the relational join. For neutrosophic relation PN (X, Y)
and QN (Y, Z) the relational join P * Q corresponding to the
neutrosophic standard max-min composition is a ternary relation
RN (X, Y, Z) defined by RN (x, y, z) = [P * Q]N (x, y, z) = min
[PN (x, y), QN (y, z)] for each x ∈ X, y ∈ Y and z ∈ Z.

This is illustrated by the following Figure 2.2.5.2.

207
x1 .7
x2 .1 .6
.5 a .5 α
.2 .3
x3 1
.8 .7
I
.2
b .2 1 β
x4 .8
c .1
.3 .5 .3 γ
x5 .2 I
.1 d
x6 .1
FIGURE: 2.2.5.2

In addition to defining a neutrosophic binary relation there


exists between two different sets, it is also possible to define
neutrosophic binary relation among the elements of a single set
X.
A neutrosophic binary relation of this type is denoted by
RN(X, X) or RN (X2) and is a subset of X × X = X2.
These relations are often referred to as neutrosophic
directed graphs or neutrosophic digraphs.

Neutrosophic binary relations RN (X, X) can be expressed


by the same forms as general neutrosophic binary relations.
However they can be conveniently expressed in terms of simple
diagrams with the following properties:

I. Each element of the set X is represented by a single


node in the diagram.
II. Directed connections between nodes indicate pairs
of elements of X for which the grade of
membership in R is non zero or indeterminate.
III. Each connection in the diagram is labeled by the
actual membership grade of the corresponding pair
in R or in indeterminacy of the relationship between
those pairs.

208
The neutrosophic membership matrix and the neutrosophic
sagittal diagram is as follows for any set X = {a, b, c, d, e}.

⎡ 0 I 0.3 0.2 0 ⎤
⎢1 0 I 0 0.3⎥⎥

⎢ I 0.2 0 0 0 ⎥.
⎢ ⎥
⎢ 0 0.6 0 0.3 I ⎥
⎢⎣ 0 0 0 I 0.2 ⎥⎦

Neutrosophic membership matrix for x is given above and the


neutrosophic sagittal diagram is given below.

a 1 a
.2
b b
.3 .3
c .2 c
.6 .3
d d

e e
.2
FIGURE 2.2.5.3

Neutrosophic diagram or graph is left for the reader as an


exercise.
The notion of reflexivity, symmetry and transitivity can be
extended for neutrosophic relations RN (X, Y) by defining them
in terms of the membership functions or indeterminacy relation.

Thus RN (X, X) is reflexive if and only if RN (x, x) = 1 for all x ∈


X. If this is not the case for some x ∈ X the relation is
irreflexive.
A weaker form of reflexivity, if for no x in X; RN(x, x) = 1 then
we call the relation to be anti-reflexive referred to as ∈-
reflexivity, is sometimes defined by requiring that
RN (x, x) ≥ ∈ where 0 < ∈ < 1.

209
A fuzzy relation is symmetric if and only if
RN (x, y) = RN (y, x) for all x, y, ∈ X.

Whenever this relation is not true for some x, y ∈ X the relation


is called asymmetric. Furthermore when RN (x, y) > 0 and RN (y,
x) > 0 implies that x = y for all x, y ∈ X the relation R is called
anti-symmetric.
A fuzzy relation RN (X, X) is transitive (or more specifically
max-min transitive) if

RN (x, z) ≥ max min [RN (x, y), RN (y, z)]


y∈Y

is satisfied for each pair (x, z) ∈ X2. A relation failing to satisfy


the above inequality for some members of X is called non-
transitive and if RN (x, x) < max min [RN(x, y), RN(y, z)] for all
y∈Y

(x, x) ∈ X , then the relation is called anti-transitive.


2

Given a relation RN(X, X) its transitive closure R NT (x, X)


can be analyzed in the following way.
The transitive closure on a crisp relation RN (X, X) is
defined as the relation that is transitive, contains

RN (X, X) < max min [RN (x, y) RN (y, z)]


y∈Y

for all (x, x) ∈ X2, then the relation is called anti-transitive.


Given a relation RN (x, x) its transitive closure RNT (X, X) can
be analyzed in the following way.
The transitive closure on a crisp relation RN (X, X) is
defined as the relation that is transitive, contains RN and has the
fewest possible members. For neutrosophic relations the last
requirement is generalized such that the elements of transitive
closure have the smallest possible membership grades, that still
allow the first two requirements to be met.
Given a relation RN (X, X) its transitive closure R NT (X, X)
can be determined by a simple algorithm.
Now we proceed on to define the notion of neutrosophic
equivalence relation.

210
DEFINITION 2.2.5.1: A crisp neutrosophic relation RN(X, X) that
is reflexive, symmetric and transitive is called an neutrosophic
equivalence relation. For each element x in X, we can define a
crisp neutrosophic set Ax which contains all the elements of X
that are related to x by the neutrosophic equivalence relation.
Formally Ax = [y | (x, y) ∈ RN (X, X)}. Ax is clearly a subset
of X. The element x is itself contained in Ax, due to the
reflexivity of R because R is transitive and symmetric each
member of Ax is related to all other members of Ax. Further no
member of Ax is related to any element of X not included in Ax.
This set Ax is clearly referred to as an neutrosophic equivalence
class of RN (X, x) with respect to x. The members of each
neutrosophic equivalence class can be considered neutrosophic
equivalent to each other and only to each other under the
relation R.

But here it is pertinent to mention that in some X; (a, b) may not


be related at all to be more precise there may be an element a ∈
X which is such that its relation with several or some elements
in X \ {a} is an indeterminate. The elements which cannot
determine its relation with other elements will be put in as
separate set.
A neutrosophic binary relation that is reflexive, symmetric
and transitive is known as a neutrosophic equivalence relation.
Now we proceed on to define Neutrosophic intersections
neutrosophic t – norms (tN – norms)
Let A and B be any two neutrosophic sets, the intersection
of A and B is specified in general by a neutrosophic binary
operation on the set N = [0, 1] ∪ I, that is a function of the form

iN: N × N → N.
For each element x of the universal set, this function takes as its
argument the pair consisting of the elements membership grades
in set A and in set B, and yield the membership grade of the
element in the set constituting the intersection of A and B. Thus,

(A ∩ B) (x) = iN [A(x), B(x)] for all x ∈ X.

211
In order for the function iN of this form to qualify as a fuzzy
intersection, it must possess appropriate properties, which
ensure that neutrosophic sets produced by iN are intuitively
acceptable as meaningful fuzzy intersections of any given pair
of neutrosophic sets. It turns out that functions known as tN-
norms, will be introduced and analyzed in this section. In fact
the class of tN- norms is now accepted as equivalent to the class
of neutrosophic fuzzy intersection. We will use the terms tN –
norms and neutrosophic intersections inter changeably.
Given a tN – norm, iN and neutrosophic sets A and B we
have to apply:
(A ∩ B) (x) = iN [A (x) , B (x)]

for each x ∈ X, to determine the intersection of A and B based


upon iN.
However the function iN is totally independent of x, it
depends only on the values A (x) and B(x). Thus we may ignore
x and assume that the arguments of iN are arbitrary numbers
a, b ∈ [0, 1] ∪ I = N in the following examination of formal
properties of tN-norm.
A neutrosophic intersection/ tN-norm iN is a binary operation
on the unit interval that satisfies at least the following axioms
for all a, b, c, d ∈ N = [0, 1] ∪ I.

1N iN (a, 1) = a
2N iN (a, I) = I
3N b ≤ d implies
iN (a, b) ≤ iN (a, d)
4N iN (a, b) = iN (b, a)
5N iN (a, iN(b, d)) = iN (a, b), d).

We call the conditions 1N to 5N as the axiomatic skeleton for


neutrosophic intersections / tN – norms. Clearly iN is a
continuous function on N \ I and iN (a, a) < a ∀a ∈ N \ I

iN (I I) = I.

If a1 < a2 and b1 < b2 implies iN (a1, b1) < iN (a2, b2).

212
Several properties in this direction can be derived as in case
of t-norms.

The following are some examples of tN –norms

1. iN (a, b) = min (a, b)


iN (a, I) = min (a, I) = I will be called as standard
neutrosophic intersection.
2. iN (a, b) = ab for a, b ∈ N \ I and iN (a, b) = I for a, b ∈
N where a = I or b = I will be called as the neutrosophic
algebraic product.
3. Bounded neutrosophic difference.
iN (a, b) = max (0, a + b – 1) for a, b ∈ I
iN (a, I) = I is yet another example of tN – norm.

1. Drastic neutrosophic intersection


2.
⎧a when b = 1
⎪b when a = 1

⎪⎪I when a = I
iN (a, b) = ⎨
⎪ or b = I
⎪ or a = b = I

⎪⎩0 otherwise.

As I is an indeterminate adjoined in tN – norms. It is not easy to


give then the graphs of neutrosophic intersections. Here also we
leave the analysis and study of these tN – norms (i.e.
neutrosophic intersections) to the reader.
The notion of neutrosophic unions closely parallels that of
neutrosophic intersections. Like neutrosophic intersection the
general neutrosophic union of two neutrosophic sets A and B is
specified by a function

µN: N × N → N where N = [0, 1] ∪ I.

The argument of this function is the pair consisting of the


membership grade of some element x in the neutrosophic set A

213
and the membership grade of that some element in the
neutrosophic set B, (here by membership grade we mean not
only the membership grade in the unit interval [0, 1] but also the
indeterminacy of the membership). The function returns the
membership grade of the element in the set A ∪ B.
Thus (A ∪ B) (x) = µN [A (x), B(x)] for all x ∈ X.
Properties that a function µN must satisfy to be initiatively
acceptable as neutrosophic union are exactly the same as
properties of functions that are known. Thus neutrosophic union
will be called as neutrosophic t-co-norm; denoted by tN – co-
norm.
A neutrosophic union / tN – co-norm µN is a binary
operation on the unit interval that satisfies at least the following
conditions for all a, b, c, d ∈ N = [0, 1] ∪ I

C1 µN (a, 0) = a
C2 µN (a, I) = I
C3 b ≤ d implies
µN (a, b) ≤ µN (a, d)
C4 µN (a, b) = µN (b, a)
C5 µN (a, µN (b, d))
= µN (µN (a, b), d).

Since the above set of conditions are essentially neutrosophic


unions we call it the axiomatic skeleton for neutrosophic unions
/ tN-co-norms.

The addition requirements for neutrosophic unions are

i. µN is a continuous functions on N \ {I}


ii. µN (a, a) > a.
iii. a1 < a2 and b1 < b2 implies µN (a1. b1) < µN (a2, b2);
a1, a2, b1, b2 ∈ N \ {I}.

We give some basic neutrosophic unions.

Let µN : [0, 1] × [0, 1] → [0, 1]

µN (a, b) = max (a, b)

214
µN (a, I) = I is called as the standard
neutrosophic union.
µN (a, b) = a + b – ab and
µN (a, I) = I .

This function will be called as the neutrosophic algebraic sum.

µN (a, b) = min (1, a + b) and µN (a, I) = I

will be called as the neutrosophic bounded sum. We define the


notion of neutrosophic drastic unions

⎧a when b = 0
⎪b when a = 0
⎪⎪
µN (a, b) = ⎨I when a = I
⎪ or b = I

⎪⎩1 otherwise.

Now we proceed on to define the notion of neutrosophic


Aggregation operators. Neutrosophic aggregation operators on
neutrosophic sets are operations by which several neutrosophic
sets are combined in a desirable way to produce a single
neutrosophic set.
Any neutrosophic aggregation operation on n neutrosophic
sets (n ≥ 2) is defined by a function hN: Nn → N where N = [0,
1] ∪ I and Nn = N × ... × N when applied to neutrosophic sets
n − times
A1, A2,…, An defined on X the function hN produces an
aggregate neutrosophic set A by operating on the membership
grades of these sets for each x ∈ X (Here also by the term
membership grades we mean not only the membership grades
from the unit interval [0, 1] but also the indeterminacy I for
some x ∈ X are included). Thus
AN (x) = hN (A1 (x), A2 (x),…, An(x))
for each x ∈ X.

215
2.3 Special Fuzzy Cognitive Models and their
Neutrosophic Analogue

In this section we define five types of special fuzzy models and


their neutrosophic analogue. This is the first time such models
are defined. They are different from the usual combined fuzzy
models like combined fuzzy cognitive maps, combined fuzzy
relational maps and so on. These models helps not only in easy
comparison also it gives a equal opportunity to study the
opinion of several experts that is why these models can be
thought of as multi expert models preserving the individual
opinion even after calculations of the resultant. Thus all special
fuzzy models are multi expert models as well as a multi models
for the same model can at the same time use a maximum of four
different models in the study. Thus this is a special feature of
these special fuzzy models. We also give the special
neutrosophic analogue of them. For in many problem we may
not have a clear cut feeling i.e., the expert may not be in a
position to give his /her opinion it can also be an indeterminate.
When the situation of indeterminacy arises fuzzy models cannot
play any role only the neutrosophic models can tackle the
situation. We also build a special mixed models which will be
both having neutrosophic models as well as fuzzy models in
their dynamical system.
In this section we define for the first time a new type of
Special Fuzzy Cognitive Models (SFCM) and their
neutrosophic analogue which is defined as Special Neutrosophic
Cognitive Models (SNCM). Further we define Special Fuzzy
and Neutrosophic Cognitive Maps (models) (SFNCM).
We illustrate them with examples from real world problems.
It is pertinent to mention here that we give only illustrative
examples not always any specific study. Now we proceed on to
define the notion of special fuzzy cognitive maps models.

DEFINITION 2.3.1: Suppose we have some m experts working


on a problem P and suppose all of them agree to work with the
same set of attributes say n attributes using only the Fuzzy
Cognitive Maps then this gives a multiexpert model called the
Special Fuzzy Cognitive Model, and if Mi denotes the fuzzy

216
connection matrix given by the ith expert using the set of n
attributes i = 1, 2, …, m then we call the special fuzzy square
matrix M = M1 ∪ M2 ∪ … ∪ Mm to be the special fuzzy
connection matrix associated with the Special Fuzzy Cognitive
Map (model) (SFCM).
Let C11 C21 … Cn1 , C12 C22 … Cn2 , … , C1m C2m … Cnm be the
special nodes of the SFCM
A = ⎡⎣ a11 a12 … a1n ⎤⎦ ∪ ⎡⎣ a12 a22 … an2 ⎤⎦ ∪ … ∪ ⎡⎣ a1m a2m … anm ⎤⎦
where a ij ∈ {0, 1}, A is called the instantaneous special state
vector and denotes the on-off position of the special node at an
instant
a ij = 0 if a ij is off and
= 1 if a ij is on I = 1, 2, …, m and j = 1, 2, …, n.
The attributes or nodes of a SFCM will be known as the special
nodes or special attributes; context wise one may use just nodes
or attributes.

The main advantage of using this model is that it can work


simultaneously at one stretch and give opinion of all the m
experts. This is clearly a multi expert model.
Now we shall illustrate the functioning of this model.
Suppose M = M1 ∪ M2 ∪ … ∪ Mm be a n × n special fuzzy
square matrix associated with the SFCM for the given problem
P i.e. M is the multiexpert special fuzzy model. Suppose we
want to study the effect of X = X1 ∪ X2 ∪ … ∪ Xm where X is
a special fuzzy row state vector where each Xi is a 1 × n fuzzy
row state vector suggested by the ith expert 1 ≤ i ≤ m with
entries from the set {0, 1} i.e. the special fuzzy row vector gives
the on or off states of the attributes associated with the problem
P that is under study. Now the effect of X on M using the
special operator described in pages 20-1 of this book is given by

XoM = (X1 ∪ X2 ∪ … ∪ Xm) o (M1 ∪ M2 ∪ … ∪ Mm)


= X1 o M1 ∪ X2 o M2 ∪ … ∪ Xm o Mm
= Y'1 ∪ Y'2 ∪ … ∪ Y'm
= Y';

217
Y' may or may not be a special fuzzy row vector with entries
from the set {0, 1}. It may also happen that the elements in Y'i;
1 ≤ i ≤ m may not belong to the set {0, 1}. Further the nodes
which were in the on state in X may not be in the on state in Y′.
To overcome all these problems we update and threshold Y′ to
Y = Y1 ∪ Y2 ∪ … ∪ Ym. Now each Yi has its entries from the
set {0, 1} and those nodes which were in the on state in X
remain to be in the on state in Y also. Now we find

YoM = (Y1 ∪ Y2 ∪ … ∪ Ym) o (M1 ∪ M2 ∪ … ∪ Mm)


= Y1 o M1 ∪ Y2 o M2 ∪ … ∪ Ym o Mm
= Z1′ ∪ Z2′ ∪ … ∪ Zm′
= Z′;

now Z′ may not be a special fuzzy row vector so we threshold


and update Z′ to Z = Z1 ∪ Z2 ∪ … ∪ Zm and now proceed on to
find Z o M, we continue this process until we arrive at a special
fixed point or a special limit cycle or a special fixed points and
limit cycles. This final resultant special fuzzy row vector will be
known as the special hidden pattern of the SFCM model.

We illustrate this by a simple real world model.

Example 2.3.1: Suppose some 5 experts are interested to study


the problem, the nation will face due to the production of more
number engineering graduates which is very disproportionate to
the job opportunities created by the nation (India) for them.
All the five experts wish to work with the five concepts
relating the unemployed engineering graduates.

E1 - Frustration
E2 - Unemployment
E3 - Increase of educated criminals
E4 - Under employment
E5 - Taking up drugs, alcohol etc.

The special fuzzy square connection matrix related with the


SFCM model given by the 5 experts be denoted by

218
M = M1 ∪ M2 ∪ M3 ∪ M4 ∪ M5

E1 E 2 E3 E 4 E 5 E1 E 2 E3 E 4 E 5
E1 ⎡ 0 1 1 1 0⎤ E1 ⎡ 0 1 0 1 0⎤
E 2 ⎢⎢1 0 1 0 1⎥ ⎥ E ⎢1 0 1 0 1 ⎥⎥
∪ 2⎢
E 3 ⎢1 1 0 1 0⎥ E3 ⎢0 1 0 1 0⎥
⎢ ⎥ ⎢ ⎥
E 4 ⎢1 0 0 0 1⎥ E 4 ⎢1 0 0 0 1⎥
E 5 ⎢⎣1 1 0 0 0 ⎥⎦ E 5 ⎢⎣ 0 1 1 0 0 ⎥⎦

E1 E 2 E3 E 4 E 5
E1 ⎡ 0 1 0 0 1⎤
E ⎢1 0 1 0 0 ⎥⎥
∪ 2⎢ ∪
E 3 ⎢1 0 0 0 1⎥
⎢ ⎥
E 4 ⎢0 0 1 0 1⎥
E 5 ⎢⎣1 0 0 1 0 ⎥⎦

E1 E 2 E3 E 4 E 5 E1 E 2 E3 E 4 E 5
E1 ⎡ 0 0 1 0 1⎤ E1 ⎡ 0 0 0 1 1⎤
E 2 ⎢⎢ 0 0 0 1 1⎥ ⎥ E 2 ⎢⎢1 0 0 0 1 ⎥⎥
∪ .
E3 ⎢0 1 0 0 1⎥ E3 ⎢0 0 0 1 1⎥
⎢ ⎥ ⎢ ⎥
E 4 ⎢0 1 1 0 1⎥ E 4 ⎢1 1 0 0 0⎥
E 5 ⎢⎣1 0 1 0 0 ⎥⎦ E 5 ⎢⎣ 0 0 1 1 0 ⎥⎦

Clearly M is a special fuzzy square matrix. Now we want to


study the effect of unemployment alone in the on state by all the
5 experts and all other nodes are in, the off state i.e. X = [0 1 0 0
0] ∪ [0 1 0 0 0] ∪ [0 1 0 0 0] ∪ [0 1 0 0 0] ∪ [0 1 0 0 0]. X is a
special fuzzy row state vector. To find the special hidden pattern
associated with X using the SFCM. Now

XoM = (X1 ∪ X2 ∪ … ∪ X5) o (M1 ∪ M2 ∪ … ∪ M5)


= X1 o M1 ∪ X2 o M2 ∪ … ∪ X5 o M5

219
⎡0 1 1 1 0⎤
⎢1 0 1 0 1 ⎥⎥

= [ 0 1 0 0 0] o ⎢1 1 0 1 0⎥ ∪
⎢ ⎥
⎢1 0 0 0 1⎥
⎢⎣1 1 0 0 0 ⎥⎦

⎡0 1 0 1 0⎤
⎢1 0 1 0 1 ⎥⎥

[ 0 1 0 0 0] o ⎢ 0 1 0 1 0⎥ ∪
⎢ ⎥
⎢1 0 0 0 1⎥
⎢⎣0 1 1 0 0 ⎥⎦

⎡0 1 0 0 1⎤
⎢1 0 1 0 0 ⎥⎥

[ 0 1 0 0 0] o ⎢1 0 0 0 1⎥ ∪
⎢ ⎥
⎢0 0 1 0 1⎥
⎢⎣1 0 0 1 0 ⎥⎦

⎡0 0 1 0 1⎤
⎢0 0 0 1 1 ⎥⎥

[ 0 1 0 0 0] o ⎢ 0 1 0 0 1⎥ ∪
⎢ ⎥
⎢0 1 1 0 1⎥
⎢⎣1 0 1 0 0 ⎥⎦

⎡0 0 0 1 1⎤
⎢1 0 0 0 1 ⎥⎥

[ 0 1 0 0 0] o ⎢ 0 0 0 1 1⎥
⎢ ⎥
⎢1 1 0 0 0⎥
⎢⎣ 0 0 1 1 0 ⎥⎦

= [1 0 1 0 1] ∪ [1 0 1 0 1] ∪ [1 0 1 0 0] ∪ [0 0 0 1 1]

220
∪ [1 0 0 0 1]
= Y'1 ∪ Y'2 ∪ Y'3 ∪ Y'4 ∪ Y'5
= Y′;
we update and threshold Y' to

Y = Y1 ∪ Y2 ∪ Y3 ∪ Y4 ∪ Y5
= [1 1 1 0 1] ∪ [1 1 1 0 1] ∪ [1 1 1 0 0] ∪ [0 1 0 1 1]
∪ [1 1 0 0 1].
Now we find the effect of Y on M i.e.

YoM = (Y1 ∪ Y2 ∪ … ∪ Y5) o (M1 ∪ M2 ∪ … ∪ M5)


= Y1 o M1 ∪ Y2 o M2 ∪ … ∪ Y5 o M5
= [3 3 2 2 1] ∪ [1 3 2 2 1] ∪ [2 1 1 0 2] ∪ [1 1 2 1 2]
∪ [1 0 1 2 2]
= Z'1 ∪ Z'2 ∪ Z'3 ∪ Z'4 ∪ Z'5
= Z';
we update and threshold Z' to obtain

Z = Z1 ∪ Z2 ∪ Z3 ∪ Z4 ∪ Z5
= [1 1 1 1 1] ∪ [1 1 1 1 1] ∪ [1 1 1 0 1] ∪ [1 1 1 1 1]
∪ [1 1 1 1 1].

Now we find the effect of Z on M i.e.

ZoM = (Z1 ∪ Z2 ∪ … ∪ Z5) o (M1 ∪ M2 ∪ … ∪ M5)


= Z1 o M1 ∪ Z2 o M2 ∪ … ∪ Z5 o M5
= [4 3 2 2 2] ∪ [2 3 2 2 2] ∪ [3 1 1 1 2] ∪ [1 2 3 1 4]
∪ [2 1 1 3 3]
= P'1 ∪ P'2 ∪ P'3 ∪ P'4 ∪ P'5
= P'.
Clearly P' is not a special fuzzy row vector; so update and
threshold P' to obtain

P = P1 ∪ P2 ∪ … ∪ P5
= [1 1 1 1 1] ∪ [1 1 1 1 1] ∪ [1 1 1 1 1] ∪ [1 1 1 1 1]
∪ [1 1 1 1 1]

221
Thus unemployment according to all experts leads to all
other problems. All experts agree upon it. We can for instance
find the effect of the special state vector like

X = [1 0 0 0 0] ∪ [0 1 0 0 0] ∪ [0 0 1 0 0] ∪ [0 0 0 1 0] ∪ [0
0 0 0 1]

where the first experts wishes to study on state of first node


frustration, second expert the on state of the second node
unemployment, the third expert the on state on the third node
viz. increase of educated criminals and the forth expert under
employments and taking up drugs alcohol etc. is taken up by the
fifth expert.

XoM = (X1 ∪ X2 ∪ … ∪ X5) o (M1 ∪ M2 ∪ … ∪ M5)


= X1 o M ∪ X2 o M2 ∪ … ∪ X5 o M5
= [0 1 1 1 0] ∪ [1 0 1 0 1] ∪ [1 0 0 0 1] ∪ [0 1 1 0 1]
∪ [0 0 1 1 0]
= Y'1 ∪ Y'2 ∪ Y'3 ∪ Y'4 ∪ Y'5
= Y'
after updating and thresholding Y' we get

Y = Y1 ∪ Y2 ∪ … ∪ Y5
= [1 1 1 1 0] ∪ [1 1 1 0 1] ∪ [1 0 1 0 1] ∪ [0 1 1 1 1]
∪ [0 0 1 1 1].
One can find Y o M and so on until one gets the special
hidden pattern.
This we have just mentioned for the reader should to know
the expert can take any special node or special nodes to be in the
on state for the study and this gives a single solution of all the
experts. The two main advantages over other multi expert
systems are

1. Stage wise comparison is possible for the same state


special vector by all experts or stage wise comparison
for different state special vector by different experts.
2. The working is not laborious as a C-program will do the
job in no time.

222
Next we proceed on to define the notion of Special Mixed
Fuzzy Cognitive maps/models (SMFCMs). This model comes
in handy when experts have different sets of attributes and the
number of attributes are also varying in number.

DEFINITION 2.3.2: Suppose we have some t experts working on


a problem P and suppose they do not agree to work on the same
set of attributes or the same attributes, they choose different sets
of attributes but they all agree to work with FCM model then we
can use the Special Mixed Fuzzy Cognitive Maps (SMFCMs)
model and get a new dynamical system which can cater to the
opinion of each and every expert simultaneously. Suppose the ith
expert works with ni attributes and the fuzzy connection matrix
associated with the FCM be given by Mi, i = 1, 2, …, t, we use
these t experts to get a special fuzzy mixed square matrix M =
M1 ∪ M2 ∪ … ∪ Mt where M corresponds to the dynamical
system which gives the opinion of all t experts and M is called
the special fuzzy connection matrix and the model related with
this M will be known as the Special Mixed Fuzzy Cognitive
Maps (SMFCMs) model associated with t experts.

Thus we see M is also a multi expert system model. It is


important to mention here that SMFCM the multi expert FCM
model is different from the Combined FCM (CFCM[108, 112])
as well as Special Fuzzy Cognitive Maps (SFCM) model. The
salient features of SMFCM model is

1. SMFCM is better than the CFCM model as the CFCM


model can function only with the same set of attributes for
the problem P.
2. Here in the SMFCM contradictory opinions do not cancel
out as in case of CFCMs.
3. In the SMFCM the opinion of each and every expert is
under display and see their opinion not as CFCMs which
gives only a collective resultant.
4. SMFCM model is better than the SFCM model as SFCM
can permit all the expert to work only with the same set of
attributes for the given problem.

223
We give an example of SMFCMs as a model for the reader to
understand this new model.

Example 2.3.2: We here give the prediction of electoral winner


or how the preference of a particular politician and so on.
Suppose we have four experts working on the problem and the
first expert wishes to work with the four attributes.

e11 - Language of the politician


1
e2 - Community of the politician
1
e3 - Service to people, public figure configuration,
personality and nature
e14 - Party’s strength and the opponents strength.

Let M1 be the fuzzy connection matrix given by the first expert,


i.e.,

e11 e12 e13 e14


e11 ⎡0 1 1 0⎤
M1 = e12 ⎢1 0 1 0 ⎥⎥ .

e13 ⎢0 0 0 1⎥
⎢ ⎥
e14 ⎣1 1 1 0⎦

Suppose the second expert wishes to work with 5 attributes


given by

e12 - Service done by the politician to people


2
e 2 - Finance and media accessibility
2
e 3 - Party’s strength and opponents strength
2
e 4 - Working member of the party
2
e 5 - His community and the locals community.

Let M2 be the fuzzy connection matrix associated with his


opinion.

224
e12 e 22 e32 e 24 e52
e12 ⎡0 1 1 0 0⎤
⎢1 0 ⎥⎥
e2 ⎢ 0 1 0
M2 = 22 .
e3 ⎢0 0 0 1 1⎥
⎢ ⎥
e 24 ⎢1 1 1 0 1⎥
e52 ⎢⎣0 0 1 1 0 ⎥⎦

Suppose the third expert wishes to work with only four


attributes given by

e13 - Community and nativity of the politician


3
e 2 - His interest for working with people and their
opinion on his personality
e33 - Amount of money he can spend on propaganda
in media
e34 - Working members of the party i.e., public mass
support.

Let M3 be the connection matrix given by the third expert using


FCM model.
e13 e32 e33 e34
e13 ⎡0 1 1 0⎤
M3 = e32 ⎢1 0 1 0 ⎥⎥ .

e33 ⎢1 0 0 1⎥
⎢ ⎥
e34 ⎣1 1 0 0⎦

Let e14 ,e 42 , … , e64 be the attributes given by the fourth expert.

e14 - Community, nativity and gender of the


politician
e 42 - The constructive and progressive work done by
him in his locality

225
e34 - His social and economic status
4
e 4 - Support of locals in propagating for his winning
the election
e54 - Money he spends on propaganda for election
4
e 6 - The strength of the party’s campaigning in
favour of him.

Let M4 denote connection matrix given by the fourth expert.

e14 e 42 e34 e44 e54 e64


e14 ⎡ 0 1 1 1 0 1⎤
⎢ 1 ⎥⎥
e 42 ⎢1 0 0 1 0
M4 = e34 ⎢ 0 0 0 0 1 1⎥ .
⎢ ⎥
e 44 ⎢1 1 0 0 0 1⎥
e54 ⎢ 0 0 1 1 0 1⎥
⎢ ⎥
e64 ⎢⎣ 0 1 1 1 1 0 ⎥⎦

Now let M = M1 ∪ M2 ∪ M3 ∪ M4 give the SMFCM which is


the multi expert opinion of the four experts. Now suppose the
four experts wish to work on the special state vector

X = [1 0 0 0] ∪ [0 1 0 0 0] ∪ [1 0 0 0] ∪ [1 0 0 0 0 0]

where the first expert wants to work with language of the


politician alone to be in the on state and the second expert wants
to work with the node finance and media accessibility alone in
the on state and all the other nodes to be in the off state; the
third expert wants to work with the node community and
nativity alone in the on state and all other nodes to be in the off
state and the forth expert wants to work with the node
community, nativity and the gender of the politician alone in the
on state. To find the special hidden pattern of X on the special
dynamical system,

M = M1 ∪ M2 ∪ M3 ∪ M4

226
⎡0 1 1 0 0⎤
⎡0 0⎤ ⎢1 0 ⎥⎥
1 1
⎢1 0 1 ⎥
0⎥ ⎢ 0 1 0
= ⎢ ∪ ⎢0 0 0 1 1⎥ ∪
⎢0 0 0 1⎥ ⎢ ⎥
⎢ ⎥ ⎢1 1 1 0 1⎥
⎣1 1 1 0⎦
⎢⎣ 0 0 1 1 0 ⎥⎦

⎡0 1 1 1 0 1⎤
⎢1 1 ⎥⎥
⎡0 1 1 0⎤ ⎢ 0 0 1 0
⎢1 0 1 1 ⎥⎥ ⎢0 0 0 0 1 1⎥
⎢ ∪ ⎢ ⎥.
⎢1 0 0 1⎥ ⎢1 1 0 0 0 1⎥
⎢ ⎥ ⎢0
⎣1 1 0 0⎦ 0 1 1 0 1⎥
⎢ ⎥
⎢⎣ 0 1 1 1 1 0 ⎥⎦

Now
XoM = X1 o M1 ∪ X2 o M2 ∪ X3 o M3 ∪ X4 o M4

⎡0 1 1 0⎤
⎢1 0 1 0 ⎥⎥
= [1 0 0 0] ⎢
⎢0 0 0 1⎥
⎢ ⎥
⎣1 1 1 0⎦

⎡0 1 1 0 0⎤
⎢1 0 1 0 0 ⎥⎥

∪ [ 0 1 0 0 0] ⎢0 0 0 1 1⎥
⎢ ⎥
⎢1 1 1 0 1⎥
⎢⎣0 0 1 1 0 ⎥⎦

⎡0 1 1 0⎤
⎢1 0 1 0 ⎥⎥
∪ [1 0 0 0] ⎢
⎢1 0 0 1⎥
⎢ ⎥
⎣1 1 0 0⎦

227
⎡0 1 1 1 0 1⎤
⎢1 0 0 1 0 1 ⎥⎥

⎢0 0 0 0 1 1⎥
∪ [1 0 0 0 0 0] ⎢ ⎥
⎢1 1 0 0 0 1⎥
⎢0 0 1 1 0 1⎥
⎢ ⎥
⎢⎣0 1 1 1 1 0 ⎥⎦

= [0 1 1 0] ∪ [1 0 1 0 0] ∪ [0 1 1 0] ∪ [0 1 1 1 0 1]
= Y'1 ∪ Y'2 ∪ Y'3 ∪ Y'4
= Y'.

After updating and thresholding Y' we get

Y = Y1 ∪ Y2 ∪ Y3 ∪ Y4
= [1 1 1 0] ∪ [1 1 1 0 0] ∪ [1 1 1 0] ∪ [1 1 1 1 0 1]
= Y.

Clearly Y is a special fuzzy row vector.


Now we find the effect of Y on the special dynamical
system M and get Z'.

YoM = [1 1 2 1] ∪ [1 1 2 1 1] ∪ [2 1 2 1] ∪ [2 3 2 3 2 4]
= Z'1 ∪ Z'2 ∪ Z'3 ∪ Z'4 ∪ Z'5
= Z'.

Let Z be the resultant vector got by updating and thresholding


Z'. i.e.,

Z = [1 1 1 1] ∪ [1 1 1 1 1] ∪ [1 1 1 1] ∪ [1 1 1 1 1 1].

It is left for the reader to verify the special hidden pattern in a


fixed point given by the special fuzzy row vector

Z = [1 1 1 1] ∪ [1 1 1 1 1] ∪ [1 1 1 1] ∪ [1 1 1 1 1 1].

Now we proceed on to define the new special neutrosophic


fuzzy cognitive model in which multi experts opinion are given.

228
DEFINITION 2.3.3: Suppose we have some n experts interested
in working with a problem P and all of them agree to work with
the same m number of attributes related with the problem and
further all of them agree to work with the neutrosophic
cognitive maps(NCMs) model; i.e., all experts feel that some of
the interrelation between the attributes is an indeterminate.
Suppose let the ith expert using the NCMs give the neutrosophic
connection matrix to be the Ni, i = 1, 2, …, n. Thus each Ni is a
m × m neutrosophic matrix and the multiexpert Special
Neutrosophic Cognitive Maps (SNCMs) model is denoted by N,
which is a special fuzzy neutrosophic square matrix and is given
by N = N1 ∪ N2 ∪ … ∪ Nn. Thus N is a special neutrosophic
connection square matrix associated with the SNCMs model.
We just mention how it works. Suppose all the n experts
give their preference of node which is to be in the on state then
the special fuzzy row vector given by them collectively be
denoted by X = X1 ∪ X2 ∪ … ∪ Xn where Xi is the special state
vector given by the ith expert; i = 1, 2, …, n. Now the effect of X
on the special fuzzy dynamical system N is given by

XoN = (X1 ∪ X2 ∪ … ∪ Xn) o (N1 ∪ N2 ∪ … ∪ Nn)


= X1 o N1 ∪ X2 o N2 ∪ … ∪ Xn o Nn
= Y'1 ∪ Y'2 ∪ … ∪Y'n
= Y'.

The working is described in page 139-141 of chapter 1. Now


after updating and thresholding Y' we get Y = Y1 ∪ Y2 ∪ … ∪
Yn. Now we find the effect of Y on N i.e.,

YoN = (Y1 ∪ Y2 ∪ … ∪ Yn) o (N1 ∪ N2 ∪ … ∪ Nn)


= Y1 o N1 ∪ Y2 o N2 ∪ … ∪ Yn o Nn
= Z'1 ∪ Z'2 ∪ … Z'n
= Z'.
We see Z' may or may not be a special fuzzy neutrosophic row
vector and further the nodes which we started to work with may
not be in the on state so we update and threshold Z' to Z and let
Z = Z1 ∪ Z2 ∪ … ∪ Zn. Clearly Z is a special fuzzy neutrosophic
row vector. We now find Z o N and so on till we arrive at a fixed
point or a limit cycle (say) T. This T will be known as the

229
special hidden pattern of the special dynamical neutrosophic
system N.

Now we see this new Special Neutrosophic Cognitive Maps


(SNCMs) model has the following advantages.

1. When all the experts want to make use of the element of


indeterminacy it serves the purpose.
2. This is also a multi expert model.

Now the reader can construct real world problems multi expert
SNCM model to study the results as the working is analogous to
the SFCM model described in definition 2.3.1 of this chapter.
Now we proceed on to define yet another new multi expert
special neutrosophic model.

DEFINITION 2.3.4: Let us take n experts who wishes to work


with the same problem P. But the experts have different sets of
attributes which they want to work with, but all of them agree to
work with the neutrosophic cognitive maps model i.e., they all
uniformly agree upon the fact that they should choose their own
attributes which involve some degrees of indeterminacy. Now
we give a multi expert model to solve this problem. Let the ith
expert choose to work with ni attributes using a NCM model and
let Ni be the neutrosophic connection matrix given by him; this
is true i = 1, 2, …, m. Let us take N = N1 ∪ N2 ∪ …∪ Nm .
Clearly N is a special fuzzy neutrosophic mixed square matrix.
This N will be known as the special dynamical system
associated with the problem and this special dynamical system
is defined as the Special Mixed Neutrosophic Cognitive Maps
(SMNCMs ) model.
We must give the functioning of the SMNCMs model. Let N
be the special fuzzy neutrosophic connection matrix associated
with the SMNCMs model. Let each of the expert(say ith expert)
give Xi to be the state row vector with which he wishes to work;
for i = 1, 2, …, m. Then the special fuzzy mixed row vector
denoted by X = X1 ∪ X2 ∪ … ∪ Xm will be the integrated
special fuzzy mixed row vector of all the experts whose effect we

230
want to study and find its resultant on the special neutrosophic
dynamical system N. Now

XoN = (X1 ∪ X2 ∪ … ∪ Xm) o (N1 ∪ N2 ∪ … ∪ Nm)


= X1 o N1 ∪ X2 o N2 ∪ … ∪ Xm o Nm.

(This operation is defined in chapter one, pages 146-151 of this


book). Let
X o N = Y'1 ∪ Y'2 ∪ … ∪ Y'm
= Y'.

Y' may or may not be a special fuzzy neutrosophic mixed row


vector so we update and threshold Y' and obtain Y = Y1 ∪ Y2 ∪
… ∪ Yn. Now we find the effect of Y on N.

XoN = (Y1 ∪ Y2 ∪ … ∪ Ym) o (N1 ∪ N2 ∪ … ∪ Nm)


= Y1 o N1 ∪ Y2 o N2 ∪ … ∪ Ym o Nm
= Z'1 ∪ Z'2 ∪ … ∪ Z'm
= Z' (say);

Z' may or may not be a special fuzzy neutrosophic mixed row


vector, we update and threshold Z' to obtain Z = Z1 ∪ Z2 ∪ …
∪ Zn. We find Z o N and so on till we obtain a fixed point or a
limit cycle. This fixed point or the limit cycle will be known as
the special fixed point or the special limit cycle and is defined
as the special hidden pattern of the special dynamical system
SMNCM.

The main advantage of this multi expert model SMNCMs are :

1. One can obtain simultaneously every expert opinion and


compare them stage by stage also.
2. SMNCMs is the dynamical system which gives the
special hidden pattern.
3. SMNCMs is best suited when the relations between
attributes involves indeterminacy.
4. SMNCMs gives the liberty to the experts to choose any
desired number of attributes which he/she chooses to
work with.

231
5. SMNCMs are better than SNCMs, SFCMs and
SMFCMs when the problem involves indeterminacy
and gives liberty to the experts to use any desired
number of attributes to work with the same problem.

Now the reader is expected to apply SMNCM model in any of


the multi expert problem and find the special hidden pattern.
The working is exactly analogous to the example 2.3.2 given in
this book.
Now we proceed on to define a new multi expert model
which is different from these four models known as Special
Fuzzy Neutrosophic Cognitive Maps model (SFNCMs-model).

DEFINITION 2.3.5: Suppose we have a problem P for which we


want to obtain a multi experts opinion. Suppose we have two
sets of experts and one set of experts are interested in only using
Fuzzy Cognitive Maps models and another set of experts keen
on using Neutrosophic Cognitive Maps models for the same
problem P and experts from both the sets demand for different
sets of attributes to be fixed by them for the problem P. Suppose
we have t1 experts who want to use FCM model with different
sets of attributes then, we use SMFCMs to model them say M1
gives the special connection matrix associated with SMFCMs of
the t1 experts, the t2 experts who choose to work with NCM
model with varying attributes is modelled using the SMNCMs
model. Let M2 give the special connection matrix of the
SMNCMs of the t2 experts. Now we define M1∪ M2 to be the
combined special connection matrix of both the models and
M1 ∪ M2= ( M 11 ∪ M 21 ∪ … ∪ M t11 ) ∪ ( M 12 ∪ M 22 ∪ …∪ M t22 )
to be the associated special connection matrix of the new
Special Fuzzy Neutrosophic Cognitive Maps (SFNCMs) model.
M1 ∪ M2 will be known as the special combined fuzzy
neutrosophic mixed square matrix.
We just give the working of the SFNCMs model. Suppose t1
experts give their on state of their nodes with which they wish to
work as X1 = X 11 ∪ X 21 ∪ … ∪ X t11 and t2 experts give their
preference nodes as X2 = X 12 ∪ X 22 ∪ … ∪ X t22 then the

232
special state vector for which we have to find the special hidden
pattern is given by
X1 ∪ X 2 = ( X 12 ∪ X 22 ∪ … ∪ X t21 )
( X 12 ∪ X 22 ∪ … ∪ X t22 )
using the SFNCMs dynamical system M 1 ∪ M 2 . The effect of
X 1 ∪ X 2 on M 1 ∪ M 2 is given by

(X1 ∪ X 2 ) o ( M 1 ∪ M 2 )
= (X1 ∪ X 2 ) o ( M 1 ∪ M 2 )
= X1 o M 1 ∪ X 2 o M 2
= ( X 21 ∪ X 21 ∪ … ∪ X t11 ) o ( M 11 ∪ M 21 ∪ … ∪ M t11 )∪
( X 12 ∪ X 22 ∪…∪ X t22 ) o ( M 12 ∪ M 22 ∪ … ∪ M t22 )
= X 11 o M 11 ∪ X 21 o M 21 ∪ X 31 o M 31 ∪ … ∪
X t11 o M t11 ∪ X 12 o M 12 ∪ X 22 o M 22 ∪ X 32 o M 32
∪ … ∪ X t2 o M t22 2

= ( Z ∪ Z ∪ … ∪ Z t21 )∪ ( T12 ∪ T22 ∪ … ∪ Tt22 )


1
2 2
2

= Z' ∪ T '.

1. Z' ∪ T' may or may not be a special fuzzy neutrosophic


mixed row vector
2. Z' ∪ T' may have the nodes with which started to work
with to be in the off state so Z' ∪ T' is thresholded and
updated to (say) P1 ∪ Q2 = ( P11 ∪ P21 ∪ … ∪ Pt11 ) ∪
( Q12 ∪ Q22 ∪ … ∪ Qt22 ).

Now we find the effect of P1 ∪ Q2 on the special dynamical


system M1 ∪ M2 i.e.,
(P1 ∪ Q2) o ( M 1 ∪ M 2 ) = P1 o M1 ∪ Q2 o M2
= ( P11 ∪ P21 ∪ … ∪ Pt11 ) o ( M 11 ∪ M 21 ∪ … ∪ M t11 ) ∪
( P12 ∪ P22 ∪ … ∪ Pt22 ) o ( M 12 ∪ M 22 ∪ … ∪ M t22 )
= ( V11 ∪ V21 ∪ … ∪ Vt11 ) ∪ ( W12 ∪ W22 ∪ … ∪ Wt22 )

233
= V2 ∪ W2 .

This resultant may not be even a special fuzzy neutrosophic


mixed row vector so we update and threshold V2 ∪ W2 to S1 ∪
S2 = ( S11 ∪ S 21 ∪ … ∪ St11 ) ∪ ( S12 ∪ S 22 ∪ … ∪ St22 ) we proceed
on to find (S1 ∪ S2) o (M1 ∪ M2) so on until we arrive at a
special fixed point or a special limit cycle. This resultant vector
will correspond to the special hidden pattern of the special
dynamical system SFNFCM.

Now we will illustrate this situation by an explicit model.

Example 2.3.3: Let the four experts work with the problem of
finding prediction of electoral winner using SFNFCM. Suppose
the first expert works with four attributes [for the attributes refer
example 2.3.2] and the fuzzy connection matrix given by him is

⎡0 1 1 0⎤
⎢1 0 1 0 ⎥⎥
M1 = ⎢ .
⎢0 0 0 1⎥
⎢ ⎥
⎣1 1 1 0⎦

The second expert uses the NCM and the connection matrix
given by him using the 5 attributes given in the example 2.3.2 is
given below
⎡0 1 1 0 0⎤
⎢1 0 1 1 0 ⎥
⎢ ⎥
M2 = ⎢ I 0 0 0 1⎥ ,
⎢ ⎥
⎢1 0 I 0 0 ⎥
⎢⎣ 0 0 1 1 0 ⎥⎦

which is neutrosophic model of the NCM. The third expert


works with four attributes and gives the following NCM model
which is a neutrosophic 4 × 4 matrix given by M3; i.e.,

234
⎡0 0 1 0⎤
⎢1 1 I 0 ⎥⎥
M3 ⎢
=
⎢0 0 0 1⎥
⎢ ⎥
⎣I 1 0 0⎦

and the fourth expert using the six attributes given in the
example 2.3.2 gives the following neutrosophic connection
matrix
⎡0 1 1 0 0 0 ⎤
⎢1 0 0 0 0 0 ⎥
⎢ ⎥
⎢0 0 0 0 1 1 ⎥
M4 = ⎢ ⎥.
⎢1 1 0 0 0 0 ⎥
⎢0 0 1 1 0 0 ⎥
⎢ ⎥
⎣⎢0 0 0 1 1 0 ⎦⎥

The special fuzzy neutrosophic mixed square matrix associated


with the SFNFCM model is given by M1 ∪ M2 ∪ M3 ∪ M4

⎡0 1 1 0 0⎤
⎡0 0⎤ ⎢1 0 0 ⎥⎥
1 1
⎢1 0 1 0 ⎥⎥ ⎢ 1 1
= ⎢ ∪ ⎢I 0 0 0 1⎥ ∪
⎢0 0 0 1⎥ ⎢ ⎥
⎢ ⎥ ⎢1 0 I 0 0⎥
⎣1 1 1 0⎦
⎢⎣ 0 0 1 1 0 ⎥⎦

⎡0 1 1 0 0 0⎤
⎢1 0 ⎥⎥
⎡0 0 1 0⎤ ⎢ 0 0 0 0
⎢1 1 I 0 ⎥⎥ ⎢0 0 0 0 1 1⎥
⎢ ∪ ⎢ ⎥.
⎢0 0 0 1⎥ ⎢1 1 0 0 0 0⎥
⎢ ⎥ ⎢0
⎣I 1 0 0⎦ 0 1 1 0 0⎥
⎢ ⎥
⎢⎣ 0 0 0 1 1 0 ⎥⎦

235
Suppose the experts wish to work with special fuzzy mixed row
input vector
X = X1 ∪ X2 ∪ X3 ∪ X4
= [1 0 0 0] ∪ [0 0 0 1 0] ∪ [0 1 0 0] ∪ [0 0 0 0 0 0 1].

Now the effect of the special input vector X on M is given by

X o M = (X1∪ X2∪ X3∪ X4) o (M1 ∪ M2 ∪ M3 ∪ M4)


= X1 o M1 ∪ X2 o M2 ∪ X3o M3 ∪ X4 o M4
⎡0 1 1 0⎤
⎢1 0 1 0 ⎥
= [1 0 0 0 ] ⎢⎢
o ⎥ ∪
0 0 0 1⎥
⎢ ⎥
⎣1 1 1 0 ⎦

⎡0 1 1 0 0⎤
⎢1 0 1 1 0 ⎥⎥

[ 0 0 0 1 0] o ⎢ I 0 0 0 1⎥ ∪
⎢ ⎥
⎢1 0 I 0 0⎥
⎢⎣ 0 0 1 1 0 ⎥⎦

⎡0 0 1 0⎤
⎢1 0 I 0 ⎥⎥
[ 0 1 0 0 ] ⎢⎢0
o
0 0 1⎥

⎢ ⎥
⎣I 1 0 0⎦

⎡0 1 1 0 0 0⎤
⎢1 0 0 0 0 0 ⎥⎥

⎢0 0 0 0 1 1⎥
[0 0 0 0 0 1] o ⎢ ⎥
⎢1 1 0 0 0 0⎥
⎢0 0 1 1 0 0⎥
⎢ ⎥
⎣⎢ 0 0 0 1 1 0 ⎦⎥

236
= [0 1 1 0] ∪ [1 0 I 0 0 ] ∪ [1 0 I 0] ∪ [0 0 0 1 1 0] .

This is updated and thresholded and we obtain

Y = Y1 ∪ Y2 ∪ Y3 ∪ Y4
= [1 1 1 0] ∪ [1 0 I 1 0 ] ∪ [1 1 I 0] ∪ [0 0 0 1 1 0] .

We can find the effect of Y on the dynamical system M. The


reader is requested to find Y o M and the special hidden pattern.

2.4 Special FRMs and NRMs and their generalizations

In this section we define for the first time the new notion of
Special Fuzzy Relational Maps (SFRM) and illustrate it by
examples. Further we define the new multi expert model known
as special mixed model known as special mixed fuzzy relational
maps, we also define the neutrosophic analogue for these two
models. Finally we define a new Special Fuzzy Neutrosophic
Relational Maps (SFNRMs). SFNRMS can be thought of as a
generalization of FCM, SFCMs, SNCMs, SMFCMs, SMNCMs
and SFNCMs. These models can be applied when the attributes
related with the problem can be divided into two different
disjoint classes. This is more time saving and space economic,
so we seek to define this model.

DEFINITION 2.4.1: Suppose n experts want to work with a


problem P and if all of them wish to work using the fuzzy
relational maps model with m attributes in the domain space
and with t attributes in the range space The ith expert gives his
opinion which is rechanged into the connection matrix of the
FRM as Mi; i = 1, 2, …, n is a m × t fuzzy matrix with entries of
Mi taken from the set S = {–1, 0, 1}. Let the collective views of
the n experts in the form of FRMs be given by M = M1 ∪ M2 ∪
…∪ Mn which is a special fuzzy rectangular matrix known as
the fuzzy relational connection matrix of the Special Fuzzy
Relational Maps (SFRMs) model.

237
We shall just describe how it functions. Given M = M1 ∪ M2
∪ … ∪ Mn to be the associated matrix of the SFRM, each of the
FRMs in the SFRM is formed or obtained using the m attributes
as the domain space and the t-attributes as the range space.
Now the domain space of the SFRM will be known as the
special domain space Ds and it contains special fuzzy row
vectors each of the fuzzy row vectors in this special fuzzy row
vector is a 1 × m matrix. Thus X = X1 ∪ X2 ∪ … ∪ Xn then each
X i = ⎡⎣ x1i , x2i , … , xmi ⎤⎦ ; 1 ≤ i ≤ n and x ij ∈ {0, 1}; 1 ≤ j ≤ m. X is a
special fuzzy row vector from the special domain space Ds.
Similarly the range space of SFRM will be known as the
special range space denoted by Rs and every element Rs is a
special fuzzy row vector of the form Y = Y1 ∪ Y2 ∪ … ∪ Yn
where each Yi = ⎡⎣ y1i , y2i , … , yti ⎤⎦ with y ij ∈ {0, 1}, i = 1, 2, …,
n; 1 ≤ j ≤ t. These special fuzzy row vectors form the special
range space or special domain space, only indicate the ON or
OFF state of the node/attribute. Any input vector of the
dynamical system of the SFRM would be a special fuzzy row
vector. Suppose X = X1 ∪ X2 ∪ … ∪ Xn from Ds be the
input special vector given by all the n experts. To find the effect
X on the SFRM special dynamical system M.

X o M = (X1 ∪ X2 ∪ … ∪ Xn) o (M1 ∪ M2 ∪ … ∪ Mn)


= X1 o M1 ∪ X2 o M2 ∪ … ∪ Xn o Mn
= Y'1 ∪ Y'2 ∪ … ∪ Y'n
= Y'.
Y' may or may not be a special fuzzy row vector. We
threshold Y' to Y and find Y o M t as Y o M is not defined as in
the FRMs, we are endorsed to use the transpose also. Each Yi is
a 1 × t fuzzy row vector (1 ≤ i ≤ n). We find

Y o M t = Z'1 ∪ Z'2 ∪ … ∪ Z'm


= Z'.
Z' may or may not be a special fuzzy row vector so we first
threshold it; now it may so happen the coordinate with which
we started in the on state would have become off so we update

238
Z' to obtain Z. Let Z = Z1 ∪ Z2 ∪ … ∪ Zn; Z ∈ Ds and Z is a
special fuzzy row vector with each Zi a 1 × m fuzzy row vector
1≤ i ≤ n.
Now we find the effect of Z on the dynamical system M, i.e.,
Z o M and so on till we arrive at a fixed point or a limit cycle or
a combination of both which will be known as the special fixed
point or special limit cycle. When the system reaches a special
fixed point or a special limit cycle we call that the special
hidden pattern of the SFRM for the special input vector X = X1
∪ X2 ∪ … ∪ Xn ∈ Ds. It is important to note that there is no
error if one takes a input special vector Y = Y1 ∪ Y2 ∪ … ∪ Yn
∈ Rs and work with the dynamical system of the SFRM. We find

Y o M t = (Y1 ∪ Y2 ∪ … ∪ Yn) o (M1 ∪ M2 ∪ … ∪ Mn)t


= (Y1 ∪ Y2 ∪ … ∪ Yn) o ( M 1t ∪ M 2t ∪ … ∪ M nt )
= Y1 o M 1t ∪ Y2o M 2t ∪ … ∪ Yn o M nt
= Z'1 ∪ Z'2 ∪ … ∪ Z'n
= Z',
(we threshold Z' and obtain Z = Z1 ∪ … ∪ Zn). Now one can
find the effect of Z on M as
(Z1 ∪ Z2 ∪ … ∪ Zn) o (M1 ∪ M2 ∪ … ∪ Mn)
= Z1 o M1 ∪ Z2 o M2 ∪ … ∪ Zn o Mn
= P'1 ∪ P'2 ∪ … ∪ P'n ;

we threshold P’ to find P = P1 ∪ P2 ∪ … ∪ Pn. Now we find P o


M t and so on until we arrive at a special hidden pattern which
may be a special fixed point or a special limit cycle. It is
important and interesting to note that in case of SFRM we get a
pair of special fuzzy row vectors (S, T) where S ∈ Ds and T ∈ Rs
to be the special hidden pattern. We see in case of SFCMs we
get only one special fuzzy row vector.

Now we illustrate this situation by an example.

Example 2.4.1: Suppose three experts want to work with the


problem of finding the relationship between the industries profit

239
and the type of salaries to be given to the employee by the
industries.
The nodes given by the three experts related to the pay
techniques for the employee and their performance is taken as
the nodes of the domain spaces which is as follows.

D1 – Pay with allowances and bonus to workers


D2 – Only pay
D3 – Pay with allowance(or bonus)
D4 – Best performance by the employee
D5 – Average performance by the employee
D6 – Poor performance by the employee
D7 – Employee works for more number of hours
D8 – Employee works for less number of hours.

Attributes related with the industry which is taken as the nodes


of the range space.

R1 – Maximum profit to the industry


R2 – Only profit to the industry
R3 – Neither profit nor loss to the industry
R4 – Loss to the industry
R5 – Heavy loss to the industry.

The connection matrix M1 of the FRM given by the first expert


related to this problem.

⎡0 0 0 0 1⎤
⎢1 0 0 0 0 ⎥⎥

⎢0 0 1 0 0⎥
⎢ ⎥
1 0 0 0 0⎥
M1 = ⎢
⎢0 1 0 0 0⎥
⎢ ⎥
⎢0 0 0 0 1⎥
⎢1 0 0 0 0⎥
⎢ ⎥
⎣⎢ 0 0 0 1 0 ⎦⎥

240
where Di’s are taken along the rows and Rj’s along the column
of the relational matrix, (1 ≤ i ≤ 8 and 1 ≤ j ≤ 5).
The relational matrix given by the second expert for the
same problem with the same set of nodes using the FRM model
is given by M2.

⎡0 0 0 1 1⎤
⎢0 0 1 0 0 ⎥⎥

⎢0 0 1 0 0⎥
⎢ ⎥
1 1 0 0 0⎥
M2 = ⎢ .
⎢0 0 1 0 0⎥
⎢ ⎥
⎢0 0 0 1 0⎥
⎢0 0 0 0 1⎥
⎢ ⎥
⎢⎣ 0 0 0 1 1 ⎥⎦

The relational matrix given by the third expert related to the


problem of pay-profit in industries using the FRM model is
given by M3.

⎡0 0 0 1 0⎤
⎢0 0 1 0 0 ⎥⎥

⎢0 1 0 0 0⎥
⎢ ⎥
1 0 0 0 0⎥
M3 = ⎢ .
⎢0 0 1 0 0⎥
⎢ ⎥
⎢0 0 0 0 1⎥
⎢0 1 0 0 0⎥
⎢ ⎥
⎢⎣ 0 0 0 0 1 ⎥⎦

Thus M = M1 ∪ M2 ∪ M3 is the special fuzzy rectangular matrix


related with the SFRM model i.e.,

241
⎡0 0 0 1⎤
0 ⎡0 0 0 1 1 ⎤
⎢1 0 0 0
0⎥ ⎥ ⎢0 0 1 0 0⎥
⎢ ⎢ ⎥
⎢0 0 1 0⎥
0 ⎢0 0 1 0 0⎥
⎢ ⎥ ⎢ ⎥
M = ⎢1 0 0 0
0⎥
∪ ⎢ 1 1 0 0 0⎥

⎢0 1 0 0⎥
0 ⎢0 0 1 0 0⎥
⎢ ⎥ ⎢ ⎥
⎢0 0 0 1⎥
0 ⎢0 0 0 1 0⎥
⎢1 0 0 0⎥
0 ⎢0 0 0 0 1 ⎥
⎢ ⎥ ⎢ ⎥
⎢⎣ 0 0 0 0 ⎥⎦
1 ⎢⎣ 0 0 0 1 1 ⎥⎦
⎡0 0 0 1 0⎤
⎢0 0 1 0 0⎥
⎢ ⎥
⎢0 1 0 0 0⎥
⎢ ⎥
⎢1 0 0 0 0 ⎥ .
⎢0 0 1 0 0⎥
⎢ ⎥
⎢0 0 0 0 1 ⎥
⎢0 1 0 0 0⎥
⎢ ⎥
⎣⎢ 0 0 0 0 1 ⎦⎥
Suppose the experts want to work with the special input vector
X = X1 ∪ X2 ∪ X3
= [1 0 0 0 0 0 0 0] ∪ [1 0 0 0 0 0 0 0] ∪ [1 0 0 0 0 0 0 0].
Now the effect of X on the dynamical system M is given by

XoM = (X1 ∪ X2 ∪ X3) o (M1 ∪ M2 ∪ M3)


= X1 o M1 ∪ X2 o M2 ∪ X3 o M3
⎡0 0 0 0 1⎤
⎢1 0 0 0 0 ⎥⎥

⎢0 0 1 0 0⎥
⎢ ⎥
1 0 0 0 0⎥
= [1 0 0 0 0 0 0 0] o ⎢⎢ ∪
0 1 0 0 0⎥
⎢ ⎥
⎢0 0 0 0 1⎥
⎢1 0 0 0 0⎥
⎢ ⎥
⎢⎣ 0 0 0 1 0 ⎥⎦

242
⎡0 0 0 1 1⎤
⎢0 0 1 0 0 ⎥⎥

⎢0 0 1 0 0⎥
⎢ ⎥
1 1 0 0 0⎥
[1 0 0 0 0 0 0 0] o ⎢⎢ ∪
0 0 1 0 0⎥
⎢ ⎥
⎢0 0 0 1 0⎥
⎢0 0 0 0 1⎥
⎢ ⎥
⎣⎢ 0 0 0 1 1 ⎦⎥

⎡0 0 0 1 0⎤
⎢0 0 1 0 0 ⎥⎥

⎢0 1 0 0 0⎥
⎢ ⎥
1 0 0 0 0⎥
[1 0 0 0 0 0 0 0 ] ⎢⎢
o
0 0 1 0 0⎥
⎢ ⎥
⎢0 0 0 0 1⎥
⎢0 1 0 0 0⎥
⎢ ⎥
⎣⎢ 0 0 0 0 1 ⎦⎥

= [0 0 0 0 1] ∪ [0 0 0 1 1] ∪ [0 0 0 1 0]
= Y1 ∪ Y2 ∪ Y3
= Y ∈ Rs.
Now we find

YoM = [1 0 0 0 0 1 0 0] ∪ [2 0 0 0 0 1 1 2] ∪ [1 0 0 0 0 0
0 0]
= Z' .

Z' is thresholded and updated and Z is obtained. Now

Z = [1 0 0 0 0 1 0 0] ∪ [1 0 0 0 0 1 1 1] ∪ [1 0 0
0 0 0 0 0].

We find

243
Z o M = [0 0 0 0 2] ∪ [0 0 0 3 3] ∪ [0 0 0 1 0]
= T'
after thresholding we get the resultant of T' as

T = [0 0 0 0 1] ∪ [0 0 0 1 1] ∪ [0 0 0 1 0]
= Y ∈ Rs.

Thus the special hidden pattern of the special dynamical system


is given by the pair of special fuzzy row vectors or the special
binary pair {[1 0 0 0 0 0 1 0 0] ∪ [1 0 0 0 0 1 1 1] ∪ [1 0 0 0 0 0
0 0], [0 0 0 0 1] ∪ [0 0 0 1 1] ∪ [0 0 0 1 0]}.

The reader can work with any special input state vector and find
the special hidden pattern as the binary pair which would be a
pair of special fuzzy row vector.

Now we proceed on to define yet another new special model viz


special mixed FRM model.

DEFINITION 2.4.2: Let P be a problem which is analyzed by a


set of m experts. The only common feature for all of the experts
is that all of them agree to work with FRM model on the same
problem but they demand their liberty to choose any set of
attributes as per their desire. Let the ith expert choose to work
with ti number domain attributes of the FRM which forms the
rows of the relational matrix and with si number of range
attributes which form the columns of the relational matrix. Let
Mi denote the connection relational matrix which is a ti × si
fuzzy matrix with entries from the set {-1, 0, 1} for i = 1, 2, …,
m. (i.e., the same procedure is adopted by all the m experts). Let
us define the Special Mixed Fuzzy Relational Maps (SMFRMs)
model to be the model represented by the special fuzzy mixed
rectangular matrix M = M1 ∪ M2 ∪ …∪ Mm. The special
domain space Ds associated with this model is represented by
the set of all special fuzzy mixed row vectors X = X1 ∪ X2 ∪
…∪ Xm where each Xi is a 1 × ti fuzzy row vector with
X i = ⎡⎣ x1i x2i … xtii ⎤⎦ and xij ∈ {0, 1}, 1 ≤ j ≤ ti, 1 ≤ i ≤ m. The

244
special range space Rs associated with the SMFRM consists of
all special fuzzy mixed row vectors of the form Y = Y1 ∪ Y2 ∪
…∪ Ym where each Yi is a 1× si fuzzy row vector with
Yi = ⎡⎣ y1i y2i … ysii ⎤⎦ ; and y ij ∈ {0, 1}; j = 1, 2, …, si and 1 ≤ i ≤
m. The special dynamical system of the SMFRM comprises of
{M = M1 ∪ M2 ∪ … ∪ Mn, Ds and Rs}.
Now we will show how the special dynamical system of the
SMFRM functions. Let X = X1∪ X2 ∪ … ∪ Xm be the special
fuzzy mixed row vector given as the input vector by all the m
experts; X ∈ Ds to find the effect of X on the special dynamical
system M.
X o M = (X1 ∪ X2 ∪ … ∪ Xn) o (M1 ∪ M2 ∪ … ∪ Mm)
= X1 o M1 ∪ X2 o M2 ∪ … ∪ Xm o Mm
= Y1' ∪ Y2' ∪ … ∪ Ym'
= Y'.

We threshold Y' to Y = Y1 ∪ Y2 ∪ … ∪ Ym, Y ∈ RS. Now we


calculate
Y o M t = (Y1 ∪ Y2 ∪ … ∪ Ym) o ( M 1t ∪ M 2t ∪ … ∪ M mt )
= Y1 o M 1t ∪ Y2 o M 2t ∪ … ∪ Ym o M mt
= Z'1 ∪ Z'2 ∪ … ∪ Z'm
= Z'.
Z' may or may not belong to Ds; we update and threshold Z’ to
Z and if Z = Z1 ∪ Z2 ∪ … ∪ Zm. We find Z o M and so on, until
we arrive at a special fixed point or a special limit cycle which
is a binary pair (T, P) with T and P special fuzzy mixed row
vectors belonging to Ds and Rs respectively.

Now we proceed on to define the new notion of special


neutrosophic relational maps (SNRM) model.

DEFINITION 2.4.3: Suppose we have a problem P at hand and


let t experts want to work with it using Neutrosophic Relational

245
Maps and all of them agree upon to work with the same set of
domain attributes.
Now if in the SFRM dynamical system the special fuzzy
rectangular matrix is replaced by the special fuzzy neutrosophic
rectangular matrix and in the special range space and in the
special domain space, the special fuzzy row vectors are
replaced by special fuzzy neutrosophic row vectors then the
resulting dynamical system will be known as the Special
Neutrosophic Relational Maps (SNRMs) model. if all the t
experts agree to work on the problem P using the same fixed
number say m attributes which forms the domain space and n
attributes forming the range space then if we denote the fuzzy
neutrosophic relational matrix given by the ith expert by Mi
where Mi is a m × n fuzzy neutrosophic rectangular matrix true
for i = 1, 2, …, t, then M = M1 ∪ M2 ∪ … ∪ Mt together with
the special range and special domain space is the dynamical
system related with all the t experts.

Now the special domain space of this system would be given by


X = X1 ∪ X2 ∪ … ∪ Xt where each X i = ⎡⎣ x1i x2i … xmi ⎤⎦ ; x ij ∈
{0, I, 1} i = 1, 2, …, t and 1 ≤ j ≤ m and X ∈ Ds, Ds denotes the
set of all X with entries from {0, I, 1}. Clearly X is a special
neutrosophic row vector.

Note: It is important to note that the special input vectors in


SNRMs will always be from the set {0, 1} i.e., it denotes the on
or off state of the node; however the resultant special row
vectors given after finding X o M can be a special neutrosophic
row vector, of the special domain space or special range space.
That is why we in the definition take the special domain space
and special range space to be a special neutrosophic row
vectors.

Now the special range space Y = Y1 ∪ Y2 ∪ … ∪ Yt where


each Yi = ( y1i yi2 … yin ) ; yij ∈ {0, I, 1}; i = 1, 2, …, t and 1 ≤ j
≤ n and Y ∈ Rs, Rs denotes the set of all Y with entries from {0,
1}. Y is also a special neutrosophic input row vector. Now the

246
triple {M = M1 ∪ M2 ∪ … ∪ Mn, Ds, Rs} forms the special
dynamical neutrosophic system and the model would be known
as the Special Neutrosophic Relational Maps (SNRM-model).

Next we proceed on to give a brief description of another


special neutrosophic model.

DEFINITION 2.4.4: Suppose in the definition 2.4.2 of the


SMFRM model if the FRMs are replaced by NRM we get the
new Special Mixed Neutrosophic Relational Maps (SMNRMs)
model i.e., the special dynamical system of the SMNRM model
is given by the triple {M = M1 ∪ M2 ∪ … ∪ Mn, Ds, Rs } where
Mi’s are fuzzy neutrosophic rectangular ti × si matrices with
entries from the set {-I, -1, 0, I, 1}; i.e., M is a special fuzzy
neutrosophic mixed rectangular matrices; where Ds and Rs are
special fuzzy neutrosophic mixed row vectors.
Here it is pertinent to mention that however when the
experts give any input special vector X = X = X1 ∪ X2 ∪ … ∪
Xn then Xi’s can take only entries from the set {0, 1}; i.e., the
dynamical set accepts only the on or off state of the nodes for no
expert can claim the existence or non existence or
indeterminacy of the node only while working with the problem
of the dynamical system with certain nodes in the on state either
from the range attributes or the domain attributes it may so
happen that the state of the node may become indeterminate
with some input vectors.

Now we proceed on to define yet another special new


model viz., Special Fuzzy Neutrosophic Relational Maps
(SFNRM) model.

DEFINITION 2.4.5: Suppose we have a problem P with which


some t experts wish to work. It so happens that out of the t
experts some m experts wants to work using only the Fuzzy
Relational Maps (FRM) model and rest (t – m) experts wants to
work using the Neutrosophic Relational Maps (NRM) model.
Here also they i.e., the m experts may work with the same set of
attributes or different sets of for their domain and range spaces

247
of the FRMs, like wise the t-m experts may work with same set
of attributes or different sets of attributes for their domain and
range spaces of the NRMs. Now this new model given by these t
experts for the problem will consists of a special dynamical
system, the special matrix which will be a mixture of
rectangular fuzzy matrices as well as rectangular neutrosophic
matrices given by M = {M1 ∪ M2 ∪ … ∪ Mm} ∪ {Mm+1 ∪ … ∪
Mt} where M1 ∪ M2 ∪ … ∪ Mm is a special fuzzy mixed
rectangular matrix and {Mm+1 ∪ … ∪ Mt} is a special
neutrosophic rectangular matrix. Thus {M1 ∪ M2 ∪ … ∪ Mm}
∪ {Mm+1 Mm+1 ∪ … ∪ Mt} will be known as the special mixed
fuzzy and neutrosophic mixed rectangular matrix. Now the
special domain space X = {X1 ∪ X2 ∪ … ∪ Xm} ∪ {Xm+1 ∪ Xm+2
∪ … ∪ Xt} will be such that {X1 ∪ X2 ∪ … ∪ Xm} is a special
fuzzy mixed row vectors and {Xm+1 ∪ Xm+2 ∪ … ∪ Xt} will be a
special fuzzy neutrosophic mixed row vectors; thus X = {X1 ∪
X2 ∪ … ∪ Xm} ∪ {Xm+1 ∪ Xm+2 ∪ … ∪ Xt} will be a special
mixed fuzzy and neutrosophic mixed row vectors. This will be
known as the Special Domain Space of the new model. Similarly
Y = {Y1 ∪ Y2 ∪ … ∪ Ym} ∪ {Ym+1 ∪ Ym+2 ∪ … ∪ Yt} will be a
special mixed fuzzy and neutrosophic mixed row vector known
as the Special Range Space of the new model.
So the Special Fuzzy and Neutrosophic Relational Maps
(SFNRM) model will consist of Special Fuzzy Neutrosophic
mixed rectangular matrices, special domain space Ds and
special range Rs where Ds will be the set of all special fuzzy
neutrosophic mixed row vectors of the form X = {X1 ∪ X2 ∪ …
∪ Xm} ∪ {Xm+1 ∪ Xm+2 ∪ … ∪ Xt} where X1 ∪ X2 ∪ … ∪ Xm are
special fuzzy mixed row vectors with entries from the set {0, 1}
and {Xm+1 ∪ Xm+2 ∪ … ∪ Xt} are special neutrosophic mixed
row vectors with entries from the set {I, 0 1}. Clearly even in
this model any input vector from the t experts say X = {X1 ∪ X2
∪ … ∪ Xm} ∪ {Xm+1 ∪ Xm+2 ∪ … ∪ Xt}will take entries only
from the set {0, 1} i.e., on or off state of the nodes or attributes.

248
Thus M = {M1 ∪ M2 ∪ … ∪ Mm} ∪ {Mm+1 ∪ …∪ Mt, Ds, Rs}
will form the special dynamical system of SFNRM model.

The main advantage of this model is that the set of experts can
choose to work either using FRM or NRM. For all the while, the
models which we have given in this section can either work in
FRM or in NRM. So in this manner this model is more powerful
or advantageous than other models.
Secondly in this model one can use any set of attributes for
the domain space and the range space, it is left for the choice of
the expert, thus the special domain space and the special range
space contain special fuzzy and neutrosophic mixed row
vectors.
This model can be realized as a generalization of SFRMs,
SMFRMs, SNRMs and SMNRMs models.
Now we proceed on to define yet another new model called
the Special Mixed Fuzzy Cognitive and Fuzzy Relational Maps
(SMFCMs) model.

DEFINITION 2.4.6: Suppose we have a problem P at hand and


some t experts want to work with this problem. Here some of the
experts want to work using FCMs and some of them prefer to
work using FRMs and they may opt for same set of attributes or
for different sets of attributes according to their choice. In this
case we use the Special Mixed Fuzzy Cognitive and Fuzzy
Relational Maps Model. This model can be thought of as a
mixture of SMFCM and SMFRM model. Let m experts choose to
work with FCM and t-m of them choose to work with FRM then
these m experts choose to work with SMFCM model and t-m of
them work with SMFRM model. Suppose M = {M1 ∪ M2 ∪ … ∪
Mm} denotes the SMFCMs associated special fuzzy mixed
square matrices and T = {Mm+1 ∪ Mm+2 ∪ … ∪ Mt}denotes the
SMFRMs associated special mixed fuzzy rectangular matrices
then M ∪ T = {M1 ∪ M2 ∪ … ∪ Mm} ∪ {Mm+1 ∪ Mm+2 ∪ … ∪
Mt} denotes the special fuzzy mixed matrices associated with the
SMFCRM-model and the special domain space is given by {X1
∪ X2 ∪ … ∪ Xm} ∪ {Xm+1 ∪ Xm+2 ∪ … ∪ Xt} the set of all
special fuzzy mixed row vectors with entries from the set {0, 1}

249
and the special range space Y = {Y1 ∪ Y2 ∪ … ∪ Ym} ∪ {Ym+1 ∪
Ym+2 ∪ … ∪ Yt} denotes the set of all special fuzzy mixed row
matrices with entries from the set {0, 1}. Thus M = {M1 ∪ M2 ∪
… ∪ Mm} ∪ {Mm+1 Mm+2 ∪ … ∪ Mt} Ds, Rs} denotes the special
fuzzy dynamical system of the SMFCFRM.

The main advantage of this special model is that for the same
problem some experts can choose to work with FCM and the
rest with FRM so this multi expert model need not model it in
two stages or as two models for now in this new model they
function as a single model with double purpose.
Now we proceed on to describe the next model viz. the
Special Mixed Neutrosophic Cognitive and Neutrosophic
Relational Maps Model (SMNCNRMs model).

DEFINITION 2.4.7: Suppose we have a problem P at hand and


all the experts agree to work using only the neutrosophic
models. One set of experts out of t experts say m experts wish to
work with the NCM model with different sets of attributes i.e.,
the SMNCM model and the t-m experts wish to work with the
NRM model they too with different sets of domain and range
space i.e., SMNRM model. Then as in case of the above
definition given in 2.4.6 we form the Special Mixed
Neutrosophic Cognitive and Neutrosophic Relational Map
Model i.e., in the definition 2.4.6 we can replace the FCM by
NCM and FRM by NRM and obtain the SMNCNRM-model. The
dynamical system associated with this model is given by M =
{M1 ∪ M2 ∪ …∪ Mm} ∪ {Mm+1 Mm+2 ∪ … ∪ Mt, Ds, Rs } where
Ds is the special domain space consisting of all special
neutrosophic mixed row vectors of the form X = {X1 ∪ X2 … ∪
Xm} ∪ {Xm+1 ∪ Xm+2 ∪ …∪ Xt}, where Xi takes its values from
the set {I, 0, 1}, i = 1, 2, …, t and the special range space Rs
consists of all special neutrosophic mixed row vectors of the
form Y = {Y1 ∪ Y2 …∪ Ym} ∪ { Ym+1 ∪ Ym+2 ∪ …∪ Yt} where Yi
takes their entries from the set {I, 0, 1}, i = 1, 2, …, t. It is
pertinent to mention here that in case of special input vectors
from the t experts would be given by P = {P1 ∪ P2 …∪ Pm} ∪

250
{Pm+1 ∪ Pm+2 ∪ … ∪ Pt} we see the entries of each Pi; 1 ≤ i ≤ t
are only are from the set {0, 1} i.e., the on or off state of the
attribute in the special range space or the special domain
space.

The main advantage of this model over the other models is that
this model can function with two types of models at a time like
the SMFCFRMs and secondly when every expert agrees to
work with neutrosophic models i.e., they feel the notion of
indeterminacy is present in the problem.
Next we proceed on to define yet a new model which can be
thought of as the generalization of all the models so far
introduced.

DEFINITION 2.4.8: Let P be the problem which is to be


investigated. Suppose t experts wish to work with the problem,
but some i.e., t1 experts wish to use FCM model, a few of them
say t2 experts use FRM model, some t3 of them want to use NCM
model as they think the problem involves some degree of
indeterminacy and the rest of experts t – (t1+ t2+ t3) of them
wish to work with NRM model. The model should be so
connected that it should work as a single unit and give the
results simultaneously. Now this new special model can be got
as the union of SMFCM ∪ SMFRM ∪ SMNCM ∪ SMNRM so
that this can be represented by the union of four sets of special
matrices given by these four special models. Suppose M denotes
the associated matrix of the new model them
M = M1 ∪ M2 ∪ M3 ∪ M4
= (M1 ∪ M2 ∪ …∪ M t1 ) ∪ { M t1 +1 ∪ M t1+2 ∪ …∪ M t2 }
∪ { Mt 2 +1
∪ Mt
1+2
∪ …∪ M t } ∪ { M t ∪ M t ∪ …∪ M t }
3 3+1 3+ 2

where M1 is a special fuzzy mixed square matrices, M2 is a


special fuzzy mixed rectangular matrices, M3 is a special fuzzy
neutrosophic mixed square matrices and M4 is a special fuzzy
neutrosophic mixed rectangular matrices related with the
models SMFCM, SMFRM, SMNCM and SMNRM respectively.
Now this special dynamical system will have three components
given by the triple {M = M1 ∪ M2 ∪ M3 ∪ M4 = {M1 ∪ M2 ∪

251
…∪ M t1 } ∪ { M t1 +1 ∪ M t1+2 ∪ … ∪ M t2 } ∪ { M t2+1 ∪ M t2+2 ∪
… ∪ M t3 } ∪ { M t3+1 ∪ M t3+2 ∪ … ∪ M t }, Ds, Rs}
where Ds is the special fuzzy neutrosophic mixed row vectors of
the form {X1 ∪ X2 ∪ … ∪ X t1 } ∪ { X t1 +1 ∪ X t1+2 ∪ … ∪ X t2 }
∪ { Xt 2 +1
∪ Xt 2+2
∪ … ∪ Xt } ∪ { Xt
3 3+1
∪ Xt 3+ 2
∪ … ∪ Xt } =
X1 ∪ X2 ∪ X3 ∪ X4 where each Xi = ⎡⎣ x1i x2i … xli ⎤⎦ where x ij ∈

{0, 1} with 1 ≤ l < t1, t2 and Xi = ⎡⎣ x1k x2k ... x kp ⎤⎦ (1 ≤ p ≤ t3, t–


(t1+ t2+ t3) and xrk ∈ {0, 1, I}. Similarly Rs is also a special
fuzzy neutrosophic mixed row vector like Ds.

Now X o M = (X1 ∪ X2 ∪ X3 ∪ X4) o (M1 ∪ M2 ∪ M3 ∪ M4) =


X1 o M1 ∪ X2 o M2 ∪ X3 o M3 ∪ X4 o M4 where each Xi o M1
works like the dynamical system SMFCM, X2 o M2 works like
the dynamical system SMFRM, X3 o M3 works like the
dynamical system SMNCM and X4 o M4 works like the
dynamical system of SMNRM respectively.
However the new SMFCRNCRM model can be realized as
generalized model of every other models so far described in
sections 2.1 to 2.4 of this book. Further this model combines the
fuzzy and indeterminacy together.
Now we will show the working of the model by an explicit
example.

Example 2.4.2: Suppose one is interested in studying the


Employee-Employers model so that the industry should run in
profit as well as the workers should be satisfied and the
performance of the industry must be good in both quality and
production. There are six experts who wish to work with the
model and each of them choose to work with a distinct model.
The set of attributes collectively put by all of them is as follows:

A1 - Employee gets pay, allowance and bonus


A2 - He gets only pay
A3 - Pay with either allowances or bonus alone
A4 - Best performance by the employee

252
A5 - Maximum profit for the employer
A6 - Average performance by the employee
A7 - Poor performance by the employee
A8 - Only profit to the employer
A9 - Employee works for more number of hours
A10 - Neither profit nor loss to the employer
A12 - Heavy loss to the employer
A13 - Employee works for less number of hours.

The first expert wishes to work with 8 attributes using the FCM
model. The connection matrix M1 given by him is as follows:

A1 A2 A3 A4 A5 A9 A11 A13
A1 ⎡0 0 0 1 1 0 0 0⎤
A2 ⎢⎢0 0 0 0 0 0 1 1 ⎥⎥
A3 ⎢0 0 0 0 0 0 1 0⎥
⎢ ⎥
M1 = A4 ⎢1 0 0 0 1 0 0 0⎥ .
A5 ⎢1 0 0 0 0 1 0 0⎥
⎢ ⎥
A9 ⎢1 0 0 0 0 1 0 0⎥
A11 ⎢0 1 0 0 0 0 1 1⎥
⎢ ⎥
A13 ⎣⎢0 1 0 0 0 0 1 0 ⎦⎥

The second expert also wants to work with FCM but with
different sets of attributes and more so only with six attributes
the connection matrix M2 given by him is a 6 × 6 matrix, M2
given by:

A1 A2 A3 A4 A8 A7
A1 ⎡0 0 0 1 0 0⎤
A2 ⎢⎢0 0 0 0 0 1 ⎥⎥
M2 = A3 ⎢0 0 0 0 1 0⎥ .
⎢ ⎥
A4 ⎢1 0 0 0 1 0⎥
A8 ⎢1 0 1 0 0 0⎥
⎢ ⎥
A7 ⎣⎢0 1 0 0 0 0 ⎦⎥

253
The third expert wants to work with a NCM model just having
only 7 attributes the connection matrix M3 given by him is as
follows. However he uses some set of attributes from the 13
attributes.

A2 A3 A6 A7 A8 A9 A10
A2 ⎡0 0 1 0 0 I 1⎤
A3 ⎢0 0 0 1 0 0 0 ⎥⎥

A6 ⎢0 0 0 0 0 0 1⎥
M3 = ⎢ ⎥.
A7 ⎢1 0 0 0 0 0 0⎥
A8 ⎢1 0 0 0 0 0 0⎥
⎢ ⎥
A9 ⎢0 1 0 0 1 0 I⎥
A10 ⎢⎣0 0 I 0 0 0 0 ⎥⎦

The fourth expert works with the FRM taking all the attributes
and forms the relational matrix M4 which is given by him is as
follows:

A5 A8 A10 A11 A12


A1 ⎡0 0 0 0 1⎤
A2 ⎢⎢1 0 0 0 0 ⎥⎥
A3 ⎢0 0 1 0 0⎥
⎢ ⎥
M4 = A4 ⎢1 0 0 0 0⎥ .
A6 ⎢0 1 0 0 0⎥
⎢ ⎥
A7 ⎢0 0 0 0 1⎥
A9 ⎢1 0 0 0 0⎥
⎢ ⎥
A13 ⎣⎢0 0 0 1 0 ⎦⎥

The fifth expert gives a relational matrix M5 using FRM model


he does not use all the attributes but uses only a fewer number
of attributes.

254
A8 A10 A11 A12
A2 ⎡0 0 1 1⎤
A3 ⎢⎢0 1 0 0 ⎥⎥
M5 = A4 ⎢0 1 0 0⎥ .
⎢ ⎥
A6 ⎢1 0 0 0⎥
A7 ⎢0 1 0 0⎥
⎢ ⎥
A13 ⎢⎣0 0 1 1 ⎥⎦

The sixth expert however using yet a different set of


attributes and uses NRM and the related matrix M6 given by
him is as follows:
A8 A10 A11 A12
A1⎡0 0 0 1⎤
A2⎢0 0 1 0 ⎥⎥

A3 ⎢ 0 0 1 0⎥
M6 = ⎢ ⎥.
A4 ⎢1 1 0 0⎥
A6 ⎢ 0 0 1 0⎥
⎢ ⎥
A7 ⎢ 0 0 0 I⎥
A13 ⎢⎣ 0 0 0 1 ⎥⎦

The six experts choose to work with the SMFCRNCRM model


given by the
(M1 ∪ M2 ∪ M3 ∪ M4 ∪ M5 ∪ M6)

⎡0 0 0 1 1 0 0 0⎤
⎢0 1 ⎥⎥
⎢ 0 0 0 0 0 1 ⎡0 0 0 1 0 0⎤
⎢0 0 0 0 0 0 1 0⎥ ⎢0 0 0 0 0 1 ⎥⎥
⎢ ⎥ ⎢
⎢1 0 0 0 1 0 0 0⎥ ⎢0 0 0 0 1 0⎥
∪ ⎢ ⎥ ∪
⎢1 ⎥

0 0 0 0 1 0 0
⎥ ⎢1 0 0 0 1 0⎥
⎢1 0 0 0 0 1 0 0⎥ ⎢1 0 1 0 0 0⎥
⎢0 ⎥ ⎢ ⎥
1 0 0 0 0 1 1 ⎢⎣ 0 1 0 0 0 0 ⎥⎦
⎢ ⎥
⎣⎢ 0 1 0 0 0 0 1 0 ⎦⎥

255
⎡0 0 0 0 1⎤
⎡0 1⎤ ⎢1 0 0 0 ⎥⎥
0 1 0 0 I
⎢0 0 0 1 0 0 0⎥ ⎥ ⎢ 0
⎢ ⎢0 0 1 0 0⎥
⎢0 0 0 0 0 0 1⎥ ⎢ ⎥
⎢ ⎥ 1 0 0 0 0⎥
⎢1 0 0 0 0 0 0⎥ ∪ ⎢ ∪
⎢0 1 0 0 0⎥
⎢1 0 0 0 0 0 0⎥ ⎢ ⎥
⎢ ⎥ ⎢0 0 0 0 1⎥
⎢0 1 0 0 1 0 I⎥ ⎢1 0 0
⎢0 0 0⎥
⎣ 0 I 0 0 0 0 ⎥⎦ ⎢ ⎥
⎢⎣ 0 0 0 1 0 ⎥⎦

⎡0 0 0 1⎤
⎡0 1⎤ ⎢0 0 0 ⎥⎥
0 1
⎢0 1 0 0 ⎥⎥ ⎢ 1
⎢ ⎢0 0 1 0⎥
⎢0 1 0 0⎥ ⎢ ⎥
⎢ ⎥ ∪ ⎢1 1 0 0⎥ ;
⎢1 0 0 0⎥
⎢0 0 1 0⎥
⎢0 1 0 0⎥ ⎢ ⎥
⎢ ⎥ ⎢0 0 0 I⎥
⎣⎢ 0 0 1 1 ⎦⎥ ⎢0 0
⎣ 0 1 ⎥⎦

is a special fuzzy-neutrosophic mixed matrix. Now the set of


special attributes or special initial domain state vector with
which the model works is

X = {⎡⎣ x 1
1 x12 … x18 ⎤⎦ ∪ ⎡⎣ x12 x 22 … x 62 ⎤⎦ ∪ ⎡⎣ x13 x 32 … x 37 ⎤⎦ ∪

}
⎡⎣ x14 x 24 … x 84 ⎤⎦ ∪ ⎡⎣ x15 x 52 … x 56 ⎤⎦ ∪ ⎡⎣ x16 x 62 … x 67 ⎤⎦
= X1 ∪ X2 ∪ X3 ∪ … ∪ X6

where X is a special fuzzy mixed row vector x ij ∈ {0, 1} or {0,


1, I}; 1 ≤ i ≤ 6 and 1 ≤ j ≤ 8. The special initial range space state
vector

Y = Y1 ∪ Y2 … ∪ Y6

256
= {⎡⎣ y1
1 y12 … y18 ⎤⎦ ∪ ⎡⎣ y12 y 22 … y 62 ⎤⎦ ∪ ⎡⎣ y13 y32 … y37 ⎤⎦ ∪

}
⎡⎣ y14 y 24 … y54 ⎤⎦ ∪ ⎡⎣ y15 y52 … y54 ⎤⎦ ∪ ⎡⎣ y16 y 62 … y 64 ⎤⎦
where y tj ∈ {0, 1, I} but if this Y is started as a initial vector
then yit ∈ {0, 1}only i.e., on or off state.
Let us suppose the experts agree and give the special initial
state vector

X = {[1 0 0 0 0 0 0 0] ∪ [0 0 0 0 1 0] ∪ [0 1 0 0 0 0 0]
∪ [0 0 0 0 0 0 0 1] ∪ [0 0 1 0 0 0] ∪ [0 0 0 1 0 0 0]}

Now
X om M = [X1 ∪ X2 ∪ … ∪ X6] o [M1 ∪ M2 ∪ … ∪ M6]
= X1 o M1 ∪ X2 o M2 ∪ … ∪ X6 o M6

= [0 0 0 1 1 0 0 0] ∪ [1 0 1 0 0 0] ∪ [0 0 0 1 0 0 0] ∪
[0 0 0 1 0] ∪ [0 1 0 0] ∪ [1 1 0 0]
= Y'1 ∪ Y'2 ∪ Y'3 ∪ Y'4 ∪ Y'5 ∪ Y'6
= Y'

after updating Y' we get

Y = Y1 ∪ Y2 ∪ Y3 ∪ Y4 ∪ Y5 ∪ Y6
= [1 0 0 1 1 0 0 0] ∪ [1 0 1 0 1 0] ∪ [0 1 0 1 0 0 0] ∪
[0 0 0 1 0] ∪ [0 1 0 0] ∪ [1 1 0 0] .

Now Y o M is not defined and M is a mixture of both square


and rectangular matrices we have to find only the special
transpose of the model and determine

Y om M = Y1 o M1 ∪ Y2 o M2 ∪ Y3 o M3 ∪ Y4 o M 4t ∪ Y5 o
M 5t ∪ Y6 o M 6t

= [2 0 0 1 2 1 0 0] ∪ [1 0 1 1 1 0] ∪ [1 0 0 1 0 0 0] ∪
[0 0 0 0 0 0 0 1] ∪ [0 1 1 0 1 0] ∪ [0 0 0 2 0 0 0]
= Z'1 ∪ Z'2 ∪ Z'3 ∪ Z'4 ∪ Z'5 ∪ Z'6

257
= Z'
after thresholding Z ' we get

Z = Z1 ∪ Z2 ∪ … ∪ Z6
= [1 0 0 1 1 1 0 0] ∪ [1 0 1 1 1 0] ∪ [1 1 0 1 0 0 0] ∪
[0 0 0 0 0 0 0 1] ∪ [0 1 1 0 1 0] ∪ [0 0 0 1 0 0 0] .

Now using the special dynamical system M we find

ZoM = Z1 o M1 ∪ Z2 o M2 ∪ Z3 o M3 ∪ Z4 o M4 ∪ Z5 o M5
∪ Z6 o M6
= [3 0 0 1 2 2 0 0] ∪ [2 0 1 1 2 0] ∪ [1 0 1 1 0 I 1] ∪
[0 0 0 1 0] ∪ [0 3 0 0] ∪ [1 1 0 0]
= T'1 ∪ T'2 ∪ … ∪ T'6.

After updating and thresholding T' we get


T = T1 ∪ T2 ∪ … ∪ T6
= [1 0 0 1 1 1 0 0] ∪ [1 0 1 1 1 0] ∪ [1 1 1 1 0 I 0] ∪
[0 0 0 1 0] ∪ [0 1 0 0] ∪ [1 1 0 0].

We see T1, T2, T4, T5 and T6 remain as fixed points. We need to


work only with M3 keeping this in mind we just work with M3
i.e., finding T3 ° M3 one gets the special limit cycle or a special
fixed point.
Now having seen the working procedure of the model we
now define yet another new special fuzzy and special neutron
neutrosophic model and generalize them in the following
section.

2.5 Special FRE models and NRE models

In this section for the first time we introduce the notion of


Special Fuzzy relational equations and their generalizations.
Any Fuzzy relational equations (FRE) model has only the
capacity to tackle a single experts opinion on any problem at a
time. A brief description of this model has been described in
page 180-6 of this book. For more about FRE please refer [106,

258
232]. The FRE model makes use of min max or max min
operators. This new model which we will be calling as special
Fuzzy Relational Equations Model can deal with any desired
number of experts at a time. This is the first time such types of
fuzzy models have been introduced.
In this section we introduce 5 types of Special fuzzy
relational equations models.

DEFINITION 2.5.1: Suppose we have a problem P at hand and


at least t experts wish to work with the same problem using the
same set of attributes both for the domain space as well as
range space then how to model this problem with multi expert
models so that the views or solution is given simultaneously and
collectively and not individually and separately. This model we
would be defining will be known as Special Fuzzy Relational
Equations (SFRE) model.
Suppose we have some t experts giving their opinion with
the same set of m domain and n-range attributes. Let Ti denote
the FRE model given by the ith experts using these m domain
attributes and n-range attributes, then Ti is a m × n fuzzy
membership matrix i.e., entries of Ti are from the unit interval
[0, 1]. Let T1, T2, …, Tt be the FRE matrices given by the t
experts. Let T = T1 ∪ T2 ∪ … ∪ Tt then T would be known as
the Special Fuzzy Relational Equation(SFRE) model i.e., all the
t experts use the same set of m row attributes and same set of n
column attributes(m need not always be equal to n).
Thus the membership matrix of the ith expert is a m × n
fuzzy matrix. The special fuzzy rectangular matrix associated
with the SFRE will be known as the special fuzzy membership
function matrix of the SFRE.
For more about FREs please refer [106, 232]. Depending
on the type of FRE and the problem we can adopt any tool say
neural net works or matrix equation or any suitable method to
solve the problem.
Now if in the definition 2.5.1 the t experts choose to work
with different sets of column attributes say mi column attributes
is chosen by the ith expert and different sets of row attributes say
ni attributes then the membership matrix would be a mi × ni
fuzzy matrix, true for i =1, 2, …, t. Now the special membership

259
matrix associated with this new model will be a special fuzzy
mixed rectangular matrix and the associated special
membership matrix will be given by M1 ∪ M2 ∪ … ∪ Mt where
each mi × ni fuzzy rectangular matrix. We call this model as
Special Mixed Fuzzy Relational Equations(SMFRE) model.
Depending on the type of FRE and the problem one can work
for the solution or the solution close to the expected solution.

The next special model is using the Neutrosophic Relational


Equations (NRE).

DEFINITION 2.5.2: Suppose we have p experts interested in


using the NRE model for the problem P(say). Further suppose
all of them agree upon to work with the same set of m row
attributes and the same set of n column attributes. Let the
neutrosophic membership matrix given by the ith expert be a m
× n neutrosophic matrix say Mi; i = 1, 2, …, p. Then we have M
= M1 ∪ M2 ∪ … ∪ Mp where M is a special fuzzy neutrosophic
rectangular matrix.
The model associated with this M will be known as the
Special Neutrosophic Relational Equations(SNRE) model. In
other words SNRE model is just the SFRE model in which
instead of working with the fuzzy membership matrices we have
to work with the fuzzy neutrosophic membership matrices. So by
replacing in SFRE model all fuzzy membership matrices by the
appropriate fuzzy neutrosophic membership matrices we get the
SNRE model. Now as in case of SMFRE if we replace each of
the mi × ni fuzzy membership matrix by a fuzzy neutrosophic
membership matrix we would get the new Special Mixed
Neutrosophic Relational Equations Model i.e., the SMNRE
model.

DEFINITION 2.5.3: Now suppose we have a set of m experts


working on a problem P where m1 of them expect to work using
the SFRE model put together or equivalently the FRE model
individually. A set m2 experts choose to work using SMFRE
model for the same problem P, m3 of the experts choose to use
NRE model i.e., they collectively work using the SNRE model

260
and the rest of the experts collectively use the SMNRE model
thus all the m experts works either on NRE and FRE models.
We call this model as the special mixed Fuzzy Neutrosophic
Relational Equations model (SMFNRE-model). This model is a
combination of all the four special models viz SMFRE, SFRE,
SNRE and SMNRE.

The main advantages of this model is:

1. This is a new multi expert special model which can


simultaneously cater to any number of experts.
2. This model allows them to work with FRE and NRE.
3. This model enables them to use any desired number of
row attributes and column attributes.

Since the very choice of the type of FRE or NRE is highly


dependent on the problem and the expected solution for the
instance it can be a multi objective optimization problem with
FRE (NRE) constraints or it may be a non linear optimization
problem with FRE(NRE) constraints or by using max min or
min max operations of FRE or use of neutral net work and so
on. For several types refer[106, 232].
Lastly we define yet a very generalized special model for
solving problems of the following type. The problem can be
thought of as the multiexpert multi model problem for we not
only have multi experts but we also have the experts wishing to
work with very many models. Then can we have a single
dynamical system to tackle the situation.
The answer is yes and we define a new Special Fuzzy-
Neutrosophic-Cognitive-Relational- Equations S-F-N-C-R-E
model.

DEFINITION 2.5.4: Suppose we have some m experts working


on a problem P. There are six sets of experts, first set of experts
wish to work with FCM, each one with the attributes of their
own choice, second set of experts wish to work with NCM each
one with the set of attributes of their own choice. The third set
of experts wants to work with FRM with a set of attributes
different for each expert. The fourth set of experts want to work

261
with NRM with a set of attributes varying from expert to expert.
The fifth set of experts want to work using FRE and with their
own set of rows and columns attributes give their fuzzy
membership matrix.
The last set of experts work using NRE. Now a model which
can cater to the needs of all these 6 sets of experts with different
types of models is termed as Super Special Hexagonal Model
denoted by SSHM = SMFCM ∪ SMNCM ∪ SMFRE ∪ SMNRE
∪ SMFRM ∪ SMNRM where ‘∪’ is just notationally describing
all the six models are present and the union of the six model
Special fuzzy and Special fuzzy neutrosophic mixed
square/rectangular matrices will be the dynamical system
associated with this super model.

Further the special operation described in chapter one will give


the desired method of working.

Note: When we use only 5 models and five sets of experts we


call it as Super Special Quintuple model, if we use only four
models with four sets of experts we call it as Super Special
Quadruple model, with three models and three sets of experts
will be known Super Special Triple model.

The main advantages of this model over the other models are as
follows :

a. This multiexpert model can simultaneously deal with both


multi experts and multi models.
b. With the advent of computer finding a program is not very
difficult for these SSHM models. When programmed
finding the solution is very easy and simple with any
number of experts and any number of models.
c. Simultaneous solution is possible using special operations.
This model can be both fuzzy as well as neutrosophic
simultaneously.
d. The special fuzzy operation mentioned in – helps in finding
the special resultant vector given any special input vector.

262
2.6 Some Programming Problems for Computer Experts

In this section we suggest some simple programming


problems for a interested computer scientist or a programmer.

a. Give a C or C++ program to find the special hidden


pattern of the following special models.

i. SFCM
ii. SMFCM
iii. SNCM
iv. SMNCM
v. SFRM
vi. SMFRM
vii. SNRM
viii. SMNRM
ix. SMFNRM
x. SMFNCRM
xi. SSHM

Hint: In case of single experts we have both Java and C program


for FCM and FRM [228, 231].

b. Find a program for calculating the min max or max min


operation for the given fuzzy row vector X = [x1 x2 …
xn]; xi ∈ [0, 1] and A = (aij); A a n × m fuzzy matrix
with aij ∈ [0, 1]; 1 ≤ i ≤ n and 1 ≤ i ≤ m; i.e., finding

max min {X,A}


⎧ ⎡ a11 a12 a1m ⎤ ⎫
⎪ ⎢a ⎪
⎪ a 22 a 2m ⎥⎥ ⎪
= max min ⎨[ x1 x 2 ...x n ] , ⎢ 21 ⎬
⎪ ⎢ ⎥⎪
⎪ ⎢ ⎥
⎩ ⎣a n1 a n 2 a nm ⎦ ⎭⎪

= ⎡⎣ max {min ( x1 ,a11 ) , min ( x 2 ,a 22 ) ,… , min ( x n ,a n1 )}


max {min ( x1 ,a12 ) , min ( x 2 ,a 22 ) ,… , min ( x n ,a n 2 )}

263
max {min ( x1 ,a1m ) , min ( x 2 ,a 2m ) ,… , min ( x n ,a nm )}⎤⎦

= ( b1 b 2 … b m ) ; bi ∈ [ 0,1] .

c. Given M = M1 ∪ M2 ∪ … ∪ Mn, M is a super fuzzy


neutrosophic mixed matrix. X = X1 ∪ X2 ∪ … ∪ Xn is
a super fuzzy mixed row vector. Using the special
sp sp
operator ° find a C or C++ program for X ° M = Y';
sp
after thresholding Y' we get Y find Y ° M = Z'; after
sp
updating and thresholding Z' we get Z find Z ° M and
so on until we get T a special fixed point or a special
limit cycle.

Hint: Refer [228, 231] for a similar procedure.

264
FURTHER READING
We have given a long list for further reading so that the
interested reader can make use of them:

1. Adamopoulos, G.I., and Pappis, C.P., Some Results on the


Resolution of Fuzzy Relation Equations, Fuzzy Sets and
Systems, 60 (1993) 83-88.

2. Adams, E.S., and Farber, D.A. Beyond the Formalism Debate:


Expert Reasoning, Fuzzy Logic and Complex Statutes,
Vanderbilt Law Review, 52 (1999) 1243-1340.
http://law.vanderbilt.edu/lawreview/vol525/adams.pdf
3. Adlassnig, K.P., Fuzzy Set Theory in Medical Diagnosis, IEEE
Trans. Systems, Man, Cybernetics, 16 (1986) 260-265.
4. Akiyama, Y., Abe, T., Mitsunaga, T., and Koga, H., A
Conceptual Study of Max-composition on the Correspondence
of Base Spaces and its Applications in Determining Fuzzy
Relations, Japanese J. of Fuzzy Theory Systems, 3 (1991) 113-
132.

5. Allen, J., Bhattacharya, S. and Smarandache, F. Fuzziness and


Funds allocation in Portfolio Optimization.
http://lanl.arxiv.org/ftp/math/papers/0203/0203073.pdf

6. Anitha, V. Application of Bidirectional Associative Memory


Model to Study Female Infanticide, Masters Dissertation,
Guide: Dr. W. B. Vasantha Kandasamy, Department of
Mathematics, Indian Institute of Technology, Chennai, March
2000.
7. Appel, K., and Haken, W., The solution of the four color map
problem, Scientific American, 237 (4), (1977) 108-121.

8. Ashbacher, C. Introduction to Neutrosophic Logic, American


Research Press, Rehoboth, 2002.
http://www.gallup.unm.edu/~smarandache/IntrodNeutLogic.pdf

265
9. Axelord, R. (ed.) Structure of Decision: The Cognitive Maps of
Political Elites, Princeton Univ. Press, New Jersey, 1976.
10. Balakrishnan, R., and Ranganathan, K., A textbook of Graph
Theory, Springer, 1999.

11. Balakrishnan, R., and Paulraja, P., Line graphs of subdivision


graphs, J. Combin. Info. and Sys. Sci., 10, (1985) 33-35.

12. Balu, M.S. Application of Fuzzy Theory to Indian Politics,


Masters Dissertation, Guide: Dr. W. B. Vasantha Kandasamy,
Department of Mathematics, Indian Institute of Technology,
April 2001.

13. Banini, G.A., and Bearman, R.A., Application of Fuzzy


Cognitive Maps to Factors Affecting Slurry Rheology, Int. J. of
Mineral Processing, 52 (1998) 233-244.

14. Bechtel, J.H., An Innovative Knowledge Based System using


Fuzzy Cognitive Maps for Command and Control, Storming
Media, Nov 1997. http://www.stormingmedia.us/cgi-
bin/32/3271/A327183.php
15. Bezdek, J.C., Pattern Recognition with Fuzzy Objective
Function Algorithm, Plenum Press, New York, 1981.
16. Birkhoff, G., Lattice Theory, American Mathematical Society,
1979.
17. Blanco, A., Delgado, M., and Requena, I., Solving Fuzzy
Relational Equations by Max-min Neural Network, Proc. 3rd
IEEE Internet Conf. On Fuzzy Systems, Orlando (1994) 1737-
1742.

18. Bohlen, M. and Mateas, M., Office Plant #1.


http://www.acsu.buffalo.edu/~mrbohlen/pdf/leonardo.pdf
19. Bondy, J.A., and Murthy, U.S.R., Graph Theory and
applications, The MacMillan Press 1976.

20. Bondy, J.A., Pancyclic graphs, J. Combin Theory Ser., 11


(1971) 80-84.

266
21. Bougon, M.G., Congregate Cognitive Maps: A Unified
Dynamic Theory of Organization and Strategy, J. of
Management Studies, 29 (1992) 369-389.

22. Brannback, M., Alback, L., Finne, T. and Rantanen, R..


Cognitive Maps: An Attempt to Trace Mind and Attention in
Decision Making, in C. Carlsson ed. Cognitive Maps and
Strategic Thinking, Meddelanden Fran Ekonomisk
Statsvetenskapliga Fakulteten vid Abo Akademi Ser. A 442
(1995) 5-25.

23. Brown, S.M. Cognitive Mapping and Repertory Grids for


Qualitative Survey Research: Some Comparative Observations,
J. of Management Studies, 29 (1992) 287-307.

24. Brubaker, D. Fuzzy Cognitive Maps, EDN ACCESS, 11 April


1996. http://www.e-
insite.net/ednmag/archives/1996/041196/08column.htm

25. Brubaker, D. More on Fuzzy Cognitive Maps, EDN ACCESS,


25 April 1996. http://www.e-
insite.net/ednmag/archives/1996/042596/09column.htm
26. Buckley, J.J., and Hayashi, Y., Fuzzy Neural Networks: A
Survey, Fuzzy Sets and Systems, 66 (1994) 1-13.

27. Carley, K. An Approach for Relating Social Structure to


Cognitive Structure, J. of Math. Sociology, 12 (1986) 137-189.

28. Carlsson, C. Cognitive Maps and Hyper-knowledge: A


Blueprint for Active Decision Support Systems. In Cognitive
Maps and Strategic Thinking, Carlsson, C. ed., Meddelanden
Fran Ekonomisk – Statesvetenkapliga Fakulteten Vid Abo
Akademi, IAMSR, Ser.A 442, (1995) 27-59.

29. Carlsson, C., and Fuller, R., Adaptive Fuzzy Cognitive Maps
for Hyper-knowledge Representation in Strategy Formation
Process In Proceedings of the International Panel Conference
on Soft and Intelligent Computing, Technical Univ. of
Budapest, (1996) 43-50. http://www.abo.fi/~rfuller/asic96.pdf

267
30. Carobs M. and Price, K., Intrusion detection systems.
http://www.cerias.purdue.edu/coast/coast-library.html

31. Carvalho, J.P., and Jose A. B. Tomè. Rule based Fuzzy


Cognitive Maps -- Fuzzy Causal Relations, Computational
Intelligence Modelling, Control and Automaton, Edited by
M.Mohammadian, 1999.
http://digitais.ist.utl.pt/uke/papers/cimca99rbfcm.pdf

32. Carvalho, J.P., and Jose A.B. Tomè. Fuzzy Mechanisms for
Causal Relations. In Proceedings of the 8th International Fuzzy
Systems Association World Congress, IFSA '99, Taiwan.
http://digitais.ist.utl.pt/uke/papers/IFSA99fmcr.pdf

33. Carvalho, J.P., and Jose A.B. Tomè. Rule based Fuzzy
Cognitive Maps: Expressing Time in Qualitative System
Dynamics.
http://digitais.ist.utl.pt/uke/papers/FUZZIEEE2001P089-
RBFCMExpressingTimeinQualitativeSystemDynamics.pdf

34. Carvalho, J.P., and Jose A.B. Tomè. Rule based Fuzzy
Cognitive Maps – Qualitative Systems Dynamics. In Proc. of
the 19th International Conference of the North American Fuzzy
Information Processing Society, NAFIPS2000, Atlanta, 2000.
http://digitais.ist.utl.pt/uke/papers/NAFIPS2000QSD.pdf

35. Carvalho, J.P., and Jose A.B. Tomè. Rule-based Fuzzy


Cognitive Maps and Fuzzy Cognitive Maps – a Comparative
Study. In Proc. of the 18th International Conference of the North
American Fuzzy Information Processing Society, by NAFIPS,
New York, (1999) 115-119.
http://digitais.ist.utl.pt/uke/papers/NAFIPS99rbfcm-fcm.pdf

36. Caudill, M. Using Neural Nets: Fuzzy Cognitive Maps,


Artificial Intelligence Expert, 6 (1990) 49-53.
37. Cechiarova, K., Unique Solvability of Max-Min Fuzzy
Equations and Strong Regularity of Matrices over Fuzzy
Algebra, Fuzzy Sets and Systems, 75 (1995) 165-177.

268
38. Chartrand, G. and Wall, C.E., On the Hamiltonian Index of a
Graph, Studia Sci. Math. Hungar, 8 (1973) 43-48.

39. Chen, S.M., Cognitive map-based decision analysis on NPN


Logics, Fuzzy Sets and Systems, 71 (1995) 155-163.
40. Cheng, L., and Peng, B., The Fuzzy Relation Equation with
Union or Intersection Preserving Operator, Fuzzy Sets and
Systems, 25 (1988) 191-204.
41. Chung, F., and Lee, T., A New Look at Solving a System of
Fuzzy Relational Equations, Fuzzy Sets and Systems, 99 (1997)
343-353.
42. Chvatal, V., and Erdos, P., A note on Hamiltonian Circuits,
Discrete Maths (2) (1972) 111-113.

43. Craiger, J.P., Causal Structure, Model Inferences and Fuzzy


Cognitive Maps: Help for the Behavioral Scientist,
International Neural Network Society, Annual Meeting World
Congress Neural Networks, June 1994.

44. Craiger, J.P., and Coovert, M.D., Modeling Dynamic Social and
Psychological Processes with Fuzzy Cognitive Maps. In Proc.
of the 3rd IEEE Conference on Fuzzy Systems, 3 (1994) 1873-
1877.

45. Craiger, J.P., Weiss, R.J., Goodman, D.F., and Butler, A.A.,
Simulating Organizational Behaviour with Fuzzy Cognitive
Maps, Int. J. of Computational Intelligence and Organization, 1
(1996) 120-123.
46. Czogala, E., Drewniak, J., and Pedrycz, W., Fuzzy relation
Applications on Finite Set, Fuzzy Sets and Systems, 7 (1982) 89-
101.
47. De Jong, K.A., An Analysis of the Behavior of a Class of
Genetic Adaptive Systems, Dissertation Abstracts Internet, 86
(1975) 5140B.

269
48. Di Nola, A., and Sessa, S., On the Set of Composite Fuzzy
Relation Equations, Fuzzy Sets and Systems, 9 (1983) 275-285.
49. Di Nola, A., On Solving Relational Equations in Brouwerian
Lattices, Fuzzy Sets and Systems, 34 (1994) 365-376.
50. Di Nola, A., Pedrycz, W., and Sessa, S., Some Theoretical
Aspects of Fuzzy Relation Equations Describing Fuzzy System,
Inform Sci., 34 (1984) 261-264.
51. Di Nola, A., Pedrycz, W., Sessa, S., and Sanchez, E., Fuzzy
Relation Equations Theory as a Basis of Fuzzy Modeling: An
Overview, Fuzzy Sets and Systems, 40 (1991) 415-429.
52. Di Nola, A., Pedrycz, W., Sessa, S., and Wang, P.Z., Fuzzy
Relation Equation under a Class of Triangular Norms: A Survey
and New Results, Stochastica, 8 (1984) 99-145.
53. Di Nola, A., Relational Equations in Totally Ordered Lattices
and their Complete Resolution, J. Math. Appl., 107 (1985) 148-
155.
54. Di Nola, A., Sessa, S., Pedrycz, W., and Sanchez, E., Fuzzy
Relational Equations and their Application in Knowledge
Engineering, Kluwer Academic Publishers, Dordrecht, 1989.

55. Diamond, J., R. McLeod, and Pedrycz, A., A Fuzzy Cognitive


System: Examination of Referential Neural Architectures, in:
Proc. of the Int. Joint Conf. on Neural Networks, 2 (1990) 617-
622.

56. Dickerson, J.A., and Kosko, B., Adaptive Cognitive Maps in


Virtual Worlds, International Neural Network Society, World
Congress on Neural Networks, June 1994.

57. Dickerson, J.A., and Kosko, B., Virtual Worlds as Fuzzy


Cognitive Maps, Presence, 3 (1994) 173-189.

58. Dickerson, J.A., Cox, Z., Wurtele, E.S., and Fulmer, A.W.,
Creating Metabolic and Regulatory Network Models using
Fuzzy Cognitive Maps.

270
http://www.botany.iastate.edu/~mash/metnetex/NAFIPS01v3a.p
df
59. Drewniak, J., Equations in Classes of Fuzzy Relations, Fuzzy
Sets and Systems, 75 (1995) 215-228.
60. Drewniak, J., Fuzzy Relation Calculus, Univ. Slaski, Katowice,
1989.
61. Dubois, D., and Prade, H., Fuzzy Relation Equations and Causal
Reasoning, Fuzzy Sets and Systems, 75 (1995) 119-134.
62. Dumford, N., and Schwartz, J.T., Linear Operators Part I,
Interscience Publishers, New York, 1958.

63. Eden C., Cognitive Mapping, European J. of Operational


Research, 36 (1988) 1-13.

64. Eden, C., On the Nature of Cognitive Maps, J. of Management


Studies, 29 (1992) 261-265.

65. Eden, C., Ackerman, F., and Cropper, S., The Analysis of Cause
Maps, Journal of Management Studies, 29 (1992) 309-323.
66. Fang, S.C., and Li, G., Solving Fuzzy Relation Equations with a
Linear Objective Function, Fuzzy Sets and Systems, 103 (1999)
107-113.
67. Fang, S.C., and Puthenpurn, S., Linear Optimization and
Extensions: Theory and Algorithm, Prentice-Hall, New Jersey,
1993.
68. Fieischner, H., Eulerian Graphs and related topics, Annals of
Disc. Math, 45, North Holland, 1990.
69. Fiorini, S., and Wilson, R.J., Edge colorings of graphs, In
Research Notes in Mathematics, 16, Pittman, London, 1971.

70. Fuzzy Thought Amplifier. The Fuzzy Cognitive Map Program,


Fuzzy Systems Engineering, USA.
http://www.fuzzysys.com/ftaprod.html

271
71. Galichet, S., and Foulloy, L., Fuzzy Controllers: Synthesis and
Equivalences, IEEE Trans. Fuzzy Sets, (1995) 140- 145.
72. Gavalec, M., Solvability and Unique Solvability of Max-min
Fuzzy Equations. Fuzzy Sets and Systems, 124 (2001) 385-393.

73. Georgopoulous, V.C., Malandrak, G.A., and Stylios, C.D., A


Fuzzy Cognitive Map Approach to Differential Diagnosis of
Specific Language Impairment, Artificial Intelligence in
Medicine, 679 (2002) 1-18.

74. Goto, K. and Yamaguchi, T., Fuzzy Associative Memory


Application to a Plant Modeling, in Proc. of the International
Conference on Artificial Neural Networks, Espoo, Finland,
(1991) 1245-1248.
75. Gottwald, S., Approximate Solutions of Fuzzy Relational
Equations and a Characterization of t-norms that Define
Matrices for Fuzzy Sets, Fuzzy Sets and Systems, 75 (1995)
189-201.
76. Gottwald, S., Approximately Solving Fuzzy Relation Equations:
Some Mathematical Results and Some Heuristic Proposals,
Fuzzy Sets and Systems, 66 (1994) 175-193.
77. Guo, S.Z., Wang, P.Z., Di Nola, A., and Sessa, S., Further
Contributions to the Study of Finite Fuzzy Relation Equations,
Fuzzy Sets and Systems, 26 (1988) 93-104.
78. Gupta, M.M., and Qi, J., Design of Fuzzy Logic Controllers
based on Generalized T-operators, Fuzzy Sets and Systems, 40
(1991) 473-489.
79. Gupta, M.M., and Qi, J., Theory of T-norms and Fuzzy
Inference, Fuzzy Sets and Systems, 40 (1991) 431-450.
80. Gupta, M.M., and Rao, D.H., On the Principles of Fuzzy Neural
Networks, Fuzzy Sets and Systems, 61 (1994) 1-18.

81. Hadjiski, M. B., Christova, N.G., and Groumpos, P.P. Design of


hybrid models for complex systems.
http://www.erudit.de/erudit/events/esit99/12773_p.pdf

272
82. Hafner, V.V., Cognitive Maps for Navigation in Open
Environments, http://citeseer.nj.nec.com/hafner00cognitive.html

83. Hagiwara, M., Extended Fuzzy Cognitive Maps, Proc. IEEE


International Conference on Fuzzy Systems, (1992) 795-801.

84. Harary, F., Graph Theory, Narosa Publications (reprint, Indian


edition), New Delhi, 1969.

85. Hart, J.A., Cognitive Maps of Three Latin American Policy


Makers, World Politics, 30 (1977) 115-140.
86. Higashi, M., and Klir, G.J., Resolution of Finite Fuzzy Relation
Equations, Fuzzy Sets and Systems, 13 (1984) 65-82.
87. Hirota, K., and Pedrycz, W., Specificity Shift in Solving Fuzzy
Relational Equations, Fuzzy Sets and Systems, 106 (1999) 211-
220.
88. Holland, J., Adaptation in Natural and Artificial Systems, The
University of Michigan Press, Ann Arbor, 1975.
89. Holton, D.A., and Sheehan, J., The Petersen Graph, Australian
Math Soc. Lecture Series 7, Cambridge Univ Press 1993.
90. Hong, D.H., and Hwang, S.Y., On the Compositional Rule of
Inference under Triangular Norms, Fuzzy Sets and Systems, 66
(1994) 25-38.
91. Hormaifar, A., Lai, S., and Qi, X., Constrained Optimization via
Genetic Algorithm, Simulation, 62 (1994) 242-254.
92. Horst, P., Matrix Algebra for Social Scientists, Holt, Rinehart
and Winston, inc, 1963.
93. Imai, H., Kikuchi, K., and Miyakoshi, M., Unattainable
Solutions of a Fuzzy Relation Equation, Fuzzy Sets and
Systems, 99 (1998) 193-196.

94. Jang, S.R., Sun, C.T., and Mizutani, E., Neuro-fuzzy and Soft
Computing, Prentice-Hall, Englewood Cliffs, NJ, 1997.

273
95. Jefferies, M.E., and Yeap, W.K., The Utility of Global
Representations in a Cognitive Map.
http://www.cs.waikato.ac.nz/~mjeff/papers/COSIT2001.pdf
96. Jenei, S., On Archimedean Triangular Norms, Fuzzy Sets and
Systems, 99 (1998) 179-186.
97. Joines, J.A., and Houck, C., On the Use of Non-stationary
Penalty Function to Solve Nonlinear Constrained Optimization
Problems with Gas, Proc. 1st IEEE Internal Conf. Evolutionary
Computation, 2 (1994) 579-584.
98. Kagei, S., Fuzzy Relational Equation with Defuzzification
Algorithm for the Largest Solution, Fuzzy Sets and Systems, 123
(2001) 119-127.

99. Kamala, R. Personality Medicine model using Fuzzy


Associative Memories, Masters Dissertation, Guide: Dr. W. B.
Vasantha Kandasamy, Department of Mathematics, Indian
Institute of Technology, March 2000.

100. Kardaras, D., and Karakostas, B., The Use of Fuzzy Cognitive
maps to Stimulate the Information Systems Strategic Planning
Process, Information and Software Technology, 41 (1999) 197-
210.

101. Kardaras, D., and Mentzas, G., Using fuzzy cognitive maps to
model and analyze business performance assessment, In Prof. of
Int. Conf. on Advances in Industrial Engineering – Applications
and Practice II, Jacob Chen and Anil Milal (eds.), (1997) 63-68.

102. Khan, M.S., M. Quaddus, A. Intrapairot, and Chong, A.,


Modelling Data Warehouse Diffusion using Fuzzy Cognitive
Maps – A Comparison with the System Dynamics Approach.
http://wawisr01.uwa.edu.au/2000/Track%204/gModelling.PDF

103. Kim, H.S., and Lee, K.C., Fuzzy Implications of Fuzzy


Cognitive Maps with Emphasis on Fuzzy Causal Relations and
Fuzzy Partially Causal Relationship, Fuzzy Sets and Systems, 97
(1998) 303-313.

274
104. Kipersztok, O., Uncertainty Propagation in FCMs for Ranking
Decision Alternatives, Proceedings of the EUFIT 97, 5th
European Congress on Intelligent Techniques and Soft
Computing, September 08-11, (1997) 1555-1560.

105. Klein, J.H., and Cooper, D.F., Cognitive maps of Decision


Makers in a Complex Game, J. of the Oper. Res. Soc., 33 (1982)
63-71.
106. Klir, G.J., and Yuan, B., Fuzzy Sets and Fuzzy Logic: Theory
and Applications, Prentice-Hall, Englewood Cliffs NJ, 1995.

107. Komathi, P.V. Application of Fuzzy Theory to study old people


problem, Masters Dissertation, Guide: Dr.W.B.Vasantha
Kandasamy, Department of Mathematics, Indian Institute of
Technology, Chennai, April 1999.

108. Kosko, B., Fuzzy Cognitive Maps, Int. J. of Man-Machine


Studies, 24 (1986) 65-75.

109. Kosko, B., Fuzzy Thinking, Hyperion, 1993.

110. Kosko, B., Heaven in a chip: Fuzzy Visions of Society and


Science in the Digital Age, Three Rivers Press, November 2000.

111. Kosko, B., Hidden Patterns in Combined and Adaptive


Knowledge Networks, Proc. of the First IEEE International
Conference on Neural Networks (ICNN-86), 2 (1988) 377-393.

112. Kosko, B., Neural Networks and Fuzzy Systems: A Dynamical


Systems Approach to Machine Intelligence, Prentice Hall of
India, 1997.
113. Kuhn, T., The Structure of Scientific Revolutions, Univ. of
Chicago Press, 1962.
114. Kurano, M., Yasuda, M., Nakatami, J., and Yoshida, Y., A
Fuzzy Relational Equation in Dynamic Fuzzy Systems, Fuzzy
Sets and Systems, 101 (1999) 439-443.

275
115. Kurano, M., Yasuda, M., Nakatami, J., and Yoshida, Y., A
Limit Theorem in Some Dynamic Fuzzy Systems, Fuzzy Sets
and Systems, 51 (1992) 83- 88.
116. Kuratowski, K., Topology I, Academic Press New York 1966.

117. Langfield-Smith, K., Exploring the Need for a Shared Cognitive


Map, J. of Management Studies, 29 (1992) 349-367.

118. Laszlo, E., Artigiani, R., Combs, A., and Csanyi, V., Changing
Visions: Human Cognitive Maps: Past, Present and Future,
Greenwood Publishing House, 1996.
119. Lee, C.C., Fuzzy Logic in Control Systems: Fuzzy Logic
Controller, Part I and II, IEEE Trans. Systems, Man and
Cybernetics, 20 (1990) 404-405.
120. Lee, C.C., Theoretical and Linguistic Aspects of Fuzzy Logic
Controller, Automatica, 15 (1979) 553-577.

121. Lee, K., Kim, S, and Sakawa, M., On-line Fault Diagnosis by
Using Fuzzy Cognitive Maps, IEICE Transactions in
Fundamentals of Electronics, Communications and Computer
Sciences (JTC-CSCC '95), Sept. 18-21 1996, v E79-A, no. 6,
June 1996, 921-927.

122. Lee, K.C., Kim, H.S., and Chu, S.C., A Fuzzy Cognitive Map
Based Bi-directional Inference Mechanism: An Application to
Stock Investment Analysis, Proc. Japan/ Korea Joint Conf. on
Expert Systems, 10 (1994), 193-196.

123. Lee, K.C., Kim, J.S., Chang, N.H., and Kwon, S.J., Fuzzy
Cognitive Map Approach to Web-mining Inference
Amplification, Expert Systems with Applications, 22 (2002)
197-211.

124. Lee, K.C., Chu, S.C., and Kim, S.H., Fuzzy Cognitive Map-
based Knowledge Acquisition Algorithm: Applications to Stock
Investment Analysis, in W.Cheng, Ed., Selected Essays on
Decision Science (Dept. of Decision Science and Managerial

276
Economics), The Chinese University of Hong Kong, (1993)
129-142.

125. Lee, K.C., Lee, W.J., Kwon, O.B., Han, J.H., Yu, P.I., A
Strategic Planning Simulation Based on Fuzzy Cognitive Map
Knowledge and Differential Game, Simulation, 71 (1998) 316-
327.

126. Lee, T.D., and Yang, C.N., Many Body Problems in Quantum
Statistical Mechanics, Physical Review, 113 (1959) 1165-1177.
127. Lettieri, A., and Liguori, F., Characterization of Some Fuzzy
Relation Equations Provided with one Solution on a Finite Set,
Fuzzy Sets and Systems, 13 (1984) 83-94.

128. Lewin, K. Principles of Topological Psychology, McGraw Hill,


New York, 1936.
129. Li, G. and Fang, S.G., On the Resolution of Finite Fuzzy
Relation Equations, OR Report No.322, North Carolina State
University, Raleigh, North Carolina, 1986.
130. Li, X., and Ruan, D., Novel Neural Algorithm Based on Fuzzy
S-rules for Solving Fuzzy Relation Equations Part I, Fuzzy Sets
and Systems, 90 (1997) 11-23.
131. Li, X., and Ruan, D., Novel Neural Algorithms Based on Fuzzy
S-rules for Solving Fuzzy Relation Equations Part II, Fuzzy Sets
and Systems, 103 (1999) 473-486.
132. Li, X., and Ruan, D., Novel Neural Algorithm Based on Fuzzy
S-rules for Solving Fuzzy Relation Equations Part III, Fuzzy
Sets and Systems, 109 (2000) 355-362.
133. Li, X., Max-min Operator Network and Fuzzy S-rule, Proc. 2nd
National Meeting on Neural Networks, Nanjing, 1991.

134. Liu, F., and F. Smarandache. Intentionally and Unintentionally.


On Both, A and Non-A, in Neutrosophy.
http://lanl.arxiv.org/ftp/math/papers/0201/0201009.pdf

277
135. Liu, F., and F. Smarandache. Logic: A Misleading Concept. A
Contradiction Study toward Agent's Logic, in Proceedings of
the First International Conference on Neutrosophy,
Neutrosophic Logic, Neutrosophic Set, Neutrosophic
Probability and Statistics, Florentin Smarandache editor,
Xiquan, Phoenix, ISBN: 1-931233-55-1, 147 p., 2002, also
published in "Libertas Mathematica", University of Texas at
Arlington, 22 (2002) 175-187.
http://lanl.arxiv.org/ftp/math/papers/0211/0211465.pdf
136. Liu, F., and Smarandache, F., Intentionally and Unintentionally.
On Both, A and Non-A, in Neutrosophy.
http://lanl.arxiv.org/ftp/math/papers/0201/0201009.pdf
137. Loetamonphing, J., and Fang, S.C., Optimization of Fuzzy
Relation Equations with Max-product Composition, Fuzzy Sets
and Systems, 118 (2001) 509-517.
138. Loetamonphing, J., Fang, S.C., and Young, R.E., Multi
Objective Optimization Problems with Fuzzy Relation
Constraints, Fuzzy Sets and Systems, 127 (2002) 147-164.
139. Lu, J., An Expert System Based on Fuzzy Relation Equations
for PCS-1900 Cellular System Models, Proc. South-eastern
INFORMS Conference, Myrtle Beach SC, Oct 1998.
140. Lu, J., and Fang, S.C., Solving Nonlinear Optimization
Problems with Fuzzy Relation Equation Constraints, Fuzzy Sets
and Systems, 119 (2001) 1-20.
141. Luo, C.Z., Reachable Solution Set of a Fuzzy Relation
Equation, J. of Math. Anal. Appl., 103 (1984) 524-532.
142. Luoh, L., Wang, W.J., Liaw, Y.K., New Algorithms for Solving
Fuzzy Relation Equations, Mathematics and Computers in
Simulation, 59 (2002) 329-333.

143. Madhan, N. Rule Based Control System to Study the


Performance Aspects of School Students, Masters Dissertation,
Guide: Dr.W.B.Vasantha Kandasamy, Department of

278
Mathematics, Indian Institute of Technology, Chennai, April
2001.

144. Meghabghab, G. Fuzzy Cognitive State Map vs. Markovian


Modeling of User's Web Behaviour, Invited Paper,
International Journal of Computation Cognition,
(http://www.YangSky.com/yangijcc.htm) 1 (Sept. 2003), 51-92.
Article published electronically on December 5, 2002).

145. Miao, Y., and Liu, Z., Dynamical Cognitive Network as an


extension of Fuzzy Cognitive Map in Proc. Int. Conf. Tools
Artificial Intelligence, Chicago, IL, November 1999.
146. Michalewicz, Z. and Janikow, Z., Genetic Algorithms for
Numerical Optimization, Stats. Comput., 1 (1991) 75-91.
147. Michalewicz, Z., and Janikow, Z., Handling Constraints in
Genetic Algorithms, Kaufmann Publishers, Los Altos CA, 1991.
148. Michalewicz, Z., Genetic Algorithms + Data Structures =
Evolution Programs, Springer, New York, 1994.
149. Miyakoshi, M., and Shimbo, M., Sets of Solution Set
Equivalent Coefficient Matrices of Fuzzy Relation Equation,
Fuzzy Sets and Systems, 35 (1990) 357-387.
150. Miyakoshi, M., and Shimbo, M., Sets of Solution Set Invariant
Coefficient Matrices of Simple Fuzzy Relation Equations, Fuzzy
Sets and Systems, 21 (1987) 59-83.
151. Miyakoshi, M., and Shimbo, M., Solutions of Fuzzy Relational
Equations with Triangular Norms, Fuzzy Sets and Systems, 16
(1985) 53-63.
152. Mizumoto, M., and Zimmermann, H.J., Comparisons of Fuzzy
Reasoning Methods, Fuzzy Sets and Systems, 8 (1982) 253-283.
153. Mizumoto, M., Are Max and Min Suitable Operators for Fuzzy
Control Methods?, Proc. 6th IFSA World Congress I, Sao Paulo,
Brazil, (1995) 497-500.

279
154. Mohr, S.T., The Use and Interpretation of Fuzzy Cognitive
Maps, Master Thesis Project, Rensselaer Polytechnic Inst. 1997,
http://www.voicenet.com/~smohr/fcm_white.html

155. Montazemi, A.R., and Conrath, D.W., The Use of Cognitive


Mapping for Information Requirements Analysis, MIS
Quarterly, 10 (1986) 45-55.
156. Mycielski, J., Surle Coloriage des graph, Colloq math, 3 (1955)
161-162.

157. Ndousse, T.D., and Okuda, T., Computational Intelligence for


Distributed Fault Management in Networks using Fuzzy
Cognitive Maps, In Proc. of the IEEE International Conference
on Communications Converging Technologies for Tomorrow's
Application, (1996) 1558-1562.
158. Neundorf, D., and Bohm, R., Solvability Criteria for Systems of
Fuzzy Relation Equations, Fuzzy Sets and Systems, 80 (1996)
345-352.
159. Ngayen, H.T., A Note on the Extension Principle for Fuzzy
Sets, J. Math Anal. Appl., 64 (1978) 369-380.

160. Nozicka, G., and Bonha, G., and Shapiro, M., Simulation
Techniques, in Structure of Decision: The Cognitive Maps of
Political Elites, R. Axelrod ed., Princeton University Press,
(1976) 349-359.
161. Oberly, D.J., and Summer, P.P., Every connected locally
connected non trivial graph with no induced claw is
Hamiltonian, J. Graph Theory, 3 (1979) 351-356.

162. Ore, O., The Four Colour Problem, Academic Press New York
1967.

163. Ozesmi, U., Ecosystems in the Mind: Fuzzy Cognitive Maps of


the Kizilirmak Delta Wetlands in Turkey, Ph.D. Dissertation
titled Conservation Strategies for Sustainable Resource use in
the Kizilirmak Delta- Turkey, University of Minnesota, (1999)

280
144-185. http://env.erciyes.edu.tr/Kizilirmak/
UODissertation/uozesmi5.pdf

164. Park, K.S., and Kim, S.H., Fuzzy Cognitive Maps Considering
Time Relationships, Int. J. Human Computer Studies, 42 (1995)
157-162.
165. Pavlica, V., and Petrovacki, D., About Simple Fuzzy Control
and Fuzzy Control Based on Fuzzy Relational Equations, Fuzzy
Sets and Systems, 101 (1999) 41-47.
166. Pedrycz, W., Fuzzy Control and Fuzzy Systems, Wiley, New
York, 1989.
167. Pedrycz, W., Fuzzy Relational Equations with Generalized
Connectives and their Applications, Fuzzy Sets and Systems, 10
(1983) 185-201.
168. Pedrycz, W., Genetic Algorithms for Learning in Fuzzy
Relational Structures, Fuzzy Sets and Systems, 69 (1995) 37-52.
169. Pedrycz, W., Inverse Problem in Fuzzy Relational Equations,
Fuzzy Sets and Systems, 36 (1990) 277-291.
170. Pedrycz, W., Processing in Relational Structures: Fuzzy
Relational Equations, Fuzzy Sets and Systems, 25 (1991) 77-
106.
171. Pedrycz, W., s-t Fuzzy Relational Equations, Fuzzy Sets and
Systems, 59 (1993) 189-195.

172. Pelaez, C.E., and Bowles, J.B., Applying Fuzzy Cognitive Maps
Knowledge Representation to Failure Modes Effects Analysis,
In Proc. of the IEEE Annual Symposium on Reliability and
Maintainability, (1995) 450-456.

173. Pelaez, C.E., and Bowles, J.B., Using Fuzzy Cognitive Maps as
a System Model for Failure Modes and Effects Analysis,
Information Sciences, 88 (1996) 177-199.

174. Praseetha, V.R. A New Class of Fuzzy Relation Equation and its
Application to a Transportation Problem, Masters Dissertation,

281
Guide: Dr. W. B. Vasantha Kandasamy, Department of
Mathematics, Indian Institute of Technology, April 2000.
175. Prevot, M., Algorithm for the Solution of Fuzzy Relation, Fuzzy
Sets and Systems, 5 (1976) 38-48.

176. Ram Kishore, M. Symptom disease model in children, Masters


Dissertation, Guide: Dr. W. B. Vasantha Kandasamy,
Department of Mathematics, Indian Institute of Technology,
Chennai, April 1999.

177. Ramathilagam, S. Mathematical Approach to the Cement


Industry problems using Fuzzy Theory, Ph.D. Dissertation,
Guide: Dr. W. B. Vasantha Kandasamy, Department of
Mathematics, Indian Institute of Technology, Madras,
November 2002.

178. Reponen, T., Parnisto, J., and Viitanen, J., Personality's Impact
on Information Management Strategy Formulation and Decision
Making, in Cognitive Maps and Strategic Thinking: Carlsson,
C., ed. Meddelanden Fran Ekonomisk Statsvetenkapliga
Fakulteten Vid Abo Akademi, IAMSR, Ser. A: 442 (1995) 115-
139.

179. Russian Status Report, The 21st Joint Coordinating Forum


IARP, Nov. 6-9, 2002, Moscow.
180. Sanchez, E., Resolution of Composite Fuzzy Relation Equation,
Inform. and Control, 30 (1976) 38-48.
181. Schitkowski, K., More Test Examples for Non-linear
Programming Codes, Lecture Notes in Economics and
Mathematical Systems: 282, Springer, New York, 1987.

182. Schneider, M., Shnaider, E., Kandel, A., and Chew, G.,
Automatic Construction of FCMs, Fuzzy Sets and Systems, 93
(1998) 161-172.
183. Sessa, S., Some Results in the Setting of Fuzzy Relation
Equation Theory, Fuzzy Sets and Systems, 14 (1984) 217-248.

282
184. Sheetz, S.D., Tegarden, D.P., Kozar, K.A., and Zigurs, I., Group
Support Systems Approach to Cognitive Mapping, Journal of
Management Information Systems, 11 (1994) 31-57.

185. Silva, P.C. Fuzzy Cognitive Maps over Possible Worlds, Proc.
of the 1995 IEEE International Conference on Fuzzy Systems, 2
(1995) 555-560.

186. Siraj, A., Bridges, S.M., and Vaughn, R.B., Fuzzy cognitive
maps for decision support in an intelligent intrusion detection
systems, www.cs.msstate.edu/~bridges/papers/nafips2001.pdf

187. Smarandache, F. (editor), Proceedings of the First International


Conference on Neutrosophy, Neutrosophic Set, Neutrosophic
Probability and Statistics, Univ. of New Mexico – Gallup,
2001.
http://www.gallup.unm.edu/~smarandache/NeutrosophicProcee
dings.pdf

188. Smarandache, F. A Unifying Field in Logics: Neutrosophic


Logic, Preface by Charles Le, American Research Press,
Rehoboth, 1999, 2000. Second edition of the Proceedings of the
First International Conference on Neutrosophy, Neutrosophic
Logic, Neutrosophic Set, Neutrosophic Probability and
Statistics, University of New Mexico, Gallup, 1-3 December
2001. http://www.gallup.unm.edu/~smarandache/eBook-
Neutrosophics2.pdf

189. Smarandache, F. Neutrosophic Logic - Generalization of the


Intuitionistic Fuzzy Logic, presented at the Special Session on
Intuitionistic Fuzzy Sets and Related Concepts, of International
EUSFLAT Conference, Zittau, Germany, 10-12 September
2003. http://lanl.arxiv.org/ftp/math/papers/0303/0303009.pdf
190. Smarandache, F., Collected Papers III, Editura Abaddaba,
Oradea, 2000.
http://www.gallup.unm.edu/~smarandache/CP3.pdf

191. Smith, E., and Eloff, J., Cognitive Fuzzy Modeling for
Enhanced Risk Assessment in Health Care Institutions, IEEE

283
Intelligent Systems and their Applications, March/April 2000,
69-75.
192. Stamou, G.B., and Tzafestas, S.G., Neural Fuzzy Relational
Systems with New Learning Algorithm, Mathematics and
Computers in Simulation, 51 (2000) 301-314.
193. Stamou, G.B., and Tzafestas, S.G., Resolution of Composite
Fuzzy Relation Equations based on Archimedean Triangular
Norms, Fuzzy Sets and Systems, 120 (2001) 395-407.
194. Steuer, R.E., Multiple Criteria Optimization Theory:
Computation and Applications, Wiley, New York, 1986.

195. Styblinski, M.A., and Meyer, B.D., Fuzzy Cognitive Maps,


Signal Flow Graphs, and Qualitative Circuit Analysis, in Proc.
of the 2nd IEEE International Conference on Neural Networks
(ICNN – 87), San Diego, California (1988) 549-556.

196. Styblinski, M.A., and Meyer B.D., Signal Flow Graphs versus
Fuzzy Cognitive Maps in Applications to Qualitative Circuit
Analysis, Int. J. of Man-machine Studies, 18 (1991) 175-186.

197. Stylios, C.D., and. Groumpos, P.P., A Soft Computing


Approach for Modelling the Supervisory of Manufacturing
Systems, Journal of Intelligent and Robotic Systems, 26 (1999)
389-403.

198. Stylios, C.D., and Groumpos, P.P., Fuzzy Cognitive Maps: a


Soft Computing Technique for Intelligent Control, in Proc. of
the 2000 IEEE International Symposium on Intelligent Control
held in Patras, Greece, July 2000, 97-102.

199. Stylios, C.D., and Groumpos, P.P., The Challenge of Modeling


Supervisory Systems using Fuzzy Cognitive Maps, J. of
Intelligent Manufacturing, 9 (1998) 339-345.

200. Stylios, C.D., Georgopoulos, V.C., and Groumpos, P.P.,


Introducing the Theory of Fuzzy Cognitive Maps in Distributed
Systems, in Proc. of the Twelfth IEEE International Symposium
on Intelligent Control, 16-18 July, Istanbul, Turkey, 55-60.

284
201. Subhaashree, S. Application of Fuzzy Logic to Unemployment
Problem, Masters Dissertation, Guide: Dr. W. B. Vasantha
Kandasamy, Department of Mathematics, Indian Institute of
Technology, April 2001.
202. Sugeno, M., Industrial Applications of Fuzzy Control, Elsevier,
New York, 1985.

203. Taber W. R., Fuzzy Cognitive Maps Model Social Systems,


Artificial Intelligence Expert, 9 (1994) 18-23.

204. Taber, W.R., Knowledge Processing with Fuzzy Cognitive


Maps, Expert System with Applications, 2 (1991) 83-87.

205. Taber, W.R., and Siegel, M.A., Estimation of Expert Weights


using Fuzzy Cognitive Maps, in Proc. of the First IEEE
International Conference on Neural Networks, (ICNN-86)
1987, 319-325.

206. Tolman, E.C., Cognitive Maps in Rats and Men, Psychological


Review, 55 (1948) 189-208.

207. Tsadiras, A.K., and Margaritis, K.G., A New Balance Degree


for Fuzzy Cognitive Maps,
http://www.erudit.de/erudit/events/esit99/12594_p.pdf

208. Tsadiras, A.K., and Margaritis, K.G., Cognitive Mapping and


Certainty Neuron Fuzzy Cognitive Maps, Information Sciences,
101 (1997) 109-130.

209. Tsadiras, A.K., and Margaritis, K.G., Introducing Memory and


Decay Factors in Fuzzy Cognitive Maps, in First International
Symposium on Fuzzy Logic (ISFL '95), Zurich, Switzerland,
May 1995, B2-B9.

210. Tsadiras, A.K., and Margaritis, K.G., Using Certainty Neurons


in Fuzzy Cognitive Maps, Neural Network World, 6 (1996) 719-
728.

285
211. Tsukamoto, Y., An Approach to Fuzzy Reasoning Methods, in:
M.Gupta, R. Ragade and R. Yager (eds.), Advances in Fuzzy Set
Theory and Applications, North-Holland, Amsterdam, (1979)
137-149.
212. Ukai, S., and Kaguei, S., Automatic Accompaniment
Performance System using Fuzzy Inference, Proc. Sino Japan
Joint Meeting: Fuzzy sets and Systems, Beijing, E1-5(1990) 1-4.
213. Ukai, S., and Kaguei, S., Automatic Generation of
Accompaniment Chords using Fuzzy Inference, J. Japan Soc.
Fuzzy Theory Systems, 3 (1991) 377-381.

214. Uma, S., Estimation of Expert Weights using Fuzzy Cognitive


Maps, Masters Dissertation, Guide: Dr. W.B.Vasantha
Kandasamy, Department of Mathematics, Indian Institute of
Technology, Chennai, March 1997.
215. Vasantha Kandasamy, W.B., Love. Life. Lust. Loss: 101 Real
Life Stories of Migration and AIDS—A Fuzzy Analysis, Tamil
Nadu State AIDS Control Society, Chennai, 2003.

216. Vasantha Kandasamy, W.B., Women. Work. Worth. Womb., 101


real life stories of rural women, Tamil Nadu State AIDS
Control Society, Chennai, 2005.

217. Vasantha Kandasamy, W.B., Public. Pity. Patients. Peace.,


Public opinion on HIV/AIDS, Tamil Nadu State AIDS Control
Society, Chennai, 2006.

218. Vasantha Kandasamy, W.B., and Minor, A., Estimation of


Production and Loss or Gain to Industries Using Matrices, Proc.
of the National Conf. on Challenges of the 21st century in
Mathematics and its allied topics, Feb. 3-4, 2001, Univ. of
Mysore, 211-218.
219. Vasantha Kandasamy, W.B., and Balu, M. S., Use of Weighted
Multi-Expert Neural Network System to Study the Indian
Politics, Varahimir J. of Math. Sci., 2 (2002) 44-53.

286
220. Vasantha Kandasamy, W.B., and Indra, V., Maximizing the
passengers comfort in the madras transport corporation using
fuzzy programming, Progress of Mat., Banaras Hindu Univ., 32
(1998) 91-134.
221. Vasantha Kandasamy, W.B., and Karuppusamy, Environmental
pollution by dyeing industries: A FAM analysis, Ultra Sci,
18(3) (2006) 541-546.

222. Vasantha Kandasamy, W.B., and Mary John, M. Fuzzy


Analysis to Study the Pollution and the Disease Caused by
Hazardous Waste From Textile Industries, Ultra Sci, 14 (2002)
248-251.

223. Vasantha Kandasamy, W.B., and Ram Kishore, M. Symptom-


Disease Model in Children using FCM, Ultra Sci., 11 (1999)
318-324.

224. Vasantha Kandasamy, W.B., and Pramod, P., Parent Children


Model using FCM to Study Dropouts in Primary Education,
Ultra Sci., 13, (2000) 174-183.
225. Vasantha Kandasamy, W.B., and Praseetha, R., New Fuzzy
Relation Equations to Estimate the Peak Hours of the Day for
Transport Systems, J. of Bihar Math. Soc., 20 (2000) 1-14.

226. Vasantha Kandasamy, W.B., and Uma, S. Combined Fuzzy


Cognitive Map of Socio-Economic Model, Appl. Sci.
Periodical, 2 (2000) 25-27.

227. Vasantha Kandasamy, W.B., and Uma, S. Fuzzy Cognitive Map


of Socio-Economic Model, Appl. Sci. Periodical, 1 (1999) 129-
136.
228. Vasantha Kandasamy, W.B., and Smarandache, F., Analysis of
social aspects of migrant labourers living with HIV/AIDS using
fuzzy theory and neutrosophic cognitive maps, Xiquan, Phoenix,
2004.

287
229. Vasantha Kandasamy, W.B., and Smarandache, F., Basic
Neutrosophic algebraic structures and their applications to
fuzzy and Neutrosophic models, Hexis, Church Rock, 2004

230. Vasantha Kandasamy, W.B., and Smarandache, F., Fuzzy and


Neutrosophic analysis of Periyar’s views on untouchability,
Hexis, Phoenix, 2005.

231. Vasantha Kandasamy, W.B., and Smarandache, F., Fuzzy


Cognitive Maps and Neutrosophic Cognitive Maps, Xiquan,
Phoenix, 2003.

232. Vasantha Kandasamy, W.B., and Smarandache, F., Fuzzy


Relational Equations and Neutrosophic Relational Equations,
Neutrosophic Book Series 3, Hexis, Church Rock, USA, 2004.

233. Vasantha Kandasamy, W.B., and Smarandache, F., Introduction


to n-adaptive fuzzy models to analyse Public opinion on AIDS,
Hexis, Phoenix, 2006.

234. Vasantha Kandasamy, W.B., and Smarandache, F., Vedic


Mathematics: ‘Vedic’ or ‘mathematics’ A Fuzzy and
Neutrosophic Analysis, automaton, Los Angeles, 2006.

235. Vasantha Kandasamy, W.B., Smarandache, F., and Ilanthenral,


K., Elementary Fuzzy matrix theory and fuzzy models for social
scientists, automaton, Los Angeles, 2006.

236. Vasantha Kandasamy, W.B., and Anitha, V., Studies on Female


Infanticide Problem using Neural Networks BAM-model, Ultra
Sci., 13 (2001) 174-183.

237. Vasantha Kandasamy, W.B., and Indra, V., Applications of


Fuzzy Cognitive Maps to Determine the Maximum Utility of a
Route, J. of Fuzzy Maths, publ. by the Int. fuzzy Mat. Inst., 8
(2000) 65-77.
238. Vasantha Kandasamy, W.B., and Victor Devadoss,
Identification of the maximum age group in which the
agricultural labourers suffer health hazards due to chemical

288
Pollution using fuzzy matrix, Dimension of pollution, 3 (2005)
1-55.

239. Vasantha Kandasamy, W.B., and Yasmin Sultana, FRM to


Analyse the Employee-Employer Relationship Model, J. Bihar
Math. Soc., 21 (2001) 25-34.

240. Vasantha Kandasamy, W.B., and Yasmin Sultana, Knowledge


Processing Using Fuzzy Relational Maps, Ultra Sci., 12 (2000)
242-245.

241. Vasantha Kandasamy, W.B., Mary John, M., and Kanagamuthu,


T., Study of Social Interaction and Woman Empowerment
Relative to HIV/AIDS, Maths Tiger, 1(4) (2002) 4-7.

242. Vasantha Kandasamy, W.B., Neelakantan, N.R., and


Ramathilagam, S., Maximize the Production of Cement
Industries by the Maximum Satisfaction of Employees using
Fuzzy Matrix, Ultra Science, 15 (2003) 45-56.
243. Vasantha Kandasamy, W.B., Neelakantan, N.R., and Kannan,
S.R., Operability Study on Decision Tables in a Chemical Plant
using Hierarchical Genetic Fuzzy Control Algorithms, Vikram
Mathematical Journal, 19 (1999) 48-59.
244. Vasantha Kandasamy, W.B., Neelakantan, N.R., and Kannan,
S.R., Replacement of Algebraic Linear Equations by Fuzzy
Relation Equations in Chemical Engineering, In Recent Trends
in Mathematical Sciences, Proc. of Int. Conf. on Recent
Advances in Mathematical Sciences held at IIT Kharagpur on
Dec. 20-22, 2001, published by Narosa Publishing House,
(2001) 161-168.
245. Vasantha Kandasamy, W.B., Neelakantan, N.R., and
Ramathilagam, S., Use of Fuzzy Neural Networks to Study the
Proper Proportions of Raw Material Mix in Cement Plants,
Varahmihir J. Math. Sci., 2 (2002) 231-246.

246. Vasantha Kandasamy, W.B., Pathinathan, and Narayan Murthy.


Child Labour Problem using Bi-directional Associative

289
Memories (BAM) Model, Proc. of the 9th National Conf. of the
Vijnana Parishad of India on Applied and Industrial
Mathematics held at Netaji Subhas Inst. of Tech. on Feb. 22-24,
2002.

247. Vasantha Kandasamy, W.B., Ramathilagam, S., and


Neelakantan, N.R., Fuzzy Optimisation Techniques in Kiln
Process, Proc. of the National Conf. on Challenges of the 21st
century in Mathematics and its allied topics, Feb. 3-4 (2001),
Univ. of Mysore, (2001) 277-287.

248. Vazquez, A., A Balanced Differential Learning Algorithm in


Fuzzy Cognitive Map
http://monet.aber.ac.uk:8080/monet/docs/pdf_files/qr_02/qr200
2alberto-vazquez.pdf

249. Venkatbabu, Indra. Mathematical Approach to the Passenger


Transportation Problem using Fuzzy Theory, Ph.D.
Dissertation, Guide: Dr. W. B. Vasantha Kandasamy,
Department of Mathematics, Indian Institute of Technology,
Chennai, June 1998.

250. Vysoký, P. Fuzzy Cognitive Maps and their Applications in


Medical Diagnostics.
http://www.cbmi.cvut.cz/lab/publikace/30/Vys98_11.doc
251. Wagenknecht, M., and Hatmasann, K., On the Existence of
Minimal Solutions for Fuzzy Equations with Tolerances, Fuzzy
Sets and Systems, 34 (1990) 237-244.
252. Wagenknecht, M., On Pseudo Transitive Approximations of
fuzzy Relations, Fuzzy Sets and Systems, 44 (1991) 45-55.
253. Wang, H.F., An Algorithm for Solving Iterated Complete
Relation Equations, Proc. NAFIPS, (1988) 242-249.
254. Wang, H.F., Wu, C.W., Ho, C.H., and Hsieh, M.J., Diagnosis of
Gastric Cancer with Fuzzy Pattern Recognition, J. Systems
Engg., 2 (1992) 151-163.

290
255. Wang, W.F., A Multi Objective Mathematical Programming
Problem with Fuzzy Relation Constraints, J. Math. Criteria
Dec. Anal., 4 (1995) 23-35.
256. Wang, X., Infinite Fuzzy Relational Equations on a Complete
Brouwerian Lattice, Fuzzy Sets and Systems, 138 (2003) 657-
666.
257. Wang, X., Method of Solution to Fuzzy Relation Equations in a
Complete Brouwerian Lattice, Fuzzy Sets and Systems, 120
(2001) 409-414.
258. Winston, W.L., Introduction to Mathematical Programming:
Applications and Algorithms, Daxbury Press, Belmont CA,
1995.

259. Wrightson, M.T., The Documentary Coding Method in R.


Axelrod ed., Structure of Decision: The Cognitive Maps of
Political Elites, Princeton Univ. Press, Princeton, NJ, (1976)
291-332.
260. Yager, R.R., Fuzzy Decision Making including Unequal
Objective, Fuzzy Sets and Systems, 1 (1978) 87-95.
261. Yager, R.R., On Ordered Weighted Averaging Operators in
Multi Criteria Decision Making, IEEE Trans. Systems, Man and
Cybernetics, 18 (1988) 183-190.

262. Yasmin Sultana, Construction of Employee-Employee


Relationship Model using Fuzzy Relational Maps, Masters
Dissertation, Guide: Dr. W. B. Vasantha Kandasamy,
Department of Mathematics, Indian Institute of Technology,
April 2000.
263. Yen, J., Langari, R., and Zadeh, L.A., Industrial Applications of
Fuzzy Logic and Intelligent Systems, IEEE Press, New York
1995.

264. Yuan, Miao and Zhi-Qiang Liu., On Causal Inference in Fuzzy


Cognitive Maps, IEEE Transactions on Fuzzy Systems, 81
(2000) 107-119.

291
265. Zadeh, L.A., A Theory of Approximate Reasoning, Machine
Intelligence, 9 (1979) 149- 194.
266. Zadeh, L.A., Similarity Relations and Fuzzy Orderings, Inform.
Sci., 3 (1971) 177-200.

267. Zhang, W.R., and Chen, S., A Logical Architecture for


Cognitive Maps, Proceedings of the 2nd IEEE Conference on
Neural Networks (ICNN-88), 1 (1988) 231-238.

268. Zhang, W.R., Chen, S.S., Wang, W., and King. R. S., A
Cognitive Map-Based Approach to the Coordination of
distributed cooperative agents, IEEE Trans. Systems Man
Cybernet, 22 (1992) 103-114.
269. Zimmermann, H.J., Fuzzy Set Theory and its Applications,
Kluwer, Boston, 1988.

292
INDEX

Absolute pure neutrosophic fuzzy number, 138


Acyclic FRM, 178
Acyclic NRMs, 203
Acyclic, 174
Acylic NCM, 197-98
Adjacency matrix, 174-5

Binary neutrosophic relations, 205

Causal relationships, 172, 175


Causalities, 172, 174
Combined FRM, 179-0
Combined Fuzzy Cognitive Maps (CFCMs), 223
Combined NCM, 199
Concepts, 172
Connection matrix, 174
Crisp binary relations, 181
Cut worthy property, 182
Cyclic, 174

Degree of false, 188


Degree of truth, 188
Diagonal entries, 174
Directed cycle of a NCM, 198

293
Directed cycle of FRM, 178
Directed cycle of NRM, 203
Directed cycle, 174
Directed edge, 174
Directed graph, 174
Domain space of a NRM, 202-3
Domain space of FRM, 177
Dynamical system of a FRM, 178
Dynamical system of a NCM, 198
Dynamical system, 175

Edge weights, 174


Edges, 172
Equilibrium state of a FRM, 179
Equilibrium state, 175
Events, 172

FCM with feed back, 175


Fixed binary pair, 23
Fixed point of a NCM, 198-99
Fixed point of NRM, 203-4
Fixed point, 21, 175
FRM with feed back, 178
Fuzzy Cognitive Maps (FCMs), 171-2
Fuzzy column vector /matrix, 9
Fuzzy compatibility relations, 180
Fuzzy interval, 7-8
Fuzzy matrices, 7-8
Fuzzy morphisms, 180
Fuzzy neutrosophic column vector/matrix, 100
Fuzzy neutrosophic interval, 99, 138
Fuzzy neutrosophic matrix, 98-9
Fuzzy neutrosophic row vector/matrix, 100, 139
Fuzzy nodes, 174
Fuzzy ordering relations, 180

294
Fuzzy relational equations, 258-9
Fuzzy Relational Maps (FRMs), 176-180
Fuzzy relations, 180
Fuzzy row vector /matrix, 9
Fuzzy sets, 174
Fuzzy unit column vector, 9-10
Fuzzy unit row vector, 9-10
Fuzzy zero column vector /matrix, 9-10
Fuzzy zero row vector /matrix, 9

Hidden pattern of a FRM, 179


Hidden pattern of a NCM, 198-99
Hidden pattern of NRM, 203-4
Hidden pattern, 175

Indeterminacy, 188
Instantaneous special state vector, 217
Instantaneous state vector, 174

Limit cycle of NRM, 203-4


Limit cycle, 21, 175
Limit point of a NCM, 198

Max min composition, 182


Max min operation, 14-18
Max min operator, 14-18
Max operation, 12-13
Maximum solution, 186
Membership matrices, 182-3
Membership neutrosophic matrix, 206
Min max operation, 7-8, 14-18

295
Min max operator, 14-18
Min operation, 12-13
Minimal solution, 186
Mixed neutrosophic integer, 99
Multiexpert special fuzzy model, 217

Neutrosophic abelian group, 192


Neutrosophic adjacency matrix, 197
Neutrosophic algebraic structures, 190
Neutrosophic algebraic sums, 215
Neutrosophic basis, 194
Neutrosophic Cognitive Maps (NCMs), 196-7
Neutrosophic column vector/matrix, 100
Neutrosophic diagonal matrix, 101
Neutrosophic dimension, 193-5
Neutrosophic drastic union, 214-5
Neutrosophic equivalence relation, 210-11
Neutrosophic equivalent, 211
Neutrosophic field of rationals, 191
Neutrosophic field of reals, 191
Neutrosophic field, 190-1
Neutrosophic intersections, 211, 213
Neutrosophic logic, 187-8
Neutrosophic matrices, 192
Neutrosophic matrix, 98-9
Neutrosophic quotient group, 193
Neutrosophic rectangular matrix, 99
Neutrosophic relation, 205
Neutrosophic Relational Maps (NRMs), 201-2
Neutrosophic relational matrix of a NRM, 202
Neutrosophic ring of integers, 99
Neutrosophic ring of rationals, 99
Neutrosophic ring of reals, 99
Neutrosophic row vector/matrix, 100
Neutrosophic sagittal diagram, 206, 209
Neutrosophic set, 205, 212
Neutrosophic space, 190

296
Neutrosophic square matrix, 99
Neutrosophic standard max-min composition, 207
Neutrosophic subfield, 191
Neutrosophic subgroup, 192
Neutrosophic unions, 213-14
Neutrosophic vector space, 191-7
Neutrosophic vector, 198-99
Neutrosophic zero matrix, 101
Nodes, 172

On-off position, 174

Policies, 172
Pure fuzzy neutrosophic number, 138
Pure neutrosophic integer, 99

Range space of FRM, 177


Range space of NRM, 202-3
Rectangular fuzzy matrix, 8
Relational matrix of FRM, 178
Resultant, 21
Revolutionary way, 175

Simple FCMs, 174


Simple FRMs, 178
Simple NCMs, 197
Simple NRMs, 202
Special attributes, 217
Special dynamical system of SFCM, 217
Special dynamical system of SMFRM, 244-5
Special fixed point of SMNCM, 230-1

297
Special fixed point of SFRM, 237-8
Special fixed point, 23, 39
Special Fuzzy Cognitive Maps (SFCMs), 216-7
Special fuzzy cognitive models, 45
Special fuzzy column matrix, 28-29
Special fuzzy dynamical system of SNCM, 229
Special fuzzy matrices, 7
Special fuzzy mixed column vector, 28-30
Special fuzzy mixed matrix, 34
Special fuzzy mixed rectangular matrix, 32
Special fuzzy mixed row vector, 28-30, 217
Special fuzzy mixed square matrix, 28-30
Special fuzzy models, 7
Special Fuzzy Neutrosophic Cognitive Maps model (SFNCMs
model), 231-2
Special fuzzy neutrosophic matrix, 138
Special fuzzy neutrosophic mixed row vector, 146
Special fuzzy neutrosophic mixed square matrix, 146
Special Fuzzy Neutrosophic Relational Maps (SFNRMs), 247
Special fuzzy operator, 39
Special fuzzy rectangular matrix, 32
Special Fuzzy Relational Equations (SFRE), 259-60
Special Fuzzy Relational Maps (SFRMs), 237-8
Special fuzzy row matrix, 27-28
Special fuzzy square matrix, 26-27
Special hidden pattern of SFRM, 237-8
Special hidden pattern of SNCM, 229
Special hidden pattern of the special dynamical system of
SMNCM, 230-1
Special hidden pattern, 23
Special limit cycle of SFRM, 237-8
Special limit cycle of SMNCM, 230-1
Special limit cycle, 23, 39
Special max min operator, 62-64, 70-72
Special min max operator, 58-1, 80-82, 85
Special Mixed Fuzzy Cognitive Fuzzy Relational Maps
(SMFCFRMs), 249-50
Special Mixed Fuzzy Cognitive Maps (SMFCMs), 223

298
Special Mixed Fuzzy Cognitive Relational Neutrosophic
Cognitive Relational Maps (SMFCRNCRMs), 250-2
Special Mixed Fuzzy Relational Equations (SMFRE), 259-60
Special Mixed Fuzzy Relational Maps (SMFRMs), 244
Special Mixed Neutrosophic Cognitive Maps (SMNCMs), 230
Special Mixed Neutrosophic Cognitive Neutrosophic Relational
Maps (SMNCNRMs), 250-1
Special Mixed Neutrosophic Relational Equations (SMNRE),
260-61
Special Mixed Neutrosophic Relational Maps (SMNRMs), 247-
8
Special Neutrosophic Cognitive Maps (SNCMs), 229
Special neutrosophic column vector, 109
Special neutrosophic mixed column vector, 109
Special neutrosophic mixed rectangular matrix, 107
Special neutrosophic mixed row vector, 108
Special neutrosophic mixed square matrix, 105-6
Special neutrosophic rectangular matrix, 105-6
Special Neutrosophic Relational Equations (SNRE), 260-61
Special Neutrosophic Relational Maps (SNRMs), 244-5
Special neutrosophic row vector, 108
Special neutrosophic square matrix, 103-4
Special nodes, 217
Special state vector of SFCM, 216-8
Special transpose of special fuzzy mixed matrix, 86-88
Square fuzzy matrix, 8
Standard neutrosophic union, 214-5
Strong neutrosophic dimension, 195
Super Special Hexagonal Model (SSHM), 261-62
Super special quadruple model, 262
Super special quintuple model, 262
Super special triple model, 262

Thresholding, 176
Transpose of a special fuzzy rectangular (mixed) matrix, 36
Transpose of a special fuzzy square (mixed) matrix, 36

299
U

Unique state vector, 175


Unit interval, 7
Updating and thresholding, 20-22
Updating, 176

Zero column vector, 9-10


Zero row vector, 9-10

300
ABOUT THE AUTHORS

Dr.W.B.Vasantha Kandasamy is an Associate Professor in the


Department of Mathematics, Indian Institute of Technology
Madras, Chennai. In the past decade she has guided 11 Ph.D.
scholars in the different fields of non-associative algebras,
algebraic coding theory, transportation theory, fuzzy groups, and
applications of fuzzy theory of the problems faced in chemical
industries and cement industries. Currently, four Ph.D. scholars
are working under her guidance.
She has to her credit 638 research papers. She has guided
over 57 M.Sc. and M.Tech. projects. She has worked in
collaboration projects with the Indian Space Research
Organization and with the Tamil Nadu State AIDS Control Society.
This is her 31st book.
On India's 60th Independence Day, Dr.Vasantha was
conferred the Kalpana Chawla Award for Courage and Daring
Enterprise by the State Government of Tamil Nadu in recognition
of her sustained fight for social justice in the Indian Institute of
Technology (IIT) Madras and for her contribution to mathematics.
(The award, instituted in the memory of Indian-American
astronaut Kalpana Chawla who died aboard Space Shuttle
Columbia). The award carried a cash prize of five lakh rupees (the
highest prize-money for any Indian award) and a gold medal.
She can be contacted at vasanthakandasamy@gmail.com
You can visit her on the web at: http://mat.iitm.ac.in/~wbv or:
http://www.vasantha.net

Dr. Florentin Smarandache is an Associate Professor of


Mathematics at the University of New Mexico in USA. He
published over 75 books and 100 articles and notes in
mathematics, physics, philosophy, psychology, literature, rebus.
In mathematics his research is in number theory, non-Euclidean
geometry, synthetic geometry, algebraic structures, statistics,
neutrosophic logic and set (generalizations of fuzzy logic and set
respectively), neutrosophic probability (generalization of classical
and imprecise probability). Also, small contributions to nuclear
and particle physics, information fusion, neutrosophy (a
generalization of dialectics), law of sensations and stimuli, etc.
He can be contacted at smarand@unm.edu

K. Ilanthenral is the editor of The Maths Tiger, Quarterly Journal of


Maths. She can be contacted at ilanthenral@gmail.com

301

Das könnte Ihnen auch gefallen