Omkar project report submitted m tec mechanical

Page 1



ACKNOWLEDGEMENT It gives me a great pleasure in presenting the Dissertation on “Evolutionary Algorithms

for

Multi-Objective

Optimization:

Modelling

and

Comparative

Evaluation�. The project work has certainly rendered me a tremendous learning. Apart from my efforts, the success of this project depends largely on the encouragement and guidelines of many others. I take this opportunity to express my gratitude to the people who have been instrumental in the successful completion of this project. First and foremost, I would like to thank to my guide, Dr. Satish S. Chinchanikar (Professor, Department of Mechanical Engineering, VIIT, Pune) for his valuable guidance and advice. My special thanks to my Co-guide, Mr. Mahendra G. Gadge (Assistant Professor, Department of Mechanical Engineering, VIIT, Pune) for his valuable guidance and advice. He inspired me greatly to work in this project. His willingness to motivate me contributed tremendously to this project. I would like to thank Dr. Atul P. Kulkarni (Associate Professor and Head, Department of Mechanical Engineering, VIIT, Pune) for allowing me to work on this project and giving me valuable guidance and advice in my Project. I am highly grateful to Dr. Bilavari S. Karkare (Principal) and I would also thank our Institution, faculties and technical staff of mechanical engineering department who helped me directly or indirectly during this project work. I also extend my heartfelt thanks to my family, siblings, all my friends and well-wishers.

Omkar Mahesh Manav

i


LIST OF FIGURE Fig. N0 Name of Figures

Page No

1.1

Hierarchy of Computational Intelligence

2

1.2

Approaches to Computational Intelligence

3

1.3

Global Search Optimization Hierarchy

6

1.4

Synergies of Computational Intelligence

7

1.5

Applied Optimization and Learning Methodology

12

3.1

Workflow for chapter 3

28

3.2

Evolutionary Model

33

3.3

NSGA II Algorithm

41

3.4

Rank and Pareto for Ra (35HRC)

50

3.5

Pareto-front for Tf (35HRC)

50

3.6

Average distance between consecutive generations (35HRC)

50

3.7

Rank and Pareto for Ra (45HRC)

52

3.8

Pareto-front for Tf (45HRC)

52

3.9

Average distance between consecutive generations (45HRC)

52

3.10

SPEA2 Algorithm

54

3.11

Rank and Pareto for Ra (35HRC)

58

3.12

Pareto-front for Tf (35HRC)

58

3.13

Average distance between consecutive generations (35HRC)

58

3.14

Rank and Pareto for Ra (45HRC)

59

3.15

Pareto-front for Tf (45HRC)

59

ii


3.16

Average distance between consecutive generations (45HRC)

60

3.17

PSO Algorithm

66

3.18

Pareto spread surface roughness and cutting force 35HRC

73

3.19

3D surface plot of optimal Ra with best position 35HRC

73

3.20

3D surface plot of Tf with best position 35HRC

74

3.21

Depth of cut influence on cutting forces 35 HRC

74

3.22

Pareto spread surface roughness and cutting force 45HRC

75

3.23

3D surface plot of optimal Ra with best position 45HRC

75

3.24

3D surface plot of Tf with best position 45 HRC

76

3.25

Depth of cut influence on cutting forces 45 HRC

76

3.26

Solution Spectrum for 35 hrc NSGA II

77

3.27

Solution Spectrum for 45 hrc NSGA II

77

3.28

Solution Spectrum for 35 hrc PSO

77

3.29

Solution Spectrum for 45 hrc PSO

77

3.30

Solution Spectrum for 35 hrc SPEA 2

77

3.31

Solution Spectrum for 45 hrc SPEA2

77

4.1

Workflow for Chapter 4

80

4.2

Simple network

82

4.3

Multi-layer feed forward network structure

85

4.4

Feed forward neural network for AISI 4340 Hard turning

88

Plots for NN 35 HRC 4.5

Performance plot of Network

89

iii


4.6

Training state of Network at each epoch

89

4.7

Training error in Ra

90

4.8

Regression fit plot for Ra

90

4.9

Training error in Ft

90

4.10

Regression fit plot for Ft

90

4.11

Training error in Fa

91

4.12

Regression fit plot for Fa

91

4.13

Training error in Fr

91

4.14

Regression fit plot for Fr

91

4.15

Training error in Tf

91

4.16

Regression fit plot for Tf

91

Plots for NN 45 HRC 4.17

Performance plot of Network

92

4.18

Training state of Network at each epoch

92

4.19

Training error in Ra

92

4.20

Regression fit plot for Ra

92

4.21

Training error in Ft

92

4.22

Regression fit plot for Ft

92

4.23

Training error in Fa

93

4.24

Regression fit plot for Fa

93

4.25

Training error in Fr

93

4.26

Regression fit plot for Fr

93

iv


4.27

Training error in Tf

93

4.28

Regression fit plot for Tf

93

4.29

ANFIS two input model

95

4.30

Applied ANFIS grid partitioning architect

101

ANFIS Grid Partitioning Plots for 35 HRC 4.31

Training Error Plots for Ra (Target vs Output)

103

4.32

Testing Error Plots for Ra (Target vs Output)

103

4.33

Validation Error Plots for Ra (Target vs Output)

103

4.34

Regression Plots for Ra (Train /Test/Validate)

103

4.35

Response Surface Plot for Ra

104

4.36

Training Error Plots for Ft (Target vs Output)

104

4.37

Testing Error Plots for Ft (Target vs Output)

104

4.38

Validation Error Plots for Ft (Target vs Output)

105

4.39

Regression Plots for Ft (Train /Test/Validate)

105

4.40

Response Surface Plot for Ft

105

4.41

Training Error Plots for Fa (Target vs Output)

105

4.42

Testing Error Plots for Fa (Target vs Output)

105

4.43

Validation Error Plots for Fa (Target vs Output)

106

4.44

Regression Plots for Fa (Train /Test/Validate)

106

4.45

Response Surface Plot for Fa

106

4.46

Training Error Plots for Fr (Target vs Output)

106

4.47

Testing Error Plots for Fr (Target vs Output

106

v


4.48

Validation Error Plots for Fr (Target vs Output)

107

4.49

Regression Plots for Fr (Train /Test/Validate)

107

4.50

Response Surface Plot for Fr

107

4.51

Training Error Plots for Tf (Target vs Output)

108

4.52

Testing Error Plots for Tf (Target vs Output)

108

4.53

Validation Error Plots for Tf (Target vs Output)

108

4.54

Regression Plots for Tf (Train /Test/Validate)

108

4.55

Response Surface Plot for Tf

108

ANFIS Grid Partitioning Plots For 45 HRC 4.56

Training Error Plots for Ra (Target vs Output)

109

4.57

Testing Error Plots for Ra (Target vs Output)

109

4.58

Validation Error Plots for Ra (Target vs Output)

109

4.59

Regression Plots for Ra (Train /Test/Validate)

109

4.60

Response Surface Plot for Ra

109

4.61

Training Error Plots for Ft (Target vs Output)

110

4.62

Testing Error Plots for Ft (Target vs Output)

110

4.63

Validation Error Plots for Ft (Target vs Output)

110

4.64

Regression Plots for Ft (Train /Test/Validate)

110

4.65

Response Surface Plot for Ft

110

4.66

Training Error Plots for Fa (Target vs Output)

111

4.67

Testing Error Plots for Fa (Target vs Output)

111

4.68

Validation Error Plots for Fa (Target vs Output)

111

vi


4.69

Regression Plots for Fa (Train /Test/Validate)

111

4.70

Response Surface Plot for Fa

111

4.71

Training Error Plots for Fr (Target vs Output)

112

4.72

Testing Error Plots for Fr (Target vs Output

112

4.73

Validation Error Plots for Fr (Target vs Output)

112

4.74

Regression Plots for Fr (Train /Test/Validate)

112

4.75

Response Surface Plot for Fr

112

4.76

Training Error Plots for Tf (Target vs Output)

113

4.77

Testing Error Plots for Tf (Target vs Output)

113

4.78

Validation Error Plots for Tf (Target vs Output)

113

4.79

Regression Plots for Tf (Train /Test/Validate)

113

4.80

Response Surface Plot for Tf

113

4.81

Developed ANFIS (Subtractive Cluster)

115

ANFIS Subtractive Cluster plots For 35 HRC 4.82

Training Error Plots for Ra (Target vs Output)

117

4.83

Testing Error Plots for Ra (Target vs Output)

117

4.84

Validation Error Plots for Ra (Target vs Output)

117

4.85

Regression Plots for Ra (Train /Test/Validate)

117

4.86

Response Surface Plot for Ra

117

4.87

Training Error Plots for Ft (Target vs Output)

118

4.88

Testing Error Plots for Ft (Target vs Output)

118

4.89

Validation Error Plots for Ft (Target vs Output)

118

vii


4.90

Regression Plots for Ft (Train /Test/Validate)

118

4.91

Response Surface Plot for Ft

118

4.92

Training Error Plots for Fa (Target vs Output)

119

4.93

Testing Error Plots for Fa (Target vs Output)

119

4.94

Validation Error Plots for Fa (Target vs Output)

119

4.95

Regression Plots for Fa (Train /Test/Validate)

119

4.96

Response Surface Plot for Fa

119

4.97

Training Error Plots for Fr (Target vs Output)

120

4.98

Testing Error Plots for Fr (Target vs Output

120

4.99

Validation Error Plots for Fr (Target vs Output)

120

4.100

Regression Plots for Fr (Train /Test/Validate)

120

4.101

Response Surface Plot for Fr

120

4.102

Training Error Plots for Tf (Target vs Output)

121

4.103

Testing Error Plots for Tf (Target vs Output)

121

4.104

Validation Error Plots for Tf (Target vs Output)

121

4.105

Regression Plots for Tf (Train /Test/Validate)

121

4.106

Response Surface Plot for Tf

121

ANFIS Subtractive Cluster plots For 45 HRC 4.107

Training Error Plots for Ra (Target vs Output)

122

4.108

Testing Error Plots for Ra (Target vs Output)

122

4.109

Validation Error Plots for Ra (Target vs Output)

122

4.110

Regression Plots for Ra (Train /Test/Validate)

122

viii


4.111

Response Surface Plot for Ra

122

4.112

Training Error Plots for Ft (Target vs Output)

123

4.113

Testing Error Plots for Ft (Target vs Output)

123

4.114

Validation Error Plots for Ft (Target vs Output)

123

4.115

Regression Plots for Ft (Train /Test/Validate)

123

4.116

Response Surface Plot for Ft

123

4.117

Training Error Plots for Fa (Target vs Output)

124

4.118

Testing Error Plots for Fa (Target vs Output)

124

4.119

Validation Error Plots for Fa (Target vs Output)

124

4.120

Regression Plots for Fa (Train /Test/Validate)

124

4.121

Response Surface Plot for Fa

124

4.122

Training Error Plots for Fr (Target vs Output)

125

4.123

Testing Error Plots for Fr (Target vs Output

125

4.124

Validation Error Plots for Fr (Target vs Output)

125

4.125

Regression Plots for Fr (Train /Test/Validate)

125

4.126

Response Surface Plot for Fr

125

4.127

Training Error Plots for Tf (Target vs Output)

126

4.128

Testing Error Plots for Tf (Target vs Output)

126

4.129

Validation Error Plots for Tf (Target vs Output)

126

4.130

Regression Plots for Tf (Train /Test/Validate)

126

4.131

Response Surface Plot for Tf

126

4.132

Developed ANFIS Fuzzy C Mean Clustering architect

128

ix


ANFIS FCM plots For 35 HRC 4.133

Training Error Plots for Ra (Target vs Output)

130

4.134

Testing Error Plots for Ra (Target vs Output)

130

4.135

Validation Error Plots for Ra (Target vs Output)

130

4.136

Regression Plots for Ra (Train /Test/Validate)

130

4.137

Response Surface Plot for Ra

130

4.138

Training Error Plots for Ft (Target vs Output)

131

4.139

Testing Error Plots for Ft (Target vs Output)

131

4.140

Validation Error Plots for Ft (Target vs Output)

131

4.141

Regression Plots for Ft (Train /Test/Validate)

131

4.142

Response Surface Plot for Ft

131

4.143

Training Error Plots for Fa (Target vs Output)

132

4.144

Testing Error Plots for Fa (Target vs Output)

132

4.145

Validation Error Plots for Fa (Target vs Output)

132

4.146

Regression Plots for Fa (Train /Test/Validate)

132

4.147

Response Surface Plot for Fa

132

4.148

Training Error Plots for Fr (Target vs Output)

133

4.149

Testing Error Plots for Fr (Target vs Output

133

4.150

Validation Error Plots for Fr (Target vs Output)

133

4.151

Regression Plots for Fr (Train /Test/Validate)

133

4.152

Response Surface Plot for Fr

133

4.153

Training Error Plots for Tf (Target vs Output)

134

x


4.154

Testing Error Plots for Tf (Target vs Output)

134

4.155

Validation Error Plots for Tf (Target vs Output)

134

4.156

Regression Plots for Tf (Train /Test/Validate)

134

4.157

Response Surface Plot for Tf

134

ANFIS FCM plots For 45 HRC 4.158

Training Error Plots for Ra (Target vs Output)

135

4.159

Testing Error Plots for Ra (Target vs Output)

135

4.160

Validation Error Plots for Ra (Target vs Output)

135

4.161

Regression Plots for Ra (Train /Test/Validate)

135

4.162

Response Surface Plot for Ra

135

4.163

Training Error Plots for Ft (Target vs Output)

136

4.164

Testing Error Plots for Ft (Target vs Output)

136

4.165

Validation Error Plots for Ft (Target vs Output)

136

4.166

Regression Plots for Ft (Train /Test/Validate)

136

4.167

Response Surface Plot for Ft

136

4.168

Training Error Plots for Fa (Target vs Output)

137

4.169

Testing Error Plots for Fa (Target vs Output)

137

4.170

Validation Error Plots for Fa (Target vs Output)

137

4.171

Regression Plots for Fa (Train /Test/Validate)

137

4.172

Response Surface Plot for Fa

137

4.173

Training Error Plots for Fr (Target vs Output)

138

4.174

Testing Error Plots for Fr (Target vs Output

138

xi


4.175

Validation Error Plots for Fr (Target vs Output)

138

4.176

Regression Plots for Fr (Train /Test/Validate)

138

4.177

Response Surface Plot for Fr

138

4.178

Training Error Plots for Tf (Target vs Output)

139

4.179

Testing Error Plots for Tf (Target vs Output)

139

4.180

Validation Error Plots for Tf (Target vs Output)

139

4.181

Regression Plots for Tf (Train /Test/Validate)

139

4.182

Response Surface Plot for Tf

139

Comparison Error Plots of Neural Network and ANFIS Prediction FOR 35 HRC 4.183

Error Estimation Plots For R a

140

4.184

Error Estimation Plots For F t

140

4.185

Error Estimation Plots For F a

141

4.186

Error Estimation Plots For F r

141

For 45 HRC 4.187

Error Estimation Plots For R a

141

4.188

Error Estimation Plots For F t

141

4.189

Error Estimation Plots For F a

141

4.190

Error Estimation Plots For F r

141

Error Plots of ANFIS (Grid Partitioning Clustering) Results

xii


FOR 35 HRC 4.191

Error Estimation Plots For R a

142

4.192

Error Estimation Plots For F t

142

4.193

Error Estimation Plots For F a

142

4.194

Error Estimation Plots For F r

142

FOR 45 HRC 4.195

Error Estimation Plots For R a

142

4.196

Error Estimation Plots For F t

142

4.197

Error Estimation Plots For F a

143

4.198

Error Estimation Plots For F r

143

Error Plots of ANFIS (Subtractive Clustering) FOR 35 HRC 4.199

Error Estimation Plots For R a

143

4.200

Error Estimation Plots For F t

143

4.201

Error Estimation Plots For F a

143

4.202

Error Estimation Plots For F r

143

FOR 45 HRC 4.203

Error Estimation Plots For R a

144

4.204

Error Estimation Plots For F t

144

4.205

Error Estimation Plots For F a

144

4.206

Error Estimation Plots For F r

144

Error Plots of ANFIS (Fuzzy C-Mean Clustering))

xiii


FOR 35 hrc 4.207

Error Estimation Plots For R a

144

4.208

Error Estimation Plots For F t

144

4.209

Error Estimation Plots For F a

145

4.210

Error Estimation Plots For F r

145

FOR 45 HRC 4.211

Error Estimation Plots For R a

145

4.212

Error Estimation Plots For F t

145

4.213

Error Estimation Plots For F a

145

4.214

Error Estimation Plots For F r

145

5.1

Chapter 5 flow chart

147

5.2

Developed NSGA–NN architect

149

NSGA-NN 35HRC 5.3

Performance plot of Network

151

5.4

Training state of Network at each epoch

151

5.5

Training error in Ra

151

5.6

Regression fit plot for Ra

151

5.7

Training error in Ft

151

5.8

Regression fit plot for Ft

151

5.10

Training error in Fa

152

5.11

Regression fit plot for Fa

152

5.12

Training error in Fr

152

xiv


5.13

Regression fit plot for Fr

152

5.14

Training error in Tf

152

5.15

Regression fit plot for Tf

152

Plots for NSGA-NN 45 HRC 5.16

Performance plot of Network

153

5.17

Training state of Network at each epoch

153

5.18

Training error in Ra

153

5.19

Regression fit plot for Ra

153

5.20

Training error in Ft

153

5.21

Regression fit plot for Ft

153

5.22

Training error in Fa

154

5.23

Regression fit plot for Fa

154

5.24

Training error in Fr

154

5.25

Regression fit plot for Fr

154

5.26

Training error in Tf

154

5.27

Regression fit plot for Tf

154

5.28

. SI-NN collaborative combination

155

5.29

Applied SI-NN synergy architect

155

Results of NSGA-NN 35 HRC 5.30

Performance plot of Network

157

5.31

Training state of Network at each epoch

157

5.32

Training error in Ra

157

xv


5.33

Regression fit plot for Ra

157

5.34

Training error in Ft

157

5.35

Regression fit plot for Ft

157

5.36

Training error in Fa

158

5.37

Regression fit plot for Fa

158

5.38

Training error in Fr

158

5.39

Regression fit plot for Fr

158

5.40

Training error in Tf

158

5.41

Regression fit plot for Tf

158

5.42

GA based ANFIS (FCM) applied Strategy

159

ANFIS-GA FCM plots For 35 HRC 5.43

Training Error Plots for Ra (Target vs Output)

160

5.44

Testing Error Plots for Ra (Target vs Output)

160

5.45

Regression Plots for Ra (Train /Test/Validate)

161

5.46

Response Surface Plot for Ra

161

5.47

Training Error Plots for Ft (Target vs Output)

161

5.48

Testing Error Plots for Ft (Target vs Output)

161

5.49

Regression Plots for Ft (Train /Test/Validate)

162

5.50

Response Surface Plot for Ft

162

5.51

Training Error Plots for Fa (Target vs Output)

162

5.52

Testing Error Plots for Fa (Target vs Output)

162

5.53

Regression Plots for Fa (Train /Test/Validate)

163

xvi


5.54

Response Surface Plot for Fa

163

5.55

Training Error Plots for Fr (Target vs Output)

163

5.56

Testing Error Plots for Fr (Target vs Output

163

5.57

Regression Plots for Fr (Train /Test/Validate)

164

5.58

Response Surface Plot for Fr

164

5.59

Training Error Plots for Tf (Target vs Output)

164

5.60

Testing Error Plots for Tf (Target vs Output)

164

5.61

Regression Plots for Tf (Train /Test/Validate)

165

5.62

Response Surface Plot for Tf

165

FCM plots For 45 HRC 5.63

Training Error Plots for Ra (Target vs Output)

166

5.64

Testing Error Plots for Ra (Target vs Output)

166

5.65

Regression Plots for Ra (Train /Test/Validate)

166

5.66

Response Surface Plot for Ra

166

5.67

Training Error Plots for Ft (Target vs Output)

167

5.68

Testing Error Plots for Ft (Target vs Output)

167

5.69

Regression Plots for Ft (Train /Test/Validate)

167

5.70

Response Surface Plot for Ft

167

5.71

Training Error Plots for Fa (Target vs Output)

168

5.72

Testing Error Plots for Fa (Target vs Output)

168

5.73

Regression Plots for Fa (Train /Test/Validate)

168

5.74

Response Surface Plot for Fa

168

xvii


5.75

Training Error Plots for Fr (Target vs Output)

169

5.76

Testing Error Plots for Fr (Target vs Output

169

5.77

Regression Plots for Fr (Train /Test/Validate)

169

5.78

Response Surface Plot for Fr

169

5.79

Training Error Plots for Tf (Target vs Output)

170

5.80

Testing Error Plots for Tf (Target vs Output)

170

5.81

Regression Plots for Tf (Train /Test/Validate)

170

5.82

Response Surface Plot for Tf

170

5.83

PS0-ANFIS applied strategy

171

PSO based ANFIS (FCM) Plots For 35 HRC 5.84

Training Error Plots for Ra (Target vs Output)

172

5.85

Testing Error Plots for Ra (Target vs Output)

172

5.86

Regression Plots for Ra (Train /Test/Validate)

172

5.87

Response Surface Plot for Ra

172

5.88

Training Error Plots for Ft (Target vs Output)

173

5.89

Testing Error Plots for Ft (Target vs Output)

173

5.90

Regression Plots for Ft (Train /Test/Validate)

173

5.91

Response Surface Plot for Ft

173

5.92

Training Error Plots for Fa (Target vs Output)

174

5.93

Testing Error Plots for Fa (Target vs Output)

174

5.94

Regression Plots for Fa (Train /Test/Validate)

174

5.95

Response Surface Plot for Fa

174

xviii


5.96

Training Error Plots for Fr (Target vs Output)

175

5.97

Testing Error Plots for Fr (Target vs Output

175

5.98

Regression Plots for Fr (Train /Test/Validate)

175

5.99

Response Surface Plot for Fr

175

5.100

Training Error Plots for Tf (Target vs Output)

176

5.101

Testing Error Plots for Tf (Target vs Output)

176

5.102

Regression Plots for Tf (Train /Test/Validate)

176

5.103

Response Surface Plot for Tf

176

PSO based ANFIS (FCM) Plots FOR 45 HRC 5.104

Training Error Plots for Ra (Target vs Output)

177

5.105

Testing Error Plots for Ra (Target vs Output)

177

5.106

Regression Plots for Ra (Train /Test/Validate)

177

5.107

Response Surface Plot for Ra

177

5.108

Training Error Plots for Ft (Target vs Output)

178

5.109

Testing Error Plots for Ft (Target vs Output)

178

5.110

Regression Plots for Ft (Train /Test/Validate)

178

5.111

Response Surface Plot for Ft

178

5.112

Training Error Plots for Fa (Target vs Output)

179

5.113

Testing Error Plots for Fa (Target vs Output)

179

5.114

Regression Plots for Fa (Train /Test/Validate)

179

5.115

Response Surface Plot for Fa

179

5.116

Training Error Plots for Fr (Target vs Output)

180

xix


5.117

Testing Error Plots for Fr (Target vs Output

180

5.118

Regression Plots for Fr (Train /Test/Validate)

180

5.119

Response Surface Plot for Fr

180

5.120

Training Error Plots for Tf (Target vs Output)

181

5.121

Testing Error Plots for Tf (Target vs Output)

181

5.122

Regression Plots for Tf (Train /Test/Validate)

181

5.123

Response Surface Plot for Tf

181

Comparison Error Plots of NN-NSGA Prediction FOR 35 HRC 5.124

Error Estimation Plots For R a

182

5.125

Error Estimation Plots For F t

182

5.126

Error Estimation Plots For F a

183

5.127

Error Estimation Plots For F r

183

For 45 HRC 5.128

Error Estimation Plots For R a

183

5.129

Error Estimation Plots For F t

183

5.130

Error Estimation Plots For F a

183

5.131

Error Estimation Plots For F r

183

Error Plots of PSO-NN Results FOR 35 HRC 5.132

Error Estimation Plots For R a

184

5.133

Error Estimation Plots For F t

184

5.134

Error Estimation Plots For F a

184

xx


5.135

Error Estimation Plots For F r

184

Error Plots of ANFIS (FCM)-GA FOR 35 HRC 5.136

Error Estimation Plots For R a

184

5.137

Error Estimation Plots For F t

184

5.138

Error Estimation Plots For F a

185

5.139

Error Estimation Plots For F r

185

Error Plots of ANFIS (FCM)-GA FOR 45 HRC 5.140

Error Estimation Plots For R a

185

5.141

Error Estimation Plots For F t

185

5.142

Error Estimation Plots For F a

185

5.143

Error Estimation Plots For F r

185

Error Plots of ANFIS (FCM)-PSO FOR 35 HRC 5.144

Error Estimation Plots For R a

186

5.145

Error Estimation Plots For F t

186

5.146

Error Estimation Plots For F a

186

5.147

Error Estimation Plots For F r

186

Error Plots of ANFIS (FCM)-PSO FOR 45 HRC 5.148

Error Estimation Plots For R a

186

5.149

Error Estimation Plots For F t

186

5.150

Error Estimation Plots For F a

187

5.151

Error Estimation Plots For F r

187

xxi


LIST OF TABLES Table No`

Name of Table

Pg. No.

3.1

Machining Constraints

30

3.2

NSGA II Setting

41

3.3

Results of NSGA II family of best solution for AISI 4340 35 HRC Steel

48

3.4

Results of NSGA II family of best solution for AISI 4340 45 HRC Steel

50

3.5

SPEA2 Parameter Setting

54

3.6

Results of SPEA II family of best solution for AISI 4340 35 HRC Steel

57

3.7

Results of SPEA II family of best solution for AISI 4340 45 HRC Steel

58

3.8

PSO Setting

67

3.9

MOPSO family of optimal solutions for 35 HRC AISI 4340 steel

72

3.10

MOPSO family of optimal solutions for 35 HRC AISI 4340 steel

74

3.11

Diversity of Evolutionary Algorithm

78

4.2.4 (a)

Description of Neural network

88

4.2.4(b)

Calibrated weights and bias above Neural Network

89

4.3.4(a)

Grid Partioning Fuzzy Structure

102

4.3.4(b)

Statistical Results of ANFIS Grid Partioning Cluster for 35 HRC and 45HRC

102

4.3.16(a)

Subtractive Fuzzy Structure

115

4.3.16(b)

Statistical Error analysis of ANFIS Subtractive clustering for 35 HRC and 45 HRC

116

xxii


4.3.26 (a)

Fuzzy structure

127

4.3.26(b)

Statistical Results of ANFIS FCM for 35 HRC and 45HRC

129

4.4.1

Statistical Comparison of Neural Network and ANFIS Prediction Results with Experimental Statistics for AISI 4340 Steel 35hrc

140

5.2.1 (a)

Description of NSGA-NN

150

5.2.1 (b)

Calibrated weigths and bias for NSGA-NN 35 HRC and 45HRC Steel

150

5.3.1 (a)

Description of PSO-NN

156

5.3.2 (b) 5.4.1 (a)

Calibrated weigths and bias for 35 HRC and 45HRCSteel Statistical Error analysis of GA and PSO based ANFIS (FCM) for 35 HRC and 45HRC

156 159

5.6.1 (a)

Statistical Comparison of Prediction Results with Experimental Statistics for AISI 4340 Steel 35HRC and 45HRC

182

6.1

Tradeoffs among forces for Surface roughness and Tool life

190

6.2

Tradeoffs among forces for Surface roughness and Tool life

192

6.3

Tradeoffs among forces for Surface roughness and Tool life

193

6.4

Tradeoffs among forces for Surface roughness and Tool life

194

6.5

Mean error and Standard Deviation between Experimental and Predicted statistics

201

6.6

Table.6.6 MSE and RMSE between Experimental and Predicted statistics

201

6.7

Mean error and Standard Deviation between Experimental and Predicted statistics

205

6.8

Mean Square Error and Root Mean Square Error between Experimental and Predicted statistics

205

xxiii


NOMENCLATURE HRC

Rockwell C Hardness

V, (vc)

Cutting Speed

f

Feed Rate

d

Depth Of Cut

Ra

Surface Roughness

Ft

Tangential Force

Fa

Axial Force

Fr

Radial Force

Tf

Tool Life

AI

Artificial Intelligence

CI

Computational Intelligence

HC

Hard Computation

SC

Soft Computing

FS

Fuzzy Systems

NN

Neural Network

EA

Evolutionary Algorithm

GSO

Global Search Optimization

SI

Swarm Intelligence

GA

Genetic Algorithm

DE

Differential Evolution

CA

Culture Algorithm

PSO

Particle Swarm Optimization

FA

Fire Fly Algorithm

BBO

Biogeography Based Optimization

xxiv


W1/W2/W3

Weights

EA-NN

Evolutionary Based Neural Network

NN-EA

Neural Network Based Evolutionary Algorithm

NSGA

Non-Sorted Domination Based Genetic Algorithm

PESA

Pareto Envelope Based Selection Algorithm

SPEA

Strength Pareto Based Evolutionary Algorithm

MF

Membership Function

ANFIS

Adaptive Neuro-Fuzzy Interface System

ACO

Ant Colony Algorithm

HLGA

Hybrid Adaptive Learning Based Genetic Algorithm

Pmt

Probability of m bits in K sting length to Mutate

Stn.d , rt n,d X t , Vt

Random number generators Particle Position and Velocity

Gt

Global Best attractor

 n ', n

Local Guide attractor

Lt

nt ,d

Swarm Potential

Sij

Layers

Membership cluster vector

d ( xi , j )

Dissimilarity Function

J q ( , u)

Cluster Function

xxv


CONTENTS Sr.

Title

Pg. No

No Acknowledgement

i

List of Figures

ii

List of Tables

xix

Nomenclature Contents Abstract

xxi xxiii xxviii

1

INTRODUCTION

1.1

Computational Intelligence (CI)

1

1.2

Approaches to Computational Intelligence

2

1.2.1

Fuzzy logic

4

1.2.2

Neural Network

4

1.2.3

Evolutionary Computing

4

1.2.4

Learning Theory

5

1.2.5

Probabilistic Methods

5

1.2.6

Swarm Intelligence

6

1.2.7

Global Search Optimization

6

1.3

Synergies of Computational Intelligence Techniques

6

1.4

Applications of Computational Intelligence

8

1.4.1

Application of NN

8

1.4.2

Application of Evolutionary Systems

9

1.4.3

Application of Fuzzy system

9

1.5

Overview of the chapter

10

1.6

Problem Statement

10

1.7

Objectives

10

1.8

Methodology

11

1.9

Thesis Outline

12

2

LITERATURE REVIEW

xxvi


2.1

Literature Review

14

2.2

Literature Summary

27

3

GLOBAL SEARCH ALGORITHMS FOR MULTI-OBJECTIVE

3.1

Introduction

28

3.2

Multi-objective Optimization

29

3.3

Application of MOOPs to Machining system

30

3.3.1

Machining Model (Surface Roughness Cutting force components material

31

3.3.2

hardness: 35 HRC). Machining Model (Surface Roughness Cutting force components work

31

3.3.3

material hardness: 45 HRC). Tool life model for 35 HRC

32

3.4

Evolutionary Algorithms (EAs)

32

3.4.1

Mathematical Formulation of Evolutionary Algorithms

33

3.4.1.

Definition of Evolutionary systems

34

3.4.1. 1 3.4.1.

Convergence Analysis of Evolutionary Algorithm

34

Criteria for mutation

35

3 3.4.1.

Criteria for Crossover

37

4 points from convergence analysis Key

40

3.4.2 3.5

3.6

NSGA II Algorithm

41

3.5.1

Initialize Variables and Evaluate Objectives

41

3.5.2

Non_dominated_sort

42

3.5.3

Selection.

45

3.5.4

Genetic Operators

45

3.5.5

Recombination of parent and off springs

47

3.5.6

Plots for NSGA II results (35 HRC)

50

3.5.7

Plots for NSGA II results (45 HRC)

52

Strength Pareto Evolutionary Algorithm 2

52

3.6.1

Initialize Variables and Evalaute Objectives

54

3.6.2

Tournament Selection

55

3.6.3

Genetic Operator

55

3.6.4

Plots for SPEA 2 results( 35 HRC)

57

3.6.5

Plots for SPEA 2 results( 45 HRC)

59

xxvii


3.7

Swarm Intelligence

60

3.8

Mathematical Formulation of PSO Algorithm

60

3.8.1

Typical Initialization strategy

62

3.8.2

Topologies of PSO

62

3.8.3

Definitions of Swarm topology

63

3.8.4

Convergence criteria

63

3.8.5

Criteria for Inertia clamping and acceleration co-efficeint

64

3.9

3.10

4

PSO Algorithm

66

3.9.

Initialize population and Evaluate fitness

67

1 3.9.

Create Grid Index

68

2 3.9.

Select Leader

69

3 3.9.

Delete extra elements

70

3.9. 4

Swarm Movement

71

6 3.9.

Plots for MOPSO results (35 HRC)

73

73.9.

Plots for MOPSO results (45 HRC)

75

8 Comparison between EA and SI technique

76

3.10.

Comparison Based on Spectrum of solution space

77

1 3.10.

Comparison Based on Diversity in solution space.

78

2 PREDICTION MODELS FOR MACHINING SYSTEM

THROUGH INTELLIGENT LEARNING TECHNIQUES 4.1

Introduction

80

4.2

Neural Network

81

4.2.1

Feed forward Neural network

81

4.2.2

Mathematical background of neural network

82

4.2.2.1

Gradient Descent Approach

83

4.2.3

Key notes form feed forward analysis

87

4.2.4

Multi-layer Perceptron for Turning of AISI 4340 Steel

87

4.2.5

Results of perceptron for 35 HRC Steel

89

4.2.6

Results of perceptron for 45 HRC Steel

92

4.3

Adaptive Neuro-Fuzzy Interference System (ANFIS)

94

4.3.1

96

Hybrid learning in ANFIS

xxviii


4.3.2

Backpropogation Learning

97

4.3.3

Fuzzy clustering Algorithms

98

4.3.4

Grid Partition clustering based Adaptive Neuro-Fuzzy Interference System

101

4.3.5

ANFIS Grid Partioning Cluster Plots For Ra 35 HRC

103

4.3.6

ANFIS Grid Partioning Cluster Plots For Ft 35 HRC

104

4.3.7

ANFIS Grid Partioning Cluster Plots For Fa 35 HRC

105

4.3.8

ANFIS Grid Partioning Cluster Plots For Fr 35 HRC

106

4.3.9

ANFIS Grid Partioning Cluster Plots For Tf 35 HRC

108

4.3.10

ANFIS Grid Partioning Cluster Plots For Ra 45 HRC

109

4.3.11

ANFIS Grid Partioning Cluster Plots For Ft 45 HRC

110

4.3.12

ANFIS Grid Partioning Cluster Plots For Fa 45 HRC

111

4.3.13

ANFIS Grid Partioning Cluster Plots For Fr 45 HRC

112

4.3.14

ANFIS Grid Partioning Cluster Plots For Tf 45 HRC HRC

113

4.3.15

Subtractive Clustering

114

4.3.16

ANFIS Substractive Cluster Plots For Ra 35 HRC

117

4.3.17

ANFIS Substractive Cluster Plots For Ft 35 HRC

118

4.3.18

ANFIS Substractive Cluster Plots For Fa 35 HRC

119

4.3.19

ANFIS Substractive Cluster Plots For Fr 35 HRC

120

4.3.20

ANFIS Substractive Cluster Plots For Tf 35 HRC

121

4.3.21

ANFIS Substractive Cluster Plots For Ra 45 HRC

122

4.3.22

ANFIS Substractive Cluster Plots For Ft 45 HRC

123

4.3.23

ANFIS Substractive Cluster Plots For Fa 45 HRC

124

4.3.24

ANFIS Substractive Cluster Plots For Fr 45 HRC

125

4.3.25

ANFIS Substractive Cluster Plots For Tf 45 HRC

126

4.3.26

Fuzzy C Mean Clustering

127

4.3.27

ANFIS

130

4.3.28

ANFIS FCM Cluster Plots For Ft 35 HRC

131

4.3.29

ANFIS FCM Cluster Plots For Fa 35 HRC

132

4.3.30

ANFIS FCM Cluster Plots For Fr 35 HRC

133

4.3.31

ANFIS FCM Cluster Plots For Tf 35 HRC

134

4.3.32

ANFIS FCM Cluster Plots For Ra 45 HRC

135

FCM Cluster Plots For Ra 35 HRC

xxix


4.4

4.3.33

ANFIS FCM Cluster Plots For Ft 45 HRC

136

4.3.34

ANFIS FCM Cluster Plots For Fa 45 HRC

137

4.3.35

ANFIS FCM ve Cluster Plots For Fr 45 HRC

138

4.3.36

ANFIS FCM Cluster Plots For Tf 45 HRC

139

Comparison of Prediction Results with Experimental Statistics

140

4.4.2

Error Plots of Neural Network Prediction Results with Experimental

140

4.4.3

Statistics for AISI 4340 Steel 35hrc Error Plots of Neural Network Prediction Results with Experimental

141

4.4.4

Statistics for AISI 4340 Steel 45HRC Error Plots of ANFIS (Grid Partitioning Clustering) Results with

142

Experimental Statistics for AISI 4340 Steel 35hrc 4.4.5

Error Plots of ANFIS (Grid Partitioning Clustering) Results with

142

Experimental Statistics for AISI 4340 Steel 45hrc 4.4.6

Error Plots of ANFIS (Subtractive Clustering) Results with Experimental

143

Statistics for AISI 4340 Steel 35hrc 4.4.7

Error Plots of ANFIS (Subtractive Clustering) Results with Experimental

144

Statistics for AISI 4340 Steel 45hrc 4.4.8

Error Plots of ANFIS (Fuzzy C-Mean Clustering)) Results with

144

Experimental Statistics for AISI 4340 Steel 35hrc 4.4.9 4.5

5

Error Plots of ANFIS (Fuzzy C-Mean Clustering)) Results with

Experimental Statistics for AISI 4340 Steel 45hrc Conclusion

145 146

HYBRIDISATION OF C.I SYNERGIES

5.1

Introduction

147

5.2

EA-NN Synergism

148

5.2.1

NSGA combined Neural Network

149

5.2.2

Results of NSGA-NN 35 HRC Steel

151

5.2.3

Results of NSGA-NN for 45 HRC Steel

153

5.3

SI-NN synergism

154

5.3.1

PSO combined Neural Network

155

5.3.2

Results of PSO-NN for 35 HRC Steel

157

xxx


5.4

5.5

5.6

Synergies of EA and ANFIS

158

5.4.1

ANFIS GA

159

5.4.2

GA based ANFIS (FCM) Plots For Ra 35 HRC

169

5.4.3

GA based ANFIS (FCM) Plots For Ft 35 HRC

161

5.4.4

GA based ANFIS (FCM) Plots For Fa 35 HRC

162

5.4.5

GA based ANFIS (FCM) Plots For Fr 35 HRC

163

5.4.6

GA based ANFIS (FCM) Plots For Tf 35 HRC

164

5.4.7

GA based ANFIS (FCM) Plots For Ra 45 HRC

166

5.4.8

GA based ANFIS (FCM) Plots For Ft 45 HRC

167

5.4.9

GA based ANFIS (FCM) Plots For Fa 45 HRC

168

5.4.10

GA based ANFIS (FCM) Plots For Fr 45 HRC

169

5.4.11

GA based ANFIS (FCM) Plots For Tf 45 HRC

170

PSO based ANFIS (FCM) 35 HRC and 45HRC

171

5.5.1

PSO based ANFIS (FCM)

171

5.5.2

PSO based ANFIS (FCM) Plots For Ra 35 HRC

172

5.5.3

PSO based ANFIS (FCM) Plots For Ft 35 HRC

173

5.5.4

PSO based ANFIS (FCM) Plots For Fa 35 HRC

174

5.5.5

PSO based ANFIS (FCM) Plots For Fr 35 HRC

175

5.5.6

PSO based ANFIS (FCM) Plots For Tf 35 HRC

176

5.5.8

PSO based ANFIS (FCM) Plots For Ra 45 HRC

177

5.5.9 5.5.10

PSO based ANFIS (FCM) Plots For Ft 45 HRC PSO based ANFIS (FCM) Plots For Fa 45 HRC

178 179

5.5.11

PSO based ANFIS (FCM) Plots For Fr 45 HRC

180

5.5.12

PSO based ANFIS (FCM) Plots For Tf 45 HRC

181

Comparison of Prediction Results with Experimental Statistics

182

5.6.2

Error Plots of NSGA-NN Prediction Results with Experimental Statistics

182

5.6.3

for AISI 4340 Steel 35HRC Error Plots of NSGA-NN Prediction Results with Experimental Statistics

183

5.6.4

for AISI 4340 Steel 45HRC Error Plots of PSO-NN Results with Experimental Statistics for AISI 4340

184

5.6.5

Steel 35HRC Error Plots of PSO-NN Results with Experimental Statistics for AISI 4340

184

Steel 45HRC

xxxi


5.6.6

Error Plots of GA based ANFIS (Fuzzy C-Mean Clustering)) Results with

185

Experimental Statistics for AISI 4340 Steel 45HRC 5.6.7

Error Plots of PSO based ANFIS (Fuzzy C-Mean Clustering)) Results with

186

5.6.8

Experimental Statistics for AISI 4340 Steel 35HRC Error Plots of PSO based ANFIS Prediction Results with Experimental

186

5.7

Statistics for AISI 4340 Steel 45HRC Conclusion

6

RESULTS AND DISCUSSION

6.1

Global Search Optimization

189

6.1.1

189

6.1.2

6.1.3

6.1.4

6.1.5 6.2

6.1.1 (a)

For AISI 4340 35HRC

189

6.1.1(b)

For AISI 4340 45HRC

191

Results for SPEA 2

192

6.1.2 (a)

For AISI 4340 35HRC

192

6.1.2 (b)

For AISI 4340 45HRC

193

Results for PSO

194

6.1.3(a)

For AISI 4340 35HRC

194

6.1.3(b)

For AISI 4340 45HRC

195

Comparison between EA and SI

196

6.1.4 (a)

196

Comparison among the Solution spectrum

Evident from the literature

197

Intelligent Learning Techniques

198

6.2.1

Neural Network

198

6.2.2

Adaptive Neuro Fuzzy Interference Technique

199

6.2.2(a)

ANFIS Grid partition

199

6.2.2 (b)

ANFIS Subtractive Clustering

200

6.2.2 (c)

ANFIS Fuzzy C-mean Clustering

200

6.2.3 6.3

NSGAII

187

Comparative Evaluation of the predictive technique on Experimental statistics

200

Synergies of CI

202

6.3.1

EA-NN

203

6.3.2

SI-NN

203

xxxii


7

6.3.3

ANFIS Synergies

204

6.3.4

Comparative Evaluation of the predictive technique on Experimental statistics

204 207

CONCLUSION

Appendix- A: Conferences and Publications Appendix- B: PG-CON Certificate Appendix- C: Certificates

xxxiii


ABSTRACT Obtaining the process parameters to optimize machining performance is vital in machining execution since they significantly affect the productivity rate, cost and quality of machining operation. Although process parameters optimization has been widely investigated for conventional machining operation, very limited work is reported on optimization of hard turning using evolutionary algorithm. In this work multi-objective optimization of hard turning with evolutionary optimization technique is attempted (i.e, NSGA II, SPEA II, PSO) during hard turning of hardened AISI 4340 Steel at different hardness level (35 and 45 HRC) with experimental based multi regression models as objective functions. The process variables are cutting speed, feed rate, and depth of cut with appropriate constraints. Further -more different intelligent learning techniques (i.e, Neural network, Adaptive Neural network based fuzzy learning) were applied using (EA-NN and NN-EA) supportive combinations to recognize the pattern of optimal solution through learning This learnt prediction model is compared with experimental statistics a comparative evaluation is made which is in good agreement with experimental data. Keywords: Multi-Objective Optimization, Adaptive Neuro-Fuzzy Optimization, Hard Turning;

xxxiv


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Chapter-1 1. Introduction The complexity of design optimization of any dynamic system has many aspects among them the major facets are the ambiguity of objectives, conflicting nature of objectives and many possible solutions this brings an issue in characterizing the difficulty of design optimization task. Any design solution has combination of values for parameters of a solution and the challenge lies in identifying the solution. Considering second issue functionality of obtained solution, the solution should be practical enough, look appealing and have moderate cost. Third issue is contributing to the ambiguity of design optimization is conflicting objectives which inhibits unidirectional solution. To tackle these complexity several attempts are made through conventional methods but the solution in general are partially satisfactory To address this issue new computational approaches are followed which has multi-agent system each agent is defined by its behavior that are classified into various categories in Computational Intelligence. 1.1 Computational Intelligence (CI) Computational Intelligence deals with the design of intelligent agents which act intelligently for goal attainment in any circumstances, flexible enough to adopt changing environment and goals. Computational Intelligence has the ability to comprehend reason, learn, and simulate intelligent behavior in systems for complete knowledge formulation. Much real-time system behavior cannot be captured exactly through classical mathematical description in spite of complex formulations; moreover complexity of mathematical description inhibits development of system model. Hence it is really advantageous to model a real time system with piece-wise linearity and non-linearity so that the highly complex and un-anticipatory models can be captured by intelligent agents. Any real time problem has uncertainties involved in it with multiple objectives and the risk in decision making should be such that the performance criteria are maintained even in drastic change, this necessity of capturing the dynamic behavior of system is replacing conventional techniques with intelligent techniques. Computational Intelligent techniques VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 1


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

are thus an illustration of alternative methods to conventional technique when the system knowledge is highly important in system modeling and control. The structure of such systems is determined by experimental evidence where direct input-output response behavior is utilized to develop system model. Intelligent systems are meant for the processes that are not properly defined, complex and stochastic in nature, time varying. The fundamental property of any intelligent system is that it must sense and reason without prior knowledge about the environment and adapt to the control action in a robust manner. Many attempts have been made to define by different researchers but the property of a system to be computationally intelligent is if it deals with numerical data and has ability of pattern recognition. CI is a subdivision of machine intelligence where subtle difference between the techniques lies in the type of computing. Machine intelligence has two constituents Artificial Intelligence based on hard computing (HC) and Computational Intelligence based on soft computing (SC) [Fuzzy Sets (FS), Neural Network (NN), Evolutionary Algorithms(EA)] (Fig. 1.1) distinguishes clearly the components of Machine intelligence and their components.

Fig. 1.1 Hierarchy of Computational Intelligence

1.2 Approaches to Computational Intelligence

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 2


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

The core of the computational intelligence is designing process or system model which is not responsive to mathematical modeling since the process exhibits following attributes 

Too complex to represent mathematical model

Models difficult to Compute

Uncertainties in operations

Nonlinear, Stochastic and disturbed in nature

The system is capable of learning to adapt to unknown situations and is able to make predictions about the process status in future time step. CI is a combination of soft computing and numerical technique with methods involving adaptive control, (Fig. 1.2) optimal control, learning theory, fuzzy logic, neural-network, evolutionary computing. All methods tuned to attain common goal set. There are five elemental methods to CI 

Fuzzy logic

Neural network

Fig. 1.2 Approaches to Computational Intelligence 

Evolutionary computing

Learning theory VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 3


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation



Probabilistic methods



Swarm Intelligence

1.2.1 Fuzzy logic In any real-time process, the measurement, process modeling, and control can never be exact to the theoretical definitions [1]. There is always a certain amount of uncertainties i.e., incompleteness, randomness of data. Fuzzy assimilates human experimental knowledge converts it into engineering model and control. Most process which are illdefined with nonlinearity and uncertainties. The fuzzy logic is more of reasoning and inference technique based on high level linguistic or semantic rules and operations. 1.2.2 Neural Network Neural network neural network is a technique adopted from the biological brain which involves a neuron as a fundamental building block [1]. These neurons receive signals from neighboring neurons through their cell body and transfer the results through a long fiber called an axon. The axons behave like signal conducting device. A similar electrical analogy of biological neural network is artificial neural network which is characterized by computational power learning of real-time data error tolerance, pattern recognition, and generalization capabilities, low-level computational algorithms which manifest good performance in numerical data processing. The learning is in different form supervised, unsupervised, and competitive and reinforcement learning. 1.2.3 Evolutionary Computing Evolutionary computing is the imitation of the process of natural selection in a search procedure based on evolutionary theory of Charles Darwin [1]. The species undergoes reproduction, gives birth to new offspring with features of combating adverse environment and survive The process of natural selection makes sure that the individuals with better fitness have opportunity to make most of the time, with expectations that the offsprings will have similar higher fitness levels. Evolutionary computation uses iterative

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 4


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

progress and development in a population. This population is then selected in random search to achieve the desired population of solution. 1.2.4 Learning Theory Learning theory is based on the human learning capabilities without much effort in a conventional sense. The mechanism of learning in humans is the process of bringing together cognitive, emotional and environmental effects to acquire enhance or change knowledge, skills. In general, learning is characterized by how the information is input, processed and stored. Learning theories fall into three framework behaviorism cognitive theories and constructivism. Behaviorism is learning based on objectively observable feature learning Cognitive learning is how learning occurs in brain. Constructivism learning is a process in which permutation of existing idea builds a new idea. In most of the machine learning four basic forms of learning si adopted i.e., supervised learning where a mapping of input to desired output is done, unsupervised learning where a set of input feature is modeled and mapping of input to output is done with similar pattern. Semi-supervised learning, combination of both learned and unlearned datasets are used to generate an appropriate classifier. Reinforcement learning involves decision making on given observation and feedback is taken from the consequence to supervise the learning process [1]. 1.2.5 Probabilistic Methods Probabilistic theory is methodology which guides in dealing with the uncertainties and imprecisions. The probabilistic methods involve a space consisting probabilities of whole system. The uncertainties of complex dynamic system are calculated and combined behavior of system is analyzed for the degree of causticity. The chaotic behavior of system is estimated by the past. In general, the chaotic behavior of system grows exponentially with time [1].

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 5


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

1.2.6 Swarm Intelligence Swarm systems are based on behavior of school of birds, insects, fireflies where a flocks of birds twisting, V-shaped structure of migrating geese, winter birds hunting for food, the synchronized flashing of fireflies are tried to imitating. The well-choreographed collective behavior without any leader is adopted to search for optimal solutions for instance, ants living in colony, their behavior is driven by the goal of colony survival instead of individual survival, While searching for food ants initially explores surrounding nests in random manner. A similar behavior is observed with flocks of birds where a leader keeps guiding the flock to updated food location [1]. 1.2.7 Global Search Optimization Both EA and SI together form a broader class of optimization driven search techniques defined as global search optimization technique as shown in Fig.1.3 below.

Fig. 1.3 Global Search Optimization Hierarchy 1.3 Synergies of Computational Intelligence Techniques The different combination of all the methodologies can be used to design intelligent systems. Though a particular technique might be excellent in approximate reasoning and modeling uncertainty but may not be so good at learning and adopting with experimental data. A combined approach with computational intelligence technique and their implementation can help in designing better intelligent agents.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 6


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Different forms of synergisms (Fig. 1.4) of fuzzy logic, neural networks and evolutionary algorithms the common forms of weakly coupled synergism of neural network and evolutionary algorithms include training designing, optimizing architecture and parameters of neural networks and feature selection scaling training data for neural network using evolutionary algorithms. A strongly coupled synergism between the two methodologies where genetic operators are represented in the form of neural network and the epochs are meant to be the generations of evolutions [1]. Synergisms of neural networks and fuzzy systems have proven to be very powerful for system modeling and learning. In weakly coupled synergism, neural networks and fuzzy system work independently towards a common goal where neural network assist fuzzy logic to form rules and tuning membership functions. In strongly coupled synergism fuzzy system assist neurons to assign weights to its membership functions where neural network learns data over the epoch. There is other synergism possible between swarm intelligence, fuzzy systems, evolutionary algorithms and neural network.

Fig. 1.4 Synergies of Computational Intelligence

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 7


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

1.4 Applications of Computational Intelligence The system designed on the basis of computational intelligence elucidates data irrespective of its orientation and it has upper hand over the classical domain based analysis where the data computing and processing becomes difficult. CI exhibits characteristic of domain independence where same methodology can be applied to different fields For instance both neural network and fuzzy logic can be applied to solve a problem the only difference would be in the performance. Neural network can be applied in five ways i.e., data analysis, classifier, clustering, pattern recognition, control strategy neural network has been successfully applied in problems behaving non-linearly whereas fuzzy logic has been applied to appliances where a module control is required the most adaptive implementation is done on stabilizing an unsteady image. Fuzzy expert systems are applied to medical systems, diagnostic, scheduling, and financial systems. 1.4.1 Application of NN 1. In aerospace neural networks are applied to high performance autopilot flight path simulation, aircraft control systems, fault detection system. 2. In automotive neural networks are used for automatic guidance system. 3. In Banking, Financial and Business it is applied for document reading, credit application evaluation, credit, and activity. 4. In defense, it is used for weapon steering, target tracking, object discrimination, facial recognition. 5. In Industrial, Manufacturing and Electronics to control, process identification, machine diagnosis, quality inspection. 6. In Medical cancer cell analysis, EEG and ECG signal analysis, optimization of transplantation.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 8


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

7. In speech, applied for speech recognition, compression, and text to speech synthesis. 8. Telecommunication applied for image and data compression, speech processing, realtime translation of spoken language. 1.4.2 Application of Evolutionary Systems 1. In automotive design, including research on composite material design and multiobjective design of automotive components for crashworthiness, weight savings and other characteristics. 2. In optimizing the structural and operational design of Industry and Manufacturing systems. For optimization of mechanical systems like heat exchanger, turbines, flywheel and computer-assisted engineering design. 3. In automotive design of mechatronic systems using bond graphs and industrial equipment design using catalogs of exemplar lever pattern. 4. In Travelling Salesman Problem (TSP) and sequences scheduling. 5. In Control-gas pipeline, pole balancing, missile evasion, design-semiconductor layout, aircraft design, keyboard configuration, communication network. 6. In Combinatorial Optimization, Scheduling applications, including job-shop scheduling with objective to schedule jobs in both sequence-dependent and non-sequence dependent for maximum production volume. 1.4.3 Application of Fuzzy system 1. Fuzzy systems are used in automobiles and vehicle subsystems such as automatic transmissions, ABS and cruise control. 2. In air conditioners, washing machines and other home appliances. 3. In digital image processing, such as edge detection and video gam artificial intelligence.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 9


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4. In Pattern recognition in remote sensing, microcontrollers and microprocessors. 6. In Hydrometeor classification for polarimetric weather radar.

1.5 Overview of the chapter In this chapter a complete overview of ideas and foundations of computational Intelligence and their methodologies are discussed (i.e., Evolutionary Systems, Neural networks and Fuzzy systems). The Possible interactions of these techniques and the type of synergism in which the limitations of one technique can be surpassed by combination of two or more CI techniques is discussed. Depending on the compatibility of individual methodologies, a better computational model can be built which could complement respective methodologies. 1.6 Problem Statement Computational Intelligence has captured the attention of major manufacturing segment for its robust and dynamic adaptability, flexibility, versatility in problem solving and decision making skill. Computational Intelligence has potential competency to capture and compare real time data. Application of Computational Intelligence is not yet explored in machining problems to its fullest potential Most of the machining problems are modeled and optimized through conventional techniques. In the present work different soft computing technique is applied over machining system to optimize machining performance and recognize machining pattern with a case study from literature [2-9] in which extensive machinability aspects of AISI 4340 alloy steel with different machining characteristics during, hard turning. Operation with different faceted is discussed. 1.7 Objectives In the present work, an attempt is made to apply the Computational Techniques and their synergies with the objective to optimize and build prediction models for conventional machining system. Different Computational methods are applied over the machining

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 10


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

system for optimization and learning of machining data Further, the optimized and learnt data are compared with the result obtained from the literature 1. To optimize hard turning of AISI 4340 steel by applying evolutionary algorithm and swarm intelligence (i.e., Non-Dominated Sorting Genetic Algorithm, Strength Pareto, Particle Swarm Optimization technique) and compare optimized results with literature. 2. To develop prediction models for hard turning of AISI 4340 steel by applying intelligent learning techniques i.e., Neural network (NN) and Adaptive NeuroFuzzy systems (ANFIS). (Tagaki-Sugeno fuzzy based Neural network) and compare the prediction model with the experimental statistics. 3. To develop supportive-combination of evolutionary based prediction neural network, furthermore the predicted model will be tested with experimental data. 4. To develop Adaptive Neuro-Fuzzy based evolutionary estimator for predicting optimized parameters for AISI 4340 steel hard turning machining operation. 1.8 Methodology The adapted methodology is developed (Fig. 1.5) to achieve the above mentioned objective with focus to optimize and develop prediction models. 1. Methodology follows two directions where in one division optimization and prediction models are applied exclusively and another division supportive– combination of optimization-prediction models is applied. 2. After obtaining results from each technique comparative evaluation of respective techniques is made with the experimental statistics of hard turning operation.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 11


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.1.5 Applied Optimization and Learning Methodology 1.9 Thesis Outline The present thesis comprises of six chapters, In chapter 1, a brief introduction of Computer Intelligence along with the various techniques and their combination is discussed, applications of CI techniques how CI can be applied to machining systems, problem statement, objectives, and methodology for the current work has been discussed. In Chapter 2, a review of literature pertaining to problem and objectives is made and conclusion from literature is drawn.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 12


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

In Chapter 3, Optimization techniques is applied to machining model of hard turned AISI 4340 steel both evolutionary and swarm intelligence based algorithm is described in brief, mathematical aspects and the pseudo code along with their results is discussed. In Chapter 4, applied predictive models i.e., neural network and adaptive neuro fuzzy based learning model are discussed. Mathematical aspects and pseudo code along with their results is discussed. In Chapter 5, synergies of CI techniques along with their mathematical background and applied pseudo code is described the results of the applied techniques are discussed. Finally, in Chapter 6 results of applied techniques relative comparison and evaluation with the experimental statistics is discussed. Conclusions of present research work and future scope is briefed.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 13


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

CHAPTER-2 2.1 LITERATURE REVIEW The literature was explored with an aim to gather research work of authors who utilized different evolutionary and learning techniques. In the literature authors have utilized optimization and predictive techniques exclusively; multi-objective machining systems were converted to single objective for performance evaluation. Many authors have optimized machining systems with unit control parameter to enhance machining performance avoiding the complex nature of conflicting objective. Literature review was emphasized on the process parameters and control parameters utilized to model machining system, applied optimization techniques and their degree of accuracy in comparison to conventional techniques. Ansalam et al. [10] investigated on improvising machining performance of hard turning operation on SS420 Steel. Process parameters considered were cutting speed, feed rate, and depth of cut to model surface roughness. RSM based regression model was built for prediction and further improvised optimization technique was applied viz., Integrated GA (IGA) which was comprehended with Conventional GA (CGA). The IGA gave better results than CGA. Hesam et al. [11] executed EDM process on DIN 1.452 stainless steel in which surface roughness and white layers were control parameters. The machining model was built on Taguchi technique and NSGA II was applied for optimization which could produce convincing results. Garg et al. [12] improvised machining turning operation of AISI 1040 Steel with surface roughness as a control parameter. Taguchi technique was applied to model surface roughness apart from that (Artificial Neural Network) ANN, (Support Vector Regression) SVR techniques were used to build regression. Genetic programming (C-GP) coupled with classifier was used as optimization technique. The results suggested that C-GP was on par with the ANN but SVR performed poorer than C-GP and ANN.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 14


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Khaider et al. [13] examined hard turning operation of AISI 52100 bearing steel with CBN (7020). The machining performance was measured in surface roughness, tool wear and material removal rate were modeled with Taguchi, RSM and grey-relation, these models were utilizes to optimize performance by applying GA and the results obtained from GA predicted parameters which gave better machining performance. Ozel and Karpat [14] investigated on enhancing the performance of AISI H13 grade steel turning with CBN tool. Prediction model for surface roughness and tool wear was built on process parameters (cutting speed, feed rate, depth of cut). Experimental data of AISI52100 steel was referred form literature and further data experiment was performed on AISI H-13 steel these data was utilized to train neural network, Regression was also carried out, two feed-forward neural networks was modeled. In the first model the input layer constituted edge geometry, hardness, cutting speed, feed rate and depth of cut to predict tool wear and surface roughness, while in the latter network material hardness, cutting speed, feed rate, depth of cut and forces were utilized to model tool wear and surface roughness. The latter network performed well than the former. Alhameri et al. [15] studied multi pass turning of austenitic AISI 302 Steel. Box-Behken design was utilized to develop model. Prediction models were also built by Regression analysis and NN to predict Tool life and machining economics with motive to minimize machine economics and maximize tool life. Abbas et al. [16] carried out research on turning operation of J steel with TungstenCarbide insert. Models were built to predict Surface roughness and Material removal rate by applying Taguchi technique. The formulated regression equations were utilized as objectives with appropriate constraints in process parameters. Multi-objective EGO algorithm was implemented to optimize machining performance. Zhenghua et al [17] investigated high speed milling aluminum alloy AlMN1CU with Carbide tipped tool. Both Linear and quadratic regression models were built and Bayesian Neural nets (BNN) was built using experimental data to predict surface roughness, the regression model built was utilized as objective functions with precise

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 15


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

constraints applied to process parameters. GA was applied to optimize parameters GA predicted optimized parameters for each surface roughness which was verified by BNN. Yunguang et al. [18] worked on micro-grinding of nickel based super alloy (DD98), surface roughness was modeled in linear and non-linear degree using CCD based RSM. GA was applied to predict control parameters for best machining performance. The results were verified experimentally and were found to show pretty good agreement. Shaharam et al. [19] examined on cellular manufacturing systems with objective to minimize cellular movement distance and machine idle time. Regression models were developed and optimization techniques viz., NSGA II, lingo, Fuzzy-GA was applied and results concluded that NSGA II gives better results than Lingo and Fuzzy-GA. N.Alberti and Perrone [20] worked on multi pass turning operation to predict least power consumption, machine economics and surface roughness for which three different modelling approaches was adopted viz., deterministic model, possibilistic model, a fuzzy possibilistic-GA model with constrained and unconstrained search space. The results established that fuzzy-possibilistic model predicted most failures and fuzzy-possibilisticGA optimized objectives to practically feasible solution. Garge et al. [21] experimented on EDM of Titanium and Inconel alloy in which surface roughness and cutting speed were control parameters modeled with process parameters. NSGA II was applied to optimize performance. Pramanic et al. [22] worked on EDM of ZrB2 where cutting speed, material removal rate, and surface roughness was modeled with process parameters by applying Taguchi technique and optimization based on Taguchi based grey relation. ANN was used to predict cutting speed and surface roughness, the predicted accuracy was checked with experimental statistics for confidence level and it gave the appreciative result. Sahali et al. [23] worked on multi point turning operation, modeling machine economics with process parameters and constrains in surface roughness, chip-tool temperature, tool life and force was applied. Optimization techniques applied were viz., Deterministic

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 16


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

technique, probabilistic technique, probabilistic- NSGA II (P-NSGAII).Results concluded that P-NSGA II outperformed deterministic and probablisitc technique. Dureja et al. [24] reviewed different optimization and modeling techniques used in hard turning operations viz., RSM, Taguchi, Regression analysis, NN, Fuzzy modeling, GA. Ganesan and kumar [25] investigated in performance enhancement of turning operation by predicting machining cost, machining time, tool wear with process parameters and GA was applied to optimize objective function. Jawahir [26] presented an analytical and numerical solution to 2D and 3D chip formation, hybrid predictive model was developed to characterize and optimize chip breakability and chip curl geometry. Furthermore, GA was used to optimize chip formation so that machining happening according to desirability, the optimized results were verified with FEM simulation results. Sundaraman et al. [27] in contrast worked on fixture design and layout of end milling; quadratic model was built using RSM and optimization done by GA and PSO. The model was built to predict work piece deformation with positions of clamp and location as parameters. The results of optimization suggested that RSM-PSO gave better solution than RSM-GA technique. Furthermore, these results were compared with FEM simulation of fixture layout. Costa [28] investigated on multi-pass turning with the objective to minimize unit production cost. The objective was constituted of actual machining cost, machining idle cost, tool replacement cost. The characteristic equation was built on process parameters viz., cutting speed, depth of cut, feed rate both in rough and finish pass, operation constraints were applied in tool life, cutting forces, power and surface roughness. A novel hybrid technique in PSO was formulated for optimization. Furthermore, this solution was compared with other techniques suggested 2.035 unit production cost, while FEGA gave 2.3057 as optimal cost SA gave 2.29, MGA gave 2.30, HC gave 2.27 and ACO gave 2.25.The hybrid PSO technique could optimize solution superior than other techniques.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 17


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Bharathi et al. [29] investigated on turning operation with diamond shape tungsten carbide tool on four different materials i.e., brass, aluminium, copper and mild steel. Machining was characterized by forces, surface roughness with cutting speed, feed rate, and depth of cut as process parameters. Furthermore, these equations were optimized using PSO technique; the optimal solution suggested the trend that higher cutting speed, lower feed, and depth of cut gave better surface roughness. The optimal surface roughness obtained from PSO for brass was 0.07μm, for copper 0.08μm, for aluminium 0.08μm and mild steel 0.08μm. Bharathi et al. [30] carried out investigation on milling operation of aluminum bar with carbide tool. Machining time and surface roughness were characterized with process parameters such as spindle speed, feed rate, and depth of cut. The characterized equations were optimized using PSO technique. The solution obtained from PSO was verified by conducting confirmation test. The solution trend showed that higher speed, lower depth of cut, lower feed rate gave better surface roughness and feed rate had a greater influence on surface roughness. The prediction ability of present approach was found to be 96 % for machining time and 85% for surface roughness. Bharathi et al. [31] investigated on modeling and optimizing both turning and grinding operation. The turning operation was done on single and multi-pass while grinding was done in the single-pass. The performance of turning was measured on machining time while grinding was done on machining time and material removal rate. Optimization technique applied were PSO, GA, and SA whose optimal solutions were comprehensively evaluated. The computational time obtained by PSO in single and multi-pass turning was 11 sec and 12 sec respectively, while for grinding 4 sec. Similarly, results of GA gave 15 sec in both single and multi- pass turning, while in grinding it gave 6 sec as optimal computational time. Likewise in SA, it was 12, 13 and 5 respectively. Optimal material removal rate in grinding was in the range of 0.17-0.44 μm, from the results it could be inferred that PSO proved to be better than GA and SA. Chandrasen et al. [32] reviewed different soft computing techniques that could be applied to machining performance prediction. Any machining system could be VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 18


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

generalized by its corresponding inputs and outputs. Input in general are process parameters, material properties, sensory feed and output of the system are concerned about machining performance i.e., dimensional deviation, cutting forces and tool wear, after machining characterizations done various soft computing techniques are applied to optimize machine model. The review concluded that best strategy to predict performance is to couple fuzzy with a neural network. Likewise, to optimize precisely GA, PSO and similar heuristic techniques are the best technique. Prabhakaran et al. [33] carried out work on machining fixture analysis where location and displacement of clamp and locator were objective functions. Regression models were developed for displacement and location and optimized using GA and ACA. Ant colony algorithm gave near optimal solution than GA. Farahnakian et al. [34] investigated end milling operation; performance was modeled in cutting force and surface roughness with process parameters as cutting speed, depth of cut, feed rate. The characterized equation was utilized to frame optimization problem, coupled PSO-NN technique was applied to optimize. The applied technique gave better Pareto-spread in solution space with good convergence. Yang et al. [35] carried out worked on multi- pass face milling operation. Performance was characterized by unit production cost with process parameters such as number of pass, depth of cut, cutting speed, feed rate. Fuzzy based multi-objective PSO was applied to optimize process parameters which gave better solution with fast convergence. Escamilla et al. [36] experimented end milling with a performance characterized in surface roughness with process parameters cutting speed, feed rate and depth of cut. Taguchi based regression equation was utilized to formulate optimization. PSO was used to optimize parameters. Li et al. [37] worked on in improving the performance of milling operation. The performance was measured such as cutting force, tool life, surface roughness and cutting power with process parameters spindle speed, feed rate, depth of cut. PSO technique was

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 19


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

applied to optimize process parameters to optimize process parameters. The Paretospread of optimal solutions was well spread and converged well. Chen and Li [38] worked on optimizing grinding operation by maximizing material removal rate. PSO was applied to optimize material removal rate. Sukla and Singh [39] investigated on Abrasive Water jet Machining (AJM) of Aluminum alloy with garnet abrasive particles. Machining model was built for kerf width and taper angle prediction with process parameters by applying Taguchi technique and Optimization techniques applied, PSO, Firefly, Simulated Annealing, Black Hole, BioGeographical, NSGA. PSO gave better results than other techniques. Asilturk and Cunkas [40] carried experimental investigation on turning operation of AISI 1040 steel with Al2O3 coated carbide insert. Tool variable considered were tool material, nose radius, and rake angle cutting edge geometry while work piece variable considered material hardness with cutting conditions such as cutting speed, feed rate and depth of cut. With these variables, a multi-regression model was developed and full factorial experimental design was built Further, ANN was developed with back propagation training algorithm to predict surface roughness. ANN and multi-regression were had close estimated of surface roughness prediction. ANN performed better with 99% regression coefficient and regression with 97% regression coefficient. Senthil et al. [41] predicted performance of cutting tool inserts using neural network. Experiments were performed on workpiece with carbide inserts with process parameters such as cutting speed, feed rate, depth of cut, material hardness, and cutting insert shape (relief angle, nose radius) to model surface roughness and flank wear. The Taguchi based ANN model was built with these process parameters as input layer. Results predicted by neural network model were compared with experimental values which predicted values close to experimental statistics. Miron et al. [42] worked on dynamic characterization and vibration analysis of lathe machining system by which the machine condition was determined. Modal analysis was

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 20


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

done to determine the natural frequencies, the frequency was compared with numerical model and validation experiment was performed. Dilbag and Venkateshwara [43] developed analytical tool wear model while turning bearing steel with ceramic tool. The model incorporated abrasion, adhesion and diffusion wear mechanism further it was validated by conducting experiment. The analytical model had capability of predicting flank wear using cutting parameter and tool geometry. Yahya et al. [44] worked on turning operation of steel at different conditions with P25 HSS tool at different working conditions. Surface roughness, flank wear, and crater wear were modeled with process parameters to determine machinability of tool steel. The relative degree of influence of each parameter on control parameters was quantified. This work can help in sorting the priority of objective functions and the contribution to overall machining performance. Hamdi et al. [45] investigated behaviour of hard turning while machining AISI H11 Steel with CBN tool. Forces and surface profile were considered as process responses. A CCD based RSM was applied to build machining model furthermore a comprehensive analysis was done on influence of process parameters over machining quality. Shihab et al. [46] conducted experiment on hard turning of AISI 52100 Steel alloy with coated carbide tool in which surface roughness and micro hardness were modelled and optimized utilizing CCD based RSM approach. The RSM based optimization technique gave satisfactory results but by reducing multi objective to single objective. Waleed et al. [47] worked on hard turning of AISI 4340 Steel with CBN tool; in his work surface roughness and tool flank wear were modelled by Taguchi technique to form multi-regression equation. This equation was used as objective functions along with constraints in process parameters; the S/N ratio analysis was done on regression to optimize control parameters. saha et al. [48] experimented on EDM Hard facing on Nano- card-11.Mathematical model was built on material removal rate, cutting speed and machining time for both

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 21


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

brass wire and zinc coated brass wire by applying both RSM and multi criteria grey relation. RSM was utilized to optimize the objectives individually for both the wires. Emeryl et al. [49] worked on hard turning of Ni-Steel alloy (62HRC) with CBN tool insert, the performance was modeled to predict cutting forces, surface roughness with process parameters by applying Taguchi method and optimization was done by Taguchi based S/N ratio. The results were in agreement with experimental data with good level of confidence. Ilhan and asks [50] worked on hard turning of AISI 4140 (51HRC) steel by carbide tool coated with Al2O3 and TiC. A three-level full factorial with Taguchi based experimental design was applied to model surface roughness by applying cutting conditions and control factors. The process variability was measured by S/N ratio. Taguchi based S/N response suggested that larger difference in S/N ratio will have more significant effect on surface roughness. The result of process variables for optimum surface roughness was 120 m/min, 0.18 mm/rev and 0.4 mm cutting speed, feed rate and depth of cut respectively. Gaurav and Choudhary [51] focused study on hard turning of EN31 bearing steel (5862HRC) with CBN tool insert. A three-level full factorial experimental design was developed; ANOVA was performed to find the relative contribution. RSM was utilized to build regression equation on cutting forces and surface roughness, further RSM optimization was done. Results showed that depth of cut had more influence on cutting forces and while speed had the least influence. Results also revealed that initially forces decreased with increasing speed later on increased along with speed due thermal softening of tool material. Ashvin and Nanavati [52] enquired on turning operation of AISI 410 steel with carbide inserts of TNMG series differing in nose radius. A three-level full factorial experimental design was done, further, RSM was utilized to model and optimize surface roughness which suggested optimal solution 225 m/min,0.1 mm/rev, 0.3 mm, 0.12 mm for cutting speed, feed rate, depth of cut and tool radius respectively.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 22


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Asilturk and Suleyman [53] investigated hard turning of AISI 304 austenite stainless steel with carbide inserts (SNMG series). A three-level full factorial Taguchi based experimental design was built. RSM based regression equation was modeled for surface parameters (Ra and Rz) using process variable, S/N ratio was determined then RSM based optimization was done. The optimized control factors setting for Ra was found to be cutting speed 50 m/min, feed rate 0.15 mm/rev, depth of cut 1.5mm and for Rz was cutting speed 150m/min, feed rate 0.15mm/rev, depth of cut 1mm. These authors applied Taguchi and RSM to model and optimize machining parameters. RSM relates response based input parameters by experimental statistics and applying regression. RSM consists of three stages design of experiment, regression, and optimization to find best optimal solution RSM is coupled with meta-heuristic technique. Aggarwal and Singh [54] reviewed different machining modeling technique for conventional machining model and types of optimization methodology to optimize and characterize machine model. Chinmaya et al. [55] experimented on hybrid machining where laser assisted machining (LAM) was coupled with turning operation. High strength alloy (Ti-6Al-4V) was machined with cobalt bound tungsten carbide, liquid nitrogen was used coolant. The LAM hybrid turning operation reduced specific cutting energy and improved surface roughness when compared to conventional machining. Wang et al. [56] worked on multi-pass turning operation of AISI 1045 with a different set of tools (TNMG carbide inserts). A hybrid model was built to predict machining performance, surface roughness, forces, and chip breakability were characterized by process parameters with operation constraints such as surface roughness, forces, and tool life. RSM was utilized to optimize process parameter. The hybrid model developed could predict slip line field accurately, which was verified by Finite Element Modelling results. Devender and Kumar [57] worked on turning of Aluminum matrix composites reinforced with SiC (Al 6061) using coated tungsten carbide tool. The effect of reinforcement on cutting forces was characterized to improve machining performance and VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 23


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

it was found that weight forces have maximum impact on the cutting forces. A quadratic model was developed using RSM. Optimal solution concluded that cutting forces was majorly affected by the type of reinforcement. EranAlsan et al. [58] investigated hard turning of AISI 4140 steel with ceramic tool mixed with Al2O3 and TiCN. Machining performance was modeled using Taguchi based RSM on flank wear and surface roughness as control parameters with process variables. The optimization results obtained from RSM technique suggested cutting speed 250m/min, feed rate 0.1mm and depth of cut 0.25-0.4 mm for surface roughness and flank wear. Hashimoto et al. [59] identified the fundamental difference in the surface integrity of hard turning and ground surface, their subsequent impact on rolling contact fatigue life. The work concluded that the mechanical deformation could play a large role during hard turning than grinding while the size effect in grinding introduced surface hardness furthermore the hard turned surface may have more than 100% longer fatigue life than a ground one with an equivalent surface finish due to very different characterization of surface integrity. The effect of turned or ground surface free of white layer was clarified a super finished turned surface may have twice a fatigue life than ground surface. Ozel et al. [60] investigated on hard turning of AISI 4340 steel with uniform and variable edge PCBN insert where the forces and tool wear was measured and 3D finite model was utilized to predict chip formation, temperature and tool wear on both type of inserts the predicted tool wear and forces were compared with experimentation. The result showed that the variable edge tool insert has advantage of less tool wear and good temperature distribution profile. Ravinder and Santram [61] investigated the effects of cutting parameters on surface roughness in turning of Al7075 hard ceramic composites and Al7075 hybrid composite using polycrystalline diamond tool (PCD) dry turning was conducted to examine the trend of roughness by using roughness tester for both composites. It was concluded that surface roughness of hybrid composite was lesser in all combination of experiment. Further VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 24


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

RSM based artificial neural network was applied to validate the results obtained during experimentation and to protect the behavior of the system under any condition within the operation range. Mia and Dhar [62] developed a predictive model of average tool-workpiece interface in hard turning of AISI 1060 steel by coated-carbide insert. Cutting condition used were cutting speed, feed rate, and depth of cut utilized to model temperature profile. Experiment was conducted in both dry and high pressure coolant environment with fullfactorial design. Temperature was measured using tool-work thermocouple. Response Surface Methodology (RSM) and Artificial Neural Network (ANN) were employed to predict the temperature. The accuracy of both RSM and ANN model were in region of acceptance. The regression coefficient of ANN for both the environment was greater than 99.8%. ANN model demonstrated a higher accuracy which was found convincing if employed for controlling cutting temperature in turning of hardened steel. Pontes et al. [63] worked on turning of AISI52100 hardened steel with multi-layered coating (Al2O3+Tic+TiN) chamfer edge. Experiments were conducted with training sets of different size to compare performance of best network in each experiment. Process parameters considered were cutting speed, feed rate, depth of cut to model performance in surface roughness. Radial base function (RBF) neural network was developed with the use of Taguchi’s orthogonal array as a tool to design parameter of network. The factors considered in designing RBF-NN were number of radial units, algorithm for selection of radial center and algorithm for selection of spread factor for evaluating performance of RBF-NN. The results revealed that algorithm for calculation of radial spread factor was most influencing among the three factors RBF-NN trained gave least mean standard deviation for worst trained case. The results suggested that DOE based RBF network are more efficient and effective than trail and error based NN architect. Gaitonde et al. [64] investigated the influence of process parameters on machinability characteristics of turning AISI D2 (cold work) tool steel with different ceramic inserts. A multi-layer feed forward Neural was developed with inputs as ceramic insert grade, cutting speed, feed rate and machining time to predict specific cutting force, surface VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 25


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

roughness and tool wear. Statistical comparison was done between the predicted results and experimental results further more interaction effects among the process parameters were studied. Wang [65] developed neural network based optimal estimator for predicting CBN tool wear during hard turning operation. Prediction model was based on fully forward connected neural network, with inputs as cutting condition and machining time and predicted output in tool flank wear. The feed forward fully connected neural network (FFCNN) based optimization was validated with experimental data. Comparison showed that the (FFCNN) estimated a close value to experimental tool wear and developed FFCNN model was found to be faster, accurate than other neural network approaches. Umbrello et al. [66] developed predictive hybrid model based on neural network and finite element method with objective to predict residual stress profile in hard turning for different combination of material properties, cutting tool geometry and cutting condition. A converse prediction of cutting conditions and geometry was made for a given residual stress profile which acted as constrained based process parameters determination. Furthermore, this model was utilized as closed feedback where the predicted residual stress of ANN were applied to simulate cutting condition in FEA and vice-versa The results obtained from ANN based FE simulation gave practical results. Ravinder and Santram [67] investigated the effect of cutting parameters (cutting speed, feed rate and approach angle) on roughness while turning Al 7075 hard ceramic based composite using polycrystalline tool diamond tool (PCD). The surface roughness was modeled by both RSM and ANN. Moreover, the influence of parameters on surface roughness was analyzed both RSM and ANN model correlated fairly to the experimental data. Yildiz [68] presented a Hybrid Differential Evolution Algorithm (HEADA) for minimizing production cost in multi-pass turning operation the algorithm was illustrated two case studies. Taguchi based differential evolution was applied to solve machining economics problem. Further hybrid differential evolution based optimization technique VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 26


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

was compared with PSO, HEGA, Scattered Search (SS), Simulated Annealing (SA), Pattern Search (PS), Floating encoding genetic algorithm (FEGA) and Hybrid Harmony Search (HSS) HEDEA outperformed all other techniques. Kara et al. [69] worked on turning of AISI 310L Stainless steel with both coated and uncoated cutting tool (TiCN+Al2O3+TiN). The cutting condition (cutting speed, feed rate, and depth of cut) were used to model tangential forces and feed force Prediction model was developed for both the responses with ANN. Two learning methods were deployed i.e., scaled conjugate learning and Lavenberg-Madquart learning. The predicted forces were accurate with error within 5%. Sener Karabulut [70] worked on milling of metal matrix composite (Aluminum Alloy 7039/Al2O3 powder metallurgy) with CVD carbide tools the process parameters were material removal rate, cutting speed, feed rate and axial depth of cut to model machining performance such as surface roughness and cutting force ANN model was developed with cutting condition to predict performance. The predicted performance was compared with experimental model which gave close results with 99.8% regression. 2.2 Literature Summary Authors have applied Evolutionary techniques on different machining systems for optimization. From the literature it could be concluded that very few authors have applied swarm intelligence techniques for optimization. While concerning to regression based learning techniques, authors have applied neural network based prediction models and few authors have utilized hybrid learning based evolutionary optimization estimators.In the literature authors, have utilized optimization and predictive techniques exclusively; multi-objective machining systems were converted to single objective for performance evaluation. Many authors have optimized machining systems with unit control parameter to enhance machining performance avoiding the complex nature of conflicting objectives. The literature motivates for research work in multi performance of machining system and multi-regression prediction models as literature lacks application of synergies of computational techniques for performance evaluation and prediction. VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 27


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

CHAPTER 3 GLOBAL SEARCH ALGORITHMS FOR MULTI-OBJECTIVE OPTIMIZATION 3.1Introduction In this chapter Multi-Objective Optimization problem (MOOPs) of machining system is solved by applying Evolutionary Algorithms (EA’s) and Swarm Intelligence (SI) which are Global Search based Optimization techniques (ref Fig.1.3). In the first section Evolutionary Based NSGA II (Non-Dominated Sort Genetic Algorithm) and SPEA 2 (Strength Pareto Evolutionary Algorithm) algorithms are applied to Machining MOOPs and a comparison of between both is made on basis of diversity of solution. In the second section Particle Swarm based Swarm Intelligence is applied to existing Machining MOOPs and optimized results are interpreted through, swarm surface and pareto plots. In the third section comparative evaluation of obtained results between EA and SI technique are analyzed and the difference in the nature of solution space between the two search optimization techniques is assessed. The workflow of this chapter is explained below,

Fig.3.1 Workflow for Chapter 3

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 28


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

3.2 Multi-objective Optimization Most of the practical problems are complex and their definition of optimality is not simple as they need to satisfy multiple competing objective functions at the same time. Moreover, some of these objectives may have conflicting relations with others, which makes the optimization difficult. Problems requiring simultaneous optimization of more than one objective function are known as multi-objective optimization problems (MOOPs). They can be defined as problems consisting of multiple objectives, which are to be minimized or maximized while maintaining some constraints. Formally, they can be defined as: Minimize/maximize f (x) Subject to objectives and constraints

g j ( x)  0 j  1, 2,3... j hk ( x)  0,

k  1,2, 3,..., K

Here, the problem optimizes g j ( x) objectives and satisfies J inequality and hk ( x) equality constraints. This type of problem has no unique perfect solution. In traditional multi-objective optimization, it is very common to simply combine all the objectives together to form a single (scalar) fitness function. But the obtained solution using a single scalar is sensitive to the weight vector used in the scaling process. This requires knowledge about the underlying problem which is not known before in most cases. Moreover, the objectives can interact or conflict with each other. Therefore, trade-offs exists when dealing with such MOOPs, rather than a single solution. Most MOOPs do not provide a single solution; rather, they offer a set of solutions. Such solutions are the ‘trade-offs’ or good compromises among the objectives. In order to generate these tradeoff solutions, an old notion of optimality called the ‘Pareto-optimum set’ is normally adopted. In multi-objective optimization, the definition of quality of solution is more complex than for single-objective optimization problems. The main challenges in a multi- objective optimization are: converge as closely as possible to the Pareto-optimal front, and VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 29


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

maintain as diverse a set of solutions as possible. The first task ensures that the obtained set of solutions is near optimal, while the second task ensures that a wide range of tradeoff solutions is obtained. 3.3 Application of MOOPs to Machining system The Machining data from hard turning of AISI 4340 steel [2] is utilized as machining objectives in which machining is performed on two different hardness and regression equations were built using RSM with process variables involving in cutting speed, feed rate and depth of cut to model surface roughness, cutting forces and tool life. The machining constraints and objectives are as follows, Table 3.1 Machining Constraints [2] Parameter

Constraints

Process parameters

Velocity(m/min)

142

265

125

175

Feed rate(mm/rev)

0.15

0.25

0.15

0.25

Depth of cut(mm)

1

2

1

2

Tangential forces(N)

337

1197

492

1296

Axial forces(N)

219

605

298

663

Radial forces(N)

197

496

256

564

Control parameters

lower Higher lower Higher bound(35) bound(35) bound(45) bound(45)

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 30


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

3.3.1 Machining Model (Surface Roughness Cutting force components and Tool Life work material hardness: 35 HRC). Ra  12.793  0.03118   v  – 28.8786   f  – 2.8599   d   0.0354   v    f   0.000236   v    d   11  f    d   0.00000381  v 2   32.039   f 2   0.2853   d 2 

 3.1

FT  373.0294  0.5308   v   788.39   f   697.2733   d   7.2420   v    f   1.9860   f    d   235   f    d   0.00075  v 2   6659.8  f 2   0.598  d 2 

 3.2 

Fa  375  2.971  v   360.24   f   76.68   d   7.9052   v    f   0.4   v    f  145   f    d   0.000398   v 2   1528.4   f 2   66.71  d 2 

 3.3

Fr  239.69  2.4094   v   755.0606   f   133.18   d   0.0559   v    f   0.2472   v    d   585   f    d   0.000415   v 2   2593.5   f 2   22.93   d 2 

 3.4 

3.3.2 Machining Model (Surface Roughness Cutting force components and Tool Life work material hardness: 45 HRC).

Ra  11.3037  0.0614   v   16.075  f   2.3075   d   0.0006   v    f   0.102   v    d   3.7   f    d   0.0000128   v 2   49   f 2   0.22   d 2 

 3.5

Ft  50.57  0.1484   v   3270   f   143.102   d   33.5   v    f 

 1.11  v    d   1175   f    d   0.01183   v 2   5909.091  f 2   38.9091 d 2 

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 31

 3.6 


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fa  260.483  0.86523   v   287.159   f   113.8068   d   6.9   v    f   1.39   v    d   695   f    d   0.01210   v 2   3477.273   f 2   41.271  d 2 

 3.7 

Fr  86.3465  1.5970   v   948.0681  f   212.0113   d   7.9   v    f   1.19   v    d   95   f    d   0.01518   v 2   2695.4545   f 2   26.95   d 2 

 3.8

3.3.3 Tool life model for 35 HRC

Tf 

(v)

0.59

423  ( f )0.4697  (d )0.47

 3.9

Tool life model for 45 HRC

Tf 

(v)

0.59

23135.13  ( f )0.4697  (d )0.47

 3.10

3.4 Evolutionary Algorithms (EAs) Optimum seeking is one of the central issue in Manufacturing system. Every problem solved is outcome of best possible choice for which a variety of tools and techniques have been developed and applied to systems for optimum seeking. Meanwhile optimum seeking in nature, biological and social systems takes place in a completely different way i.e., natural evolution they have adapted themselves to a constantly shifting and changing environment in order to survive. Those weaker and lesser fit members of species tend to die away leaving create stronger and fitter to mate create to create offspring and ensure the containing survival of species and it is upon this dictated idea

that evolutionary computing is based on. Evolutionary computing is

emulation of the process of natural selection in search procedure (as shown in Fig.3.1).

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 32


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

 ( p  1)  [ ( ( p))] Fig 3.2 Evolutionary Model The current EAs applied in Multi-Objective Optimization Problem (MOOPs) and the combination became known as a multi-objective evolutionary algorithm (MOEA). An MOEA will be considered good only if both the goals of convergence and diversity are satisfied simultaneously. The MOEA’s population-based approach helps to preserve and utilize the non-dominated diverse set of solutions in a population. The MOEA converges to a Pareto-optimal front with a good spread of solutions in some fixed number of generations. Most MOEAs use the concept of domination to attain the set of Paretooptimal solutions. 3.4.1 Mathematical Formulation of Evolutionary Algorithms Evolutionary algorithm is stochastic in nature. The probability of finding best solution or no solution is equally likely, if the parameters of genetic operator are not appropriate depending upon nature of problem in hand. A prior convergence analysis is essential for favorable working of Evolutionary algorithm So at each genetic operations the probability of obtaining best solution and its heritance in the subsequent operation is essential.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 33


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

3.4.1.1 Definition of Evolutionary systems Considering a generic function f  x   0 on X. with no constraints imposed on X. further EA system which does not use any specific properties of the set X, the only condition is that the function f should be defined at every point of X i.e, X  R n . Then the problem of optimization can be defined as,

Max / Min xX  f ( x)

where

f : SD  R

f  x   0 X  Rn Let S be a space of binary string and C(X) be an encoding function: C : X  S and we discretize search space X to XD by a simple binary code as encoding function the optimization problem is converted to finite set SD  S , SD  C ( X ) .

max SD S f (s)

where

f : SD  R

There are Variety of Evolutionary system with different types of selection, crossover and mutation. This section discusses most generalized terms with no bias in different genetic operation. 3.4.1.2 Convergence Analysis of Evolutionary Algorithm Consider the following events in evolutionary system with their respective properties let P be a population and n be the size of population

P  A | H   Probability that the population does not contain solution after mutation provided it contained after crossover.

P A | H  Probability that the no solution is found after mutation and crossover



P H  Probability that solution will be found after a crossover.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 34


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

_ P  H   No solution after crossover.  

max (f(S)) where f  0, S  S S is a finite S  2m m

Encoding capacity (No. of bits).

3.4.1.3 Criteria for mutation * Let Pmt be the probability of mutation s  S be selected individual, then the probability

of individual not getting mutated be P S x  mut  PD 

 and not giving rise

S * in the

population  m / 2 integral part of quotient m/2. and P S  S n  is probability of mutation that S mutates to S * . Let K be the no. of different bits in S & S *

 0  k  m

then the probability that S

mutates in S * is given by binomial theorem.

P

 P , P   PS  S   1 P  

k

k

m

mt

k

m

Pm  Pmt 

1  k mk P 1  Pmt   : Pmk is The probability of ‘K’ mutation k  m   Cm

1  Pmt 

mk

 is the probability that the remaining (m-k ) bits do not mutate

1  probability that precisely the necessary K-bits mutate but not other bits. Cmk

Then P{s  s '  s*}  1  P{s  s*}  1 

m pmt is the probability that s does not mutate Cm[ m/2]

in S*

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 35


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Now, the probability that best solution is obtained after mutation is

P S S

*

Pmtm 1 k 1 m mk  k Pmt 1  Pmt   k Pmt  ( m /2) Cm Cm Cm

For an individual to get mutated the minimum condition is that half of its bit length must participate in mutation and in contradictory assuming that mutation does not happen then,

P S *  mut  P   P{  S  S '  S *} SP

  Pmt m   P     P S  S  S *    1  ( m /2)   1  mmt/2  s p s p  Cm   Cm 

'

(3.2)

The above condition is valid when the probability of mutation pmt <0.5, if the probability of mutation exceeds 0.5 then the probability that the population does not contain solution S* is given as follows.

 (1  Pmt )  P S *  mut ( P)  1   Cmm /2  

P S *  mut ( P)

Now evaluating binary strings over the objective f  P k    max  ( f ( S )) if elite method  SSp   is utilized for selection then, f x  f ( Pk )  f ( Pk 1 )  ...........  f ( P0 ) Where f ( P K ) which are randomly calculated

and subsequently we get, f f p p 'f

 

0 f ( p kf )  f ( s ) f p p kf  Pf0

Re sults Pr obability

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 36


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Here Pfk  Probability that f ( P k )  f after k th iteration and  Pfk  is the probability that f ( P k )  f * after the k th iteration now that the required expectation probability does not

decrease, E  P k    f .Pfk , k  0,1, 2..., E  P k   E  P k 1   E  P k 2  .......  E  P0 

Now if l  individual with fitness f *( x) in the general population S (i.e. optimization) problem has ‘l’ solution

S

 1

,........., Sl

if elistism is applied then the expectation is

unchanged i.e, E[ P K ]  f * Consider situation when A’ =A that no solution is found after the first iteration, consequently suppose that a hypothesis H stating at least one of the solutions results from  _  _ crossover, Then possible events P  A  P  A | H   P H   P  A | H   P  H     

3.4.1.4 Criteria for Crossover Estimating P{H} from the above which only differ from the pmt by the fact that population contains one solution before mutation, where as it does not contain any in second case. Applying from the above, PA | H  P

S .......S 

_   P A | H   P  

*

*

1

l

 P   mut cross( P 0 )  1  mmt/2   Cm 

S .......S  *

*

1

l

  Pmt m    1  m /2  C m 

n l

  Pmt m    1  m /2  C m 

 P   mut cross( P )  1  mmt/2   Cm 

0

While Estimating P  H  we say that pairs

 S1 , S2 

n

n l

n

is good if it yield solution after

crossover it can be concluded that a pair is good if both the individuals in the pair contains fragments of the same solution as sub-string. VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 37


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Consider the following events B  Event of at least a good pair will be chosen.

C  Event that all the pair is chosen for crossover are good.

P H   P H | B P B  P H | B  P H | C This can be concluded from the fact that if all the pairs are good then the probability is maximum compared to other events, So the probability that a good pair (s1,s2) yields a solution .can be written as

q pc  pc q  m  1, m 1 P{s1  s2  s*}  1  pc

P{s1  s2  s*} 

In contrary P{H | C}, the probability that a solution does not arise after a crossover, provided that all pairs are good:

P{H | C} 

( S1 , S2 )Cross ( P )

P{s1  s2  s1* ,...sl }   (1  pc )  (1  pc ) n

P{H | C}  1  (1  pc ) n Assuming that atleast ‘n’ pairs take part in the crossover. Then probability of good pairs after crossover

P{H }  P{H | C}  1  P{H | C}  (1  (1  pc )n Now if Event A happens then the possibility of event A can be written as follows

P{A}  P{A | H }P{H }  P{A | H }P{H }  (1 

m n pmt pmt n n   p   ) (1 (1 ) ) (1 )n *1 c [ m /2] [ m /2] CM Cm

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 38


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

S  (1 

m pmt )n (2  (1  pc ) n ) [ m /2] CM

(3.3)

Therefore probability in kth iteration can be written through mathematical induction as, P{ Ak } = P{AK | Ak 1}P{AK 1}  SS K 1  S K , where P{ AK | Ak 1} is the probability

That no solution arises after the kth iteration now from above frame works the expectation of solution after the kth iteration

E[ P K ]   fp kf  f * p kf  f * (1  S k )  f * With this following conclusions can be drawn from the expectation regarding the parameters which influence mean convergence rate 1. AS pc ( 0  pc  1 ) increases S also increases 2. As pmt (0  pmt  1) increases S decreases 3. As m increases S also increases 4. The dependence of S on n can be drawn from the below expression

(1 

m pmt )n  0 as (n infinity) [ m /2] CM

n and (2  (1  pc ) )  2 as (n infinity)

Minimum value of S =1 for n=0 In order extract the extreme limits of convergence we consider the extreme of the function S(n). Finding optimal parameters for above events we use eq. m pmt Let a  1  [ m /2] CM

b  1  pc

S (n)  a n (2  bn ) VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 39


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

For extremum S ' (n) =0 S ' (n)  a n (2  bn ) ln a  bn ln b

The optimum n is given by n= log b

2log a ln ab

3.4.2 Key points from convergence analysis 

With increase in population size the string length increases, reaches optimum peek and then decreases. The least string length possible for encoding is 1 for zero size population.

From above result it can be concluded that n exists and is real and S attains its maximum i.e, as n increase, S first increases and then decrease.

For accelerating convergence pc should be minimum while pmt should be maximum but the drawback of having pc minimum is that the best solution are not inherited to next generation and high pmt destroys best solution space.

With elitism the expectation of solution getting transferred to next generation increases moreover if crossover probability is increased the density of best solution increases including good solution in every iteration.

If the mutation probability is decreased then the string length participating decrease which decreases the passivity of killing best solutions, but if mutation is accurately tuned then there is quite a possibility that worst strings could give good solution.

If the crossover size is increased the convergence decelerates but the chances of obtaining good solution increases. In contrast for accelerating convergence if the mutation rate is increased then good solutions are lost leading to no solution so a good balance between convergence time and crossover-mutation rate is essential.

It is always suggested that in order to obtain good solution if convergence rate is allowed to float freely whenever possible and when convergence rate is strict criteria then it is suggested that the string length is kept minimum so that humming effect and relative degrees of change in string character is merely small.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 40


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

With the key points in mind evolutionary algorithm is chosen with properties which could overcome these drawbacks A fast elite NSGA II and SPEA2 algorithms are utilized to optimize our machining system. 3.5 NSGA II Algorithm

Fig. 3.3 NSGA II Algorithm [69-70] Table 3.2 NSGA II Setting Population size Generation Crossover probability Crossover constant Mutation probability Mutation constant

1000 100 0.8 0.1 0.1 0.2

3.5.1 Initialize Variables and Evaluate Objectives Initialize Variables uses the bounds of ‘V’ decision variables and randomly generates N number of population over the bound ‘range’ and each objective ‘M’ is evaluated over this population ‘pop’ for fitness through Evaluate Objectives. Initialize Variables (N, M, V, range) 1. For i :[1-N] 2. For j :[1-V] 3. V[i,j]= Rmin+ random(0,1)*Range VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 41


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4. Pop: Initialize Variables 5. Evaluate Objectives [Pop, M, V, N] 6. For i :[1-N] 7. For j :[1-V] 8. Objectives[I,j]= f (v) 3.5.2 Non_dominated_sort Now the evaluated objective is sorted using Non_dominated_sort technique where population is sorted on the basis of domination front and over this front the population. Procedure for Non-dominate sorting 1. For each individual p in main population P do the following 2. Initialize Sp = ∅. This set would contain all the individuals that is being dominated by p.

3. Initialize np = 0. This would be the number of individuals that dominate p. 4. for each individual q in P if p dominated q then add q to the set Sp i.e. Sp = Sp

∪ {q} else if q dominates p then

5. Increment the domination counter for p i.e. np = np + 1 6. If np= 0 i.e. number of individuals dominate p then p belongs to the first front; Set rank of individual p to one i.e. prank = 1. Update the first front set by adding p to front one i.e F1 = F1 ∪ {p}

This is carried out for all the individuals in main population P. 7. Initialize the front counter to one. i = 1 following is carried out while the ith front is nonempty i.e. F[ ƒ]= ∅.

8. Q = ∅. The set for storing the individuals for (i + 1)th front.

9. for each individual p in front Fi for each individual q in Sp (Sp is the set of individuals dominated by p).

10. nq = nq −1, decrement the domination count for individual q.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 42


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

11. if nq = 0 then none of the individuals in the subsequent fronts would dominate q. Hence set qrank = i + 1. 12. Update the set Q with individual q i.e. Q = Q âˆŞ q.

13. Increment the front counter by one.

14. Now the set Q is the next front and hence Fi = Q. Non_dominated_sort (pop,M,V,N) 1. For i :[1-N]//Initialize domination set 2. Initiate Pop[i].domination set [] 3. Pop[i].domination count=0; 4. //Initialize empty front F[1]=[] 5. For i :[1-N] 6. For j :[i+1-N] 7. P=pop[i] q=pop[j] //Consequetive population 8. //Check for domination 9. If dominates(p,q) 10. d=dominates(p,q) 11. d=all(p<=q)&&any(p<q) 12. P.dominates set=[P.dominates set ,j] 13. q.dominates count=[q.dominated count +1] 14. if dominates (q.cost p.cost) 15. q.dominate set=[q.dominates ,i] 16. p.dominate set=p.dominated count+1 17. Exchange pop[i] with p and pop[j] with q 18. If pop[i].dominated count==0 19. F[i] = [F[i],i] &pop[i].rank=1 20. While (~front not empty) 21. //calculated the subsequent fronts 22. Exchange p=pop[F[i] & q=pop[j] 23. q.dominated count =q.dominated count-1 24. q.dominated count ==0 VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 43


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

25. Q=[q,j] 26. q.rank=pop(F[i])+1 27. exchange q with pop[j] 28. F[pop[i]+1]=Q Once the non-dominated sort is complete the crowding distance is assigned. Since the individuals are selected based on rank and crowding distance all the individuals in the population are assigned a crowding distance value. Crowding distance is assigned front wise and comparing the crowding distance between two individuals in different front is meaningless. The crowing distance is calculated as below 15. For each front Fi, n is the number of individuals. 16. Initialize the distance to be zero for all the individuals i.e. Fi (dj ) = 0, where j corresponds to the jth individual in front Fi. 17. for each objective function m 18. Sort the individuals in front Fi based on objective m i.e. I = Sort (Fi, m). 19. Assign infinite distance to boundary values for each individual in Fi i.e. I (d k ) = ∞ and I (dn) =∞ 20. for k = 2 to (n − 1) I (d k )  I (d k ) 

I (k  1).m  I ( K  1).m f mmax  f mmin

I(k).m is the value of the mth objective function of the kth individual in I Distance (M, V, N) 1. For front : [i-length[(F)]-1] //Calculate distance 2. //Initiate disyance d=0 3. //Index of fronts F[p,q] 4. S.up=S.up+1 5. S.down=S. up 6. Push[S,pop[i]] 7. I=sort(pop(push(S,pop[i]) VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 44

 3.4 


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

8. Set Fmax=sort(pop[I],first]=inf && Fmin=sort(pop[I],last)=inf I (d k )  I (d k ) 

I (k  1).m  I ( K  1).m f mmax  f mmin

9. Return fronts and distance 3.5.3 Selection Once the individuals are sorted based on non-domination and with crowding distance assigned, the selection is carried out using a crowded- comparison-operator Non-domination rank prank i.e. individuals in front Fi will have their rank Crowding distance Fi(dj ) 1. p <n q if 2. prank < qrank 3. or if p and q belong to the same front Fi then Fi(dp) > Fi(dq) i.e. the crowing distance should be more. The individuals are selected by using a binary tournament selection with crowedcomparison-operator Tournament selection (pop, Toursize, V, M, N) 1. For i :[1-N] 2. For j :[1-Tour size] 3. I:random(N,Tour size) 4. Get [i1,i2]:I(j) 5. //Check rank and distance of candidate 6. [I_Rank, I_distance]=pop[min(pop(Rank),max(pop(distance))] 7. I_min : pop[find(min(pop(rank)&&max(pop(distance))] 3.5.4 Genetic Operators

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 45


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Real-coded GA’s use Simulated Binary Crossover (SBX) operator for crossover and polynomial mutation Simulated Binary Crossover. Simulated binary crossover simulates the binary crossover simulated. Simulated Binary Crossover (SBX) c1, k  0.5 1   k

 p1, k

 1   k

 p2, k



c2, k  0.5 1   k  p1, k  1   k  p 2, k 

 3.5  3.6

Where ci, k is the ith child with kth component, pi,k is the selected parent and βk (≥ 0) is a sample from a random number generated having the density.

p( )  0.5(c  1)1/  c 2

if 0    1

p( )  0.5(c  1) c 2

if 0    1

p( )  0.5(c  1)1/  c 2

if  1

3.7  3.8 3.9

This distribution can be obtained from a uniformly sampled random number u between (0, 1). ηc is the distribution index for crossover

 (u)  (2u)1/( 1) c

1/(c 1)

 (u)  (1/ 2(1  u))

Crossover (parent pop, M, V, Rang, PC) 1. For i:[1-N] 2. If Random(0,1)<PC 3. //Child initiation child1 and child 2 4. Select parents 5. P1: round [N*random(0,1)] 6. P2: round[N*random(0,1)] 7. Parent 1=parent pop[P1,:] 8. Parent 2=parent pop[P2,:] 9. //Simulated Binary Crossover VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 46

3.10 3.11


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

10. Ui[i] : random(0,1) 11. If Uj[i] <0.5 1

  (2U i )1 Pc 12. Else



1 2(1  U i ) Pc 1

13. Evaluate objective(child1,M,V) 14. Evaluate objective(child2,M,V) 15. Polynomial Mutation ck  pk  ( p u k  pkl ) k

 k  (2rk )1/

m 1

1

if  0.5

 3.12 

1/m 1

 k  1  [2(1  rk )]

16. Mutate (parent pop, M, V, Range, Pm) 17. For i :[1-N] 18. P3=round[N*random(0,1)] 19. Child3=parentpop(P3,1) 20. m(i)=random(0,1) 21. if m(i)<0.5 1

 (i )  2 * m(i ) pm 1 else

 (i ) 

1 2(1  m(i )) pc 1

22. Child 3=child3[i]+δ(i) 23. evaluate objective(child 3,M,V,N) Where, rK is a uniformly sampled random number between (0, 1) and ηm is mutation rate. 3.5.5 Recombination of parent and off springs

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 47


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

The best parents from former generations and off springs in current generations are combined to preserve parents so that the parents are conserved in the consecutive generations Replace pop (Intermediate pop,M,V,N) 1. For i : [1-N] 2. Sort pop 3. Max_rank=Intermediate pop[find(maxRank)] 4. For i: 1 to max_rank 5. J=max(find[sorted pop(max>rank)==i) 6. If (j>N) 7. //sorted with rank 8. //find the number of individuals with current rank 9. k=j-N 10. p=sorted_pop(k:N) 11. //Sort according to distance 12. For j : [1-N] 13. F[N+K:]=p(N:j) 14. Elseif j<N 15. F[N:j]=sorted_pop[j:N] 16. For i :[1-Kmax] 17. Pool=round(Np/2),tour=2 18. Parent pop=Tournament selection (pop, Tour) 19. [Child 1,Child 2]=Crossover (Parent pop, M, V, Range, PC) 20. [Child 3]=Mutate(Parent pop, M, V, Range, Pm) 21. Offspring pop=[Child 1, Child 2, Child 3] 22. Intermediate pop =[pop, offspring] 23. Replace pop(Intermediate pop, M, V, N) Table.3.3 Results of NSGA II family of best solution for AISI 4340 35 HRC Steel Vc

f

d

Ra

Ft

Fa

Fr

Tf

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 48

R

D


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

265

0.15

2

1.440575

689.4132

257.6872

376.0769

27.67777

1

65535

265

0.15

1

1.732035

481.3859

108.6272

196.3489

38.33675

1

65535

265

0.25

2

2.972375

889.7312

341.015

436.8659

21.77353

1

65535

169.434

0.15

1

4.023594

412.8885

152.2218

231.225

49.91373

1

65535

261.1212

0.15

1

1.823688

475.9385

108.9812

196.2864

38.67171

1

65535

265

0.25

2

2.972375

889.7312

341.015

436.8659

21.77353

1

65535

264.9902

0.25

2

2.972568

889.7437

341.0119

436.8632

21.774

1

65535

264.9983

0.241889

1.927966

2.753395

851.9404

321.1578

419.3945

22.49811

1

0.02106

231.6619

0.150082

1

2.523533

442.0419

115.6496

199.9704

41.49104

1

0.017777

256.2502

0.157696

1.086001

1.906649

498.2758

125.4584

216.8083

36.74274

1

0.016073

265

0.241178

1.922463

2.735899

848.8689

319.5894

417.9773

22.55941

1

0.015979

255.922

0.160701

1.078365

1.924486

500.1209

127.8311

218.4533

36.56701

1

0.01586

264.9996

0.246009

1.921864

2.81867

861.1237

322.9102

422.1392

22.35356

1

0.015844

185.7026

0.15

1

3.628576

414.8737

139.6661

219.9264

47.28547

1

0.015625

177.0656

0.15

1

3.838041

413.3254

146.0696

225.6509

48.63301

1

0.015374

265

0.197867

1.329979

1.873269

621.6293

191.4162

291.9741

29.43819

1

0.014759

264.961

0.161601

1.172733

1.681103

531.3475

137.4695

234.1206

34.35057

1

0.014673

218.6432

0.150107

1

2.834868

431.2042

120.7764

203.8858

42.92803

1

0.014661

264.9999

0.187048

1.311578

1.792259

598.7628

178.8963

278.8227

30.42464

1

0.014372

264.8573

0.241097

1.982864

2.795903

862.7214

331.3484

426.7589

22.24438

1

0.014369

174.3167

0.15

1

3.904826

413.0673

148.2322

227.6029

49.08405

1

0.014283

265

0.244146

1.891249

2.756463

849.2464

315.7876

416.0992

22.60344

1

0.014065

265

0.198594

1.349524

1.881922

627.2379

194.4536

295.5071

29.18677

1

0.014015

225.3303

0.150002

1

2.674819

436.3252

117.9001

201.6127

42.18555

1

0.013776

185.282

0.15

1

3.638763

414.7724

139.9641

220.1908

47.34876

1

0.013629

186.0766

0.15

1

3.619519

414.966

139.4022

219.6925

47.22937

1

0.013486

264.9959

0.240879

1.953051

2.760167

855.148

325.316

422.2024

22.40593

1

189.9865

0.15

1

3.524895

416.0567

136.7103

217.3171

46.65346

1

0.013037

264.9613

0.150352

1.247266

1.609129

533.0857

133.4604

236.7962

34.52022

1

0.01294

264.5977

0.171423

1.103768

1.747466

529.6774

141.0782

232.7329

34.4049

1

0.01288

232.9983

0.150418

1

2.491644

443.7511

115.4963

199.9651

41.30708

1

0.012815

265

0.246117

1.759588

2.664722

823.9152

293.6823

399.6592

23.29502

1

0.012699

208.8492

0.150014

1

3.070018

424.5623

125.443

207.6562

44.11734

1

0.01264

Table 3.3 consists optimized

results for 35 HRC process parameters in first three

columns (vc,f,d) and latter columns contains optimized results for objective functions (Ra VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 49


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

,Ft ,Fa ,Fr ,Tf ).Out of the 1000 chromosome solutions few best chromosomes

are listed in the above table. 3.5.6 Plots for NSGA II results (35 HRC)

Fig. 3.4 Rank and Pareto for Ra(35HRC)

Fig. 3.5 Pareto-front for Tf (35HRC)

Fig. 3.6 Average distance between consecutive generations (35HRC) Table.3.4 Results of NSGA II family of best solution for AISI 4340 45 HRC Steel Vc

f

d

Ra

Ft

Fa

Fr

Tf

R

D

175

0.2500

2

5.4703

1.1135e+03

576.9927

494.4395

8.9403

1

65535

135.9606

0.1500

1

3.9895

519.5461

331.3059

282.9869

28.2786

1

65535

130.8263

0.1500

1

4.0725

524.2045

331.6248

282.5873

30.0149

1

65535

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 50


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

175

0.1500

1

3.5798

504.5262

349.7478

312.2032

19.1319

1

65535

175

0.2500

2

5.4703

1.1135e+03

576.9927

494.4395

8.9403

1

65535

171.7381

0.1500

1

3.5991

504.4006

346.7949

307.9906

19.6973

1

65535

170.6747

0.1512

1

3.6100

505.0240

346.2216

307.2963

19.8126

1

0.0195

175.0000

0.2487

1.9979

5.4454

1.1089e+03

574.4013

492.7024

8.9698

1

0.0172

160.3800

0.1500

1.0066

3.6897

508.6927

339.0356

296.4326

21.7907

1

0.0168

174.9624

0.2458

1.9982

5.3980

1.1015e+03

569.8832

489.6478

9.0210

1

0.0160

175

0.2373

2

5.2635

1.0800e+03

556.7698

480.8408

9.1639

1

0.0156

171.7067

0.1500

1.0037

3.6009

505.9857

346.9955

308.2302

19.6489

1

0.0154

167.9552

0.1514

1.0066

3.6322

508.2227

344.5538

304.7724

20.2004

1

0.0154

168.5030

0.1500

1.0865

3.6577

541.3019

350.1705

311.2217

19.0637

1

0.0146

175

0.2306

2

5.1613

1.0631e+03

546.5998

474.0485

9.2885

1

0.0145

175

0.1857

1.4158

4.0521

716.6107

404.1508

368.3416

13.3300

1

0.0144

136.3303

0.1500

1

3.9838

519.2347

331.3076

283.0465

28.1599

1

0.0143

174.4113

0.1512

1.1067

3.6398

551.0508

356.2463

320.0333

17.7614

1

0.0142

174.8377

0.1511

1.0501

3.6089

526.7210

352.9048

316.2096

18.4048

1

0.0141

145.7619

0.1500

1

3.8500

512.3850

332.4684

285.9718

25.3898

1

0.0141

174.9083

0.1500

1

3.5803

504.5192

349.6613

312.0804

19.1474

1

0.0140

175

0.2393

2

5.2960

1.0853e+03

559.9760

482.9890

9.1264

1

0.0139

174.9987

0.2313

1.5566

4.6773

856.1857

468.7186

420.8659

11.1902

1

0.0137

174.9734

0.2064

1.1378

4.0238

611.7845

388.5914

358.3491

14.9399

1

0.0137

171.7164

0.1528

1.0761

3.6445

538.6585

352.6485

315.3194

18.4855

1

0.0136

175

0.1504

1.0131

3.5873

510.3848

350.6039

313.3299

18.9233

1

0.0136

175

0.2458

2

5.4006

1.1023e+03

570.2226

489.8743

9.0119

1

0.0135

175.0000

0.2028

1.1931

4.0322

634.4821

392.0288

360.3824

14.5323

1

0.0135

174.9971

0.2481

2

5.4377

1.1083e+03

573.8392

492.3111

8.9735

1

0.0133

175

0.1500

1.0286

3.5935

516.9707

351.4236

314.2841

18.7316

1

0.0132

168.5690

0.1500

1.0653

3.6481

532.3890

348.6946

309.5442

19.3321

1

0.0131

175

0.2295

1.0450

4.1928

593.8836

396.2472

370.1817

15.1377

1

0.0131

Table 3.4.consists optimized

results for 45 HRC process parameters in first three

columns (vc ,f, d) and latter columns contains optimized results for objective functions (Ra ,Ft , Fa , Fr , Tf ).Out of the 1000 chromosome solutions few best chromosomes are listed in the above table. VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 51


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

3.5.7 Plots for NSGA II results (45 HRC)

Fig. 3.7 Rank and Pareto for Ra (45HRC)

Fig. 3.8 Pareto-front for Tf (45HRC)

Fig. 3.9 Average distance between consecutive generations (45HRC) 3.6 Strength Pareto Evolutionary Algorithm (Type 2) Initialize population structure with fields in position cost fitness variables, dominance field and cumulative fitness. Then objectives are evaluated over random pop. Position with initial fitness, pop.cost. A fitness subset ‘archive’ for best individuals is initiated, first, all non-dominated population members are copied to the archive; any dominated individuals or duplicates are removed from the archive during this update operation. If the size of the updated archive exceeds a predefined limit, further archive members are deleted by a clustering technique which preserves the characteristics of the nonVISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 52


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

dominated front. Afterwards, fitness values are assigned to both archive and population members: Each individual i in the archive is assigned a strength value S(i) ∈ [0, 1), which at the same

time represents its fitness value F (i).and S(i) is the number of population members ‘j’

that are dominated by or equal to i with respect to the objective values, divided by the population size plus one. The fitness F (j) of an individual j in the population is calculated by summing the strength values S(i) of all archive members i that dominate or are equal to j, and adding one at the end. To avoid the situation that individuals dominated by the same archive members have identical fitness values, for each individual both dominating and dominated solutions are taken into account. Each individual i in the archive P t and the population Pt is assigned a strength value S(i), representing the number of solutions it dominates.

S  i   {j j  P.t  Pt  i  j} | R(i) 

S ( j)

j Pt  Pt , j i

On the basis of the S values, the raw fitness R(i) of an individual i is calculated: That is the raw fitness is determined by the strengths of its dominators in both archive and population. fitness is to be minimized here, i.e., R(i) =0 corresponds to a nondominated individual, while a high R(i) value means that i is dominated by many individuals. Additional density information is incorporated to discriminate between individuals having identical raw fitness values. The density estimation utilizes kth nearest neighbor method, where the density at any point is a (decreasing) function of the distance to the kth nearest data point. Here the inverse of the distance to the kth nearest neighbor is used for as density estimate. For each individual i the distances (in objective space) to all individuals j in archive and population VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 53


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

are calculated and stored in a list. After sorting the list in increasing order, the kth element gives the distance sought, denoted as σk. The kth nearest parameter is defined as square root of the sample size,

NN

D(i) 

and density distribution as 1 k  2

(3.13)

F (i)  R(i)  D(i)

(3.14)

In the denominator, two is added to ensure that its value is greater than zero and that D(i) < 1. Finally, adding D(i) to the raw fitness value R(i) of an individual i yields its fitness F (i).

Fig. 3.10 SPEA2 Algorithm[71] Table 3.5 SPEA2 Parameter Setting Population size Generation Archive size

Crossover probability Crossover constant Mutation probability Mutation constant

1000 100 300 0.7 0.1 0.1 0.2

3.6.1 Initialize Variables and Evalaute Objectives Initialize {pop.position, pop.cost, pop.S, pop.R, pop.σ, pop.D, pop.F} 1. For i : [1-N] VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 54


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

2. Pop.position :random(range,N) 3. Pop.cost=evalautefunction(pop.position) 4. //Initialize archive 5. Archive={ } 3.6.2 Tournament Selection The tournament selection is similar to that applied in NSGA II Binary tournament selection(archive,[archive.F],N) 6. I=random(N,2) 7. I1=I(1) 8. I2=I(2) 9. If F(i1)<F(i2) 10. P=pop[i1] 11. Else 12. P=pop[i2]

3.6.3 Genetic Operator Genetic operator applied is similar to that of NSGA II with slight variation in mutation technique. Crossover (p1,p2,crossover parameters) 13. Parametrs :(Υ,range) 14. α:random(-Υ,1+Υ,N) y1   * p1  (1   )* p 2

15. y 2   * p 2  (1   )* p1 16. Y1=min[max(y1,range)] 17. Y2=min[max(y2,range)] 18. Mutate (p3,mutation parameters, range) 19. Parameters : , range 20. Rmin=min(range)

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 55


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

21. Rmax=max(range) 22. dr=Rmax-Rmin 23. σ: *dr 24. y=p1+σ*random(Nm) 25. y=min(max(y,range)) 26. Main learning 27. Do until( max iteration IT) 28. P=[pop,archive] 29. //check for domination 30. [dom,p.S]=dominates (p[i]) 31. S=[p.S] 32. P[I].R=sum(S[dom]) 33. Q=[p.cost] 34. σ : Euclidiean distance[q] 35. σ : sort[σ] 36. p[i].σ =σ 37. p[i].σ[k]=p[i].σ[k]

p[i].D 

1 p[i]. [k  2]

38. 39. //Fitness Calculation

40. P[i].F=p[i].R+p[I].D 41. Fit =sum(find(p.R==0)) 42. P.F=[fit] 1. Archive =p[size[p.R]] 2. While[min(σ)==max(σ(k)&&k<size(σ))] 3. Pareto front=archive[archive.R==0] 4. [p1,p2]=binary tournament selection(archive,[archive.F],N) 5. Popc.cost=evalautefitness[child1,child2] 6. [child 1 child2]=crossover(p1,p2,crossoverparameters) 7. [p3]=binarytournamentselection(archive,[archive.F],N)

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 56


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

8. [child 3]=mutate(p3,mutation parameters,N) 9. Popm.cost=evalautefitness[child3] 10. pop=[popc.cost popm.cost] Table 3.6 Results of SPEA II family of best solution for AISI 4340 35 HRC Steel

Position

Cost

S

R

[259.15 0.17 1.07]

[1.88 514.11 136.22 225.95 35.51]

[9.00]

[0.00]

0.375 [0.32] [0.32]

[265.00 0.16 1.84]

[1.60 674.93 238.28 350.68 27.63]

[16.00] [0.00]

0.3563 [0.32] [0.32]

[265.00 0.24 1.81]

[2.57 813.37 295.99 398.32 23.38]

[17.00] [0.00]

0.2863 [0.34] [0.34]

[265.00 0.20 1.96]

[2.05 757.72 292.73 391.58 24.60]

[9.00]

[0.00]

0.305 [0.33] [0.33]

[264.60 0.15 1.94]

[1.51 683.95 250.23 366.11 27.68]

[6.00]

[0.00]

0.4035 [0.32] [0.32]

[264.83 0.17 1.74]

[1.64 661.12 224.65 335.27 28.00]

[26.00] [0.00]

0.3147 [0.33] [0.33]

[264.98 0.21 1.77]

[2.11 737.27 265.23 367.09 25.17]

[26.00] [0.00]

0.2558 [0.34] [0.34]

[265.00 0.19 1.79]

[1.89 706.15 252.72 356.92 26.15]

[20.00] [0.00]

0.3483 [0.33] [0.33]

[264.23 0.21 1.92]

[2.25 776.56 295.04 392.75 24.13]

[11.00] [0.00]

0.2511 [0.34] [0.34]

[264.85 0.22 1.47]

[2.09 689.39 226.10 330.51 26.92]

[34.00] [0.00]

0.3906 [0.32] [0.32]

[265.00 0.24 1.69]

[2.57 802.55 280.54 388.16 23.84]

[16.00] [0.00]

0.2971 [0.34] [0.34]

[264.99 0.18 1.88]

[1.81 711.36 262.25 367.37 26.09]

[17.00] [0.00]

0.3296 [0.33] [0.33]

[264.72 0.25 1.93]

[2.91 874.97 328.11 427.52 22.13]

[6.00]

[0.00]

0.2971 [0.34] [0.34]

[256.63 0.16 1.09]

[1.90 505.03 129.90 221.26 36.21]

[13.00] [0.00]

0.3225 [0.33] [0.33]

[264.97 0.19 1.45]

[1.83 636.05 200.63 303.73 28.72]

[39.00] [0.00]

0.3818 [0.32] [0.32]

[263.42 0.20 1.31]

[1.91 616.67 188.86 289.03 29.75]

[30.00] [0.00]

0.368 [0.32] [0.32]

[264.62 0.21 1.91]

[2.16 763.63 289.27 387.95 24.44]

[12.00] [0.00]

0.2511 [0.34] [0.34]

[264.57 0.23 1.41]

[2.16 695.37 225.90 331.84 26.94]

[30.00] [0.00]

0.3986 [0.32] [0.32]

[265.00 0.21 1.97]

[2.28 789.31 306.06 402.04 23.74]

[8.00]

0.2477 [0.34] [0.34]

[0.00]

sigmaK D

F

Table 3.6 has results of optimized solution obtained from SPEA2 algorithm for 35HRC position matrix represents the process parameters (vc,f,d), cost matrix represents optimized solution for objectives (Ra ,Ft ,Fa ,Fr ,Tf ). and the remaining cells correspond to SPAE2 parameters. Out of the 300 solutions in archive only few are listed in the above table. 3.6.4 Plots for SPEA 2 results( 35 HRC)

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 57


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig. 3.12 Pareto-front for Tf (35HRC)

Fig. 3.11 Rank and Pareto for Ra (35HRC)

Fig. 3.13 Average distance between consecutive generations (35HRC) Table 3.7 Results of SPEA II family of best solution for AISI 4340 45 HRC Steel

Position

Cost

S

R

sigmaK D

[175.00 0.22 1.90]

[4.92 997.97 515.49 452.74 9.82]

[14.00] [0.00]

0.238 [0.35]

[0.35]

[174.72 0.23 1.54]

[4.67 849.67 466.88 419.61 11.30]

[10.00] [0.00]

0.271 [0.34]

[0.34]

[174.92 0.18 1.34]

[3.93 674.57 389.76 355.44 14.19]

[29.00] [0.00]

0.427 [0.31]

[0.31]

[174.83 0.23 1.70]

[4.80 920.84 489.93 435.20 10.54]

[14.00] [0.00]

0.247 [0.35]

[0.35]

[174.85 0.18 1.55]

[4.14 773.13 418.81 380.47 12.54]

[21.00] [0.00]

0.299 [0.33]

[0.33]

[175.00 0.16 1.33]

[3.83 656.51 380.43 346.35 14.81]

[26.00] [0.00]

0.482 [0.30]

[0.30]

[174.99 0.19 1.70]

[4.37 852.70 448.58 404.23 11.40]

[21.00] [0.00]

0.253 [0.34]

[0.34]

[175.00 0.19 1.44]

[4.14 737.76 413.56 376.50 12.91]

[24.00] [0.00]

0.373 [0.32]

[0.32]

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 58

F


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

[174.99 0.22 1.47]

[4.45 793.05 442.97 401.34 11.96]

[11.00] [0.00]

0.278 [0.34]

[0.34]

[175.00 0.21 1.84]

[4.64 934.49 483.32 430.12 10.45]

[20.00] [0.00]

0.238 [0.35]

[0.35]

[175.00 0.22 1.51]

[4.54 819.43 453.20 409.14 11.62]

[13.00] [0.00]

0.26 [0.34]

[0.34]

[173.87 0.19 1.34]

[4.06 693.36 401.38 366.13 13.76]

[26.00] [0.00]

0.435 [0.31]

[0.31]

[175.00 0.21 1.40]

[4.23 738.85 419.98 382.61 12.77]

[23.00] [0.00]

0.348 [0.33]

[0.33]

[175.00 0.24 1.29]

[4.51 735.25 436.39 400.33 12.70]

[7.00]

[0.00]

0.346 [0.33]

[0.33]

[168.21 0.16 1.02]

[3.67 519.42 348.53 310.79 19.45]

[15.00] [0.00]

1.069 [0.33]

[0.33]

[175.00 0.24 1.19]

[4.48 688.80 426.53 395.02 13.40]

[8.00]

[0.00]

0.425 [0.31]

[0.31]

[174.95 0.23 1.26]

[4.33 698.11 419.46 385.62 13.28]

[16.00] [0.00]

0.347 [0.33]

[0.33]

[175.00 0.22 1.70]

[4.68 901.42 478.44 426.62 10.73]

[18.00] [0.00]

0.266 [0.34]

[0.34]

[175.00 0.24 1.82]

[5.07 1000.50 524.52 459.35 9.80]

[11.00] [0.00]

0.353 [0.32]

[0.32]

Table 3.7 has results of optimized solution obtained from SPEA2 algorithm for 45HRC position matrix represents the process parameters, (vc, f, d) cost matrix represents optimized solution for objectives (Ra ,Ft ,Fa , Fr , Tf ) and the remaining cells correspond to SPAE2 parameters Out of the 300 solutions in archive only few are listed in the above table 3.6.5 Plots for SPEA 2 results( 45 HRC)

Fig. 3.14 Rank and Pareto for Ra (45HRC)

Fig. 3.15 Pareto-front for Tf (45HRC)

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 59


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig. 3.16 Average distance between consecutive generations (45HRC)

3.7 Swarm Intelligence Swarm systems are based on behavior of school of birds, insects, fireflies where a flocks of birds twisting ,V-shaped structure of migrating geese, winter birds hunting for food, the synchronized flashing of fireflies are tried to imitating. The well-choreographed collective behavior without any leader is adopted to search for optimal solutions. For instance Ants living in colony, their behavior is driven by the goal of colony survival instead of individual survival, while searching for food ants initially explores surrounding nests. In random manner A similar behavior is observed with flocks of birds where a leader keeps guiding the flock to updated food location. 3.8

Mathematical Formulation of PSO Algorithm

Assuming a swarm ‘S’ of N(n) particle moves through dimension ‘D” in search space RD Let ‘f’ be objectives of our optimization problem the function ‘f’ is defined over the discretized space as f : R D  R over the space ‘R’. The definition of swarm can be condensed to S   St tN  X t ,Vt , Lt , Gt tN  [ X 0 ,V0 , L0 , G0  ,  X1 ,V1 , L1 , G1  ...............  X n , Vn , Ln , Gn ] o o

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 60


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Where X t  X tn,d 1  n  Nt  & 1  d  D 

(dth coordinate of the velocity of particle ‘n’ after time step ‘t’) Vt  Vt n,d 1  n  Nt  & 1  d  D 

(dth coordinate of the velocity of particle ‘n’ after time step ‘t’) Lt  Lnt ,d 1  n  Nt  & 1  d  D 

(dth coordinate of local best (attraction) of particle ‘n’ after time step ‘t’) Gt  Gtn,d 1  n  Nt  & 1  d  D 

(dth coordinate of global best (attraction) of particle ‘n’ after time step ‘t’) n 1 N ,d l ,d Furthermore Gtn ,d = Gt if (n<N) and satisfying initial and final conditions Gt  Gt 1

with given distribution for initial position and velocity (x0,v0) Initial grid index G0I is determined by minimum argument of function.

G0l  arg min1  n  N{ f ( X 0n )}atl0  x0

St 1   X t 1 ,Vt 1 , Lt 1 , Gt 1 which is determined by movement equation.

vtn,1d  vtn,1d  C1.r1n,d Lnt ,d  X tn,d  C2 .S1n,d Gtn,d  X tn,d

(3.15)

Where  C1 & C2  control the influence of personal best of particle and the common knowledge of swarm also known as acceleration co-efficient.

Stn.d & rtn,d are randomness which are drawn uniformly at random [0,1]

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 61


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

If a particle velocity component exceed a certain interval [-Vmax, Vmax]. It is set back to interval found. The movement equation is altered by clamping inertia.

vtn,d  1.vtn,d  C1.rt n,d Lnt ,d  X tn,d  C2 .Stn,d Gtn,d  X tn,d

and position is altered as

X tn,1d  X tn,d  vtn,1d   0,1   0.72 C1  C2  1.49

3.8.1

Typical Initialization strategy

1) Random initiation of position, velocity 2) Initiate velocity with zero matrix movement v n,d   .vn,d  C1rand (0,1)  Ln,d  X n,d   C2 rand (0,1) G n,d  X n,d 

3) Update position along with the personal and global best X n,d  X n,d  v n,d

If f  X n   L  Ln  then Ln  X n

 

f X n  L  G  then G  X n

3.8.2 Topologies of PSO To better predict social learning process the global best particle (Gd) is replaced by the local guide particle (Ld) Topology is typically represented as graph whose nodes are particles and edges connect neighboring particles. The edge connections between any two grid points n1 & n2 as a particle for its own local guide, the edge connections are determined by minimum argument of function Pt n  Euclidean(arg min f ( x)

 n ', n

where

x  [ Lt

| n'  N (n) |]

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 62


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

 n ', n

Lt

 is local guide attraction of particle ‘n’ at the time step ‘t’ when particle ‘n’

makes its move.  n ', n

L

  n ' n ',n  Lt    n ' L If n’  n  L t 1 

Attempts to form neighborhood topology depend on Inclined distance of particle in search space. In general neighborhood topology is choose independent of particle position in search space. 3.8.3 Definitions of Swarm topology For a Swarm, definition of topology and its potential drives solution space, leader of swarm both locally and global depends on the topology, though topology remains static the co-ordinates of particle attractors monitors the distribution of swarm. In general the global and local attractor is determined by fitness augments i.e.

G0l  arg min1  n  N{ f ( X 0n )}atl0  x0 Lnt  arg min[ X tn1 , Lnt ] t ,d For a given swarm S in stochastic process  X t ,Vt , Lt , Gt  the potential function n in

dimension d is determines the swarm fitness level at interval time step t. It is determined by global best and personal best of swam at all interval.

nt ,d 

n 1

 ( | vtn ,d |  | Gtn,1d  X tn ,d |)  '

n' 1

'

n N

 ( | v

n' , d t 1

'

|  | Gtn,1d  X tn1,d |)

(3.16)

n'  n 1

3.8.4 Convergence criteria It is important to control swarm topology for determining the desired solution space which depends on the movement constants applied on swarm so for a swarm to converge

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 63


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

the inertia damping co –efficient which keeps the swarm in bound and the acceleration co–efficient keep drives the co-ordinates of local and global attractors. 3.8.5 Criteria for Inertia clamping and acceleration co-efficeint At the topological development both global and local attractors are bound to same value

X t1,1  X 01,1  v1,1s For consecutive iterations t

X t1,1  X 01,1   v1,1 s t 0

t 1

 X 01,1  v0   s t 0

Lt s  X

1,1 t

 Lt s  [ X

1,1 0

v

t 1



s

]

0 t 0

1 ]  X 01,1  v0 [ 1 

So for swarm to be in bounds the inertia damping co-efficient should be between [0<  <1]. Now for acceleration co-efficient the movement equation is analyzed

vtn,1d  1.vtn,d  C1.rt n ,d  Lnt ,d  X tn ,d   C2 .Stn,1d  Gtn ,d  X tn ,d  X tn,1d  X tn,d  vtn,1d [ X tn,1d , vtn ,d ]tN The consecutive velocities cab be relates as follows

vtn,1d  X tn,1d  X tn ,d vtn ,d  X tn ,d  X tn,1d vtn,1d  vtn ,d  [ X tn,1d  X tn,1d ] and position can be re written as

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 64


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

X tn,1d  [1    (C1.rtn1,d  C2 .Stn,1d ) X tn,d   X tn,1d  C1.rtn,d Lnt ,d  C2 .Stn,1d Gtn,d ] Now to calculate the error in position between the expected and attained the expentancy operator is applied on position vector.

E[ X tn,1d ]  E[ X tn,d ][1    0.5*(C1  C2 )]   E[ X tn,1d ]  0.5*[C1.Lnt ,d  C2 .Gtn,d ]

(3.17)

Assuming the expectance to be  , now for a particle in swarm it is expecte that the calculated and expected positions are equal so for optimum goal attainment the expectancy is equal to zero which converts the position expectancy equation to

 2  [1    0.5*(C1  C2 )]   =0 Finding roots of expectancy

1,2  0.5*([1    0.5*(C1  C)]  [1    0.5*(C1  C2 )]2  4 Now for expectancy is always a positive real

[1    0.5*(C1  C2 )]2  4.  0

(3.18)

We arrive at optimal co-efficient criteria for better convergence. 0  (C1  C2 )  4(1   )

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 65


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

3.9 PSO Algorithm

Fig.3.17 PSO Algorithm[74] Table 3.8 PSO Setting MOPSO Definition Stopping/convergence

100

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 66


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Total particle population Max no of repository elements Inertia weight ( w) Inertia weight damping rate (w.damp) Personal learning co efficient (a1) Global learning co-efficient (a2) No of grid in each dimension Inflation rate ( á˝°) Leader selection pressure ( ) Deletion selection pressure ( ) Mutation rate (mu)

1000 500 0.5 0.99 1 2 7 0.1 2 2 0.1

3.9.1 Initialize population and Evaluate fitness Initialize population structure with fields Particle. Evaluate fitness over particle position and determine the dominance levels of each particle by dominates function //Initialize particle structure {Particle.Position,

Particle.Velocity,

Particle.Cost,

Particle.Best

Position,

Particle.Bestcost, Particle.Is Dominated, Particle.Grid Index, Particle.Grid Subindex} //Evaluate particle.position and cost 1. For i :[1-N] 2. Pop[i].position=random[Range,N] 3. Pop[i].velocity : zeros[N] 4. Pop[i].cost :evaluate (pop[i].position) 5. //update personal best 6. Exchange pop[i].best position with pop[i].position 7. Exchange popi[i].best cost with pop[i].cost Initiate repository element which is subset of all particles with best position and cost then select leader for swarm at every iteration through select leader function and i=update the particle structure for current leader swarm. //determine domination level VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 67


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Pop : domination(pop,N) 1. For i:[1-N] 2. For j:[i+1-N] 3. If dominates (pop[i], pop[j]) 4. True(pop[j] Is dominated) 5. Else if dominates(pop[j],pop[i]) 6. True(pop[i] Is dominated) 7. b=dominates(pop[i],pop[j]) 8. b= all(x<y)&&any(x<y) 9. //Initiate repository element 10. Rep=pop[~dominated pop) Apply mutation operator on the updated particle structure then calculate the dominance level for current structure. Create neighborhood for swarm by initiating grid topology for swarm. Update the swarm in the repository element with current dominance level. 3.9.2 Create Grid Index Now the topology is built for swarm which was initialized through particle structure topology is static and remains unchanged at every generation. Naumann /Grid based topology is built which utilizes Euclidean co-ordinates based on position of each swarm //create grid index Grid=create grid(rep,ngrid,Îą) 1. P=[pop.cost] 2. Rmin=min(p,[],2) 3. Rmax=max(p,[],2) 4. dr=Rmax-Rmin

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 68


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Rmin  Rmin   * dr 5. Rmax  Rmin   * dr 6. //Initiate grid index 7. Grid.L : [] Grid.H :[] 8. Object : size[p,1] 9. For j: [1-object] 10. p.object[equal spacing in R] 11. Grid[j].L =[-∞, p.object] 12. Grid[j].H=[p.object, ∞] 13. For i :[1-size(Rep)] 14. Rep[i]=find Grid index (Rep[i],Grid) 15. Obj=size(particle.cost) 16. Grid size=size(Grid.L) 17. Particle.Grid sub Index :zeros(p.object) 18. For i : [1-obj] 19. Particle.Grid sub Index[i]=find(particle.cost[i]<grid] 20. Particle.Grid Index=N*Grid size*particle grid index+ particle grid sub index 21. //Initiate repository element 22. Rep=pop[~dominated pop)

3.9.3 Select Leader After building topology each particle in repository is recognized with its position and velocity co–ordinates the identity of each particle is recognized through this co –ordinate Swarm is lead most fittest particle/particles At each generation the swarm changes its leader according to the swarm velocities and position evaluated through swarm movement equation. //select leader Select leader(rep, )

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 69


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

1. //Grid Index of all repository 2. I=[rep.Grid Index] 3. //occupied cells 4. C=select unique cells(I) 5. Q=find(I==C) 6. //selection probability 7. p=exp(- *N)

P

p

p 8. 9. s=select(P>random(0,1)) 10. //select cell 11. Sc=unique (s) 12. Find(I==s) 13. Leader=Rep(sc) 14. Mutation(pop,pm,Range) 15. If pm<random[0,1] 16. Pop[i]=R+dr*R 17. dr=pm*(Rmax-Rmin) 3.9.5 Delete extra elements Excess particles in the repository are either deleted or replaced by better fit particles in each generation if repository exceed its size then the interia damping factor reduces the velocity of swarm resulting in poor convergence hence at aevery generation repository is checked for its size. //Delete extra elements Delete Rep member(Rep,Υ) [1] Grid Index =(Rep.Grid Index) [2] Deletion=select leader (Rep,Υ)

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 70


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

[3] Deletion={ } 3.9.6 Swarm Movement Now that the swarm and its topology is built the swarm is allowed to move over the topology with governing movement equation followed by a slight mutation in fit particles which would accelerate motion of swarm by randomly changing the velocity and position coordinates. Main Learning Do until max (IT) For i : [1-N] //Select Leader 1. leader=Select leader(rep, )

pop  i  .Velocity  w * pop  i  .Velocity c1* rand VarSize  . *  pop  i  .Best.Position  pop  i  .Position   c 2* rand VarSize  . *  leader.Position  pop  i  .Position  pop  i  .Position  pop  i  .Position  pop  i  .Velocity pop  i  .Position  max  pop  i  .Position, VarMin  pop  i  .Position  min  pop  i  .Position, VarMax  pop  i  .Cost  Evaluate  pop  i  .Position  2. Newpop=mutatute(pop,pm,Range) 3. newpop.cost=evaluate(newpop.position) 4. Determine domination(Rep) 5. If dominates(New.pop.position, pop.position) 6. True(Is dominated pop.position) 7. Else if dominates(pop.position, New.pop.position) 8. True(Is dominated New.pop.position) 9. Grid=Create Grid(Rep, grid size,α) 10. Check if resize>maxrep

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 71


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

11. Rep=delete rep members (Rep,Î¥). Table 3.9 MOPSO family of optimal solutions for 35 HRC AISI 4340 steel Grid Position

Velocity

[265.00 0.16

Cost

Best Position

Best Cost

[1.55 702.05 266.48 379.44

[263.49 0.22

[2.01 595.25 181.77 277.57

2.00]

[1.62 -0.06 1.07]

26.93]

1.01]

31.89]

[212.34 0.15

[-43.26 -0.04 -

[2.99 426.70 123.66 206.18

[255.60 0.19

[2.03 659.89 217.50 320.77

1.00]

0.97]

43.69]

1.59]

28.39]

[193.76 0.15

[137.29 -0.02 -

[3.43 417.33 134.23 215.15

[193.76 0.15

[3.43 417.33 134.23 215.15

1.00]

0.35]

46.12]

1.00]

46.12]

[174.19 0.15

[-85.70 -0.01 -

[3.91 413.06 148.34 227.70

[174.19 0.15

[3.91 413.06 148.34 227.70

1.00]

0.24]

[265.00 0.15

49.11]

1.00]

49.11]

[1.45 652.07 221.19 340.52

[265.00 0.15

[1.45 652.07 221.19 340.52

1.82]

[7.93 -0.01 0.20]

28.93]

1.82]

28.93]

[174.72 0.15

[-32.71 -0.00 -

[3.89 413.10 147.91 227.31

[207.43 0.15

[3.10 427.11 127.30 209.77

1.00]

0.01]

49.02]

1.01]

44.08]

[215.30 0.15

[-30.50 -0.03 -

[2.92 428.66 122.23 205.01

[215.30 0.15

[2.92 428.66 122.23 205.01

1.00]

1.23]

[211.21 0.15

43.34]

1.00]

43.34]

[3.01 425.98 124.22 206.64

[211.21 0.15

[3.01 425.98 124.22 206.64

1.00]

[-1.61 -0.04 -0.73]

43.83]

1.00]

43.83]

[194.42 0.15

[-14.35 -0.05 -

[3.42 417.57 133.80 214.78

[194.42 0.15

[3.42 417.57 133.80 214.78

1.00]

0.63]

[265.00 0.20

46.02]

1.00]

46.02]

[1.95 694.04 238.22 342.10

[265.00 0.20

[1.95 694.04 238.22 342.10

1.66]

[37.22 -0.00 0.20]

26.52]

1.66]

26.52]

[171.25 0.15

[-32.00 -0.03 -

[3.98 412.91 150.71 229.85

[203.25 0.18

[3.13 673.17 231.63 322.24

1.00]

0.62]

[265.00 0.15

49.60]

1.60]

33.15]

[1.66 509.07 120.80 217.67

[265.00 0.15

[1.66 509.07 120.80 217.67

1.13]

[32.08 -0.01 0.13]

36.15]

1.13]

36.15]

[265.00 0.15

[26.00 -0.05 -

[1.49 603.83 180.32 296.71

[265.00 0.15

[1.49 603.83 180.32 296.71

1.59]

0.03]

[249.95 0.16 1.00]

[8.34 -0.09 -0.99]

[265.00 0.16 2.00]

[32.06 0.00 1.15]

[265.00 0.16 1.55]

[2.84 -0.08 0.06]

[265.00 0.15 1.62]

[1.51 -0.02 0.03]

[265.00 0.20

30.84]

1.59]

30.84]

[2.10 480.93 125.34 211.22

[249.95 0.16

[2.10 480.93 125.34 211.22

37.96]

1.00]

37.96]

[1.50 696.28 262.59 377.89

[265.00 0.16

[1.50 696.28 262.59 377.89

27.25]

2.00]

27.25]

[1.57 608.71 184.51 296.17

[265.00 0.24

[2.38 749.62 248.99 359.27

30.28]

1.49]

25.43]

[1.48 610.04 185.18 302.22

[265.00 0.15

[1.49 603.83 180.32 296.71

30.57]

1.59]

30.84]

[1.96 698.30 241.07 344.82

[265.00 0.20

[1.96 698.30 241.07 344.82

1.68]

[10.04 0.01 0.04]

26.38]

1.68]

26.38]

[265.00 0.17

[125.17 -0.00 -

[1.74 530.80 141.76 233.24

[265.00 0.17

[1.74 530.80 141.76 233.24

1.10]

0.30]

34.33]

1.10]

34.33]

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 72

Index

GridSubIndex

10668

[2,6,6,7,3]

33631

[6,2,2,2,7]

40282

[7,2,3,3,7]

46844

[8,2,3,3,8]

9850

[2,5,5,6,4]

46844

[8,2,3,3,8]

27070

[5,2,2,2,7]

33631

[6,2,2,2,7]

40282

[7,2,3,3,7]

17139

[3,6,5,6,3]

46844

[8,2,3,3,8]

14684

[3,3,2,3,5]

9031

[2,4,4,5,4]

21236

[4,3,2,2,5]

10659

[2,6,6,6,3]

9031

[2,4,4,5,4]

9031

[2,4,4,5,4]

17139

[3,6,5,6,3]

15494

[3,4,3,3,5]


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

[265.00 0.16

[1.61 708.49 270.60 381.20

[265.00 0.15

[1.49 603.83 180.32 296.71

2.00]

[0.68 0.01 0.47]

26.59]

1.59]

30.84]

[265.00 0.24

[28.76 -0.01 -

[2.70 846.43 323.67 419.78

[265.00 0.24

[2.70 846.43 323.67 419.78

1.96]

0.04]

[265.00 0.15 2.00]

[3.61 0.00 0.67]

[250.88 0.17 1.00]

[-7.54 -0.02 -0.11]

22.55]

1.96]

22.55]

[1.48 694.34 261.24 377.37

[265.00 0.15

[1.48 694.34 261.24 377.37

27.37]

2.00]

27.37]

[2.09 486.86 128.60 214.65

[250.88 0.17

[2.09 486.86 128.60 214.65

37.52]

1.00]

37.52]

10668

[2,6,6,7,3]

31169

[5,7,7,8,2]

10659

[2,6,6,6,3]

21326

[4,3,3,3,5]

Table 3.9 has results obtained from MOPSO for 35 HRC the position, Best position matrix represent process parameters (vc,f,d),. and Cost, Best Cost represents objective fitness (Ra ,Ft ,Fa ,Fr ,Tf ) and the latter columns are corresponding to grid index and grid sub index of PSO topology Out of the 500 repository elements few are listed in above table. 3.9.7 Plots for MOPSO results (35 HRC)

Fig 3.18 Pareto spread surface roughness and

Fig 3.19 3D surface plot of optimal Ra with

cutting force 35HRC

best position 35HRC

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 73


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig 3.20 3D surface plot of Tf with best

Fig 3.21. Depth of cut influence on cutting

position 35HRC

forces 35 HRC

Table 3.10 MOPSO family of optimal solutions for 35 HRC AISI 4340 steel Table 1 Position

Repository elements Solution for Population Velocity

[131.18 0.15 1.00]

[-38.59 -0.02 -0.73]

[145.95 0.15 1.00]

[-29.05 -0.08 -1.19]

[175.00 0.24 2.00]

[2.94 0.05 0.95]

[175.00 0.15 1.52]

[18.14 -0.00 0.06]

[175.00 0.19 2.00]

[2.26 0.01 0.10]

[157.18 0.15 1.00]

[-17.82 -0.04 -1.28]

[175.00 0.19 2.00]

[56.89 -0.05 0.12]

[175.00 0.24 1.68]

[4.45 -0.01 0.22]

[132.82 0.15 1.00]

[-53.49 -0.04 -0.47]

[149.38 0.15 1.00]

[-7.05 -0.03 -1.52]

[154.57 0.15 1.00]

[-20.43 -0.12 -0.45]

[162.55 0.15

Cost

Best Position

Best Cost

[4.07 523.87 331.58 282.59

[169.76 0.15

[3.91 730.05 393.60 357.11

29.89]

1.55]

14.41]

[3.85 512.27 332.51 286.06

[145.95 0.15

[3.85 512.27 332.51 286.06

25.34]

1.00]

25.34]

[5.24 1075.52 554.08 479.04

[172.92 0.18

[4.53 953.91 482.02 431.27

9.20]

2.00]

10.62]

[3.88 720.39 390.67 356.83

[175.00 0.15

[3.88 720.39 390.67 356.83

13.99]

1.52]

13.99]

[4.67 977.77 494.94 440.17

[175.00 0.19

[4.67 977.77 494.94 440.17

10.11]

2.00]

10.11]

[3.72 506.91 336.75 293.12

[157.18 0.15

[3.72 506.91 336.75 293.12

22.59]

1.00]

22.59]

[4.62 969.60 489.97 436.99

[136.93 0.15

[4.01 617.69 362.52 315.19

10.21]

1.26]

23.55]

[4.99 944.81 506.19 447.74

[175.00 0.24

[4.99 944.81 506.19 447.74

10.29]

1.68]

10.29]

[4.04 522.32 331.42 282.65

[132.82 0.15

[4.04 522.32 331.42 282.65

29.32]

1.00]

29.32]

[3.80 510.32 333.49 287.81

[149.38 0.15

[3.80 510.32 333.49 287.81

24.44]

1.00]

24.44]

[3.75 507.89 335.50 291.14

[154.57 0.15

[3.75 507.89 335.50 291.14

23.19]

1.00]

23.19]

Grid Index

GridSubIndex

20510

[4,2,2,2,8]

13948

[3,2,2,2,7]

45011

[7,8,7,7,2]

15501

[3,4,3,4,3]

31070

[5,7,6,6,2]

13947

[3,2,2,2,6]

31070

[5,7,6,6,2]

36911

[6,6,6,7,2]

20510

[4,2,2,2,8]

13947

[3,2,2,2,6]

13947

[3,2,2,2,6]

[3.67 505.40 339.86 297.86

[166.04 0.15

[3.64 504.79 342.25 301.40

1.00]

[-3.48 -0.02 -0.06]

21.45]

1.00]

20.75]

7395

[2,2,2,3,6]

[173.21 0.15

[0.62 -0.06 -0.09]

[3.64 549.52 354.83 317.92

[172.59 0.20

[4.02 635.55 389.86 357.26

8204

[2,3,3,3,5]

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 74


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

1.10] [175.00 0.19 1.42]

[4.58 0.01 0.22]

[175.00 0.23 1.57]

[1.33 -0.02 -0.43]

[175.00 0.23 1.61]

[0.30 0.01 0.40]

[175.00 0.20 1.21]

[0.75 -0.05 -0.42]

[175.00 0.24 2.00]

[1.67 0.02 1.18]

[175.00 0.15 1.10]

[2.22 -0.01 0.10]

[131.18 0.15 1.00]

[-38.59 -0.02 -0.73]

[145.95 0.15 1.00]

[-29.05 -0.08 -1.19]

[175.00 0.24 2.00]

[2.94 0.05 0.95]

18.04]

1.19]

14.92]

[4.07 721.13 405.90 369.87

[175.00 0.19

[4.07 721.13 405.90 369.87

13.24]

1.42]

13.24]

[4.70 864.69 471.99 423.23

[175.00 0.23

[4.70 864.69 471.99 423.23

11.10]

1.57]

11.10]

[4.70 876.32 474.30 424.45

[175.00 0.23

[4.70 876.32 474.30 424.45

10.97]

1.61]

10.97]

[4.04 639.51 392.88 360.94

[175.00 0.20

[4.04 639.51 392.88 360.94

14.45]

1.21]

14.45]

[5.28 1082.79 558.47 481.98

[172.40 0.25

[4.61 740.79 441.72 405.77

9.14]

1.25]

13.04]

[3.63 548.62 356.03 319.83

[175.00 0.15

[3.63 548.62 356.03 319.83

17.79]

1.10]

17.79]

[4.07 523.87 331.58 282.59

[169.76 0.15

[3.91 730.05 393.60 357.11

29.89]

1.55]

14.41]

[3.85 512.27 332.51 286.06

[145.95 0.15

[3.85 512.27 332.51 286.06

25.34]

1.00]

25.34]

[5.24 1075.52 554.08 479.04

[172.92 0.18

[4.53 953.91 482.02 431.27

9.20]

2.00]

10.62]

22143

[4,4,4,4,3]

36822

[6,6,5,6,3]

36822

[6,6,5,6,3]

14854

[3,3,4,4,4]

45020

[7,8,7,8,2]

8204

[2,3,3,3,5]

20510

[4,2,2,2,8]

13948

[3,2,2,2,7]

45011

[7,8,7,7,2]

Table 3.10 has results obtained for MOPSO for 45 HRC the position, Best position matrix represent process parameters (vc,f,d),. and Cost, Best Cost represents objective fitness (Ra ,Ft ,Fa ,Fr ,Tf ) and the latter columns are corresponding grid index and gris sub index of PSO topology Out of the 500 repository elements few are listed in above table.

3.9.8 Plots for MOPSO results (45 HRC)

Fig 3.22 Pareto spread surface roughness Fig 3.23 and cutting force 45HRC

3D surface plot of optimal Ra

with best position 45HRC

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 75


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.3.24 3D surface plot of Tf with best

Fig.3.25 Depth of cut influence on cutting

position 45 HRC

forces 45 HRC

3.10 Comparison between EA and SI technique

Fig.3.26. Solution Spectrum for 35 hrc Fig.3.27. Solution Spectrum for 45 hrc

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 76


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.3.28 Solution Spectrum for 35 hrc Fig.3.29 Solution Spectrum for 45 hrc PSO

Fig 3.30. Solution Spectrum for 35 hrc Fig. 3.31 Solution Spectrum for 45 hrc

3.10.1 Comparison Based on Spectrum of solution space 1. The search exploration for solution in both EA and SI varies in demography of population size and density which can be observed from the spectrum distribution of solution space. 2. The spectrum of solution space reveals the demographic changes in solution space at each generation hence it is crucial to compare saturation levels in solution spectrum. 3. Solution Spectrum in NSGA II :

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 77


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

3.1 The solution spectrum distribution in NSGA II (as shown in Fig. (3.26-3.27)) has attained uniform amplitude with periodic crusts and troughs in all the objectives exhibiting quit a good saturation level in the demography in solution space for both hardness levels. 3.2 Solution Spectrum in SPEA 2 While in SPEA 2 the demography of solution space (as shown Fig. (3.30-3.31)) is quite different from the NSGA II as the solution space is built on niche Pareto sharing 3.3 The spectrum is pretty disturbed with low levels of saturation and crusts and troughs varying throughout the wavelength of the data. for both the hardness levels. 4.

Solution Spectrum in PSO

4.1 The trend in PSO shows moderate level of disturbance in spectrum of solution space (as shown Fig. (3.28-3.29)) but the level of saturation is appreciable when compared to the SPEA 2 and change in demography is not as amplifying as SPEA 2 while in comparison with the NSGA II it is inferior in terms of saturation levels. From the nature of spectrum in solution space conclusion can be condensed to as follows. NSGA II relatively better compared to PSO and SPAE 2, PSO is better compared to SPEA 2. 3.10.2 Comparison Based on Diversity in solution space. 1. The diversity of solution space is evaluated through the average Pareto spread in each generation, diversity measures exploration potential in search space. 2. Higher diversity in generations gives greater chances of solutions getting retained from varying locals of search space hence increasing the strength of solution. Table 3.11Diversity of Evolutionary Algorithm Evolutionary Algorithm NSGA II SPEA 2

Diversity [0.01-1] [0.27-0.37]

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 78


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

3 The diversity of NSGA II is superior when compared to SPEA 2 the range of search exploration between the generations in NSGA II varied from Euclidean spread of [0.01 to 1] (Fig.3.6 and Fig. 3.9) which shows that the exploration happened between two extremum and in mid generations solution spread curled towards mean solution and drifted away from the mean solution. 4 While in the SPEA 2 the Euclidean spread was in a short range of 0.27-0.37 (as shown in (Fig.3.26 and Fig. 3.27). with all the equal dominating and non-dominated solutions in the spread, the diversity oscillated in short range at each generation with most of the generations between average spread [0.3-0.35]. 5 The overall analysis between EA and PSO suggests that NSGA II performs well in terms of diverse solution while SPEA 2 and PSO performs well when the solutions spread is short In NSGA II more diverse solutions are preserved while in PSO and SPEA 2 neighborhood solutions are preserved. Further potentials of both EA and SI are explored with synergism with prediction models in chapter 5 where EA and PSO are utilized to enhance learning in prediction models.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 79


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

CHAPTER 4 PREDICTION MODELS USING INTELLIGENT LEARNING TECHNIQUES 4.1 Introduction This chapter illustrates the development of prediction modeling using intelligent learning techniques

on machining system. Learning algorithm with their

mathematical framework is extensively discussed and applied to existing machining system. In the first segment prediction model by applying neural network is developed for both steels. In the second segment adaptive learning techniques are developed and the third section developed models are analyzed over machining statistics. Further extensive statistical analysis is done between the experimental da ta and the prediction results for evaluating the accuracy among the developed models. The objective of this chapter is explained through the following work flow.

Fig.4.1 Workflow for Chapter 4 VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 80


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.2 Neural Network Neural network consists of nodes connected by direct links each link has numeric weight wij associated with it which determines the strength and configuration between links. A threshold function ‘activation function’ f(.) is applied to model which alters the topology of links, these connection between nodes forms layer pattern so called network architecture. Depending on the direction of propagation of weights in the link neural model is classified into feed forward and feed backward /recurrent network. 4.2.1 Feed forward Neural network Feed forward networks are arranged in layers such that each unit receives input only from the units in the preceding layer. The architect of the feed forward network has following composition. 

First layer is input layer receiver of data or input from the external stimuli, the incoming data is then sent to the next layer where the number of layers can be more than one.

Second layer consists of hidden layer in which the number of nodes depends on the complexity and non-linearity of data to be handled with weights defining connections between node and bias at each node A single hidden layer constituents a network activated by threshold /activation function which takes augment of weights and bias matrix from the net This augmented net is propagated to the subsequent layers depending number of hidden layers.

Data processed in hidden layers are routed to the output layer. This layer plays a role in determining the validity of data that are analyzed based on the existing limits in the activation function.

Neural network runs training examples through the net one at a time, adjusting the weights slightly after each example to reduce the error. Each cycle through the examples called epoch.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 81


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.2.2 Mathematical background of neural network Since the essential elements of neural network are discussed now the working matrix by which each layer learns is discussed through simple network.

Fig.4.2 Simple network Let X be input vector and Y be output layer vector with weight matrix mapping between the input and output layer. Then the neural network is characterized by the learning model n

Yi   Wx Xi  bi i 1

Where the input layer and output layers are defined by the vectors

X  {x1 , x2 , x3 ,......xl } Y  { y1 , y2 , y3 ,....... yl } Now for training and mapping between the input and output layer learning law which describes that prediction accuracy to increase the weights of nodes should be correlated to attain minimum error in predictor consequently to store a prototype

( xi , yi ) .The weights are altered by weight matrix

w  .yi x iT Where

is learning factor and is generally kept positive, elements of weight matrix w

starts from zero to a perfectly associative neuron weight. And inverse mapping is possible at any instance of learning stage by recalling the weight matrix as

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 82


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

L

W   Yi .Xi T i 1

W . X K  YK

K  1, 2...l

Now that the weight matrix is defined and one to one mapping between the input and output layer is established the next step is to minimize the error of mapping in weight matrix by applying gradient descent approach. 4.2.2.1 Gradient Descent Approach According to gradient descent the mean square E(w) associated gradient of expected error .The error gradient ( points in the direction in which E(W) will decrease at fastest possible rate w  k  1  w  k   (E)

α arbitrary constant similarly the lest mean square error for predicted output and weight is calculated as 1 ek [w(k), y(k)]  [| w(k).y(k) | w  k .y  k ] 2

To minimize this error the error gradient showed be headed in the direction where error is least ek  w  k  y  k   w  k 

1  y  k  .Tb  w  k   .y  k   y  k  2

w  k  1

From extremum of error gradient



Tb  w  k    1  1

0

we get the if

w(k).y(k)  0

if

w(k).y(k)  0

By fixed incrementing weights at every epoch we get

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 83


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

w  k  1  w  k  

 y  k   y  k  Tb  w  k   y  k  2

Now validating weights for desired output d(k), the augments w(k) and y(k)determines the rate at which the desired output is approached . ek  d  k   w  k  * y  k 

Evaluating least squared errors for the error function. 2 T e2k  d  k   w  k  .y  k   d  k   w  k  .y  k   

2

To minimize the least square error expectation operator E[ek2:]is applied over the squared gradient of error k 

e2k  2e k y  k  w

 

E k  2E  ek y  k   E(K )  2E  y  k  .yT  k  .w  k   d  k  .y  k 

By estimating the mean of the gradient the direction of least error propagation can be known K  2E  y  k  .yT  k  .w  k   d  k  .y  k 

the expectancy of gradient is expected to be zero and with few manipulations weight matrix which has least error propagation is obtained.as K  0 P  y  k  .yT  k  . Q  d  k  .y  k  W*  Q1.P

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 84


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.4.3 Multi-layer feed forward network structure Now that the weights matrix for a single layer is established an extension of weight matrix for consecutive layers are evaluated for the above multi-layer network. E k 

2 1 m   y j  k   d  k  2 j1

N

ET   E  k  k 1

Where ET is the expectation of all the layers similarly evaluating the gradients of expectation E  k  dy j

 yj  dj

s j   yi wil   j

The activation function or transfer function characterizes the input output relationship

y j  f j (s j ) Most common choice of activation function is sigmoidal function which satisfies the continuous differential function and continuous everywhere.

yj 

1  1  exp( s j )

1 n

1  exp[ ( wij yi(1)   j )] i 1

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 85


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Now for the above function the error propagation for every input to layer can be written by chain rule as E  k  s j

E  k  y j  y j s j

After few assumptions in the activation function and applying the above chain rule following error propagation in unit layer j can be expressed as

dyi 1 d ( )  y j (1  y j )  ds j ds j 1  exp( s j ) E (k )  ( y j  d j ) y j (1  y j ) s j Where j are the threshold value generally referred as bias and yj in the consecutive layers is determined by transfer function.so at each layer the targets changes with the weights which are expected to approach to desired matrix. Thus the error gradient at each layer with respect to the weights can be written by chain rule as E  k  w m

E  k  yl si   yl sl w ml

Summing up all errors in all the layers by chain rule we arrive at E  k  yi

 j

E  k  s j

w ij

There are two approaches to apply gradient descent method to the training method of a multi-layer feed forward neural network .The first is based on periodic updating and second is based on continuous updating. In both the cases the weights are repeatedly monitored either sequentially or randomly until the convergence criteria is satisfied ET E(k) E(k) E(k) E(k) T [ , , .... ] w w1 w 2 w 3 w m VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 86


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

With m donating the number of weights in the network and weights are updated only once every epoch after all the training patterns are evaluated the weights get updated by the generalized fixed increment/decrement rule. w new  w old  n

E  k  w

Where η is small constant referred to as learning rate. Wnew and wold are weights vector at epoch k+1 and k respectively. 4.2.3 Key notes form feed forward analysis In order to build accurate prediction model sufficient input and output vectors is necessary to be included in the network. A reasonably sufficient amount of exemplars is essential for a prediction model to work accurately. There is no hard rule for selecting number of nodes and layers it is purely a trail error based mapping technique. If the developed model is accurate enough for a given set of nodes and layers and satisfies the stopping criteria then the network hypothesis is acceptable. 4.2.4 Multi-layer Perceptron for Turning of AISI 4340 Steel With the above fundamentals a prediction model for machining system [1] is built for both the steels i.e., 35 HRC and 45HRC with process parameters as input vectors and cutting speed, forces and tool life as output vectors. In the Fig a schematic description of developed prediction model is explained in detail. This model is adopted for predicting both the steels.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 87


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.4.4 Feed forward neural network for AISI 4340 Hard turning Table 4.2.4 (a) Description of Neural network Neural Network Type Training function No of neurons in Hidden layer No of neurons in output

Feed forward neural network Levenberg-Marqaurdt 10

Weights in hidden layer Weights in output layer Training samples

5 30 [3×10] 50[5×10] [700 3]

Testing samples

[150 3]

Validation samples Transfer function Training performance

[150 3] Tan-sigmoid function 2.861*10-4 (35HRC) NN

Testing performance Validation performance

4.147*10-4 (35HRC) NN 1.777*10-4 (35HRC) NN

7.930*104 (45HRC) NN 0.00106(45HRC) 0.00118(45HRC) NN

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 88


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Table 4.2.4 (b) Calibrated weights and bias above Neural Network Calibrated weights and bias for 35 HRC Steel Hidden layer Definition(sij)

Output layer Definition(sij)

Bias(i)

W1(vc)

W2(f)

W3(d)

bias(i)

W1(Ra)

W2(Ft)

W3(Fa)

W4(Fr)

W5(Tf)

0.660 -0.660

-0.135

-0.086

0.005

4.388

-3.260

-2.039

-8.270

-6.538

-4.691

0.0024

-0.034

0.1812

-0.80

1.545

-3.425

3.573

2.873

3.957

0.661

0.002

0.148

-0.099

11.57

4.809

-0.977

1.910

-5.394

0.541

0.901

-0.710 0.002

0.535 -0.180

0.138 -0.119

10.89 12.74

0.429 6.030

0.058 0.239

-0.07 0.002

0.042 -3.204

-0.035 3.619

-0.677 0.167

-0.307

-0.020

0.295

1.690

0.597

0.981

-0.136

-0.025

-0.180

-0.441

0.334

0.086

0.489

0.539

-0.648

0.390

-0.115

-0.672

-0.134 -0.255 -0.465

-0.005 -0.0165 0.351

-0.07 0.2469 0.0912

-0.067 1.143 0.6855

-6.242 2.986 1.169

6.290 4.917 -1.4074

6.749 -0.703 0.850

10.459 0.274 -0.113

-0.99003 -1.376

Caliberated weigths and bias for 45 HRC Steel 2.991 0.541 1.972 -2.656 0.245 0.276 0.134 0.323 -2.418 2.11 1.853 1.524

0.746 0.635 0.974

0.176 -0.127 0.200

0.0013 -1.326 0.0059

-0.007 0.707 0.0128

-0.027 1.233 0.0091

-0.017 -1.894 -0.010

0.210

0.180

0.099

0.428

0.620

0.860

2.276

0.0923

0.0118

0.566

0.519

-2.88

-1.639

-0.557

0.261

0.676

-0.0322

-0.0670

-0.039

-0.100

0.615 -1.914

2.144 -0.230

1.797 -0.261

1.122 -2.013

0.728 0.198

-0.0492 -0.063

-0.088 0.0148

-0.0488 0.0102

-0.106 0.072

0.493 6.36 0.733

0.646

-0.628

0.776

-0.298

0.0085

-0.029

-0.188

-0.062

-0.051 -1.295

-0.0027 -1.324

-0.004 -1.956

-0.011 -1.324

-0.08 -0.01

3.857359 0.268918

1.379878 -0.332

2.528731 -0.4

4.2.4 Results of perceptron for 35 HRC Steel

Fig.4.5 Performance plot of Network

Fig.4.6 Training state of Network at each

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 89


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.4.7 Training error in Ra

Fig.4.8 Regression fit plot for Ra

Fig.4.9 Training error in Ft

Fig.4.10 Regression fit plot for Ft

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 90


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.4.11 Training error in Fa

Fig.4.12 Regression fit plot for Fa

Fig.4.13 Training error in Fr Fig.4.14 Regression fit plot for Fr

Fig.4.15 Training error in Tf Fig.4.16 Regression fit plot for Tf

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 91


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.2.6 Results of perceptron for 45 HRC Steel

Fig.4.17 Performance plot of Network

Fig.4.18 Training state of Network at each epoch

Fig.4.19 Training error in Ra

Fig.4.20 Regression fit plot for Ra

Fig.4.21 Training error in Ft

Fig.4.22 Regression fit plot for Ft

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 92


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.4.23 Training error in Fa

Fig.4.25 Training error in Fr

Fig.4.27 Training error in Tf

Fig.4.24 Regression fit plot for Fa

Fig.4.26 Regression fit plot for Fr

Fig.4.28 Regression fit plot for Tf

The hypothesis of network for both the steels were quite accurate with the man square error approaching to order of 10 -3.Error gradients and mean error gradients in VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 93


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

learning stage is found to converge to minimum criteria. Error analysis for each target output is evaluated at learning stages i.e., training testing and validating and targets are fir to linear regression model with regression co–efficient approach to 1. 4.3 Adaptive Neuro-Fuzzy Interference System (ANFIS) A neuro-fuzzy interference system utilizes the human ability of recognizing pattern in modeling information either numeric or linguistic by employing fuzzy membership function and fuzzy if-then rules combined with neural network architect. ANFIS shares methodology of fuzzy sets and neural network for building learning model by interpreting the fuzzy system in terms of neural nodes with slight modification in the fuzzy rule base which is replaced by weights instead of linguistic rules. There are two ways of implying fuzzy neural systems. In the first method the fuzzy rules are modified with no change in the input and output membership functions. In the second method fuzzy neural systems with learning algorithms such as backpropagation or hybrid learning are applied to learn and adjust the membership function parameters. Different combinations of fuzzy neural systems are possible with varying input output membership functions. The applied adaptive neuro fuzzy interference system is explained through a two input-single output model utilizing Sugeno-type fuzzy system also known as Takagi-Sugeno-Kang type fuzzy system where rule base is replaced by neural network weights and output membership are defined by linear function instead of fuzzy linguistic model.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 94


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.4.29 ANFIS two input model Layer 1 Every node in this layer is defined by membership function where each node gives membership values after evaluating inputs over the membership function. The applied membership function can be linear or exponential with each input defined by desired subsets in membership function In most cases Gaussian membership function is applied

 A,i  exp[(

x  ci 2 ) ] 2ai

O1,i   A,i ( x),

i  1, 2..k

O1,i   B ,i ( y ),

i  1, 2..k

Layer 2 In this layer the node is fixed and takes the fuzzified value as input from the layer 1. The output of this node is the result of fuzzy multiplication of membership function which goes into the next node. Each node represents the firing strength of each rule in the second layer. The T-norm operator with and operation is applied to obtain the output. This layer is known as andecent layer

O2,i  wi   A,i ( x)* B,i ( y)

i  1, 2..k

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 95


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Layer 3 In this layer the normalized weight for each firing strength with respect to all the cumulative fire strength is calculated. The resulting weights is called normalized firing strength

O3,i  wi 

wi  wi

Layer 4 Every node in this layer is adaptive node known as consequent layer and gives output with node function defined as O4,i  wi fi

O4,i = wi ( pi x  qi y  ri ) Layer 5 Single node fixed node that calculates overall output of from consequent layer O5,i   wi fi

4.3.1 Hybrid learning in ANFIS The adaptive layers i.e., first and fourth layer contain parameters which can be modified at every iteration. The antecedent and consequent parameters can be updated through learning method. There are two paths of learning forward and backward path. In the forward path recursive least square method is used to alter consequent parameters. While in the backward path the antecedent parameters are changed through gradient descent method.at each iteration which is also called epochs. Forward Learning In the forward learning consequent parameters are adjusted f  w1 f1  w2 f 2  w1 ( p1 x  q1 y  r1 )  w2 ( p2 x  q2 y  r2 )  ( w1 x) p1  ( w1 y )q1  w1r1  ( w2 x) p2  ( w2 y )q2  w2 r2

When N training data are given as input vector the n the consequent function changes to VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 96


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

( w1 x)1 p1  ( w1 y )1 q1  ( w1r1 )1  ( w2 x)2 p2  ( w2 y )2 q2  ( w2 r2 )3  f1 ( w1 x)n p1  ( w1 y) n q1  ( w1r1 ) n  ( w2 x) n p2  ( w2 y ) n q2  ( w2 r2 ) n  f n The above equation is simplified and expressed in matrix form A  y

Where  is the vector M×1 M is the number of elements that the consequent parameters is set and A is the vector P* M where P is the number of N data training provided to the adaptive network and y is the output vector P*1 whose elements are N number of output data of an adaptive network. The optimum solution for  is defined as

 *  ( AT A)1. AT y T 1 Where AT is the inverse of A and if not singular, ( A A) is pseudo inverse of A by

using recursive LSE method then

i 1  i  Pi 1ai 1 ( yiT1  aiT1i ) Where a a row is vector of matrix A and Pi sometimes called a covariance matrix and is defined by

Pi  ( AT A)1 4.3.2 Back propogation Learning The parameters in Gaussian parameters are trained for minimizing error For a given adaptive network where the network consists of five layers and has total of N (L) node in layer L then the square error in the L layer to n data is 1  n  N the error at each node can be written as N (l )

En   d k  X kL,n k 1

Where dk is the k-th component of the vector of the desired output while X kL,n is k-th component of the vector of actual output generated by adaptive network with input VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 97


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

from the input vector n. The main aim of adaptive learning system is to reduce error that occurs .i.e.,

 L ,i 

En  2(di ,n  X iL,n )  iL,n

Applying chain rule for consecutive layers for error propagation we get l 1 N ( l 1) En En X m, p   l 1 l 1 X l ,i m 1 X m , n X m , n

With 0  l  L  1 internal node error is a cumulative node error in the layer l+1. For a specific node in adaptive layer the error rate corresponding to parameter

is given as

En E x  n  xS x 

Where S is the set of nodes containing the parameter

the error specific to this

parameters is given as

En En =  n 1  N

4.3.3 Fuzzy clustering Algorithms Fuzzy clustering algorithms are utilized to discretize the membership function into subsets in the input vectors so that each input vector is sub divided into topologies defined by densities of points in respective region. Building a fuzzy set requires following key points. Selection of inputs-outputs vectors choice of specific type of fuzzy interference system the no of mf’s function and subsets in mf’s Generating the antecedent and consequent rule a)

Choosing an appropriate family of parameters mf’s. To illustrate fuzzy

clustering a fuzzy set with following definition is declared. Let u(t), y(t), x(t) denote the input, output and state of a system S at time t

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 98


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

X (t  1)  f ( x(t ), u (t )) Y (t  1)  f ( x(t ), u (t )) f &g f : X *U  X G : X *U  Y

  [m1T , m2T ,.....mmT ]T b) Clustering approaches: A clustering approach can be applied to estimate the data distribution and resulting clustering which produces the membership function c) Clustering in mf’s In any clustering technique the goal is to estimate  that characterizes the best cluster for input vector X. The parameter vector  is sensitive to the shape of clusters. To define the topology of clusters a set of m points mi in the l-dimensional space is required which corresponds to a cluster

  [m1T , m2T ,.....mmT ]T / C (c1T , r1 , c2T , r2 ,.......cmT , rm ) d) Definition of cluster Let X be data set, for which m clustering is defined in R partition of X into m sets so that the following three conditions are met

X  {x1 , x2 ,......xn }

  [m1T , m2T ,.....mmT ]T Conditions for clustering

1.Ci   , i  1,........m m

2. Ci  X i 1

3.Ci  C j   , i, j  1,......, m The alternative definition is in terms of fuzzy sets is characterized by m functions u j where VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 99


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

u j : X  [0,1], J  1,....m And m

 u j ( xi )  1, 2,.....N , j 1

N

0   u j ( xi )  N , j  1, 2,....m i 1

These are called membership functions. The value of fuzzy membership function is a mathematical characterization of clusters which is not precisely defined and each vector x belongs to more than one cluster simultaneously. X  {x1 , x2 ,......xn } Ci   , i  1,........m m

Ci  X

i 1

Ci  C j   , i, j  1,......, m u j : X  [0,1], J  1,....m m

 u ( x )  1, 2,.....N , j 1

j

i

N

0   u j ( xi )  N , j  1, 2,....m i 1

e) Proximity measures: This parameter measure will quantify the similarities and dissimilarities between the two clusters and within the clusters with no bias in selected clusters that each cluster should contributed equally with no domination among each other. The proximity measure has two property functions which measure dissimilarity and similarity between two vectors. Dissimilarity measure

d : X *X R d 0  R :   d 0  d ( x, )

,

x, 

d ( xi , j ) f) Clustering Algorithm Having adopted proximity measure clustering criteria is applied to choose specific algorithm scheme that forms the clustering structure Most of the fuzzy clustering algorithm are derived by minimizing functions of form

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 100


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

N

m

J q ( , u )   uijq d ( xi , j ) t 1 j 1

Where J q ( , u) is clustering structure and uijq is membership function for each input vector and d ( xi , j ) is the proximity measure between the clusters. Now that the fuzzy clustering is defined the clusters are applied to adaptive layer of neuro fuzzy interference model. 4.3.4 Grid Partition clustering based Adaptive Neuro-Fuzzy Interference System Grid portioning is morphological clustering technique where the membership is sub divided into grid elements. Depending on the fuzzy rules R a fuzzy portioning is done on input membership function with cluster definitions N

m

J q ( , u )   uijq d ( xi , j ) t 1 j 1

Fig.4.30 Applied ANFIS grid partitioning architect VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 101


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Table 4.3.4 (a)Fuzzy Structure ANFIS type Fuzzy system And method Or method Defuzzification method Implication method Aggregation method Membership function Fuzzy rule Input Mf Type Output Mf type Max Epoch Error Goal Initial Step Step size Decrease rate Step size Increament rate

Grid Partitioning Sugeno Prod Max Weigthed average Prod Max Five 125 Gaussmf Linear 200 0 0.01 0.9 1.1

Table 4.3.4 b)Statistical Results of ANFIS Grid Partioning Cluster for 35 HRC and 45HRC Statistical Results of ANFIS Grid Partioning Cluster for 35 HRC ANFIS Grid35

Error mean(µ)

Results of ANFIS Grid Partioning Cluster for 45 HRC

Error STD(σ)

MSE

RMSE

Error mean(µ)

Error STD(σ)

MSE

RMSE

Surface roughness (Ra) Train Ra (µm)

7.74*10-4

6.41*10-4

4.11*10-7

6.41*10-4

5.49*10-7

6.185*10-4

3.82*10-7

6.181*10-4

Test Ra (µm)

1.74*10-4

0.0020

3.99*10-6

0.0020

1.09*10-4

0.0019

3.44*10-6

0.0019

Validation Ra (µm)

2.63*10

-5

9.74*10

-4

9.84*10

-7

9.74*10

-4

1.56*10

-5

9.17*10

-4

8.41*10

-7

9.17*10-4

Tangential force (Ft) 0.1352

0.0183

0.1351

5.69*10-5

0.0788

0.0062

0.0787

-0.0501

0.5101

0.2610

0.5109

7.612*10-4

0.3501

0.128

0.3148

-0.0023

0.714

0.41

0.57

1.62*10-4

0.153

0.0235

0.153

Axial force (Fa) Train Fa(N)

1.54*10-5

0.0847

0.0072

0.0847

3.76*10-5

0.0675

0.0045

0.0674

Test Fa(N)

0.0203

0.3453

0.118

0.3448

0.0055

0.1399

0.0195

0.1396

Validation Fa(N)

0.0031

0.1547

0.0239

01547

8.58*10-4

0.0824

0.0068

0.0824

2.4*10--5

0.082

0.0823

3.59*10-5

0.0654

0.00043

0.0654

Train Ft(N)

4.49*10

Test Ft(N) Validation Ft(N)

-5

Radial force(Fr) Train Fr(N)

0.0068

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 102


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Test Fr(N)

0.0567

0.3802

0.1468

0.3831

-0.0207

0.2173

0.0473

0.0654

Validation Fr(N)

0.0085

0.1665

0.0278

0.1666

-0.0031

0.1036

0.0107

0.1036

Tool Life (Tf) Train Tf (min)

7.56*10-6

0.0120

1.43*10-4

0.0120

4.44*10-6

0.0074

5.409*10-5

0.0074

Test Tf (min)

-6.69*10-4

0.0271

7.28*10-4

0.0270

0.0014

0.0235

5.487*10-4

0.0234

Validation Tf (min)

-9.34*10-4

0.0152

2.313*10-4

0.015

2.206*10-4

0.011

1.282*10-4

0.0113

4.3.5 ANFIS Grid Partioning Cluster Plots For Ra 35 HRC

Fig.4.31 Training Error Plots for Ra (Target vs

Fig.4.32 Testing Error Plots for Ra (Target vs Output)

Fig.4.33 Validation Error Plots for Ra (Target vs

Fig.4.34 Regression Plots for Ra (Train

Output)

/Test/Validate)

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 103


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.4.35 Response Surface Plot for Ra 4.3.6 ANFIS Grid Partioning Cluster Plots For Ft 35 HRC

Fig.4.36 Training Error Plots for Ft (Target vs

Fig.4.37 Testing Error Plots for Ft (Target vs

Output)

Output)

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 104


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.4.38 Validation Error Plots for Ft (Target vs

Fig.4.39 Regression Plots for Ft (Train /Test/Validate)

Output)

Fig.4.40 Response Surface Plot for Ft 4.3.7 ANFIS Grid Partioning Cluster Plots For Fa 35 HRC

Fig.4.41 Training Error Plots for Fa (Target vs

Fig.4.42 Testing Error Plots for Fa (Target vs

Output)

Output)

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 105


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.4.43 Validation Error Plots for Fa (Target vs

Fig.4.44 Regression Plots for Fa (Train

Output)

/Test/Validate)

Fig.4.45 Response Surface Plot for Fa 4.3.8 ANFIS Grid Partioning Cluster Plots For Fr 35 HRC

Fig.4.46 Training Error Plots for Fr (Target vs Output)

Fig.4.47 Testing Error Plots for Fr (Target vs Output)

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 106


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.4.48 Validation Error Plots for Fr (Target vs Fig.4.49 Regression Plots for F (Train /Test/Validate) r Output)

Fig.4.50 Response Surface Plot for Fr

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 107


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.9 ANFIS Grid Partioning Cluster Plots For Tf 35 HRC

Fig.4.51 Training Error Plots for Tf (Target vs

Fig.4.52 Testing Error Plots for Tf (Target

Output)

vs Output)

Fig.4.53 Validation Error Plots for Tf (Target vs

Fig.4.54 Regression Plots for Tf (Train

Output)

/Test/Validate)

Fig.4.55 Response Surface Plot for Tf VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 108


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.10 ANFIS Grid Partioning Cluster Plots For Ra 45 HRC

Fig.4.56 Training Error Plots for Ra (Target

Fig.4.57 Testing Error Plots for Ra (Target

vs Output)

vs Output)

Fig.4.58 Validation Error Plots for Ra (Target vs Output)

Fig.4.59 Regression Plots for Ra (Train /Test/Validate)

Fig.4.60 Response Surface Plot for Ra

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 109


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.11 ANFIS Grid Partioning Cluster Plots For Ft 45 HRC

Fig.4.61 Training Error Plots for Ft (Target vs

Fig.4.62 Testing Error Plots for Ft

Output)

(Target vs Output)

Fig.4.63 Validation Error Plots for Ft (Target vs Fig.4.64 Regression Plots for Ft (Train Output

/Test/Validate)

Fig.4.65 Response Surface Plot for Ft a

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 110


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.12 ANFIS Grid Partioning Cluster Plots For Fa 45 HRC

Fig.4.66 Training Error Plots for Fa (Target

Fig.4.67 Testing Error Plots for Fa

vs Output)

(Target vs Output)

Fig.4.68 Validation Error Plots for Fa (Target

Fig.4.69 Regression Plots for Fa (Train

vs Output

/Test/Validate)

Fig.4.70 Response Surface Plot for Fa a VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 111


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.13 ANFIS Grid Partioning Cluster Plots For Fr 45 HRC

Fig.4.71 Training Error Plots for Fr (Target

Fig.4.72 Testing Error Plots for Fr (Target

vs Output)

vs Output)

Fig.4.73 Validation Error Plots for Fr (Target vs Output

Fig.4.74 Regression Plots for Fr (Train /Test/Validate)

Fig.4.75 Response Surface Plot for Fr

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 112


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.14 ANFIS Grid Partioning Cluster for TF Plots For Tf 45 HRC

Fig.4.76 Training Error Plots for Tf (Target vs

Fig.4.77 Testing Error Plots for Tf

Output)

(Target vs Output)

Fig.4.78 Validation Error Plots for Tf (Target

Fig.4.79 Regression Plots for Tf (Train

vs Output)

/Test/Validate)

Fig.4.80 Response Surface Plot for Tf VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 113


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.15 Subtractive Clustering In subtractive clusetering each point in membership function forms ac luster center and the point with higher influence on any cluster center will take in the that cluster. Recalling the cluster function N

m

J q ( , u )   uijq d ( xi , j ) t 1 j 1

For subtractive clustering the dissimilarity function is d ( xi , j ) is defined by exponential function N

d ( xi , j )   exp[ J 1

|| xi  x j ||

 j (ra / 2)2

]

Where  j (ra / 2) is the radius of xi & x j are membership function points if a point has many other points surrounded around itself then point has highest density point. The highest density point is as first cluster center xc and in the consecutive the iteration the density measure of each point is obtained by subtracting cluster points by applying the following equation. N

di  di  d . exp[ J 1

|| xi  x j ||

 j (ra / 2)2

]

This is continued until the points in membership functions are exhausted. After the calculation of the dissimilarity at each point first cluster center is identified as the point having highest density. Eliminate all points in the vicinity around the first cluster center of its defined radius value. For the next iteration update the dissimilarity and apply the cluster function.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 114


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.4.81 Developed ANFIS (Subtractive Cluster) for machining system Table 4.3.15 (a) Fuzzy Structure ANFIS type

Substractive Clustering

Fuzzy system And method Or method Defuzzification method Implication method Aggregation method Cluster Radius No of clusters Max Epoch Error Goal Initial Step Step size Decrease rate Step size Increament rate

Sugeno Prod Probor Weigthed average Prod Max 0.5 12 100 0 0.01 0.9 1.1

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 115


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Table 4.3.15 (b) Statistical Error analysis of ANFIS Subtractive clustering for 35 HRC and 45 HRC Statistical Results of ANFIS Subtractive Cluster for 35 HRC ANFIS Grid35

Error mean(µ)

Error STD(σ)

MSE

RMSE

Results of ANFIS Subtractive Cluster for 45 HRC Error Error MSE RMSE mean(µ) STD(σ)

Surface roughness (Ra) Train Ra (µm)

1.0458*10-7

0.0044

1.922*10-5

0.0044

6.64*10-6

4.32*10-4

2.63*10-7

5.12*10-4

Test Ra (µm)

-2.82*10-5

0.0052

2.86*10-5

0.0052

1.165*10-5

0.0039

3.48*10-5

5.89*10-3

0.0045

2.036*10-5

0.0045

5.58*10-5

0.0035

1.25*10-5

0.0035

Validation Ra -4.121*10-6 (µm) Tangential force (Ft) Train Ft(N)

-4.28*10-6

0.5928

0.3510

0.5924

-1.29*10-6

0.4917

0.2415

0.4914

Test Ft(N)

-0.0135

0.6604

0.433

0.658

0.0128

0.6335

0.3988

0.6315

Validation -0.0020 Ft(N) Axial force (Fa) Train Fa(N) 2.98*10-6

0.6031

0.3633

0.6028

0.0019

0.515

0.2651

0.5149

0.3191

0.1017

0.318

3.621*10-6

0.2806

0.07806

0.2805

Test Fa(N)

0.4351

0.1886

0.434

0.0414

0.4176

0.1749

0.4183

0.3389

0.1147

0.3387

0.0062

0.305

0.0931

0.3051

-0.0233

Validation -0.35 Fa(N) Radial force(Fr) Train Fr(N)

-3.86*10-6

0.3432

0.1176

0.3430

4.85*10-6

0.3020

0.0911

0.3013

Test Fr(N)

-0.0484

0.380

0.1458

0.3818

-0.0051

0.352

0.1233

0.3512

Validation Fr(N) Tool Life (Tf)

-0.0073

0.3849

0.121

0.349

-7.62*10-4

0.3099

0.096

0.3098

Train Tf (min)

1.49*10-6

0.0325

0.011

0.0325

7.78*10-4

0.0249

6.19*10-4

0.0249

Test Tf (min)

0.0048

0.0358

0.0013

0.0360

0.0018

0.0283

7.96*10-4

0.0282

Validation Tf (min)

7.24*10-4

0.0311

0.0011

0.0311

2.776*10-4

0.0254

6.45*10-4

0.0254

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 116


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.16 ANFIS Subtractive clustering Plots For Ra 35 HRC

Fig.4.82 Training Error Plots for Ra (Target

Fig.4.83 Testing Error Plots for Ra (Target

vs Output)

vs Output)

Fig.4.84 Validation Error Plots for Ra

Fig.4.85 Regression Plots for Ra (Train

(Target vs Output)

/Test/Validate)

Fig.4.86 Response Surface Plot for Ra VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 117


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.17 ANFIS Subtractive clustering Plots For Ft 35 HRC

Fig.4.87 Training Error Plots for Ft (Target vs Fig.4.88 Testing Error Plots for Output)

(Target vs Output)

Fig.4.89 Validation Error Plots for Ft (Target vs

Fig.4.90 Regression Plots for Ft

Output)

(Train /Test/Validate

Fig.4.91 Response Surface Plot for Ft

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 118

Ft


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.18 ANFIS Subtractive clustering Plots For Fa 35 HRC

Fig.4.92 Training Error Plots for Fa (Target vs

Fig.4.93 Testing Error Plots for Fa

Output)

(Target vs Output)

Fig.4.94 Validation Error Plots for Fa (Target vs Output)

Fig.4.95 Regression Plots for (Train /Test/Validate)

Fig.4.96 Response Surface Plot for Fa

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 119

Fa


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.19 ANFIS Subtractive clustering Plots For Fr 35 HRC

Fig.4.97 Training Error Plots for Fr (Target vs Output)

Fig.4.99 Validation Error Plots for Fr (Target vs Output)

Fig.4.98 Testing Error Plots for

Fr

(Target vs Output)

Fig.4.100 Regression Plots for (Train /Test/Validate)

Fig.4.101 Response Surface Plot for Fr

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 120

Fr


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.20 ANFIS Subtractive clustering Plots For Tf 35 HRC

Fig.4.102 Training Error Plots for Tf (Target

Fig.4.103 Testing Error Plots for Tf

vs Output)

(Target vs Output)

Fig.4.104 Validation Error Plots for Tf (Target

Fig.4.105 Regression Plots for Tf (Train

vs Output)

/Test/Validate)

Fig.4.106 Response Surface Plot for Tf

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 121


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.21 ANFIS Subtractive clustering Plots For Ra 45 HRC

Fig.4.107 Training Error Plots for Ra

Fig.4.108 Testing Error Plots for Ra (Target

(Target vs Output)

vs Output)

Fig.4.109 Validation Error Plots for Ra

Fig.4.110 Regression Plots for Ra (Train

(Target vs Output)

/Test/Validate)

Fig.4.111 Response Surface Plot for Ra

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 122


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.22 ANFIS Subtractive clustering Plots For Ft 45 HRC

Fig.4.112 Training Error Plots for Ft

Fig.4.113 Testing Error Plots for Ft

(Target vs Output)

(Target vs Output)

Fig.4.114 Validation Error Plots for Ft

Fig.4.115 Regression Plots for Ft (Train

(Target vs Output)

/Test/Validate)

Fig.4.116 Response Surface Plot for Ft VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 123


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.23 ANFIS Subtractive clustering Plots For Fa 45 HRC

Fig.4.117 Training Error Plots for Fa (Target

Fig.4.118 Testing Error Plots for Fa (Target vs

vs Output)

Output)

Fig.4.119 Validation Error Plots for Fa (Target

Fig.4.120 Regression Plots for Fa (Train

vs Output)

/Test/Validate)

Fig.4.121 Response Surface Plot for Fa

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 124


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.24 ANFIS Subtractive clustering Plots For Fr 45 HRC

Fig.4.122 Training Error Plots for Fr (Target vs

Fig. 4.123 Testing Error Plots for Fr (Target

Output)

vs Output)

Fig.4.124 Validation Error Plots for Fr (Target vs Output)

Fig.4.125 Regression Plots for /Test/Validate)

Fig.4.126 Response Surface Plot for Fr VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 125

Fr (Train


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.25 ANFIS Subtractive clustering Plots For Tf 45 HRC

Fig.4.127 Training Error Plots for

Tf

(Target vs Output)

Fig.4.129 Validation Error Plots for TF (Target vs Output)

Fig.4.128 Testing Error Plots for Tf (Target vs Output)

Fig.4.130 Regression Plots for (Train /Test/Validate

Fig.4.131 Response Surface Plot for Tf VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 126

Tf


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.26 Fuzzy C Mean Clustering

Fuzzy C-Mean clustering is another circular invariant clustering technique the radius of clusters is calculated by membership functions. Recall the cluster function N

m

J q ( , u )   uijq d ( xi , j ) t 1 j 1

N

Compute the cluster mean i 

u j 1 M

m

ij

 (u j 1

* xj

ij

)m

Compute dissimilarity function d ( xi , j ) || xi   j ||

Update the member partition matrix uij by uij 

1 M

k 1

dik

(d

) 2/ m1

kj

Evaluate cluster function J q ( , u) and do until the cluster criteria is reached Table 4.3.26 a) Fuzzy structure

ANFIS type Fuzzy system And method Or method Defuzzification method Implication method Aggregation method No of clusters Partition matrix Exponenet Maximum iteration Improvemenet level Max Epoch Error Goal Initial Step Step size Decrease rate Step size Increament rate

Fuzzy C Mean Sugeno Prod Probor Weigthed average Prod Max 15 2 200 1*10-5 200 0 0.01 0.9 1.1

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 127


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.4.132 Developed ANFIS Fuzzy C Mean Clustering architect

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 128


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Table 4.3.26 b) Statistical Results of ANFIS FCM for 35 HRC and 45 HRC Statistical Results of ANFIS Fuzzy C -Mean for 35 HRC

Results of ANFIS Fuzzy C -Mean Cluster

ANFIS Grid-35

for 45 HRC Error Error mean(µ) STD(σ)

Error mean(µ)

Error STD(σ)

MSE

RMSE

MSE

RMSE

Surface roughness (Ra) Train (µm)

Ra 7.014*10-8 0.0046

2.077*10-5

0.0046

5.58*10-8

0.0033

1.076*1 0-5

0.0033

Test Ra (µm)

1.165*10-4 0.0053

2.78*10-5

0.0053

-0.0032

0.0457

0.0021

0.0456

Validation Ra (µm)

1.753*10-5 0.0047

2.182*10-5

0.0047

2.64*10-5

8.89*10-

7.34*10-

8.56*10-

4

7

4

Tangential force (Ft) Train Ft(N)

-1.02*10-

1.2296

1.5101

1.228

-2.93*10-6

1.2581

1.581

1.2574

Test Ft(N)

-0.0224

1.897

3.577

1.893

-0.1034

1.4454

2.085

1.44

1.3498

1.8202

1.3491

-0.0155

1.287

1.656

1.287

5

Validation -0.0034 Ft(N) Axial force (Fa) Train Fa(N)

-3.10*10-

0.7534

0.5669

0.753

6.36*10-6

0.466

0.2172

0.4661

Test Fa(N)

0.0835

0.9753

0.953

0.976

-0.0626

0.59206

0.3500

0.5916

Validation Fa(N)

0.0125

0.7908

0.625

0.7905

-0.0094

0.4871

0.2371

0.487

6

Radial force(Fr) Train Fr(N)

-2.149*10-

0.863

0.744

0.862

-3.36*10-7

2.12*10-4

0.0041

0.064

Test Fr(N)

0.1296

0.935

0.886

0.941

-0.0112

0.559

0.3132

0.559

Validation Fr(N)

0.0194

0.875

0.765

0.875

-0.0017

0.559

0.3132

0.5597

0.0027

0.0524

1.347*10-6

0.0460

0.0021

0.046

0.0053

0.0729

-0.0032

0.0457

0.0021

0.0456

0.0031

00560

-4.80*10-4

0.0459

0.0021

0.0459

6

Tool Life (Tf) Train Tf 2.648*10-6 (min) Test Tf (min) 0.0058 Validation Tf (min)

8.8668*10-4

0.052 9 0.072 9 0.056 0

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 129


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.27 ANFIS Fuzzy C -Mean Clustering Plots For Ra 35 HRC

Fig. 4.133 Training Error Plots for Ra

Fig.4.134 Testing Error Plots for Ra

(Target vs Output)

(Target vs Output)

Fig.4.135 Validation Error Plots for Ra (Target vs Output)

Fig.4.136 Regression Plots for Ra (Train /Test/Validate

Fig.4.137 Response Surface Plot for

Ra

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 130


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.28 ANFIS Fuzzy C -Mean Clustering Plots For Ft 35 HRC

Fig.4.138 Training Error Plots for Ft (Target

Fig.4.139 Testing Error Plots for Ft

vs Output)

(Target vs Output)

Fig.4.140 Validation Error Plots for Ft (Target vs Output)

Fig.4.141 Regression Plots for (Train /Test/Validate

Fig.4.142 Response Surface Plot for

Ft

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 131

Ft


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.29 ANFIS Fuzzy C -Mean Clustering Plots For Fa 35 HRC

Fig.4.143 Training Error Plots for

Fa

(Target vs Output)

Fig.4.145 Validation Error Plots for Fa (Target vs Output)

Fig.4.144 Testing Error Plots for

Fa

(Target vs Output)

Fig4.146. Regression Plots for (Train /Test/Validate

Fig.4.147 Response Surface Plot for Fa

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 132

Fa


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.30 ANFIS Fuzzy C -Mean Clustering Plots For Fr 35 HRC

Fig.4.148 Training Error Plots for

Fr

(Target vs Output)

Fig.4.150 Validation Error Plots for Fr (Target vs Output)

Fig.4.149 Testing Error Plots for

Fr

(Target vs Output)

Fig.4.151 Regression Plots for (Train /Test/Validate

Fig.4152 Response Surface Plot for Fr

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 133

Fr


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.31 ANFIS Fuzzy C -Mean Clustering Plots For Tf 35 HRC

Fig.4.153 Training Error Plots for

Tf

(Target vs Output)

Fig.4.155 Validation Error Plots for Tf (Target vs Output)

Fig.4.154 Testing Error Plots for

Tf

(Target vs Output)

Fig.4.156 Regression Plots for (Train /Test/Validate)

Fig.4.157 Response Surface Plot for Tf

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 134

Tf


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.32 ANFIS Fuzzy C -Mean Clustering Plots For Ra 45 HRC

Fig.4.158 Training Error Plots for Ra

Fig.4.159 Testing Error Plots for

Ra

(Target vs Output)

(Target vs Output)

Fig.4.160 Validation Error Plots for Ra

Fig.4.161 Regression Plots for Ra (Train

(Target vs Output)

/Test/Validate

Fig.4.162 Response Surface Plot for Ra

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 135


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.33 ANFIS Fuzzy C -Mean Clustering Plots For Ft 45 HRC

Fig.4.163 Training Error Plots for Ft

Fig.4.164 Testing Error Plots for

Ft

(Target vs Output)

(Target vs Output)

Fig.4.165 Validation Error Plots for Ft

Fig.4.166 Regression Plots for Ft (Train

(Target vs Output)

/Test/Validate

Fig.4.167 Response Surface Plot for

Ft

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 136


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.34 ANFIS Fuzzy C -Mean Clustering Plots For Fa 45 HRC

Fig.4.168 Training Error Plots for

Fa (Target

Fig.4.169 Testing Error Plots for

vs Output)

Output)

Fig.4.170 Validation Error Plots for Fa (Target vs Output)

Fa (Target vs

Fig.4.171 Regression Plots for Fa (Train /Test/Validate)

Fig.4.172 Response Surface Plot for Fa

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 137


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.35 ANFIS Fuzzy C -Mean Clustering Plots For Fr 45 HRC

Fig.4.173 Training Error Plots for Fr

Fig.4.174 Testing Error Plots for Fr

(Target vs Output)

(Target vs Output)

Fig.4.175 Validation Error Plots for Fr

Fig.4.176 Regression Plots for Fr (Train

(Target vs Output)

/Test/Validate)

Fig.4.177 Response Surface Plot for Fr

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 138


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.3.36 ANFIS Fuzzy C -Mean Clustering Plots For Tf 45 HRC

Fig.4.178 Training Error Plots for

Tf

Fig.4.179 Testing Error Plots for

(Target vs Output)

(Target vs Output)

Fig.4.180 Validation Error Plots for Tf

Fig.4.181 Regression Plots for Tf

(Target vs Output)

(Train /Test/Validate)

Fig.4.182 Response Surface Plot for Tf

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 139

Tf


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.4 Comparison of Prediction Results with Experimental Statistics 4.4.1 Statistical Comparison of Neural Network and ANFIS Prediction Results with Experimental Statistics for AISI 4340 Steel 35hrc NN35 Error Error MSE RMSE NN VS Experimental 35 HRC STD(σ) Ra 0.0442 1.0814 1.1130 1.0550 Ft (N) 0.186.6 126.4 594.81 24.38 4 Fa (N) 103.98 33.77 1.897*10 109.07 Fr (N) 1.1318 93.95 93.95 9.69 ANFIS grid partitioning VS Experimental 35 HRC Ra -0.2308 0.5964 0.3912 0.6255 4 Ft (N) 17.49 106.9 1.116*10 105.66 Fa (N) 157 80 3.09*104 176 Fr (N) 32.57 41.29 2.6*103 51.78 ANFIS Subtractive VS Experimental 35 HRC Ra -0.226 0.5964 0.3875 0.6225 Ft (N) 15.29 105.36 1.07*104 103 4 Fa (N) 157.79 82.3 3.133*10 177.016 Fr (N) 27.89 40.81 2.3604*103 48.58 ANFIS FCM VS Experimental 35 HRC Ra -0.2339 0.5964 0.392 0.6266 Ft (N) 18.19 104.9 1.08*104 103.9 Fa (N) 157 79.42 3.073*104 175.32 Fr (N) 32.23 42.20 2.73*103 52.25

Error Error MSE RMSE NN VS Experimental STD(σ) 45 HRC -0.1700 0.2769 0.1018 0.3190 -57.23 128.3 1.8929*104 137.5 54.67 33.57 4.0607*103 63.72 0.415 14.0720 188.29 13.72 ANFIS grid partitioning VS Experimental -0.9278 0.5964 1.198 1.0949 3 -22.35 77.58 6.128*10 78.85 -7.5709 40.08 1.5841*103 39.8 -7.69 38.35 1.457*103 38.17 ANFIS Subtractive VS Experimental 45 -0.947 0.594 1.23 1.109 -23.106 76.64 6.1143*103 78.19 -6.9912 39.58 1.537*103 39.213 -8.525 36.57 1.343*103 36.65 ANFIS FCM VS Experimental 45 HRC -0.856 0.596 1.0714 1.0351 -21.7 78.8 6.38*103 79.8 -6.8003 38.45 6.36*10-6 0.4664 -6.63 36.89 1.33*103 36.56

4.4.2 Error Plots of Neural Network Prediction Results with Experimental Statistics for AISI 4340 Steel 35hrc

Fig.4.183 Error Estimation Plots For R a

Fig4.184.Error Estimation Plots For F t

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 140


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.4.185.Error Estimation Plots For F a

Fig.4.186 Error Estimation Plots For F r

4.4.3 Error Plots of Neural Network Prediction Results with Experimental Statistics for AISI 4340 Steel 45hrc

Fig.4.187.Error Estimation Plots For Ra

Fig.4.188 Error Estimation Plots For F t

Fig.4.189 Error Estimation Plots For F a

Fig.4.190 Error Estimation Plots For F r

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 141


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.4.4 Error Plots of ANFIS (Grid Partitioning Clustering) Results with Experimental Statistics for AISI 4340 Steel 35hrc

Fig.4.191 Error Estimation Plots For R a

Fig.4.193 Error Estimation Plots For F a

Fig.4.192 Error Estimation Plots For F t

Fig.4.194 Error Estimation Plots For F r

4.4.5 Error Plots of ANFIS (Grid Partitioning Clustering) Results with Experimental Statistics for AISI 4340 Steel 45hrc

Fig.4.195 Error Estimation Plots For R a

Fig.4.196 Error Estimation Plots For F t

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 142


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.4.197 Error Estimation Plots For F a

Fig.4.198 Error Estimation Plots For F r

4.4.6 Error Plots of ANFIS (Subtractive Clustering) Results with Experimental Statistics for AISI 4340 Steel 35hrc

Fig.4.199 Error Estimation Plots For R a

Fig.4.200 Error Estimation Plots For F t

Fig.4.201 Error Estimation Plots For F a

Fig.4.202 Error Estimation Plots For F r

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 143


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.4.7 Error Plots of ANFIS (Subtractive Clustering) Results with Experimental Statistics for AISI 4340 Steel 45hrc

Fig.4.203 Error Estimation Plots For R a

Fig.4.204 Error Estimation Plots For F t

Fig.4.205 Error Estimation Plots For F a

Fig.4.206 Error Estimation Plots For F r

4.4.8 Error Plots of ANFIS (Fuzzy C-Mean Clustering)) Results with Experimental Statistics for AISI 4340 Steel 35hrc

Fig.4.207 Error Estimation Plots For R a

Fig.4.208 Error Estimation Plots For F t

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 144


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.4.209 Error Estimation Plots For F a

Fig.4.210 Error Estimation Plots For F r

4.4.9 Error Plots of ANFIS (Fuzzy C-Mean Clustering)) Results with Experimental Statistics for AISI 4340 Steel 45hrc

Fig.4.211 Error Estimation Plots For R a

Fig.4.213 Error Estimation Plots For F a

Fig.4.212 Error Estimation Plots For F t

Fig.4.214 Error Estimation Plots For F r

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 145


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

4.5 Conclusion 

From the Comparison PLOTS and table the mean error and RMS of neural network was found to be lower compared to ANFIS models

Though the errors in neural network was less the prediction curve for surface roughness should poor match with the experimental curve while in ANFIS the curve match for surface roughness was better in comparison to neural network.

 With this results learning techniques for were further attempted to improve with synergies.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 146


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

CHAPTER 5 HYBRIDIZATION OF C.I SYNERGIES 5.1 Introduction In the previous chapters optimization and prediction techniques were applied exclusively and the results obtained from them were found to be pretty convincing when compared with the experimental statistics In this chapter hybridization of these techniques is applied to our current machining problem with objective to improve the ability of techniques through mutual assistance and improve prediction and optimization ability of exclusive techniques. A exposition of adapted synergies is illustrated in brief. In the first section the combinations of Neuro-Evolutionay and Neuro-Swarm techniques is implemented , in the second section combination of Evolutionary - Neuro fuzzy and Swarm-Neuro fuzzy is exercised and in the third segment a .comparison is made between the predicted results obtained from synergism and experimental statistics. The objective of this chapter is depicted through the flow chart.

Fig.5.1 Chapter flow chart

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 147


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.2 EA-NN Synergism Various types of EAs and NNs synergies are possible which can be broadly classified in three combinations supportive combination, collaborative combination and amalgamated combination 

EA uses a population of entire solution space of optimization problem and in contrast NN uses these optimized results as exemplars for training and the learning then converges depending on the learning parameters and topology of NN. The performance of both EA and NN can be improved by accelerating the convergence if an appropriate population of data sets and strategic learning parameters is applied.

In a supportive combination EAs and NNs are used sequentially where one is primary problem solver and other is secondary.

In collaborative combination they are used simultaneously where both EAs and NNs solve the problem together and in amalgamated combination the EA search technique and NN as pattern model.

Collaborative Combination In collaborative learning both EAs and NNs are used simultaneously using the result of one to prepare data set for other .In other words one technique plays primary role of solving problem and the other technique is a supportive to solve the problem.

Finding an appropriate topology of NN for a given problem is trial and error task. Synergies between EAs and NN assists in determining optimal network architecture and then evaluate neural network. A typical procedure for EA-NN synergy can be outlines as below 1. Create an initial population of individuals which would go into Evolutionary algorithm and generate a optimal population which will be data set for Neural Network architect.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 148


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

2. Set up the training data for neural network as received from the EA results by permuting train and target sets and dividing them for training testing and validation 3. Apply learning criteria for network and evaluate the weights and bias for the targets. Test the targets against the expected outputs 4. Evaluate the training errors and fitness of network for current learning, transfer back weights and bias along with the targets to EA. Repeat the step2-4 until the convergence or max generation is reached. 5.2.1 NSGA combined Neural Network

Fig.5.2 Developed NSGA–NN architect

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 149


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Table 5.2.1 (a) Description of NSGA-NN

Population size Generation Crossover probability Crossover constant Mutation probability Mutation constant Neural Network Type Training function No of neurons in Hidden layer No of neurons in output layer Weights in hidden layer Weights in output layer Training samples Testing samples Validation samples Transfer function Training performance Testing performance Validation performance

1000 100 0.8 0.1 0.1 0.2 Feed forward neural network Levenberg-Marqaurdt 10 5 30 [3×10] 50[5×10] [700 3] [150 3] [150 3] Tan-sigmoid function 2.861e-04(35HRC) 1.861e-04(45HRC) 4.147e-04(35HRC) 5.847e-04(45HRC) 1.777e-04(35HRC) 8.777e-04(45HRC)

5.2.1 (b) Calibrated weigths and bias for NSGA-NN architect Calibrated weights and bias for 35 HRC Steel Hidden layer Definition(sij) Bias( W1(v W2(f) W3(d -0.771

0.242

-0.284

0.444

Output layer Definition(sij) bias W1(R W2(F W3(F W4(Fr  3.444 -2.004 0.010 0.068 -0.154

-0.410

-0.040

-0.039

-0.248

1.881

0.511

-1.764

0.516

-4.511

1.559

-0.734

0.065

0.155

-0.224

1.784

-1.875

2.474

1.788

1.7166

0.607

-0.279

-0.208

0.206

0.0684

3.254

3.329

2.564

0.611

-0.616

0.637

-1.215

-0.633

0.895

0.274

3.678

0.772

0.0842

-0.040

0.0041

-0.136

W5(Tf -0.138

0.746

0.164

0.173

-0.102

0.5492

0.377

-2.206

-0.811

-0.746

0.895

-0.051

-0.037

-0.306

-3.277

-0.713

-3.158

-2.335

-0.383

0.870

-0.0149 0.164

0.227 -0.16

0.050 0.2133

-1.800 1.23

-1.466 2.071

3.607 1.67

-1.017 -2.687

-2.811 0.373

-0.23

-0.21

-1.002

1.107

-0.036

0.1082

-0.077

0.5352

0.55 -1.672

Calibrated weights and bias for 45 HRC Steel -0.810

-0.856

-2.308

1.099

0.874

-0.651

-2.535

-0.896

-0.856

-2.308

1.099

1.1200

0.528

-0.651

-2.535

-0.896

-1.368

-0.039

-0.020

2.240 0.017 4 2.414

-1.434

0.0990

0.0157

0.731

-0.103

0.478

0.099

0.0157

-0.0209

-0.0158

-0.0248

-0.793

3.064

-0.1033

0.4782

-0.793

1.200

-0.135

1.867

-3.515

1.823

2.756

3.370

1.867

1.823

2.7567

2.046

0.589

2.358 -0.733

0.005

-0.001

0.002

0.0051

-0.001

0.0002

0.0021

0.0335

-0.013

-0.773

0.496

-0.013

-0.773

0.496

1.115

0.933

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 150


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

-6.767 1.321 0.118

-0.0140

0.024

-0.020

-0.014

0.0242

-0.0207

-0.009

0.261

-0.444

-0.415

-1.545

-0.692

-1.112

-0.749

0.275 0.111

-0.109

0.164

0.697

-0.938

0.5015

1.499

-0.848

-0.583

5.2.2 Results of NSGA-NN 35 HRC Steel

Fig.5.3 Performance plot of Network

Fig.5.4 Training state of Network at each epoch

Fig.5.5 Training error in Ra

Fig.5.6 Regression fit plot for Ra

Fig.5.7 Training error in Ft

Fig.5.8 Regression fit plot for Ft

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 151


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.5.10 Training error in Fa

Fig.5.11 Regression fit plot for Fa

Fig.5.12 Training error in Fr

Fig.5.13 Regression fit plot for Fr

Fig.5.14 Training error in Tf

Fig.5.15 Regression fit plot for Tf

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 152


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.2.3 Results of NSGA-NN for 45 HRC Steel

Fig.5.16 Performance plot of Network

Fig.5.17 Training state of Network at each epoch

Fig.5.18 Training error in Ra

Fig.5.19 Regression fit plot for Ra

Fig.5.20 Training error in Ft

Fig.5.21 Regression fit plot for Ft

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 153


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.5.22 Training error in Fa

Fig.5.24 Training error in Fr

Fig.5.26 Training error in Tf

Fig.5.23 Regression fit plot for Fa

Fig.5.25 Regression fit plot for Fr

Fig.5.27 Regression fit plot for Tf

5.3 SI-NN synergism Similar to EA-NN synergies, SI-NN is has three class of synergies which differ in the degree of coupling and interdependency in working towards solution. The collaborative

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 154


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

combination strategy is analogues to EA-NN strategy. The figure below illustrates the collaborative combination of SI and NN.

Fig.5.28. SI-NN collaborative combination 5.3.1 PSO combined Neural Network

Fig.5.29 Applied SI-NN synergy architect

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 155


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Table 5.3.1 (a) Description of PSO-NN Total particle population Max no of repository elements Inertia weight ( w) Inertia weight damping rate (w.damp) Personal learning co efficient (a1) Global learning co-efficient (a2) No of grid in each dimension Inflation rate ( ὰ) Leader selection pressure ( ) Deletion selection pressure ( ) Mutation rate (mu) Neural Network Type Training function No of neurons in Hidden layer No of neurons in output layer Weights in hidden layer Weights in output layer Training samples Testing samples Validation samples Transfer function Training performance Testing performance Validation performance

1000 500 0.5 0.99 1 2 7 0.1 2 2 0.1 Feed forward neural network Levenberg-Marqaurdt 10 5 30 [3×10] 50[5×10] [700 3] [150 3] [150 3] Tan-sigmoid function 1.5224e-05 2.089e-05 2.03e-05

5.3.2 (b) Calibrated weights and bias for 35 SI-NN Hidden layer Definition(sij) Bias(i) W1(vc) W2(f) -0.904 1.656 1.671 -3.013 2.294 -1.05 0.829 -0.131 0.344 -3.295 1.724 1.051 -1.695 0.330 0.163 -0.369 -0.369 0.143 2.0164 -2.643 3.987 -1.248 2.022 -0.81 -0.629 0.403 0.346 -0.705 3.124 -1.47 -0.633 -0.304 -0.02 0.3125 0.390 -0.23 0.896 2.234 -1.43 -0.362 -0.010 -0.09 1.834 -0.298 -0.63

W3(d) -0.155 0.092 0.001 0.0081 -0.869 0.166 -0.614 3.412 -0.012 1.981 -0.185 0.075 1.056 -0.21 -1.286

Output layer Definition(sij) bias(i) W1(Ra) W2(Ft) 1.100 -0.027 -0.001 2.257 -0.121 -0.006 1.922 -0.200 -0.453 0.426 0.111 0.061 0.227 -0.292 0.017 3.723 2.250 0.001 0.0006 0.001 -0.001 0.44 0.556 -0.018 -0.002 0.833 -2.606 0.375 -0.425 0.087 0.010 0.540 -0.184 -0.166 0.0019

W3(Fa) -0.003 0.006 0.908 0.057 0.243 0.860 -0.0001 -0.003 0.616 0.0009 1.609 1.281 -0.005 -1.76 -0.0005

W4(Fr) 0.0011 -0.005 -0.629 0.049 0.167 -1.183 0.00007 -0.0002 0.436 0.0004 0.649 -1.364 -0.008 -3.427 -0.0023

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 156

W5(Tf) 0.004 -0.03 -1.07 -0.01 0.2031 0.1944 0.0015 0.0024 0.122 -0.001 2.208 0.150 0.007 0.285 -0.002


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

-4.050 -0.983 1.326 1.975 3.971

-4.123 -0.091 2.797 -0.015 1.211

2.584 -0.05 -1.62 0.364 -1.51

1.262 0.440 -0.057 0.611 -0.028

-1.35 -0.911 0.107 -1.441 0.4813

0.284 0.0689 0.0142 -0.074 -1.663

0.0736 1.9151 -0.0097 0.12064 -0.3358

-0.914 1.440 -0.029 0.341 -0.469

-2.361 -0.087 -0.024 -0.802 0.126

5.3.2 Results of PSO-NN for 35 HRC Steel

Fig.5.30 Performance plot of Network

Fig.5.31 Performance plot of Network

Fig.5.32 Training error in Ra

Fig.5.33 Regression fit plot for Ra

Fig.5.34 Training error in Ft

Fig.5.35 Regression fit plot for Ft

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 157


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.5.36 Training error in Fa

Fig.5.38 Training error in Fr

Fig.5.40 Training error in Tf

Fig.5.37 Regression fit plot for Fa

Fig.5.39 Regression fit plot for Fr

Fig.5.41 Regression fit plot for Tf

5.4 Synergies of EA and ANFIS The synergism between EA and ANFIS is strongly coupled here the EA technique is toapplied on the membership function which optimizes fitness value of fuzzy output. The VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 158


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

below flow chart outlines the applied strategy in synergism .Membership function are clustered through Fuzzy C-mean clustering and the fuzzy structure utilized is same as previously applied (refer chapter 4 table (4.3.27 (a) )). A detailed description is given in appendix. 5.4.1 ANFIS GA

Fig.5.42 GA based ANFIS (FCM) applied Strategy [72] 5.4.1 (a) Statistical Error analysis of GA based ANFIS (FCM) for 35 HRC and 45HRC Results of GA based ANFIS (FCM) for 35 HRC ANFIS Grid-

Error

Error

35

mean(µ)

STD(σ)

MSE

Results of GA based ANFIS (FCM) for 45 HRC RMSE

Error

Error

mean(µ)

STD(σ)

MSE

RMSE

Surface roughness (Ra) Train

Ra

Test Ra (µm)

-1.5*10-15

0.1016

0.0103

0.1015

3.14*10-16

0.0682

0.0046

0.0681

-0.0020

0.1004

0.0100

0.1002

0.0014

0.069

0.0048

0.0692

Tangential force (Ft) 9.22*10-3

Train Ft(N)

Test Ft(N) -0.3057 Axial force (Fa)

23.7

561.42

23.69

-2.69*10-13

18.31

335.08

18.30

25.68

657.6

25.64

1.3611

19.58

384

19.59

Train Fa(N)

1.21*10-13

11.57

133.7

11.56

-1.69*10-13

8.42

70.8

8.4176

Test Fa(N)

0.0194

10.79

116.09

10.77

1.681

9.79

98.45

9.92

Radial force(Fr) VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 159


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation 4.037*10-13 -0.0449

Train Fr(N) Test Fr(N)

7.568 7.90

57.2 62.314

7.56 7.894

-4.3*10-13 1.082

6.266 7.41

38.70 55.97

6.22 7.481

Tool Life (Tf) Train Tf (min)

-3.06*10-14

0.7850

0.6153

0.7844

2.69*10-14

0.6304

0.39

0.62

Test Tf (min)

0.034

0.80

0.6431

0.0802

-0.0157

0.16

0.38

0.618

Results of PSO based ANFIS (FCM) (35 HRC)

Results of PSO based ANFIS (FCM) 45 HRC

Surface roughness (Ra) Train Ra (µm) -8.43*10-17 Test Ra (µm) -4.79*10-4 Tangential force (Ft) Train Ft (N) -3.3*10-13 Test Ft (N) -2.214 Axial force (Fa) 1.02*10-13

Train Fa (N)

Test Fa (N) -0.2655 Radial force(Fr) Train Fr (N) -2.36*1013

1.4*10-15 1.4*10-15

1.4*10-15 1.4*10-15

0.0988 0.1010

0.2168 0.0102

0.4656 0.1008

0.0042 0.0048

0.0681 0.0692

23.87

589.7

23.86

-4.98*10-13

18.57

344.4

23.79

569.38

23.86

1.178

18.90

357.6

10.98

120.46

10.96

-5.59*10-13

8.78

77.08

8.78

11.98

143.22

11.96

0.080

9.045

81.54

9.03

7.57

57.22

7.56

-3.23*10-13

6.57

42.79

6.54

12.55

Test Fr (N) Tool Life (Tf) Train Tf (min)

0.221

7.71

59.42

7.708

-0.1446

6.73

45.25

6.72

3.03*10-14

0.796

0.6335

0.795

2.90*10-14

0.6186

45.25

6.72

Test Tf (min)

0.0042

0.88

0.774

0.8817

-0.0036

0.64

0.416

0.645

5.4.2 GA based ANFIS (FCM) Plots For Ra 35 HRC

Fig.5.43 Training Error Plots for Ra (Target vs Output)

Fig.5.44 Testing Error Plots for Ra (Target vs Output)

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 160


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.5.45 Regression Plots for Ra (Train /Test/Validate)

Fig.5.46 Response Surface Plot for Ra 5.4.3 GA based ANFIS (FCM) Plots For Ft 35 HRC

Fig.5.47 Training Error Plots for Ft (Target vs Output)

Fig.5.48 Testing Error Plots for Ft (Target vs Output)

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 161


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.5.49 Regression Plots for Ft (Train /Test/Validate)

Fig.5..50 Regression Plots for Ft (Train /Test/Validate) 5.4.4 GA based ANFIS (FCM) Plots For Fa 35 HRC

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 162


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.5.51 Training Error Plots for Fa (Target vs Output)

Fig.5.52 Testing Error Plots for Fa (Target vs Output)

Fig.5.53 Testing Error Plots for Fa (Target vs Output)

Fig.5.54 Response Surface Plot for Fa 5.4.5 GA based ANFIS (FCM) Plots For Fr 35 HRC

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 163


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.5.55 Training Error Plots for Fr (Target vs Output)

Fig.5.56 Testing Error Plots for Fr (Target vs Output)

Fig.5.57 Regression Plots for Fr (Train /Test/Validate)

Fig.5.58 Response Surface Plot for Fr 5.4.6 GA based ANFIS (FCM) Plots For Tf 35 HRC

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 164


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.5.59 Training Error Plots for Tf (Target Fig.5.60 Testing Error Plots for Tf (Target vs Output) vs Output)

Fig.5.61 Regression Plots for Tf (Train /Test)

Fig.5.62 Response Surface Plot for Tf

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 165


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.4.7 GA based ANFIS (FCM) Plots For Ra 45 HRC

Fig.5.63 Training Error Plots for Ra (Target vs Output)

Fig.5.64 Testing Error Plots for Ra (Target vs Output)

Fig.5.65 Regression Plots for Ra (Train /Test)

Fig.5.66 Response Surface Plot for Ra

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 166


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.4.8 GA based ANFIS (FCM) Plots For Ft 45 HRC

Fig.5.67 Training Error Plots for Ft (Target vs Output)

Fig.5.68 Training Error Plots for Ft (Target vs Output)

Fig.5.69 Regression Plots for Ft (Train /TesT

Fig.5.70 Response Surface Plot for Ft a VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 167


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.4.9 GA based ANFIS (FCM) Plots For Fa 45 HRC

Fig.5.71 Training Error Plots for Fa (Target vs Output)

Fig.5.72 Testing Error Plots for Fa (Target vs Output)

Fig.5.73 Regression Plots for Fa (Train /Test)

Fig.5.74 Response Surface Plot for Fa a

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 168


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.4.10 GA based ANFIS (FCM) Plots For Fr 45 HRC

Fig.5.75 Training Error Plots for Fr (Target vs Output)

Fig.5.76 Testing Error Plots for Fr (Target vs Output)

Fig.5.77 Regression Plots for Fr (Train /Test/Validate)

Fig.5.78 Response Surface Plot for Fr a VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 169


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.4.11 GA based ANFIS (FCM) Plots For Tf 45 HRC

Fig.5.79 Training Error Plots for Tf (Target vs Output)

Fig.5.80 Testing Error Plots for Tf (Target vs Output)

Fig.5.81 Regression Plots for Tf (Train /Test)

Fig.5.82 Response Surface Plot for Tf a VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 170


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.5 PSO based ANFIS (FCM) 35 HRC and 45HRC Steel The synergism between SI and ANFIS is also strongly coupled where the SI technique is applied on the membership function to optimize fitness value of fuzzy output. The below flow chart outlines the applied strategy in synergism .Membership function are clustered through Fuzzy Cmean clustering and the fuzzy structure utilized is same as previously applied (refer chapter 4 table (4.3.27 (a) )). A detailed description is given in appendix. 5.5.1 PSO based ANFIS (FCM)

Fig.5.83 PS0-ANFIS applied strategy [72]

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 171


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.5.2 PSO based ANFIS (FCM) Plots For Ra 35 HRC

Fig.5.84 Training Error Plots for Ra (Target vs Output)

Fig.5.85 Testing Error Plots for Ra (Target vs Output)

Fig.5.86 Regression Plots for Ra (Train /Test)

Fig.5.87 Response Surface Plot for Ra

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 172


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.5.3 PSO based ANFIS (FCM) Plots For Ft 35 HRC

Fig.5.88 Training Error Plots for Ft (Target vs Output)

Fig.5.89 Testing Error Plots for Ft (Target vs Output)

Fig.5.90 Regression Plots for Ft (Train /TesT)

Fig.5.91 Response Surface Plot for Ft

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 173


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.5.4 PSO based ANFIS (FCM) Plots For Fa 35 HRC

Fig.5.92 Training Error Plots for Fa (Target vs Output)

Fig.5.94 Regression Plots for

Fig.5.93 Testing Error Plots for Fa (Target vs Output)

Fa (Train /Test)

Fig.5.95 Response Surface Plot for Fa

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 174


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.5.5 PSO based ANFIS (FCM) Plots For Fr 35 HRC

Fig.5.96 Training Error Plots for Fr (Target vs Output)

Fig.5.97 Testing Error Plots for (Target vs Output)

Fig.5.98 Regression Plots for Fr

(Train /Test/)

Fig.5.99 Response Surface Plot for Fr

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 175

Fr


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.5.6 PSO based ANFIS (FCM) Plots For Tf 35 HRC

Fig.5.100 Training Error Plots for Tf (Target vs Output)

Fig.5.101 Testing Error Plots for Tf (Target vs Output)

Fig.5.102 Regression Plots for Tf (Train /Test)

Fig.5.103 Response Surface Plot for Tf VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 176


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.5.7 PSO based ANFIS (FCM) Plots For Ra 45 HRC

Fig.5.104 Training Error Plots for Ra (Target vs Output)

Fig.5.105 Testing Error Plots for Ra (Target vs Output)

Fig.5.106 Regression Plots for Ra (Train /Test)

Fig.5.107 Response Surface Plot for Ra VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 177


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.5.8 PSO based ANFIS (FCM) Plots For Ft 45 HRC

Fig.5.108 Training Error Plots for Ft (Target vs Output)

Fig.5.109 Testing Error Plots for Ft (Target vs Output)

Fig.5.110 Regression Plots for Ft (Train /Test)

Fig.5.111 Response Surface Plot for Ft VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 178


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.5.9 PSO based ANFIS (FCM) Plots For Fa 45 HRC

Fig.5.112 Training Error Plots for Fa (Target vs Output)

Fig.5.113 Testing Error Plots for Fa (Target vs Output)

Fig.5.114 Regression Plots for Fa (Train /Test)

Fig.5.115 Response Surface Plot for Fa VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 179


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.5.10 PSO based ANFIS (FCM) Plots For Fr 45 HRC

Fig.5.116 Training Error Plots for (Target vs Output)

Fr

Fig.5.118 Regression Plots for

Fig.5.117 Testing Error Plots for Fr (Target vs Output)

Fr (Train /Test/Validate

Fig.5.119 Response Surface Plot for Fr VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 180


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.5.11 PSO based ANFIS (FCM) Plots For Tf 45 HRC

Fig.5.120 Training Error Plots for vs Output)

Tf (Target

Fig.5.121 Testing Error Plots for Tf (Target vs Output)

Fig.5.122 Regression Plots for Tf (Train /Test

Fig.5.123 Response Surface Plot for Tf VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 181


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.6 Comparison of Prediction Results with Experimental Statistics 5.6.1 NSGA-NN 5.6.1 (a) Statistical Comparison of

Prediction Results with Experimental

Statistics for AISI 4340 Steel 35HRC and 45HRC NN35 Error Error MSE NSGA-NN VS Experimental STD(σ) 35 HRC Ra --0.221 0.979 0.960 Ft (N) 42.67 86.21 8.88*103 Fa (N) 156.2 49.56 2.675*104 Fr (N) -0.723 14.77 206.55 PSO NN VS Experimental 35 HRC Ra 1.749 0.732 3.56 Ft (N) -53.25 104.6 1.32*14 Fa (N) 138.22 40.44 2.06*104 Fr (N) 40.08 64.63 5.57*103 ANFIS GA VS Experimental 35 HRC Ra -0.1572 0.4507 0.2177 Ft (N) 17.6 57.06 3.4*103 Fa (N) 136.152 51.87 2.109*104 Fr (N) 11.065 20.54 523 ANFIS PSO VS Experimental 35 HRC Ra 0.1578 0.4494 0.2168 Ft (N) 17.13 57.16 3.398*103 Fa (N) 135.9 52 2.11*104 Fr (N) 10.89 70.79 517

RMSE Error Error MSE RMSE NSGA-NN VS Experimental STD(σ) 45 HRC 0.980 -0.355 0.378 0.262 0.512 94.24 -57.23 128.36 1.89*104 137.58 114.49 -13.571 38.046 1.55*103 39.488 14.37 -11.74 35.99 1.36*103 36.99 PSO NN VS Experimental 45 HRC 1.88 115.04 43.73 74.66 ANFIS GA VS Experimental 45 HRC 0.466 0.0371 0.2796 0.075 0.2751 4 58.3 -21.84 107.08 1.64*10 107.76 145.2 -45.95 41.46 3.74*103 61.19 22.87 -0.91 36.86 1.29*103 35.93 ANFIS PSO VS Experimental 45 HRC 0.4656 0.0372 0.2798 0.075 0.275 58.29 -27.34 106.19 1.14*104 107.05 145.4 -46.83 41.55 3.83*103 61.91 22.75 -0.747 36.77 1.28*103 35.84

5.6.2 Error Plots of NSGA-NN Prediction Results with Experimental Statistics for AISI 4340 Steel 35hrc

Fig.5.124 Error Estimation Plots For R a

Fig.5.125 Error Estimation Plots For F t

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 182


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.5.126 Error Estimation Plots For F a

Fig.5.127 Error Estimation Plots For Fr

5.6.3 Error Plots of NSGA-NN Prediction Results with Experimental Statistics for AISI 4340 Steel 45hrc

Fig.5.128 Error Estimation Plots For R a

Fig.5.129 Error Estimation Plots For F t

Fig.5.130 Error Estimation Plots For F a

Fig.5.131 Error Estimation Plots For F r

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 183


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.6.4 Error Plots of PSO-NN Results with Experimental Statistics for AISI 4340 Steel 35 HRC

Fig.5.132 Error Estimation Plots For R a

Fig.5.133 Error Estimation Plots For F t

Fig.5134 Error Estimation Plots For F a

Fig.5.135 Error Estimation Plots For F r

5.6.5 Error Plots of GA based ANFIS (Fuzzy C-Mean Clustering)) Results with Experimental Statistics for AISI 4340 Steel 45hrc

Fig.5.136 Error Estimation Plots For R a

Fig.5.137 Error Estimation Plots For F t

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 184


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.5.138 Error Estimation Plots For F a

Fig.5.139 Error Estimation Plots For F r

5.6.6 Error Plots of GA based ANFIS (Fuzzy C-Mean Clustering)) Results with Experimental Statistics for AISI 4340 Steel 45hrc

Fig.5.140 Error Estimation Plots For R a

Fig.5.141 Error Estimation Plots For F t

Fig.5.142 Error Estimation Plots For F a

Fig.5.143 Error Estimation Plots For F r

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 185


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

5.6.7 Error Plots of PSO based ANFIS (Fuzzy C-Mean Clustering)) Results with Experimental Statistics for AISI 4340 Steel 35hrc

Fig.5.144 Error Estimation Plots For R a

Fig.5.145 Error Estimation Plots For F t

Fig.5.146 Error Estimation Plots For F a

Fig.5.147 Error Estimation Plots For F r

5.6.8 Error Plots of PSO based ANFIS Prediction Results with Experimental Statistics for AISI 4340 Steel 45hrc

Fig.5.148 Error Estimation Plots For R a

Fig.5.149 Error Estimation Plots For F t

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 186


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fig.5.150 Error Estimation Plots For F a

Fig.5.151 Error Estimation Plots For F r

5.7 Conclusion An extensive statistical error analysis of developed prediction model through possible synergies is performed and at all the three learning stages the errors were monitored i.e, while training testing and validating. The learning converged well for all applied techniques. Furthermore the predicted results are tested against experimental statistics for evaluating the prediction accuracy of each learning model. The prediction models demonstrated relatively varying accuracy results which are tabulated in each sections. Following crucial observations can be made through statistical error analysis. 1. The results of EA –NN were more accurate than SI-NN and less accurate when compare to the synergies of ANFIS. 2. EA-NN accuracy was convincing but the accuracy of ANFIS-EA and ANFIS-SI were better than the EA-NN and SI-NN prediction models which can be clearly commented from the statistical error table. 3. Between the ANFIS –EA and ANFIS-SI the relative difference of accuracy is negligible as both the techniques demonstrated almost similar results in prediction 4. The Adaptive neuro fuzzy combination proved to be better than the neuro computing combination this difference is possible due to the adaptive layers introduced in the neuro-fuzzy inference. Though in both the techniques back VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 187


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

propagation learning algorithm is applied for error minimization in learning, the adaptive layers and clustering technique in membership function of fuzzy structure improved the learning ability of prediction model. 5. However the synergism of neural network gave better accuracy than the exclusive techniques exhibiting improvement in learnability of pattern

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 188


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

CHAPTER 6 RESULTS AND DISCUSSION The results from optimization and predictive techniques are elaborately discussed and comparison with experimental statistics is made. 6.1 Global Search Optimization Global Search optimization technique are explained extensively in chapter 3 and implemented strategy is shown in fig.3.1 both techniques from EA and SI were utilized to optimize machining performance 6.1.1 NSGAII NSGA II utilized non-domination technique for its population selection and mating, the fitness of individuals were determined by two criteria (a) Pareto-Individual rank (b) Pseudo Euclidean distance According to criteria an individual with least rank and maximum Euclidean distance is the fittest and these individual/individuals represents globally optimized solutions. The degree of optimality depends on their relative function of rank and distance which represents diversity of individuals in each generation. The results of both the steels AISI 4340 35HRC and 45HRC had initial population of 1000 and elitism was applied at each generation. The Table 3.3 and Table 3.4 records results of family of optimized individuals for both steels after hundred generations arranged in descending order of fronts and fitness levels which are calculated by rank and distance. Only the first 20 solutions are listed out of the 1000 population in the tables. 6.1.1 (a) For AISI 4340 35HRC From Table 3.3 the first front has maximum distance and minimum rank the solutions and these solutions shows a good tradeoffs between the surface roughness and tool life among them fifth individual with fitness which gives both good surface finish and VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 189


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

maximum tool life .This individual corresponds to process parameters (cutting speed ,feed rate and depth of cut) 170m/min, 0.15mm/rev, 1mm respectively with machining objectives(Ra, Ft, Fa, Fr, Tf) 4.0µm, 412N, 152N, 231N and 50 minutes. The subsequent individual with process parameter (cutting speed, feed rate and depth of cut) 231m/min,0.15mm/rev,1mm corresponding to machining performance (Ra, Ft, Fa, Fr, Tf) 2.5µm, 442N, 115N, 199N and 42.49 minutes can be picked out as optimal solution. Though the fitness of first front is fittest among others, the individuals from subsequent front can also be chosen from the solution space with permitted tradeoffs. The Fig.3.3-3.5 constitutes plots of NSGA II for 35HRC, in the Fig.3.3 the first subplot shows the rank of individual at each generation, from the figure it can be inferred that the rank converges to one for most of the population as non-dominated technique is applied for sorting, in the second subplot of Fig 3.3 Pareto plots between the objectives are plotted to evaluate relative tradeoffs between two objectives. Pareto front depicts plots for elite members which can be between two or three objectives that are non-inferior. The subplots 2, 3 and 4 in the figure 3.3 are pareto plots between force and surface roughness, the second subplot represents pareto-front between Ra-Ft, in third and fourth sub plot pareto front between Ra-Fa and Ra-Fr. The Fig 3.4 contains pareto-fronts between tool life and forces; first subplot is paretofront between Tf -Ft, second subplot between Tf-Fa and third sub plot between Tf-Fr. From the pareto fronts for minimum surface roughness and maximum tool life following region range of forces were found to be favorable (approximately) optimal surface profile. Table.6.1 Tradeoffs among forces for Surface roughness and Tool life Forces Tangential Forces Axial Forces Radial forces

Surface Roughness(1-2.5 µm) 800-600 N 300-200 400-300

Tool life (Tf>40 mins) <470N <160N <250N

In fig. 3.5 average pareto diversity plots in the consecutive generations is plotted which is determined by average pareto (Euclidean) distance among individuals The diversity of

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 190


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

population varied from 1-0.001 The population in overall generations exhibited vivid diversity. Initially the average distance between the generations was maximum after 20 generations a shift in distance was observed with distance spread to 0.6 in mid generation and as final generations are approached the distance between generations converges to 0.001 showing a good migration of individuals across generations. 6.1.1 (b) For AISI 4340 45HRC For AISI 4340 45 HRC steel the family of solutions is listed in table 3.4 out of the 1000 elite solutions only 20 solutions are tabulated. The best individual in the first front is corresponding to process parameters (cutting speed, feed rate and depth of cut) 130.8m/min, 0.15 mm/rev, 1mm and objective fitness (Ra, Ft, Fa, Fr, Tf) 4.0 Âľm, 542N, 331N, 282N, 30 minutes is fittest. Similarly preceding individual with process parameters (cutting speed, feed rate and depth of cut) 135m/min, 0.15mm/rev, 1mm corresponding to objective fitness (Ra, Ft, Fa, Fr, Tf) 3.98, 519N, 331N, 282N, 28 minutes can be picked as best solution. Likewise favorable solutions can be chosen with the degree of tradeoffs permitted among the objectives. The Fig.3.6 - 3.8 are plots of NSGA II for 45HRC, in the Fig.3.6 the first subplot depicts rank of individual at each generation and the rank converges to one for most of the population, in the second subplot of Fig.3.7 Pareto plots between the objectives are graphed and in Fig.3.8 diversity plot at each generation is drawn. The subplots 2, 3 and 4 in the figure 3.6 are Pareto plots between surface roughness and force, the second subplot represents pareto-front between Ra-Ft, the third and fourth sub plot shows pareto front between Ra-Fa and Ra-Fr. The Fig.3.7 contains pareto-fronts between tool life and forces; first subplot is paretofront between Tf-Ft, second subplot between Tf-Fa and third sub plot between Tf-Fr. From the pareto fronts for minimum surface roughness and maximum tool life following region range of forces were found to be favorable (approximately) for optimal solution in 45HRC machining.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 191


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Table.6.2 Tradeoffs among forces for Surface roughness and Tool life Forces Tangential Forces Axial Forces

Surface Roughness(3.5-4µm) 600-400 N 360-300N

Tool life (Tf>10 mins) <500N <160N

Radial forces

300-330N

<350N

In Fig.3.8 diversity of generations is plotted, at initial generations the average pareto distance was maximum latter on at mid generations a sharp drop is observed between 2535 generations and then at final generations the distance converged to as low as 0.001 showing good migration ability and diversity. 6.1.2 Results for SPEA 2 6.1.2 (a) For AISI 4340 35 HRC Since SPEA 2 is a combination of pareto envelope and niche–pareto the rank and fitness strength is calculated to both best elements in archive and individuals in population through dominance level. The proposition rule of determining best solution is by calculating effective fitness value which is sum of raw fitness and density distribution about individuals. Table 3.6 holds optimal solution for AISI 4340 steel 35 HRC which contains 300 archive elements. The matrix F contains the fitness strength of each individual which calculated by summation of R (rank) and D (distance /density function) matrix (in table 3.6), the distance D is calculated by K nearest cluster algorithm .The matrix S is strength value which is evaluated by dominance count. The position matrix represents process parameters and the cost matrix represents machining objectives. The individual with least strength value represents the fittest individual i.e.,

the

individual in first front with strength fitness S 0.32 can be picked as best individual with position (process parameters :cutting speed, feed rate and depth of cut) 259m/min, 0.17mm/rev, 1.07mm corresponding to machining objectives (Ra, Ft, Fa, Fr, Tf) 1.88µm, 514N, 136N, 225N, 35.51 minutes.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 192


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

In Fig.3.10 subplots have individual ranks and pareto fronts between surface roughnessforces, the subplots 2, 3, 4 depicts the pareto front between Ra-Ft, Ra-Fa, Ra-Fr, which are archive element containing only fittest individual. Similarly the pareto fronts between tool life and forces are plotted in fig.3.11 the subplots 1,2,3 are pareto fronts for Tf- Ft, Tf- Fa, Tf- Fr are represented. Though the solutions suggested by SPEA 2 is different individuals from the NSGA II the force constraint obtained is same as NSGA II for minimum surface roughness and maximum tool life Table.6.3 Tradeoffs among forces for Surface roughness and Tool life Forces Tangential Forces Axial Forces Radial forces

Surface Roughness(1-2.5µm) 780-590 N 300-200N 300-330N

Tool life (Tf>35 mins) <500N <160N <250N

In Fig.3.12 diversity among generation is plotted the average pareto distance is different from the NSGA II and the spread is limited between 0.37-0.27 the spread is random representing a good mix in individual among generations but the combination is restricted between limited distance. 6.1.2 (b) For AISI 4340 45 HRC The results of 45 HRC steel is recorded in the Table 3.7 with same definition of elements. From the table the strength fitness is least for third individual with F 0.31 the position (process parameters: cutting speed, feed rate and depth of cut) of this individual is 174 m/min,0.23 mm/rev, 1.7 mm with machining performance (Ra, Ft, Fa, Fr, Tf) 3.93µm, 674N, 389N, 355N, 14.19 minutes and next fittest individual is in consequent front with strength fitness S 0.33 with process parameters168m/min,0.16mm/rev,1.02mm for machining performance 3.67µm,519.42N,384N,310N,19 minutes. Pareto fronts for 45 HRC objectives are in Fig.3.13 and Fig.3.14,in fig 3.13 the sub plot 2,3,4 represents the pareto front between Ra-Ft, Ra-Fa, Ra-Fr, and subplots in Fig.3.14 shows subplots of Tf-Ft, Tf - Fa, Tf- Fr respectively for minimum surface roughness and maximum tool life the force constraints should be as follows

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 193


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Table.6.4 Tradeoffs among forces for Surface roughness and Tool life Forces Tangential Forces Axial Forces Radial forces

Surface Roughness(3.5-4Âľm) 600-400 N 360-300N 300-330N

Tool life (Tf>10 mins) <500N <160N <350N

In Fig.3.15 diversity among generation is determined the average pareto distance is limited between 0.4-0.26 the spread is random representing a good combinations in individual among generations but this trend is limited. 6.1.3 Results for PSO 6.1.3 (a) For AISI 4340 35HRC The PSO algorithm utilizes swarm movement to find optimal solution. The applied swarm constitutes of particle structure with position, cost and their best position, best cost associated with it a pseudo velocity and acceleration associated for each particle. Table 3.9 comprises of optimal solution for AISI 4340 steel where the position matrix holds process parameters cost holds the machining performance for each position best cost ,best position is defined for each position. The velocity matrix is which correspond to the position matrix is utilized to move swarm in search space and changes at each transition. The Grid Index matrix contains the topology of swarm at each transition and grid sub index contains the neighborhood topology of grid index. The table consists of 500 swarm particles which are in repository element out of 1000 swarm. In Fig.3.17 pareto front between the surface roughness and tangential force is graphed which gives similar results as pareto front of Evolutionary Techniques. Fig.3.18-3.19 are swam surfaces for surface roughness and tool life respectively which shows the potential of swarm. Each particle on the swarm surface is defined by particle position in Fig.3.18 the particles at the declination of surface are fittest particles which converged to minimum surface roughness values and in Fig.3.19 the particles at the projection of swarm surface represents the optimal tool life showing convergence at the foot of swarm surface.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 194


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

In Fig.3.20 the influence of depth of cut on force component on optimal particles is plotted. The subplots 1, 2 and 3 show the force gradients with respect to depth of cut for best particles. The change in force about the mean line is shows that best particles are about the mean line presenting favorable solutions for surface roughness and tool life. From the table 3.9 optimal solution is determined by leaders among swarm the first leader corresponding is to gird index 40282 with best position 193.76m/min,0.15mm/rev, 1 mm with machining performance (Ra, Ft, Fa, Fr, Tf) 3.43 Âľm, 417N, 134N, 215N, 46.12 minutes the subsequent solution is corresponding to particle with grid index 46844 similarly other solutions are equally fit depending on the permitted tradeoffs solutions can be chosen. 6.1.3 (b) For AISI 4340 45HRC The solution for AISI 4340 45 HRC steel has same entities which are listed in table 3.10.The pareto-fronts and swarm surface are plotted in fig 3.21 and fig 3.22-3.23 respectively. The swarm surface of 45HRC showed similar trends as to 35 HRC steel the 45HRC steel shows similar trends in swarm surface. In Fig.3.21 the pareto front between surface roughness and tangential force is drawn which gives force constraints similar to NSGA II 45 HRC fronts .In fig 3.22 and fig 3.23 the swarm surface for surface roughness and tool life is drawn. In fig 3.24 the influence of depth of cut on force component on optimal particles is plotted. The subplots 1, 2 and 3 show the force gradients with respect to depth of cut for best particles.. From the table 3.10 optimal solution is determined by swarm leaders among them the first leader corresponds to gird index 20510 with best position 136 m/min, 0.15mm/rev, 1mm with machining performance (Ra, Ft, Fa, Fr, Tf) 4.04 Âľm, 522N, 134N, 331N, 29.32 minutes the subsequent solution is corresponding to particle with grid index 31070 similarly other solutions are equally fit depending on the allowed tradeoffs solutions can be chosen.3

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 195


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

6.1.4 Comparison between EA and SI The Fig.3.25 - Fig.3.30 are plots of solution spectrum obtained from applied global search optimization technique. Each plot has solutions for five objectives i.e, Surface roughness, tool life, radial force, axial force and tangential force. In fig 3.25 and fig 3.26 the solution space obtained from NSGA II for 35 HRC and 45 HRC steels is drawn. Clearly distinction can be made the graphs that in 35 HRC steel radial forces are dominating axial force while in 45 HRC steel axial force is dominating radial force. A similar trend can be inferred from solutions of SPEA2 and PSO. Fig 3.27 and Fig 3.28 are solution space obtained from the PSO technique for 35 HRC and 45 HRC steel respectively and Fig.3.29 and Fig.3.30 are solution space obtained from the SPEA2 for 35HRC and 45HRC respectively. For 35 HRC the sequence of solution for each objectives are represented in following order surface roughness, tool life, axial force, radial force and tangential force ,while for 35 HRC the solution for each objective is in order of surface roughness, tool life ,radial force, axial force and tangential force. 6.1.4 (a) Comparison among the Solution spectrum The solution space of NSGA II (fig 3.25 & 3.26) has thousand populations which is a combination of parents and offspring formed by tournament selection. The solution space is distributed with constant amplitude in solution space. While the solution space of PSO and SPEA2 (fig.3.27-fig.3.30) vary in amplitude in solution space at each generation. The solution space of PSO (Fig 3.27 & fig 3.28) has five hundred best solutions which are elite solutions. The solution trend is random when compared to NSGAII showing varying amplitude across wavelength with immediate change in crest and troughs in local and global minima and maxima. However the solution space of SPEA2 has three hundred best elite solutions which are highly disturbed amplitude when compared to NSGAII and PSO the jumps in maxima and minima is uneven exhibiting unsaturation in local and global minima . VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 196


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

From the solution space analysis the NSGAII showed a better saturation in local global maxima minima while PSO should a moderate saturation and SPEA2 exhibited lower saturation levels in maxima and minima. The saturation levels also depicts tradeoffs in objectives from which an following inference can be made. The solution obtained from the NSGAII solution gave better tradeoffs among objectives which can be observed from each front in solution space. While PSO gave a moderate trade off with the solution space was favoring lower surface roughness by reducing tool life for few swarm particles though the overall trade off was found to be good enough. The solution space of SPEA2 gave unbalance tradeoff between objectives favoring surface roughness and reducing tool life for many individuals. From above discussion it can be concluded that the solution space provided by NSGAII is better than other two while the solution space provided by the PSO is better than SPEA2. 6.1.5 Evident from the literature The solutions obtained by fittest individuals in NSGAII and PSO keeps the flank wear of tool under working limit. The tool used for current hard turning was multi-layer coated carbide inserts and the cutting speed suggested by both the optimization techniques suggested cutting speed below 200 m/min and feed rate, depth of cut in between LFLD (low federate low depth of cut) and HLHD (high feed high depth of cut) at this condition the flank wear is less than 0.15 mm for tool life greater than 40 minutes [2]. This cutting condition keeps flank wear under appreciable level, restricting sharp rise in cutting forces due to high flank wear rate [3]. If the flank wear is inhibited above this then machining chatter due to excessive forces is controlled and better surface finish is obtained. In contrast SPEA2 suggests cutting speed close to 260m/min (250-260 m/min) and feed rate, depth of cut near (HLHD condition) for which flank wear about 0.2mm for time cutting greater than 35 minutes.[2]. At this flank wear the forces tend to increase sharply which may increase machining vibrations resulting in poor surface finish. The tradeoff between the surface roughness and tool life should be picked wisely for a successful hard turning. VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 197


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

6.2 Intelligent Learning Techniques Intelligent learning systems were applied to recognize machining pattern sequence which is discernible in the machining statistics. In fig 4.1 the adopted strategy for chapter 4 is illustrated and accordingly the results were discussed. Learned networks were utilized to predict machining performance on experimental runs and obtained results were analyzed for statistical error. An extensive mathematical framework is discussed for the all the applied learning techniques. 6.2.1 Neural Network The applied network architect for both steels is illustrated in fig 4.4. A multi-layer feed forward perceptron type network was used to for both steels which require many exemplars for mapping multiple vectors. From the regression model thousand machining data were generated for each vector model for better interpolation of target vectors in learning space. These data sets were randomly permutated and split for training, testing and validating samples. Out of the thousand sets 700 samples were utilized for training 150 sample for training and 150 for validation. The description of applied neural network for AISI 4340 steel 35 HRC is listed in table 4.2.4 (a) and the calibrated weights and bias for the targets is given in table 4.2.4 (b). The network performance plot in Fig.4.5 shows drop in mean square error from 104 to 10-3 in 1000 epoch while learning (training, testing and validating stage). The error gradient Fig.4.6 between the targets and output converged to 10-2 at 1000 epoch from gradient of 105. For each objective an error histogram and regression plots were drawn in (Fig. 3.7Fig. 3.16) in each error plot the mean error was close enough to the zero error line depicting that the error minimized to its least possible value. Linear regression fit was obtained between targets and output while learning each objective with regression co efficient close to 1. Same network was utilized for leaning machine statistics of 45HRC steel. The network description and calibrated weights and bias are listed in Table 4.2.5(c) and 4.2.5(d).The mean square error while learning converged to similar order (Fig.5.17) to that of 35HRC. VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 198


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

While in training state the error gradient gave sharp slopes though at final epoch the gradient converged to 1.125 with zero validation fails. The error histogram and regression plot for each vector is in (Fig.4.19-Fig.4.28).The error for each machining vector was in order of 10-2. 6.2.2 Adaptive Neuro Fuzzy Interference Technique In this network two adaptive layers with learnable parameters were used viz., antecedent and consequent layer. Three different clustering techniques were applied on membership function i.e., Grid portioning clustering, Subtractive clustering and Fuzzy C- mean clustering. Each cluster techniques had different weights (connections) and the fuzzy structure of each technique varied. The analytical description of hybrid neuro-fuzzy techniques is discussed in brief. For each machining vector fuzzy structure learns machining data sequentially. From the RSM model thousand machining data for learning were generated. These vectors were randomly permutated and split into training (850 samples), testing (150 samples) and validated on complete machining vector. 6.2.2 (a) ANFIS Grid partition The applied structure is illustrated in fig.4.30 and fuzzy structure implemented is tabulated in table 4.3.4 (a) fuzzy structure. For each machining vector following plots were drawn. (i)

Simultaneous plots between expected target and output.

(ii)

Errors between the target-output.

(iii)

Normal density fit between mean error and standard deviation.

At each epoch the error between targets and output while learning was calculated and error gradients were minimized through hybrid leaning and back propagation algorithm. In table 4.3.4 (b) statistical errors in training, testing and validating for each objective in 35HRC is listed and mean error for each vector were as low of order 10-5 these statistical results are plotted in Fig.4.36-Fig.4.60.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 199


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Same architect and fuzzy structure is utilized for learning machining pattern for 45 HRC. The statistical error in training, testing and validation for each objective is tabulated in table 4.3.4 (c) the error for was as low as low as 10-5 and plots for each objective are illustrated in between Fig.4.61-Fig.4.85. 6.2.2 (b) ANFIS Subtractive Clustering The developed structure for subtractive clustering is illustrated in fig.4.86, fuzzy structure implemented is tabulated in table 4.3.16 (a) and in table 4.3.16 (b) &4.3.16 (C) statistical errors in training, testing and validating for objective in 35 HRC and 45HRC respectively is listed and these statistical results are plotted for each objective. In comparison to grid partition clustering the mean error in subtractive clustering is reduced to order of 10-7.for 35 HRC and 10-6 in 45 HRC showing improvement in leaning ability. The plots for each objective and their corresponding figures for 35HRC and 45HRC steel acquired from ANFIS subtractive Clustering are illustrated in figures Fig.4.87-Fig.4.136. 6.2.2 (c) ANFIS Fuzzy C-mean Clustering The developed structure for FCM clustering is shown in fig.4.137 and fuzzy structure implemented is tabulated in table 4.3.27 (a). In table 4.3.27 (b) & 4.3.27 (C) statistical errors in training, testing and validating for both steels is listed. The learning error in FCM was further reduced to order of 10-8 in both the steels though not for all vectors The mean error in all vectors was of similar order (10-6) showing similar results as that of subtractive clustering. The plots for each objectives and their corresponding figures for 35HRC and 45HRC steel obtained from ANFIS FCM are shown in Fig.4.138-Fig.4.187. 6.2.3 Comparative Evaluation of the predictive technique on Experimental statistics The developed predictive model was tested on Experimental statistics and a statistical analysis of the errors in predictions was made for each model. Comparison graphs for each objective in each technique were plotted (fig.4.188-fig.4.219). The Mean errors and

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 200


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

their standard deviations of each objectives in both the steel are listed in table (6.5) and the means squared error (MSE), root mean squared error (RMSE) is tabulated in table. Table.6.5 Mean error and Standard Deviation between Experimental and Predicted statistics Machining Objective

Neural Network

ANFIS Grid

Error Measure

ANFIS Sub

ANFIS FCM

Neural Network

Mean Error

ANFIS Grid

ANFIS Sub

ANFIS FCM

Standard Deviations

Comparison Errors for 35 HRC AISI 4340 Steel 0.0442 -0.2308 -0.226 -0.2339 1.0814 Ra(µm)

0.5964

0.5964

0.5964

Ft (N)

0.186.6

17.49

15.29

18.19

126.4

106.9

105.36

104.9

Fa (N)

103.98

157

157.79

157

33.77

80

82.3

79.42

Fr (N)

1.1318

32.57

27.89

32.23

93.95

41.29

40.81

42.20

Comparison Errors for 45 HRC AISI 4340 Steel Ra(µm)

-0.1700

-0.9278

-0.947

-0.856

0.2769

0.5964

0.594

0.596

Ft (N)

-57.23

-22.35

-23.106

-21.7

128.3

77.58

76.64

78.8

Fa (N)

54.67

-7.5709

-6.9912

-6.8003

33.57

40.08

39.58

38.45

Fr (N)

0.415

-7.69

-8.525

-6.63

14.0720

38.35

36.57

36.89

Table.6.6 MSE and RMSE between Experimental and Predicted statistics Machini ng Objectiv e Error Measure

Neural Networ k

ANFIS

ANFIS

ANFIS

Grid

Sub

FCM

Mean Square Error

Neural Network

ANFIS

ANFIS

ANFIS

Grid

Sub

FCM

Root Mean Square Error

Comparison Errors for 35 HRC AISI 4340 Steel Ra(µm) 1.1130 0.3912 0.3875 0.392

1.0550

0.6255

0.6225

0.6266

Ft (N)

594.81

1.116*104

1.07*104

1.08*104

24.38

105.66

103

103.9

Fa (N)

1.89*10

3.09*104

3.13*104

3.07*104

109.07

176

177.016

175.32

4

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 201


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fr (N)

2.6*103

93.95

2.36*103

2.73*103

9.69

51.78

48.58

52.25

Comparison Errors for 45 HRC AISI 4340 Steel Ra(Âľm)

0.1018

1.198

1.23

1.0714

0.3190

1.0949

1.109

1.0351

Ft (N)

1.89*10

6.128*103

6.11*103

6.38*103

137.5

78.85

78.19

79.8

Fa (N)

4.06*10

1.58*103

1.53*103

6.36*10-6

63.72

39.8

39.213

0.4664

Fr (N)

188.29

1.45*103

1.33*103

1.33*103

13.72

38.17

36.65

36.56

4

3

Though the statistical errors in neural network was relatively in comparison to ANFIS models the relative change in error at each prediction points in neural network is higher than the ANFIS models and the curve traced by neural network in both the steels did not match well when compared to curved traced by ANFIS models. This observation can be inferred by analyzing corresponding comparison graphs for each objectives in specific techniques. The error plots for each technique are drawn in Fig.4.188-4.219. The results obtained from the prediction models were accurate enough to predict around the experimental statistics, though the relative degree of accuracy varied for different learning technique. The mean error of neural network was less than the neuro-fuzzy but the prediction curve could not trace well with the experimental curve in converse the mean error of ANFIS was comparatively larger but the prediction curve traced well with the experimental curve.

6.3 Synergies of CI The combination of optimization techniques were applied to prediction model to further improve results in pattern learning of current machining statistics The combination of synergies applied is shown in Fig.5.1 results of each combination is discussed individually.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 202


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

6.3.1 EA-NN For EA-NN combination NSGA II optimization technique was applied to optimize weights. The applied architect is shown in Fig.5.2. The network description, calibrated weights of combined network is listed in table 5.2.2 (a) and 5.2.3 (b).The optimized populations from NSGAII were given input to neural network at each epoch and depending on drop in error gradients the weights were adjusted and fed to NSGA II for optimization. The optimized solution space (1000 population) of NSGA II are permutated randomly and split into training (700) testing (150) and validating samples. These samples would adjust weights accordingly at every epoch. The mean square error for network converged to order of 10-3 just in 300 epoch exhibiting good convergence in network error. The plot for network performance is in Fig.5.3.The training state plots is illustrated in Fig.5.4 the error gradient dropped to 0.194 which is lesser than the neural network. Likewise the error plots and regression for each vector is in Fig.5.5-Fig.5.15. Same architect is utilized for 45 HRC steel the learning performance was better than the exclusive neural network with mean square error as low as 10-4 in training, testing, validation stages. The performance plot is drawn in Fig.5.16 and the training state plots are in Fig.5.17, error gradients were also found to be minimized to gradient as low as 10 -, error and regression plots for each vector are in Fig.5.16-Fig.5.41. 6.3.2 SI-NN Applied architect for PSO-NN combination is illustrated in Fig.5.29. The best solutions in archive elements of PSO were utilized for network learning. The archive had 500 elements out of which 350 samples were utilized for training, 75 samples for testing and 75 samples for validation. The network description and calibrated weights are tabulated in table 5.3.1 (a) and table.5.3.1 (b). Two hidden layers were defined for network to work accurately since the examples provided by PSO were less compared to NSGA II. For each input and output vectors 20 weights were utilized for mapping and 10 bias were added at each hidden layer. VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 203


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

The network performance of combined SI-NN was better than NN with mean square error dropping as low as 10-4 in all the learning state and error gradient going as low as 10-4 at final epoch. The performance plot of network is illustrated in Fig.5.30 and training state plots are in Fig.5.31.The overall network performance was better than the exclusive neural network. 6.3.3 ANFIS Synergies To enhance ANFIS, GA and PSO are applied to optimize membership function, FCM structure from previously developed model was utilized for combination. The applied strategy utilized for ANFIS-GA is explained in Fig.5.42, the architect utilized is same as that of ANFIS-FCM. Thousand learning samples were generated from RSM model these exemplars were randomly permuted and used for leaning. Out of thousand exemplars 700 were split for training sets and 300 for testing. Since fuzzy maps only one vector per training the multi objective NSGA was reduced to single objective GA with no change in genetic and selection operators. From the error table 5.3.1 (a) the mean error for each vector of 35HRC dropped to order of 10-13 which is way better compared to exclusively applied ANFIS (10 -6) demonstrating excellent improvement in pattern learning ability. The error plots for each vector are drawn in Fig.5.44-Fig.5.64.Similarly for 45 HRC from the Table.5.3.7 mean error for each vector dropped to order of 10-13 exhibiting better leaning than exclusive ANFIS. Developed strategy for ANFIS-PSO technique is shown in Fig.5.84, the thousand samples employed in ANFIS–GA is used for ANFIS-PSO learning. From table 5.4.7 the learning error was reduced to order of 10-14 exhibiting better learning ability The error plots for both the techniques is represented in Fig.5.44-Fig.5.125. 6.3.4 Comparative Evaluation of the predictive technique on Experimental statistics The developed predictive model was tested on Experimental statistics and a statistical analysis of the errors in predictions was made for each model. Comparison graphs for each objective in each technique were plotted (fig.4.188-fig.4.219).The Mean errors and

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 204


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

their standard deviations of each objectives in both the steel are listed in table 6.7 and the means squared error(MSE), root mean squared error (RMSE) is tabulated in table 6.8 Table.6.7 Mean error and Standard Deviation between Experimental and Predicted statistics Machinin g Objective Error Measure

NSG A-NN

PSO-NN

ANFISGA

ANFISPSO

NSGA -NN

Mean Error

PSO-NN

ANFIS ANFIS-GA PSO

Standard Deviation

Comparison Errors for 35 HRC AISI 4340 Steel -0.221 1.749 -0.1572 0.1578 Ra(µm) 42.67 -53.25 17.6 17.13 Ft (N) 156.2 138.22 136.152 135.9 Fa (N) -0.723 40.08 11.065 10.89 Fr (N) Comparison Errors for 45 HRC AISI 4340 Steel -0.355 0.0371 0.0372 Ra(µm) -57.23 -21.84 -27.34 Ft (N) -13.57 -45.95 -46.83 Fa (N) -11.74 -0.91 -0.747 Fr (N)

0.979 86.21 49.56 14.77

0.732 104.6 40.44 64.63

0.4507 57.06 51.87 20.54

0.4494 57.16 52 70.79

0.378 128.36 38.046 35.99

0.4507 57.06 51.87 20.54

0.2796 107.08 41.46 36.86

0.2798 106.19 41.55 36.77

Table.6.8 Mean Square Error and Root Mean Square Error between Experimental and Predicted statistics Machini ng Objectiv e Error Measure

NSGANN

PSO-NN

ANFISGA

ANFIS-PSO

NSGANN

Mean Square Error

Comparison Errors for 35 HRC AISI 4340 Steel 3.56 0.2177 0.2168 Ra(µm 0.960 ) 8.88*103 1.32*104 3.4*103 3.398*103 Ft (N) 4 4 4 2.06*10 2.109*10 2.11*104 Fa (N) 2.67*10 5.57*103 523. 517 Fr (N) 206.55 Comparison Errors for 45 HRC AISI 4340 Steel 0.075 0.075 Ra(µm 0.262 ) 1.64*104 1.14*104 Ft (N) 1.89*104 3.74*103 3.83*103 Fa (N) 1.55*103 3 1.29*103 1.28*103 Fr (N) 1.36*10

PSONN

ANFISGA

ANFI S-PSO

Root Mean Square Error

0.980

1.88

0.466

0.4656

94.24 114.49 14.37

115.04 43.73 74.66

58.3 145.2 22.87

58.29 145.4 22.75

0.512

0.2751

0.275

137.58 39.488 36.99

107.76 61.19 35.93

107.05 61.91 35.84

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 205


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

From table 6.7 the values of mean error and standard deviations of ANFIS combinations were less than the Neural Network synergies. Similarly (table 6.8) the mean square error and root mean square error for ANFIS synergies were lesser compared to Neural Network Synergies. The amplitude of error between experimental and predicted statistics was shorter compared to Neural Network. The ANFIS combination performed better than the NN combination this enhancement is possibly due to the optimization of adaptive layers. Though the statistical error is not very large when compared in magnitude the curve traced by ANFIS synergies were found to better than the curve traced by NN synergies. The error plots for each technique are illustrated in Fig.5.126-Fig.5.153. In brief the combinations of ANFIS performed better than the combinations of Neural Network. However the developed prediction techniques were accurate enough to learn experimental machining statistics.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 206


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

CHAPTER 7 CONCLUSION On an overview the work dealt with optimizing machining performance in hard turning of AISI 4340 steel and building accurate predictive models by applying intelligence learning techniques. 7.1 Optimization Trends 

From the results of optimization the best tradeoffs among machining objectives was obtained by NSGA II solution space, followed by PSO while solution space of SPEA2 was biased toward surface roughness recommending lower tool life in most of its best solution.

The difference in tradeoffs recommended by algorithms is due to the degree of elitism demonstrated by algorithms Elitism is similar to a pseudo memory that is associated to algorithm by which it recognizes the best individual so that it doesn’t search for same individual in again in consecutive reducing exploration time for searching best individual.

The elitism in NSGA II was implemented by tournament selection where consecutive best individuals selected at each tournament are preserved by replacement of chromosome in intermediate solution. Elitism was performed on single set of population i.e. both fittest individuals and intermediate chromosomes were members of chromosome. Though achieving elitism by single set of chromosome slowed down the convergence while recursion. This drawback was overcome by providing randomness in genetic operator. The mutation operator acted as agent of entropy in solution space when elite individuals enveloped to local minimums.

While in PSO the elitism was carried out by introducing pbest and gbest population sets where the local minimum were stored in pbest and global minimum in gbest. The repository element in PSO was elite members which contained gbest populations. The elite mechanism in PSO was dynamic compared to NSGA II in VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 207


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

PSO, the gbest and pbest attractor in neighborhood topology acted as initiator for gathering elite members. The problem of local minimum was overcome by the swarm movement equation Unlike NSGA, PSO doesn’t have genetic operator so when the size of elite members exceeded inertia damping co efficient controlled it, for better convergence a pseudo mutation on particles were introduced. 

Similarity in solution space in NSGA II and PSO could be justified by finding analogies between both algorithms. The best parents in NSGA acted as pseudo particle attractor in recognizing elite members which is similar to the gbest and pbest attractor in grid topology and the genetic operator is similar to that of swarm movement operator Since the elite members in both the algorithms were same (i.e., 500) which justly supports analogy.

But when solution space of SPEA 2 is compared the fittest individuals is quite different from NSGA and PSO algorithm it shows inclination towards one vector weakening other vector even though elitism is fairly applied. This behavior is drawn from the niche behavior of fit individuals in archive (elite members).

7.2 Prediction Trends 

Two variants of learning techniques were applied for recognizing pattern in machining statistics. Initially neural network and ANFIS were applied exclusively for learning machining examples. The accuracy of learnt networks were tested on experimental statistics both techniques gave similar results in prediction. Neural network gave relatively lower errors in comparison to ANFIS.

Among the different clustering techniques used in ANFIS the FCM clustering technique gave relatively lesser errors in learning hence demonstrating better leaning ability than among three techniques. To further improve the learning ability combination of optimization and learning techniques were utilized Both EA and SI coupled learning techniques were applied.

For both the leaning techniques (NN & ANFIS) collaborative combination was utilized. The optimization technique was secondary technique while the leaning technique was primary. In neural network the NSGA II and PSO optimized were

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 208


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

used to optimize the weights in hidden layer and reduce the error between targets and outputs the prediction results obtained from the NSGA-NN were better in comparison to PSO-NN. 

The ANFIS combination Technique exhibited better learning trends than the ANFIS applied exclusively. Both the combinations ANFIS-GA and ANFIS-PSO gave better results in comparison to experimental statistics. The accuracy of both techniques was similar with minor difference in learning error. In comparison to neural network synergies the ANFIS synergies gave better results on experimental statistics and illustrated enhanced learning.



To summarize the combined predictive models performed better in comparison to the exclusive techniques and in optimization techniques the NSGA II and PSO algorithm gave relatively good tradeoffs in MOOPS when compared to SPEA2 algorithm.

Future Scope The unexplored Meta heuristic techniques can be applied for much better tradeoffs in MOOPS and further enhancement in learning technique is possible by applying further introducing stronger coupling among the prediction and optimization techniques. The applied strategy for current machining system can be generalized to other conventional and non-conventional machining systems.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 209


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

REFERENCES [1] Andries P. Engelbrecht Computational Intelligence An Introduction John Wiley & Sons Ltd ISBN 978-0-470-03561-0. [2] Satish Chinchanikar, S.K. Choudhury, Effect of work material hardness and cutting parameters on performance of coated carbide tool when turning hardened steel: An optimization approach Measurement 46 (2013) pp. 1572–1584. [3] Wear behaviours of single-layer and multi-layer coated carbide inserts in high speed machining of hardened AISI 4340 steel

Journal of Mechanical Science and

Technology 27 (5) (2013) 1451~1459 [4] Satish Chinchanikar, S.K. Choudhury, Hard turning using HiPIMS-coated carbide tools: Wear behaviour under dry and minimum quantity lubrication (MQL) Measurement 55 (2014) 536–548. [5] Satish Chinchanikar, S.K. Choudhury Investigations on machinability aspects of hardened AISI 4340 steel at different levels of hardness using coated carbide tools Int. Journal of Refractory Metals and Hard Materials 38 (2013) 124–133. [6] Gaurav Bartarya S.K. Choudhury State of the art in hard turning International Journal of Machine Tools & Manufacture 53 (2012) 1–14. [7] Satish Chinchanikar, S.K. Choudhury Machining of hardened steel Experimental investigations performance cooling techniques: A review International Journal of Machine Tools & Manufacture 89 (2015) 95–109. [8] Satish Chinchanikar, S.K. Choudhury Cutting force modeling considering tool wear effect during turning of hardened AISI 4340 alloy steel using multi-layer TiCN/Al2O3/TiN-coated carbide tools Int. J. Adv. Manuf. Technol. (2016) 83:1749– 1762 DOI 10.1007/s00170-015-7662-5 [9] Predictive modeling for flank wear progression of coated carbide tool in turning hardened steel under practical machining conditions

Int. J Adv. Manuf. Technol.

(2015) 76:1185–1201 DOI 10.1007/s00170-014-6285-6 [10]

T. G. Ansalam Raj & V. N. Narayanan Namboothiri, An Improved genetic

algorithm for the prediction of surface finish in dry turning of SS 420 materials Int J Adv Manuf Technol (2010) 47:313–324 DOI 10.1007/s00170-009-2197-2. VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 210


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

[11]

Hesam Shahali, M Reza Soleymani Yazdi, Aminollah Mohammadi and Ehsan

Iimanian, Optimization of surface roughness and thickness of white layer in wire electrical discharge machining of DIN 1.4542 stainless steel using micro genetic algorithm and signal to noise ratio techniques Proc IMechE Part B: J Engineering Manufacture 226(5) 803–812_ IMechE 2012. [12]

A. Garg & L. Rachmawati & K. Tai Classification-driven model selection

approach of genetic programming in modelling of turning process Int J Adv Manuf Technol (2013) 69, pp. 1137–1151. [13]

Khaider Bouacha, Mohamed Athmane Yallese, Samir Khamel, Salim Belhadi

Analysis and optimization of hard turning operation using cubic boron nitride tool Int. Journal of Refractory Metals and Hard Materials 45 (2014), pp. 160–178. [14]

Yiğit Karpat & Tuğrul Özel, Multi-objective optimization for turning processes

using neural network modelling and dynamic-neighbourhood particle swarm optimization Int J Adv Manuf Technol (2007) 35, pp. 234–247. [15]

M. Al-Ahmari Prediction and optimisation models for turning Operations ISSN:

0020-7543, pp. 1366-588 [16]

Adel T. Abbas Karim Hamza Mohamed F. Aly and Essam A. Al-Bahkali,

Multiobjective Optimization of Turning Cutting Parameters for J-Steel Material Advances in Materials Science and Engineering Volume 2016, Article ID 6429160, 8 pages Hindawi Publishing. [17]

ZhenhuaWang,Juntang Yuan, Zengbin Yin and Chao Li, Study on high-speed

cutting parameters optimization of AlMn1Cu based on neural network and genetic algorithm Advances in Mechanical Engineering 2016, Vol. 8(4), pp. 1–12 . [18]

Yunguang Zhou & Yadong Gong& Zongxiao Zhu & Qi Gao & Xuelong Wen

Modelling and optimisation of surface roughness from micro grinding of nickel-based single crystal super alloy using the response surface methodology and genetic algorithm Int J Adv Manuf Technol (2016) 85:2607 2622 [19]

Shahram Saeidi & Maghsud Solimanpur & Iraj Mahdavi & Nikbakhsh Javadian

A multi-objective genetic algorithm for solving cell formation problem using a fuzzy goal programming approach Int J Adv Manuf Technol (2014) 70, pp. 1635–1652.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 211


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

[20]

N Alberti1 and G Perrone2 Multi-pass machining optimization by using fuzzy

Possibilistic programming and genetic algorithms Proc Instn Mech Engrs Vol 213 Part B. [21]

Mohinder P Garg, Ajai Jain and Gian Bhushan Modelling and multi-objective

optimization of process parameters of wire electrical discharge machining using nondominated sorting genetic algorithm-II Proc IMechE Part B: J Engineering Manufacture 226(12), pp. 1986–2001. [22]

A. Pramanick, N. Saha, P. P. Dey and P. K. Das Wire EDM Process Modelling

With Artificial Neural Network And Optimization By Grey Entropy-Based Taguchi Technique

For Machining Pure Zirconium Di-boride Journal of Manufacturing

Technology Research ISSN: 1943-8095. [23]

M. A. Sahali & I. Belaidi & R. Serra, New approach for robust multi-objective

optimization of turning parameters using probabilistic genetic algorithm Int J Adv Manuf Technol (2016) 83:1265–1279. [24]

JS Dureja, VKGupta, Vishal S Sharma, Manu Dogra and Manpreet S Bhatti, A

review of empirical modelling techniques to optimize machining parameters for hard turning applications Proc IMechE Part B: J Engineering Manufacture 2016, Vol. 230(3), pp. 389–404. [25]

H. Ganesan G. Mohankumar, Optimization of Machining Techniques in CNC

Turning Centre Using Genetic Algorithm Arab J Sci. Eng. (2013) 38, pp. 1529–1538. [26]

I.S. Jawahir, X. Wang, Development of hybrid predictive models and

optimization techniques for machining operations Journal of Materials Processing Technology 185 (2007) 46–59 25. [27]

KA Sundararaman, K. P Padmanaban and M Sabareeswaran, Optimization of

machining fixture layout using integrated response surface methodology and evolutionary techniques, Proc IMechE Part C:J Mechanical Engineering Science 2016, Vol. 230(13) 2245–2259 . [28]

Antonio Costa & Giovanni Celano & Sergio Fichera, Optimization of multi-pass

turning economies through a hybrid particle swarm optimization technique, Int. J Adv. Manuf. Technol. (2011) 53:421–433. DOI 10.1007/s00170-010-2861-6.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 212


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

[29]

S. Bharathi Raja, N. Baskar,

Application of Particle Swarm Optimization

technique for achieving desired milled surface roughness in minimum machining time, Expert Systems with Applications 39 (2012) 5982. [30]

S. Bharathi Raja, N. Baskar Optimization techniques for machining operations:a

retrospective research based on various mathematical models, Int. J. Adv. Manuf. Technol. (2010) 48:1075–1090. [31]

S. Bharathi Raja, N. Baskar, Particle swarm optimization technique for

determining optimal machining parameters of different work piece materials in turning operation Int. J Adv. Manuf. Technol. (2011) 54:445–463. DOI 10.1007/s00170-010-2958. [32]

M. Chandrasekaran& M. Muralidhar&C. Murali Krishna & U. S. Dixit

Application of soft computing techniques in machining performance prediction and optimization: a literature review Int. J. Adv. Manuf. Technol. (2010) 46:445–464, DOI 10.1007/s00170-009-2104. [33]

G. Prabhaharan , K. P. Padmanaban, R. Krishnakumar Machining fixture layout

optimization using FEM and evolutionary techniques, Int. J Adv. Manuf. Technol. (2007) 32: 1090–1103, DOI 10.1007/s00170-006. [34]

Farahnakian M., Razfar, M. R., Moghri, M., &Asadnia, M. (2011),The selection

of milling parameters by the PSO-based neural network modeling method. International Journal of Advanced Manufacturing Technology, 1-12. [35]

Yang, W., Guo, Y., & Liao, W. (2011a). Optimization of multi-pass face milling

using a fuzzy particle swarm optimization algorithm. International Journal of Advanced Manufacturing Technology, 54(1-4), 45. [36]

Escamilla, I., Perez, P., Torres, L., Zambrano, P., & Gonzalez, B. (2009),

Optimization using neural network modeling and swarm intelligence in the machining of titanium (ti 6al 4v) alloy. International Conference on Artificial Intelligence, MICAI 2009, 33-38. [37]

Li, J. G., Yao, Y. X., Gao, D., Liu, C. Q., & Yuan, Z. J. (2008), Cutting

parameters optimization by using particle swarm optimization (PSO), Applied Mechanics and Materials Vols. 10-12 (2008), pp. 879-883.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 213


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

[38]

Chen, Z. and Li, Y. (2008). An improved particle swarm algorithm and its

application in grinding process optimization, Proceedings of the 27th Chinese Control Conference. 2-5 [39]

Rajkamal Shukla, Dinesh Singh Experimentation investigation of abrasive water

jet machining parameters using Taguchi and Evolutionary optimization techniques Swarm and Evolutionary Computation journal .via Taguchi method-based response surface analysis. Measurement 45 (2012) 785–794 [40]

Ilhan Asiltürk a, Mehmet Çunkas, Modeling and prediction of surface roughness

in turning operations using artificial neural network and multiple regression method Expert Systems with Applications 38 (2011) 5826–5832 [41]

N Senthilkumar, T Tamizharasan, Flank wear and surface roughness prediction

in hard turning Via artificial neural network and multiple regressions. [42]

Miron Zapciu

Jean-Yves K’nevez

Alain Gérard

Olivier Cahuc

Florinel Bisu Dynamic characterization of machining systems

Claudiu

Int. J. Adv. Manuf.

Technol. (2011) 57:73–83DOI 10.1007/s00170-011-3277-7. [43]

Dilbag Singh & P. Venkateswara Rao, Flank wear prediction of ceramic tools in

hard turning Int J Adv Manuf Technol (2010) 50:479–493 DOI 10.1007/s00170-0102550-5 [44]

Yahya Isik, Investigating the machinability of tool steels in turning operations

Materials and Design 28 (2007) pp. 1417–1424. [45]

Hamdi Aouici, Mohamed Athmane Yallese, Kamel Chaoui, Tarek Mabrouki,

Jean-François Rigal. Analysis of surface roughness and cutting force components in hard turning with CBN tool: Prediction model and cutting conditions optimization Measurement 45 (2012), pp. 344–353. [46]

Suha K. Shihab, Zahid A. Khan, Aas Mohammad and Arshad Noor Siddiquee,

Investigations on the Effect of CNC Dry Hard Turning Process Parameters on Surface [47]

Waleed Bin Rashid & Saurav Goel & J. Paulo Davim & Shrikrishna N.Joshi

Parametric design optimization of hard turning of AISI 4340 steel (69 HRC) Int J Adv Manuf Technol (2016) 82, pp. 451–462.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 214


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

[48]

Abhijit Saha, Subhas Chandra Mondal, Multi-objective optimization in WEDM

process of nanostructured hard-facing materials through hybrid techniques Measurement 94 (2016) pp. 46–59. [49]

Emre Yu cel and Mustafa Gu¨nay Modelling and optimization of the cutting

conditions in hard turning of high-alloy white cast iron (Ni-Hard) Proc IMech E Part C: J Mechanical Engineering Science 227(10) 2280–2290. [50]

Ilhan Asiltürk, Süleyman Neseli, Multi-responseoptimization of CNC turning

parameters via Taguchi method-based response surface analysis, Measurement 45 (2012), pp. 785–794. [51]

Gaurav Bartarya S.K. Choudhury State of the art in hard turning International

Journal of Machine Tools & Manufacture 53 (2012) 1–14. [52]

Ashvin J. Makadia, J. I. Nanavati, Optimisation of machining parameters for

turning operations based on response surface methodology Measurement 46 (2013), pp. 1521–1529. [53]

Ilhan Asilturk, Harun Akku, Determining the effect of cutting parameters on

surface roughness in hard turning using the Taguchi method. Measurement 44 (2011) pp. 1697–1704. [54]

Aman Aggarwal and Hari Singh, Optimization of machining techniques – A

retrospective and literature review Sadhana Vol. 30, Part 6, December 2005, pp. 699– 711. [55]

Chinmaya R. Dandekar, Yung C. Shin, JohnBarnes, Machinability improvement

of titanium alloy (Ti–6Al–4V) via LAM and hybrid machining, International Journal of Machine Tools & Manufacture 50 (2010) 174–182. [56]

X. Wang, Z.J. Da, A.K. Balaji, I.S. Jawahir, Performance-Based Predictive

Models and Optimization Methods for Turning Operations and Applications: Part 3— Optimum Cutting Conditions and Selection of Cutting Tools, Journal of Manufacturing Processes Vol. 9/No. 1 2007. [57]

Devinder Priyadarshi, and Rajesh Kumar Sharma, Effect of type and percentage

of reinforcement for optimization of the cutting force in turning of Aluminum matrix

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 215


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Nano composites using response surface methodologies

Journal of Mechanical

Science and Technology 30 (3) (2016) 1095~1101. [58]

Ersan Aslan, Necip Camus cua, Burak Birgoren, Design optimization of cutting

parameters when turning hardened AISI 4140 steel (63 HRC) with Al2O3 + TiCN mixed ceramic tool, Materials and Design 28 (2007) 1618–1622. [59]

Hashimoto, Y.B. Guo, A.W. Warren, Surface Integrity Difference between Hard

Turned and Ground Surfaces and Its Impact on Fatigue Life The Timken Company, Canton, Ohio 44706, USA [60]

T. Ozel, Y. Karpat, A. Srivastava, Hard turning with variable micro-geometry

PcBN tools CIRP Annals - Manufacturing Technology 57 (2008) pp. 73–76. [61]

Ravinder Kumar, Santram Chauhan, Study on surface roughness measurement for

turning of Al 7075/10/SiCp and Al 7075 hybrid composites by using response surface methodology (RSM) and artificial neural networking (ANN). [62]

Mozammel Mia, Nikhil Ranjan Dhar Prediction of surface roughness in hard

turning under high pressure coolant using Artificial Neural Network Measurement 92 (2016) 464–474 [63]

Fabrício José Pontes b, Anderson Paulo de Paiva a, Pedro Paulo Balestrassi a,

João Roberto Ferreira a,⇑,Messias Borges da Silva b Optimization of Radial Basis Function neural network employed for prediction of surface roughness in hard turning process using Taguchi’s orthogonal arrays Expert Systems with Applications 39 (2012) 7776–7787. [64]

Vinayak Neelakanth Gaitonde S. R. Karnik & Luis Figueira & J. Paulo Davim

Performance comparison of conventional and wiper ceramicinserts in hard turning through artificial neuralnetwork modeling Int J Adv Manuf Technol (2011) 52:101– 114 DOI 10.1007/s00170-010-2714-3. [65]

Xiaoyu Wang Wen Wang Yong Huang Nhan Nguyen Design of neural network-

based estimator for tool wear modeling in hard turning Intell Manuf (2008) 19:383– 396 DOI 10.1007/s10845-008-0090-8.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 216


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

[66]

D. Umbrello a,∗, G. Ambrogioa, L. Filice a, R. Shivpuri ba An ANN approach for

predicting subsurface residual stresses and the desired cutting conditions during hard turning Journal of Materials Processing Technology 189 (2007) 143–15 [67]

Ravinder Kumar, Santram Chauhan, Study on surface roughness measurement for

turning of Al 7075/10/SiCp and Al 7075 hybrid composites by using response surface methodology (RSM) and artificial neural networking (ANN). [68]

Ali R. Yildiz Hybrid Taguchi-differential evolution algorithm for optimization of

multi-pass turning operations Applied Soft Computing 13 (2013) 1433–1439. [69]

F. Kara • K. Aslantas • A. Cicek ANN and multiple regression method-based

modelling of cutting forces in orthogonal machining of AISI 316L stainless steel Neural Comput & Applic (2015) 26:237–250 DOI 10.1007/s00521-014-1721[70]

Sener Karabulut Optimization of surface roughness and cutting force

duringAA7039/Al2O3 metal matrix composites milling using neural networks and Taguchi method Measurement 66 (2015) 139–149. [71]

Deb, K. (2008) Multi-objective Optimisation using Evolutionary Algorithms, 2nd

ed., John Wiley & Sons, Chichester. Deb, K., Pratap, A., Agarwal, S. and Meyarivan, T. (2002) A fast and elitist multi-objective genetic algorithm: NSGA-II, IEEE Transactions on Evolutionary Computation, 6(2), 182–197. [72]

Deb, K., Pratap, A., Agarwal, S. and Meyarivan, T. (2002) A fast and elitist

multi-objective genetic algorithm: NSGA-II, IEEE Transactions on Evolutionary Computation, 6(2), 182–197. [73]

Zitzler, E., Laumanns, M. and Thiele, L. (2001) SPEA2: Improving the strength

Pareto evolutionary algorithm. Proceedings of the Evolutionary Methods for Design, Optimization, and Control with Applications to Industrial Problems, EUROGEN 2001, Athens, Greece, pp. 95–100. [74]

S.Milad.Nayyer Sabeti 1 and MR.Deevband 2 Hybrid Evolutionary Algorithms

based on PSO-GA for Training ANFIS Structure International Journal of Computer Science Issues, Volume 12, Issue 5, September 2015.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 217


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

[75]

Yosra JARRAYA, Souhir BOUAZIZ, Adel M. ALIMI Ajith ABRAHAM Fuzzy

Modeling System based on Hybrid Evolutionary Approach Machine Intelligence Research Labs, WA, USA. [76]

J. Kennedy, R. Eberhart, “Particle Swarm Optimization”, From Proc. IEEE Int’l.

Conf. on Neural Networks (Perth, Australia), IEEE Service Centre, Piscataway, NJ, IV:1942–1948, 1995.

APPENDIX: A CONFERENCES AND PUBLICATION Conferences [1] Presented Paper in “VISHWACON 2016-2017” at Vishwakarma Institute of Information Technology, Pune on 17th Feb. 2017. [2] Presented Paper in “ICMMM-2017” at VIT-University, Vellore on 10th March 2017. [3] Presented Paper in “MECHPGCON-2017” held at Zeal College of Engineering and Research, Pune on 20th June 2017. [4] Presented Paper in “ICMTS-2017” held at Indian Institute of Technology –Madras, Chennai on 7th July 2017. [5] Paper presented in “IconAMMA-2017” held at Amrita Vishwa Vidyapeetham University on Aug 2017.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 218


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Selected Paper for conferences and publication [1] A research paper selected at “IconAMMA-2017” held at Amrita Vishwa Vidyapeetham University on Aug 2017 and will be published in Materials Today: Proceedings.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 219


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

APPENDIX: B CERTIFICATES [1] Certificate of presented paper in “MECHPGCON-2017” held at Zeal College of Engineering and Research, Pune on 20th June 2017.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 220


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Reviewer & Evaluation Report MECHPGCON-2017

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 221


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

[2] Certificate of presented paper in “VISHWACON 2016-2017” at Vishwakarma Institute of Information Technology, Pune on 17th Feb. 2017.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 222


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

[3] Certificate of presented paper in “ICMMM-2017” at VIT-University, Vellore on 10th March 2017.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 223


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 224


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

[4] Certificate of presented paper in “ICMTS-2017” held at Indian Institute of Technology –Madras, Chennai on 7th July 2017.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 225


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

[5] Certificate of poster presentation in “AVISHKAR-2016” held at Vishwakarma Institute of Information Technology, Pune on 2nd December 2016.

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 226


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Appendix

ANFIS Algorithm //Load data [Inputs ,Targets]=Load Machining data //Shuffle data S=randompermutation([data]) [Inputs, Targets]=[Inputs(S,:),Targets(S,:)] Train inputs=[Inputs, Targets] Train(Inputs, Targets), Test(Inputs, outputs)]=[Input(S,:) Targets(S,: )] //Generate ANFIS structure Do  Training objectives fis =Create Initial Fis (data) Create Initial fis(data) Switch case Case 1 Grid Partitioning ANFIS Params No of mf nmfs : 5 Input mf type Gauss mf Output mf type : Linear Fis structure= genfis(Train Input , Train targets, nmfs, gauss, Linear) Data size =size(data,1) VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 227


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

In_n =size(data, 2)-1 Input mfs_Type= In_mfs type Rule_n=prod(nmfs) Fis.name=anfis Fis.and method=’prod’ Fis.or method=’ max’ Fis.defuzzification method=’weight average’ Fis.implication method=’prod’ Fis.agrregation=’max’ Case 2 Cluster method Influence Radius Fis=generate fis (Train inputs, Train targets ,Radius) Case 3 C-mean cluster method FCM option structure No of clusters (ncluster) Portioning matrix component Maximum no of iteration Fis = generate fis(Train input, Train target ,ncluster, FCM options,Optimization method) Training params [max epoch, error goal,initial step size, step size decreament rate, step size increament rate] Output =evaluate fis (Inputs,fis)

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 228


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Train output= output [Train(Inputs)] Test output=output[Test (Inputs)] //Statistical Analysis Train error =Train targets-Train outputs Train MSE =Mean(Train error) Train RMSE =

MSE

Train Mean error=Mean(Train error) Train Error STD= STD [Train error] Test Error calculation Test error = Test targets- Test outputs Test MSE =Mean(Test error) Test RMSE =

MSE

Test Mean error=Mean(Test error) Test Error STD= STD [Test error]

ANFIS-GA/PSO Algorithm ANFIS GA/PSO //Load data [Inputs ,Targets]=Load Machining data //Shuffle data S=randompermutation([data]) [Inputs, Targets]=[Inputs(S,:),Targets(S,:)] Train inputs=[Inputs, Targets] VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 229


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Train(Inputs, Targets), Test(Inputs, outputs)]=[Input(S,:) Targets(S,: )] //Generate ANFIS structure fis =Create Initial Fis (data) Switch case Case 1 Train ANFIS using GA (fis,data) Case 2 Train ANFIS using PSO(fis ,data) Train output: evaluate fis(data.train inputs, fis) Test output : evaluate fis(data.test inputs, fis) Create Initial fis(data) Switch case Case 1 Grid Partitioning ANFIS Params No of mf nmfs : 5 Input mf type Gauss mf Output mf type : Linear Fis structure= genfis(Train Input , Train targets, nmfs, gauss, Linear) Data size =size(data,1) In_n =size(data, 2)-1 Input mfs_Type= In_mfs type Rule_n=prod(nmfs) VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 230


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Fis.name=anfis Fis.and method=’prod’ Fis.or method=’ max’ Fis.defuzzification method=’weight average’ Fis.implication method=’prod’ Fis.agrregation=’max’ Case 2 Cluster method Influence Radius Fis=generate fis (Train inputs, Train targets ,Radius) Case 3 C-mean cluster method FCM option structure No of clusters (ncluster) Portioning matrix component Maximum no of iteration Fis = generate fis(Train input, Trin target ,ncluster, FCM options) Train fis cost(x,fis,data) P0=Get fis params(fis) P=x*p0 Fis =Set fis params(fis,p) X=data.Train inputs t=data.Train targets VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 231


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Y=evaluate(x,fis) // calculate Statistical error Error :[t]-[y] MSE : mean(Error) Get fis params(fis) P=[] N=size[fis.input] For i [1-N] Nmfs =size[fis.Input[i].mf] For j [1-nmfs] P=[p, fis.Input[j].mf.params] Noutput=size[fis.output] For i [1-noutputs] Nmfs =size[fis.output[i].mfs.params] For j [1-nmfs] P=[p, fis.Input[j].mf.params] Set fis params(fis,p) P0=Get fis parms P0=[] X= data.Train Inputs Y=data.Train Targets Y=evaluate fis (x, fis) Train Anfis using GA(fis, data) VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 232


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

P0=Get fis params(fis) Evalaute fitness function= Train fis cost(x,fis,data) Nvar=size(p0) Range [min , max] Max it=50 N=25 // no of pop Pc=[0.4-0.8] //crossover percentage Nc=2*round(pc*N/2) Pm=[0.2-0.4] //mutation percentage Nm =round(pm*N) ÎĽ=0.7 Mu=0.15 //mutation rate //Initialize population Pop.position=[] Pop.cost=[] For i [1-N] If it>1 Pop[i].position =random(range ,N) //Evalute Pop[i].cost =Train fis cost(x,fis, data) //sort Population [cost ,sort order]= sort[pop.cost] Pop=pop[sort order] VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 233


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

//Best cost Best pop=pop[1] Worst pop=pop[end] Main Loop Do until (max it) p  exp( P

  * best cos t ) worst cos t

p N

p n 1

[p1,p2]=roulette wheel selection(P) [child 1 child 2]=crossover (p1,p2, crossover params, N) //Evaluate off spring Popc=Train fis cost (child1, fis, data) //Mutate P3=roulette wheel selection(P) Child 3=muatate (p3,mutation params, N) Popm =Train fis cost (child 3, fis, data) Pop=[pop,popc,popm) Crossover (p1,p2,crossover parameters) Parametrs :(Υ,range) α:random(-Υ,1+Υ,N) y1   * p1  (1   )* p 2 y 2   * p 2  (1   )* p1

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 234


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Y1=min[max(y1,range)] Y2=min[max(y2,range)] Mutate(p3,mutation parameters, range) Parameters : , range Rmin=min(range) Rmax=max(range) dr=Rmax-Rmin Ďƒ: *dr y=p1+Ďƒ*random(Nm) y=min(max(y,range)) Train anfis using PSO (fis ,data) P0=Get fis params(fis) Evaluate fitness function = Train fis cost (x, fis ,data) nvar =size(p0,1) range(min, max) max it =50 N=25 //no of pop W=1, Wdamp=0.99, C1=1, c2 =2 //Initialize Particle structure {Particle.Position, Particle.Velocity, Particle.Cost, Particle.Best Position, Particle.Bestcost, Particle.Is Dominated, Particle.Grid Index, Particle.Grid Subindex.} //Evaluate particle.position and cost For i :[1-N]

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 235


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Pop[i].position=random[Range,N] Pop[i].velocity : zeros[N] Pop[i].cost :Train fis cost (pop[i].position) //update personal best Exchange pop[i].best position with pop[i].position Exchange popi[i].best cost with pop[i].cost //determine domination level Pop : domination(pop,N) For i:[1-N] For j:[i+1-N] If dominates (pop[i], pop[j]) True(pop[j] Is dominated) Else if dominates(pop[j],pop[i]) True(pop[i] Is dominated) b=dominates(pop[i],pop[j]) b= all(x<y)&&any(x<y) Do until max (IT) For i : [1-N] //Select Leader leader=Select leader(rep, )

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 236


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation pop  i  .Velocity  w * pop  i  .Velocity c1* rand VarSize  . *  pop  i  .Best.Position  pop  i  .Position   c 2* rand VarSize  . *  leader.Position  pop  i  .Position  pop  i  .Position  pop  i  .Position  pop  i  .Velocity pop  i  .Position  max  pop  i  .Position, VarMin  pop  i  .Position  min  pop  i  .Position, VarMax  pop  i  .Cost  Train fis cost  pop  i  .Position 

New pop=mutatute(pop,pm,Range) New pop.cost= Train fis cost (newpop.position) Determine domination(pop) If dominates(New.pop.position, pop.position) True(Is dominated pop.position) Else if dominates(pop.position, New.pop.position) True(Is dominated New.pop.position) Pop[i].Position = max(particle[i].Position,Rmin); pop[i].Position = min(particle[i].Position,Rmax); if particle(i).Cost<particle(i).Best.Cost pop(i).Best.Position=particle(i).Position; pop(i).Best.Cost=particle(i).Cost; if pop(i).Best.Cost<BestSol.Cost BestSol=pop(i).Best

Algorithm for Neural Network Neural network //Load data [Train Inputs Target Inputs]= Load data VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 237


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

//Define network structure Def Train function type :Train func Hidden layers //network type Net=fitnet (hidden layer, size, Train func) Fitnet (hidden layer ,size,Train func) nnetparamsInfo.hiddensize nnetparamsInfo.hidden layers nnetparamsInfo.nntype nnetparamsInfo.Train func net.input.process func : {mapinput vectors} net.output processfunc: {mapoutput vectors} net.divide params.train ratio=0.7 net.divide params.validate ratio=0.15 net.divide params.testratio=0.15 //performance function Net.perform fucn=’MSE’ [net ,train]=train(net, x,t) //test network Y=net(x) E=[t]-[y] Performance =perform(net ,t, y) EANNHRC(x) VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 238


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

//Input layer 1 X1_step1_xoffset=[Input] X1_step1_gain=[‘e’] X1_step1_ymin=-1 //Layer 1 B1=[bais for each neuron] IW1=[weights for each neouron] //Layer2 B2=[bais for each neuron] IW2=[weights for each neuron] //output layer Y1_step1_ymin=-1 Y1_step1_gain=[] Y1_step1_xoffset=[outputs] //simulation For i:[1-size(output)] //Input1 Xp1=mapminmax_apply([x1,i],x1_step1_gain, x1_step1_xoffset,x1_step1_ymin) //Layer 1 A1=tansig_apply((b1,1,q)+IW1*xp1) //layer 2 A2=[b1,1,q]+lw1*a1 //output 2 VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 239


Evolutionary Algorithms for Multi-Objective Optimization: Modelling and Comparative Evaluation

Y(1,t,s)=mapminmax(a2,y1_step1_gain,y1_step1_xoffset,y1_step1_ymin) Mapminmax_apply(x,settings_gain,settings_xoffset,setting_ymin) Y=bsxfun(@minus,x,setting_x_offset) Y=bsxfun(@times,y,setting_gain) Y=bsx(@plus,y,setting_ymin) //sigmoid symmetric transfer function Tansig_apply(x) //mapminimumand maximum output Mapminmax_reverse(y,setting_gain,setting_xoffset,setting_ymin) X=bsxfun(@minus,y,setting_x_offset) x=bsxfun(@times,x,setting_gain) x=bsx(@plus,x,setting_ymin)

VISHWAKARMA INSTITUTE OF INFORMATION TECHNOLOGY, PUNE M.E. (Mechanical) (Design Engineering) 240


Turn static files into dynamic content formats.

Create a flipbook
Issuu converts static files into: digital portfolios, online yearbooks, online catalogs, digital photo albums and more. Sign up and create your flipbook.