码迷,mamicode.com
首页 > 其他好文 > 详细

BP神经网络

时间:2018-06-23 17:14:20      阅读:227      评论:0      收藏:0      [点我收藏+]

标签:std   计算   put   sys   void   mes   layer   delay   com   

头文件

 1 #pragma once
 2 
 3 #include <iostream>
 4 #include <cmath>
 5 #include <vector>
 6 #include <stdlib.h>
 7 #include <time.h>
 8 
 9 using namespace std;
10 
11 #define innode 2        //输入结点数
12 #define hidenode 4      //隐含结点数
13 #define hidelayer 1     //隐含层数
14 #define outnode 1       //输出结点数
15 #define learningRate 0.9//学习速率,alpha
16 
17 // --- -1~1 随机数产生器 --- 
18 inline double get_11Random()    // -1 ~ 1
19 {
20     return ((2.0*(double)rand()/RAND_MAX) - 1);
21 }
22 
23 // --- sigmoid 函数 --- 
24 inline double sigmoid(double x)
25 {
26     double ans = 1 / (1+exp(-x));
27     return ans;
28 }
29 
30 // --- 输入层节点。包含以下分量:--- 
31 // 1.value:     固定输入值; 
32 // 2.weight:    面对第一层隐含层每个节点都有权值; 
33 // 3.wDeltaSum: 面对第一层隐含层每个节点权值的delta值累积
34 typedef struct inputNode
35 {
36     double value;
37     vector<double> weight, wDeltaSum;
38 }inputNode;
39 
40 // --- 输出层节点。包含以下数值:--- 
41 // 1.value:     节点当前值; 
42 // 2.delta:     与正确输出值之间的delta值; 
43 // 3.rightout:  正确输出值
44 // 4.bias:      偏移量
45 // 5.bDeltaSum: bias的delta值的累积,每个节点一个
46 typedef struct outputNode   // 输出层节点
47 {
48     double value, delta, rightout, bias, bDeltaSum;
49 }outputNode;
50 
51 // --- 隐含层节点。包含以下数值:--- 
52 // 1.value:     节点当前值; 
53 // 2.delta:     BP推导出的delta值;
54 // 3.bias:      偏移量
55 // 4.bDeltaSum: bias的delta值的累积,每个节点一个
56 // 5.weight:    面对下一层(隐含层/输出层)每个节点都有权值; 
57 // 6.wDeltaSum: weight的delta值的累积,面对下一层(隐含层/输出层)每个节点各自积累
58 typedef struct hiddenNode   // 隐含层节点
59 {
60     double value, delta, bias, bDeltaSum;
61     vector<double> weight, wDeltaSum;
62 }hiddenNode;
63 
64 // --- 单个样本 --- 
65 typedef struct sample
66 {
67     vector<double> in, out;
68 }sample;
69 
70 // --- BP神经网络 --- 
71 class BpNet
72 {
73 public:
74     BpNet();    //构造函数
75     void forwardPropagationEpoc();  // 单个样本前向传播
76     void backPropagationEpoc();     // 单个样本后向传播
77     // 更新 weight, bias
78     void training (static vector<sample> sampleGroup, double threshold);
79     // 神经网络预测
80     void predict  (vector<sample>& testGroup);                          
81     // 设置学习样本输入
82     void setInput (static vector<double> sampleIn); 
83     // 设置学习样本输出        
84     void setOutput(static vector<double> sampleOut);    
85 
86 public:
87     double error;
88     inputNode* inputLayer[innode];// 输入层(仅一层)
89     outputNode* outputLayer[outnode];// 输出层(仅一层)
90     // 隐含层(可能有多层)
91     hiddenNode* hiddenLayer[hidelayer][hidenode];
92 };

主程序:main

 1 #include "BP.h"
 2 
 3 int main()
 4 {
 5     BpNet testNet;
 6 
 7     // 学习样本
 8     vector<double> samplein[4];
 9     vector<double> sampleout[4];
10     samplein[0].push_back(0); samplein[0].push_back(0); sampleout[0].push_back(0); 
11     samplein[1].push_back(0); samplein[1].push_back(1); sampleout[1].push_back(1); 
12     samplein[2].push_back(1); samplein[2].push_back(0); sampleout[2].push_back(1); 
13     samplein[3].push_back(1); samplein[3].push_back(1); sampleout[3].push_back(0); 
14     sample sampleInOut[4];
15     for (int i = 0; i < 4; i++)
16     {
17         sampleInOut[i].in = samplein[i];
18         sampleInOut[i].out = sampleout[i];
19     }
20     vector<sample> sampleGroup(sampleInOut, sampleInOut + 4);
21     testNet.training(sampleGroup, 0.0001);
22 
23     // 测试数据
24     vector<double> testin[4];
25     vector<double> testout[4];
26     testin[0].push_back(0.1);   testin[0].push_back(0.2);
27     testin[1].push_back(0.15);  testin[1].push_back(0.9);
28     testin[2].push_back(1.1);   testin[2].push_back(0.01);
29     testin[3].push_back(0.88);  testin[3].push_back(1.03);
30     sample testInOut[4];
31     for (int i = 0; i < 4; i++) testInOut[i].in = testin[i];
32     vector<sample> testGroup(testInOut, testInOut + 4);
33 
34     // 预测测试数据,并输出结果
35     testNet.predict(testGroup);
36     for (int i = 0; i < testGroup.size(); i++)
37     {
38         for (int j = 0; j < testGroup[i].in.size(); j++) 
39         cout << testGroup[i].in[j] << "\t";
40         cout << "-- prediction :";
41         for (int j = 0; j < testGroup[i].out.size(); j++) 
42         cout << testGroup[i].out[j] << "\t";
43         cout << endl;
44     }
45 
46     system("pause");
47     return 0;
48 }

源程序:bp.cpp

  1 #include "BP.h"
  2 
  3 using namespace std;
  4 
  5 BpNet::BpNet()
  6 {
  7     srand((unsigned)time(NULL)); // 随机数种子    
  8     error = 100.f;      // error初始值,极大值即可
  9 
 10     // 初始化输入层
 11     for (int i = 0; i < innode; i++)
 12     {
 13         inputLayer[i] = new inputNode();
 14         for (int j = 0; j < hidenode; j++) 
 15         {
 16             inputLayer[i]->weight.push_back(get_11Random());
 17             inputLayer[i]->wDeltaSum.push_back(0.f);
 18         }
 19     }
 20 
 21     // 初始化隐藏层
 22     for (int i = 0; i < hidelayer; i++)
 23     {
 24         if (i == hidelayer - 1)
 25         {
 26             for (int j = 0; j < hidenode; j++)
 27             {
 28                 hiddenLayer[i][j] = new hiddenNode();
 29                 hiddenLayer[i][j]->bias = get_11Random();
 30                 for (int k = 0; k < outnode; k++) 
 31                 {
 32                     hiddenLayer[i][j]->weight.push_back(get_11Random());
 33                     hiddenLayer[i][j]->wDeltaSum.push_back(0.f);
 34                 }
 35             }
 36         }
 37         else
 38         {
 39             for (int j = 0; j < hidenode; j++)
 40             {
 41                 hiddenLayer[i][j] = new hiddenNode();
 42                 hiddenLayer[i][j]->bias = get_11Random();
 43                 for (int k = 0; k < hidenode; k++) 
 44                 {
 45                     hiddenLayer[i][j]->weight.push_back(get_11Random());
 46                 }
 47             }
 48         }
 49     }
 50 
 51     // 初始化输出层
 52     for (int i = 0; i < outnode; i++)
 53     {
 54         outputLayer[i] = new outputNode();
 55         outputLayer[i]->bias = get_11Random();
 56     }
 57 }
 58 
 59 void BpNet::forwardPropagationEpoc()
 60 {
 61     // forward propagation on hidden layer
 62     for (int i = 0; i < hidelayer; i++)
 63     {
 64         if (i == 0)
 65         {
 66             for (int j = 0; j < hidenode; j++)
 67             {
 68                 double sum = 0.f;
 69                 for (int k = 0; k < innode; k++) 
 70                 {
 71                     sum += inputLayer[k]->value * inputLayer[k]->weight[j];
 72                 }
 73                 sum += hiddenLayer[i][j]->bias;
 74                 hiddenLayer[i][j]->value = sigmoid(sum);
 75             }
 76         }
 77         else
 78         {
 79             for (int j = 0; j < hidenode; j++)
 80             {
 81                 double sum = 0.f;
 82                 for (int k = 0; k < hidenode; k++) 
 83                 {
 84                     sum += hiddenLayer[i-1][k]->value * hiddenLayer[i-1][k]->weight[j];
 85                 }
 86                 sum += hiddenLayer[i][j]->bias;
 87                 hiddenLayer[i][j]->value = sigmoid(sum);
 88             }
 89         }
 90     }
 91 
 92     // forward propagation on output layer
 93     for (int i = 0; i < outnode; i++)
 94     {
 95         double sum = 0.f;
 96         for (int j = 0; j < hidenode; j++)
 97         {
 98             sum += hiddenLayer[hidelayer-1][j]->value * hiddenLayer[hidelayer-1][j]->weight[i];
 99         }
100         sum += outputLayer[i]->bias;
101         outputLayer[i]->value = sigmoid(sum);
102     }
103 }
104 
105 void BpNet::backPropagationEpoc()
106 {
107     // backward propagation on output layer
108     // -- compute delta
109     for (int i = 0; i < outnode; i++)
110     {
111         double tmpe = fabs(outputLayer[i]->value-outputLayer[i]->rightout);
112         error += tmpe * tmpe / 2;
113 
114         outputLayer[i]->delta 
115             = (outputLayer[i]->value-outputLayer[i]->rightout)*(1-outputLayer[i]->value)*outputLayer[i]->value;
116     }
117 
118     // backward propagation on hidden layer
119     // -- compute delta
120     for (int i = hidelayer - 1; i >= 0; i--)    // 反向计算
121     {
122         if (i == hidelayer - 1)
123         {
124             for (int j = 0; j < hidenode; j++)
125             {
126                 double sum = 0.f;
127                 for (int k=0; k<outnode; k++)
128                 {
129                     sum += outputLayer[k]->delta * hiddenLayer[i][j]->weight[k];
130                 }
131                 hiddenLayer[i][j]->delta
132                     = sum * (1 - hiddenLayer[i][j]->value) * hiddenLayer[i][j]->value;
133             }
134         }
135         else
136         {
137             for (int j = 0; j < hidenode; j++)
138             {
139                 double sum = 0.f;
140                 for (int k=0; k<hidenode; k++)
141                 {
142                     sum += hiddenLayer[i + 1][k]->delta * hiddenLayer[i][j]->weight[k];
143                 }
144                 hiddenLayer[i][j]->delta 
145                     = sum * (1 - hiddenLayer[i][j]->value) * hiddenLayer[i][j]->value;
146             }
147         }
148     }
149 
150     // backward propagation on input layer
151     // -- update weight delta sum
152     for (int i = 0; i < innode; i++)
153     {
154         for (int j = 0; j < hidenode; j++)
155         {
156             inputLayer[i]->wDeltaSum[j] += inputLayer[i]->value * hiddenLayer[0][j]->delta;
157         }
158     }
159 
160     // backward propagation on hidden layer
161     // -- update weight delta sum & bias delta sum
162     for (int i = 0; i < hidelayer; i++)
163     {
164         if (i == hidelayer - 1)
165         {
166             for (int j = 0; j < hidenode; j++)
167             {
168                 hiddenLayer[i][j]->bDeltaSum += hiddenLayer[i][j]->delta;
169                 for (int k = 0; k < outnode; k++)
170                 { 
171                     hiddenLayer[i][j]->wDeltaSum[k] += hiddenLayer[i][j]->value * outputLayer[k]->delta; 
172                 }
173             }
174         }
175         else
176         {
177             for (int j = 0; j < hidenode; j++)
178             {
179                 hiddenLayer[i][j]->bDeltaSum += hiddenLayer[i][j]->delta;
180                 for (int k = 0; k < hidenode; k++)
181                 { 
182                     hiddenLayer[i][j]->wDeltaSum[k] += hiddenLayer[i][j]->value * hiddenLayer[i+1][k]->delta; 
183                 }
184             }
185         }
186     }
187 
188     // backward propagation on output layer
189     // -- update bias delta sum
190     for (int i = 0; i < outnode; i++) outputLayer[i]->bDeltaSum += outputLayer[i]->delta;
191 }
192 
193 void BpNet::training(static vector<sample> sampleGroup, double threshold)
194 {
195     int sampleNum = sampleGroup.size();
196 
197     while(error > threshold)
198     //for (int curTrainingTime = 0; curTrainingTime < trainingTime; curTrainingTime++)
199     {
200         cout << "training error: " << error << endl;
201         error = 0.f;
202         // initialize delta sum
203         for (int i = 0; i < innode; i++) inputLayer[i]->wDeltaSum.assign(inputLayer[i]->wDeltaSum.size(), 0.f);
204         for (int i = 0; i < hidelayer; i++){
205             for (int j = 0; j < hidenode; j++) 
206             {
207                 hiddenLayer[i][j]->wDeltaSum.assign(hiddenLayer[i][j]->wDeltaSum.size(), 0.f);
208                 hiddenLayer[i][j]->bDeltaSum = 0.f;
209             }
210         }
211         for (int i = 0; i < outnode; i++) outputLayer[i]->bDeltaSum = 0.f;
212 
213         for (int iter = 0; iter < sampleNum; iter++)
214         {
215             setInput(sampleGroup[iter].in);
216             setOutput(sampleGroup[iter].out);
217 
218             forwardPropagationEpoc();
219             backPropagationEpoc();
220         }
221 
222         // backward propagation on input layer
223         // -- update weight
224         for (int i = 0; i < innode; i++)
225         {
226             for (int j = 0; j < hidenode; j++) 
227             {
228                 inputLayer[i]->weight[j] -= learningRate * inputLayer[i]->wDeltaSum[j] / sampleNum;
229             }
230         }
231 
232         // backward propagation on hidden layer
233         // -- update weight & bias
234         for (int i = 0; i < hidelayer; i++)
235         {
236             if (i == hidelayer - 1)
237             {
238                 for (int j = 0; j < hidenode; j++)
239                 { 
240                     // bias
241                     hiddenLayer[i][j]->bias -= learningRate * hiddenLayer[i][j]->bDeltaSum / sampleNum;
242 
243                     // weight
244                     for (int k = 0; k < outnode; k++) 
245                     { hiddenLayer[i][j]->weight[k] -= learningRate * hiddenLayer[i][j]->wDeltaSum[k] / sampleNum; }
246                 }
247             }
248             else
249             {
250                 for (int j = 0; j < hidenode; j++)
251                 {
252                     // bias
253                     hiddenLayer[i][j]->bias -= learningRate * hiddenLayer[i][j]->bDeltaSum / sampleNum;
254 
255                     // weight
256                     for (int k = 0; k < hidenode; k++) 
257                     { hiddenLayer[i][j]->weight[k] -= learningRate * hiddenLayer[i][j]->wDeltaSum[k] / sampleNum; }
258                 }
259             }
260         }
261 
262         // backward propagation on output layer
263         // -- update bias
264         for (int i = 0; i < outnode; i++)
265         { outputLayer[i]->bias -= learningRate * outputLayer[i]->bDeltaSum / sampleNum; }
266     }
267 }
268 
269 void BpNet::predict(vector<sample>& testGroup)
270 {
271     int testNum = testGroup.size();
272 
273     for (int iter = 0; iter < testNum; iter++)
274     {
275         testGroup[iter].out.clear();
276         setInput(testGroup[iter].in);
277 
278         // forward propagation on hidden layer
279         for (int i = 0; i < hidelayer; i++)
280         {
281             if (i == 0)
282             {
283                 for (int j = 0; j < hidenode; j++)
284                 {
285                     double sum = 0.f;
286                     for (int k = 0; k < innode; k++) 
287                     {
288                         sum += inputLayer[k]->value * inputLayer[k]->weight[j];
289                     }
290                     sum += hiddenLayer[i][j]->bias;
291                     hiddenLayer[i][j]->value = sigmoid(sum);
292                 }
293             }
294             else
295             {
296                 for (int j = 0; j < hidenode; j++)
297                 {
298                     double sum = 0.f;
299                     for (int k = 0; k < hidenode; k++) 
300                     {
301                         sum += hiddenLayer[i-1][k]->value * hiddenLayer[i-1][k]->weight[j];
302                     }
303                     sum += hiddenLayer[i][j]->bias;
304                     hiddenLayer[i][j]->value = sigmoid(sum);
305                 }
306             }
307         }
308 
309         // forward propagation on output layer
310         for (int i = 0; i < outnode; i++)
311         {
312             double sum = 0.f;
313             for (int j = 0; j < hidenode; j++)
314             {
315                 sum += hiddenLayer[hidelayer-1][j]->value * hiddenLayer[hidelayer-1][j]->weight[i];
316             }
317             sum += outputLayer[i]->bias;
318             outputLayer[i]->value = sigmoid(sum);
319             testGroup[iter].out.push_back(outputLayer[i]->value);
320         }
321     }
322 }
323 
324 void BpNet::setInput(static vector<double> sampleIn)
325 {
326     for (int i = 0; i < innode; i++)
327     inputLayer[i]->value = sampleIn[i];
328 }
329 
330 void BpNet::setOutput(static vector<double> sampleOut)
331 {
332     for (int i = 0; i < outnode; i++) 
333     outputLayer[i]->rightout = sampleOut[i];
334 }

 

BP神经网络

标签:std   计算   put   sys   void   mes   layer   delay   com   

原文地址:https://www.cnblogs.com/hsy1941/p/9217398.html

(0)
(0)
   
举报
评论 一句话评论(0
登录后才能评论!
© 2014 mamicode.com 版权所有  联系我们:gaon5@hotmail.com
迷上了代码!