Advertisement
Guest User

Untitled

a guest
Feb 22nd, 2018
66
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
Python 0.55 KB | None | 0 0
  1. loss_list= []
  2. x =[]
  3. u = 0.001 # константа для скорости обучения
  4.  
  5. def delta_L(y, x, w, reg = 0.001): #градиент
  6.     return np.sum((-X*(y[:,np.newaxis]))/(1+np.exp(y[:,np.newaxis]*(X.dot(w[:,np.newaxis])))) + reg * w,axis=0)
  7. for epoch in range(10000): #10000 эпох для обучения
  8.     loss = 0
  9.  
  10.     loss = np.sum(np.log(1+np.exp(-y[:,np.newaxis]*(X.dot(w[:,np.newaxis])))),axis = 0)/len(X)
  11.  
  12.     w -=  u* delta_L(y,X,w)
  13.     x.append(epoch)
  14.     loss /= len(X)
  15.     loss_list.append(loss)
  16.  
  17. plt.plot(x,loss_list)
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement