怎么在python中实现SVM 线性分类模型?很多新手对此不是很清楚,为了帮助大家解决这个难题,下面小编将为大家详细讲解,有这方面需求的人可以来学习下,希望你能有所收获。
创新互联建站专注于企业营销型网站建设、网站重做改版、邻水网站定制设计、自适应品牌网站建设、HTML5建站、商城网站建设、集团公司官网建设、外贸营销网站建设、高端网站制作、响应式网页设计等建站业务,价格优惠性价比高,为邻水等各大城市提供网站开发制作服务。导入对应的包和数据
import matplotlib.pyplot as plt import numpy as np from sklearn import datasets,linear_model,cross_validation,svm def load_data_regression(): diabetes = datasets.load_diabetes() return cross_validation.train_test_split(diabetes,diabetes.target,test_size=0.25,random_state=0) def load_data_classfication(): iris = datasets.load_iris() X_train = iris.data y_train = iris.target return cross_validation.train_test_split(X_train,y_train,test_size=0.25,random_state=0,stratify=y_train)
#线性分类SVM def test_LinearSVC(*data): X_train,X_test,y_train,y_test = data cls = svm.LinearSVC() cls.fit(X_train,y_train) print('Coefficients:%s,intercept%s'%(cls.coef_,cls.intercept_)) print('Score:%.2f'%cls.score(X_test,y_test)) X_train,X_test,y_train,y_test = load_data_classfication() test_LinearSVC(X_train,X_test,y_train,y_test)
def test_LinearSVC_loss(*data): X_train,X_test,y_train,y_test = data losses = ['hinge','squared_hinge'] for loss in losses: cls = svm.LinearSVC(loss=loss) cls.fit(X_train,y_train) print('loss:%s'%loss) print('Coefficients:%s,intercept%s'%(cls.coef_,cls.intercept_)) print('Score:%.2f'%cls.score(X_test,y_test)) X_train,X_test,y_train,y_test = load_data_classfication() test_LinearSVC_loss(X_train,X_test,y_train,y_test)
#考察罚项形式的影响 def test_LinearSVC_L12(*data): X_train,X_test,y_train,y_test = data L12 = ['l1','l2'] for p in L12: cls = svm.LinearSVC(penalty=p,dual=False) cls.fit(X_train,y_train) print('penalty:%s'%p) print('Coefficients:%s,intercept%s'%(cls.coef_,cls.intercept_)) print('Score:%.2f'%cls.score(X_test,y_test)) X_train,X_test,y_train,y_test = load_data_classfication() test_LinearSVC_L12(X_train,X_test,y_train,y_test)
#考察罚项系数C的影响 def test_LinearSVC_C(*data): X_train,X_test,y_train,y_test = data Cs = np.logspace(-2,1) train_scores = [] test_scores = [] for C in Cs: cls = svm.LinearSVC(C=C) cls.fit(X_train,y_train) train_scores.append(cls.score(X_train,y_train)) test_scores.append(cls.score(X_test,y_test)) fig = plt.figure() ax = fig.add_subplot(1,1,1) ax.plot(Cs,train_scores,label = 'Training score') ax.plot(Cs,test_scores,label = 'Testing score') ax.set_xlabel(r'C') ax.set_xscale('log') ax.set_ylabel(r'score') ax.set_title('LinearSVC') ax.legend(loc='best') plt.show() X_train,X_test,y_train,y_test = load_data_classfication() test_LinearSVC_C(X_train,X_test,y_train,y_test)
#非线性分类SVM #线性核 def test_SVC_linear(*data): X_train, X_test, y_train, y_test = data cls = svm.SVC(kernel='linear') cls.fit(X_train,y_train) print('Coefficients:%s,intercept%s'%(cls.coef_,cls.intercept_)) print('Score:%.2f'%cls.score(X_test,y_test)) X_train,X_test,y_train,y_test = load_data_classfication() test_SVC_linear(X_train,X_test,y_train,y_test)
#考察高斯核 def test_SVC_rbf(*data): X_train, X_test, y_train, y_test = data ###测试gamm### gamms = range(1, 20) train_scores = [] test_scores = [] for gamm in gamms: cls = svm.SVC(kernel='rbf', gamma=gamm) cls.fit(X_train, y_train) train_scores.append(cls.score(X_train, y_train)) test_scores.append(cls.score(X_test, y_test)) fig = plt.figure() ax = fig.add_subplot(1, 1, 1) ax.plot(gamms, train_scores, label='Training score', marker='+') ax.plot(gamms, test_scores, label='Testing score', marker='o') ax.set_xlabel(r'$\gamma$') ax.set_ylabel(r'score') ax.set_ylim(0, 1.05) ax.set_title('SVC_rbf') ax.legend(loc='best') plt.show() X_train,X_test,y_train,y_test = load_data_classfication() test_SVC_rbf(X_train,X_test,y_train,y_test)
#考察sigmoid核 def test_SVC_sigmod(*data): X_train, X_test, y_train, y_test = data fig = plt.figure() ###测试gamm### gamms = np.logspace(-2, 1) train_scores = [] test_scores = [] for gamm in gamms: cls = svm.SVC(kernel='sigmoid',gamma=gamm,coef0=0) cls.fit(X_train, y_train) train_scores.append(cls.score(X_train, y_train)) test_scores.append(cls.score(X_test, y_test)) ax = fig.add_subplot(1, 2, 1) ax.plot(gamms, train_scores, label='Training score', marker='+') ax.plot(gamms, test_scores, label='Testing score', marker='o') ax.set_xlabel(r'$\gamma$') ax.set_ylabel(r'score') ax.set_xscale('log') ax.set_ylim(0, 1.05) ax.set_title('SVC_sigmoid_gamm') ax.legend(loc='best') #测试r rs = np.linspace(0,5) train_scores = [] test_scores = [] for r in rs: cls = svm.SVC(kernel='sigmoid', gamma=0.01, coef0=r) cls.fit(X_train, y_train) train_scores.append(cls.score(X_train, y_train)) test_scores.append(cls.score(X_test, y_test)) ax = fig.add_subplot(1, 2, 2) ax.plot(rs, train_scores, label='Training score', marker='+') ax.plot(rs, test_scores, label='Testing score', marker='o') ax.set_xlabel(r'r') ax.set_ylabel(r'score') ax.set_ylim(0, 1.05) ax.set_title('SVC_sigmoid_r') ax.legend(loc='best') plt.show() X_train,X_test,y_train,y_test = load_data_classfication() test_SVC_sigmod(X_train,X_test,y_train,y_test)
看完上述内容是否对您有帮助呢?如果还想对相关知识有进一步的了解或阅读更多相关文章,请关注创新互联行业资讯频道,感谢您对创新互联的支持。