模拟退火算法之特征选择的python实现(一)

目录

1. 模拟退火算法实现步骤 

2. python实现

3. 实验结果

4. 参考文献


模拟退火算法的基本原理在这里就不一一赘述了, 关于原理,可以参考百度百科博客1博客2

在本节按照基本实现步骤实现模拟退火算法, 对于模拟退火算法的高级封装(类封装), 可以参考模拟退火算法之特征选择的python实现(二)

1. 模拟退火算法实现步骤 

2. python实现

import numpy as np
from sklearn.preprocessing import StandardScaler
from sklearn.neural_network import MLPRegressor
from sklearn.metrics import mean_squared_error
import scipy.io as sio
from sklearn.model_selection import train_test_split


def get_data_subset(x_data, columns):
    return x_data[:, columns]


def get_initial_solution(feature_size, selected_feature):
    sol = np.arange(feature_size-1)
    np.random.shuffle(sol)
    return sol[:selected_feature]


def get_neighbor(current_solution, feature_size, temperature):
    all_features = range(feature_size-1)
    selected = current_solution
    not_selected = np.setdiff1d(all_features, selected)

    # swap one selected feature with one non-selected feature
    num_swaps = int(min(np.ceil(np.abs(np.random.normal(0, 0.1*len(selected)*temperature))), np.ceil(0.1*len(selected))))
    feature_out = np.random.randint(0, len(selected), num_swaps)  # 产生num_swaps个样本索引(从range(len(selected))中)
    selected = np.delete(selected, feature_out)

    feature_in = np.random.randint(0, len(not_selected), num_swaps)  # 产生num_swaps个样本索引(从range(len(not_selected))中)
    selected = np.append(selected, not_selected[feature_in])

    return selected


def get_cost(solution, x_train, x_test, y_train, y_test):
    x_scale, y_scale = StandardScaler(), StandardScaler()
    x_train_scaled = x_scale.fit_transform(x_train)
    x_test_scaled = x_scale.transform(x_test)
    y_train_scaled = y_scale.fit_transform(y_train.reshape(-1, 1))
    y_test_scaled = y_scale.transform(y_test.reshape(-1, 1))
    limited_train_data = get_data_subset(x_train_scaled, solution)
    limited_test_data = get_data_subset(x_test_scaled, solution)
    model = MLPRegressor(hidden_layer_sizes=43)
    model.fit(limited_train_data, y_train_scaled.ravel())
    y_test_pred = model.predict(limited_test_data)
    y_test_pred = y_scale.inverse_transform(y_test_pred)
    return round(mean_squared_error(y_test, y_test_pred), 4)


def get_probability(temperature, delta_cost):
    return np.exp(delta_cost/temperature)


def simulated_annealing(initT, minT, alpha, iterations, features, x_train, x_test, y_train, y_test):
    temperature = initT  # 当前温度
    solution = get_initial_solution(feature_size=features_size, selected_feature=sel_feature)
    cost = get_cost(solution, x_train, x_test, y_train, y_test)
    temp_history = [temperature]
    best_cost_history = []
    best_solution_history = []
    best_cost = cost
    best_solution = solution
    while temperature > minT:
        for k in range(iterations):
            next_solution = get_neighbor(solution, features, temperature)
            next_cost = get_cost(next_solution, x_train, x_test, y_train, y_test)

            probability = 0
            if next_cost > cost:  # 计算向差方向移动的概率 (即移动后的解比当前解要差)
                probability = get_probability(temperature, cost-next_cost)
            if next_cost < cost or np.random.random() < probability:  # 朝着最优解移动或以一定概率向差方向移动
                cost = next_cost
                solution = next_solution
            if next_cost < best_cost:  # 最优值和最优解
                best_cost = cost
                best_solution = solution
        print("当前温度:", round(temperature, 2))
        print("当前温度下最好的得分:", best_cost)
        print("当前温度下波长数量:", len(solution))
        temperature *= alpha
        temp_history.append(temperature)
        best_cost_history.append(best_cost)
        best_solution_history.append(best_solution)
    return temp_history, best_cost_history, best_solution_history


# 1.数据获取
mat = sio.loadmat('NDFNDF_smote.mat')
data = mat['NDFNDF_smote']
x, y = data[:, :50], data[:, 50]
print('原始数据大小:', x.shape, y.shape)

# 2.样本集划分和预处理
x_train, x_test, y_train, y_test = train_test_split(x, y, test_size=0.2, random_state=1)

features_size = x.shape[1]
sel_feature = 5
temp_his, best_cost_his, best_solution_his = simulated_annealing(initT=100,
                                                                 minT=1,
                                                                 alpha=0.95,
                                                                 iterations=50,
                                                                                 
features=features_size,
                                                                 x_train=x_train,
                                                                 x_test=x_test,
                                                                 y_train=y_train,
                                                                 y_test=y_test)

3. 实验结果

4. 参考文献

谢云. 模拟退火算法的原理及实现. 1999. PhD Thesis.

陈华根, 吴健生, 王家林, & 陈冰. (2004). 模拟退火算法机理研究 (Doctoral dissertation).

  • 2
    点赞
  • 39
    收藏
    觉得还不错? 一键收藏
  • 3
    评论
评论 3
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值