发布时间:2023-05-14 17:30
这次的任务是节点分类。使用的是Cora数据集,该数据集由2708篇论文,及它们之间的引用关系构成的5429条边组成。这些论文根据主题被划分为7类,分别是神经网络、牵强化学习、规则学习、概率方法、遗传算法、理论研究、案例相关。每篇论文的特征是通过词袋模型得到的,维度为1433,每一维表示一个词,1表示该词在这篇文章中出现过,0表示未出现。
导入需要的包:
import itertools
import os
import os.path as osp
import pickle
import urllib
from collections import namedtuple
import numpy as np
import scipy.sparse as sp
import torch
import torch.nn as nn
import torch.nn.functional as F
import torch.nn.init as init
import torch.optim as optim
import matplotlib.pyplot as plt
%matplotlib inline
CoraData类用来对数据进行预处理,主要包括下载数据、规范化数据并进行缓存以备重复使用。
###################################################################################
# 数据准备 #
###################################################################################
## 用于保存处理好的数据
Data = namedtuple('Data', ['x', 'y', 'adjacency', 'train_mask', 'val_mask','test_mask'])
class CoraData(object):
download_url = "http://github.com/kimiyoung/planetoid/raw/master/data"
filenames = ["ind.cora.{}".format(name) for name in ['x', 'tx', 'allx', 'y', 'ty', 'ally', 'graph','test.index']]
######### 数据准备部分——数据下载 ##################################
def __init__(self, data_root = "cora", rebuild = False):
"""
包括数据下载、处理、加载等功能
当数据的缓存文件存在时,使用缓存文件,否则将下载、处理、并缓存到磁盘
Args:
-------
data_root: string, optional
存放数据的目录,原始数据路径:{data_root}/raw
缓存存放路径:{data_root}/processed_cora.pkl
rebuild: boolean, optional
是否需要重新构建数据集,当设为True时,如果缓存数据存在也会重建数据
"""
self.data_root = data_root
save_file = osp.join(self.data_root, "processed_cora.pkl")
if osp.exists(save_file) and not rebuild:
print("Using Cached file:{}".format(save_file))
self._data = pickle.load(open(save_file, "rb"))
else:
self.maybe_download()
self._data = self.process_data()
with open(save_file, "wb") as f:
pickle.dump(self.data, f)
print("Cached file: {}".format(save_file))
@property
def data(self):
"""
返回Data数据对象,包括x,y,adjacency, train_mask, valid_mask, test_mask
"""
return self._data
def maybe_download(self):
save_path = osp.join(self.data_root, "raw")
for name in self.filenames:
if not osp.exists(osp.join(save_path, name)):
self.download_data("{}/{}".format(self.download_url, name), save_path)
@staticmethod
def download_data(url, save_path):
"""
数据下载工具,当原始数据不存在时将会进行下载
"""
if not osp.exists(save_path):
os.makedirs(save_path)
data = urllib.request.urlopen(url)
filename = osp.basename(url)
with open(osp.join(save_path, filename), 'wb') as f:
f.write(data.read())
return True
######### 下面是数据处理的部分 ##################################
def process_data(self):
"""
处理数据,得到节点特征和标签、邻接矩阵、训练集、验证集以及测试集
"""
print("Process data ...")
_, tx, allx, y, ty, ally, graph, test_index = [self.read_data(osp.join(self.data_root,"raw", name)) for name in self.filenames]
train_index = np.arange(y.shape[0]) #
val_index = np.arange(y.shape[0], y.shape[0]+500)
sorted_test_index = sorted(test_index)
#将训练集和测试集结合得到总的数据集
x = np.concatenate((allx, tx), axis = 0)
# 将训练集和测试集结合得到总的数据集,并且得到标签(ally和ty实质上是y的向量
y = np.concatenate((ally, ty), axis = 0).argmax(axis = 1)
x[test_index] = x[sorted_test_index]
y[test_index] = y[sorted_test_index]
num_nodes = x.shape[0]
train_mask = np.zeros(num_nodes, dtype = np.bool)
val_mask = np.zeros(num_nodes, dtype = np.bool)
test_mask = np.zeros(num_nodes, dtype = np.bool)
train_mask[train_index] = True
val_mask[val_index] = True
test_mask[test_index] = True
adjacency = self.build_adjacency(graph)
print("Node's feature shape: ", x.shape)
print("Node's label shape: ", y.shape)
print("Adjacency's shape: ", adjacency.shape)
print("Number of training nodes: ", train_mask.sum())
print("Number of validation nodes: ", val_mask.sum())
print("Number of test nodes: ", test_mask.sum())
return Data(x=x, y=y, adjacency= adjacency, train_mask=train_mask, val_mask=val_mask, test_mask=test_mask)
@staticmethod
def build_adjacency(adj_dict):
"""
根据邻接列表创建邻接矩阵
"""
edge_index = []
num_nodes = len(adj_dict)
for src,dst in adj_dict.items():
edge_index.extend([src, v] for v in dst)
edge_index.extend([v, src] for v in dst)
# 由于上述得到的结果中存在重复的边,删掉这些重复的边
edge_index = list(k for k, _ in itertools.groupby(sorted(edge_index)))
edge_index = np.asarray(edge_index)
adjacency = sp.coo_matrix((np.ones(len(edge_index)), (edge_index[:, 0], edge_index[:, 1])),shape=(num_nodes, num_nodes), dtype = "float32")
return adjacency
@staticmethod
def read_data(path):
"""
使用不同的方式读取原始数据以进一步处理
"""
name = osp.basename(path)
if name == "ind.cora.test.index":
out = np.genfromtxt(path, dtype = "int64")
return out
else:
out = pickle.load(open(path, 'rb'), encoding = "latin1")
out = out.toarray() if hasattr(out, "toarray") else out
return out
# 规范化邻接矩阵
def normalization(adjacency):
"""
计算 L=D^-0.5 * (A+I) * D^-0.5
"""
adjacency += sp.eye(adjacency.shape[0]) #增加自连接
degree = np.array(adjacency.sum(1))
d_hat = sp.diags(np.power(degree, -0.5).flatten())
return d_hat.dot(adjacency).dot(d_hat).tocoo()
# 参数设置
device = "cuda" if torch.cuda.is_available() else "cpu"
# 加载数据,并转换为torch.Tensor
print('='*20)
print('加载数据')
dataset = CoraData().data
x = dataset.x / dataset.x.sum(1, keepdims = True) #归一化数据,使得每一行和为1
tensor_x = torch.from_numpy(x).to(device)
tensor_y = torch.from_numpy(dataset.y).to(device)
tensor_train_mask = torch.from_numpy(dataset.train_mask).to(device)
tensor_val_mask = torch.from_numpy(dataset.val_mask).to(device)
tensor_test_mask = torch.from_numpy(dataset.test_mask).to(device)
normalization_adjacency = normalization(dataset.adjacency) #规范化邻接矩阵
indices = torch.from_numpy(np.asarray([normalization_adjacency.row, normalization_adjacency.col]).astype('int64')).long()
values = torch.from_numpy(normalization_adjacency.data.astype(np.float32))
tensor_adjacency = torch.sparse.FloatTensor(indices, values, (2708, 2708)).to(device)
Python namedtuple
namedtuple是python元组的升级版本——具名元组。类似于简化定义一个类及其属性。
在这里,数据集被预处理为包括以下几个部分的数据形式:
scipy.sparse.coo_matrix
;数据集:https://github.com/kimiyoung/planetoid
下载的数据集中:
tx
:测试实例的特征向量ty
:测试实例的标签的one-hot编码test.index
:the indices of test instances in graph
, for the inductive settingally
: allx
中所有实例的标签allx
, the feature vectors of both labeled and unlabeled training instances (a superset of x)graph
, 一个字典,其格式是{index: [index_of_neighbor_nodes]}
根据GCN的定义 X ′ = σ ( L ~ s y m X W ) X'=\sigma(\tilde{L}_{sym}XW) X′=σ(L~symXW)来定义GCN层。由于邻接矩阵是稀疏矩阵,为了提高运算效率,使用稀疏矩阵乘法。
###################################################################################
# GCN层定义 #
###################################################################################
class GraphConvolution(nn.Module):
def __init__(self, input_dim, output_dim, use_bias = True):
"""
图卷积:L*X*\theta
Args:
-------
input_dim: int
节点输入特征的维度
output_dim: int
输出特征维度
use_bias: bool, optional
是否使用偏置
"""
super(GraphConvolution, self).__init__()
self.input_dim = input_dim
self.output_dim = output_dim
self.use_bias = use_bias
self.weight = nn.Parameter(torch.Tensor(input_dim, output_dim))
if self.use_bias:
self.bias = nn.Parameter(torch.Tensor(output_dim))
else:
self.register_parameter('bias', None)
self.reset_parameters()
def reset_parameters(self):
init.kaiming_uniform_(self.weight)
if self.use_bias:
init.zeros_(self.bias)
def forward(self, adjacency, input_feature):
"""
邻接矩阵是系数矩阵,因此在计算时使用稀疏矩阵乘法
Args:
-------
adjacency: torch.sparse.FloatTensor
邻接矩阵
input_feature: torch.Tensor
输入特征
"""
support = torch.mm(input_feature, self.weight)
output = torch.sparse.mm(adjacency, support)
if self.use_bias:
output += self.bias
return output
有了数据和GCN层,就可以构建模型进行训练了。我们定义一个两层的GCN,其中输入维度为1433,隐藏层维度设置为16,最后一层GCN将输出维度变为类别7,激活函数使用ReLU。
###################################################################################
# 模型构建 #
###################################################################################
class GcnNet(nn.Module):
"""
定义一个包含两层GraphConvolution的模型
"""
def __init__(self, input_dim = 1433):
super(GcnNet,self).__init__()
self.gcn1 = GraphConvolution(input_dim, 16)
self.gcn2 = GraphConvolution(16, 7)
def forward(self, adjacency, feature):
h = F.relu(self.gcn1(adjacency, feature))
logits = self.gcn2(adjacency, h)
return logits
#超参数定义
learning_rate = 0.1
weight_decay = 5e-4
epoches = 200
# 模型定义(模型实例化、损失函数定义、优化器定义)
model = GcnNet().to(device)
criterion = nn.CrossEntropyLoss().to(device)
optimizer = optim.Adam(model.parameters(), lr = learning_rate, weight_decay = weight_decay)
###################################################################################
# 模型训练与测试 #
###################################################################################
def train():
loss_history = []
val_acc_history = []
train_acc_history = []
model.train()
train_y = tensor_y[tensor_train_mask]
for epoch in range(epoches):
logits = model(tensor_adjacency, tensor_x) #前向传播
train_mask_logits = logits[tensor_train_mask] # 只选择训练节点进行监督
loss = criterion(train_mask_logits, train_y) # 计算损失值
optimizer.zero_grad()
loss.backward() # 反向传播计算参数的梯度
optimizer.step() # 使用优化方法进行梯度更新
train_acc = test(tensor_train_mask) # 计算当前模型在训练集上的准确率
val_acc = test(tensor_val_mask) # 计算当前模型在验证集上的准确率
# 记录训练过程中的损失值和准确率的变化,用于画图
loss_history.append(loss.item())
train_acc_history.append(train_acc.item())
val_acc_history.append(val_acc.item())
print("Epoch {:03d}: Loss {:.4f}, TrainAcc {:.4f}, ValAcc {:.4f}".format(epoch, loss.item(), train_acc.item(), val_acc.item()))
return loss_history, train_acc_history, val_acc_history
def test(mask):
model.eval()
with torch.no_grad():
logits = model(tensor_adjacency, tensor_x)
test_mask_logits = logits[mask]
predict_y = test_mask_logits.max(1)[1]
accuracy = torch.eq(predict_y, tensor_y[mask]).float().mean()
return accuracy
def plot_loss_with_acc(loss_history, val_acc_history):
fig = plt.figure()
ax1 = fig.add_subplot(111)
ax1.plot(range(len(loss_history)), loss_history,
c=np.array([255, 71, 90]) / 255.)
plt.ylabel('Loss')
ax2 = fig.add_subplot(111, sharex=ax1, frameon=False)
ax2.plot(range(len(val_acc_history)), val_acc_history,
c=np.array([79, 179, 255]) / 255.)
ax2.yaxis.tick_right()
ax2.yaxis.set_label_position("right")
plt.ylabel('ValAcc')
plt.xlabel('Epoch')
plt.title('Training Loss & Validation Accuracy')
plt.show()
############# 开始训练 ###################
print('='*15)
print('开始训练')
loss, train_acc, val_acc = train()#每个epoch 模型在训练集上的loss 和验证集上的准确率
# 可视化展示结果
plot_loss_with_acc(loss, val_acc)
#计算最后训练好的模型在测试集上准确率
test_acc = test(tensor_test_mask)
print("Test accuarcy: ", test_acc.item())
python字典中添加、修改数据、删除数据和遍历数据、enumerate函数和公共方法
linux克隆虚拟机、开机修改系统相关配置、修改 /etc/hostname ,修改主机名、查看network服务和NetworkManager服务
kafka笔记3--快速部署KRaft版本的kafka3.1.1
笔记:STM32启动文件startup_stm32f10x_xx.s
MacBook M1安装Numpy,Pandas,Matplotlib
让控件如此丝滑Scroller和VelocityTracker的API讲解与实战——Android高级UI