import torch import numpy as np import matplotlib.pyplot as plt from torch import nn, optim from torch.autograd import Variable x_data = np.random.rand(100) noise = np.random.normal(0, 0.01, x_data.shape) y_data = 0.1*x_data+0.2+noise # plt.scatter(x_data, y_data) # plt.show() x_data = x_data.reshape(-1, 1) y_data = y_data.reshape(-1, 1) # 把numpy数据变成张量tensor数据 x_data = torch.FloatTensor(x_data) y_data = torch.FloatTensor(y_data) # 构建网络模型 inputs = Variable(x_data) target = Variable(y_data) class LinearRegression(nn.Module): # 初始化,定义网络结构 # 一般把网络中具有可学习的参数的层放在初始化的里面,__int__()中 def __int__(self): super(LinearRegression, self).__init__() self.fc = nn.Linear(1, 1) # 定义网络计算 def forward(self, x): out = self.fc(x) return out # 实例化模型 model = LinearRegression() # 定义代价函数 mse_loss = nn.MSELoss() # 定义优化器 optimizer = optim.SGD(model.parameters(), lr=0.1) # 查看模型参数 for name, parameters in model.named_parameters(): print('name:{},parameters:{}'.format(name, parameters))
时间:2023-03-06 07:23:06 浏览:61
这段代码使用了Python中的一些库和模块,包括torch、numpy和matplotlib.pyplot,还有torch中的nn、optim模块和Variable函数。
首先,通过numpy库生成了一个包含100个随机数的数组x_data,同时也生成了一些符合正态分布的噪声noise。然后,根据公式y_data = 0.1*x_data + 0.2*noise,生成了相应的y_data数组。
接下来,使用torch中的nn模块定义神经网络模型,使用optim模块定义优化器,然后使用autograd中的Variable函数将x_data和y_data转化为可以进行自动求导的变量。这些步骤通常是构建神经网络模型并进行训练的前置准备。
最后,这段代码可以用来训练一个神经网络模型,以便预测y_data值。
相关问题
import torch import torch.nn as nn import numpy as np import matplotlib.pyplot as plt from torch import autograd """ 用神经网络模拟微分方程,f(x)'=f(x),初始条件f(0) = 1 """ class Net(nn.Module): def __init__(self, NL, NN): # NL n个l(线性,全连接)隐藏层, NN 输入数据的维数, # NL是有多少层隐藏层 # NN是每层的神经元数量 super(Net, self).__init__() self.input_layer = nn.Linear(1, NN) self.hidden_layer = nn.Linear(NN,int(NN/2)) ## 原文这里用NN,我这里用的下采样,经过实验验证,“等采样”更优。更多情况有待我实验验证。 self.output_layer = nn.Linear(int(NN/2), 1) def forward(self, x): out = torch.tanh(self.input_layer(x)) out = torch.tanh(self.hidden_layer(out)) out_final = self.output_layer(out) return out_final net=Net(4,20) # 4层 20个 mse_cost_function = torch.nn.MSELoss(reduction='mean') # Mean squared error 均方误差求 optimizer = torch.optim.Adam(net.parameters(),lr=1e-4) # 优化器 def ode_01(x,net): y=net(x) y_x = autograd.grad(y, x,grad_outputs=torch.ones_like(net(x)),create_graph=True)[0] return y-y_x # y-y' = 0 # requires_grad=True).unsqueeze(-1) plt.ion() # 动态图 iterations=200000 for epoch in range(iterations): optimizer.zero_grad() # 梯度归0 ## 求边界条件的损失函数 x_0 = torch.zeros(2000, 1) y_0 = net(x_0) mse_i = mse_cost_function(y_0, torch.ones(2000, 1)) # f(0) - 1 = 0 ## 方程的损失函数 x_in = np.random.uniform(low=0.0, high=2.0, size=(2000, 1)) pt_x_in = autograd.Variable(torch.from_numpy(x_in).float(), requires_grad=True) # x 随机数 pt_y_colection=ode_01(pt_x_in,net) pt_all_zeros= autograd.Variable(torch.from_numpy(np.zeros((2000,1))).float(), requires_grad=False) mse_f=mse_cost_function(pt_y_colection, pt_all_zeros) # y-y' = 0 loss = mse_i + mse_f loss.backward() # 反向传播 optimizer.step() # 优化下一步。This is equivalent to : theta_new = theta_old - alpha * derivative of J w.r.t theta if epoch%1000==0: y = torch.exp(pt_x_in) # y 真实值 y_train0 = net(pt_x_in) # y 预测值 print(epoch, "Traning Loss:", loss.data) print(f'times {epoch} - loss: {loss.item()} - y_0: {y_0}') plt.cla() plt.scatter(pt_x_in.detach().numpy(), y.detach().numpy()) plt.scatter(pt_x_in.detach().numpy(), y_train0.detach().numpy(),c='red') plt.pause(0.1)
这是一段 Python 代码,主要是在导入一些库和定义一个神经网络模型类 "Net"。代码中导入了 torch、torch.nn、numpy 和 matplotlib.pyplot 库,并在 "Net" 类中进行了一些初始化。代码还提到了一个微分方程:f(x)' = f(x), 初始条件f(0) = 1, 用神经网络模拟。
帮我把这段代码从tensorflow框架改成pytorch框架: import tensorflow as tf import os import numpy as np import matplotlib.pyplot as plt os.environ["CUDA_VISIBLE_DEVICES"] = "0" base_dir = 'E:/direction/datasetsall/' train_dir = os.path.join(base_dir, 'train_img/') validation_dir = os.path.join(base_dir, 'val_img/') train_cats_dir = os.path.join(train_dir, 'down') train_dogs_dir = os.path.join(train_dir, 'up') validation_cats_dir = os.path.join(validation_dir, 'down') validation_dogs_dir = os.path.join(validation_dir, 'up') batch_size = 64 epochs = 50 IMG_HEIGHT = 128 IMG_WIDTH = 128 num_cats_tr = len(os.listdir(train_cats_dir)) num_dogs_tr = len(os.listdir(train_dogs_dir)) num_cats_val = len(os.listdir(validation_cats_dir)) num_dogs_val = len(os.listdir(validation_dogs_dir)) total_train = num_cats_tr + num_dogs_tr total_val = num_cats_val + num_dogs_val train_image_generator = tf.keras.preprocessing.image.ImageDataGenerator(rescale=1. / 255) validation_image_generator = tf.keras.preprocessing.image.ImageDataGenerator(rescale=1. / 255) train_data_gen = train_image_generator.flow_from_directory(batch_size=batch_size, directory=train_dir, shuffle=True, target_size=(IMG_HEIGHT, IMG_WIDTH), class_mode='categorical') val_data_gen = validation_image_generator.flow_from_directory(batch_size=batch_size, directory=validation_dir, target_size=(IMG_HEIGHT, IMG_WIDTH), class_mode='categorical') sample_training_images, _ = next(train_data_gen) model = tf.keras.models.Sequential([ tf.keras.layers.Conv2D(16, 3, padding='same', activation='relu', input_shape=(IMG_HEIGHT, IMG_WIDTH, 3)), tf.keras.layers.MaxPooling2D(), tf.keras.layers.Conv2D(32, 3, padding='same', activation='relu'), tf.keras.layers.MaxPooling2D(), tf.keras.layers.Conv2D(64, 3, padding='same', activation='relu'), tf.keras.layers.MaxPooling2D(), tf.keras.layers.Flatten(), tf.keras.layers.Dense(256, activation='relu'), tf.keras.layers.Dense(2, activation='softmax') ]) model.compile(optimizer='adam', loss=tf.keras.losses.BinaryCrossentropy(from_logits=True), metrics=['accuracy']) model.summary() history = model.fit_generator( train_data_gen, steps_per_epoch=total_train // batch_size, epochs=epochs, validation_data=val_data_gen, validation_steps=total_val // batch_size ) # 可视化训练结果 acc = history.history['accuracy'] val_acc = history.history['val_accuracy'] loss = history.history['loss'] val_loss = history.history['val_loss'] epochs_range = range(epochs) model.save("./model/timo_classification_128_maxPool2D_dense256.h5")
import torch import torch.nn as nn import torch.optim as optim from torch.utils.data import DataLoader from torchvision import datasets, transforms import os BATCH_SIZE = 64 EPOCHS = 50 IMG_HEIGHT = 128 IMG_WIDTH = 128 train_transforms = transforms.Compose([ transforms.Resize((IMG_HEIGHT,IMG_WIDTH)), transforms.RandomHorizontalFlip(), transforms.ToTensor(), transforms.Normalize([0.5,0.5,0.5], [0.5,0.5,0.5])]) test_transforms = transforms.Compose([ transforms.Resize((IMG_HEIGHT,IMG_WIDTH)), transforms.ToTensor(), transforms.Normalize([0.5,0.5,0.5], [0.5,0.5,0.5])]) base_dir = 'E:/direction/datasetsall/' train_dir = os.path.join(base_dir, 'train_img/') validation_dir = os.path.join(base_dir, 'val_img/') train_cats_dir = os.path.join(train_dir, 'down') train_dogs_dir = os.path.join(train_dir, 'up') validation_cats_dir = os.path.join(validation_dir, 'down') validation_dogs_dir = os.path.join(validation_dir, 'up') train_dataset = datasets.ImageFolder(train_dir, transform=train_transforms) train_loader = DataLoader(train_dataset, batch_size=BATCH_SIZE, shuffle=True) test_dataset = datasets.ImageFolder(validation_dir, transform=test_transforms) test_loader = DataLoader(test_dataset, batch_size=BATCH_SIZE, shuffle=False) device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu") model = nn.Sequential( nn.Conv2d(3, 16, 3, padding=1), nn.ReLU(), nn.MaxPool2d(2), nn.Conv2d(16, 32, 3, padding=1), nn.ReLU(), nn.MaxPool2d(2), nn.Conv2d(32, 64, 3, padding=1), nn.ReLU(), nn.MaxPool2d(2), nn.Flatten(), nn.Linear(64 * (IMG_HEIGHT // 8) * (IMG_WIDTH // 8), 256), nn.ReLU(), nn.Linear(256, 2), nn.Softmax(dim=1) ) model.to(device) criterion = nn.CrossEntropyLoss() optimizer = optim.Adam(model.parameters(), lr=0.001) for epoch in range(EPOCHS): train_loss = 0.0 train_acc = 0.0 model.train() for images, labels in train_loader: images = images.to(device) labels = labels.to(device) optimizer.zero_grad() outputs = model(images) loss = criterion(outputs, labels) loss.backward() optimizer.step() train_loss += loss.item() * images.size(0) _, preds = torch.max(outputs, 1) train_acc += torch.sum(preds == labels.data) train_loss = train_loss / len(train_loader.dataset) train_acc = train_acc / len(train_loader.dataset) print('Epoch: {} \tTraining Loss: {:.6f} \tTraining Accuracy: {:.6f}'.format(epoch+1, train_loss,train_acc)) with torch.no_grad(): test_loss = 0.0 test_acc = 0.0 model.eval() for images, labels in test_loader: images = images.to(device) labels = labels.to(device) outputs = model(images) loss = criterion(outputs, labels) test_loss += loss.item() * images.size(0) _, preds = torch.max(outputs, 1) test_acc += torch.sum(preds == labels.data) test_loss = test_loss / len(test_loader.dataset) test_acc = test_acc / len(test_loader.dataset) print('Test Loss: {:.6f} \tTest Accuracy: {:.6f}'.format(test_loss,test_acc))