import torch import pandas as pd from sklearn.preprocessing import LabelEncoder, OneHotEncoder
时间: 2023-10-26 15:04:42 浏览: 41
这段代码导入了三个库:torch、pandas和sklearn。torch是PyTorch深度学习框架的Python接口,pandas是一个用于数据处理和分析的Python库,sklearn则是Python中流行的机器学习库。这段代码还使用了sklearn库中的两个类:LabelEncoder和OneHotEncoder。LabelEncoder用于将文本标签转换为数字标签,OneHotEncoder则用于将数字标签转换为独热编码。这两个类通常被用于数据预处理中的特征编码。
相关问题
import numpy as np import matplotlib.pyplot as plt import pandas as pd import torch import torch.nn as nn from setuptools.sandbox import save_path from sklearn.preprocessing import StandardScaler, MinMaxScaler from torch.utils.data import TensorDataset from tqdm import tqdm
这段代码是在导入一些常用的 Python 库和模块,包括:
- numpy:用于科学计算的 Python 库,提供了高性能的多维数组(ndarray)对象和相关的计算功能;
- matplotlib:用于绘制图表和数据可视化的 Python 库;
- pandas:用于数据处理和分析的 Python 库,提供了高性能、易用的数据结构和数据分析工具;
- torch:PyTorch 深度学习框架的 Python 接口,提供了张量(tensor)计算和自动求导功能;
- torch.nn:PyTorch 深度学习框架的神经网络模块,提供了常用的神经网络层和模型;
- StandardScaler 和 MinMaxScaler:用于数据标准化和归一化的类;
- TensorDataset:用于封装张量数据集的类;
- tqdm:用于实现进度条的 Python 模块。
这些库和模块在深度学习领域经常被使用,可以方便地进行数据处理、模型构建和训练等任务。
将冒号后面的代码改写成一个nn.module类:import pandas as pd import numpy as np from sklearn.preprocessing import MinMaxScaler import matplotlib.pyplot as plt from keras.models import Sequential from keras.layers import Dense, LSTM data1 = pd.read_csv("终极1.csv", usecols=[17], encoding='gb18030') df = data1.fillna(method='ffill') data = df.values.reshape(-1, 1) scaler = MinMaxScaler(feature_range=(0, 1)) data = scaler.fit_transform(data) train_size = int(len(data) * 0.8) test_size = len(data) - train_size train, test = data[0:train_size, :], data[train_size:len(data), :] def create_dataset(dataset, look_back=1): dataX, dataY = [], [] for i in range(len(dataset)-look_back-1): a = dataset[i:(i+look_back), 0] dataX.append(a) dataY.append(dataset[i + look_back, 0]) return np.array(dataX), np.array(dataY) look_back = 30 trainX, trainY = create_dataset(train, look_back) testX, testY = create_dataset(test, look_back) trainX = np.reshape(trainX, (trainX.shape[0], 1, trainX.shape[1])) testX = np.reshape(testX, (testX.shape[0], 1, testX.shape[1])) model = Sequential() model.add(LSTM(50, input_shape=(1, look_back), return_sequences=True)) model.add(LSTM(50)) model.add(Dense(1)) model.compile(loss='mean_squared_error', optimizer='adam') model.fit(trainX, trainY, epochs=6, batch_size=1, verbose=2) trainPredict = model.predict(trainX) testPredict = model.predict(testX) trainPredict = scaler.inverse_transform(trainPredict) trainY = scaler.inverse_transform([trainY]) testPredict = scaler.inverse_transform(testPredict) testY = scaler.inverse_transform([testY])
以下是将代码改写成一个nn.module类:
```python
import torch.nn as nn
import pandas as pd
import numpy as np
from sklearn.preprocessing import MinMaxScaler
import matplotlib.pyplot as plt
from keras.models import Sequential
from keras.layers import Dense, LSTM
class LSTMModel(nn.Module):
def __init__(self, input_size, hidden_size, output_size, num_layers, look_back):
super().__init__()
self.hidden_size = hidden_size
self.num_layers = num_layers
self.look_back = look_back
self.lstm1 = nn.LSTM(input_size, hidden_size, num_layers, batch_first=True)
self.lstm2 = nn.LSTM(hidden_size, hidden_size, num_layers, batch_first=True)
self.fc = nn.Linear(hidden_size, output_size)
def forward(self, x):
h0 = torch.zeros(self.num_layers, x.size(0), self.hidden_size).to(device)
c0 = torch.zeros(self.num_layers, x.size(0), self.hidden_size).to(device)
x = x.view(-1, self.look_back, 1)
out, _ = self.lstm1(x, (h0, c0))
out, _ = self.lstm2(out, (h0, c0))
out = self.fc(out[:, -1, :])
return out
data1 = pd.read_csv("终极1.csv", usecols=[17], encoding='gb18030')
df = data1.fillna(method='ffill')
data = df.values.reshape(-1, 1)
scaler = MinMaxScaler(feature_range=(0, 1))
data = scaler.fit_transform(data)
train_size = int(len(data) * 0.8)
test_size = len(data) - train_size
train, test = data[0:train_size, :], data[train_size:len(data), :]
def create_dataset(dataset, look_back=1):
dataX, dataY = [], []
for i in range(len(dataset)-look_back-1):
a = dataset[i:(i+look_back), 0]
dataX.append(a)
dataY.append(dataset[i + look_back, 0])
return np.array(dataX), np.array(dataY)
look_back = 30
trainX, trainY = create_dataset(train, look_back)
testX, testY = create_dataset(test, look_back)
trainX = np.reshape(trainX, (trainX.shape[0], trainX.shape[1], 1))
testX = np.reshape(testX, (testX.shape[0], testX.shape[1], 1))
device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
model = LSTMModel(input_size=1, hidden_size=50, output_size=1, num_layers=2, look_back=look_back).to(device)
criterion = nn.MSELoss()
optimizer = torch.optim.Adam(model.parameters(), lr=0.001)
num_epochs = 6
for epoch in range(num_epochs):
for i, (inputs, labels) in enumerate(zip(trainX, trainY)):
inputs = torch.from_numpy(inputs).float().to(device)
labels = torch.from_numpy(labels).float().to(device)
outputs = model(inputs)
loss = criterion(outputs, labels)
optimizer.zero_grad()
loss.backward()
optimizer.step()
if (i+1) % 100 == 0:
print(f'Epoch [{epoch+1}/{num_epochs}], Step [{i+1}/{len(trainX)}], Loss: {loss.item():.4f}')
model.eval()
trainPredict = scaler.inverse_transform(model(torch.from_numpy(trainX).float().to(device)).detach().cpu().numpy())
testPredict = scaler.inverse_transform(model(torch.from_numpy(testX).float().to(device)).detach().cpu().numpy())
trainY = scaler.inverse_transform([trainY])
testY = scaler.inverse_transform([testY])
```
这个类接受5个参数,分别是:
- input_size: LSTM的输入维度,对于这个例子来说,input_size=1
- hidden_size: LSTM的隐藏层维度
- output_size: LSTM的输出维度,对于这个例子来说,output_size=1
- num_layers: LSTM的层数
- look_back: 每个样本的序列长度,对于这个例子来说,look_back=30