everyone, I am using LSTM to predict the stock index of someday using the ones of 30 days before it as the input only. I think in this example, the size of LSTM input should be [10,30,1],so I use t_x=x.view(10,30,1)
to reshape the input. But there is an RuntimeError( shape '[10, 30, 1]' is invalid for input of size 150
) when I run the code below, could you please help me find what's the problem? Thank you:)
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import torch
import torch.nn as nn
import torch.nn.functional as F
from torch.utils.data import Dataset, DataLoader
from torch.utils.data import TensorDataset
dataread_df=pd.read_csv('D:/Desktop/399300.csv')
dataread_series=pd.Series(dataread_df['value'].values)
plt.plot(dataread_series)
plt.show()
def generate_data_df(series, n):
if len(series) <= n:
raise Exception("The Length of series is %d, while affect by (n=%d)." % (len(series), n))
df = pd.DataFrame()
for i in range(n):
df['x%d' % i] = series.tolist()[i:-(n - i)]
df['y'] = series.tolist()[n:]
return df
data_df = generate_data_df(dataread_series, 30)
data_numpy=np.array(data_df)
mean=np.mean(data_numpy)
std=np.std(data_numpy)
data_numpy = (data_numpy-mean)/std
train_size=int(len(data_numpy)*0.7)
test_size=len(data_numpy)-train_size
trainset_np=data_numpy[:train_size]
testset_np=data_numpy[train_size:]
train_x_np=trainset_np[:,:30]
train_y_np=trainset_np[:,30:]
test_x_np=testset_np[:,:30]
test_y_np=testset_np[:,30:]
train_x=torch.Tensor(train_x_np)
train_y=torch.Tensor(train_y_np)
test_x=torch.Tensor(test_x_np)
test_y=torch.Tensor(test_y_np)
trainset=TensorDataset(train_x,train_y)
testset=TensorDataset(test_x,test_y)
trainloader = DataLoader(trainset, batch_size=10, shuffle=True)
testloader=DataLoader(testset,batch_size=10,shuffle=True)
class Net(nn.Module):
def __init__(self):
super(Net,self).__init__()
self.rnn=nn.LSTM(input_size=1,hidden_size=64,num_layers=1,batch_first=True)
self.out=nn.Sequential(nn.Linear(64,1))
def forward(self,x):
r_out,(h_n,h_c)=self.rnn(x,None)
out=self.out(r_out[:,-1,:])
return out
rnn = Net()
print(rnn)
optimizer = torch.optim.Adam(rnn.parameters(), lr=0.0001)
criterion = nn.MSELoss()
train_correct=0
test_correct=0
train_total=0
test_total=0
prediction_list=[]
for epoch in range(10):
running_loss_train=0
running_loss_test=0
for i,(x1,y1) in enumerate(trainloader):
t_x1=x1.view(10,30,1)
output=rnn(t_x1)
loss_train=criterion(output,y1)
optimizer.zero_grad()
loss_train.backward()
optimizer.step()
running_loss_train+=loss_train.item()
for i,(x2,y2) in enumerate(testloader):
t_x2=x2.view(10,30,1)
prediction=rnn(t_x2)
loss_test=criterion(prediction,y2)
running_loss_test+=loss_test.item()
prediction_list.append(prediction)
print('Epoch {} Train Loss:{}, Test Loss:{}'.format(epoch+1,running_loss_train,running_loss_test))
prediction_list_plot=np.array(prediction_list)
plt.plot(test_y_np.flatten(),'r-',linewidth=0.1,label='real data')
plt.plot(prediction_list_plot.flatten(),'b-',linewidth=0.1,label='predicted data')
plt.show()
print('Finish training')
RuntimeError:
RuntimeError Traceback (most recent call last)
<ipython-input-3-fb8cb4c93775> in <module>
71 running_loss_test=0
72 for i,(x1,y1) in enumerate(trainloader):
---> 73 t_x1=x1.view(10,30,1)
74 output=rnn(t_x1)
75 loss_train=criterion(output,y1)
RuntimeError: shape '[10, 30, 1]' is invalid for input of size 150
Change this
t_x1=x1.view(10,30,1)
as
t_x1=x1.view(150,150,1)
and try it
From the documentation of view() method, 'The returned tensor shares the same data and must have the same number of elements , but may have a different size.'
x1 = torch.randn((150,))
t_x1 = x1.view(10,30,1)
RuntimeError: shape '[10, 30, 1]' is invalid for input of size 150
This is because 150 != 10 * 30. If you want to use 30 time-steps then your sample size should be 150 / 30 = 5. So, the correct way is
t_x1 = x1.view(5,30,1)
Given that you use batch_first=True
and assume a batch size of 10, (10, 30, 1)
as shape of the input is correct since it's (batch_size, seq_len, input_size)
.
The question is where the 150
comes from. What is the shape of x1
before trying to apply the .view(...)
? Can you check the following:
for i,(x1,y1) in enumerate(trainloader):
print(x1.shape)
...
Intuitively, it should something like (10, ???)
since you set 10 as batch size. Right now I assume that something with you training and test data is off.
The technical post webpages of this site follow the CC BY-SA 4.0 protocol. If you need to reprint, please indicate the site URL or the original address.Any question please contact:yoyou2525@163.com.