| import torch
|
| import torch.nn as nn
|
| import torch.optim as optim
|
|
|
|
|
| class SimpleNN(nn.Module):
|
| def __init__(self, input_size, hidden_size, output_size):
|
| super(SimpleNN, self).__init__()
|
| self.hidden = nn.Linear(input_size, hidden_size)
|
| self.activation = nn.ReLU()
|
| self.output = nn.Linear(hidden_size, output_size)
|
|
|
| def forward(self, x):
|
| x = self.hidden(x)
|
| x = self.activation(x)
|
| x = self.output(x)
|
| return x
|
|
|
|
|
| input_size = 10
|
| hidden_size = 4
|
| output_size = 4
|
|
|
| model = SimpleNN(input_size, hidden_size, output_size)
|
| print(model)
|
|
|
|
|
| criterion = nn.CrossEntropyLoss()
|
| optimizer = optim.Adam(model.parameters(), lr=0.01)
|
|
|
|
|
| X_train = torch.rand(10, input_size)
|
| y_train = torch.tensor([0, 1, 2, 1, 0, 1, 2, 1, 0, 1])
|
|
|
|
|
| for i in range(1):
|
| optimizer.zero_grad()
|
| outputs = model(X_train)
|
| loss = criterion(outputs, y_train)
|
| loss.backward()
|
| optimizer.step()
|
|
|
|
|
| print(f"Output:\n{outputs}")
|
| print(f"Loss: {loss.item()}")
|
|
|