paranormal-or-skeptic-ISI-p.../Feed-Forward.py

132 lines
3.3 KiB
Python

#!/usr/bin/env python
# coding: utf-8
# In[2]:
import torch
import gensim.downloader as downloader
import pandas as pd
import csv
from nltk.tokenize import word_tokenize as tokenize
import numpy as np
# In[7]:
class NeuralNetworkModel(torch.nn.Module):
def __init__(self, input_size, hidden_size, num_classes):
super(NeuralNetworkModel, self).__init__()
self.fc1 = torch.nn.Linear(input_size,hidden_size)
self.fc2 = torch.nn.Linear(hidden_size,num_classes)
def forward(self,x):
x = self.fc1(x)
x = torch.relu(x)
x = self.fc2(x)
x = torch.sigmoid(x)
return x
# In[4]:
w2v = downloader.load('word2vec-google-news-300')
# In[9]:
#model + settings
nn = NeuralNetworkModel(300,300,1)
crit = torch.nn.BCELoss()
opti = torch.optim.SGD(nn.parameters(), lr=0.08)
BATCH_SIZE = 5
epochs = 5
# In[12]:
#trening
#wczytanie danych
train_data_in = pd.read_csv('train/in.tsv.xz', compression='xz', header=None, error_bad_lines=False, quoting=csv.QUOTE_NONE, sep='\t', nrows=3000)
train_data_ex = pd.read_csv('train/expected.tsv', header=None, error_bad_lines=False, quoting=csv.QUOTE_NONE, sep='\t', nrows=3000)
#preprocessing
train_in = train_data_in[0].str.lower()
train_in = [tokenize(line) for line in train_in]
train_in = [np.mean([w2v[x] for x in data if x in w2v] or [np.zeros(300)], axis=0) for data in train_in]
train_ex = train_data_ex[0]
for epoch in range(epochs):
nn.train()
for i in range(0,train_data_ex.shape[0],BATCH_SIZE):
x = train_in[i:i + BATCH_SIZE]
x = torch.tensor(x)
y = train_ex[i:i + BATCH_SIZE]
y = torch.tensor(y.astype(np.float32).to_numpy()).reshape(-1,1)
opti.zero_grad()
y_pred = nn(x.float())
loss = crit(y_pred,y)
loss.backward()
opti.step()
# In[27]:
#dev-0 predict
#wczytanie danych
dev0_data = pd.read_csv('dev-0/in.tsv.xz', compression='xz', header=None, error_bad_lines=False, quoting=csv.QUOTE_NONE, sep='\t')
dev0_data = dev0_data[0].str.lower()
dev0_data = [tokenize(line) for line in dev0_data]
dev0_data = [np.mean([w2v[x] for x in data if x in w2v] or [np.zeros(300)], axis=0) for data in dev0_data]
dev0_y=[]
nn.eval()
with torch.no_grad():
for i in range(0, len(dev0_data), BATCH_SIZE):
x = dev0_data[i:i + BATCH_SIZE]
x = torch.tensor(x)
dev0_y_pred = nn(x.float())
dev0_y_prediction = (dev0_y_pred > 0.5)
dev0_y.extend(dev0_y_prediction)
#zapis wyników
np.asarray(dev0_y, dtype=np.int32).tofile('dev-0/out.tsv', sep='\n')
# In[28]:
#test-A predict
#wczytanie danych
testA_data = pd.read_csv('test-A/in.tsv.xz', compression='xz', header=None, quoting=csv.QUOTE_NONE, sep='\t')
testA_data = testA_data[0].str.lower()
testA_data = [tokenize(line) for line in testA_data]
testA_data = [np.mean([w2v[x] for x in data if x in w2v] or [np.zeros(300)], axis=0) for data in testA_data]
testA_y=[]
nn.eval()
with torch.no_grad():
for i in range(0, len(testA_data), BATCH_SIZE):
x = testA_data[i:i + BATCH_SIZE]
x = torch.tensor(x)
testA_y_pred = nn(x.float())
testA_y_prediction = (testA_y_pred > 0.5)
testA_y.extend(testA_y_prediction)
#zapis wyników
np.asarray(testA_y, dtype=np.int32).tofile('test-A/out.tsv', sep='\n')