-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathexp_IMDB_PAIRS.py
More file actions
61 lines (44 loc) · 1.64 KB
/
exp_IMDB_PAIRS.py
File metadata and controls
61 lines (44 loc) · 1.64 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
import numpy as np
import torch
import torch.nn as nn
from datasets import load_dataset
from scipy.stats import norm
from torch.utils.data import DataLoader
from transformers import AutoTokenizer
from core import PAIRS, multi_evaluate, exp_aggregator, IMDBModel, IMDBDataset
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
def experiment(num_neurons, batch_size):
imdb_dataset = load_dataset("imdb")
tokenizer = AutoTokenizer.from_pretrained("tokenizer")
train_dataset = IMDBDataset(imdb_dataset["train"], tokenizer, device)
test_dataset = IMDBDataset(imdb_dataset["test"], tokenizer, device)
train_loader = DataLoader(train_dataset, batch_size=batch_size, shuffle=True)
val_loader = DataLoader(test_dataset, batch_size=batch_size, shuffle=False)
layer = nn.Linear(250 * 250, num_neurons).to(device)
model = IMDBModel(layer)
optimal_bias = norm.ppf(1 / batch_size) * np.sqrt(250 * 250)
with torch.no_grad():
layer.weight.data.normal_()
layer.bias.data.fill_(optimal_bias)
PAIRS(
layer=model.fc1,
train_dataloader=train_loader,
batch_size=batch_size,
n_neurons=num_neurons,
)
return multi_evaluate(
model=model,
val_dataloader=val_loader,
batch_size=batch_size,
num_neurons=num_neurons,
eval_iters=10
)
def main():
file_name = "results_IMDB_PAIRS.csv"
torch.manual_seed(42)
runs_per_setting = 10
layer_sizes = [1000]
batch_sizes = [20, 50, 100, 200]
exp_aggregator(file_name, experiment, layer_sizes, batch_sizes, runs_per_setting)
if __name__ == "__main__":
main()