-
Notifications
You must be signed in to change notification settings - Fork 29
intRVFL #98
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Merged
Merged
intRVFL #98
Changes from all commits
Commits
Show all changes
16 commits
Select commit
Hold shift + click to select a range
88db019
Draft first version of intRVFL classification
denkle 8eddbc9
[github-action] formatting fixes
github-actions[bot] 04e5b0c
Reworked code targeting modularity but still there is a number of iss…
denkle d3eee30
[github-action] formatting fixes
github-actions[bot] 89feff8
Revised code to improve modularity
denkle 765681c
[github-action] formatting fixes
github-actions[bot] 39bcc22
Continue improving benchmark. Refactored several components
denkle a8fc793
[github-action] formatting fixes
github-actions[bot] 98972bf
Further improve the code logic and allocation accross the library
denkle 44a3e94
Merge conflicts
denkle 6541c71
[github-action] formatting fixes
github-actions[bot] 0dcde3a
Allow more types of models in Density encoding
mikeheddes d3836d4
Refactor benchmark
mikeheddes f6a287f
Merge branch 'main' into intRVFL
mikeheddes 9e9d7ba
Refactor benchmark example
mikeheddes b75ac69
[github-action] formatting fixes
github-actions[bot] File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
There are no files selected for viewing
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -16,4 +16,5 @@ torchhd.embeddings | |
Thermometer | ||
Circular | ||
Projection | ||
Sinusoid | ||
Sinusoid | ||
Density |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,326 @@ | ||
import torch | ||
import torch.nn as nn | ||
import torch.utils.data as data | ||
from torch import Tensor | ||
from tqdm import tqdm | ||
|
||
# Note: this example requires the torchmetrics library: https://torchmetrics.readthedocs.io | ||
import torchmetrics | ||
import torchhd | ||
from torchhd.datasets import UCIClassificationBenchmark | ||
|
||
|
||
# Function for performing min-max normalization of the input data samples | ||
def create_min_max_normalize(min: Tensor, max: Tensor): | ||
def normalize(input: Tensor) -> Tensor: | ||
return torch.nan_to_num((input - min) / (max - min)) | ||
|
||
return normalize | ||
|
||
|
||
# Function that forms the classifier (readout matrix) with the ridge regression | ||
def classifier_ridge_regression( | ||
train_ld: data.DataLoader, | ||
dimensions: int, | ||
num_classes: int, | ||
lamb: float, | ||
encoding_function, | ||
data_type: torch.dtype, | ||
device: torch.device, | ||
): | ||
|
||
# Get number of training samples | ||
num_train = len(train_ld.dataset) | ||
# Collects high-dimensional represetations of data in the train data | ||
total_samples_hv = torch.zeros( | ||
num_train, | ||
dimensions, | ||
dtype=data_type, | ||
device=device, | ||
) | ||
# Collects one-hot encodings of class labels | ||
labels_one_hot = torch.zeros( | ||
num_train, | ||
num_classes, | ||
dtype=data_type, | ||
device=device, | ||
) | ||
|
||
with torch.no_grad(): | ||
count = 0 | ||
for samples, labels in tqdm(train_ld, desc="Training"): | ||
|
||
samples = samples.to(device) | ||
labels = labels.to(device) | ||
# Make one-hot encoding | ||
labels_one_hot[torch.arange(count, count + samples.size(0)), labels] = 1 | ||
|
||
# Make transformation into high-dimensional space | ||
samples_hv = encoding_function(samples) | ||
total_samples_hv[count : count + samples.size(0), :] = samples_hv | ||
|
||
count += samples.size(0) | ||
|
||
# Compute the readout matrix using the ridge regression | ||
Wout = ( | ||
torch.t(labels_one_hot) | ||
@ total_samples_hv | ||
@ torch.linalg.pinv( | ||
torch.t(total_samples_hv) @ total_samples_hv | ||
+ lamb * torch.diag(torch.var(total_samples_hv, 0)) | ||
) | ||
) | ||
|
||
return Wout | ||
|
||
|
||
# Specify a model to be evaluated | ||
class IntRVFLRidge(nn.Module): | ||
"""Class implementing integer random vector functional link network (intRVFL) model as described in `Density Encoding Enables Resource-Efficient Randomly Connected Neural Networks <https://doi.org/10.1109/TNNLS.2020.3015971>`_. | ||
|
||
Args: | ||
dataset (torchhd.datasets.CollectionDataset): Specifies a dataset to be evaluted by the model. | ||
num_feat (int): Number of features in the dataset. | ||
device (torch.device, optional): Specifies device to be used for Torch. | ||
""" | ||
|
||
# These values of hyperparameters were found via the grid search for intRVFL model as described in the article. | ||
INT_RVFL_HYPER = { | ||
"abalone": (1450, 32, 15), | ||
"acute-inflammation": (50, 0.0009765625, 1), | ||
"acute-nephritis": (50, 0.0009765625, 1), | ||
"adult": (1150, 0.0625, 3), | ||
"annealing": (1150, 0.015625, 7), | ||
"arrhythmia": (1400, 0.0009765625, 7), | ||
"audiology-std": (950, 16, 3), | ||
"balance-scale": (50, 32, 7), | ||
"balloons": (50, 0.0009765625, 1), | ||
"bank": (200, 0.001953125, 7), | ||
"blood": (50, 16, 7), | ||
"breast-cancer": (50, 32, 1), | ||
"breast-cancer-wisc": (650, 16, 3), | ||
"breast-cancer-wisc-diag": (1500, 2, 3), | ||
"breast-cancer-wisc-prog": (1450, 0.01562500, 3), | ||
"breast-tissue": (1300, 0.1250000, 1), | ||
"car": (250, 32, 3), | ||
"cardiotocography-10clases": (1350, 0.0009765625, 3), | ||
"cardiotocography-3clases": (900, 0.007812500, 15), | ||
"chess-krvk": (800, 4, 1), | ||
"chess-krvkp": (1350, 0.01562500, 3), | ||
"congressional-voting": (100, 32, 15), | ||
"conn-bench-sonar-mines-rocks": (1100, 0.01562500, 3), | ||
"conn-bench-vowel-deterding": (1350, 8, 3), | ||
"connect-4": (1100, 0.5, 3), | ||
"contrac": (50, 8, 7), | ||
"credit-approval": (200, 32, 7), | ||
"cylinder-bands": (1100, 0.0009765625, 7), | ||
"dermatology": (900, 8, 3), | ||
"echocardiogram": (250, 32, 15), | ||
"ecoli": (350, 32, 3), | ||
"energy-y1": (650, 0.1250000, 3), | ||
"energy-y2": (1000, 0.0625, 7), | ||
"fertility": (150, 32, 7), | ||
"flags": (900, 32, 15), | ||
"glass": (1400, 0.03125000, 3), | ||
"haberman-survival": (100, 32, 3), | ||
"hayes-roth": (50, 16, 1), | ||
"heart-cleveland": (50, 32, 15), | ||
"heart-hungarian": (50, 16, 15), | ||
"heart-switzerland": (50, 8, 15), | ||
"heart-va": (1350, 0.1250000, 15), | ||
"hepatitis": (1300, 0.03125000, 1), | ||
"hill-valley": (150, 0.01562500, 1), | ||
"horse-colic": (850, 32, 1), | ||
"ilpd-indian-liver": (1200, 0.25, 7), | ||
"image-segmentation": (650, 8, 1), | ||
"ionosphere": (1150, 0.001953125, 1), | ||
"iris": (50, 4, 3), | ||
"led-display": (50, 0.0009765625, 7), | ||
"lenses": (50, 0.03125000, 1), | ||
"letter": (1500, 32, 1), | ||
"libras": (1250, 0.1250000, 3), | ||
"low-res-spect": (1400, 8, 7), | ||
"lung-cancer": (450, 0.0009765625, 1), | ||
"lymphography": (1150, 32, 1), | ||
"magic": (800, 16, 3), | ||
"mammographic": (150, 16, 7), | ||
"miniboone": (650, 0.0625, 15), | ||
"molec-biol-promoter": (1250, 32, 1), | ||
"molec-biol-splice": (1000, 8, 15), | ||
"monks-1": (50, 4, 3), | ||
"monks-2": (400, 32, 1), | ||
"monks-3": (50, 4, 15), | ||
"mushroom": (150, 0.25, 3), | ||
"musk-1": (1300, 0.001953125, 7), | ||
"musk-2": (1150, 0.007812500, 7), | ||
"nursery": (1000, 32, 3), | ||
"oocytes_merluccius_nucleus_4d": (1500, 1, 7), | ||
"oocytes_merluccius_states_2f": (1500, 0.0625, 7), | ||
"oocytes_trisopterus_nucleus_2f": (1450, 0.003906250, 3), | ||
"oocytes_trisopterus_states_5b": (1450, 2, 7), | ||
"optical": (1100, 32, 7), | ||
"ozone": (50, 0.003906250, 1), | ||
"page-blocks": (800, 0.001953125, 1), | ||
"parkinsons": (1200, 0.5, 1), | ||
"pendigits": (1500, 0.1250000, 1), | ||
"pima": (50, 32, 1), | ||
"pittsburg-bridges-MATERIAL": (100, 8, 1), | ||
"pittsburg-bridges-REL-L": (1200, 0.5, 1), | ||
"pittsburg-bridges-SPAN": (450, 4, 7), | ||
"pittsburg-bridges-T-OR-D": (1000, 16, 1), | ||
"pittsburg-bridges-TYPE": (50, 32, 7), | ||
"planning": (50, 32, 1), | ||
"plant-margin": (1350, 2, 7), | ||
"plant-shape": (1450, 0.25, 3), | ||
"plant-texture": (1500, 4, 7), | ||
"post-operative": (50, 32, 15), | ||
"primary-tumor": (950, 32, 3), | ||
"ringnorm": (1500, 0.125, 3), | ||
"seeds": (550, 32, 1), | ||
"semeion": (1400, 32, 15), | ||
"soybean": (850, 1, 3), | ||
"spambase": (1350, 0.0078125, 15), | ||
"spect": (50, 32, 1), | ||
"spectf": (1100, 0.25, 15), | ||
"statlog-australian-credit": (200, 32, 15), | ||
"statlog-german-credit": (500, 32, 15), | ||
"statlog-heart": (50, 32, 7), | ||
"statlog-image": (950, 0.125, 1), | ||
"statlog-landsat": (1500, 16, 3), | ||
"statlog-shuttle": (100, 0.125, 7), | ||
"statlog-vehicle": (1450, 0.125, 7), | ||
"steel-plates": (1500, 0.0078125, 3), | ||
"synthetic-control": (1350, 16, 3), | ||
"teaching": (400, 32, 3), | ||
"thyroid": (300, 0.001953125, 7), | ||
"tic-tac-toe": (750, 8, 1), | ||
"titanic": (50, 0.0009765625, 1), | ||
"trains": (100, 16, 1), | ||
"twonorm": (1100, 0.0078125, 15), | ||
"vertebral-column-2clases": (250, 32, 3), | ||
"vertebral-column-3clases": (200, 32, 15), | ||
"wall-following": (1200, 0.00390625, 3), | ||
"waveform": (1400, 8, 7), | ||
"waveform-noise": (1300, 0.0009765625, 15), | ||
"wine": (850, 32, 1), | ||
"wine-quality-red": (1100, 32, 1), | ||
"wine-quality-white": (950, 8, 3), | ||
"yeast": (1350, 4, 1), | ||
"zoo": (400, 8, 7), | ||
} | ||
|
||
def __init__( | ||
self, | ||
dataset: torchhd.datasets.CollectionDataset, | ||
num_feat: int, | ||
device: torch.device = None, | ||
): | ||
super(IntRVFLRidge, self).__init__() | ||
self.device = device | ||
self.num_feat = num_feat | ||
# Fetch the hyperparameters for the corresponding dataset | ||
hyper_param = self.INT_RVFL_HYPER[dataset.name] | ||
# Dimensionality of vectors used when transforming input data | ||
self.dimensions = hyper_param[0] | ||
# Regularization parameter used for ridge regression classifier | ||
self.lamb = hyper_param[1] | ||
# Parameter of the clipping function used as the part of transforming input data | ||
self.kappa = hyper_param[2] | ||
# Number of classes in the dataset | ||
self.num_classes = len(dataset.classes) | ||
# Initialize the classifier | ||
self.classify = nn.Linear(self.dimensions, self.num_classes, bias=False) | ||
self.classify.weight.data.fill_(0.0) | ||
# Set up the encoding for the model as specified in "Density" | ||
self.hypervector_encoding = torchhd.embeddings.Density( | ||
self.num_feat, self.dimensions | ||
) | ||
|
||
# Specify encoding function for data samples | ||
def encode(self, x): | ||
return self.hypervector_encoding(x).clipping(self.kappa) | ||
|
||
# Specify how to make an inference step and issue a prediction | ||
def forward(self, x): | ||
# Make encodings for all data samples in the batch | ||
encodings = self.encode(x) | ||
# Get similarity values for each class assuming implicitly that there is only one prototype per class. This does not have to be the case in general. | ||
logit = self.classify(encodings) | ||
# Form predictions | ||
predictions = torch.argmax(logit, dim=-1) | ||
return predictions | ||
|
||
# Train the classfier | ||
def fit( | ||
self, | ||
train_ld: data.DataLoader, | ||
): | ||
# Gets classifier (readout matrix) via the ridge regression | ||
Wout = classifier_ridge_regression( | ||
train_ld, | ||
self.dimensions, | ||
self.num_classes, | ||
self.lamb, | ||
self.encode, | ||
self.hypervector_encoding.key.weight.dtype, | ||
self.device, | ||
) | ||
# Assign the obtained classifier to the output | ||
with torch.no_grad(): | ||
self.classify.weight.copy_(Wout) | ||
|
||
|
||
# Specify device to be used for Torch. | ||
device = torch.device("cuda" if torch.cuda.is_available() else "cpu") | ||
print("Using {} device".format(device)) | ||
# Specifies batch size to be used for the model. | ||
batch_size = 10 | ||
# Specifies how many random initializations of the model to evaluate for each dataset in the collection. | ||
repeats = 5 | ||
|
||
|
||
# Get an instance of the UCI benchmark | ||
benchmark = UCIClassificationBenchmark("../data", download=True) | ||
# Perform evaluation | ||
for dataset in benchmark.datasets(): | ||
print(dataset.name) | ||
|
||
# Number of features in the dataset. | ||
num_feat = dataset.train[0][0].size(-1) | ||
# Number of classes in the dataset. | ||
num_classes = len(dataset.train.classes) | ||
|
||
# Get values for min-max normalization and add the transformation | ||
min_val = torch.min(dataset.train.data, 0).values.to(device) | ||
max_val = torch.max(dataset.train.data, 0).values.to(device) | ||
transform = create_min_max_normalize(min_val, max_val) | ||
dataset.train.transform = transform | ||
dataset.test.transform = transform | ||
|
||
# Set up data loaders | ||
train_loader = data.DataLoader(dataset.train, batch_size=batch_size, shuffle=True) | ||
test_loader = data.DataLoader(dataset.test, batch_size=batch_size) | ||
|
||
# Run for the requested number of simulations | ||
for r in range(repeats): | ||
# Creates a model to be evaluated. The model should specify both transformation of input data as weel as the algortihm for forming the classifier. | ||
model = IntRVFLRidge( | ||
getattr(torchhd.datasets, dataset.name), num_feat, device | ||
).to(device) | ||
|
||
# Obtain the classifier for the model | ||
model.fit(train_loader) | ||
accuracy = torchmetrics.Accuracy("multiclass", num_classes=num_classes) | ||
|
||
with torch.no_grad(): | ||
for samples, targets in tqdm(test_loader, desc="Testing"): | ||
samples = samples.to(device) | ||
# Make prediction | ||
predictions = model(samples) | ||
accuracy.update(predictions.cpu(), targets) | ||
|
||
benchmark.report(dataset, accuracy.compute().item()) | ||
|
||
# Returns a dictionary with names of the datasets and their respective accuracy that is averaged over folds (if applicable) and repeats | ||
benchmark_accuracy = benchmark.score() | ||
print(benchmark_accuracy) |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Oops, something went wrong.
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Uh oh!
There was an error while loading. Please reload this page.