We had created a R notebook version of the first portion of movielens python notebook from the Fastai Deep Learning for Coders (Part 1) where high level fastai functions were used to build and fit the model. This notebook tries to create the R version of the second portion of the movielens python notebook where Jeremy creates the collaborative filtering model form scratch.
This content is covered in videos of lecture 5 and lecture 6. It will be helpful to listen to the lectures before going through this notebook since the concepts of the model and approach are discussed in the lecture and this notebook is just a replication attempt of the material from the course using R.
Initial Setup
# import R libraries
library(reticulate)
library(ggplot2)
library(dplyr)
Attaching package: ‘dplyr’
The following objects are masked from ‘package:stats’:
filter, lag
The following objects are masked from ‘package:base’:
intersect, setdiff, setequal, union
Notes about python setup and machine used are covered in this R notebook.
use_python("/home/paperspace/anaconda3/envs/fastai/bin/python", required = TRUE)
use_condaenv("fastai")
py_config()
python: /home/paperspace/anaconda3/envs/fastai/bin/python
libpython: /home/paperspace/anaconda3/envs/fastai/lib/libpython3.6m.so
pythonhome: /home/paperspace/anaconda3/envs/fastai:/home/paperspace/anaconda3/envs/fastai
version: 3.6.4 |Anaconda, Inc.| (default, Dec 21 2017, 21:42:08) [GCC 7.2.0]
numpy: /home/paperspace/anaconda3/envs/fastai/lib/python3.6/site-packages/numpy
numpy_version: 1.13.3
NOTE: Python version was forced by use_python function
main = import_main()
bi = import_builtins()
# get relevant python imports
fstai_learner = import_from_path("fastai.learner", "../../fastai")
fstai_coldata = import_from_path("fastai.column_data", "../../fastai")
py_run_string("
from fastai.learner import *
from fastai.column_data import *
")
Get Data
The ratings dataset has the ratings for different users and movies. The movies dataset has the movie title information.
datapath = "../../data/ml-latest-small/"
ratings = read.csv(paste0(datapath, "ratings.csv"), stringsAsFactors = FALSE)
head(ratings)
movies = read.csv(paste0(datapath, "movies.csv"), stringsAsFactors = FALSE)
head(movies)
Working with PyTorch Tensors
Amat = matrix(c(1.0, 2.0, 3.0, 4.0), byrow = TRUE, ncol = 2)
Bmat = matrix(c(2.0, 2.0, 10.0, 10.0), byrow = TRUE, ncol = 2)
a = py$T(Amat)
a
1 2
3 4
[torch.cuda.FloatTensor of size 2x2 (GPU 0)]
b = py$T(Bmat)
b
2 2
10 10
[torch.cuda.FloatTensor of size 2x2 (GPU 0)]
In python when we use a*b
, it works fine but won’t work when calling from R since R doesn’t know how to use *
operator with Torch tensors. So here I have used the pyTorch mul
function for multiplication
a$mul(b)
2 4
30 40
[torch.cuda.FloatTensor of size 2x2 (GPU 0)]
a$mul(b)$sum(1L) # note the use of 1L instead of 1 to ensure that we are passing integer
6
70
[torch.cuda.FloatTensor of size 2 (GPU 0)]
In PyTorch, we need to define a model as a Python class that inherits from nn.Module
. It has a specific method forward
which gives the recipe for computing the prediction given inputs. For example, if the prediction r
is a dot product of input vectors u
and m
, then it would be defined in the following class.
py_run_string("
class DotProduct(nn.Module):
def forward(self, u, m): return (u*m).sum(1)
")
A model can then be defined as the instance of the class and calling the forward method with inputs with give the prediction.
model = main$DotProduct()
model$forward(a, b)
6
70
[torch.cuda.FloatTensor of size 2 (GPU 0)]
Data Preparation for Use in PyTorch Model
The ratings dataset has userId and movieId fields. But for passing to pyTorch, we create a sequential index for both.
u_unique = unique(ratings$userId)
user2idx = as.integer(seq(0, length(u_unique) - 1))
names(user2idx) = u_unique
user2idx[1:10]
1 2 3 4 5 6 7 8 9 10
0 1 2 3 4 5 6 7 8 9
m_unique = unique(ratings$movieId)
movie2idx = as.integer(seq(0, length(m_unique) - 1))
names(movie2idx) = m_unique
movie2idx[1:10]
31 1029 1061 1129 1172 1263 1287 1293 1339 1343
0 1 2 3 4 5 6 7 8 9
ratings$userIdx = user2idx[as.character(ratings$userId)]
ratings$movieIdx = movie2idx[as.character(ratings$movieId)]
n_users = length(u_unique)
n_movies = length(m_unique)
n_users; n_movies
[1] 671
[1] 9066
Model 1
Each user is \(i\) represent by an embedding vector \(u_i\) consisting of n_factor
values. Similarly a movie \(j\) is represented by an embedding vector \(m_j\) consisting of n_factor
values. The model of rating \(r_{ij}\) given by user \(i\) to movie \(j\) is: \[ r_{ij} = u_i^Tv_j\] The class below defines the model. The constructor for class passes other inputs and initializes model parameters.
py_run_string("
class EmbeddingDot(nn.Module):
def __init__(self, n_users, n_movies, n_factors):
super().__init__()
self.u = nn.Embedding(n_users, n_factors)
self.m = nn.Embedding(n_movies, n_factors)
self.u.weight.data.uniform_(0,0.05)
self.m.weight.data.uniform_(0,0.05)
def forward(self, cats, conts):
users,movies = cats[:,0],cats[:,1]
u,m = self.u(users),self.m(movies)
return (u*m).sum(1)
")
The x dataframe just includes the sequential user id and movie id. y is a numpy array of ratings with type float32.
x = ratings[, c("userIdx", "movieIdx")]
y = np_array(ratings$rating)$astype(py$np$float32)
The list of validation data rows are selected and n_factors
is set to 50.
val_idxs = py$get_cv_idxs(nrow(ratings))
val_idxs = as.integer(val_idxs)
n_factors = 50L
The data loader object is created
data = py$ColumnarModelData$from_data_frame(datapath, val_idxs, r_to_py(x), y, c("userIdx", "movieIdx"), 64L)
Model is defined using the defined class EmbeddingDot
. Based on the model, the optimizer opt
is defined.
wd=1e-5
model = py$EmbeddingDot(n_users, n_movies, n_factors)$cuda()
opt = py$optim$SGD(model$parameters(), 1e-1, weight_decay=wd, momentum=0.9)
The model details are listed below
model
EmbeddingDot(
(u): Embedding(671, 50)
(m): Embedding(9066, 50)
)
Model is fit. In Jupyter notebook, a widget shows a nice output of progress. That output doesn’t render properly within RStudio and the html generated document has too much output. For now, I have turned off output and am just storing the final validation loss metric (MSE loss in this case)
mdlfit = py$fit(model, data, 3L, opt, py$F$mse_loss)
paste0("MSE of validation set = ", round(mdlfit[[1]], 3))
[1] "MSE of validation set = 1.219"
Model run for some more epochs
py$set_lrs(opt, 0.01)
mdlfit = py$fit(model, data, 3L, opt, py$F$mse_loss)
paste0("MSE of validation set = ", round(mdlfit[[1]], 3))
[1] "MSE of validation set = 1.131"
Alternate way to check validation MSE is to explicitly get predictions from model for validation data and compare to validation data
yval_preds = py$predict(model, data$val_dl)
yval=data$val_y[,1]
mse_loss = mean((yval_preds - yval)**2)
mse_loss
[1] 1.131446
Model 2
This is model 1 with added bias term for users and movies. Each user is \(i\) represent by an embedding vector \(u_i\) consisting of n_factor
values and a user bias value \(ub_i\). Similarly a movie \(j\) is represented by an embedding vector \(m_j\) consisting of n_factor
values and movie bias value \(mb_j\). In addtion, the output is contrained to be between minimum rating r_min
and maximum rating r_max
using a sigmoid function. The model of rating \(r_{ij}\) given by user \(i\) to movie \(j\) is: \[ r_{ij} = \frac{e^{u_i^Tv_j + b_i + m_j}}{e^{u_i^Tv_j + b_i + m_j} + 1}(r_{max} - r_{min}) + r_{min} \]
py_run_string("
def get_emb(ni,nf):
e = nn.Embedding(ni, nf)
e.weight.data.uniform_(-0.01,0.01)
return e
class EmbeddingDotBias(nn.Module):
def __init__(self, n_users, n_movies, n_factors, min_rating, max_rating):
super().__init__()
(self.u, self.m, self.ub, self.mb) = [get_emb(*o) for o in [
(n_users, n_factors), (n_movies, n_factors), (n_users,1), (n_movies,1)
]]
self.max_rating = max_rating
self.min_rating = min_rating
def forward(self, cats, conts):
users,movies = cats[:,0],cats[:,1]
um = (self.u(users)* self.m(movies)).sum(1)
res = um + self.ub(users).squeeze() + self.mb(movies).squeeze()
res = F.sigmoid(res) * (self.max_rating-self.min_rating) + self.min_rating
return res
")
The model object opt
is defined and fit
wd=2e-4
min_rating = min(ratings$rating)
max_rating = max(ratings$rating)
model = py$EmbeddingDotBias(n_users, n_movies, n_factors, min_rating, max_rating)$cuda()
opt = py$optim$SGD(model$parameters(), 1e-1, weight_decay=wd, momentum=0.9)
The model details are listed below
model
EmbeddingDotBias(
(u): Embedding(671, 50)
(m): Embedding(9066, 50)
(ub): Embedding(671, 1)
(mb): Embedding(9066, 1)
)
mdlfit = py$fit(model, data, 3L, opt, py$F$mse_loss)
paste0("MSE of validation set = ", round(mdlfit[[1]], 3))
[1] "MSE of validation set = 0.807"
Change the learning rate and refit
py$set_lrs(opt, 0.01)
mdlfit = py$fit(model, data, 3L, opt, py$F$mse_loss)
paste0("MSE of validation set = ", round(mdlfit[[1]], 3))
[1] "MSE of validation set = 0.801"
Model 3
This model is a deep learning model with the following layers:
- Input - Concatenation of user embedding vector and movie embedding vector (size =
2*n_factors
)
- Dropout with dropout rate
p1=0.05
- Linear fully connected layer with output size
nh = 10
- Relu
- Dropout with dropout rate
p2=0.5
- Linear fully connected layer with output size 1.
- Apply sigmoid function and scale to be between min and max rating
py_run_string("
class EmbeddingNet(nn.Module):
def __init__(self, n_users, n_movies, n_factors, min_rating, max_rating, nh=10, p1=0.05, p2=0.5):
super().__init__()
(self.u, self.m) = [get_emb(*o) for o in [
(n_users, n_factors), (n_movies, n_factors)]]
self.lin1 = nn.Linear(n_factors*2, nh)
self.lin2 = nn.Linear(nh, 1)
self.drop1 = nn.Dropout(p1)
self.drop2 = nn.Dropout(p2)
self.min_rating = min_rating
self.max_rating = max_rating
def forward(self, cats, conts):
users,movies = cats[:,0],cats[:,1]
x = self.drop1(torch.cat([self.u(users),self.m(movies)], dim=1))
x = self.drop2(F.relu(self.lin1(x)))
return F.sigmoid(self.lin2(x)) * (self.max_rating-self.min_rating+1) + self.min_rating-0.5
")
Model is defined and fit.
wd=1e-5
model = py$EmbeddingNet(n_users, n_movies, n_factors, min_rating, max_rating)$cuda()
opt = py$optim$Adam(model$parameters(), 1e-3, weight_decay=wd)
The model details are listed below
model
EmbeddingNet(
(u): Embedding(671, 50)
(m): Embedding(9066, 50)
(lin1): Linear(in_features=100, out_features=10)
(lin2): Linear(in_features=10, out_features=1)
(drop1): Dropout(p=0.05)
(drop2): Dropout(p=0.5)
)
mdlfit = py$fit(model, data, 3L, opt, py$F$mse_loss)
paste0("MSE of validation set = ", round(mdlfit[[1]], 3))
[1] "MSE of validation set = 0.784"
Learning rate is changed and model refit
py$set_lrs(opt, 0.01)
mdlfit = py$fit(model, data, 3L, opt, py$F$mse_loss)
paste0("MSE of validation set = ", round(mdlfit[[1]], 3))
[1] "MSE of validation set = 0.828"
Summary
This shows how a model can be developed from scratch. But the model would need to be defined as python class. I am guessing if somebody is developing a model from scratch, it might be better to do it just in python and create wrapper functions which can then by used by R.
sessionInfo()
R version 3.4.3 (2017-11-30)
Platform: x86_64-pc-linux-gnu (64-bit)
Running under: Ubuntu 16.04.3 LTS
Matrix products: default
BLAS: /usr/lib/libblas/libblas.so.3.6.0
LAPACK: /home/paperspace/anaconda3/envs/fastai/lib/libmkl_intel_lp64.so
locale:
[1] LC_CTYPE=en_US.UTF-8 LC_NUMERIC=C LC_TIME=en_US.UTF-8
[4] LC_COLLATE=en_US.UTF-8 LC_MONETARY=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8
[7] LC_PAPER=en_US.UTF-8 LC_NAME=C LC_ADDRESS=C
[10] LC_TELEPHONE=C LC_MEASUREMENT=en_US.UTF-8 LC_IDENTIFICATION=C
attached base packages:
[1] stats graphics grDevices utils datasets methods base
other attached packages:
[1] dplyr_0.7.4 ggplot2_2.2.1 reticulate_1.6
loaded via a namespace (and not attached):
[1] Rcpp_0.12.14 assertthat_0.2.0 R6_2.2.2 grid_3.4.3 plyr_1.8.4
[6] jsonlite_1.5 gtable_0.2.0 magrittr_1.5 scales_0.5.0 pillar_1.1.0
[11] rlang_0.1.6 lazyeval_0.2.1 bindrcpp_0.2 tools_3.4.3 glue_1.2.0
[16] munsell_0.4.3 yaml_2.1.18 compiler_3.4.3 pkgconfig_2.0.1 colorspace_1.3-2
[21] bindr_0.1.1 knitr_1.20 tibble_1.4.2
---
title: "Fastai Collaborative Filtering (from Scratch) with R and Reticulate"
output: html_notebook
editor_options: 
  chunk_output_type: inline
---

We had created a [R notebook](https://notesofdabbler.github.io/fastai_dl1_withR/movieLens.nb.html) version of the first portion of movielens [python notebook](https://github.com/fastai/fastai/blob/master/courses/dl1/lesson5-movielens.ipynb) from the [Fastai Deep Learning for Coders (Part 1)](http://course.fast.ai/) where high level fastai functions were used to build and fit the model. This notebook tries to create the R version of the second portion of the movielens python notebook where Jeremy creates the collaborative filtering model form scratch. 

This content is covered in videos of [lecture 5](http://course.fast.ai/lessons/lesson5.html) and [lecture 6](http://course.fast.ai/lessons/lesson6.html). It will be helpful to listen to the lectures before going through this notebook since the concepts of the model and approach are discussed in the lecture and this notebook is just a replication attempt of the material from the course using R.

## Initial Setup

```{r}
# import R libraries
library(reticulate)
library(ggplot2)
library(dplyr)

```

Notes about python setup and machine used are covered in this [R notebook](https://notesofdabbler.github.io/fastai_dl1_withR/movieLens.nb.html).

```{r}

use_python("/home/paperspace/anaconda3/envs/fastai/bin/python", required = TRUE)
use_condaenv("fastai")
py_config()

main = import_main()
bi = import_builtins()
```

```{r}
# get relevant python imports
fstai_learner = import_from_path("fastai.learner", "../../fastai")
fstai_coldata = import_from_path("fastai.column_data", "../../fastai")
```

```{r}
py_run_string("
from fastai.learner import *
from fastai.column_data import *
              ")
```

## Get Data

The ratings dataset has the ratings for different users and movies. The movies dataset has the movie title information.

```{r}
datapath = "../../data/ml-latest-small/"

ratings = read.csv(paste0(datapath, "ratings.csv"), stringsAsFactors = FALSE)
head(ratings)

movies = read.csv(paste0(datapath, "movies.csv"), stringsAsFactors = FALSE)
head(movies)
```

## Working with PyTorch Tensors
```{r}
Amat = matrix(c(1.0, 2.0, 3.0, 4.0), byrow = TRUE, ncol = 2)
Bmat = matrix(c(2.0, 2.0, 10.0, 10.0), byrow = TRUE, ncol = 2)

a = py$T(Amat)
a

b = py$T(Bmat)
b
```
In python when we use `a*b`, it works fine but won't work when calling from R since R doesn't know how to use `*` operator with Torch tensors. So here I have used the pyTorch `mul` function for multiplication
```{r}
a$mul(b)
a$mul(b)$sum(1L) # note the use of 1L instead of 1 to ensure that we are passing integer
```

In PyTorch, we need to define a model as a Python class that inherits from `nn.Module`. It has a specific method `forward` which gives the recipe for computing the prediction given inputs. For example, if the prediction `r` is a dot product of input vectors `u` and `m`, then it would be defined in the following class.
```{r}
py_run_string("
class DotProduct(nn.Module):
    def forward(self, u, m): return (u*m).sum(1)              
              ")
```
A model can then be defined as the instance of the class and calling the forward method with inputs with give the prediction.
```{r}
model = main$DotProduct()
model$forward(a, b)
```

## Data Preparation for Use in PyTorch Model

The ratings dataset has userId and movieId fields. But for passing to pyTorch, we create a sequential index for both.

```{r}
u_unique = unique(ratings$userId)
user2idx = as.integer(seq(0, length(u_unique) - 1))
names(user2idx) = u_unique
user2idx[1:10]

m_unique = unique(ratings$movieId)
movie2idx = as.integer(seq(0, length(m_unique) - 1))
names(movie2idx) = m_unique
movie2idx[1:10]
```

```{r}
ratings$userIdx = user2idx[as.character(ratings$userId)]
ratings$movieIdx = movie2idx[as.character(ratings$movieId)]

n_users = length(u_unique)
n_movies = length(m_unique)

n_users; n_movies
```

## Model 1

Each user is $i$ represent by an embedding vector $u_i$ consisting of `n_factor` values. Similarly a movie $j$ is represented by an embedding vector $m_j$ consisting of `n_factor` values. The model of rating $r_{ij}$ given by user $i$ to movie $j$ is:
$$ r_{ij} = u_i^Tv_j$$
The class below defines the model. The constructor for class passes other inputs and initializes model parameters. 
```{r}
py_run_string("
class EmbeddingDot(nn.Module):
    def __init__(self, n_users, n_movies, n_factors):
        super().__init__()
        self.u = nn.Embedding(n_users, n_factors)
        self.m = nn.Embedding(n_movies, n_factors)
        self.u.weight.data.uniform_(0,0.05)
        self.m.weight.data.uniform_(0,0.05)
        
    def forward(self, cats, conts):
        users,movies = cats[:,0],cats[:,1]
        u,m = self.u(users),self.m(movies)
        return (u*m).sum(1)              
              ")
```

The x dataframe just includes the sequential user id and movie id. y is a numpy array of ratings with type float32.
```{r}
x = ratings[, c("userIdx", "movieIdx")]
y = np_array(ratings$rating)$astype(py$np$float32)
```

The list of validation data rows are selected and `n_factors` is set to 50.
```{r}
val_idxs = py$get_cv_idxs(nrow(ratings))
val_idxs = as.integer(val_idxs)
n_factors = 50L
```

The data loader object is created
```{r}
data = py$ColumnarModelData$from_data_frame(datapath, val_idxs, r_to_py(x), y, c("userIdx", "movieIdx"), 64L)
```
Model is defined using the defined class `EmbeddingDot`. Based on the model, the optimizer `opt` is defined.
```{r}
wd=1e-5
model = py$EmbeddingDot(n_users, n_movies, n_factors)$cuda()
opt = py$optim$SGD(model$parameters(), 1e-1, weight_decay=wd, momentum=0.9)
```

The model details are listed below
```{r}
model
```
Model is fit. In Jupyter notebook, a widget shows a nice output of progress. That output doesn't render properly within RStudio and the html generated document has too much output. For now, I have turned off output and am just storing the final validation loss metric (MSE loss in this case)
```{r results="hide"}
mdlfit = py$fit(model, data, 3L, opt, py$F$mse_loss)
```
```{r}
paste0("MSE of validation set = ", round(mdlfit[[1]], 3))
```

Model run for some more epochs
```{r results="hide"}
py$set_lrs(opt, 0.01)
mdlfit = py$fit(model, data, 3L, opt, py$F$mse_loss)
```
```{r}
paste0("MSE of validation set = ", round(mdlfit[[1]], 3))
```

Alternate way to check validation MSE is to explicitly get predictions from model for validation data and compare to validation data
```{r}
yval_preds = py$predict(model, data$val_dl)
yval=data$val_y[,1]
mse_loss = mean((yval_preds - yval)**2)
mse_loss
```

## Model 2
This is model 1 with added bias term for users and movies. Each user is $i$ represent by an embedding vector $u_i$ consisting of `n_factor` values and a user bias value $ub_i$. Similarly a movie $j$ is represented by an embedding vector $m_j$ consisting of `n_factor` values and movie bias value $mb_j$. In addtion, the output is contrained to be between minimum rating `r_min` and maximum rating `r_max` using a sigmoid function. The model of rating $r_{ij}$ given by user $i$ to movie $j$ is:
$$ r_{ij} = \frac{e^{u_i^Tv_j + b_i + m_j}}{e^{u_i^Tv_j + b_i + m_j} + 1}(r_{max} - r_{min}) + r_{min} $$
```{r}
py_run_string("
def get_emb(ni,nf):
    e = nn.Embedding(ni, nf)
    e.weight.data.uniform_(-0.01,0.01)
    return e
              
class EmbeddingDotBias(nn.Module):
    def __init__(self, n_users, n_movies, n_factors, min_rating, max_rating):
        super().__init__()
        (self.u, self.m, self.ub, self.mb) = [get_emb(*o) for o in [
              (n_users, n_factors), (n_movies, n_factors), (n_users,1), (n_movies,1)
              ]]
        self.max_rating = max_rating
        self.min_rating = min_rating
              
    def forward(self, cats, conts):
        users,movies = cats[:,0],cats[:,1]
        um = (self.u(users)* self.m(movies)).sum(1)
        res = um + self.ub(users).squeeze() + self.mb(movies).squeeze()
        res = F.sigmoid(res) * (self.max_rating-self.min_rating) + self.min_rating
        return res              
              ")
```

The model object `opt` is defined and fit
```{r}
wd=2e-4
min_rating = min(ratings$rating)
max_rating = max(ratings$rating)
model = py$EmbeddingDotBias(n_users, n_movies, n_factors, min_rating, max_rating)$cuda()
opt = py$optim$SGD(model$parameters(), 1e-1, weight_decay=wd, momentum=0.9)
```

The model details are listed below
```{r}
model
```

```{r results = "hide"}
mdlfit = py$fit(model, data, 3L, opt, py$F$mse_loss)
```
```{r}
paste0("MSE of validation set = ", round(mdlfit[[1]], 3))
```
Change the learning rate and refit
```{r results = "hide"}
py$set_lrs(opt, 0.01)
mdlfit = py$fit(model, data, 3L, opt, py$F$mse_loss)
```
```{r}
paste0("MSE of validation set = ", round(mdlfit[[1]], 3))
```

## Model 3

This model is a deep learning model with the following layers:

1. Input - Concatenation of user embedding vector and movie embedding vector (size = `2*n_factors`)
2. Dropout with dropout rate `p1=0.05`
3. Linear fully connected layer with output size `nh = 10`
4. Relu
5. Dropout with dropout rate `p2=0.5`
6. Linear fully connected layer with output size 1.
7. Apply sigmoid function and scale to be between min and max rating

```{r}
py_run_string("
class EmbeddingNet(nn.Module):
    def __init__(self, n_users, n_movies, n_factors, min_rating, max_rating, nh=10, p1=0.05, p2=0.5):
        super().__init__()
        (self.u, self.m) = [get_emb(*o) for o in [
            (n_users, n_factors), (n_movies, n_factors)]]
        self.lin1 = nn.Linear(n_factors*2, nh)
        self.lin2 = nn.Linear(nh, 1)
        self.drop1 = nn.Dropout(p1)
        self.drop2 = nn.Dropout(p2)
        self.min_rating = min_rating
        self.max_rating = max_rating
        
    def forward(self, cats, conts):
        users,movies = cats[:,0],cats[:,1]
        x = self.drop1(torch.cat([self.u(users),self.m(movies)], dim=1))
        x = self.drop2(F.relu(self.lin1(x)))
        return F.sigmoid(self.lin2(x)) * (self.max_rating-self.min_rating+1) + self.min_rating-0.5
           ")
```

Model is defined and fit.
```{r}
wd=1e-5
model = py$EmbeddingNet(n_users, n_movies, n_factors, min_rating, max_rating)$cuda()
opt = py$optim$Adam(model$parameters(), 1e-3, weight_decay=wd)
```

The model details are listed below
```{r}
model
```

```{r results = "hide"}
mdlfit = py$fit(model, data, 3L, opt, py$F$mse_loss)
```
```{r}
paste0("MSE of validation set = ", round(mdlfit[[1]], 3))
```
Learning rate is changed and model refit
```{r results="hide"}
py$set_lrs(opt, 0.01)
mdlfit = py$fit(model, data, 3L, opt, py$F$mse_loss)
```
```{r}
paste0("MSE of validation set = ", round(mdlfit[[1]], 3))
```

## Summary

This shows how a model can be developed from scratch. But the model would need to be defined as python class. I am guessing if somebody is developing a model from scratch, it might be better to do it just in python and create wrapper functions which can then by used by R. 

```{r}
sessionInfo()
```