Session-based Recommendation with Graph Neural Networks
04 Sep 2022
TL;DR, it doesn’t really work well in real world.
사실 gated GNN도 별게 없는게, 기존의 gru를 graph를 통해 propagate하고 accumluate한다 라고 이해하면 됩니다. Session-based GNN은 gated gnn에다가 global preference를 이해하기 위해 attention pooling layer를 추가했다 로 요약(?)해볼 수 있습니다.
SRGNN의 코드중 이해하기 어려웠던 것 그래프를 만드는 부분이었습니다. sequence가 [12,3,1,3,12]라고 있다고 해봅시다. 그러면 각 element에 대해 index로 맵핑을 시켜줍니다. (12는 1, 3은 2, 1은 3) 그러면, 그래프는 1-2-3-2-1 로 표현될 수 있고, in/out node에 대해 따로 따로 matrix를 만들어주고 합쳐주면 됩니다.
그래프를 만드는 것에서 오버헤드가 생겨서, 학습전에 sparse matrix 형태로 변환해주어 저장했더니 out of memory 문제가 발생하지 않았습니다.
그리고 이전에 인턴으로 일하면서, DQN에서 user의 historical한 state를 modeling하기 위한 user encoder로 SR-GNN을 사용해봤습니다. 그냥 GRU를 encoder로 사용할때 보다 성능 향상이 거의 없었습니다. 그 이유에 대해 다음과 같은 두가지 가설을 세워보았습니다.
-
paper에서 실험한 데이터셋은 id-based dataset이라서 각 item을 더 explicit하게 학습할 수 있다. 하지만 실제 서비스에서 사용한 feature들은 categorical한 feature 및 vgg를 통과한 continuous한 feature 등으로 이루어져 있었다.
-
실제 적용하려는 서비스에서는 repetitive item이 아니라 하나의 시퀀스에는 unique한 아이템만 있는 서비스 였습니다. (하나의 시퀀스에서 한 아이템은 한번만 나옵니다.)
이러한 이유로 RNN계열의 GRU가 더 잘 작동한것 같습니다.
추가로, official code에서 다음과 같이 잘못된 코드가 있었습니다.
hy = newgate + inputgate * (hidden - newgate)
이게 아니라 아래코드와 같이 작성해야 맞습니다.
hy = hidden + inputgate * (newgate - hidden)
해당 repo에서 pr을 닫아서? 수정할 순 없었지만, 위와 같은 코드에 대해 issue에서 다른사람들도 지적하고 있는것을 볼수 있습니다.
Prior Knowledge
Before explaining SRGNN, I will skim through Session, Session based Recsys, GNN(from Scarselli et al), Gated GNN.
Session
(cite) A session is defined as a series of related browser requests that come from the same client during a certain time period. Session tracking ties together a series of browser requests—think of these requests as pages—that may have some meaning as a whole, such as a shopping cart application.
Session based Recsys
(cite) Different from other Recsys such as content-based RSs and collaborative filtering-based RSs which usually model long-term yet static user preferences, SBRSs aim to capture short-term but dynamic user preferences to provide more timely and accurate recommendations sensitive to the evolution of their session contexts.
Graph Neural Networks(from Scarselli et al)
(cite lecture) From This lecture, basically GNN has two key model (1) Propagation model (2) Output model (possibly outdated but this is from author’s presentation).
-
Propagation model
Node representation for node v at propagation step $t:h_v^{(t)}$
\[h_v^{t}=\sum_{v'\in IN(v)}f(h_{v'}^{(t-1)},l_{(v',v)})+\sum_{v'\in OUT(v)}f(h_{v'}^{(t-1)},l_{(v,v')})\]This equation is for propagation and summation of node representations. Propagate representations along edges, allow multiple edge types and propagation on both directions. And When They finished propagation, then they summation of all node representation. ($f$: transformation function which transform nodes by their edge type and directions.)
\[Example: f(h_{v'}^{(t-1)},l_{(v',v)})=A^{(l_{(v,v')})}h_{v'}^{(t-1)}+b^{(l_{(v,v')})}\] -
Output model For each node v, compute an output based on final node representation. $g$ can be a neural net.
When we consider GNN as RNN (propagate each time step), Back propagation through time is expensive. GNN restrict the propagation model so that the propagation function is a contraction map which have a unique fixed point. (I don’t any of Banach fixed-point theorem(contraction mapping theorem) but anyway It restrict propagate.) And then Run the propagation until convergence.
Gated Graph Neural Networks
This main key-point (same cite from GNN) (cite lecture)
- Unroll recurrence for a fixed number of steps and just use back propagation through time with modern optimization methods.
- Also changed the propagation model a bit to use gating mechanisms like in LSTMs and GRUs.
From This architecture We can get these Benefits.
- No restriction on the propagation model, does not need to be a contraction map. (Author said that It can have more capacity and Power to solve more complicated model)
- Initialization matters now so problem specific information can be fed in as the input.
- Learning to compute representations within a fixed budget. When we train or test model, we just propagate fixed number of steps, so computation cost decrease.
- Gating makes the propagation model better.
This is Initialization Parts.
- Problem specific node annotations in $h_v^{(0)}$.
-
Q. Example reachability problem: can we go from A to B?
It is easy to learn a propagation model that copies and adds the first bit to a node’s neighbor.
It is easy to learn an output model that outputs ‘yes’ if it seems the [red, green] pattern, otherwise no.
- In practice, we pad node annotations with extra 0’s to add capacity h. \(h_v^{(0)}=[l_v^T,0^T]^T\) This means problem specific node annotations.
This is Propagation Model part for Gated GNN.
GNN propagation model with gating and other minor differences. You can thinks as Feed-Forward Network which is special connection structure. This is sort of Sparse structure and each connection share parameters dependent on edge’s type and direction.
\[a^{t}= \left[ \begin{matrix} a^{(OUT)} \\ a^{(IN)} \\ \end{matrix} \right]\]$a$ : concatenate all node representation in to big vector for matrix operation.
A stands for transformation Matrix. In matrix, each block can share parameter dependent on graph structure.
Let’s talk about equation of Gated GNN.
\[a^{(t)} = Ah^{(t-1)}+b\] \[h_v^{(t)} = tanh(Wa_v^{(t)})\]It’s look’s like vanilla RNN. And In this equation above, we add equation about gate.
\[a^{(t)} = Ah^{(t-1)}+b\] \[Reset \space gate: r_v^{t}=\sigma(W^ra_v^{(t)}+U^rh_v^{(t-1)})\] \[Update \space gate: z_v^{t}=\sigma(W^za_v^{(t)}+U^zh_v^{(t-1)})\] \[h_v^{(t)} = tanh(Wa_v^{(t)}+ U(r_v^t\odot h_v^{(t-1)}))\] \[h_v^{(t)} = (1-z_v^t)\odot h_v^{(t-1)}+z_v^t\odot h_v^{(t)}\]This is Output Models for Gated GNN.
- Per node output same as in GNNs.
- Node selection output $o_v=g(h_v^{(T)},l_v)$ computes scores for each node, then take softmax over all nodes to select one.
-
Graph level output Graph representation vector
\[h_g=\sum_{v\in g}\sigma(i(h_v^{(T)},l_v))\cdot h_v^{(T)}\]This vector can be used to do graph level classification, regression etc.
There is one more architecture which called Graph Graph Sequence Neural Networks
Prior Works before SRGNN
- Session-based recommendations with recurrent neural networks(Hidasi et al. 2016a) : proposes RNN approach at first.
- Improved recurrent neural networks for session-based recommendations(Tan, Xu, and Liu 2016) : Imporoved first one by data augmentation and considering temporal shift
- Neural attentive session-based recommendation,NARM(Li et al. 2017a) : design a global and local RNN recommendation to capture users’s sequential behavior and main purposes simultaneously.
- Stamp: Short-term attention/memory priority model for session-based recommendation(Liu et al. 2018) : Similar to NARM, STAMP also captured user’s general interests and current interests by employing simple MLP networks and an attentive net.
Prior Works Limitation
- Firstly, without adequate user behavior in one session, these methods have difficulty in estimating user representations. Sessions are mostly anonymous and numerous, and user behavior implicated in session clicks is often limited. It is thus difficult to accurately estimate the representation of each user from each session.
- Secondly, previous work reveals that patterns of item transitions are important and can be used as a local factor (Li et al. 2017a; Liu et al. 2018) in session-based recommendation, but these methods always model single way transitions between consecutive items and neglect the transitions among the contexts, i.e. other items in the session.
How Session-based Recommendation with Graph Neural Networks, SR-GNN solved these problem?
- We model separated session sequences into graph- structured data and use graph neural networks to capture complex item transitions.
- To generate session-based recommendations, we do not rely on user representations, but use the session embedding, which can be obtained merely based on latent vectors of items involved in each single session.
- Extensive experiments conducted on real-world datasets show that SR-GNN evidently outperforms the state-of-art methods.
How does SRGNN works?
- (input) At first, all session sequences are modeled as directed session graphs, where each session sequence can be treated as a subgraph.
- Then, each session graph is proceeded successively and the latent vectors for all nodes involved in each graph can be obtained through gated graph neural networks.
- After that, we represent each session as a composition of the global preference and the current interest of the user in that session, where these global and local session embedding vectors are both composed by the latent vectors of nodes.
- (output) Finally, for each session, we predict the probability of each item to be the next click.