Skip to content

Gyu-Seok0/Upitt_RS_Project

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

52 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Upitt RS Project

GOAL

How to recommend contents to students given context

Data

Contents : Wikipedia or Youtube
Context : Textbook pdf

Approach

(1) Create a graph
(2) Get a embedding for each node
(3) Link prediction

Methods

(1) Create a graph: WikiAPI + Keyword Network + Sentence Simiarity Network
(2) Get Embedding: MetaPath2Vec
(3) Link Predcition: Heterogenous GraphSage

Codes

[Python]

Before you proceed it, you should set up two conda environments by using environment_make_graph.yml and environment_hinsage.yml.

1. create a graph 
Wikipedia.py (environment_make_graph.yml must be set up)

2. get a embedding for each node 
make_embedding.py (environment_make_graph.yml must be set up)

3. link prediction
Hinsage.py (environment_hinsage.yml must be set up)

[Jupyter Notebook]

1. create a graph
Preprocess.ipynb (make a csv file to get all of information, csv_keywords_df.csv)
Wikipedia.ipynb (make four csv file to link csv and wiki, (csv_dict.pickle, wiki_dict.pickle, csv2wiki.pickle, wiki2csv.pickle))

2. get a embedding for each node 
MetaPath2Vec.ipynb (make a graph and get embedding of each node, (Embedding, csv_wiki_graph))

3. link prediction
weighted_link_prediction.ipynb

[Note]

  • We save all files as pickle
  • We use libraries and frameworks like pytorch, tensorflow, dgl, stellargraph and so on.
  • Keyword Network and Sentence Simiarity Network are based on BERT, which are provided from HuggingFace.
  • Useful Document: Link Predcition by using Heterogenous GraphSage

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published