Register
Login
Resources
Docs Blog Datasets Glossary Case Studies Tutorials & Webinars
Product
Data Engine LLMs Platform Enterprise
Pricing Explore
Connect to our Discord channel
Integration:  git github
d4e927bd6a
evaluating with beir
4 months ago
892dcc235a
extracting readmes with elixir
4 months ago
93530dbac0
parsing Python AST in Rust
9 months ago
59ab7dbb3c
graph similarity model for label which also is a graph
1 year ago
892dcc235a
extracting readmes with elixir
4 months ago
37820bef8a
code2doc - return filenames, BEIR notebook
1 week ago
afd04ee803
postprocessing in generation pipeline
5 months ago
4a6ee424ab
README generation report
5 days ago
org
175df3ab85
code2documentation - generating READMEs from code
2 weeks ago
aa42f940fb
Python repository imports
3 years ago
3 years ago
0dcfcee20a
generating READMEs with code2doc - pipeline
1 week ago
a5b52709f0
project cleanup, working poetry
5 months ago
a5b52709f0
project cleanup, working poetry
5 months ago
175df3ab85
code2documentation - generating READMEs from code
2 weeks ago
7ca0a14e31
README, ploomber config refactor
2 years ago
4f85d29009
training with readmes
1 year ago
dc442194e5
data download from GCP
4 years ago
892dcc235a
extracting readmes with elixir
4 months ago
daa81ccce2
refactoring preprocessing, tests
4 years ago
50a4b5910e
guild changes
4 years ago
0dcfcee20a
generating READMEs with code2doc - pipeline
1 week ago
6bdd909614
moved some functions from notebooks to packagages
4 years ago
0dcfcee20a
generating READMEs with code2doc - pipeline
1 week ago
a5b52709f0
project cleanup, working poetry
5 months ago
77414a025b
pipeline for fetching import data
2 years ago
8f01f1bfc2
nbdev setup
3 years ago
Storage Buckets

README.md

You have to be logged in to leave a comment. Sign In

github-search

Repository for "Searching Github Python repositories with machine learning" masters thesis

Jakub Bartczuk

Running this project

Prerequisites

Preprocessing steps were tested on a machine with 64GB RAM.

For training Graph Neural Networks CUDA GPU is required.

General remarks

The project uses nbdev to create Python files from Jupyter notebook. To "make" project run

nbdev_build_lib; pip install -e .

in the root directory.

We use ploomber for managing training and data preprocessing.

For example to create csv files with extracted READMEs run

ploomber build --partial make_readmes --skip-upstream --force

Relevant definitions can be found in pipeline.yaml and env.yaml

TODO Downloading data

TODO Model checkpoints

Training models

Ploomber step:

run_gnn_experiment

TODO Using models

Tip!

Press p or to see the previous file or, n or to see the next file

About

No description

Collaborators 1

Comments

Loading...