Skip to content

imdeepmind/CodeBot

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

37 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

CodeBot

An LSTM based Deep Learning model that generates random C/C++ code.

Note: Due to limited computational power, I have trained the model only on limited data

Table of contents:

Introduction

Computer codes are complicated, hard to learn for Humans. Here in this project, I attempted to make a Deep Learning model that learns to code in C/C++ and generates small snippets of random code.

Dataset

For training such a language model, I need a huge amount of codes. So after some research, I decided to use Google's Chromium project. Chromium is an open-sourced web browser and contains a huge amount of code that is mostly written in C/C++.

To get the data, I cloned the repo and find a list of all C/C++ codes.

Model

Here I have used a simple LSTM based model. The first layer in the model is a Embedding layer.

The second layer is the LSTM cell. LSTM or Long Short Term Memory Recurrent Neural Network is a type of Neural Network, that is used for sequenced data.

Finally a Dense layer with Softmax activation function for the output layer.

Sample Results

Here are some of the text that the model generated after training.

Generated text with temperature 0.2:

#define UP(GPU_EXT_FALSERT_SH_CONER_CONER_H_
#define in the context the static context the context the context the context context in the context reserver the context reserver base the context the context the context the context the source context reserved reserved reserved request the context the std::web::Context the source context reserved by a BSD-style license the process the context in the context the context the context the context reserved.
 // Use a devel server is source context a not the conte

Generated text with temperature 0.5:

#define UProwiteStatus reserver.h"
#include "fuchsia///wrbidages_stroute.h"

#include "base/file.h"
#include "base/fuchsia::web::Values::Oner context_str_context_prowerned.h"
#include "base/ins/crommand_binding_data);
      GpuContext kManding::Strings::Context::Attrib3f chrolor);
}

void GetPathApligator() {
    const :base::FileResure::Creater::Commandler::Commanding();
  }

  // Context bever the LICENSE file.

#include "base/fuchsia/versione/gtest/cline/service/chandle_thes[]lore.h"
#include "base/loo

Generated text with temperature 1.0:

#define UPURS_, "tessatil.h"
#include "gpu/verve/gpund.h"
#inclling(EntElpoNiteNapeInt3> nulling_achment_requos)));
  if (drawullong_tebaivabjeasyis* bat));
      EXPECT_EQ(base::TestHa_context(ortes.set(";
  fucts.h>

#inclindeb mole Version_state const Requert wensur target) {
    cast::Bind(lierabler_.seturf Crequener::FramebyttenTeakRewroder::webIn inglefer_fuchSwable(conplage_contexts.set.contextUNechRewerters ::nerlia:::Valse.UinsFalsPull params.

// If ther a leder -yy's BONE D;
      cideouen enar

Generated text with temperature 1.2:

#define UPablig(unpus-PoinPoindBingon,
                Componationgs::IfNEXL_X*(11,
      wirlo* case_phate_lase_uctor);
  ZHCERGOR())).DisCompoReqerOfErner[pRativeValGet::BogllilWRewTest-mover1TextoryBU Ray pbeck.
// O thher hend lowsss to fuxnmemmate
// fuc fuchsina he
// erraogs |&mptors5 biquilterter.h"
#include_ubs(        maip_type)
     "ructure_swout|h() proller gl::maratern thilpe) {
  cmds_: TRGPURARSE_NA_YSION))
  stchuich("5_FASI, kGraMemontedProatelGpUnifor_muse.esing(tpath));

  fuchsioler_s

Dependencies

The project is developed using Python 3.6. Along with that, I have used Tensorflow 2.0, NLTK, SQLite, etc.

For details information about the libraries that I have used, please check the requirements.txt file.

File Structure

The following are the files in the project.

  • callback.py - This class is used for generating some samples results after each epoch
  • model.py - This file contains the class Model that performs everything
  • main.py - In this file, I'm calling all the methods in the class

Running Locally

To run the project in your local machine, first, install the libraries mentioned in the requirements.txt file, then run the main.py file.

The project expects a data folder in the root directory on the project, where you need to put the entire Chromium codebase. The path should be something like this: ./data/chromium-master/.

Future Improvements

  • The model is trained only on a fraction of the available data, so training the model on entire data can improve the accuracy of the model.
  • LSTM cells are simple and naive. Some of the latest models, like GPT and/or Transformer models can do a better job.

Acknowledgments