Skip to content

berdon/sif

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

16 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

sif

Node.JS tokenizer/lexer that works similar to flex-bison/lex-yacc.

Example

See ./examples for some basic examples using sif.

Here's a simple example simple parsing an XML file:

const grammar = new Grammar();
grammar
    .for("ROOT",         p => p.expect("ELE"))
    .for("ELE",          p => p.expect("ELE_TAG ELE_TAG_TAIL"))
    .for("ELE_TAG_TAIL", p => p.expect("GTHAN ELE[] END_START IDENTIFIER GTHAN"))
    .for("ELE_TAG_TAIL", p => p.expect("FSLASH GTHAN"))
    .for("ELE_TAG",      p => p.expect("LTHAN IDENTIFIER PROPERTY[]"))
    .for("PROPERTY",     p => p.expect("IDENTIFIER EQUAL VALUE"))
    .for("VALUE",        p => p.expect("STRING"))
    .for("VALUE",        p => p.expect("INTEGER"))
    .for("VALUE",        p => p.expect("FLOAT"))

const tokenizer = Tokenizer.fromJson(data);
const lexer = new Lexer('ROOT', grammar, tokenizer);
lexer.parse(process.argv[2]);

And here's a more involved example that converts an XML file into JSON:

import { Lexer, Tokenizer, Grammar } from "../../out";
import data from "./token.json";

let root: any = {};
const grammar = new Grammar();
grammar
    .for("ROOT", p => p.expect("ELE", ctx => {
        return root = ctx.value
    }))
    .for("ELE", p => p
        .expect("ELE_TAG", ctx => ctx.bag.node = ctx.value)
        .expect("ELE_TAG_TAIL").do(ctx => { ctx.bag.node.children = ctx.value.length > 0 ? ctx.value : undefined; return ctx.bag.node }))
    .for("ELE_TAG_TAIL", p => p
        .expect("GTHAN", "ELE[]", ctx => ctx.bag.children = ctx.value)
        .expect("END_START IDENTIFIER GTHAN", ctx => ctx.bag.children))
    .for("ELE_TAG_TAIL", p => p.expect("FSLASH GTHAN", ctx => []))
    .for("ELE_TAG", p => p
        .expect("LTHAN IDENTIFIER", ctx => ctx.bag.tag = ctx.value)
        .expect("PROPERTY[]", ctx => ({ tag: ctx.bag.tag, properties: ctx.value.length > 0 ? ctx.value : undefined, children: [] })))
    .for("PROPERTY", p => p
        .expect("IDENTIFIER", ctx => ctx.bag.name = ctx.value)
        .expect("EQUAL PROPERTY_VALUE", ctx => ({ [ctx.bag.name]: ctx.value })))
    .for("PROPERTY_VALUE", p => p.expect("STRING", ctx => ctx.value.substring(1, ctx.value.length - 1)))
    .for("PROPERTY_VALUE", p => p.expect("INTEGER", ctx => Number(ctx.value)))
    .for("PROPERTY_VALUE", p => p.expect("FLOAT", ctx => Number(ctx.value)))

const tokenizer = Tokenizer.fromJson(data);
const lexer = new Lexer('ROOT', grammar, tokenizer);
lexer.parse(process.argv[2]);

console.log(JSON.stringify(root, null, 4));

And parsing JSON:

const grammar = new Grammar();
grammar
    .for("ROOT",         p => p.expect("OBJECT"))
    .for("ROOT",         p => p.expect("ARRAY"))
    
    // Objects
    .for("OBJECT",       p => p.expect("LBRACE OBJECT_BODY"))
    .for("OBJECT_BODY",  p => p.expect("PROPERTY OBJECT_END"))
    .for("OBJECT_END",   p => p.expect("COMMA OBJECT_BODY"))
    .for("OBJECT_END",   p => p.expect("RBRACE"))
    .for("PROPERTY",     p => p.expect("STRING COLON VALUE"))
    
    // Arrays
    .for("ARRAY",        p => p.expect("LBRACKET ARRAY_BODY"))
    .for("ARRAY_BODY",   p => p.expect("VALUE ARRAY_END"))
    .for("ARRAY_END",    p => p.expect("COMMA ARRAY_BODY"))
    .for("ARRAY_END",    p => p.expect("RBRACKET"))

    // Values
    .for("VALUE",        p => p.expect("STRING"))
    .for("VALUE",        p => p.expect("INTEGER"))
    .for("VALUE",        p => p.expect("OBJECT"))
    .for("VALUE",        p => p.expect("ARRAY"))

const tokenizer = Tokenizer.fromJson(data);
const lexer = new Lexer('ROOT', grammar, tokenizer);
lexer.parse(process.argv[2]);

About

Node.JS tokenizer/lexer that works similar to flex-bison/lex-yacc.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published