tokenizer works

This commit is contained in:
jmosrael@gmail.com
2024-05-17 19:59:17 -07:00
parent 3094ae3a08
commit 401330d842
2 changed files with 55 additions and 85 deletions

View File

@@ -1,35 +1,31 @@
import Tokenizr from 'tokenizr'
import ASTY from 'asty'
import { Op } from 'sequelize'
export const lexr = new Tokenizr()
export class OrderByParser {
lexer = new Tokenizr()
lexr.rule(/[AND|OR]/, (ctx, m) => {
ctx.accept('conjunction', m[0])
})
constructor() {
this.lexer.rule(/,/, (ctx, m) => {
ctx.accept('spacer')
})
export const orderByLexr = new Tokenizr()
this.lexer.rule(/ASC|DESC/, (ctx, m) => {
ctx.accept('direction', m[0])
})
orderByLexr.rule(/,/, (ctx, m) => {
ctx.accept('spacer')
})
this.lexer.rule(/[a-zA-Z]+/, (ctx, m) => {
ctx.accept('column', m[0])
})
orderByLexr.rule(/ASC|DESC/, (ctx, m) => {
ctx.accept('direction', m[0])
})
this.lexer.rule(/\s/, (ctx, m) => {
ctx.ignore()
})
}
orderByLexr.rule(/[a-zA-Z]+/, (ctx, m) => {
ctx.accept('column', m[0])
})
orderByLexr.rule(/\s/, (ctx, m) => {
ctx.ignore()
})
export function parseOrderByString(orderBy: string) {
parse(orderBy: string) {
const output = []
let holding = []
orderByLexr
this.lexer
.input(orderBy)
.tokens()
.forEach((token) => {
@@ -48,6 +44,7 @@ export function parseOrderByString(orderBy: string) {
output.push(holding)
}
return output
}
}
const openGroupRegex = /\(/
@@ -61,55 +58,47 @@ const opperatorMap = {
'!=': Op.ne,
'<=': Op.lte,
'>=': Op.gte,
':': Op.like
'>': Op.gt,
'<': Op.lt,
':': Op.like,
AND: Op.and,
OR: Op.or
}
export class FilterParser {
asty: ASTY = new ASTY()
lexer: Tokenizr = this.createLexer()
lexer: Tokenizr
constructor() {}
constructor() {
this.lexer = new Tokenizr()
private createLexer(): Tokenizr {
const tkz = new Tokenizr()
tkz.rule(openGroupRegex, (ctx, m) => {
this.lexer.rule(openGroupRegex, (ctx, m) => {
ctx.accept('opengroup')
})
tkz.rule(closeGroupRegex, (ctx, m) => {
this.lexer.rule(closeGroupRegex, (ctx, m) => {
ctx.accept('closegroup')
})
tkz.rule(conjunctinoRegex, (ctx, m) => {
this.lexer.rule(conjunctinoRegex, (ctx, m) => {
ctx.accept('conjunction', m[0])
})
tkz.rule(equalityRegex, (ctx, m) => {
this.lexer.rule(equalityRegex, (ctx, m) => {
ctx.accept('column', m[1])
ctx.accept('opperator', m[2])
ctx.accept('value', m[3])
})
tkz.rule(spacerRegex, (ctx, m) => {
this.lexer.rule(spacerRegex, (ctx, m) => {
ctx.ignore()
})
return tkz
}
parseFilter(filter: string) {
parse(filter: string) {
this.lexer.input(filter)
console.log(`parsing ${filter}`)
this.lexer.begin()
this.lexer.tokens().forEach((token) => {
console.log(token)
})
this.lexer.rollback()
let block = this.parseBlock('AND')
this.lexer.consume('EOF')
this.lexer.reset()
return block
}
@@ -122,12 +111,8 @@ export class FilterParser {
() => this.parseEquality(),
() => this.parseGroup(),
() => {
console.log("Conjunct")
let conToken = this.lexer.consume('conjunction')
console.log("potato")
if (items.length === 1) {
activeCon = conToken.value
}
@@ -137,42 +122,36 @@ export class FilterParser {
} else {
return this.parseBlock(conToken.value)
}
}, () => {}
},
() => this.parseEmpty()
)
if (nextItem === undefined) {
console.log("breaking")
break
}
items.push(nextItem)
}
return this.asty.create('conjunction').set('type', activeCon).add(items)
return { [opperatorMap[activeCon]]: items }
}
private parseConjunction() {}
private parseEquality() {
console.log("doing an equality")
let columnToken = this.lexer.consume('column')
let opperatorToken = this.lexer.consume('opperator')
let valueToken = this.lexer.consume('value')
let node = this.asty.create('equality')
if (opperatorToken.value === ":") {
node.pos(columnToken.line, columnToken.column, columnToken.pos)
return {
[columnToken.value]: { [opperatorMap[opperatorToken.value]]: `%${valueToken.value.toString()}%` }
}
} else {
node.set(columnToken.type, columnToken.value)
node.set(opperatorToken.type, opperatorToken.value)
node.set(valueToken.type, valueToken.value)
node.set('expression', {
[columnToken.value]: {
[opperatorMap[opperatorToken.value]]: valueToken.value
}
})
return node
return {
[columnToken.value]: { [opperatorMap[opperatorToken.value]]: valueToken.value.toString() }
}
}
}
private parseGroup() {
@@ -182,4 +161,8 @@ export class FilterParser {
return block
}
private parseEmpty() {
return undefined
}
}