Skip to content

Commit

Permalink
Added assign sample, parses fine, now onto some rules =D
Browse files Browse the repository at this point in the history
  • Loading branch information
fasterthanlime committed Sep 28, 2009
1 parent f3e8dad commit 21f6cfd
Show file tree
Hide file tree
Showing 4 changed files with 14 additions and 6 deletions.
7 changes: 6 additions & 1 deletion Makefile
@@ -1,10 +1,15 @@
TESTFILE=source/Lexer.ooc
GDBCOMMANDS_PATH=/tmp/gdbcommands
.PHONY:all clean mrproper .PHONY:all clean mrproper


all: all:
ooc meta -sourcepath=source/:../rock/source/ -g ${OOC_FLAGS} ooc meta -sourcepath=source/:../rock/source/ -g ${OOC_FLAGS}


test: test:
make all && ./meta make all && ./meta ${TESTFILE}

debug:
echo run ${TESTFILE} >> ${GDBCOMMANDS_PATH} && gdb ./meta -command=${GDBCOMMANDS_PATH} && rm ${GDBCOMMANDS_PATH}


clean: clean:
rm -rf ooc_tmp/ rm -rf ooc_tmp/
Expand Down
1 change: 1 addition & 0 deletions samples/assign.ooc
@@ -0,0 +1 @@
var1=var2
8 changes: 5 additions & 3 deletions source/Lexer.ooc
Expand Up @@ -16,18 +16,20 @@ Lexer: class {
rules: ArrayList<Rule> rules: ArrayList<Rule>


init: func { init: func {
rules := ArrayList<Rule> new()
rules = ArrayList<Rule> new()

} }


parse: func (path: String) { parse: func (path: String) {


sReader := SourceReader getReaderFromPath(path) sReader := SourceReader getReaderFromPath(path)
tokens := Tokenizer new() parse(sReader) tokens := Tokenizer new() parse(sReader)


"Got tokens: " println()
for(token: Token in tokens) { for(token: Token in tokens) {
printf("%s ", token toString()) printf("%s ", token toString(sReader))
} }
println()


for(rule in rules) { for(rule in rules) {


Expand Down
4 changes: 2 additions & 2 deletions source/rules/TokenRule.ooc
@@ -1,7 +1,7 @@
import frontend/SourceReader import frontend/SourceReader
import MetaNode import ../Lexer


TokenMetaNode: class extends MetaNode { TokenRule: class extends MetaNode {


string : String string : String


Expand Down

0 comments on commit 21f6cfd

Please sign in to comment.