mirror of
https://github.com/jorenchik/mdemory.git
synced 2026-03-22 00:26:21 +00:00
Refactoring and custom deck file extension
This commit is contained in:
@@ -16,15 +16,17 @@ func main() {
|
||||
}
|
||||
fileContents := string(file)
|
||||
|
||||
tokens, err := tokenize([]rune(fileContents))
|
||||
tokens, err := tokenizeMdem([]rune(fileContents))
|
||||
if (err != nil) {
|
||||
fmt.Printf("%s\n", err.Error())
|
||||
return
|
||||
}
|
||||
if (true) {
|
||||
log.Println("Lexer output:")
|
||||
for _, el := range tokens {
|
||||
fmt.Print(el.ToString())
|
||||
}
|
||||
}
|
||||
|
||||
automata = parserAutomata()
|
||||
err = validateGrammar()
|
||||
@@ -36,10 +38,12 @@ func main() {
|
||||
if (err != nil) {
|
||||
log.Fatal(err.Error())
|
||||
}
|
||||
if (true) {
|
||||
log.Println("Parser output:")
|
||||
for _, element := range questions {
|
||||
fmt.Printf("%s", element.ToString())
|
||||
}
|
||||
}
|
||||
|
||||
log.Println("Compilation completed")
|
||||
}
|
||||
|
||||
@@ -1,4 +1,3 @@
|
||||
|
||||
[MathTest]
|
||||
items = {graph_theory.networks, linear_algebra.*}
|
||||
base_interval = 2
|
||||
@@ -1,5 +1,3 @@
|
||||
module github.com/jorenchik/mdemory/src/compiler
|
||||
|
||||
go 1.22.5
|
||||
|
||||
require golang.org/x/text v0.16.0
|
||||
|
||||
@@ -1,2 +0,0 @@
|
||||
golang.org/x/text v0.16.0 h1:a94ExnEXNtEwYLGJSIUxnWoxoRz/ZcCsV63ROupILh4=
|
||||
golang.org/x/text v0.16.0/go.mod h1:GhwF1Be+LQoKShO3cGOHzqOgRrGaYc9AvblQOmPVHnI=
|
||||
|
||||
@@ -5,7 +5,6 @@ import (
|
||||
"strings"
|
||||
)
|
||||
|
||||
var tokens []Token
|
||||
var buffer []rune
|
||||
var row int32 = 1
|
||||
var column int32 = 1
|
||||
@@ -37,7 +36,6 @@ type Token struct {
|
||||
column int32;
|
||||
}
|
||||
|
||||
|
||||
func (token Token)ToString() string {
|
||||
content := token.content
|
||||
if (token.tokenType == TextFragment) {
|
||||
@@ -57,6 +55,7 @@ func (token Token)ToString() string {
|
||||
)
|
||||
}
|
||||
|
||||
var tokens []Token
|
||||
|
||||
func makePostTextToken(ttype TokenType, tokenLen int32, textType TokenType) {
|
||||
if (len(strings.Trim(string(buffer), " \n\t")) - 1 > 0) {
|
||||
@@ -89,7 +88,7 @@ func makePostTextToken(ttype TokenType, tokenLen int32, textType TokenType) {
|
||||
buffer = []rune{}
|
||||
}
|
||||
|
||||
func tokenize(fileRunes []rune) ( []Token, error ) {
|
||||
func tokenizeMdem(fileRunes []rune) ( []Token, error ) {
|
||||
tokens = []Token{}
|
||||
buffer = []rune{}
|
||||
|
||||
@@ -187,29 +186,17 @@ func tokenize(fileRunes []rune) ( []Token, error ) {
|
||||
|
||||
func ToString (ttype *TokenType) string {
|
||||
switch *ttype {
|
||||
case TextFragment:
|
||||
return "TextFragment"
|
||||
case QuestionEnd:
|
||||
return "QuestionEnd"
|
||||
case ElementDashStart:
|
||||
return "ElementDashStart"
|
||||
case ElementPlusStart:
|
||||
return "ElementPlusStart"
|
||||
case Identifier:
|
||||
return "Identifier"
|
||||
case IdentifierStart:
|
||||
return "IdentifierStart"
|
||||
case IdentifierEnd:
|
||||
return "IdentifierEnd"
|
||||
case SectionIdentifierStart:
|
||||
return "SectionIdentifierStart"
|
||||
case SectionStart:
|
||||
return "SectionStart"
|
||||
case SectionEnd:
|
||||
return "SectionEnd"
|
||||
case EOF:
|
||||
return "EndOfFile"
|
||||
default:
|
||||
return "NOT_DEFINED"
|
||||
case TextFragment: return "TextFragment"
|
||||
case QuestionEnd: return "QuestionEnd"
|
||||
case ElementDashStart: return "ElementDashStart"
|
||||
case ElementPlusStart: return "ElementPlusStart"
|
||||
case Identifier: return "Identifier"
|
||||
case IdentifierStart: return "IdentifierStart"
|
||||
case IdentifierEnd: return "IdentifierEnd"
|
||||
case SectionIdentifierStart: return "SectionIdentifierStart"
|
||||
case SectionStart: return "SectionStart"
|
||||
case SectionEnd: return "SectionEnd"
|
||||
case EOF: return "EndOfFile"
|
||||
default: return "NOT_DEFINED"
|
||||
}
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user