From 6fd6acc71b5b3f7599fa5db035c4aeef1d9a39fa Mon Sep 17 00:00:00 2001 From: Eugen Wissner Date: Fri, 30 May 2025 19:51:18 +0200 Subject: [PATCH] Handle ASCII codes > 128 in the tokenizer --- source/Lexer.mod | 190 +++++++++++++------------- source/Transpiler.mod | 304 ++++++++++++++++++++++-------------------- 2 files changed, 255 insertions(+), 239 deletions(-) diff --git a/source/Lexer.mod b/source/Lexer.mod index b4f54ca..0262b8e 100644 --- a/source/Lexer.mod +++ b/source/Lexer.mod @@ -72,6 +72,8 @@ VAR transitions: ARRAY[1..16] OF TransitionClasses; PROCEDURE initialize_classification(); +VAR + i: CARDINAL; BEGIN classification[1] := transitionClassEof; (* NUL *) classification[2] := transitionClassInvalid; (* SOH *) @@ -200,255 +202,261 @@ BEGIN classification[125] := transitionClassSingle; (* | *) classification[126] := transitionClassOther; (* } *) classification[127] := transitionClassSingle; (* ~ *) - classification[128] := transitionClassInvalid (* DEL *) + classification[128] := transitionClassInvalid; (* DEL *) + + i := 129; + WHILE i <= 256 DO + classification[i] := transitionClassOther; + i := i + 1 + END END initialize_classification; PROCEDURE compare_keyword(Keyword: ARRAY OF CHAR; TokenStart: PLexerBuffer; TokenEnd: PLexerBuffer): BOOLEAN; VAR - Result: BOOLEAN; - Index: CARDINAL; + result: BOOLEAN; + index: CARDINAL; BEGIN - Index := 0; - Result := TRUE; + index := 0; + result := TRUE; - WHILE (Index < Length(Keyword)) AND (TokenStart <> TokenEnd) AND Result DO - Result := (Keyword[Index] = TokenStart^) OR (Lower(Keyword[Index]) = TokenStart^); + WHILE (index < Length(Keyword)) AND (TokenStart <> TokenEnd) AND result DO + result := (Keyword[index] = TokenStart^) OR (Lower(Keyword[index]) = TokenStart^); INC(TokenStart); - INC(Index) + INC(index) END; - Result := (Index = Length(Keyword)) AND (TokenStart = TokenEnd) AND Result; - RETURN Result + result := (index = Length(Keyword)) AND (TokenStart = TokenEnd) AND result; + RETURN result END compare_keyword; (* Reached the end of file. *) -PROCEDURE transition_action_eof(lexer: PLexer; AToken: PLexerToken); +PROCEDURE transition_action_eof(lexer: PLexer; token: PLexerToken); BEGIN - AToken^.Kind := lexerKindEof + token^.Kind := lexerKindEof END transition_action_eof; (* Add the character to the token currently read and advance to the next character. *) -PROCEDURE transition_action_accumulate(lexer: PLexer; AToken: PLexerToken); +PROCEDURE transition_action_accumulate(lexer: PLexer; token: PLexerToken); BEGIN INC(lexer^.Current) END transition_action_accumulate; (* The current character is not a part of the token. Finish the token already * read. Don't advance to the next character. *) -PROCEDURE transition_action_finalize(lexer: PLexer; AToken: PLexerToken); +PROCEDURE transition_action_finalize(lexer: PLexer; token: PLexerToken); BEGIN IF lexer^.Start^ = ':' THEN - AToken^.Kind := lexerKindColon + token^.Kind := lexerKindColon END; IF lexer^.Start^ = '>' THEN - AToken^.Kind := lexerKindGreaterThan + token^.Kind := lexerKindGreaterThan END; IF lexer^.Start^ = '<' THEN - AToken^.Kind := lexerKindLessThan + token^.Kind := lexerKindLessThan END; IF lexer^.Start^ = '(' THEN - AToken^.Kind := lexerKindLeftParen + token^.Kind := lexerKindLeftParen END; IF lexer^.Start^ = '-' THEN - AToken^.Kind := lexerKindLeftParen + token^.Kind := lexerKindLeftParen END; IF lexer^.Start^ = '.' THEN - AToken^.Kind := lexerKindDot + token^.Kind := lexerKindDot END END transition_action_finalize; (* An action for tokens containing multiple characters. *) -PROCEDURE transition_action_composite(lexer: PLexer; AToken: PLexerToken); +PROCEDURE transition_action_composite(lexer: PLexer; token: PLexerToken); BEGIN IF lexer^.Start^ = '<' THEN IF lexer^.Current^ = '>' THEN - AToken^.Kind := lexerKindNotEqual + token^.Kind := lexerKindNotEqual END; IF lexer^.Current^ = '=' THEN - AToken^.Kind := lexerKindLessEqual + token^.Kind := lexerKindLessEqual END END; IF (lexer^.Start^ = '>') AND (lexer^.Current^ = '=') THEN - AToken^.Kind := lexerKindGreaterEqual + token^.Kind := lexerKindGreaterEqual END; IF (lexer^.Start^ = '.') AND (lexer^.Current^ = '.') THEN - AToken^.Kind := lexerKindRange + token^.Kind := lexerKindRange END; IF (lexer^.Start^ = ':') AND (lexer^.Current^ = '=') THEN - AToken^.Kind := lexerKindAssignment + token^.Kind := lexerKindAssignment END; INC(lexer^.Current) END transition_action_composite; (* Skip a space. *) -PROCEDURE transition_action_skip(lexer: PLexer; AToken: PLexerToken); +PROCEDURE transition_action_skip(lexer: PLexer; token: PLexerToken); BEGIN INC(lexer^.Current); INC(lexer^.Start) END transition_action_skip; (* Delimited string action. *) -PROCEDURE transition_action_delimited(lexer: PLexer; AToken: PLexerToken); +PROCEDURE transition_action_delimited(lexer: PLexer; token: PLexerToken); BEGIN IF lexer^.Start^ = '(' THEN - AToken^.Kind := lexerKindComment + token^.Kind := lexerKindComment END; IF lexer^.Start^ = '"' THEN - AToken^.Kind := lexerKindCharacter + token^.Kind := lexerKindCharacter END; IF lexer^.Start^ = "'" THEN - AToken^.Kind := lexerKindString + token^.Kind := lexerKindString END; INC(lexer^.Current) END transition_action_delimited; (* Finalize keyword OR identifier. *) -PROCEDURE transition_action_key_id(lexer: PLexer; AToken: PLexerToken); +PROCEDURE transition_action_key_id(lexer: PLexer; token: PLexerToken); BEGIN - AToken^.Kind := lexerKindIdentifier; + token^.Kind := lexerKindIdentifier; - AToken^.identifierKind[1] := lexer^.Current - lexer^.Start; - MemCopy(lexer^.Start, ORD(AToken^.identifierKind[1]), ADR(AToken^.identifierKind[2])); + token^.identifierKind[1] := lexer^.Current - lexer^.Start; + MemCopy(lexer^.Start, ORD(token^.identifierKind[1]), ADR(token^.identifierKind[2])); IF compare_keyword('PROGRAM', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindProgram + token^.Kind := lexerKindProgram END; IF compare_keyword('IMPORT', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindImport + token^.Kind := lexerKindImport END; IF compare_keyword('CONST', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindConst + token^.Kind := lexerKindConst END; IF compare_keyword('VAR', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindVar + token^.Kind := lexerKindVar END; IF compare_keyword('IF', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindIf + token^.Kind := lexerKindIf END; IF compare_keyword('THEN', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindThen + token^.Kind := lexerKindThen END; IF compare_keyword('ELSIF', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindElsif + token^.Kind := lexerKindElsif END; IF compare_keyword('ELSE', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindElse + token^.Kind := lexerKindElse END; IF compare_keyword('WHILE', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindWhile + token^.Kind := lexerKindWhile END; IF compare_keyword('DO', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindDo + token^.Kind := lexerKindDo END; IF compare_keyword('proc', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindProc + token^.Kind := lexerKindProc END; IF compare_keyword('BEGIN', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindBegin + token^.Kind := lexerKindBegin END; IF compare_keyword('END', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindEnd + token^.Kind := lexerKindEnd END; IF compare_keyword('TYPE', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindType + token^.Kind := lexerKindType END; IF compare_keyword('RECORD', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindRecord + token^.Kind := lexerKindRecord END; IF compare_keyword('UNION', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindUnion + token^.Kind := lexerKindUnion END; IF compare_keyword('NIL', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindNull + token^.Kind := lexerKindNull END; IF compare_keyword('AND', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindAnd + token^.Kind := lexerKindAnd END; IF compare_keyword('OR', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindOr + token^.Kind := lexerKindOr END; IF compare_keyword('RETURN', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindReturn + token^.Kind := lexerKindReturn END; IF compare_keyword('DEFINITION', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindDefinition + token^.Kind := lexerKindDefinition END; IF compare_keyword('TO', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindTo + token^.Kind := lexerKindTo END; IF compare_keyword('CASE', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindCase + token^.Kind := lexerKindCase END; IF compare_keyword('OF', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindOf + token^.Kind := lexerKindOf END; IF compare_keyword('FROM', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindFrom + token^.Kind := lexerKindFrom END; IF compare_keyword('MODULE', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindModule + token^.Kind := lexerKindModule END; IF compare_keyword('IMPLEMENTATION', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindImplementation + token^.Kind := lexerKindImplementation END; IF compare_keyword('POINTER', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindPointer + token^.Kind := lexerKindPointer END; IF compare_keyword('ARRAY', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindArray + token^.Kind := lexerKindArray END; IF compare_keyword('TRUE', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindBoolean; - AToken^.booleanKind := TRUE + token^.Kind := lexerKindBoolean; + token^.booleanKind := TRUE END; IF compare_keyword('FALSE', lexer^.Start, lexer^.Current) THEN - AToken^.Kind := lexerKindBoolean; - AToken^.booleanKind := FALSE + token^.Kind := lexerKindBoolean; + token^.booleanKind := FALSE END END transition_action_key_id; (* Action for tokens containing only one character. The character cannot be * followed by other characters forming a composite token. *) -PROCEDURE transition_action_single(lexer: PLexer; AToken: PLexerToken); +PROCEDURE transition_action_single(lexer: PLexer; token: PLexerToken); BEGIN IF lexer^.Current^ = '&' THEN - AToken^.Kind := lexerKindAnd + token^.Kind := lexerKindAnd END; IF lexer^.Current^ = ';' THEN - AToken^.Kind := lexerKindSemicolon + token^.Kind := lexerKindSemicolon END; IF lexer^.Current^ = ',' THEN - AToken^.Kind := lexerKindComma + token^.Kind := lexerKindComma END; IF lexer^.Current^ = ',' THEN - AToken^.Kind := lexerKindComma + token^.Kind := lexerKindComma END; IF lexer^.Current^ = ')' THEN - AToken^.Kind := lexerKindRightParen + token^.Kind := lexerKindRightParen END; IF lexer^.Current^ = '[' THEN - AToken^.Kind := lexerKindLeftSquare + token^.Kind := lexerKindLeftSquare END; IF lexer^.Current^ = ']' THEN - AToken^.Kind := lexerKindRightSquare + token^.Kind := lexerKindRightSquare END; IF lexer^.Current^ = '^' THEN - AToken^.Kind := lexerKindHat + token^.Kind := lexerKindHat END; IF lexer^.Current^ = '=' THEN - AToken^.Kind := lexerKindEqual + token^.Kind := lexerKindEqual END; IF lexer^.Current^ = '+' THEN - AToken^.Kind := lexerKindPlus + token^.Kind := lexerKindPlus END; IF lexer^.Current^ = '/' THEN - AToken^.Kind := lexerKindDivision + token^.Kind := lexerKindDivision END; IF lexer^.Current^ = '%' THEN - AToken^.Kind := lexerKindRemainder + token^.Kind := lexerKindRemainder END; IF lexer^.Current^ = '@' THEN - AToken^.Kind := lexerKindAt + token^.Kind := lexerKindAt END; IF lexer^.Current^ = '|' THEN - AToken^.Kind := lexerKindPipe + token^.Kind := lexerKindPipe END; INC(lexer^.Current) END transition_action_single; (* Handle an integer literal. *) -PROCEDURE transition_action_integer(lexer: PLexer; AToken: PLexerToken); +PROCEDURE transition_action_integer(lexer: PLexer; token: PLexerToken); BEGIN - AToken^.Kind := lexerKindInteger + token^.Kind := lexerKindInteger END transition_action_integer; PROCEDURE set_default_transition(CurrentState: TransitionState; DefaultAction: TransitionAction; NextState: TransitionState); VAR @@ -747,7 +755,7 @@ VAR CurrentClass: TransitionClass; CurrentState: TransitionState; CurrentTransition: Transition; - Result: LexerToken; + result: LexerToken; BEGIN lexer^.Current := lexer^.Start; CurrentState := transitionStateStart; @@ -757,15 +765,15 @@ BEGIN CurrentTransition := transitions[ORD(CurrentState) + 1][ORD(CurrentClass) + 1]; IF CurrentTransition.Action <> NIL THEN - CurrentTransition.Action(lexer, ADR(Result)) + CurrentTransition.Action(lexer, ADR(result)) END; CurrentState := CurrentTransition.NextState END; - RETURN Result + RETURN result END lexer_current; PROCEDURE lexer_lex(lexer: PLexer): LexerToken; VAR - Result: LexerToken; + result: LexerToken; BEGIN IF lexer^.Length = 0 THEN lexer^.Length := ReadNBytes(lexer^.Input, CHUNK_SIZE, lexer^.Buffer); @@ -773,8 +781,8 @@ BEGIN END; lexer^.Start := lexer^.Current; - Result := lexer_current(lexer); - RETURN Result + result := lexer_current(lexer); + RETURN result END lexer_lex; PROCEDURE lexer_destroy(lexer: PLexer); BEGIN diff --git a/source/Transpiler.mod b/source/Transpiler.mod index 97a806c..6df07bd 100644 --- a/source/Transpiler.mod +++ b/source/Transpiler.mod @@ -71,69 +71,69 @@ BEGIN END transpile_import_part; PROCEDURE transpile_constant(context: PTranspilerContext; lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; written_bytes: CARDINAL; BEGIN WriteString(' '); - Token := lexer_current(lexer); + token := lexer_current(lexer); written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); WriteString(' = '); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); write_semicolon() END transpile_constant; PROCEDURE transpile_constant_part(context: PTranspilerContext; lexer: PLexer): BOOLEAN; VAR - Token: LexerToken; + token: LexerToken; result: BOOLEAN; BEGIN - Token := lexer_current(lexer); - result := Token.Kind = lexerKindConst; + token := lexer_current(lexer); + result := token.Kind = lexerKindConst; IF result THEN WriteString('CONST'); WriteLn(); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); - WHILE Token.Kind = lexerKindIdentifier DO + WHILE token.Kind = lexerKindIdentifier DO transpile_constant(context, lexer); - Token := transpiler_lex(lexer) + token := transpiler_lex(lexer) END END; RETURN result END transpile_constant_part; PROCEDURE transpile_module(context: PTranspilerContext; lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; written_bytes: CARDINAL; BEGIN - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); - IF Token.Kind = lexerKindDefinition THEN + IF token.Kind = lexerKindDefinition THEN WriteString('DEFINITION '); - Token := transpiler_lex(lexer) + token := transpiler_lex(lexer) END; - IF Token.Kind = lexerKindImplementation THEN + IF token.Kind = lexerKindImplementation THEN WriteString('IMPLEMENTATION '); - Token := transpiler_lex(lexer) + token := transpiler_lex(lexer) END; WriteString('MODULE '); (* Write the module name and end the line with a semicolon and newline. *) - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); write_semicolon(); WriteLn(); (* Write the module body. *) - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); transpile_import_part(context, lexer); IF transpile_constant_part(context, lexer) THEN WriteLn() @@ -147,32 +147,32 @@ BEGIN WriteString('END '); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); Write('.'); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); WriteLn() END transpile_module; PROCEDURE transpile_type_fields(context: PTranspilerContext; lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; written_bytes: CARDINAL; BEGIN - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); - WHILE Token.Kind <> lexerKindEnd DO + WHILE token.Kind <> lexerKindEnd DO WriteString(' '); written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); WriteString(': '); transpile_type_expression(context, lexer); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); - IF Token.Kind = lexerKindSemicolon THEN - Token := transpiler_lex(lexer); + IF token.Kind = lexerKindSemicolon THEN + token := transpiler_lex(lexer); Write(';') END; WriteLn() @@ -180,7 +180,7 @@ BEGIN END transpile_type_fields; PROCEDURE transpile_record_type(context: PTranspilerContext; lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; BEGIN WriteString('RECORD'); WriteLn(); @@ -189,32 +189,32 @@ BEGIN END transpile_record_type; PROCEDURE transpile_pointer_type(context: PTranspilerContext; lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; written_bytes: CARDINAL; BEGIN - Token := lexer_current(lexer); + token := lexer_current(lexer); WriteString('POINTER TO '); - IF Token.Kind = lexerKindPointer THEN - Token := transpiler_lex(lexer) + IF token.Kind = lexerKindPointer THEN + token := transpiler_lex(lexer) END; transpile_type_expression(context, lexer) END transpile_pointer_type; PROCEDURE transpile_array_type(context: PTranspilerContext; lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; written_bytes: CARDINAL; BEGIN WriteString('ARRAY'); - Token := lexer_current(lexer); + token := lexer_current(lexer); - IF Token.Kind = lexerKindArray THEN - Token := transpiler_lex(lexer) + IF token.Kind = lexerKindArray THEN + token := transpiler_lex(lexer) END; - IF Token.Kind <> lexerKindOf THEN + IF token.Kind <> lexerKindOf THEN WriteString('[1..'); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); Write(']') END; WriteString(' OF '); @@ -222,50 +222,50 @@ BEGIN END transpile_array_type; PROCEDURE transpile_enumeration_type(context: PTranspilerContext; lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; written_bytes: CARDINAL; BEGIN WriteString('('); WriteLn(); WriteString(' '); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); - WHILE Token.Kind = lexerKindComma DO + WHILE token.Kind = lexerKindComma DO Write(','); WriteLn(); WriteString(' '); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); - Token := transpiler_lex(lexer) + token := transpiler_lex(lexer) END; WriteLn(); WriteString(' )') END transpile_enumeration_type; PROCEDURE transpile_union_type(context: PTranspilerContext; lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; END transpile_union_type; PROCEDURE transpile_procedure_type(context: PTranspilerContext; lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; written_bytes: CARDINAL; BEGIN - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); WriteString('PROCEDURE('); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); - WHILE Token.Kind <> lexerKindRightParen DO + WHILE token.Kind <> lexerKindRightParen DO written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); - Token := transpiler_lex(lexer); - IF Token.Kind = lexerKindComma THEN - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); + IF token.Kind = lexerKindComma THEN + token := transpiler_lex(lexer); WriteString(', ') END END; @@ -273,42 +273,42 @@ BEGIN END transpile_procedure_type; PROCEDURE transpile_type_expression(context: PTranspilerContext; lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; written_bytes: CARDINAL; BEGIN - Token := transpiler_lex(lexer); - IF Token.Kind = lexerKindRecord THEN + token := transpiler_lex(lexer); + IF token.Kind = lexerKindRecord THEN transpile_record_type(context, lexer) END; - IF Token.Kind = lexerKindLeftParen THEN + IF token.Kind = lexerKindLeftParen THEN transpile_enumeration_type(context, lexer) END; - IF (Token.Kind = lexerKindArray) OR (Token.Kind = lexerKindLeftSquare) THEN + IF (token.Kind = lexerKindArray) OR (token.Kind = lexerKindLeftSquare) THEN transpile_array_type(context, lexer) END; - IF Token.Kind = lexerKindHat THEN + IF token.Kind = lexerKindHat THEN transpile_pointer_type(context, lexer) END; - IF Token.Kind = lexerKindProc THEN + IF token.Kind = lexerKindProc THEN transpile_procedure_type(context, lexer) END; - IF Token.Kind = lexerKindIdentifier THEN + IF token.Kind = lexerKindIdentifier THEN written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start) END END transpile_type_expression; PROCEDURE transpile_type_declaration(context: PTranspilerContext; lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; written_bytes: CARDINAL; BEGIN WriteString(' '); - Token := lexer_current(lexer); + token := lexer_current(lexer); written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); WriteString(' = '); transpile_type_expression(context, lexer); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); write_semicolon(); END transpile_type_declaration; PROCEDURE transpile_type_part(context: PTranspilerContext; lexer: PLexer); @@ -331,35 +331,35 @@ BEGIN END transpile_type_part; PROCEDURE transpile_variable_declaration(context: PTranspilerContext; lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; written_bytes: CARDINAL; BEGIN WriteString(' '); - Token := lexer_current(lexer); + token := lexer_current(lexer); written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); WriteString(': '); transpile_type_expression(context, lexer); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); write_semicolon() END transpile_variable_declaration; PROCEDURE transpile_variable_part(context: PTranspilerContext; lexer: PLexer): BOOLEAN; VAR - Token: LexerToken; + token: LexerToken; result: BOOLEAN; BEGIN - Token := lexer_current(lexer); - result := Token.Kind = lexerKindVar; + token := lexer_current(lexer); + result := token.Kind = lexerKindVar; IF result THEN WriteString('VAR'); WriteLn(); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); - WHILE Token.Kind = lexerKindIdentifier DO + WHILE token.Kind = lexerKindIdentifier DO transpile_variable_declaration(context, lexer); - Token := transpiler_lex(lexer) + token := transpiler_lex(lexer) END END; RETURN result @@ -408,14 +408,14 @@ BEGIN RETURN result END transpile_procedure_heading; -PROCEDURE transpile_expression(context: PTranspilerContext; lexer: PLexer; TrailingToken: LexerKind); +PROCEDURE transpile_expression(context: PTranspilerContext; lexer: PLexer; trailing_token: LexerKind); VAR token: LexerToken; written_bytes: CARDINAL; BEGIN token := transpiler_lex(lexer); - WHILE (token.Kind <> TrailingToken) AND (token.Kind <> lexerKindEnd) DO + WHILE (token.Kind <> trailing_token) AND (token.Kind <> lexerKindEnd) DO written_bytes := 0; IF token.Kind = lexerKindNull THEN WriteString('NIL '); @@ -456,7 +456,7 @@ BEGIN END transpile_if_statement; PROCEDURE transpile_while_statement(context: PTranspilerContext; lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; written_bytes: CARDINAL; BEGIN WriteString(' WHILE '); @@ -466,7 +466,7 @@ BEGIN WriteLn(); transpile_statements(context, lexer); WriteString(' END'); - Token := transpiler_lex(lexer) + token := transpiler_lex(lexer) END transpile_while_statement; PROCEDURE transpile_assignment_statement(context: PTranspilerContext; lexer: PLexer); BEGIN @@ -475,26 +475,70 @@ BEGIN END transpile_assignment_statement; PROCEDURE transpile_call_statement(context: PTranspilerContext; lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; written_bytes: CARDINAL; BEGIN WriteString('('); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); - WHILE (Token.Kind <> lexerKindSemicolon) AND (Token.Kind <> lexerKindEnd) DO + WHILE (token.Kind <> lexerKindSemicolon) AND (token.Kind <> lexerKindEnd) DO written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); - Token := transpiler_lex(lexer) + token := transpiler_lex(lexer) END END transpile_call_statement; -PROCEDURE transpile_return_statement(context: PTranspilerContext; lexer: PLexer); +PROCEDURE transpile_designator_expression(context: PTranspilerContext; lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; + written_bytes: CARDINAL; +BEGIN + WriteString(' '); + written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); + token := transpiler_lex(lexer); + + WHILE token.Kind = lexerKindLeftSquare DO + Write('['); + token := transpiler_lex(lexer); + WHILE token.Kind <> lexerKindRightSquare DO + written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); + token := transpiler_lex(lexer) + END; + Write(']'); + token := transpiler_lex(lexer) + END; + IF token.Kind = lexerKindHat THEN + Write('^'); + token := transpiler_lex(lexer) + END; + IF token.Kind = lexerKindDot THEN + Write('.'); + token := transpiler_lex(lexer); + written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); + token := transpiler_lex(lexer) + END; + IF token.Kind = lexerKindHat THEN + Write('^'); + token := transpiler_lex(lexer) + END; + WHILE token.Kind = lexerKindLeftSquare DO + Write('['); + token := transpiler_lex(lexer); + WHILE token.Kind <> lexerKindRightSquare DO + written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); + token := transpiler_lex(lexer) + END; + Write(']'); + token := transpiler_lex(lexer) + END +END transpile_designator_expression; +PROCEDURE transpile_return_statement(context: PTranspilerContext; lexer: PLexer); +VAR + token: LexerToken; written_bytes: CARDINAL; BEGIN WriteString(' RETURN '); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); - Token := transpiler_lex(lexer) + token := transpiler_lex(lexer) END transpile_return_statement; PROCEDURE transpile_statement(context: PTranspilerContext; lexer: PLexer); VAR @@ -513,44 +557,8 @@ BEGIN transpile_return_statement(context, lexer) END; IF token.Kind = lexerKindIdentifier THEN - WriteString(' '); - written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); - token := transpiler_lex(lexer); - - WHILE token.Kind = lexerKindLeftSquare DO - Write('['); - token := transpiler_lex(lexer); - WHILE token.Kind <> lexerKindRightSquare DO - written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); - token := transpiler_lex(lexer) - END; - Write(']'); - token := transpiler_lex(lexer); - END; - IF token.Kind = lexerKindHat THEN - Write('^'); - token := transpiler_lex(lexer) - END; - IF token.Kind = lexerKindDot THEN - Write('.'); - token := transpiler_lex(lexer); - written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); - token := transpiler_lex(lexer); - END; - IF token.Kind = lexerKindHat THEN - Write('^'); - token := transpiler_lex(lexer) - END; - WHILE token.Kind = lexerKindLeftSquare DO - Write('['); - token := transpiler_lex(lexer); - WHILE token.Kind <> lexerKindRightSquare DO - written_bytes := WriteNBytes(StdOut, ADDRESS(lexer^.Current - lexer^.Start), lexer^.Start); - token := transpiler_lex(lexer) - END; - Write(']'); - token := transpiler_lex(lexer); - END; + transpile_designator_expression(context, lexer); + token := lexer_current(lexer); IF token.Kind = lexerKindAssignment THEN transpile_assignment_statement(context, lexer) @@ -562,15 +570,15 @@ BEGIN END transpile_statement; PROCEDURE transpile_statements(context: PTranspilerContext; lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; BEGIN - Token := lexer_current(lexer); + token := lexer_current(lexer); - WHILE Token.Kind <> lexerKindEnd DO + WHILE token.Kind <> lexerKindEnd DO transpile_statement(context, lexer); - Token := lexer_current(lexer); + token := lexer_current(lexer); - IF Token.Kind = lexerKindSemicolon THEN + IF token.Kind = lexerKindSemicolon THEN Write(';') END; WriteLn() @@ -578,10 +586,10 @@ BEGIN END transpile_statements; PROCEDURE transpile_statement_part(context: PTranspilerContext; lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; BEGIN - Token := lexer_current(lexer); - IF Token.Kind = lexerKindBegin THEN + token := lexer_current(lexer); + IF token.Kind = lexerKindBegin THEN WriteString('BEGIN'); WriteLn(); transpile_statements(context, lexer) @@ -589,40 +597,40 @@ BEGIN END transpile_statement_part; PROCEDURE transpile_procedure_declaration(context: PTranspilerContext; lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; seen_part: BOOLEAN; written_bytes: CARDINAL; BEGIN - Token := transpile_procedure_heading(context, lexer); + token := transpile_procedure_heading(context, lexer); seen_part := transpile_constant_part(context, lexer); seen_part := transpile_variable_part(context, lexer); transpile_statement_part(context, lexer); WriteString('END '); - written_bytes := WriteNBytes(StdOut, ORD(Token.identifierKind[1]), ADR(Token.identifierKind[2])); + written_bytes := WriteNBytes(StdOut, ORD(token.identifierKind[1]), ADR(token.identifierKind[2])); - Token := transpiler_lex(lexer); + token := transpiler_lex(lexer); write_semicolon(); - Token := transpiler_lex(lexer) + token := transpiler_lex(lexer) END transpile_procedure_declaration; PROCEDURE transpile_procedure_part(context: PTranspilerContext; lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; BEGIN - Token := lexer_current(lexer); + token := lexer_current(lexer); - WHILE Token.Kind = lexerKindProc DO + WHILE token.Kind = lexerKindProc DO transpile_procedure_declaration(context, lexer); - Token := lexer_current(lexer); + token := lexer_current(lexer); WriteLn() END END transpile_procedure_part; PROCEDURE transpile(lexer: PLexer); VAR - Token: LexerToken; + token: LexerToken; written_bytes: CARDINAL; - Context: TranspilerContext; + context: TranspilerContext; BEGIN - transpile_module(ADR(Context), lexer) + transpile_module(ADR(context), lexer) END transpile; END Transpiler.