org.antlr.codegen
Class JavaScriptTarget
java.lang.Object
org.antlr.codegen.Target
org.antlr.codegen.JavaScriptTarget
public class JavaScriptTarget
- extends Target
| Methods inherited from class org.antlr.codegen.Target |
genRecognizerFile, genRecognizerHeaderFile, getMaxCharValue, getTargetCharLiteralFromANTLRCharLiteral, getTargetStringLiteralFromANTLRStringLiteral, getTargetStringLiteralFromString, getTargetStringLiteralFromString, getTokenTypeAsTargetLabel, isValidActionScope, performGrammarAnalysis, postProcessAction |
| Methods inherited from class java.lang.Object |
clone, equals, finalize, getClass, hashCode, notify, notifyAll, toString, wait, wait, wait |
JavaScriptTarget
public JavaScriptTarget()
encodeIntAsCharEscape
public String encodeIntAsCharEscape(int v)
- Convert an int to a JavaScript Unicode character literal.
The current JavaScript spec (ECMA-262) doesn't provide for octal
notation in String literals, although some implementations support it.
This method overrides the parent class so that characters will always
be encoded as Unicode literals (e.g. ).
- Overrides:
encodeIntAsCharEscape in class Target
getTarget64BitStringFromValue
public String getTarget64BitStringFromValue(long word)
- Convert long to two 32-bit numbers separted by a comma.
JavaScript does not support 64-bit numbers, so we need to break
the number into two 32-bit literals to give to the Bit. A number like
0xHHHHHHHHLLLLLLLL is broken into the following string:
"0xLLLLLLLL, 0xHHHHHHHH"
Note that the low order bits are first, followed by the high order bits.
This is to match how the BitSet constructor works, where the bits are
passed in in 32-bit chunks with low-order bits coming first.
Note: stole the following two methods from the ActionScript target.
- Overrides:
getTarget64BitStringFromValue in class Target
Copyright © 2011. All Rights Reserved.