var searchIndex = {}; searchIndex["luthor"] = {"doc":"","items":[[0,"lexers","luthor","",null,null],[0,"json","luthor::lexers","",null,null],[5,"lex","luthor::lexers::json","",null,{"inputs":[{"name":"str"}],"output":{"name":"vec"}}],[0,"xml","luthor::lexers","",null,null],[5,"lex","luthor::lexers::xml","",null,{"inputs":[{"name":"str"}],"output":{"name":"vec"}}],[0,"ruby","luthor::lexers","",null,null],[5,"lex","luthor::lexers::ruby","",null,{"inputs":[{"name":"str"}],"output":{"name":"vec"}}],[0,"default","luthor::lexers","",null,null],[5,"lex","luthor::lexers::default","",null,{"inputs":[{"name":"str"}],"output":{"name":"vec"}}],[0,"token","luthor","",null,null],[3,"Token","luthor::token","",null,null],[12,"lexeme","","",0,null],[12,"category","","",0,null],[4,"Category","","",null,null],[13,"Whitespace","","",1,null],[13,"Identifier","","",1,null],[13,"Keyword","","",1,null],[13,"Brace","","",1,null],[13,"Bracket","","",1,null],[13,"Parenthesis","","",1,null],[13,"Operator","","",1,null],[13,"Integer","","",1,null],[13,"Float","","",1,null],[13,"String","","",1,null],[13,"Boolean","","",1,null],[13,"Text","","",1,null],[13,"Comment","","",1,null],[13,"Method","","",1,null],[11,"clone","","",1,null],[11,"fmt","","",1,null],[11,"eq","","",1,null],[11,"clone","","",0,null],[11,"fmt","","",0,null],[11,"eq","","",0,null],[11,"ne","","",0,null],[0,"tokenizer","luthor","",null,null],[3,"StateFunction","luthor::tokenizer","",null,null],[12,"0","","",2,null],[3,"Tokenizer","","The Tokenizer type is used to produce and store\ntokens for the various language and format lexers.",null,null],[12,"states","","",3,null],[5,"new","","Initializes a new tokenizer with the given data.",null,{"inputs":[{"name":"str"}],"output":{"name":"tokenizer"}}],[11,"tokens","","Returns a copy of the tokens processed to date.",3,null],[11,"advance","","Moves to the next character in the data.\nDoes nothing if there is no more data to process.",3,null],[11,"current_char","","Returns the character at the current position,\nunless all of the data has been processed.",3,null],[11,"starts_with","","Whether or not the remaining data starts\nwith the specified string.",3,null],[11,"tokenize","","Creates and stores a token with the given category containing any\ndata processed using `advance` since the last call to this method.",3,null],[11,"tokenize_next","","Creates and stores a token with the given category and the\nnext `amount` characters of the data. Before doing this, it\ntokenizes any previously processed characters with the generic\nCategory::Text category.",3,null]],"paths":[[3,"Token"],[4,"Category"],[3,"StateFunction"],[3,"Tokenizer"]]}; initSearch(searchIndex);