| /src/tools/m4/bootstrap/ |
| Makefile | 3 FILES=parser.h parser.c tokenizer.c 17 tokenizer.c: ${DIST}/tokenizer.l 18 ${LEX} ${DIST}/tokenizer.l
|
| /src/tools/m4/ |
| Makefile | 17 tokenizer.c: ${.CURDIR}/bootstrap/tokenizer.c
|
| /src/external/bsd/atf/dist/tools/ |
| parser_test.cpp | 118 // Tests for the "tokenizer" class. 134 class tokenizer : public tools::parser::tokenizer< std::istream > { class in namespace:minimal 136 tokenizer(std::istream& is, bool skipws) : function in class:minimal::tokenizer 137 tools::parser::tokenizer< std::istream > 154 class tokenizer : public tools::parser::tokenizer< std::istream > { class in namespace:delims 156 tokenizer(std::istream& is, bool skipws) : function in class:delims::tokenizer 157 tools::parser::tokenizer< std::istream > 177 class tokenizer : public tools::parser::tokenizer< std::istream > class in namespace:keywords 179 tokenizer(std::istream& is, bool skipws) : function in class:keywords::tokenizer 198 class tokenizer : public tools::parser::tokenizer< std::istream > { class in namespace:quotes 200 tokenizer(std::istream& is, bool skipws) : function in class:quotes::tokenizer [all...] |
| reader.cpp | 90 class tokenizer : public tools::parser::tokenizer< std::istream > { class in namespace:atf_tps 92 tokenizer(std::istream& is, size_t curline) : function in class:atf_tps::tokenizer 93 tools::parser::tokenizer< std::istream > 120 read_timeval(tools::parser::parser< atf_tps::tokenizer >& parser) 213 tools::parser::parser< tokenizer >& p = 214 *reinterpret_cast< tools::parser::parser< tokenizer >* > 232 tools::parser::parser< tokenizer >& p = 233 *reinterpret_cast< tools::parser::parser< tokenizer >* > 304 tools::parser::parser< tokenizer >& p [all...] |
| parser.cpp | 212 // The header tokenizer. 225 class tokenizer : public impl::tokenizer< std::istream > { class in namespace:header 227 tokenizer(std::istream& is, size_t curline) : function in class:header::tokenizer 228 impl::tokenizer< std::istream > 239 impl::parser< header::tokenizer >& 240 read(impl::parser< header::tokenizer >& p, impl::header_entry& he) 319 header::tokenizer tkz(is, curline); 320 impl::parser< header::tokenizer > p(tkz);
|
| config_file.cpp | 56 class tokenizer : public tools::parser::tokenizer< std::istream > { class in namespace:__anon4079::atf_config 58 tokenizer(std::istream& is, size_t curline) : function in class:__anon4079::atf_config::tokenizer 59 tools::parser::tokenizer< std::istream > 161 tokenizer tkz(m_is, hml.first); 162 tools::parser::parser< tokenizer > p(tkz);
|
| parser.hpp | 117 // The "tokenizer" class. 121 //! \brief A stream tokenizer. 124 //! tokenizer. It is only able to recognize one character-long delimiters, 131 class tokenizer { class in namespace:tools::parser 154 tokenizer(IS&, bool, const token_type&, const token_type&, 168 tokenizer< IS >::tokenizer(IS& p_is, function in class:tools::parser::tokenizer 186 tokenizer< IS >::lineno(void) 194 tokenizer< IS >::add_delim(char delim, const token_type& type) 202 tokenizer< IS >::add_keyword(const std::string& keyword [all...] |
| atffile.cpp | 66 class tokenizer : public tools::parser::tokenizer< std::istream > { class in namespace:atf_atffile 68 tokenizer(std::istream& is, size_t curline) : function in class:atf_atffile::tokenizer 69 tools::parser::tokenizer< std::istream > 135 tokenizer tkz(m_is, hml.first); 136 tools::parser::parser< tokenizer > p(tkz);
|
| test-program.cpp | 89 class tokenizer : public tools::parser::tokenizer< std::istream > { class in namespace:__anon4086::atf_tp 91 tokenizer(std::istream& is, size_t curline) : function in class:__anon4086::atf_tp::tokenizer 92 tools::parser::tokenizer< std::istream > 497 tokenizer tkz(m_is, hml.first); 498 tools::parser::parser< tokenizer > p(tkz);
|
| /src/usr.bin/m4/ |
| Makefile | 11 SRCS= parser.y tokenizer.l eval.c expr.c look.c main.c misc.c gnum4.c trace.c 23 tokenizer.o: parser.h 25 CLEANFILES+=parser.c parser.h tokenizer.o
|
| /src/external/gpl3/gcc.old/dist/libcc1/ |
| findcomp.cc | 80 class tokenizer class 84 tokenizer (const char *str) function in class:tokenizer 131 tokenizer dirs (cpath);
|
| /src/external/gpl3/gcc/dist/libcc1/ |
| findcomp.cc | 80 class tokenizer class 84 tokenizer (const char *str) function in class:tokenizer 131 tokenizer dirs (cpath);
|
| /src/lib/libedit/ |
| Makefile | 20 terminal.c tokenizer.c tokenizern.c tty.c vi.c 143 COPTS.tokenizer.c+= -Wno-cast-qual
|
| histedit.h | 242 typedef struct tokenizer Tokenizer; 247 Tokenizer *tok_init(const char *); 248 void tok_end(Tokenizer *); 249 void tok_reset(Tokenizer *); 250 int tok_line(Tokenizer *, const LineInfo *, 252 int tok_str(Tokenizer *, const char *, 307 /* Wide character tokenizer support */
|
| tokenizer.c | 1 /* $NetBSD: tokenizer.c,v 1.29 2023/05/30 11:53:40 christos Exp $ */ 38 static char sccsid[] = "@(#)tokenizer.c 8.1 (Berkeley) 6/4/93"; 40 __RCSID("$NetBSD: tokenizer.c,v 1.29 2023/05/30 11:53:40 christos Exp $"); 46 * tokenize.c: Bourne shell like tokenizer 85 struct TYPE(tokenizer) { 97 static void FUN(tok,finish)(TYPE(Tokenizer) *); 101 * Finish a word in the tokenizer. 104 FUN(tok,finish)(TYPE(Tokenizer) *tok) 118 * Initialize the tokenizer 120 TYPE(Tokenizer) * [all...] |
| /src/usr.bin/fgen/ |
| fgen.l | 91 static int tokenizer = 0; 285 tokenizer\[ { ltoken.type = TOK_BEGTOK; ltoken.text = yytext; 291 \]tokenizer { ltoken.type = TOK_ENDTOK; ltoken.text = yytext; 324 * tokenizer aliases. 1095 if (tokenizer) { 1638 /* This is actually a tokenizer directive. */ 1648 * Another tokenizer directive. 1943 /* This is actually a tokenizer directive. */ 1958 /* Tokenizer directives */ 1961 tokenizer = 1 [all...] |
| /src/external/gpl3/gcc.old/dist/libcpp/ |
| lex.cc | 3340 /* Maybe tell the tokenizer we expect a header-name down the 5338 Start up the tokenizer. */
|
| /src/external/gpl3/gcc/dist/libcpp/ |
| lex.cc | 3613 /* Maybe tell the tokenizer we expect a header-name down the 5658 Start up the tokenizer. */
|
| /src/external/public-domain/sqlite/dist/ |
| sqlite3.c | 13180 ** Tokenize text using the tokenizer belonging to the FTS5 table. 13329 ** token. It is the output of the tokenizer module. For tokendata=1 13343 ** It is the output of the tokenizer module. For tokendata=1 tables, this 13401 ** Applications may also register custom tokenizer types. A tokenizer 13408 ** This function is used to allocate and initialize a tokenizer instance. 13409 ** A tokenizer instance is required to actually tokenize text. 13415 ** containing the tokenizer arguments, if any, specified following the 13416 ** tokenizer name as part of the CREATE VIRTUAL TABLE statement used 13420 ** should be set to point to the new tokenizer handle and SQLITE_O 252555 fts5_tokenizer tokenizer; \/* Parent tokenizer module *\/ member in struct:PorterTokenizer [all...] |