Skip site navigation (1)Skip section navigation (2)

FreeBSD Man Pages

Man Page or Keyword Search:
Man Architecture
Apropos Keyword Search (all sections) Output format
home | help
FLEX(1)								       FLEX(1)

NAME
       flex - fast lexical analyzer generator

SYNOPSIS
       flex [-bcdfhilnpstvwBFILTV78+? -C[aefFmr] -ooutput -Pprefix -Sskeleton]
       [--help --version] [filename ...]

OVERVIEW
       This manual describes flex, a tool for generating programs that perform
       pattern-matching	on text.  The manual includes both tutorial and	refer-
       ence sections:

	   Description
	       a brief overview	of the tool

	   Some	Simple Examples

	   Format Of The Input File

	   Patterns
	       the extended regular expressions	used by	flex

	   How The Input Is Matched
	       the rules for determining what has been matched

	   Actions
	       how to specify what to do when a	pattern	is matched

	   The Generated Scanner
	       details regarding the scanner that flex produces;
	       how to control the input	source

	   Start Conditions
	       introducing context into	your scanners, and
	       managing	"mini-scanners"

	   Multiple Input Buffers
	       how to manipulate multiple input	sources; how to
	       scan from strings instead of files

	   End-of-file Rules
	       special rules for matching the end of the input

	   Miscellaneous Macros
	       a summary of macros available to	the actions

	   Values Available To The User
	       a summary of values available to	the actions

	   Interfacing With Yacc
	       connecting flex scanners	together with yacc parsers

	   Options
	       flex command-line options, and the "%option"
	       directive

	   Performance Considerations
	       how to make your	scanner	go as fast as possible

	   Generating C++ Scanners
	       the (experimental) facility for generating C++
	       scanner classes

	   Incompatibilities With Lex And POSIX
	       how flex	differs	from AT&T lex and the POSIX lex
	       standard

	   Diagnostics
	       those error messages produced by	flex (or scanners
	       it generates) whose meanings might not be apparent

	   Files
	       files used by flex

	   Deficiencies	/ Bugs
	       known problems with flex

	   See Also
	       other documentation, related tools

	   Author
	       includes	contact	information

DESCRIPTION
       flex is a tool for generating scanners: programs	which recognize	 lexi-
       cal  patterns  in text.	flex reads the given input files, or its stan-
       dard input if no	file names are given, for a description	of  a  scanner
       to  generate.   The  description	 is  in	 the  form of pairs of regular
       expressions and C code, called rules. flex  generates  as  output  a  C
       source  file,  lex.yy.c,	which defines a	routine	yylex().  This file is
       compiled	and linked with	the -ll	 library  to  produce  an  executable.
       When  the  executable  is run, it analyzes its input for	occurrences of
       the regular expressions.	 Whenever it finds one,	it executes the	corre-
       sponding	C code.

SOME SIMPLE EXAMPLES
       First some simple examples to get the flavor of how one uses flex.  The
       following flex input specifies a	scanner	which whenever	it  encounters
       the string "username" will replace it with the user's login name:

	   %%
	   username    printf( "%s", getlogin()	);

       By  default,  any  text	not matched by a flex scanner is copied	to the
       output, so the net effect of this scanner is to copy its	input file  to
       its output with each occurrence of "username" expanded.	In this	input,
       there is	just one rule.	"username" is the pattern and the "printf"  is
       the action.  The	"%%" marks the beginning of the	rules.

       Here's another simple example:

	   %{
		   int num_lines = 0, num_chars	= 0;
	   %}

	   %%
	   \n	   ++num_lines;	++num_chars;
	   .	   ++num_chars;

	   %%
	   main()
		   {
		   yylex();
		   printf( "# of lines = %d, # of chars	= %d\n",
			   num_lines, num_chars	);
		   }

       This scanner counts the number of characters and	the number of lines in
       its input (it produces no output	other than the	final  report  on  the
       counts).	   The	first  line  declares  two  globals,  "num_lines"  and
       "num_chars", which are accessible both inside yylex() and in the	main()
       routine declared	after the second "%%".	There are two rules, one which
       matches a newline ("\n")	and increments both the	 line  count  and  the
       character  count, and one which matches any character other than	a new-
       line (indicated by the "." regular expression).

       A somewhat more complicated example:

	   /* scanner for a toy	Pascal-like language */

	   %{
	   /* need this	for the	call to	atof() below */
	   #include <math.h>
	   %}

	   DIGIT    [0-9]
	   ID	    [a-z][a-z0-9]*

	   %%

	   {DIGIT}+    {
		       printf( "An integer: %s (%d)\n",	yytext,
			       atoi( yytext ) );
		       }

	   {DIGIT}+"."{DIGIT}*	      {
		       printf( "A float: %s (%g)\n", yytext,
			       atof( yytext ) );
		       }

	   if|then|begin|end|procedure|function	       {
		       printf( "A keyword: %s\n", yytext );
		       }

	   {ID}	       printf( "An identifier: %s\n", yytext );

	   "+"|"-"|"*"|"/"   printf( "An operator: %s\n", yytext );

	   "{"[^}\n]*"}"     /*	eat up one-line	comments */

	   [ \t\n]+	     /*	eat up whitespace */

	   .	       printf( "Unrecognized character:	%s\n", yytext );

	   %%

	   main( argc, argv )
	   int argc;
	   char	**argv;
	       {
	       ++argv, --argc;	/* skip	over program name */
	       if ( argc > 0 )
		       yyin = fopen( argv[0], "r" );
	       else
		       yyin = stdin;

	       yylex();
	       }

       This is the beginnings of a simple scanner for a	language like  Pascal.
       It  identifies  different  types	 of  tokens and	reports	on what	it has
       seen.

       The details of this example will	be explained  in  the  following  sec-
       tions.

FORMAT OF THE INPUT FILE
       The  flex  input	 file  consists	of three sections, separated by	a line
       with just %% in it:

	   definitions
	   %%
	   rules
	   %%
	   user	code

       The definitions section contains	declarations of	 simple	 name  defini-
       tions  to simplify the scanner specification, and declarations of start
       conditions, which are explained in a later section.

       Name definitions	have the form:

	   name	definition

       The "name" is a word beginning with a letter  or	 an  underscore	 ('_')
       followed	by zero	or more	letters, digits, '_', or '-' (dash).  The def-
       inition is taken	to begin at the	first non-white-space  character  fol-
       lowing  the name	and continuing to the end of the line.	The definition
       can subsequently	be referred to using "{name}", which  will  expand  to
       "(definition)".	For example,

	   DIGIT    [0-9]
	   ID	    [a-z][a-z0-9]*

       defines	"DIGIT"	 to  be	 a  regular  expression	which matches a	single
       digit, and "ID" to be a regular expression which	matches	a letter  fol-
       lowed by	zero-or-more letters-or-digits.	 A subsequent reference	to

	   {DIGIT}+"."{DIGIT}*

       is identical to

	   ([0-9])+"."([0-9])*

       and  matches  one-or-more digits	followed by a '.' followed by zero-or-
       more digits.

       The rules section of the	flex input contains a series of	rules  of  the
       form:

	   pattern   action

       where  the  pattern must	be unindented and the action must begin	on the
       same line.

       See below for a further description of patterns and actions.

       Finally,	the user code section is simply	copied to  lex.yy.c  verbatim.
       It is used for companion	routines which call or are called by the scan-
       ner.  The presence of this section is optional; if it is	 missing,  the
       second %% in the	input file may be skipped, too.

       In  the	definitions  and  rules	 sections,  any	 indented text or text
       enclosed	in %{ and %} is	copied verbatim	to the output (with the	 %{}'s
       removed).  The %{}'s must appear	unindented on lines by themselves.

       In  the	rules  section,	 any indented or %{} text appearing before the
       first rule may be used to declare variables  which  are	local  to  the
       scanning	 routine and (after the	declarations) code which is to be exe-
       cuted whenever the scanning routine is entered.	Other indented or  %{}
       text in the rule	section	is still copied	to the output, but its meaning
       is not well-defined and it may well  cause  compile-time	 errors	 (this
       feature	is present for POSIX compliance; see below for other such fea-
       tures).

       In the definitions section (but not in the  rules  section),  an	 unin-
       dented comment (i.e., a line beginning with "/*") is also copied	verba-
       tim to the output up to the next	"*/".

PATTERNS
       The patterns in the input are written using an extended set of  regular
       expressions.  These are:

	   x	      match the	character 'x'
	   .	      any character (byte) except newline
	   [xyz]      a	"character class"; in this case, the pattern
			matches	either an 'x', a 'y', or a 'z'
	   [abj-oZ]   a	"character class" with a range in it; matches
			an 'a',	a 'b', any letter from 'j' through 'o',
			or a 'Z'
	   [^A-Z]     a	"negated character class", i.e., any character
			but those in the class.	 In this case, any
			character EXCEPT an uppercase letter.
	   [^A-Z\n]   any character EXCEPT an uppercase	letter or
			a newline
	   r*	      zero or more r's,	where r	is any regular expression
	   r+	      one or more r's
	   r?	      zero or one r's (that is,	"an optional r")
	   r{2,5}     anywhere from two	to five	r's
	   r{2,}      two or more r's
	   r{4}	      exactly 4	r's
	   {name}     the expansion of the "name" definition
		      (see above)
	   "[xyz]\"foo"
		      the literal string: [xyz]"foo
	   \X	      if X is an 'a', 'b', 'f',	'n', 'r', 't', or 'v',
			then the ANSI-C	interpretation of \x.
			Otherwise, a literal 'X' (used to escape
			operators such as '*')
	   \0	      a	NUL character (ASCII code 0)
	   \123	      the character with octal value 123
	   \x2a	      the character with hexadecimal value 2a
	   (r)	      match an r; parentheses are used to override
			precedence (see	below)

	   rs	      the regular expression r followed	by the
			regular	expression s; called "concatenation"

	   r|s	      either an	r or an	s

	   r/s	      an r but only if it is followed by an s.	The
			text matched by	s is included when determining
			whether	this rule is the "longest match",
			but is then returned to	the input before
			the action is executed.	 So the	action only
			sees the text matched by r.  This type
			of pattern is called trailing context".
			(There are some	combinations of	r/s that flex
			cannot match correctly;	see notes in the
			Deficiencies / Bugs section below regarding
			"dangerous trailing context".)
	   ^r	      an r, but	only at	the beginning of a line	(i.e.,
			which just starting to scan, or	right after a
			newline	has been scanned).
	   r$	      an r, but	only at	the end	of a line (i.e., just
			before a newline).  Equivalent to "r/\n".

		      Note that	flex's notion of "newline" is exactly
		      whatever the C compiler used to compile flex
		      interprets '\n' as; in particular, on some DOS
		      systems you must either filter out \r's in the
		      input yourself, or explicitly use	r/\r\n for "r$".

	   <s>r	      an r, but	only in	start condition	s (see
			below for discussion of	start conditions)
	   <s1,s2,s3>r
		      same, but	in any of start	conditions s1,
			s2, or s3
	   <*>r	      an r in any start	condition, even	an exclusive one.

	   <<EOF>>    an end-of-file
	   <s1,s2><<EOF>>
		      an end-of-file when in start condition s1	or s2

       Note that inside	of a character class, all regular expression operators
       lose their special meaning except escape	('\') and the character	 class
       operators, '-', ']', and, at the	beginning of the class,	'^'.

       The  regular  expressions  listed above are grouped according to	prece-
       dence, from highest precedence at the top  to  lowest  at  the  bottom.
       Those grouped together have equal precedence.  For example,

	   foo|bar*

       is the same as

	   (foo)|(ba(r*))

       since  the  '*'	operator has higher precedence than concatenation, and
       concatenation higher than alternation ('|').   This  pattern  therefore
       matches either the string "foo" or the string "ba" followed by zero-or-
       more r's.  To match "foo" or zero-or-more "bar"'s, use:

	   foo|(bar)*

       and to match zero-or-more "foo"'s-or-"bar"'s:

	   (foo|bar)*

       In addition to characters and ranges of characters,  character  classes
       can  also  contain  character class expressions.	 These are expressions
       enclosed	inside [: and :]  delimiters  (which  themselves  must	appear
       between	the  '['  and  ']'  of the character class; other elements may
       occur inside the	character class, too).	The valid expressions are:

	   [:alnum:] [:alpha:] [:blank:]
	   [:cntrl:] [:digit:] [:graph:]
	   [:lower:] [:print:] [:punct:]
	   [:space:] [:upper:] [:xdigit:]

       These expressions all designate a set of	characters equivalent  to  the
       corresponding standard C	isXXX function.	 For example, [:alnum:]	desig-
       nates those characters for which	isalnum() returns  true	 -  i.e.,  any
       alphabetic  or  numeric.	 Some systems don't provide isblank(), so flex
       defines [:blank:] as a blank or a tab.

       For example, the	following character classes are	all equivalent:

	   [[:alnum:]]
	   [[:alpha:][:digit:]]
	   [[:alpha:]0-9]
	   [a-zA-Z0-9]

       If your scanner is case-insensitive (the	-i flag), then	[:upper:]  and
       [:lower:] are equivalent	to [:alpha:].

       Some notes on patterns:

       -      A	 negated  character  class  such as the	example	"[^A-Z]" above
	      will match a  newline  unless  "\n"  (or	an  equivalent	escape
	      sequence)	 is  one  of  the characters explicitly	present	in the
	      negated character	class (e.g., "[^A-Z\n]").  This	is unlike  how
	      many  other  regular  expression	tools  treat negated character
	      classes, but unfortunately  the  inconsistency  is  historically
	      entrenched.   Matching  newlines means that a pattern like [^"]*
	      can match	the entire input unless	there's	another	quote  in  the
	      input.

       -      A	 rule  can  have at most one instance of trailing context (the
	      '/' operator or the '$' operator).  The  start  condition,  '^',
	      and "<<EOF>>" patterns can only occur at the beginning of	a pat-
	      tern, and, as well as with '/' and '$', cannot be	grouped	inside
	      parentheses.   A	'^' which does not occur at the	beginning of a
	      rule or a	'$' which does not occur at the	end of	a  rule	 loses
	      its special properties and is treated as a normal	character.

	      The following are	illegal:

		  foo/bar$
		  <sc1>foo<sc2>bar

	      Note that	the first of these, can	be written "foo/bar\n".

	      The  following will result in '$'	or '^' being treated as	a nor-
	      mal character:

		  foo|(bar$)
		  foo|^bar

	      If what's	wanted is a "foo" or a bar-followed-by-a-newline,  the
	      following	 could	be  used  (the special '|' action is explained
	      below):

		  foo	   |
		  bar$	   /* action goes here */

	      A	similar	trick will work	for matching a foo  or	a  bar-at-the-
	      beginning-of-a-line.

HOW THE	INPUT IS MATCHED
       When  the  generated  scanner is	run, it	analyzes its input looking for
       strings which match any of its patterns.	 If it	finds  more  than  one
       match,  it  takes  the one matching the most text (for trailing context
       rules, this includes the	length of the trailing part,  even  though  it
       will  then  be returned to the input).  If it finds two or more matches
       of the same length, the rule listed first in the	 flex  input  file  is
       chosen.

       Once  the  match	 is  determined,  the  text corresponding to the match
       (called the token) is made available in the  global  character  pointer
       yytext, and its length in the global integer yyleng.  The action	corre-
       sponding	to the matched pattern	is  then  executed  (a	more  detailed
       description  of	actions	 follows),  and	 then  the  remaining input is
       scanned for another match.

       If no match is found, then the default rule is executed:	the next char-
       acter  in  the  input  is considered matched and	copied to the standard
       output.	Thus, the simplest legal flex input is:

	   %%

       which generates a scanner that simply copies its	input  (one  character
       at a time) to its output.

       Note  that  yytext  can	be  defined in two different ways: either as a
       character pointer or as a character array.  You can control which defi-
       nition flex uses	by including one of the	special	directives %pointer or
       %array in the first (definitions) section  of  your  flex  input.   The
       default is %pointer, unless you use the -l lex compatibility option, in
       which case yytext will be an array.  The	advantage of using %pointer is
       substantially faster scanning and no buffer overflow when matching very
       large tokens (unless you	run out	of dynamic memory).  The  disadvantage
       is  that	 you are restricted in how your	actions	can modify yytext (see
       the next	section), and calls  to	 the  unput()  function	 destroys  the
       present	contents  of  yytext,  which  can  be  a  considerable porting
       headache	when moving between different lex versions.

       The advantage of	%array is that you can	then  modify  yytext  to  your
       heart's	content,  and  calls  to  unput()  do  not destroy yytext (see
       below).	Furthermore, existing lex  programs  sometimes	access	yytext
       externally using	declarations of	the form:
	   extern char yytext[];
       This  definition	 is erroneous when used	with %pointer, but correct for
       %array.

       %array defines yytext to	 be  an	 array	of  YYLMAX  characters,	 which
       defaults	 to  a	fairly large value.  You can change the	size by	simply
       #define'ing YYLMAX to a different value in the first  section  of  your
       flex input.  As mentioned above,	with %pointer yytext grows dynamically
       to accommodate large tokens.  While this	means  your  %pointer  scanner
       can  accommodate	 very  large tokens (such as matching entire blocks of
       comments), bear in mind that each time the scanner must	resize	yytext
       it  also	 must  rescan the entire token from the	beginning, so matching
       such tokens can prove slow.  yytext presently does not dynamically grow
       if  a  call  to	unput()	 results  in  too much text being pushed back;
       instead,	a run-time error results.

       Also note that you cannot use %array with C++ scanner classes (the  c++
       option; see below).

ACTIONS
       Each  pattern  in  a  rule has a	corresponding action, which can	be any
       arbitrary C statement.  The  pattern  ends  at  the  first  non-escaped
       whitespace  character; the remainder of the line	is its action.	If the
       action is empty,	then when the pattern is matched the  input  token  is
       simply discarded.  For example, here is the specification for a program
       which deletes all occurrences of	"zap me" from its input:

	   %%
	   "zap	me"

       (It will	copy all other characters in the input	to  the	 output	 since
       they will be matched by the default rule.)

       Here  is	 a program which compresses multiple blanks and	tabs down to a
       single blank, and throws	away whitespace	found at the end of a line:

	   %%
	   [ \t]+	 putchar( ' ' );
	   [ \t]+$	 /* ignore this	token */

       If the action contains a	'{', then the action spans till	the  balancing
       '}'  is	found,	and  the  action may cross multiple lines.  flex knows
       about C strings and comments and	won't be fooled	by braces found	within
       them,  but  also	 allows	actions	to begin with %{ and will consider the
       action to be all	the text up to the next	 %}  (regardless  of  ordinary
       braces inside the action).

       An  action consisting solely of a vertical bar ('|') means "same	as the
       action for the next rule."  See below for an illustration.

       Actions can include arbitrary C code, including	return	statements  to
       return  a  value	to whatever routine called yylex().  Each time yylex()
       is called it continues processing tokens	from where it  last  left  off
       until it	either reaches the end of the file or executes a return.

       Actions	are  free  to  modify yytext except for	lengthening it (adding
       characters to its end--these will overwrite  later  characters  in  the
       input  stream).	 This  however	does  not apply	when using %array (see
       above); in that case, yytext may	be freely modified in any way.

       Actions are free	to modify yyleng except	they should not	do so  if  the
       action also includes use	of yymore() (see below).

       There  are  a number of special directives which	can be included	within
       an action:

       -      ECHO copies yytext to the	scanner's output.

       -      BEGIN followed by	the name of a start condition places the scan-
	      ner in the corresponding start condition (see below).

       -      REJECT  directs  the  scanner to proceed on to the "second best"
	      rule which matched the input (or a prefix	of  the	 input).   The
	      rule is chosen as	described above	in "How	the Input is Matched",
	      and yytext and yyleng set	up appropriately.  It  may  either  be
	      one which	matched	as much	text as	the originally chosen rule but
	      came later in the	flex input file, or  one  which	 matched  less
	      text.   For  example, the	following will both count the words in
	      the input	and call the  routine  special()  whenever  "frob"  is
	      seen:

			  int word_count = 0;
		  %%

		  frob	      special(); REJECT;
		  [^ \t\n]+   ++word_count;

	      Without  the  REJECT,  any  "frob"'s  in	the input would	not be
	      counted as words,	since the scanner normally executes  only  one
	      action per token.	 Multiple REJECT's are allowed,	each one find-
	      ing the next best	choice to  the	currently  active  rule.   For
	      example,	when  the following scanner scans the token "abcd", it
	      will write "abcdabcaba" to the output:

		  %%
		  a	   |
		  ab	   |
		  abc	   |
		  abcd	   ECHO; REJECT;
		  .|\n	   /* eat up any unmatched character */

	      (The first three rules share the fourth's	action since they  use
	      the  special  '|'	 action.)   REJECT is a	particularly expensive
	      feature in terms of scanner performance; if it is	used in	any of
	      the  scanner's  actions  it  will	slow down all of the scanner's
	      matching.	 Furthermore, REJECT cannot be used with  the  -Cf  or
	      -CF options (see below).

	      Note  also  that	unlike	the other special actions, REJECT is a
	      branch; code immediately following it in the action will not  be
	      executed.

       -      yymore() tells the scanner that the next time it matches a rule,
	      the corresponding	token should  be  appended  onto  the  current
	      value  of	 yytext	 rather	than replacing it.  For	example, given
	      the input	"mega-kludge" the  following  will  write  "mega-mega-
	      kludge" to the output:

		  %%
		  mega-	   ECHO; yymore();
		  kludge   ECHO;

	      First  "mega-"  is  matched  and	echoed	to  the	 output.  Then
	      "kludge" is matched, but the previous "mega-" is	still  hanging
	      around  at  the beginning	of yytext so the ECHO for the "kludge"
	      rule will	actually write "mega-kludge".

       Two notes regarding use of yymore().  First, yymore()  depends  on  the
       value  of yyleng	correctly reflecting the size of the current token, so
       you must	not modify yyleng if you  are  using  yymore().	  Second,  the
       presence	 of  yymore()  in the scanner's	action entails a minor perfor-
       mance penalty in	the scanner's matching speed.

       -      yyless(n)	returns	all but	the first n characters of the  current
	      token  back  to  the  input stream, where	they will be rescanned
	      when the scanner looks for the next match.   yytext  and	yyleng
	      are  adjusted appropriately (e.g., yyleng	will now be equal to n
	      ).  For example, on the input "foobar" the following will	 write
	      out "foobarbar":

		  %%
		  foobar    ECHO; yyless(3);
		  [a-z]+    ECHO;

	      An  argument  of 0 to yyless will	cause the entire current input
	      string to	be scanned again.  Unless you've changed how the scan-
	      ner  will	subsequently process its input (using BEGIN, for exam-
	      ple), this will result in	an endless loop.

       Note that yyless	is a macro and can only	be  used  in  the  flex	 input
       file, not from other source files.

       -      unput(c)	puts  the  character c back onto the input stream.  It
	      will be the next character scanned.  The following  action  will
	      take  the	current	token and cause	it to be rescanned enclosed in
	      parentheses.

		  {
		  int i;
		  /* Copy yytext because unput() trashes yytext	*/
		  char *yycopy = strdup( yytext	);
		  unput( ')' );
		  for (	i = yyleng - 1;	i >= 0;	--i )
		      unput( yycopy[i] );
		  unput( '(' );
		  free(	yycopy );
		  }

	      Note that	since each unput() puts	the given  character  back  at
	      the  beginning of	the input stream, pushing back strings must be
	      done back-to-front.

       An important potential problem when using unput() is that  if  you  are
       using  %pointer	(the default), a call to unput() destroys the contents
       of yytext, starting with	its  rightmost	character  and	devouring  one
       character  to the left with each	call.  If you need the value of	yytext
       preserved after a call to unput() (as in	the above example),  you  must
       either  first  copy  it	elsewhere,  or build your scanner using	%array
       instead (see How	The Input Is Matched).

       Finally,	note that you cannot put back EOF to attempt to	mark the input
       stream with an end-of-file.

       -      input()  reads  the  next	 character from	the input stream.  For
	      example, the following is	one way	to eat up C comments:

		  %%
		  "/*"	      {
			      register int c;

			      for ( ; ;	)
				  {
				  while	( (c = input())	!= '*' &&
					  c != EOF )
				      ;	   /* eat up text of comment */

				  if ( c == '*'	)
				      {
				      while ( (c = input()) == '*' )
					  ;
				      if ( c ==	'/' )
					  break;    /* found the end */
				      }

				  if ( c == EOF	)
				      {
				      error( "EOF in comment" );
				      break;
				      }
				  }
			      }

	      (Note that if the	scanner	is compiled using C++, then input() is
	      instead referred to as yyinput(),	in order to avoid a name clash
	      with the C++ stream by the name of input.)

       -      YY_FLUSH_BUFFER flushes the scanner's internal  buffer  so  that
	      the  next	 time  the  scanner attempts to	match a	token, it will
	      first refill the buffer using YY_INPUT (see The Generated	 Scan-
	      ner,  below).  This action is a special case of the more general
	      yy_flush_buffer()	function, described below in the section  Mul-
	      tiple Input Buffers.

       -      yyterminate()  can  be  used in lieu of a	return statement in an
	      action.  It terminates the scanner and returns a 0 to the	 scan-
	      ner's  caller, indicating	"all done".  By	default, yyterminate()
	      is also called when an end-of-file  is  encountered.   It	 is  a
	      macro and	may be redefined.

THE GENERATED SCANNER
       The  output  of	flex is	the file lex.yy.c, which contains the scanning
       routine yylex(),	a number of tables used	by it for matching tokens, and
       a  number  of  auxiliary	 routines  and macros.	By default, yylex() is
       declared	as follows:

	   int yylex()
	       {
	       ... various definitions and the actions in here ...
	       }

       (If your	environment supports function prototypes, then it will be "int
       yylex(  void  )".)   This  definition  may  be  changed by defining the
       "YY_DECL" macro.	 For example, you could	use:

	   #define YY_DECL float lexscan( a, b ) float a, b;

       to give the scanning routine the	name lexscan, returning	a  float,  and
       taking two floats as arguments.	Note that if you give arguments	to the
       scanning	routine	using a	K&R-style/non-prototyped function declaration,
       you must	terminate the definition with a	semi-colon (;).

       Whenever	 yylex() is called, it scans tokens from the global input file
       yyin (which defaults to stdin).	It continues until it  either  reaches
       an  end-of-file	(at  which point it returns the	value 0) or one	of its
       actions executes	a return statement.

       If the scanner reaches an end-of-file, subsequent calls	are  undefined
       unless  either yyin is pointed at a new input file (in which case scan-
       ning continues from that	file), or yyrestart() is called.   yyrestart()
       takes  one  argument, a FILE * pointer (which can be nil, if you've set
       up YY_INPUT to scan from	a source other	than  yyin),  and  initializes
       yyin  for  scanning from	that file.  Essentially	there is no difference
       between just assigning yyin to a	new input file or using	yyrestart() to
       do so; the latter is available for compatibility	with previous versions
       of flex,	and because it can be used to switch input files in the	middle
       of  scanning.   It  can	also  be  used to throw	away the current input
       buffer, by calling it with an argument of yyin; but better  is  to  use
       YY_FLUSH_BUFFER	(see above).  Note that	yyrestart() does not reset the
       start condition to INITIAL (see Start Conditions, below).

       If yylex() stops	scanning due to	executing a return statement in	one of
       the  actions,  the  scanner may then be called again and	it will	resume
       scanning	where it left off.

       By default (and for purposes of efficiency), the	 scanner  uses	block-
       reads  rather  than  simple  getc() calls to read characters from yyin.
       The nature of how it gets its input can be controlled by	 defining  the
       YY_INPUT	     macro.	  YY_INPUT's	  calling      sequence	    is
       "YY_INPUT(buf,result,max_size)".	 Its action is to place	up to max_size
       characters  in  the character array buf and return in the integer vari-
       able result either the  number  of  characters  read  or	 the  constant
       YY_NULL	(0  on	Unix  systems)	to indicate EOF.  The default YY_INPUT
       reads from the global file-pointer "yyin".

       A sample	definition of YY_INPUT (in  the	 definitions  section  of  the
       input file):

	   %{
	   #define YY_INPUT(buf,result,max_size) \
	       { \
	       int c = getchar(); \
	       result =	(c == EOF) ? YY_NULL : (buf[0] = c, 1);	\
	       }
	   %}

       This definition will change the input processing	to occur one character
       at a time.

       When the	scanner	receives an end-of-file	indication from	 YY_INPUT,  it
       then  checks  the yywrap() function.  If	yywrap() returns false (zero),
       then it is assumed that the function has	gone ahead and set up yyin  to
       point  to  another  input  file,	and scanning continues.	 If it returns
       true (non-zero),	then  the  scanner  terminates,	 returning  0  to  its
       caller.	 Note  that  in	 either	 case,	the  start  condition  remains
       unchanged; it does not revert to	INITIAL.

       If you do not supply your own version of	yywrap(), then you must	either
       use  %option  noyywrap  (in  which  case	 the scanner behaves as	though
       yywrap()	returned 1), or	you must link with -ll to obtain  the  default
       version of the routine, which always returns 1.

       Three routines are available for	scanning from in-memory	buffers	rather
       than files: yy_scan_string(),  yy_scan_bytes(),	and  yy_scan_buffer().
       See the discussion of them below	in the section Multiple	Input Buffers.

       The scanner writes its ECHO output to the yyout global  (default,  std-
       out), which may be redefined by the user	simply by assigning it to some
       other FILE pointer.

START CONDITIONS
       flex provides a mechanism for conditionally activating rules.  Any rule
       whose  pattern  is  prefixed  with  "<sc>" will only be active when the
       scanner is in the start condition named "sc".  For example,

	   <STRING>[^"]*	{ /* eat up the	string body ...	*/
		       ...
		       }

       will be active only when	the scanner is in the  "STRING"	 start	condi-
       tion, and

	   <INITIAL,STRING,QUOTE>\.	   { /*	handle an escape ... */
		       ...
		       }

       will  be	 active	 only when the current start condition is either "INI-
       TIAL", "STRING",	or "QUOTE".

       Start conditions	are declared in	the definitions	(first)	section	of the
       input using unindented lines beginning with either %s or	%x followed by
       a list of names.	 The former declares inclusive start  conditions,  the
       latter  exclusive  start	 conditions.   A  start	condition is activated
       using the BEGIN action.	Until the next BEGIN action is executed, rules
       with  the  given	 start	condition  will	be active and rules with other
       start conditions	will be	inactive.  If the start	 condition  is	inclu-
       sive,  then  rules with no start	conditions at all will also be active.
       If it is	exclusive, then	only rules qualified with the start  condition
       will  be	active.	 A set of rules	contingent on the same exclusive start
       condition describe a scanner which is independent of any	of  the	 other
       rules  in  the flex input.  Because of this, exclusive start conditions
       make it easy to specify "mini-scanners"	which  scan  portions  of  the
       input  that are syntactically different from the	rest (e.g., comments).

       If the distinction between inclusive and	exclusive start	conditions  is
       still  a	little vague, here's a simple example illustrating the connec-
       tion between the	two.  The set of rules:

	   %s example
	   %%

	   <example>foo	  do_something();

	   bar		  something_else();

       is equivalent to

	   %x example
	   %%

	   <example>foo	  do_something();

	   <INITIAL,example>bar	   something_else();

       Without the <INITIAL,example> qualifier,	the bar	pattern	in the	second
       example	wouldn't be active (i.e., couldn't match) when in start	condi-
       tion example.  If we just used <example>	to qualify bar,	 though,  then
       it  would  only	be  active in example and not in INITIAL, while	in the
       first example it's active in both, because in  the  first  example  the
       example startion	condition is an	inclusive (%s) start condition.

       Also  note that the special start-condition specifier <*> matches every
       start condition.	 Thus, the above example could also have been written;

	   %x example
	   %%

	   <example>foo	  do_something();

	   <*>bar    something_else();

       The  default  rule  (to ECHO any	unmatched character) remains active in
       start conditions.  It is	equivalent to:

	   <*>.|\n     ECHO;

       BEGIN(0)	returns	to the original	state where only  the  rules  with  no
       start conditions	are active.  This state	can also be referred to	as the
       start-condition "INITIAL", so BEGIN(INITIAL) is equivalent to BEGIN(0).
       (The  parentheses  around the start condition name are not required but
       are considered good style.)

       BEGIN actions can also be given as indented code	at  the	 beginning  of
       the  rules  section.  For example, the following	will cause the scanner
       to enter	the "SPECIAL" start condition whenever yylex() is  called  and
       the global variable enter_special is true:

		   int enter_special;

	   %x SPECIAL
	   %%
		   if (	enter_special )
		       BEGIN(SPECIAL);

	   <SPECIAL>blahblahblah
	   ...more rules follow...

       To  illustrate  the  uses  of start conditions, here is a scanner which
       provides	two different interpretations of a string like "123.456".   By
       default	it  will  treat	 it  as	three tokens, the integer "123", a dot
       ('.'), and the integer "456".  But if the string	is preceded earlier in
       the  line  by  the  string "expect-floats" it will treat	it as a	single
       token, the floating-point number	123.456:

	   %{
	   #include <math.h>
	   %}
	   %s expect

	   %%
	   expect-floats	BEGIN(expect);

	   <expect>[0-9]+"."[0-9]+	{
		       printf( "found a	float, = %f\n",
			       atof( yytext ) );
		       }
	   <expect>\n		{
		       /* that's the end of the	line, so
			* we need another "expect-number"
			* before we'll recognize any more
			* numbers
			*/
		       BEGIN(INITIAL);
		       }

	   [0-9]+      {
		       printf( "found an integer, = %d\n",
			       atoi( yytext ) );
		       }

	   "."	       printf( "found a	dot\n" );

       Here is a scanner which recognizes  (and	 discards)  C  comments	 while
       maintaining a count of the current input	line.

	   %x comment
	   %%
		   int line_num	= 1;

	   "/*"		BEGIN(comment);

	   <comment>[^*\n]*	   /* eat anything that's not a	'*' */
	   <comment>"*"+[^*/\n]*   /* eat up '*'s not followed by '/'s */
	   <comment>\n		   ++line_num;
	   <comment>"*"+"/"	   BEGIN(INITIAL);

       This scanner goes to a bit of trouble to	match as much text as possible
       with each rule.	In general, when  attempting  to  write	 a  high-speed
       scanner	try to match as	much possible in each rule, as it's a big win.

       Note that start-conditions names	are really integer values and  can  be
       stored  as  such.   Thus,  the above could be extended in the following
       fashion:

	   %x comment foo
	   %%
		   int line_num	= 1;
		   int comment_caller;

	   "/*"		{
			comment_caller = INITIAL;
			BEGIN(comment);
			}

	   ...

	   <foo>"/*"	{
			comment_caller = foo;
			BEGIN(comment);
			}

	   <comment>[^*\n]*	   /* eat anything that's not a	'*' */
	   <comment>"*"+[^*/\n]*   /* eat up '*'s not followed by '/'s */
	   <comment>\n		   ++line_num;
	   <comment>"*"+"/"	   BEGIN(comment_caller);

       Furthermore, you	can access the current start condition using the inte-
       ger-valued  YY_START macro.  For	example, the above assignments to com-
       ment_caller could instead be written

	   comment_caller = YY_START;

       Flex provides YYSTATE as	an alias for YY_START (since  that  is	what's
       used by AT&T lex).

       Note  that  start conditions do not have	their own name-space; %s's and
       %x's declare names in the same fashion as #define's.

       Finally,	here's an example of how to match C-style quoted strings using
       exclusive  start	 conditions,  including	expanded escape	sequences (but
       not including checking for a string that's too long):

	   %x str

	   %%
		   char	string_buf[MAX_STR_CONST];
		   char	*string_buf_ptr;

	   \"	   string_buf_ptr = string_buf;	BEGIN(str);

	   <str>\"	  { /* saw closing quote - all done */
		   BEGIN(INITIAL);
		   *string_buf_ptr = '\0';
		   /* return string constant token type	and
		    * value to parser
		    */
		   }

	   <str>\n	  {
		   /* error - unterminated string constant */
		   /* generate error message */
		   }

	   <str>\\[0-7]{1,3} {
		   /* octal escape sequence */
		   int result;

		   (void) sscanf( yytext + 1, "%o", &result );

		   if (	result > 0xff )
			   /* error, constant is out-of-bounds */

		   *string_buf_ptr++ = result;
		   }

	   <str>\\[0-9]+ {
		   /* generate error - bad escape sequence; something
		    * like '\48' or '\0777777'
		    */
		   }

	   <str>\\n  *string_buf_ptr++ = '\n';
	   <str>\\t  *string_buf_ptr++ = '\t';
	   <str>\\r  *string_buf_ptr++ = '\r';
	   <str>\\b  *string_buf_ptr++ = '\b';
	   <str>\\f  *string_buf_ptr++ = '\f';

	   <str>\\(.|\n)  *string_buf_ptr++ = yytext[1];

	   <str>[^\\\n\"]+	  {
		   char	*yptr =	yytext;

		   while ( *yptr )
			   *string_buf_ptr++ = *yptr++;
		   }

       Often, such as in some of the examples above, you  wind	up  writing  a
       whole bunch of rules all	preceded by the	same start condition(s).  Flex
       makes this a little easier and cleaner by introducing a notion of start
       condition scope.	 A start condition scope is begun with:

	   <SCs>{

       where  SCs is a list of one or more start conditions.  Inside the start
       condition scope,	every rule automatically has the prefix	_SCs_  applied
       to it, until a '}' which	matches	the initial '{'.  So, for example,

	   <ESC>{
	       "\\n"   return '\n';
	       "\\r"   return '\r';
	       "\\f"   return '\f';
	       "\\0"   return '\0';
	   }

       is equivalent to:

	   <ESC>"\\n"  return '\n';
	   <ESC>"\\r"  return '\r';
	   <ESC>"\\f"  return '\f';
	   <ESC>"\\0"  return '\0';

       Start condition scopes may be nested.

       Three  routines	are  available for manipulating	stacks of start	condi-
       tions:

       void yy_push_state(int new_state)
	      pushes the current start condition onto the  top	of  the	 start
	      condition	stack and switches to new_state	as though you had used
	      BEGIN new_state (recall that  start  condition  names  are  also
	      integers).

       void yy_pop_state()
	      pops the top of the stack	and switches to	it via BEGIN.

       int yy_top_state()
	      returns  the  top	of the stack without altering the stack's con-
	      tents.

       The start condition stack grows dynamically and so has no built-in size
       limitation.  If memory is exhausted, program execution aborts.

       To  use	start  condition  stacks,  your	scanner	must include a %option
       stack directive (see Options below).

MULTIPLE INPUT BUFFERS
       Some scanners (such as those which  support  "include"  files)  require
       reading from several input streams.  As flex scanners do	a large	amount
       of buffering, one cannot	control	where the next input will be read from
       by  simply  writing  a YY_INPUT which is	sensitive to the scanning con-
       text.  YY_INPUT is only called when the scanner reaches the end of  its
       buffer,	which may be a long time after scanning	a statement such as an
       "include" which requires	switching the input source.

       To negotiate these sorts	of problems, flex  provides  a	mechanism  for
       creating	and switching between multiple input buffers.  An input	buffer
       is created by using:

	   YY_BUFFER_STATE yy_create_buffer( FILE *file, int size )

       which takes a FILE pointer and a	size and creates a  buffer  associated
       with  the  given	file and large enough to hold size characters (when in
       doubt, use YY_BUF_SIZE for the size).   It  returns  a  YY_BUFFER_STATE
       handle,	which  may  then be passed to other routines (see below).  The
       YY_BUFFER_STATE type is a pointer to an opaque  struct  yy_buffer_state
       structure,  so  you  may	safely initialize YY_BUFFER_STATE variables to
       ((YY_BUFFER_STATE) 0) if	you wish, and also refer to the	opaque	struc-
       ture  in	order to correctly declare input buffers in source files other
       than that of your scanner.  Note	that the FILE pointer in the  call  to
       yy_create_buffer	is only	used as	the value of yyin seen by YY_INPUT; if
       you redefine YY_INPUT so	it no longer uses yyin,	then  you  can	safely
       pass  a	nil FILE pointer to yy_create_buffer.  You select a particular
       buffer to scan from using:

	   void	yy_switch_to_buffer( YY_BUFFER_STATE new_buffer	)

       switches	the scanner's input buffer so subsequent tokens	will come from
       new_buffer.  Note that yy_switch_to_buffer() may	be used	by yywrap() to
       set things up for continued scanning, instead of	opening	a new file and
       pointing	yyin at	it.  Note also that switching input sources via	either
       yy_switch_to_buffer() or	yywrap() does not change the start  condition.

	   void	yy_delete_buffer( YY_BUFFER_STATE buffer )

       is  used	to reclaim the storage associated with a buffer.  ( buffer can
       be nil, in which	case the routine does nothing.)	 You  can  also	 clear
       the current contents of a buffer	using:

	   void	yy_flush_buffer( YY_BUFFER_STATE buffer	)

       This  function  discards	 the  buffer's	contents, so the next time the
       scanner attempts	to match a token from the buffer, it will  first  fill
       the buffer anew using YY_INPUT.

       yy_new_buffer()	is  an alias for yy_create_buffer(), provided for com-
       patibility with the C++ use of new and delete for creating and destroy-
       ing dynamic objects.

       Finally,	 the  YY_CURRENT_BUFFER	macro returns a	YY_BUFFER_STATE	handle
       to the current buffer.

       Here is an example of using these features for writing a	scanner	 which
       expands include files (the <<EOF>> feature is discussed below):

	   /* the "incl" state is used for picking up the name
	    * of an include file
	    */
	   %x incl

	   %{
	   #define MAX_INCLUDE_DEPTH 10
	   YY_BUFFER_STATE include_stack[MAX_INCLUDE_DEPTH];
	   int include_stack_ptr = 0;
	   %}

	   %%
	   include	       BEGIN(incl);

	   [a-z]+	       ECHO;
	   [^a-z\n]*\n?	       ECHO;

	   <incl>[ \t]*	     /*	eat the	whitespace */
	   <incl>[^ \t\n]+   { /* got the include file name */
		   if (	include_stack_ptr >= MAX_INCLUDE_DEPTH )
		       {
		       fprintf(	stderr,	"Includes nested too deeply" );
		       exit( 1 );
		       }

		   include_stack[include_stack_ptr++] =
		       YY_CURRENT_BUFFER;

		   yyin	= fopen( yytext, "r" );

		   if (	! yyin )
		       error( ... );

		   yy_switch_to_buffer(
		       yy_create_buffer( yyin, YY_BUF_SIZE ) );

		   BEGIN(INITIAL);
		   }

	   <<EOF>> {
		   if (	--include_stack_ptr < 0	)
		       {
		       yyterminate();
		       }

		   else
		       {
		       yy_delete_buffer( YY_CURRENT_BUFFER );
		       yy_switch_to_buffer(
			    include_stack[include_stack_ptr] );
		       }
		   }

       Three  routines are available for setting up input buffers for scanning
       in-memory strings instead of files.  All	of them	 create	 a  new	 input
       buffer	for   scanning	 the   string,	 and  return  a	 corresponding
       YY_BUFFER_STATE handle (which you should	delete with yy_delete_buffer()
       when  done  with	 it).	They  also  switch  to	the  new  buffer using
       yy_switch_to_buffer(), so the next call to yylex() will start  scanning
       the string.

       yy_scan_string(const char *str)
	      scans a NUL-terminated string.

       yy_scan_bytes(const char	*bytes,	int len)
	      scans  len bytes (including possibly NUL's) starting at location
	      bytes.

       Note that both of these functions create	and scan a copy	of the	string
       or  bytes.  (This may be	desirable, since yylex() modifies the contents
       of the buffer it	is scanning.)  You can avoid the copy by using:

       yy_scan_buffer(char *base, yy_size_t size)
	      which scans in place the buffer starting at base,	consisting  of
	      size   bytes,   the   last   two	 bytes	 of   which   must  be
	      YY_END_OF_BUFFER_CHAR (ASCII NUL).  These	last two bytes are not
	      scanned;	  thus,	  scanning   consists	of   base[0]   through
	      base[size-2], inclusive.

	      If you fail to set up base in  this  manner  (i.e.,  forget  the
	      final  two  YY_END_OF_BUFFER_CHAR	 bytes), then yy_scan_buffer()
	      returns a	nil pointer instead of creating	a new input buffer.

	      The type yy_size_t is an integral	type to	which you can cast  an
	      integer expression reflecting the	size of	the buffer.

END-OF-FILE RULES
       The special rule	"<<EOF>>" indicates actions which are to be taken when
       an end-of-file is encountered  and  yywrap()  returns  non-zero	(i.e.,
       indicates  no  further  files  to  process).  The action	must finish by
       doing one of four things:

       -      assigning	yyin to	a new input  file  (in	previous  versions  of
	      flex,  after  doing  the	assignment you had to call the special
	      action YY_NEW_FILE; this is no longer necessary);

       -      executing	a return statement;

       -      executing	the special yyterminate() action;

       -      or, switching to a new  buffer  using  yy_switch_to_buffer()  as
	      shown in the example above.

       <<EOF>>	rules  may  not	 be used with other patterns; they may only be
       qualified with a	list of	start conditions.  If an  unqualified  <<EOF>>
       rule  is	given, it applies to all start conditions which	do not already
       have <<EOF>> actions.  To specify an <<EOF>> rule for only the  initial
       start condition,	use

	   <INITIAL><<EOF>>

       These  rules are	useful for catching things like	unclosed comments.  An
       example:

	   %x quote
	   %%

	   ...other rules for dealing with quotes...

	   <quote><<EOF>>   {
		    error( "unterminated quote"	);
		    yyterminate();
		    }
	   <<EOF>>  {
		    if ( *++filelist )
			yyin = fopen( *filelist, "r" );
		    else
		       yyterminate();
		    }

MISCELLANEOUS MACROS
       The macro YY_USER_ACTION	can be defined to provide an action  which  is
       always  executed	 prior	to the matched rule's action.  For example, it
       could be	#define'd to call a routine to convert yytext  to  lower-case.
       When YY_USER_ACTION is invoked, the variable yy_act gives the number of
       the matched rule	(rules are numbered starting  with  1).	  Suppose  you
       want to profile how often each of your rules is matched.	 The following
       would do	the trick:

	   #define YY_USER_ACTION ++ctr[yy_act]

       where ctr is an array to	hold the counts	for the	different rules.  Note
       that  the macro YY_NUM_RULES gives the total number of rules (including
       the default rule, even if you use -s), so a correct declaration for ctr
       is:

	   int ctr[YY_NUM_RULES];

       The  macro  YY_USER_INIT	 may  be defined to provide an action which is
       always executed before the first	scan (and before the scanner's	inter-
       nal initializations are done).  For example, it could be	used to	call a
       routine to read in a data table or open a logging file.

       The macro yy_set_interactive(is_interactive) can	 be  used  to  control
       whether	the  current buffer is considered interactive.	An interactive
       buffer is processed more	slowly,	but must be used  when	the  scanner's
       input  source is	indeed interactive to avoid problems due to waiting to
       fill buffers (see the discussion	of the -I  flag	 below).   A  non-zero
       value  in  the macro invocation marks the buffer	as interactive,	a zero
       value as	non-interactive.   Note	 that  use  of	this  macro  overrides
       %option	interactive  ,	%option	 always-interactive  or	%option	never-
       interactive (see	Options	below).	 yy_set_interactive() must be  invoked
       prior to	beginning to scan the buffer that is (or is not) to be consid-
       ered interactive.

       The macro yy_set_bol(at_bol) can	be used	to control whether the current
       buffer's	scanning context for the next token match is done as though at
       the beginning of	 a  line.   A  non-zero	 macro	argument  makes	 rules
       anchored	with

       The  macro  YY_AT_BOL() returns true if the next	token scanned from the
       current buffer will have	'^' rules active, false	otherwise.

       In the generated	scanner, the actions are all  gathered	in  one	 large
       switch  statement and separated using YY_BREAK, which may be redefined.
       By default, it is simply	a "break", to separate each rule's action from
       the  following  rule's.	 Redefining  YY_BREAK allows, for example, C++
       users to	#define	YY_BREAK to do nothing (while being very careful  that
       every  rule ends	with a "break" or a "return"!) to avoid	suffering from
       unreachable statement warnings where because a rule's action ends  with
       "return", the YY_BREAK is inaccessible.

VALUES AVAILABLE TO THE	USER
       This section summarizes the various values available to the user	in the
       rule actions.

       -      char *yytext holds the text of the current  token.   It  may  be
	      modified but not lengthened (you cannot append characters	to the
	      end).

	      If the special directive %array appears in the first section  of
	      the  scanner  description,  then yytext is instead declared char
	      yytext[YYLMAX], where YYLMAX is a	macro definition that you  can
	      redefine	in  the	 first	section	 if you	don't like the default
	      value (generally 8KB).  Using %array results in somewhat	slower
	      scanners,	 but  the  value  of yytext becomes immune to calls to
	      input() and unput(), which potentially destroy  its  value  when
	      yytext  is  a  character	pointer.   The	opposite  of %array is
	      %pointer,	which is the default.

	      You cannot use %array when generating C++	scanner	 classes  (the
	      -+ flag).

       -      int yyleng holds the length of the current token.

       -      FILE *yyin is the	file which by default flex reads from.	It may
	      be redefined but doing  so  only	makes  sense  before  scanning
	      begins or	after an EOF has been encountered.  Changing it	in the
	      midst of	scanning  will	have  unexpected  results  since  flex
	      buffers  its input; use yyrestart() instead.  Once scanning ter-
	      minates because an end-of-file has been  seen,  you  can	assign
	      yyin  at	the  new input file and	then call the scanner again to
	      continue scanning.

       -      void yyrestart( FILE *new_file ) may be called to	point yyin  at
	      the  new input file.  The	switch-over to the new file is immedi-
	      ate (any previously buffered-up input is lost).  Note that call-
	      ing  yyrestart()	with  yyin as an argument thus throws away the
	      current input buffer and continues scanning the same input file.

       -      FILE  *yyout is the file to which	ECHO actions are done.	It can
	      be reassigned by the user.

       -      YY_CURRENT_BUFFER	returns	a YY_BUFFER_STATE handle to  the  cur-
	      rent buffer.

       -      YY_START	returns	 an integer value corresponding	to the current
	      start condition.	You can	subsequently use this value with BEGIN
	      to return	to that	start condition.

INTERFACING WITH YACC
       One  of the main	uses of	flex is	as a companion to the yacc parser-gen-
       erator.	yacc parsers expect to call a routine named  yylex()  to  find
       the  next  input	 token.	 The routine is	supposed to return the type of
       the next	token as well as putting any associated	value  in  the	global
       yylval.	 To use	flex with yacc,	one specifies the -d option to yacc to
       instruct	it to generate the file	y.tab.h	containing definitions of  all
       the %tokens appearing in	the yacc input.	 This file is then included in
       the flex	scanner.  For example, if one of the tokens  is	 "TOK_NUMBER",
       part of the scanner might look like:

	   %{
	   #include "y.tab.h"
	   %}

	   %%

	   [0-9]+	 yylval	= atoi(	yytext ); return TOK_NUMBER;

OPTIONS
       flex has	the following options:

       -b     Generate	backing-up  information	to lex.backup.	This is	a list
	      of scanner states	which require backing up and the input charac-
	      ters  on which they do so.  By adding rules one can remove back-
	      ing-up states.  If all backing-up	states are eliminated and  -Cf
	      or  -CF  is used,	the generated scanner will run faster (see the
	      -p flag).	 Only users who	wish to	squeeze	every last  cycle  out
	      of  their	 scanners need worry about this	option.	 (See the sec-
	      tion on Performance Considerations below.)

       -c     is a do-nothing, deprecated option included  for	POSIX  compli-
	      ance.

       -d     makes  the generated scanner run in debug	mode.  Whenever	a pat-
	      tern is recognized and  the  global  yy_flex_debug  is  non-zero
	      (which  is the default), the scanner will	write to stderr	a line
	      of the form:

		  --accepting rule at line 53 ("the matched text")

	      The line number refers to	the location of	the rule in  the  file
	      defining	the  scanner  (i.e.,  the  file	that was fed to	flex).
	      Messages are also	generated when the scanner backs  up,  accepts
	      the  default  rule,  reaches  the	 end  of  its input buffer (or
	      encounters a NUL;	at this	point, the two look the	same as	far as
	      the scanner's concerned),	or reaches an end-of-file.

       -f     specifies	 fast scanner.	No table compression is	done and stdio
	      is bypassed.  The	result is large	 but  fast.   This  option  is
	      equivalent to -Cfr (see below).

       -h     generates	 a "help" summary of flex's options to stdout and then
	      exits.  -?  and --help are synonyms for -h.

       -i     instructs	flex to	generate a case-insensitive scanner.  The case
	      of letters given in the flex input patterns will be ignored, and
	      tokens in	the input will be matched  regardless  of  case.   The
	      matched text given in yytext will	have the preserved case	(i.e.,
	      it will not be folded).

       -l     turns on maximum compatibility with the original AT&T lex	imple-
	      mentation.   Note	 that  this  does not mean full	compatibility.
	      Use of this option costs a considerable amount  of  performance,
	      and  it cannot be	used with the -+, -f, -F, -Cf, or -CF options.
	      For details on the compatibilities it provides, see the  section
	      "Incompatibilities  With Lex And POSIX" below.  This option also
	      results in the name YY_FLEX_LEX_COMPAT being  #define'd  in  the
	      generated	scanner.

       -n     is another do-nothing, deprecated	option included	only for POSIX
	      compliance.

       -p     generates	a performance report to	stderr.	 The  report  consists
	      of comments regarding features of	the flex input file which will
	      cause a serious loss of performance in  the  resulting  scanner.
	      If you give the flag twice, you will also	get comments regarding
	      features that lead to minor performance losses.

	      Note that	the use	of  REJECT,  %option  yylineno,	 and  variable
	      trailing	context	 (see  the  Deficiencies / Bugs	section	below)
	      entails a	substantial performance	penalty; use of	yymore(),  the
	      ^	 operator, and the -I flag entail minor	performance penalties.

       -s     causes the default rule (that unmatched scanner input is	echoed
	      to  stdout)  to  be suppressed.  If the scanner encounters input
	      that does	not match any of its rules, it aborts with  an	error.
	      This option is useful for	finding	holes in a scanner's rule set.

       -t     instructs	flex to	write the scanner  it  generates  to  standard
	      output instead of	lex.yy.c.

       -v     specifies	 that flex should write	to stderr a summary of statis-
	      tics regarding the scanner it generates.	Most of	the statistics
	      are  meaningless	to  the	 casual	 flex user, but	the first line
	      identifies the version of	flex (same as reported by -V), and the
	      next  line the flags used	when generating	the scanner, including
	      those that are on	by default.

       -w     suppresses warning messages.

       -B     instructs	flex to	generate a  batch  scanner,  the  opposite  of
	      interactive  scanners  generated by -I (see below).  In general,
	      you use -B when you are certain that your	scanner	will never  be
	      used  interactively,  and	you want to squeeze a little more per-
	      formance out of it.  If your goal	is instead to  squeeze	out  a
	      lot  more	 performance,  you  should   be	 using	the -Cf	or -CF
	      options (discussed below), which turn on -B  automatically  any-
	      way.

       -F     specifies	 that  the fast	scanner	table representation should be
	      used (and	stdio bypassed).  This representation is about as fast
	      as the full table	representation (-f), and for some sets of pat-
	      terns will be considerably smaller (and for others, larger).  In
	      general,	if  the	 pattern  set  contains	 both "keywords" and a
	      catch-all, "identifier" rule, such as in the set:

		  "case"    return TOK_CASE;
		  "switch"  return TOK_SWITCH;
		  ...
		  "default" return TOK_DEFAULT;
		  [a-z]+    return TOK_ID;

	      then you're better off using the full table representation.   If
	      only  the	 "identifier"  rule is present and you then use	a hash
	      table or some such to detect the	keywords,  you're  better  off
	      using -F.

	      This  option  is	equivalent  to -CFr (see below).  It cannot be
	      used with	-+.

       -I     instructs	flex to	generate an interactive	scanner.  An  interac-
	      tive  scanner  is	one that only looks ahead to decide what token
	      has been matched if it  absolutely  must.	  It  turns  out  that
	      always  looking  one  extra character ahead, even	if the scanner
	      has already seen enough text to disambiguate the current	token,
	      is  a  bit  faster  than only looking ahead when necessary.  But
	      scanners that always look	ahead give dreadful  interactive  per-
	      formance;	 for  example,	when a user types a newline, it	is not
	      recognized as a newline token until they	enter  another	token,
	      which often means	typing in another whole	line.

	      Flex  scanners  default to interactive unless you	use the	-Cf or
	      -CF table-compression options (see below).   That's  because  if
	      you're  looking  for high-performance you	should be using	one of
	      these options, so	if you didn't, flex assumes you'd rather trade
	      off  a  bit  of  run-time	 performance for intuitive interactive
	      behavior.	 Note also that	you cannot use -I in conjunction  with
	      -Cf or -CF.  Thus, this option is	not really needed; it is on by
	      default for all those cases in which it is allowed.

	      Note that	if isatty() returns false for the scanner input,  flex
	      will  revert  to batch mode, even	if -I was specified.  To force
	      interactive mode no matter what, use %option  always-interactive
	      (see Options below).

	      You  can	force a	scanner	to not be interactive by using -B (see
	      above).

       -L     instructs	flex not to generate #line directives.	 Without  this
	      option, flex peppers the generated scanner with #line directives
	      so error messages	in the actions will be correctly located  with
	      respect  to  either  the original	flex input file	(if the	errors
	      are due to code in the input file), or lex.yy.c (if  the	errors
	      are  flex's  fault -- you	should report these sorts of errors to
	      the email	address	given below).

       -T     makes flex run in	trace mode.  It	will generate a	 lot  of  mes-
	      sages  to	stderr concerning the form of the input	and the	resul-
	      tant non-deterministic and deterministic finite automata.	  This
	      option is	mostly for use in maintaining flex.

       -V     prints  the  version number to stdout and	exits.	--version is a
	      synonym for -V.

       -7     instructs	flex to	generate a 7-bit scanner, i.e.,	one which  can
	      only  recognize 7-bit characters in its input.  The advantage of
	      using -7 is that the scanner's tables can	be up to half the size
	      of  those	generated using	the -8 option (see below).  The	disad-
	      vantage is that such scanners often hang or crash	if their input
	      contains an 8-bit	character.

	      Note,  however,  that unless you generate	your scanner using the
	      -Cf or -CF table compression options, use	of -7 will save	only a
	      small  amount of table space, and	make your scanner considerably
	      less portable.  Flex's default behavior is to generate an	 8-bit
	      scanner  unless  you  use	 the  -Cf  or  -CF, in which case flex
	      defaults to generating  7-bit  scanners  unless  your  site  was
	      always  configured  to generate 8-bit scanners (as will often be
	      the case with non-USA sites).  You can tell whether flex	gener-
	      ated  a 7-bit or an 8-bit	scanner	by inspecting the flag summary
	      in the -v	output as described above.

	      Note that	if you use  -Cfe  or  -CFe  (those  table  compression
	      options,	but  also  using  equivalence classes as discussed see
	      below), flex still defaults  to  generating  an  8-bit  scanner,
	      since  usually  with these compression options full 8-bit	tables
	      are not much more	expensive than 7-bit tables.

       -8     instructs	flex to	generate an 8-bit scanner, i.e., one which can
	      recognize	 8-bit characters.  This flag is only needed for scan-
	      ners generated using -Cf or -CF, as otherwise flex  defaults  to
	      generating an 8-bit scanner anyway.

	      See  the	discussion of -7 above for flex's default behavior and
	      the tradeoffs between 7-bit and 8-bit scanners.

       -+     specifies	that you want flex to generate a  C++  scanner	class.
	      See the section on Generating C++	Scanners below for details.

       -C[aefFmr]
	      controls	the  degree  of	table compression and, more generally,
	      trade-offs between small scanners	and fast scanners.

	      -Ca ("align") instructs flex to trade off	larger tables  in  the
	      generated	scanner	for faster performance because the elements of
	      the tables are better aligned for	memory access and computation.
	      On  some RISC architectures, fetching and	manipulating longwords
	      is more efficient	than with smaller-sized	units such  as	short-
	      words.   This  option  can double	the size of the	tables used by
	      your scanner.

	      -Ce directs flex to construct equivalence	classes, i.e., sets of
	      characters which have identical lexical properties (for example,
	      if the only appearance of	digits in the flex  input  is  in  the
	      character	 class "[0-9]" then the	digits '0', '1', ..., '9' will
	      all be put in the	same equivalence class).  Equivalence  classes
	      usually  give dramatic reductions	in the final table/object file
	      sizes (typically a factor	of 2-5)	and are	pretty	cheap  perfor-
	      mance-wise (one array look-up per	character scanned).

	      -Cf specifies that the full scanner tables should	be generated -
	      flex should not compress the tables by taking advantages of sim-
	      ilar transition functions	for different states.

	      -CF  specifies  that  the	 alternate fast	scanner	representation
	      (described above under the -F flag) should be used.  This	option
	      cannot be	used with -+.

	      -Cm  directs  flex  to construct meta-equivalence	classes, which
	      are sets of equivalence classes (or characters,  if  equivalence
	      classes  are  not	 being	used) that are commonly	used together.
	      Meta-equivalence classes are often a big	win  when  using  com-
	      pressed tables, but they have a moderate performance impact (one
	      or two "if" tests	and one	array look-up per character  scanned).

	      -Cr  causes  the generated scanner to bypass use of the standard
	      I/O library (stdio) for input.  Instead of  calling  fread()  or
	      getc(),  the  scanner will use the read()	system call, resulting
	      in a performance gain which varies from system to	system,	but in
	      general  is probably negligible unless you are also using	-Cf or
	      -CF.  Using -Cr can cause	strange	behavior if, for example,  you
	      read from	yyin using stdio prior to calling the scanner (because
	      the scanner will miss whatever text your previous	reads left  in
	      the stdio	input buffer).

	      -Cr  has	no  effect  if	you define YY_INPUT (see The Generated
	      Scanner above).

	      A	lone -C	specifies that the scanner tables should be compressed
	      but  neither  equivalence	 classes  nor meta-equivalence classes
	      should be	used.

	      The options -Cf or -CF and -Cm do	 not  make  sense  together  -
	      there  is	no opportunity for meta-equivalence classes if the ta-
	      ble is not being	compressed.   Otherwise	 the  options  may  be
	      freely mixed, and	are cumulative.

	      The  default  setting  is	-Cem, which specifies that flex	should
	      generate equivalence classes and meta-equivalence	classes.  This
	      setting  provides	 the highest degree of table compression.  You
	      can trade	off faster-executing scanners at the  cost  of	larger
	      tables with the following	generally being	true:

		  slowest & smallest
			-Cem
			-Cm
			-Ce
			-C
			-C{f,F}e
			-C{f,F}
			-C{f,F}a
		  fastest & largest

	      Note  that  scanners with	the smallest tables are	usually	gener-
	      ated and compiled	the quickest, so during	development  you  will
	      usually want to use the default, maximal compression.

	      -Cfe  is often a good compromise between speed and size for pro-
	      duction scanners.

       -ooutput
	      directs flex to write the	scanner	to the file output instead  of
	      lex.yy.c.	  If you combine -o with the -t	option,	then the scan-
	      ner is written to	stdout but its #line directives	 (see  the  -L
	      option above) refer to the file output.

       -Pprefix
	      changes the default yy prefix used by flex for all globally-vis-
	      ible variable and	function names	to  instead  be	 prefix.   For
	      example,	-Pfoo  changes the name	of yytext to footext.  It also
	      changes the name of the default output  file  from  lex.yy.c  to
	      lex.foo.c.  Here are all of the names affected:

		  yy_create_buffer
		  yy_delete_buffer
		  yy_flex_debug
		  yy_init_buffer
		  yy_flush_buffer
		  yy_load_buffer_state
		  yy_switch_to_buffer
		  yyin
		  yyleng
		  yylex
		  yylineno
		  yyout
		  yyrestart
		  yytext
		  yywrap

	      (If   you	 are  using  a	C++  scanner,  then  only  yywrap  and
	      yyFlexLexer are affected.)  Within your scanner itself, you  can
	      still  refer  to the global variables and	functions using	either
	      version of their name; but externally, they  have	 the  modified
	      name.

	      This option lets you easily link together	multiple flex programs
	      into the same executable.	 Note, though, that using this	option
	      also  renames  yywrap(), so you now must either provide your own
	      (appropriately-named) version of the routine for	your  scanner,
	      or  use %option noyywrap,	as linking with	-ll no longer provides
	      one for you by default.

       -Sskeleton_file
	      overrides	the default skeleton file from which  flex  constructs
	      its  scanners.   You'll  never  need  this option	unless you are
	      doing flex maintenance or	development.

       flex also provides a mechanism for controlling options within the scan-
       ner specification itself, rather	than from the flex command-line.  This
       is done by including %option directives in the  first  section  of  the
       scanner	specification.	You can	specify	multiple options with a	single
       %option directive, and multiple directives in the first section of your
       flex input file.

       Most options are	given simply as	names, optionally preceded by the word
       "no" (with no intervening whitespace) to	negate their meaning.  A  num-
       ber are equivalent to flex flags	or their negation:

	   7bit		   -7 option
	   8bit		   -8 option
	   align	   -Ca option
	   backup	   -b option
	   batch	   -B option
	   c++		   -+ option

	   caseful or
	   case-sensitive  opposite of -i (default)

	   case-insensitive or
	   caseless	   -i option

	   debug	   -d option
	   default	   opposite of -s option
	   ecs		   -Ce option
	   fast		   -F option
	   full		   -f option
	   interactive	   -I option
	   lex-compat	   -l option
	   meta-ecs	   -Cm option
	   perf-report	   -p option
	   read		   -Cr option
	   stdout	   -t option
	   verbose	   -v option
	   warn		   opposite of -w option
			   (use	"%option nowarn" for -w)

	   array	   equivalent to "%array"
	   pointer	   equivalent to "%pointer" (default)

       Some %option's provide features otherwise not available:

       always-interactive
	      instructs	 flex to generate a scanner which always considers its
	      input "interactive".  Normally, on each new input	file the scan-
	      ner  calls isatty() in an	attempt	to determine whether the scan-
	      ner's input source is interactive	and  thus  should  be  read  a
	      character	at a time.  When this option is	used, however, then no
	      such call	is made.

       main   directs flex to provide a	default	main() program for  the	 scan-
	      ner,  which  simply calls	yylex().  This option implies noyywrap
	      (see below).

       never-interactive
	      instructs	flex to	generate a scanner which never	considers  its
	      input  "interactive" (again, no call made	to isatty()).  This is
	      the opposite of always-interactive.

       stack  enables the use of start condition stacks	(see Start  Conditions
	      above).

       stdinit
	      if  set  (i.e.,  %option	stdinit) initializes yyin and yyout to
	      stdin and	stdout,	instead	of the default of nil.	Some  existing
	      lex programs depend on this behavior, even though	it is not com-
	      pliant with ANSI C, which	does not require stdin and  stdout  to
	      be compile-time constant.

       yylineno
	      directs  flex to generate	a scanner that maintains the number of
	      the current line read from its  input  in	 the  global  variable
	      yylineno.	 This option is	implied	by %option lex-compat.

       yywrap if  unset	 (i.e.,	 %option noyywrap), makes the scanner not call
	      yywrap() upon an end-of-file, but	simply assume that  there  are
	      no  more files to	scan (until the	user points yyin at a new file
	      and calls	yylex()	again).

       flex scans your rule actions to determine whether you use the REJECT or
       yymore()	 features.   The  reject  and  yymore options are available to
       override	its decision as	to whether you use the options,	either by set-
       ting  them  (e.g.,  %option  reject)  to	indicate the feature is	indeed
       used, or	unsetting them to indicate it  actually	 is  not  used	(e.g.,
       %option noyymore).

       Three options take string-delimited values, offset with '=':

	   %option outfile="ABC"

       is equivalent to	-oABC, and

	   %option prefix="XYZ"

       is equivalent to	-PXYZ.	Finally,

	   %option yyclass="foo"

       only  applies  when  generating a C++ scanner ( -+ option).  It informs
       flex that you have derived foo as a subclass of	yyFlexLexer,  so  flex
       will  place your	actions	in the member function foo::yylex() instead of
       yyFlexLexer::yylex().  It also generates	a yyFlexLexer::yylex()	member
       function	 that  emits a run-time	error (by invoking yyFlexLexer::Lexer-
       Error())	if called.  See	Generating C++ Scanners, below,	for additional
       information.

       A number	of options are available for lint purists who want to suppress
       the appearance of unneeded routines in the generated scanner.  Each  of
       the following, if unset (e.g., %option nounput ), results in the	corre-
       sponding	routine	not appearing in the generated scanner:

	   input, unput
	   yy_push_state, yy_pop_state,	yy_top_state
	   yy_scan_buffer, yy_scan_bytes, yy_scan_string

       (though yy_push_state() and friends won't appear	anyway unless you  use
       %option stack).

PERFORMANCE CONSIDERATIONS
       The main	design goal of flex is that it generate	high-performance scan-
       ners.  It has been optimized for	dealing	well with large	sets of	rules.
       Aside  from  the	 effects  on scanner speed of the table	compression -C
       options outlined	above, there are a  number  of	options/actions	 which
       degrade performance.  These are,	from most expensive to least:

	   REJECT
	   %option yylineno
	   arbitrary trailing context

	   pattern sets	that require backing up
	   %array
	   %option interactive
	   %option always-interactive

	   '^' beginning-of-line operator
	   yymore()

       with  the  first	three all being	quite expensive	and the	last two being
       quite cheap.  Note also that unput() is implemented as a	 routine  call
       that  potentially  does quite a bit of work, while yyless() is a	quite-
       cheap macro; so if just putting back some excess	text you scanned,  use
       yyless().

       REJECT  should  be  avoided at all costs	when performance is important.
       It is a particularly expensive option.

       Getting rid of backing up is messy and often may	be an enormous	amount
       of  work	 for a complicated scanner.  In	principal, one begins by using
       the -b flag to generate a lex.backup file.  For example,	on the input

	   %%
	   foo	      return TOK_KEYWORD;
	   foobar     return TOK_KEYWORD;

       the file	looks like:

	   State #6 is non-accepting -
	    associated rule line numbers:
		  2	  3
	    out-transitions: [ o ]
	    jam-transitions: EOF [ \001-n  p-\177 ]

	   State #8 is non-accepting -
	    associated rule line numbers:
		  3
	    out-transitions: [ a ]
	    jam-transitions: EOF [ \001-`  b-\177 ]

	   State #9 is non-accepting -
	    associated rule line numbers:
		  3
	    out-transitions: [ r ]
	    jam-transitions: EOF [ \001-q  s-\177 ]

	   Compressed tables always back up.

       The first few lines tell	us that	there's	a scanner state	 in  which  it
       can  make  a  transition	 on an 'o' but not on any other	character, and
       that in that state the currently	scanned	text does not match any	 rule.
       The  state occurs when trying to	match the rules	found at lines 2 and 3
       in the input file.  If the scanner is in	 that  state  and  then	 reads
       something  other	 than  an  'o',	it will	have to	back up	to find	a rule
       which is	matched.  With a bit of	headscratching one can see  that  this
       must  be	 the  state it's in when it has	seen "fo".  When this has hap-
       pened, if anything other	than another 'o' is  seen,  the	 scanner  will
       have to back up to simply match the 'f' (by the default rule).

       The  comment regarding State #8 indicates there's a problem when	"foob"
       has been	scanned.  Indeed, on any character  other  than	 an  'a',  the
       scanner	will  have to back up to accept	"foo".	Similarly, the comment
       for State #9 concerns when "fooba" has been scanned and an 'r' does not
       follow.

       The  final  comment  reminds  us	that there's no	point going to all the
       trouble of removing backing up from the rules unless we're using	-Cf or
       -CF,  since  there's no performance gain	doing so with compressed scan-
       ners.

       The way to remove the backing up	is to add "error" rules:

	   %%
	   foo	       return TOK_KEYWORD;
	   foobar      return TOK_KEYWORD;

	   fooba       |
	   foob	       |
	   fo	       {
		       /* false	alarm, not really a keyword */
		       return TOK_ID;
		       }

       Eliminating backing up among a list of keywords can also	be done	 using
       a "catch-all" rule:

	   %%
	   foo	       return TOK_KEYWORD;
	   foobar      return TOK_KEYWORD;

	   [a-z]+      return TOK_ID;

       This is usually the best	solution when appropriate.

       Backing	up  messages tend to cascade.  With a complicated set of rules
       it's not	uncommon to get	hundreds of messages.	If  one	 can  decipher
       them,  though, it often only takes a dozen or so	rules to eliminate the
       backing up (though it's easy to make a mistake and have an  error  rule
       accidentally  match a valid token.  A possible future flex feature will
       be to automatically add rules to	eliminate backing up).

       It's important to keep in mind that you gain the	benefits of  eliminat-
       ing  backing  up	 only  if  you eliminate every instance	of backing up.
       Leaving just one	means you gain nothing.

       Variable	trailing context (where	both the leading and trailing parts do
       not  have  a  fixed length) entails almost the same performance loss as
       REJECT (i.e., substantial).  So when possible a rule like:

	   %%
	   mouse|rat/(cat|dog)	 run();

       is better written:

	   %%
	   mouse/cat|dog	 run();
	   rat/cat|dog		 run();

       or as

	   %%
	   mouse|rat/cat	 run();
	   mouse|rat/dog	 run();

       Note that here the special '|' action does not provide any savings, and
       can even	make things worse (see Deficiencies / Bugs below).

       Another	area  where the	user can increase a scanner's performance (and
       one that's easier to implement) arises from the fact  that  the	longer
       the  tokens  matched, the faster	the scanner will run.  This is because
       with long tokens	the processing of most input characters	takes place in
       the  (short) inner scanning loop, and does not often have to go through
       the additional work of  setting	up  the	 scanning  environment	(e.g.,
       yytext) for the action.	Recall the scanner for C comments:

	   %x comment
	   %%
		   int line_num	= 1;

	   "/*"		BEGIN(comment);

	   <comment>[^*\n]*
	   <comment>"*"+[^*/\n]*
	   <comment>\n		   ++line_num;
	   <comment>"*"+"/"	   BEGIN(INITIAL);

       This could be sped up by	writing	it as:

	   %x comment
	   %%
		   int line_num	= 1;

	   "/*"		BEGIN(comment);

	   <comment>[^*\n]*
	   <comment>[^*\n]*\n	   ++line_num;
	   <comment>"*"+[^*/\n]*
	   <comment>"*"+[^*/\n]*\n ++line_num;
	   <comment>"*"+"/"	   BEGIN(INITIAL);

       Now instead of each newline requiring the processing of another action,
       recognizing the newlines	is "distributed" over the other	rules to  keep
       the  matched text as long as possible.  Note that adding	rules does not
       slow down the scanner!  The speed of the	scanner	is independent of  the
       number of rules or (modulo the considerations given at the beginning of
       this section) how complicated the rules are with	 regard	 to  operators
       such as '*' and '|'.

       A  final	 example  in  speeding	up a scanner: suppose you want to scan
       through a file containing identifiers and keywords, one	per  line  and
       with no other extraneous	characters, and	recognize all the keywords.  A
       natural first approach is:

	   %%
	   asm	    |
	   auto	    |
	   break    |
	   ... etc ...
	   volatile |
	   while    /* it's a keyword */

	   .|\n	    /* it's not	a keyword */

       To eliminate the	back-tracking, introduce a catch-all rule:

	   %%
	   asm	    |
	   auto	    |
	   break    |
	   ... etc ...
	   volatile |
	   while    /* it's a keyword */

	   [a-z]+   |
	   .|\n	    /* it's not	a keyword */

       Now, if it's guaranteed that there's exactly one	word per line, then we
       can  reduce  the	 total	number	of matches by a	half by	merging	in the
       recognition of newlines with that of the	other tokens:

	   %%
	   asm\n    |
	   auto\n   |
	   break\n  |
	   ... etc ...
	   volatile\n |
	   while\n  /* it's a keyword */

	   [a-z]+\n |
	   .|\n	    /* it's not	a keyword */

       One has to be careful here, as we have now reintroduced backing up into
       the scanner.  In	particular, while we know that there will never	be any
       characters in the input stream other than  letters  or  newlines,  flex
       can't figure this out, and it will plan for possibly needing to back up
       when it has scanned a token like	"auto" and then	the next character  is
       something  other	 than a	newline	or a letter.  Previously it would then
       just match the "auto" rule and be done, but now it has no "auto"	 rule,
       only  a	"auto\n" rule.	To eliminate the possibility of	backing	up, we
       could either duplicate all rules	but without final newlines, or,	 since
       we never	expect to encounter such an input and therefore	don't how it's
       classified, we can introduce one	more catch-all rule,  this  one	 which
       doesn't include a newline:

	   %%
	   asm\n    |
	   auto\n   |
	   break\n  |
	   ... etc ...
	   volatile\n |
	   while\n  /* it's a keyword */

	   [a-z]+\n |
	   [a-z]+   |
	   .|\n	    /* it's not	a keyword */

       Compiled	 with -Cf, this	is about as fast as one	can get	a flex scanner
       to go for this particular problem.

       A final note: flex is slow when matching	 NUL's,	 particularly  when  a
       token  contains	multiple  NUL's.  It's best to write rules which match
       short amounts of	text if	it's anticipated  that	the  text  will	 often
       include NUL's.

       Another	final  note  regarding	performance: as	mentioned above	in the
       section How the Input is	Matched, dynamically resizing yytext to	accom-
       modate huge tokens is a slow process because it presently requires that
       the (huge) token	be rescanned from the beginning.  Thus if  performance
       is  vital,  you	should attempt to match	"large"	quantities of text but
       not "huge" quantities, where the	cutoff between the two is at about  8K
       characters/token.

GENERATING C++ SCANNERS
       flex provides two different ways	to generate scanners for use with C++.
       The first way is	to simply compile a scanner generated by flex using  a
       C++  compiler  instead  of  a C compiler.  You should not encounter any
       compilations errors (please report any you find to  the	email  address
       given  in the Author section below).  You can then use C++ code in your
       rule actions instead of C code.	Note that the default input source for
       your  scanner remains yyin, and default echoing is still	done to	yyout.
       Both of these remain FILE * variables and not C++ streams.

       You can also use	flex to	generate a C++ scanner	class,	using  the  -+
       option  (or,  equivalently, %option c++), which is automatically	speci-
       fied if the name	of the flex executable ends in a '+', such as  flex++.
       When  using this	option,	flex defaults to generating the	scanner	to the
       file lex.yy.cc instead of lex.yy.c.  The	generated scanner includes the
       header  file  FlexLexer.h,  which  defines  the	interface  to  two C++
       classes.

       The first class,	FlexLexer, provides an abstract	 base  class  defining
       the  general scanner class interface.  It provides the following	member
       functions:

       const char* YYText()
	      returns the text of the most recently matched token, the equiva-
	      lent of yytext.

       int YYLeng()
	      returns  the  length  of	the  most  recently matched token, the
	      equivalent of yyleng.

       int lineno() const
	      returns the current input	line number (see %option yylineno), or
	      1	if %option yylineno was	not used.

       void set_debug( int flag	)
	      sets the debugging flag for the scanner, equivalent to assigning
	      to yy_flex_debug (see the	Options	section	above).	 Note that you
	      must  build the scanner using %option debug to include debugging
	      information in it.

       int debug() const
	      returns the current setting of the debugging flag.

       Also provided are member	functions equivalent to	yy_switch_to_buffer(),
       yy_create_buffer()  (though  the	 first	argument is an istream*	object
       pointer and not a FILE*),  yy_flush_buffer(),  yy_delete_buffer(),  and
       yyrestart() (again, the first argument is a istream* object pointer).

       The  second  class  defined  in	FlexLexer.h  is	 yyFlexLexer, which is
       derived from FlexLexer.	It defines  the	 following  additional	member
       functions:

       yyFlexLexer( istream* arg_yyin =	0, ostream* arg_yyout =	0 )
	      constructs  a  yyFlexLexer  object  using	 the given streams for
	      input and	output.	 If not	specified, the streams default to  cin
	      and cout,	respectively.

       virtual int yylex()
	      performs	the  same role is yylex() does for ordinary flex scan-
	      ners: it scans the  input	 stream,  consuming  tokens,  until  a
	      rule's  action returns a value.  If you derive a subclass	S from
	      yyFlexLexer and want to access the member	 functions  and	 vari-
	      ables  of	 S  inside  yylex(),  then  you	 need  to  use %option
	      yyclass="S" to inform flex that you will be using	that  subclass
	      instead  of  yyFlexLexer.	  In this case,	rather than generating
	      yyFlexLexer::yylex(), flex generates S::yylex() (and also	gener-
	      ates a dummy yyFlexLexer::yylex()	that calls yyFlexLexer::Lexer-
	      Error() if called).

       virtual void switch_streams(istream* new_in = 0,
	      ostream* new_out = 0) reassigns yyin to new_in (if non-nil)  and
	      yyout  to	new_out	(ditto), deleting the previous input buffer if
	      yyin is reassigned.

       int yylex( istream* new_in, ostream* new_out = 0	)
	      first switches the input	streams	 via  switch_streams(  new_in,
	      new_out )	and then returns the value of yylex().

       In  addition, yyFlexLexer defines the following protected virtual func-
       tions which you can redefine in derived classes to tailor the scanner:

       virtual int LexerInput( char* buf, int max_size )
	      reads up to max_size characters into buf and returns the	number
	      of  characters read.  To indicate	end-of-input, return 0 charac-
	      ters.  Note that "interactive"  scanners	(see  the  -B  and  -I
	      flags)  define  the  macro YY_INTERACTIVE.  If you redefine Lex-
	      erInput()	and  need  to  take  different	actions	 depending  on
	      whether  or  not	the  scanner  might be scanning	an interactive
	      input source, you	can test for the presence  of  this  name  via
	      #ifdef.

       virtual void LexerOutput( const char* buf, int size )
	      writes  out  size	 characters  from the buffer buf, which, while
	      NUL-terminated, may also contain "internal" NUL's	if  the	 scan-
	      ner's rules can match text with NUL's in them.

       virtual void LexerError(	const char* msg	)
	      reports  a  fatal	 error	message.   The default version of this
	      function writes the message to the stream	cerr and exits.

       Note that a yyFlexLexer object  contains	 its  entire  scanning	state.
       Thus  you  can  use such	objects	to create reentrant scanners.  You can
       instantiate multiple instances of the same yyFlexLexer class,  and  you
       can also	combine	multiple C++ scanner classes together in the same pro-
       gram using the -P option	discussed above.

       Finally,	note that the %array feature is	not available to  C++  scanner
       classes;	you must use %pointer (the default).

       Here is an example of a simple C++ scanner:

	       // An example of	using the flex C++ scanner class.

	   %{
	   int mylineno	= 0;
	   %}

	   string  \"[^\n"]+\"

	   ws	   [ \t]+

	   alpha   [A-Za-z]
	   dig	   [0-9]
	   name	   ({alpha}|{dig}|\$)({alpha}|{dig}|[_.\-/$])*
	   num1	   [-+]?{dig}+\.?([eE][-+]?{dig}+)?
	   num2	   [-+]?{dig}*\.{dig}+([eE][-+]?{dig}+)?
	   number  {num1}|{num2}

	   %%

	   {ws}	   /* skip blanks and tabs */

	   "/*"	   {
		   int c;

		   while((c = yyinput()) != 0)
		       {
		       if(c == '\n')
			   ++mylineno;

		       else if(c == '*')
			   {
			   if((c = yyinput()) == '/')
			       break;
			   else
			       unput(c);
			   }
		       }
		   }

	   {number}  cout << "number " << YYText() << '\n';

	   \n	     mylineno++;

	   {name}    cout << "name " <<	YYText() << '\n';

	   {string}  cout << "string " << YYText() << '\n';

	   %%

	   int main( int /* argc */, char** /* argv */ )
	       {
	       FlexLexer* lexer	= new yyFlexLexer;
	       while(lexer->yylex() != 0)
		   ;
	       return 0;
	       }
       If  you	want to	create multiple	(different) lexer classes, you use the
       -P flag (or the prefix= option) to  rename  each	 yyFlexLexer  to  some
       other  xxFlexLexer.   You  then can include <FlexLexer.h> in your other
       sources once per	lexer class, first renaming yyFlexLexer	as follows:

	   #undef yyFlexLexer
	   #define yyFlexLexer xxFlexLexer
	   #include <FlexLexer.h>

	   #undef yyFlexLexer
	   #define yyFlexLexer zzFlexLexer
	   #include <FlexLexer.h>

       if, for example,	you used %option prefix="xx" for one of	your  scanners
       and %option prefix="zz" for the other.

       IMPORTANT:  the	present	form of	the scanning class is experimental and
       may change considerably between major releases.

INCOMPATIBILITIES WITH LEX AND POSIX
       flex is a rewrite of the	AT&T Unix lex tool (the	two implementations do
       not  share  any	code, though), with some extensions and	incompatibili-
       ties, both of which are of concern to those who wish to write  scanners
       acceptable  to either implementation.  Flex is fully compliant with the
       POSIX lex specification,	except that when using %pointer	(the default),
       a  call to unput() destroys the contents	of yytext, which is counter to
       the POSIX specification.

       In this section we discuss all of the known  areas  of  incompatibility
       between flex, AT&T lex, and the POSIX specification.

       flex's  -l option turns on maximum compatibility	with the original AT&T
       lex implementation, at the cost of a major loss in the generated	 scan-
       ner's  performance.  We note below which	incompatibilities can be over-
       come using the -l option.

       flex is fully compatible	with lex with the following exceptions:

       -      The undocumented lex scanner internal variable yylineno  is  not
	      supported	unless -l or %option yylineno is used.

	      yylineno should be maintained on a per-buffer basis, rather than
	      a	per-scanner (single global variable) basis.

	      yylineno is not part of the POSIX	specification.

       -      The input() routine is not redefinable, though it	may be	called
	      to  read	characters  following  whatever	 has been matched by a
	      rule.  If	input()	encounters an end-of-file the normal  yywrap()
	      processing  is  done.   A	 ``real''  end-of-file	is returned by
	      input() as EOF.

	      Input is instead controlled by defining the YY_INPUT macro.

	      The flex restriction that	input()	 cannot	 be  redefined	is  in
	      accordance  with	the POSIX specification, which simply does not
	      specify any way of controlling the scanner's input other than by
	      making an	initial	assignment to yyin.

       -      The  unput() routine is not redefinable.	This restriction is in
	      accordance with POSIX.

       -      flex scanners are	not as reentrant as lex	scanners.  In particu-
	      lar, if you have an interactive scanner and an interrupt handler
	      which long-jumps out of the scanner, and the scanner  is	subse-
	      quently called again, you	may get	the following message:

		  fatal	flex scanner internal error--end of buffer missed

	      To reenter the scanner, first use

		  yyrestart( yyin );

	      Note  that this call will	throw away any buffered	input; usually
	      this isn't a problem with	an interactive scanner.

	      Also note	that flex C++ scanner classes  are  reentrant,	so  if
	      using  C++  is  an  option for you, you should use them instead.
	      See "Generating C++ Scanners" above for details.

       -      output() is not supported.  Output from the ECHO macro  is  done
	      to the file-pointer yyout	(default stdout).

	      output() is not part of the POSIX	specification.

       -      lex  does	 not  support  exclusive start conditions (%x),	though
	      they are in the POSIX specification.

       -      When definitions are expanded, flex encloses them	 in  parenthe-
	      ses.  With lex, the following:

		  NAME	  [A-Z][A-Z0-9]*
		  %%
		  foo{NAME}?	  printf( "Found it\n" );
		  %%

	      will  not	 match	the  string  "foo"  because  when the macro is
	      expanded the rule	is equivalent to "foo[A-Z][A-Z0-9]*?"  and the
	      precedence  is such that the '?' is associated with "[A-Z0-9]*".
	      With flex, the rule will be expanded  to	"foo([A-Z][A-Z0-9]*)?"
	      and so the string	"foo" will match.

	      Note that	if the definition begins with ^	or ends	with $ then it
	      is not expanded with parentheses,	to allow  these	 operators  to
	      appear  in  definitions  without	losing their special meanings.
	      But the <s>, /, and <<EOF>> operators cannot be used in  a  flex
	      definition.

	      Using  -l	 results  in the lex behavior of no parentheses	around
	      the definition.

	      The POSIX	specification is that the definition  be  enclosed  in
	      parentheses.

       -      Some  implementations of lex allow a rule's action to begin on a
	      separate line, if	the rule's pattern has trailing	whitespace:

		  %%
		  foo|bar<space	here>
		    { foobar_action(); }

	      flex does	not support this feature.

       -      The lex %r (generate a Ratfor scanner) option is not  supported.
	      It is not	part of	the POSIX specification.

       -      After  a	call  to  unput(),  yytext is undefined	until the next
	      token is matched,	unless the scanner  was	 built	using  %array.
	      This  is	not the	case with lex or the POSIX specification.  The
	      -l option	does away with this incompatibility.

       -      The precedence of	the {} (numeric	range) operator	is  different.
	      lex  interprets  "abc{1,3}"  as "match one, two, or three	occur-
	      rences of	'abc'",	whereas	flex interprets	it as "match 'ab' fol-
	      lowed  by	one, two, or three occurrences of 'c'".	 The latter is
	      in agreement with	the POSIX specification.

       -      The precedence of	the ^ operator is different.   lex  interprets
	      "^foo|bar" as "match either 'foo'	at the beginning of a line, or
	      'bar' anywhere", whereas flex interprets	it  as	"match	either
	      'foo'  or	 'bar'	if they	come at	the beginning of a line".  The
	      latter is	in agreement with the POSIX specification.

       -      The special table-size declarations such as %a supported by  lex
	      are not required by flex scanners; flex ignores them.

       -      The  name	 FLEX_SCANNER  is #define'd so scanners	may be written
	      for  use	with  either  flex  or	lex.   Scanners	 also  include
	      YY_FLEX_MAJOR_VERSION and	YY_FLEX_MINOR_VERSION indicating which
	      version of flex generated	the scanner (for example, for the  2.5
	      release, these defines would be 2	and 5 respectively).

       The following flex features are not included in lex or the POSIX	speci-
       fication:

	   C++ scanners
	   %option
	   start condition scopes
	   start condition stacks
	   interactive/non-interactive scanners
	   yy_scan_string() and	friends
	   yyterminate()
	   yy_set_interactive()
	   yy_set_bol()
	   YY_AT_BOL()
	   <<EOF>>
	   <*>
	   YY_DECL
	   YY_START
	   YY_USER_ACTION
	   YY_USER_INIT
	   #line directives
	   %{}'s around	actions
	   multiple actions on a line

       plus almost all of the flex flags.  The last feature in the list	refers
       to  the	fact  that  with flex you can put multiple actions on the same
       line, separated with semi-colons, while with lex, the following

	   foo	  handle_foo();	++num_foos_seen;

       is (rather surprisingly)	truncated to

	   foo	  handle_foo();

       flex does not truncate the action.  Actions that	are  not  enclosed  in
       braces are simply terminated at the end of the line.

DIAGNOSTICS
       warning,	rule cannot be matched indicates that the given	rule cannot be
       matched because it follows other	rules that will	always match the  same
       text  as	 it.   For  example,  in the following "foo" cannot be matched
       because it comes	after an identifier "catch-all"	rule:

	   [a-z]+    got_identifier();
	   foo	     got_foo();

       Using REJECT in a scanner suppresses this warning.

       warning,	-s option given	but default rule can be	matched	means that  it
       is  possible  (perhaps  only  in	a particular start condition) that the
       default rule (match any single character) is the	 only  one  that  will
       match  a	 particular input.  Since -s was given,	presumably this	is not
       intended.

       reject_used_but_not_detected undefined or  yymore_used_but_not_detected
       undefined - These errors	can occur at compile time.  They indicate that
       the scanner uses	REJECT or yymore() but that flex failed	to notice  the
       fact,  meaning  that  flex  scanned  the	first two sections looking for
       occurrences of these actions and	failed to find any,  but  somehow  you
       snuck  some  in (via a #include file, for example).  Use	%option	reject
       or %option yymore to indicate to	flex that you really do	use these fea-
       tures.

       flex  scanner  jammed  -	 a scanner compiled with -s has	encountered an
       input string which wasn't matched by any	of its rules.  This error  can
       also occur due to internal problems.

       token  too  large, exceeds YYLMAX - your	scanner	uses %array and	one of
       its rules matched a string longer than the YYLMAX constant (8K bytes by
       default).  You can increase the value by	#define'ing YYLMAX in the def-
       initions	section	of your	flex input.

       scanner requires	-8 flag	to use the character 'x' - Your	scanner	speci-
       fication	 includes  recognizing the 8-bit character 'x' and you did not
       specify the -8 flag, and	your scanner defaulted to  7-bit  because  you
       used  the  -Cf or -CF table compression options.	 See the discussion of
       the -7 flag for details.

       flex scanner push-back overflow - you used unput() to push back so much
       text that the scanner's buffer could not	hold both the pushed-back text
       and the current token in	yytext.	 Ideally the  scanner  should  dynami-
       cally resize the	buffer in this case, but at present it does not.

       input buffer overflow, can't enlarge buffer because scanner uses	REJECT
       - the scanner was working on matching  an  extremely  large  token  and
       needed  to  expand  the	input buffer.  This doesn't work with scanners
       that use	REJECT.

       fatal flex scanner internal error--end of  buffer  missed  -  This  can
       occur in	an scanner which is reentered after a long-jump	has jumped out
       (or over) the scanner's activation frame.  Before reentering the	 scan-
       ner, use:

	   yyrestart( yyin );

       or, as noted above, switch to using the C++ scanner class.

       too many	start conditions in __ construct! - you	listed more start con-
       ditions in a <> construct than exist (so	you must have listed at	 least
       one of them twice).

FILES
       -ll    library with which scanners must be linked.

       lex.yy.c
	      generated	scanner	(called	lexyy.c	on some	systems).

       lex.yy.cc
	      generated	C++ scanner class, when	using -+.

       _FlexLexer.h_
	      header  file defining the	C++ scanner base class,	FlexLexer, and
	      its derived class, yyFlexLexer.

       flex.skl
	      skeleton scanner.	 This file is only used	 when  building	 flex,
	      not when flex executes.

       lex.backup
	      backing-up  information for -b flag (called lex.bck on some sys-
	      tems).

DEFICIENCIES / BUGS
       Some trailing context patterns cannot be	properly matched and  generate
       warning	messages  ("dangerous  trailing	context").  These are patterns
       where the ending	of the first part of the rule matches the beginning of
       the  second  part, such as "zx*/xy*", where the 'x*' matches the	'x' at
       the beginning of	the trailing context.	(Note  that  the  POSIX	 draft
       states that the text matched by such patterns is	undefined.)

       For  some trailing context rules, parts which are actually fixed-length
       are not recognized as such, leading to the above	mentioned  performance
       loss.   In  particular,	parts  using '|' or {n}	(such as "foo{3}") are
       always considered variable-length.

       Combining trailing context with the special '|' action  can  result  in
       fixed  trailing	context	 being turned into the more expensive variable
       trailing	context.  For example, in the following:

	   %%
	   abc	    |
	   xyz/def

       Use of unput() invalidates yytext and yyleng, unless the	%array	direc-
       tive or the -l option has been used.

       Pattern-matching	 of  NUL's is substantially slower than	matching other
       characters.

       Dynamic resizing	of the input buffer is slow, as	it entails  rescanning
       all the text matched so far by the current (generally huge) token.

       Due  to	both  buffering	 of  input and read-ahead, you cannot intermix
       calls to	<stdio.h> routines, such as, for example, getchar(), with flex
       rules and expect	it to work.  Call input() instead.

       The  total  table  entries listed by the	-v flag	excludes the number of
       table entries needed to determine what rule has been matched.  The num-
       ber of entries is equal to the number of	DFA states if the scanner does
       not use REJECT, and somewhat greater than the number of	states	if  it
       does.

       REJECT cannot be	used with the -f or -F options.

       The flex	internal algorithms need documentation.

SEE ALSO
       lex(1), yacc(1),	sed(1),	awk(1).

       John Levine, Tony Mason,	and Doug Brown,	Lex _ Yacc, O'Reilly and Asso-
       ciates.	Be sure	to get the 2nd edition.

       M. E. Lesk and E. Schmidt, LEX -	Lexical	Analyzer Generator

       Alfred Aho, Ravi	Sethi and Jeffrey Ullman, Compilers: Principles, Tech-
       niques  and Tools, Addison-Wesley (1986).  Describes the	pattern-match-
       ing techniques used by flex (deterministic finite automata).

AUTHOR
       Vern Paxson, with the help of many ideas	and much inspiration from  Van
       Jacobson.  Original version by Jef Poskanzer.  The fast table represen-
       tation is a partial implementation of a design done  by	Van  Jacobson.
       The implementation was done by Kevin Gong and Vern Paxson.

       Thanks  to  the	many flex beta-testers,	feedbackers, and contributors,
       especially Francois Pinard, Casey Leedom, Robert	Abramovitz, Stan Ader-
       mann, Terry Allen, David	Barker-Plummer,	John Basrai, Neal Becker, Nel-
       son H.F.	Beebe, benson@odi.com, Karl Berry, Peter A. Bigot, Simon Blan-
       chard,  Keith  Bostic,  Frederic	 Brehm,	 Ian  Brockbank, Kin Cho, Nick
       Christopher, Brian Clapper, J.T.	Conklin,  Jason	 Coughlin,  Bill  Cox,
       Nick  Cropper,  Dave  Curtis,  Scott David Daniels, Chris G. Demetriou,
       Theo Deraadt, Mike Donahue, Chuck Doucette,  Tom	 Epperly,  Leo	Eskin,
       Chris  Faylor,  Chris Flatters, Jon Forrest, Jeffrey Friedl, Joe	Gayda,
       Kaveh R.	Ghazi, Wolfgang	Glunz, Eric  Goldman,  Christopher  M.	Gould,
       Ulrich  Grepel,	Peer Griebel, Jan Hajic, Charles Hemphill, NORO	Hideo,
       Jarkko Hietaniemi, Scott	Hofmann, Jeff Honig, Dana Hudes, Eric  Hughes,
       John  Interrante,  Ceriel  Jacobs, Michal Jaegermann, Sakari Jalovaara,
       Jeffrey R. Jones, Henry Juengst,	Klaus Kaempf, Jonathan I. Kamens, Ter-
       rence  O	 Kane,	Amir  Katz,  ken@ken.hilco.com,	 Kevin B. Kenny, Steve
       Kirsch, Winfried	Koenig,	Marq Kole, Ronald Lamprecht, Greg  Lee,	 Rohan
       Lenard,	Craig  Leres,  John Levine, Steve Liddle, David	Loffredo, Mike
       Long, Mohamed  el  Lozy,	 Brian	Madsen,	 Malte,	 Joe  Marshall,	 Bengt
       Martensson,  Chris  Metcalf,  Luke  Mewburn, Jim	Meyering, R. Alexander
       Milowski, Erik Naggum, G.T. Nicol,  Landon  Noll,  James	 Nordby,  Marc
       Nozell, Richard Ohnemus,	Karsten	Pahnke,	Sven Panne, Roland Pesch, Wal-
       ter Pelissero, Gaumond Pierre, Esmond Pitt, Jef Poskanzer, Joe  Rahmeh,
       Jarmo  Raiha,  Frederic	Raimbault,  Pat	Rankin,	Rick Richardson, Kevin
       Rodgers,	Kai Uwe	Rommel,	Jim Roskind, Alberto Santini, Andreas Scherer,
       Darrell	Schiebel,  Raf Schietekat, Doug	Schmidt, Philippe Schnoebelen,
       Andreas Schwab, Larry Schwimmer,	Alex Siegel, Eckehard Stolz,  Jan-Erik
       Strvmquist,  Mike  Stump,  Paul Stuart, Dave Tallman, Ian Lance Taylor,
       Chris Thewalt, Richard M. Timoney, Jodi Tsai, Paul Tuinenga, Gary Weik,
       Frank  Whaley,  Gerhard	Wilhelms,  Kent	Williams, Ken Yap, Ron Zellar,
       Nathan Zelle, David Zuhn, and those whose names have slipped my margin-
       al  mail-archiving  skills  but whose contributions are appreciated all
       the same.

       Thanks to Keith Bostic, Jon Forrest, Noah Friedman, John	Gilmore, Craig
       Leres,  John  Levine,  Bob  Mulcahy, G.T.  Nicol, Francois Pinard, Rich
       Salz,  and  Richard  Stallman  for  help	 with	various	  distribution
       headaches.

       Thanks  to Esmond Pitt and Earle	Horton for 8-bit character support; to
       Benson Margulies	and Fred Burke for C++ support;	to Kent	 Williams  and
       Tom Epperly for C++ class support; to Ove Ewerlid for support of	NUL's;
       and to Eric Hughes for support of multiple buffers.

       This work was primarily done when I was	with  the  Real	 Time  Systems
       Group at	the Lawrence Berkeley Laboratory in Berkeley, CA.  Many	thanks
       to all there for	the support I received.

       Send comments to	vern@ee.lbl.gov.

Version	2.5			  April	1995			       FLEX(1)

NAME | SYNOPSIS | OVERVIEW | DESCRIPTION | SOME SIMPLE EXAMPLES | FORMAT OF THE INPUT FILE | PATTERNS | HOW THE INPUT IS MATCHED | ACTIONS | THE GENERATED SCANNER | START CONDITIONS | MULTIPLE INPUT BUFFERS | END-OF-FILE RULES | MISCELLANEOUS MACROS | VALUES AVAILABLE TO THE USER | INTERFACING WITH YACC | OPTIONS | PERFORMANCE CONSIDERATIONS | GENERATING C++ SCANNERS | INCOMPATIBILITIES WITH LEX AND POSIX | DIAGNOSTICS | FILES | DEFICIENCIES / BUGS | SEE ALSO | AUTHOR

Want to link to this manual page? Use this URL:
<http://www.freebsd.org/cgi/man.cgi?query=lex&sektion=1&manpath=FreeBSD+4.6-RELEASE>

home | help