o
    'hZt                     @   sL  U d Z ddlZddlZddlmZ ddlmZ ddlm	Z	 ddl
mZ ddlmZ dd	lmZ ejr<ddlZdd
lmZ edZejejdf ed< edZedZedejZedejej B Z!edejej B Z"e	dZ#e	dZ$e	dZ%e	dZ&e	dZ'e	dZ(e	dZ)e	dZ*e	dZ+e	dZ,e	dZ-e	dZ.e	dZ/e	d Z0e	d!Z1e	d"Z2e	d#Z3e	d$Z4e	d%Z5e	d&Z6e	d'Z7e	d(Z8e	d)Z9e	d*Z:e	d+Z;e	d,Z<e	d-Z=e	d.Z>e	d/Z?e	d0Z@e	d1ZAe	d2ZBe	d3ZCe	d4ZDe	d5ZEe	d6ZFe	d7ZGe	d8ZHe	d9ZIe	d:ZJe	d;ZKe	d<ZLe	d=ZMe	d>ZNe	d?ZOe	d@ZPe	dAZQe	dBZRe	dCZSi dDe#dEe;dFe'dGe*dHe3dIe2dJe6dKe<dLe.dMe8dNe/dOe9dPe-dQe7dRe)dSe4dTe+e,e0e1e$e(e%e5e&e:dU	ZTdVdW eTU D ZVeWeTeWeVksJ dXedNdYXdZd[ eYeTd\d] d^D  dOZZe[eIeKeJe=eNeOePgZ\e[e=eQeKePgZ]d_e^d`e^fdadbZ_dcddd`e^fdedfZ`dge^d`e^fdhdiZadje^d`ebfdkdlZcdmdnd`ejdeje^e^f  fdodpZeG dqdr drZfG dsdd ddejgZhG dtdu duZiG dvdw dwZjddxdyZkG dzd{ d{elZmG d|d} d}ejgZnG d~d dZodS )zImplements a Jinja / Python combination lexer. The ``Lexer`` class
is used to do some preprocessing. It filters out invalid operators like
the bitshift operators we don't allow in templates. It separates
template code and python code in expressions.
    N)literal_eval)deque)intern   )pattern)TemplateSyntaxError)LRUCache)Environment2   Lexer_lexer_cachez\s+z(\r\n|\r|\n)z7('([^'\\]*(?:\\.[^'\\]*)*)'|"([^"\\]*(?:\\.[^"\\]*)*)")z
    (
        0b(_?[0-1])+ # binary
    |
        0o(_?[0-7])+ # octal
    |
        0x(_?[\da-f])+ # hex
    |
        [1-9](_?\d)* # decimal
    |
        0(_?0)* # decimal zero
    )
    z
    (?<!\.)  # doesn't start with a .
    (\d+_)*\d+  # digits, possibly _ separated
    (
        (\.(\d+_)*\d+)?  # optional fractional part
        e[+\-]?(\d+_)*\d+  # exponent part
    |
        \.(\d+_)*\d+  # required fractional part
    )
    addassigncoloncommadivdoteqfloordivgtgteqlbracelbracketlparenltlteqmodmulnepipepowrbracerbracketrparen	semicolonsubtilde
whitespacefloatintegernamestringoperatorblock_begin	block_endvariable_beginvariable_end	raw_beginraw_endcomment_begincomment_endcommentlinestatement_beginlinestatement_endlinecomment_beginlinecomment_endlinecommentdatainitialeof+-/z//*%z**~[](){}z==z!=>)	z>=<z<==.:|,;c                 C   s   i | ]\}}||qS  rR   ).0kvrR   rR   H/var/www/html/olx_land/venv/lib/python3.10/site-packages/jinja2/lexer.py
<dictcomp>   s    rW   zoperators droppedrO   c                 c   s    | ]}t |V  qd S N)reescaperS   xrR   rR   rV   	<genexpr>       r]   c                 C   s
   t |  S rX   )lenr\   rR   rR   rV   <lambda>   s   
 ra   )key
token_typereturnc                 C   sL   | t v rt |  S tdtdtdtdtdtdtdtdt	dt
d	td
tdi| | S )Nzbegin of commentzend of commentr5   zbegin of statement blockzend of statement blockzbegin of print statementzend of print statementzbegin of line statementzend of line statementztemplate data / textzend of template)reverse_operatorsTOKEN_COMMENT_BEGINTOKEN_COMMENT_ENDTOKEN_COMMENTTOKEN_LINECOMMENTTOKEN_BLOCK_BEGINTOKEN_BLOCK_ENDTOKEN_VARIABLE_BEGINTOKEN_VARIABLE_ENDTOKEN_LINESTATEMENT_BEGINTOKEN_LINESTATEMENT_END
TOKEN_DATA	TOKEN_EOFget)rc   rR   rR   rV   _describe_token_type   s"   rs   tokenTokenc                 C   s   | j tkr| jS t| j S )z#Returns a description of the token.)type
TOKEN_NAMEvaluers   )rt   rR   rR   rV   describe_token   s   

ry   exprc                 C   s8   d| v r|  dd\}}|tkr|S t|S | }t|S )z0Like `describe_token` but for token expressions.rN   r   )splitrw   rs   )rz   rv   rx   rR   rR   rV   describe_token_expr   s   r|   rx   c                 C   s   t t| S )zsCount the number of newline characters in the string.  This is
    useful for extensions that filter a stream.
    )r_   
newline_refindall)rx   rR   rR   rV   count_newlines      r   environmentr	   c                 C   s   t j}t| jt|| jft| jt|| jft| jt|| jfg}| j	dur8|
t| j	td|| j	 f | jdurM|
t| jtd|| j f dd t|ddD S )zACompiles all the rules from the environment into a list of rules.Nz	^[ \t\v]*z(?:^|(?<=\S))[^\S\r\n]*c                 S   s   g | ]}|d d qS )r   NrR   r[   rR   rR   rV   
<listcomp>   s    z!compile_rules.<locals>.<listcomp>T)reverse)rY   rZ   r_   comment_start_stringrf   block_start_stringrj   variable_start_stringrl   line_statement_prefixappendrn   line_comment_prefixTOKEN_LINECOMMENT_BEGINsorted)r   erulesrR   rR   rV   compile_rules   s:   

r   c                   @   sL   e Zd ZdZefdedeje ddfddZde	d	ej
e dd
fddZdS )FailurezjClass that raises a `TemplateSyntaxError` if called.
    Used by the `Lexer` to specify known errors.
    messageclsrd   Nc                 C   s   || _ || _d S rX   )r   error_class)selfr   r   rR   rR   rV   __init__  s   
zFailure.__init__linenofilenamezte.NoReturnc                 C   s   |  | j||rX   )r   r   )r   r   r   rR   rR   rV   __call__	  s   zFailure.__call__)__name__
__module____qualname____doc__r   strtTyper   intOptionalr   rR   rR   rR   rV   r      s    
 r   c                   @   sX   e Zd ZU eed< eed< eed< defddZdedefdd	Zd
edefddZ	dS )ru   r   rv   rx   rd   c                 C      t | S rX   )ry   r   rR   rR   rV   __str__     zToken.__str__rz   c                 C   s2   | j |krdS d|v r|dd| j | jgkS dS )zTest a token against a token expression.  This can either be a
        token type or ``'token_type:token_value'``.  This can only test
        against string values and types.
        TrN   r   F)rv   r{   rx   r   rz   rR   rR   rV   test  s
   
z
Token.testiterablec                    s   t  fdd|D S )z(Test against multiple token expressions.c                 3   s    | ]}  |V  qd S rX   )r   )rS   rz   r   rR   rV   r]   &  r^   z!Token.test_any.<locals>.<genexpr>)any)r   r   rR   r   rV   test_any$  s   zToken.test_anyN)
r   r   r   r   __annotations__r   r   boolr   r   rR   rR   rR   rV   ru     s   
 c                   @   s2   e Zd ZdZdddZddd	Zdefd
dZdS )TokenStreamIteratorz`The iterator for tokenstreams.  Iterate over the stream
    until the eof token is reached.
    streamTokenStreamrd   Nc                 C   s
   || _ d S rX   )r   )r   r   rR   rR   rV   r   .  s   
zTokenStreamIterator.__init__c                 C   s   | S rX   rR   r   rR   rR   rV   __iter__1  s   zTokenStreamIterator.__iter__c                 C   s.   | j j}|jtu r| j   tt| j  |S rX   )r   currentrv   rq   closeStopIterationnextr   rt   rR   rR   rV   __next__4  s   


zTokenStreamIterator.__next__)r   r   rd   N)rd   r   )r   r   r   r   r   r   ru   r   rR   rR   rR   rV   r   )  s
    

r   c                   @   s   e Zd ZdZdeje deje deje fddZ	de
fdd	Zdefd
dZedefddZdeddfddZdefddZd#deddfddZdedeje fddZdedefddZdefddZd$dd Zdedefd!d"ZdS )%r   zA token stream is an iterable that yields :class:`Token`\s.  The
    parser however does not iterate over it but calls :meth:`next` to go
    one token ahead.  The current active token is stored as :attr:`current`.
    	generatorr*   r   c                 C   s>   t || _t | _|| _|| _d| _tdtd| _	t
|  d S )NFr    )iter_iterr   _pushedr*   r   closedru   TOKEN_INITIALr   r   )r   r   r*   r   rR   rR   rV   r   E  s   
zTokenStream.__init__rd   c                 C   r   rX   )r   r   rR   rR   rV   r   S  r   zTokenStream.__iter__c                 C   s   t | jp
| jjtuS rX   )r   r   r   rv   rq   r   rR   rR   rV   __bool__V  s   zTokenStream.__bool__c                 C   s   |  S )z Are we at the end of the stream?rR   r   rR   rR   rV   eosY  s   zTokenStream.eosrt   Nc                 C   s   | j | dS )z Push a token back to the stream.N)r   r   r   rR   rR   rV   push^  s   zTokenStream.pushc                 C   s"   t | }| j}| | || _|S )zLook at the next token.)r   r   r   )r   	old_tokenresultrR   rR   rV   lookb  s
   
zTokenStream.lookr   nc                 C   s   t |D ]}t|  qdS )zGot n tokens ahead.N)ranger   )r   r   _rR   rR   rV   skipj  s   
zTokenStream.skiprz   c                 C   s   | j |r
t| S dS )zqPerform the token test and return the token if it matched.
        Otherwise the return value is `None`.
        N)r   r   r   r   rR   rR   rV   next_ifo  s   zTokenStream.next_ifc                 C   s   |  |duS )z8Like :meth:`next_if` but only returns `True` or `False`.N)r   r   rR   rR   rV   skip_ifx  s   zTokenStream.skip_ifc                 C   s\   | j }| jr| j | _ |S | j jtur,z	t| j| _ W |S  ty+   |   Y |S w |S )z|Go one token ahead and return the old one.

        Use the built-in :func:`next` instead of calling this directly.
        )	r   r   popleftrv   rq   r   r   r   r   )r   rvrR   rR   rV   r   |  s   
zTokenStream.__next__c                 C   s&   t | jjtd| _td| _d| _dS )zClose the stream.r   rR   TN)ru   r   r   rq   r   r   r   r   rR   rR   rV   r     s   

zTokenStream.closec                 C   sn   | j |s3t|}| j jtu rtd|d| j j| j| jtd|dt	| j | j j| j| jt
| S )z}Expect a given token type and return it.  This accepts the same
        argument as :meth:`jinja2.lexer.Token.test`.
        z%unexpected end of template, expected rM   zexpected token z, got )r   r   r|   rv   rq   r   r   r*   r   ry   r   r   rR   rR   rV   expect  s    
zTokenStream.expect)r   )rd   N)r   r   r   r   r   Iterableru   r   r   r   r   r   r   r   propertyr   r   r   r   r   r   r   r   r   r   rR   rR   rR   rV   r   ?  s*    
	
r   c                 C   sZ   | j | j| j| j| j| j| j| j| j| j	| j
| jf}t|}|du r+t|  t|< }|S )z(Return a lexer which is probably cached.N)r   block_end_stringr   variable_end_stringr   comment_end_stringr   r   trim_blockslstrip_blocksnewline_sequencekeep_trailing_newliner   rr   r   )r   rb   lexerrR   rR   rV   	get_lexer  s"   
r   c                       s$   e Zd ZdZdZ fddZ  ZS )OptionalLStripzWA special tuple for marking a point in the state that can have
    lstrip applied.
    rR   c                    s   t  | |S rX   )super__new__)r   memberskwargs	__class__rR   rV   r     s   zOptionalLStrip.__new__)r   r   r   r   	__slots__r   __classcell__rR   rR   r   rV   r     s    r   c                   @   sN   e Zd ZU eje ed< ejeejedf eje	 f ed< ej
e ed< dS )_Ruler   .tokenscommandN)r   r   r   r   Patternr   r   UnionTupler   r   rR   rR   rR   rV   r     s   
 $r   c                   @   s   e Zd ZdZdddZdedefd	d
Z			ddedeje deje deje de	f
ddZ
		ddejejeeef  deje deje deje fddZ		ddedeje deje deje dejejeeef  f
ddZdS )r   a  Class that implements a lexer for a given environment. Automatically
    created by the environment class, usually you don't have to do that.

    Note that the lexer is not automatically bound to an environment.
    Multiple environments can share the same lexer.
    r   r	   rd   Nc                 C   s&  t j}dtdtjt fdd}tttd ttt	d tt
td tttd tttd tttd g}t|}||j}||j}||j}||j}	|jrLdnd}
|j| _|j| _|j| _d| d| d	| d
}d|gdd |D  }dt|d| dttddt|dtd gtt|d| d| d	| |
 d
tt fdt|dt!dfd gt"t|d| d| d	| |
 dt#dg| t$t|d|	 d	|	 t%dg| t&t|d| d| d| d	| |
 d

ttt'dt|dt!dfd gt(t|dt)dg| t*t|dt+t,fdgi| _-d S )Nr\   rd   c                 S   s   t | t jt jB S rX   )rY   compileMSr`   rR   rR   rV   c  s   zLexer.__init__.<locals>.cz\n?r   z(?P<raw_begin>z(\-|\+|)\s*raw\s*(?:\-z\s*|z))rO   c                 S   s"   g | ]\}}d | d| dqS )z(?P<rJ   z	(\-|\+|))rR   )rS   r   rrR   rR   rV   r   
  s   " z"Lexer.__init__.<locals>.<listcomp>rootz(.*?)(?:rG   #bygroupz.+z(.*?)((?:\+z|\-#popz(.)zMissing end of comment tagz(?:\+z\-z	(.*?)((?:z(\-|\+|))\s*endraw\s*(?:\+zMissing end of raw directivez	\s*(\n|$)z(.*?)()(?=\n|$)).rY   rZ   r   r   r   r   whitespace_reTOKEN_WHITESPACEfloat_reTOKEN_FLOAT
integer_reTOKEN_INTEGERname_rerw   	string_reTOKEN_STRINGoperator_reTOKEN_OPERATORr   r   r   r   r   r   r   r   r   joinr   rp   rf   rh   rg   r   rj   rk   rl   rm   TOKEN_RAW_BEGINTOKEN_RAW_ENDrn   ro   r   ri   TOKEN_LINECOMMENT_ENDr   )r   r   r   r   	tag_rulesroot_tag_rulesblock_start_reblock_end_recomment_end_revariable_end_reblock_suffix_reroot_raw_reroot_parts_rerR   rR   rV   r     s   










		
zLexer.__init__rx   c                 C   s   t | j|S )z`Replace all newlines with the configured sequence in strings
        and template data.
        )r}   r%   r   )r   rx   rR   rR   rV   _normalize_newlinesV  r   zLexer._normalize_newlinessourcer*   r   statec                 C   s&   |  ||||}t| |||||S )z:Calls tokeniter + tokenize and wraps it in a token stream.)	tokeniterr   wrap)r   r  r*   r   r  r   rR   rR   rV   tokenize\  s   zLexer.tokenizer   c           
      c   sT   |D ]\}}}|t v rq|}|tkrt}n|tkrt}n|ttfv r$q|tkr.| |}nr|dkr5|}nk|t	krG|}|
 sFtd|||nY|tkr}z| |dd ddd}W nB ty| } zt|dd  }	t|	||||d	}~ww |tkrt|d
dd}n|tkrt|d
d}n|tkrt| }t|||V  qd	S )zThis is called with the stream as returned by `tokenize` and wraps
        every token in a :class:`Token` and converts the value.
        keywordzInvalid character in identifierr   asciibackslashreplacezunicode-escaperN   Nr   r   r   )ignored_tokensrn   rj   ro   rk   r   r   rp   r  rw   isidentifierr   r   encodedecode	Exceptionr   r{   stripr   r   replacer   r   r   	operatorsru   )
r   r   r*   r   r   rt   	value_strrx   r   msgrR   rR   rV   r
  g  sR   	z
Lexer.wrapc                 c   s"   t |ddd }| js|d dkr|d= d|}d}d}dg}|dur:|dkr:|d	v s3J d
||d  | j|d  }	t|}
g }d}d}	 |	D ]\}}}|||}|du r_qN|ri|tt	t
fv riqNt|tr2| }t|tr|d }tdd |ddd D }|dkr| }|t|d d}|g|dd }n.|dkr| jr| ts|dd }|dks|rt||r|d| g|dd }t|D ]W\}}t|tr||||dkr|  D ]\}}|dur
|||fV  ||d7 } nqt|dq|| }|s|tvr%|||fV  ||d| 7 }d}qni| }|tkr|dkrF|d n@|dkrQ|d n5|dkr\|d n*|dv r|sot d| d||||! }||krt d| d| d||||s|tvr|||fV  ||d7 }| dd dk}|" }|dur|dkr|!  n*|dkr|  D ]\}}|dur||  n	qt|dn|| | j|d  }	n||krt|d|} n||
krdS t d || d!| |||qL)"a  This method tokenizes the text and returns the tokens in a
        generator. Use this method if you just want to tokenize a template.

        .. versionchanged:: 3.0
            Only ``\n``, ``\r\n`` and ``\r`` are treated as line
            breaks.
        N   r  r   
r   r   r   )variableblockzinvalid state_beginTc                 s   s    | ]	}|d ur|V  qd S rX   rR   )rS   grR   rR   rV   r]     s    z"Lexer.tokeniter.<locals>.<genexpr>r?   r>   r   z= wanted to resolve the token dynamically but no group matchedrH   rI   rF   rG   rD   rE   )rI   rG   rE   zunexpected ''z', expected 'r   zA wanted to resolve the new state dynamically but no group matchedz* yielded empty string without stack changezunexpected char z at )#r}   r{   r   r   r   r   r_   matchrm   rk   ro   
isinstancetuplegroupsr   r   rstripcountr   	groupdictrr   rl   rfindr   	fullmatch	enumerater   itemsRuntimeErrorignore_if_emptygroupr   r   popend)r   r  r*   r   r  linesposr   stackstatetokenssource_lengthbalancing_stacknewlines_strippedline_startingregexr   	new_statemr$  text
strip_signstrippedl_posidxrt   rb   rx   r;   expected_oppos2rR   rR   rV   r	    s   




















 zLexer.tokeniter)r   r	   rd   N)NNN)NN)r   r   r   r   r   r   r  r   r   r   r  r   r   r   Iteratorru   r
  r	  rR   rR   rR   rV   r     sT    
w	

:)r   r	   rd   r   )pr   rY   typingr   astr   collectionsr   sysr   _identifierr   r   
exceptionsr   utilsr   TYPE_CHECKINGtyping_extensionster   r	   r   MutableMappingr   r   r   r   r}   r   r   
IGNORECASEVERBOSEr   r   	TOKEN_ADDTOKEN_ASSIGNTOKEN_COLONTOKEN_COMMA	TOKEN_DIV	TOKEN_DOTTOKEN_EQTOKEN_FLOORDIVTOKEN_GT
TOKEN_GTEQTOKEN_LBRACETOKEN_LBRACKETTOKEN_LPARENTOKEN_LT
TOKEN_LTEQ	TOKEN_MOD	TOKEN_MULTOKEN_NE
TOKEN_PIPE	TOKEN_POWTOKEN_RBRACETOKEN_RBRACKETTOKEN_RPARENTOKEN_SEMICOLON	TOKEN_SUBTOKEN_TILDEr   r   r   rw   r   r   rj   rk   rl   rm   r   r   rf   rg   rh   rn   ro   r   r   ri   rp   r   rq   r  r+  re   r_   r   r   r   	frozensetr  r-  r   rs   ry   r|   r   r   Listr   r   
NamedTupleru   r   r   r   r#  r   r   r   rR   rR   rR   rV   <module>   s.   




	
&
")
m