ok

Mini Shell

Direktori : /proc/thread-self/root/proc/thread-self/root/usr/lib/python2.7/site-packages/jinja2/
Upload File :
Current File : //proc/thread-self/root/proc/thread-self/root/usr/lib/python2.7/site-packages/jinja2/lexer.pyc

�
��Rc	@s�dZddlZddlmZddlmZddlmZddlm	Z	ddl
mZmZm
Z
mZmZe	d�Zejd	ej�Zejd
ej�Zejd�Zyedd
d�Wn ek
r�ejd�Zn0XddlmZejdejejf�Zejd�Zejd�Zed�Zed�Z ed�Z!ed�Z"ed�Z#ed�Z$ed�Z%ed�Z&ed�Z'ed�Z(ed�Z)ed�Z*ed �Z+ed!�Z,ed"�Z-ed#�Z.ed$�Z/ed%�Z0ed&�Z1ed'�Z2ed(�Z3ed)�Z4ed*�Z5ed+�Z6ed,�Z7ed-�Z8ed.�Z9ed/�Z:ed0�Z;ed1�Z<ed2�Z=ed3�Z>ed4�Z?ed5�Z@ed6�ZAed7�ZBed8�ZCed9�ZDed:�ZEed;�ZFed<�ZGed=�ZHed>�ZIed?�ZJed@�ZKedA�ZLedB�ZMedC�ZNedD�ZOiedE6e7dF6e#dG6e&dH6e/dI6e.dJ6e2dK6e8dL6e*dM6e4dN6e+dO6e5dP6e)dQ6e3dR6e%dS6e0dT6e'dU6e(dV6e,dW6e-dX6e dY6e$dZ6e!d[6e1d\6e"d]6e6d^6ZPeQgeeP�D]\ZRZSeSeRf^q[�ZTeUeP�eUeT�ks�tVd_��ejd`d\jWda�eXePdbdc��D���ZYeZeEeGeFe9e9eJeKeLg�Z[eZe9eMeGeLg�Z\dd�Z]de�Z^df�Z_dg�Z`dh�Zadiebfdj��YZcdkedfdl��YZee
dmebfdn��Y�Zfe
doebfdp��Y�Zgdq�Zhdrebfds��YZidS(ts�
    jinja2.lexer
    ~~~~~~~~~~~~

    This module implements a Jinja / Python combination lexer. The
    `Lexer` class provided by this module is used to do some preprocessing
    for Jinja.

    On the one hand it filters out invalid operators like the bitshift
    operators we don't allow in templates. On the other hand it separates
    template code and python code in expressions.

    :copyright: (c) 2010 by the Jinja Team.
    :license: BSD, see LICENSE for more details.
i����N(t
itemgetter(tdeque(tTemplateSyntaxError(tLRUCache(tnextt	iteritemstimplements_iteratort	text_typetinterni2s\s+s7('([^'\\]*(?:\\.[^'\\]*)*)'|"([^"\\]*(?:\\.[^"\\]*)*)")s\d+sföös	<unknown>tevals\b[a-zA-Z_][a-zA-Z0-9_]*\b(t_stringdefss	[%s][%s]*s(?<!\.)\d+\.\d+s(\r\n|\r|\n)taddtassigntcolontcommatdivtdotteqtfloordivtgttgteqtlbracetlbrackettlparentlttlteqtmodtmultnetpipetpowtrbracetrbrackettrparent	semicolontsubttildet
whitespacetfloattintegertnametstringtoperatortblock_begint	block_endtvariable_begintvariable_endt	raw_begintraw_endt
comment_begintcomment_endtcommenttlinestatement_begintlinestatement_endtlinecomment_begintlinecomment_endtlinecommenttdatatinitialteoft+t-t/s//t*t%s**t~t[t]t(t)t{t}s==s!=t>s>=t<s<=t=t.t:t|t,t;soperators droppeds(%s)ccs|]}tj|�VqdS(N(tretescape(t.0tx((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pys	<genexpr>�stkeycCst|�S(N(tlen(RS((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyt<lambda>�scCsx|tkrt|Sidt6dt6dt6dt6dt6dt6dt6dt6dt	6d	t
6d
t6dt6j
||�S(Nsbegin of commentsend of commentR3sbegin of statement blocksend of statement blocksbegin of print statementsend of print statementsbegin of line statementsend of line statementstemplate data / textsend of template(treverse_operatorstTOKEN_COMMENT_BEGINtTOKEN_COMMENT_ENDt
TOKEN_COMMENTtTOKEN_LINECOMMENTtTOKEN_BLOCK_BEGINtTOKEN_BLOCK_ENDtTOKEN_VARIABLE_BEGINtTOKEN_VARIABLE_ENDtTOKEN_LINESTATEMENT_BEGINtTOKEN_LINESTATEMENT_ENDt
TOKEN_DATAt	TOKEN_EOFtget(t
token_type((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyt_describe_token_type�s 
cCs#|jdkr|jSt|j�S(s#Returns a description of the token.R((ttypetvalueRf(ttoken((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pytdescribe_token�scCsGd|kr7|jdd�\}}|dkr=|Sn|}t|�S(s0Like `describe_token` but for token expressions.RLiR((tsplitRf(texprRgRh((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pytdescribe_token_expr�scCsttj|��S(ssCount the number of newline characters in the string.  This is
    useful for extensions that filter a stream.
    (RUt
newline_retfindall(Rh((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pytcount_newlines�scCstj}t|j�d||j�ft|j�d||j�ft|j�d||j�fg}|jd
k	r�|jt|j�dd||j�f�n|j	d
k	r�|jt|j	�dd||j	�f�ngt
|dt�D]}|d	^q�S(sACompiles all the rules from the environment into a list of rules.R3tblocktvariablet
linestatements	^[ \t\v]*R8s(?:^|(?<=\S))[^\S\r\n]*treverseiN(RPRQRUtcomment_start_stringtblock_start_stringtvariable_start_stringtline_statement_prefixtNonetappendtline_comment_prefixtsortedtTrue(tenvironmenttetrulesRS((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyt
compile_rules�s	tFailurecBs#eZdZed�Zd�ZRS(sjClass that raises a `TemplateSyntaxError` if called.
    Used by the `Lexer` to specify known errors.
    cCs||_||_dS(N(tmessageterror_class(tselfR�tcls((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyt__init__�s	cCs|j|j||��dS(N(R�R�(R�tlinenotfilename((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyt__call__�s(t__name__t
__module__t__doc__RR�R�(((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyR��stTokencBs`eZdZdZd�ed�D�\ZZZd�Zd�Z	d�Z
d�Zd�ZRS(	sToken class.ccs!|]}tt|��VqdS(N(tpropertyR(RRRS((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pys	<genexpr>�sicCs%tj||tt|��|f�S(N(ttuplet__new__Rtstr(R�R�RgRh((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyR��scCs7|jtkrt|jS|jdkr0|jS|jS(NR((RgRWRh(R�((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyt__str__�s
cCsE|j|krtSd|krA|jdd�|j|jgkStS(s�Test a token against a token expression.  This can either be a
        token type or ``'token_type:token_value'``.  This can only test
        against string values and types.
        RLi(RgR}RkRhtFalse(R�Rl((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyttest�s
"cGs(x!|D]}|j|�rtSqWtS(s(Test against multiple token expressions.(R�R}R�(R�titerableRl((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyttest_any�s
cCsd|j|j|jfS(NsToken(%r, %r, %r)(R�RgRh(R�((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyt__repr__s((
R�R�R�t	__slots__trangeR�RgRhR�R�R�R�R�(((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyR��s			
	tTokenStreamIteratorcBs)eZdZd�Zd�Zd�ZRS(s`The iterator for tokenstreams.  Iterate over the stream
    until the eof token is reached.
    cCs
||_dS(N(tstream(R�R�((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyR�scCs|S(N((R�((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyt__iter__scCsE|jj}|jtkr4|jj�t��nt|j�|S(N(R�tcurrentRgRctcloset
StopIterationR(R�Ri((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyt__next__s

(R�R�R�R�R�R�(((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyR�s		tTokenStreamcBs�eZdZd�Zd�Zd�ZeZed�dd�Zd�Z	d�Z
d	d
�Zd�Zd�Z
d
�Zd�Zd�ZRS(s�A token stream is an iterable that yields :class:`Token`\s.  The
    parser however does not iterate over it but calls :meth:`next` to go
    one token ahead.  The current active token is stored as :attr:`current`.
    cCsYt|�|_t�|_||_||_t|_tdt	d�|_
t|�dS(Nit(titert_iterRt_pushedR(R�R�tclosedR�t
TOKEN_INITIALR�R(R�t	generatorR(R�((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyR�(s			cCs
t|�S(N(R�(R�((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyR�1scCst|j�p|jjtk	S(N(tboolR�R�RgRc(R�((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyt__bool__4scCs|S(N((RS((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyRV8stdocs Are we at the end of the stream?cCs|jj|�dS(s Push a token back to the stream.N(R�Rz(R�Ri((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pytpush:scCs/t|�}|j}|j|�||_|S(sLook at the next token.(RR�R�(R�t	old_tokentresult((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pytlook>s
	
	icCs%xt|�D]}t|�q
WdS(sGot n tokens ahead.N(R�R(R�tnRS((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pytskipFscCs |jj|�rt|�SdS(sqPerform the token test and return the token if it matched.
        Otherwise the return value is `None`.
        N(R�R�R(R�Rl((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pytnext_ifKscCs|j|�dk	S(s8Like :meth:`next_if` but only returns `True` or `False`.N(R�Ry(R�Rl((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pytskip_ifRscCst|j}|jr'|jj�|_nI|jjtk	rpyt|j�|_Wqptk
rl|j�qpXn|S(s)Go one token ahead and return the old one(	R�R�tpopleftRgRcRR�R�R�(R�trv((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyR�Vs		
cCs1t|jjtd�|_d|_t|_dS(sClose the stream.R�N(R�R�R�RcRyR�R}R�(R�((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyR�bs	cCs�|jj|�s�t|�}|jjtkrXtd||jj|j|j��ntd|t	|j�f|jj|j|j��nz|jSWdt
|�XdS(s}Expect a given token type and return it.  This accepts the same
        argument as :meth:`jinja2.lexer.Token.test`.
        s(unexpected end of template, expected %r.sexpected token %r, got %rN(R�R�RmRgRcRR�R(R�RjR(R�Rl((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pytexpecths		(R�R�R�R�R�R�t__nonzero__R�teosR�R�R�R�R�R�R�R�(((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyR�!s										cCs�|j|j|j|j|j|j|j|j|j|j	|j
|jf}tj
|�}|dkr�t|�}|t|<n|S(s(Return a lexer which is probably cached.N(Rvtblock_end_stringRwtvariable_end_stringRutcomment_end_stringRxR{ttrim_blockst
lstrip_blockstnewline_sequencetkeep_trailing_newlinet_lexer_cacheRdRytLexer(R~RTtlexer((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyt	get_lexer}s"
R�cBsPeZdZd�Zd�Zdddd�Zddd�Zddd�ZRS(s
Class that implements a lexer for a given environment. Automatically
    created by the environment class, usually you don't have to do that.

    Note that the lexer is not automatically bound to an environment.
    Multiple environments can share the same lexer.
    cCs4d�}tj}ttdfttdfttdft	t
dfttdft
tdfg}t|�}|jr{dp~d}i}|jr�|d�}|d||j��}	|	j|j�}
||
r�d||
jd��p�d7}|	j|j�}
||
r$d||
jd��p'd7}|d||j��}|j|j�}
|
rud||
jd��pxd}d	}
d
|
||j�|||j�f}d|
||j�|||j�f}||d<||d
<nd||j�}|j|_|j|_i|ddjd||j�|||j�||j�fgg|D]+\}}d|||j||�f^qZ��tdfdf|d�tdfgd6|d||j�||j�|f�ttfdf|d�t d�fdfgt!6|d||j�||j�|f�t"dfg|t#6|d||j$�||j$�f�t%dfg|t&6|d||j�|||j�||j�|f�tt'fdf|d�t d�fdfgt(6|d�t)dfg|t*6|d�t+t,fdfgt-6|_.dS( NcSstj|tjtjB�S(N(RPtcompiletMtS(RS((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyRV�ss\n?R�R<s^%s(.*)s|%sis(?!%s)s^[ \t]*s%s%s(?!%s)|%s\+?s%s%s%s|%s\+?RqR3s%ss(.*?)(?:%s)RMs4(?P<raw_begin>(?:\s*%s\-|%s)\s*raw\s*(?:\-%s\s*|%s))s(?P<%s_begin>\s*%s\-|%s)s#bygroups.+troots(.*?)((?:\-%s\s*|%s)%s)s#pops(.)sMissing end of comment tags(?:\-%s\s*|%s)%ss
\-%s\s*|%ss1(.*?)((?:\s*%s\-|%s)\s*endraw\s*(?:\-%s\s*|%s%s))sMissing end of raw directives	\s*(\n|$)s(.*?)()(?=\n|$)(/RPRQt
whitespace_retTOKEN_WHITESPACERytfloat_retTOKEN_FLOATt
integer_ret
TOKEN_INTEGERtname_ret
TOKEN_NAMEt	string_retTOKEN_STRINGtoperator_retTOKEN_OPERATORR�R�R�RvtmatchRutgroupRwR�R�tjoinR�RdRbR�RZRYR�RXR]R\R�R_R^t
TOKEN_RAW_ENDtTOKEN_RAW_BEGINRaR`R[tTOKEN_LINECOMMENT_ENDtTOKEN_LINECOMMENT_BEGINR�(R�R~tcRt	tag_rulestroot_tag_rulestblock_suffix_ret	prefix_retno_lstrip_ret
block_difftmtcomment_difftno_variable_ret	lstrip_retblock_prefix_retcomment_prefix_reR�tr((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyR��s�				))%

:
"

"cCstj|j|�S(s@Called for strings and template data to normalize it to unicode.(RnR#R�(R�Rh((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyt_normalize_newlinesscCs7|j||||�}t|j|||�||�S(sCCalls tokeniter + tokenize and wraps it in a token stream.
        (t	tokeniterR�twrap(R�tsourceR(R�tstateR�((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyttokenizesc	cs�x�|D]�\}}}|tkr(qn�|dkr=d}np|dkrRd}n[|dkrdqnI|dkr�|j|�}n+|dkr�|}n|d	kr�t|�}n�|d
kr^y/|j|dd!�jd
d�jd�}WnGtk
r6}t|�jd�dj�}t||||��nXyt|�}Wq�t	k
rZq�XnO|dkryt
|�}n4|dkr�t|�}n|dkr�t|}nt
|||�VqWdS(s�This is called with the stream as returned by `tokenize` and wraps
        every token in a :class:`Token` and converts the value.
        R4R+R5R,R/R0R9tkeywordR(R)ii����tasciitbackslashreplacesunicode-escapeRLR'R&R*N(R/R0(tignored_tokensR�R�tencodetdecodet	ExceptionRktstripRtUnicodeErrortintR&t	operatorsR�(	R�R�R(R�R�RiRhRtmsg((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyR�$sD			


c	csrt|�}|j�}|jr[|r[x1d D]&}|j|�r.|jd�Pq.q.Wndj|�}d}d}dg}	|dk	r�|dkr�|d!ks�td
��|	j|d�nd}|j|	d}
t	|�}g}xxy|
D]>\}
}}|
j
||�}|dkr)q�n|rA|d"krAq�nt|t�rMx
t
|�D]�\}}|jtkr�|||��q]|dkr�x�t|j��D]=\}}|dk	r�|||fV||jd�7}Pq�q�Wtd|
��q]|j|d�}|s"|tkr3|||fVn||jd�7}q]Wn|j�}|dkr'|dkr�|jd�q'|dkr�|jd�q'|dkr�|jd�q'|d#kr'|s�td||||��n|j�}||kr$td||f|||��q$q'n|s9|tkrJ|||fVn||jd�7}|j�}|dk	r|dkr�|	j�nl|dkr�x]t|j��D])\}}|dk	r�|	j|�Pq�q�Wtd|
��n
|	j|�|j|	d}
n||kr-td|
��n|}Pq�W||krHdStd|||f|||��q�dS($s�This method tokenizes the text and returns the tokens in a
        generator.  Use this method if you just want to tokenize a template.
        s
s
s
R�iiR�RrRqs
invalid statet_begini����R.R,R5s#bygroups?%r wanted to resolve the token dynamically but no group matchedR*RFRGRDRERBRCsunexpected '%s'sunexpected '%s', expected '%s's#popsC%r wanted to resolve the new state dynamically but no group matcheds,%r yielded empty string without stack changeNsunexpected char %r at %d(s
s
s
(svariablesblock(R.s	block_endslinestatement_end(RGRERC(Rt
splitlinesR�tendswithRzR�RytAssertionErrorR�RUR�t
isinstanceR�t	enumeratet	__class__R�Rt	groupdicttcounttRuntimeErrorR�tignore_if_emptyRtpoptend(R�R�R(R�R�tlinestnewlinetposR�tstacktstatetokenst
source_lengthtbalancing_stacktregexttokenst	new_stateR�tidxRiRTRhR9texpected_optpos2((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyR�Qs�

		







N(	R�R�R�R�R�RyR�R�R�(((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyR��s	�	-(jR�RPR*RtcollectionsRtjinja2.exceptionsRtjinja2.utilsRtjinja2._compatRRRRRR�R�tUR�R�R�R�tSyntaxErrorR�tjinja2R
t	xid_starttxid_continueR�Rnt	TOKEN_ADDtTOKEN_ASSIGNtTOKEN_COLONtTOKEN_COMMAt	TOKEN_DIVt	TOKEN_DOTtTOKEN_EQtTOKEN_FLOORDIVtTOKEN_GTt
TOKEN_GTEQtTOKEN_LBRACEtTOKEN_LBRACKETtTOKEN_LPARENtTOKEN_LTt
TOKEN_LTEQt	TOKEN_MODt	TOKEN_MULtTOKEN_NEt
TOKEN_PIPEt	TOKEN_POWtTOKEN_RBRACEtTOKEN_RBRACKETtTOKEN_RPARENtTOKEN_SEMICOLONt	TOKEN_SUBtTOKEN_TILDER�R�R�R�R�R�R\R]R^R_R�R�RXRYRZR`RaR�R�R[RbR�RcR�tdicttktvRWRUR�R�R|R�t	frozensetR�RRfRjRmRpR�tobjectR�R�R�R�R�R�R�(((s0/usr/lib/python2.7/site-packages/jinja2/lexer.pyt<module>s�(	

1$ 							
+[	

Zerion Mini Shell 1.0