File: //lib64/python3.6/test/__pycache__/test_tokenize.cpython-36.pyc
3
\k� � @ s: d dl mZ d dlmZmZmZmZmZmZmZm Z m
Z
mZmZm
ZmZmZmZ d dlmZ d dlZd dlmZmZ d dlmZmZ d dlZd dlZdd� ZG d d
� d
e�Zdd� ZG d
d� de�ZG dd� de�Z G dd� de�Z!G dd� de�Z"G dd� de�Z#G dd� de�Z$G dd� de�Z%e&dk�r6ej'� dS )� )�support)�tokenize� _tokenize�
untokenize�NUMBER�NAME�OP�STRING� ENDMARKER�ENCODING�tok_name�detect_encoding�open�Untokenizer�generate_tokens�NEWLINE)�BytesION)�TestCase�mock)�VALID_UNDERSCORE_LITERALS�INVALID_UNDERSCORE_LITERALSc
C s� g }t |j� �}|d dk}xh| D ]`\}}}}} |tkr:P |rT|tkrT|d |krTq"t| }|jd|d�d|d�d|� d|� �� q"W |S ) N� z
r z Z10� Z13���)�len�
splitlinesr
r r �append)
Ztoken_generatorZ
source_string�resultZ num_linesZmissing_trailing_nl�type�token�start�end�line� r# �%/usr/lib64/python3.6/test_tokenize.py�stringify_tokens_from_source s *r% c @ s� e Zd Zdd� Zdd� Zdd� Zdd� Zd d
� Zdd� Zd
d� Z dd� Z
dd� Zdd� Zdd� Z
dd� Zdd� Zdd� Zdd� Zdd � Zd!d"� Zd#d$� Zd%d&� Zd'd(� Zd)S )*�TokenizeTestc C s<