]> git.ipfire.org Git - thirdparty/Python/cpython.git/commitdiff
bpo-36654: Add examples for using tokenize module programmatically (GH-12947)
authorMiss Islington (bot) <31488909+miss-islington@users.noreply.github.com>
Sat, 25 Jan 2020 19:36:04 +0000 (11:36 -0800)
committerBerker Peksag <berker.peksag@gmail.com>
Sat, 25 Jan 2020 19:36:04 +0000 (22:36 +0300)
(cherry picked from commit 4b09dc79f4d08d85f2cc945563e9c8ef1e531d7b)

Co-authored-by: Windson yang <wiwindson@outlook.com>
Doc/library/tokenize.rst

index 4c0a0ceef7dc4e7bfac203d4dca8937bac34992f..4dd56f9e7c8bf6ff58b582d20f8eba8231421577 100644 (file)
@@ -267,3 +267,22 @@ The exact token type names can be displayed using the :option:`-e` option:
     4,10-4,11:          RPAR           ')'
     4,11-4,12:          NEWLINE        '\n'
     5,0-5,0:            ENDMARKER      ''
+
+Example of tokenizing a file programmatically, reading unicode
+strings instead of bytes with :func:`generate_tokens`::
+
+    import tokenize
+
+    with tokenize.open('hello.py') as f:
+        tokens = tokenize.generate_tokens(f.readline)
+        for token in tokens:
+            print(token)
+
+Or reading bytes directly with :func:`.tokenize`::
+
+    import tokenize
+
+    with open('hello.py', 'rb') as f:
+        tokens = tokenize.tokenize(f.readline)
+        for token in tokens:
+            print(token)