ACIL FM
Dark
Refresh
Current DIR:
/opt/alt/python313/lib64/python3.13/test/test_free_threading
/
opt
alt
python313
lib64
python3.13
test
test_free_threading
Upload
Zip Selected
Delete Selected
Pilih semua
Nama
Ukuran
Permission
Aksi
__pycache__
-
chmod
Open
Rename
Delete
test_code.py
912 B
chmod
View
DL
Edit
Rename
Delete
test_dict.py
5.44 MB
chmod
View
DL
Edit
Rename
Delete
test_func_annotations.py
2.55 MB
chmod
View
DL
Edit
Rename
Delete
test_gc.py
1.45 MB
chmod
View
DL
Edit
Rename
Delete
test_list.py
1.77 MB
chmod
View
DL
Edit
Rename
Delete
test_monitoring.py
6.42 MB
chmod
View
DL
Edit
Rename
Delete
test_set.py
1.08 MB
chmod
View
DL
Edit
Rename
Delete
test_slots.py
1001 B
chmod
View
DL
Edit
Rename
Delete
test_str.py
1.97 MB
chmod
View
DL
Edit
Rename
Delete
test_tokenize.py
2.16 MB
chmod
View
DL
Edit
Rename
Delete
test_type.py
3.57 MB
chmod
View
DL
Edit
Rename
Delete
__init__.py
225 B
chmod
View
DL
Edit
Rename
Delete
Edit file: /opt/alt/python313/lib64/python3.13/test/test_free_threading/test_tokenize.py
import io import time import unittest import tokenize from functools import partial from threading import Thread from test.support import threading_helper @threading_helper.requires_working_threading() class TestTokenize(unittest.TestCase): def test_tokenizer_iter(self): source = io.StringIO("for _ in a:\n pass") it = tokenize._tokenize.TokenizerIter(source.readline, extra_tokens=False) tokens = [] def next_token(it): while True: try: r = next(it) tokens.append(tokenize.TokenInfo._make(r)) time.sleep(0.03) except StopIteration: return threads = [] for _ in range(5): threads.append(Thread(target=partial(next_token, it))) for thread in threads: thread.start() for thread in threads: thread.join() expected_tokens = [ tokenize.TokenInfo(type=1, string='for', start=(1, 0), end=(1, 3), line='for _ in a:\n'), tokenize.TokenInfo(type=1, string='_', start=(1, 4), end=(1, 5), line='for _ in a:\n'), tokenize.TokenInfo(type=1, string='in', start=(1, 6), end=(1, 8), line='for _ in a:\n'), tokenize.TokenInfo(type=1, string='a', start=(1, 9), end=(1, 10), line='for _ in a:\n'), tokenize.TokenInfo(type=11, string=':', start=(1, 10), end=(1, 11), line='for _ in a:\n'), tokenize.TokenInfo(type=4, string='', start=(1, 11), end=(1, 11), line='for _ in a:\n'), tokenize.TokenInfo(type=5, string='', start=(2, -1), end=(2, -1), line=' pass'), tokenize.TokenInfo(type=1, string='pass', start=(2, 2), end=(2, 6), line=' pass'), tokenize.TokenInfo(type=4, string='', start=(2, 6), end=(2, 6), line=' pass'), tokenize.TokenInfo(type=6, string='', start=(2, -1), end=(2, -1), line=' pass'), tokenize.TokenInfo(type=0, string='', start=(2, -1), end=(2, -1), line=' pass'), ] tokens.sort() expected_tokens.sort() self.assertListEqual(tokens, expected_tokens) if __name__ == "__main__": unittest.main()
Simpan
Batal
Isi Zip:
Unzip
Create
Buat Folder
Buat File
Terminal / Execute
Run
Chmod Bulk
All File
All Folder
All File dan Folder
Apply