You can not select more than 25 topics
			Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
		
		
		
		
		
			
		
			
				
					
					
						
							184 lines
						
					
					
						
							6.6 KiB
						
					
					
				
			
		
		
	
	
							184 lines
						
					
					
						
							6.6 KiB
						
					
					
				| import os
 | |
| import time
 | |
| import threading
 | |
| import logging
 | |
| import json
 | |
| from pathlib import Path
 | |
| from openpilot.system.hardware.hw import Paths
 | |
| 
 | |
| from openpilot.common.swaglog import cloudlog
 | |
| from openpilot.system.loggerd.uploader import main, UPLOAD_ATTR_NAME, UPLOAD_ATTR_VALUE
 | |
| 
 | |
| from openpilot.system.loggerd.tests.loggerd_tests_common import UploaderTestCase
 | |
| 
 | |
| 
 | |
| class FakeLogHandler(logging.Handler):
 | |
|   def __init__(self):
 | |
|     logging.Handler.__init__(self)
 | |
|     self.reset()
 | |
| 
 | |
|   def reset(self):
 | |
|     self.upload_order = list()
 | |
|     self.upload_ignored = list()
 | |
| 
 | |
|   def emit(self, record):
 | |
|     try:
 | |
|       j = json.loads(record.getMessage())
 | |
|       if j["event"] == "upload_success":
 | |
|         self.upload_order.append(j["key"])
 | |
|       if j["event"] == "upload_ignored":
 | |
|         self.upload_ignored.append(j["key"])
 | |
|     except Exception:
 | |
|       pass
 | |
| 
 | |
| log_handler = FakeLogHandler()
 | |
| cloudlog.addHandler(log_handler)
 | |
| 
 | |
| 
 | |
| class TestUploader(UploaderTestCase):
 | |
|   def setup_method(self):
 | |
|     super().setup_method()
 | |
|     log_handler.reset()
 | |
| 
 | |
|   def start_thread(self):
 | |
|     self.end_event = threading.Event()
 | |
|     self.up_thread = threading.Thread(target=main, args=[self.end_event])
 | |
|     self.up_thread.daemon = True
 | |
|     self.up_thread.start()
 | |
| 
 | |
|   def join_thread(self):
 | |
|     self.end_event.set()
 | |
|     self.up_thread.join()
 | |
| 
 | |
|   def gen_files(self, lock=False, xattr: bytes = None, boot=True) -> list[Path]:
 | |
|     f_paths = []
 | |
|     for t in ["qlog", "rlog", "dcamera.hevc", "fcamera.hevc"]:
 | |
|       f_paths.append(self.make_file_with_data(self.seg_dir, t, 1, lock=lock, upload_xattr=xattr))
 | |
| 
 | |
|     if boot:
 | |
|       f_paths.append(self.make_file_with_data("boot", f"{self.seg_dir}", 1, lock=lock, upload_xattr=xattr))
 | |
|     return f_paths
 | |
| 
 | |
|   def gen_order(self, seg1: list[int], seg2: list[int], boot=True) -> list[str]:
 | |
|     keys = []
 | |
|     if boot:
 | |
|       keys += [f"boot/{self.seg_format.format(i)}.zst" for i in seg1]
 | |
|       keys += [f"boot/{self.seg_format2.format(i)}.zst" for i in seg2]
 | |
|     keys += [f"{self.seg_format.format(i)}/qlog.zst" for i in seg1]
 | |
|     keys += [f"{self.seg_format2.format(i)}/qlog.zst" for i in seg2]
 | |
|     return keys
 | |
| 
 | |
|   def test_upload(self):
 | |
|     self.gen_files(lock=False)
 | |
| 
 | |
|     self.start_thread()
 | |
|     # allow enough time that files could upload twice if there is a bug in the logic
 | |
|     time.sleep(1)
 | |
|     self.join_thread()
 | |
| 
 | |
|     exp_order = self.gen_order([self.seg_num], [])
 | |
| 
 | |
|     assert len(log_handler.upload_ignored) == 0, "Some files were ignored"
 | |
|     assert not len(log_handler.upload_order) < len(exp_order), "Some files failed to upload"
 | |
|     assert not len(log_handler.upload_order) > len(exp_order), "Some files were uploaded twice"
 | |
|     for f_path in exp_order:
 | |
|       assert os.getxattr((Path(Paths.log_root()) / f_path).with_suffix(""), UPLOAD_ATTR_NAME) == UPLOAD_ATTR_VALUE, "All files not uploaded"
 | |
| 
 | |
|     assert log_handler.upload_order == exp_order, "Files uploaded in wrong order"
 | |
| 
 | |
|   def test_upload_with_wrong_xattr(self):
 | |
|     self.gen_files(lock=False, xattr=b'0')
 | |
| 
 | |
|     self.start_thread()
 | |
|     # allow enough time that files could upload twice if there is a bug in the logic
 | |
|     time.sleep(1)
 | |
|     self.join_thread()
 | |
| 
 | |
|     exp_order = self.gen_order([self.seg_num], [])
 | |
| 
 | |
|     assert len(log_handler.upload_ignored) == 0, "Some files were ignored"
 | |
|     assert not len(log_handler.upload_order) < len(exp_order), "Some files failed to upload"
 | |
|     assert not len(log_handler.upload_order) > len(exp_order), "Some files were uploaded twice"
 | |
|     for f_path in exp_order:
 | |
|       assert os.getxattr((Path(Paths.log_root()) / f_path).with_suffix(""), UPLOAD_ATTR_NAME) == UPLOAD_ATTR_VALUE, "All files not uploaded"
 | |
| 
 | |
|     assert log_handler.upload_order == exp_order, "Files uploaded in wrong order"
 | |
| 
 | |
|   def test_upload_ignored(self):
 | |
|     self.set_ignore()
 | |
|     self.gen_files(lock=False)
 | |
| 
 | |
|     self.start_thread()
 | |
|     # allow enough time that files could upload twice if there is a bug in the logic
 | |
|     time.sleep(1)
 | |
|     self.join_thread()
 | |
| 
 | |
|     exp_order = self.gen_order([self.seg_num], [])
 | |
| 
 | |
|     assert len(log_handler.upload_order) == 0, "Some files were not ignored"
 | |
|     assert not len(log_handler.upload_ignored) < len(exp_order), "Some files failed to ignore"
 | |
|     assert not len(log_handler.upload_ignored) > len(exp_order), "Some files were ignored twice"
 | |
|     for f_path in exp_order:
 | |
|       assert os.getxattr((Path(Paths.log_root()) / f_path).with_suffix(""), UPLOAD_ATTR_NAME) == UPLOAD_ATTR_VALUE, "All files not ignored"
 | |
| 
 | |
|     assert log_handler.upload_ignored == exp_order, "Files ignored in wrong order"
 | |
| 
 | |
|   def test_upload_files_in_create_order(self):
 | |
|     seg1_nums = [0, 1, 2, 10, 20]
 | |
|     for i in seg1_nums:
 | |
|       self.seg_dir = self.seg_format.format(i)
 | |
|       self.gen_files(boot=False)
 | |
|     seg2_nums = [5, 50, 51]
 | |
|     for i in seg2_nums:
 | |
|       self.seg_dir = self.seg_format2.format(i)
 | |
|       self.gen_files(boot=False)
 | |
| 
 | |
|     exp_order = self.gen_order(seg1_nums, seg2_nums, boot=False)
 | |
| 
 | |
|     self.start_thread()
 | |
|     # allow enough time that files could upload twice if there is a bug in the logic
 | |
|     time.sleep(1)
 | |
|     self.join_thread()
 | |
| 
 | |
|     assert len(log_handler.upload_ignored) == 0, "Some files were ignored"
 | |
|     assert not len(log_handler.upload_order) < len(exp_order), "Some files failed to upload"
 | |
|     assert not len(log_handler.upload_order) > len(exp_order), "Some files were uploaded twice"
 | |
|     for f_path in exp_order:
 | |
|       assert os.getxattr((Path(Paths.log_root()) / f_path).with_suffix(""), UPLOAD_ATTR_NAME) == UPLOAD_ATTR_VALUE, "All files not uploaded"
 | |
| 
 | |
|     assert log_handler.upload_order == exp_order, "Files uploaded in wrong order"
 | |
| 
 | |
|   def test_no_upload_with_lock_file(self):
 | |
|     self.start_thread()
 | |
| 
 | |
|     time.sleep(0.25)
 | |
|     f_paths = self.gen_files(lock=True, boot=False)
 | |
| 
 | |
|     # allow enough time that files should have been uploaded if they would be uploaded
 | |
|     time.sleep(1)
 | |
|     self.join_thread()
 | |
| 
 | |
|     for f_path in f_paths:
 | |
|       fn = f_path.with_suffix(f_path.suffix.replace(".zst", ""))
 | |
|       uploaded = UPLOAD_ATTR_NAME in os.listxattr(fn) and os.getxattr(fn, UPLOAD_ATTR_NAME) == UPLOAD_ATTR_VALUE
 | |
|       assert not uploaded, "File upload when locked"
 | |
| 
 | |
|   def test_no_upload_with_xattr(self):
 | |
|     self.gen_files(lock=False, xattr=UPLOAD_ATTR_VALUE)
 | |
| 
 | |
|     self.start_thread()
 | |
|     # allow enough time that files could upload twice if there is a bug in the logic
 | |
|     time.sleep(1)
 | |
|     self.join_thread()
 | |
| 
 | |
|     assert len(log_handler.upload_order) == 0, "File uploaded again"
 | |
| 
 | |
|   def test_clear_locks_on_startup(self):
 | |
|     f_paths = self.gen_files(lock=True, boot=False)
 | |
|     self.start_thread()
 | |
|     time.sleep(0.25)
 | |
|     self.join_thread()
 | |
| 
 | |
|     for f_path in f_paths:
 | |
|       lock_path = f_path.with_suffix(f_path.suffix + ".lock")
 | |
|       assert not lock_path.is_file(), "File lock not cleared on startup"
 | |
| 
 |