You can not select more than 25 topics
			Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
		
		
		
		
		
			
		
			
				
					
					
						
							1284 lines
						
					
					
						
							37 KiB
						
					
					
				
			
		
		
	
	
							1284 lines
						
					
					
						
							37 KiB
						
					
					
				| import os
 | |
| import sys
 | |
| import glob
 | |
| import json
 | |
| import time
 | |
| import struct
 | |
| import tempfile
 | |
| import threading
 | |
| import xml.etree.ElementTree as ET
 | |
| import numpy as np
 | |
| if sys.version_info >= (3,0):
 | |
|   import queue
 | |
|   import pickle
 | |
|   from io import BytesIO as StringIO
 | |
| else:
 | |
|   import Queue as queue
 | |
|   import cPickle as pickle
 | |
|   from cStringIO import StringIO
 | |
| 
 | |
| import subprocess
 | |
| from aenum import Enum
 | |
| from lru import LRU
 | |
| from functools import wraps
 | |
| from concurrent.futures import ThreadPoolExecutor, as_completed
 | |
| 
 | |
| from tools.lib.cache import cache_path_for_file_path
 | |
| from tools.lib.exceptions import DataUnreadableError
 | |
| try:
 | |
|   from xx.chffr.lib.filereader import FileReader
 | |
| except ImportError:
 | |
|   from tools.lib.filereader import FileReader
 | |
| from tools.lib.file_helpers import atomic_write_in_dir
 | |
| from tools.lib.mkvparse import mkvindex
 | |
| from tools.lib.route import Route
 | |
| 
 | |
| H264_SLICE_P = 0
 | |
| H264_SLICE_B = 1
 | |
| H264_SLICE_I = 2
 | |
| 
 | |
| HEVC_SLICE_B = 0
 | |
| HEVC_SLICE_P = 1
 | |
| HEVC_SLICE_I = 2
 | |
| 
 | |
| SLICE_I = 2 # hevc and h264 are the same :)
 | |
| 
 | |
| class FrameType(Enum):
 | |
|   raw = 1
 | |
|   h265_stream = 2
 | |
|   h264_mp4 = 3
 | |
|   h264_pstream = 4
 | |
|   ffv1_mkv = 5
 | |
|   ffvhuff_mkv = 6
 | |
| 
 | |
| def fingerprint_video(fn):
 | |
|   with FileReader(fn) as f:
 | |
|     header = f.read(4)
 | |
|   if len(header) == 0:
 | |
|     raise DataUnreadableError("%s is empty" % fn)
 | |
|   elif header == b"\x00\xc0\x12\x00":
 | |
|     return FrameType.raw
 | |
|   elif header == b"\x00\x00\x00\x01":
 | |
|     if 'hevc' in fn:
 | |
|       return FrameType.h265_stream
 | |
|     elif os.path.basename(fn) in ("camera", "acamera"):
 | |
|       return FrameType.h264_pstream
 | |
|     else:
 | |
|       raise NotImplementedError(fn)
 | |
|   elif header == b"\x00\x00\x00\x1c":
 | |
|     return FrameType.h264_mp4
 | |
|   elif header == b"\x1a\x45\xdf\xa3":
 | |
|     return FrameType.ffv1_mkv
 | |
|   else:
 | |
|     raise NotImplementedError(fn)
 | |
| 
 | |
| 
 | |
| def ffprobe(fn, fmt=None):
 | |
|   cmd = ["ffprobe",
 | |
|     "-v", "quiet",
 | |
|     "-print_format", "json",
 | |
|     "-show_format", "-show_streams"]
 | |
|   if fmt:
 | |
|     cmd += ["-format", fmt]
 | |
|   cmd += [fn]
 | |
| 
 | |
|   try:
 | |
|     ffprobe_output = subprocess.check_output(cmd)
 | |
|   except subprocess.CalledProcessError as e:
 | |
|     raise DataUnreadableError(fn)
 | |
| 
 | |
|   return json.loads(ffprobe_output)
 | |
| 
 | |
| 
 | |
| def vidindex(fn, typ):
 | |
|   vidindex_dir = os.path.join(os.path.dirname(os.path.realpath(__file__)), "vidindex")
 | |
|   vidindex = os.path.join(vidindex_dir, "vidindex")
 | |
| 
 | |
|   subprocess.check_call(["make"], cwd=vidindex_dir, stdout=open("/dev/null","w"))
 | |
| 
 | |
|   with tempfile.NamedTemporaryFile() as prefix_f, \
 | |
|        tempfile.NamedTemporaryFile() as index_f:
 | |
|     try:
 | |
|       subprocess.check_call([vidindex, typ, fn, prefix_f.name, index_f.name])
 | |
|     except subprocess.CalledProcessError as e:
 | |
|       raise DataUnreadableError("vidindex failed on file %s" % fn)
 | |
|     with open(index_f.name, "rb") as f:
 | |
|       index = f.read()
 | |
|     with open(prefix_f.name, "rb") as f:
 | |
|       prefix = f.read()
 | |
| 
 | |
|   index = np.frombuffer(index, np.uint32).reshape(-1, 2)
 | |
| 
 | |
|   assert index[-1, 0] == 0xFFFFFFFF
 | |
|   assert index[-1, 1] == os.path.getsize(fn)
 | |
| 
 | |
|   return index, prefix
 | |
| 
 | |
| 
 | |
| def cache_fn(func):
 | |
|   @wraps(func)
 | |
|   def cache_inner(fn, *args, **kwargs):
 | |
|     cache_prefix = kwargs.pop('cache_prefix', None)
 | |
|     cache_path = cache_path_for_file_path(fn, cache_prefix)
 | |
| 
 | |
|     if cache_path and os.path.exists(cache_path):
 | |
|       with open(cache_path, "rb") as cache_file:
 | |
|         cache_value = pickle.load(cache_file)
 | |
|     else:
 | |
|       cache_value = func(fn, *args, **kwargs)
 | |
| 
 | |
|       if cache_path:
 | |
|         with atomic_write_in_dir(cache_path, mode="wb", overwrite=True) as cache_file:
 | |
|           pickle.dump(cache_value, cache_file, -1)
 | |
| 
 | |
|     return cache_value
 | |
| 
 | |
|   return cache_inner
 | |
| 
 | |
| @cache_fn
 | |
| def index_stream(fn, typ):
 | |
|   assert typ in ("hevc", "h264")
 | |
| 
 | |
|   with FileReader(fn) as f:
 | |
|     assert os.path.exists(f.name), fn
 | |
|     index, prefix = vidindex(f.name, typ)
 | |
|     probe = ffprobe(f.name, typ)
 | |
| 
 | |
|   return {
 | |
|     'index': index,
 | |
|     'global_prefix': prefix,
 | |
|     'probe': probe
 | |
|   }
 | |
| 
 | |
| @cache_fn
 | |
| def index_mp4(fn):
 | |
|   with FileReader(fn) as f:
 | |
|     return vidindex_mp4(f.name)
 | |
| 
 | |
| @cache_fn
 | |
| def index_mkv(fn):
 | |
|   with FileReader(fn) as f:
 | |
|     probe = ffprobe(f.name, "matroska")
 | |
|     with open(f.name, "rb") as d_f:
 | |
|       config_record, index = mkvindex.mkvindex(d_f)
 | |
|   return {
 | |
|     'probe': probe,
 | |
|     'config_record': config_record,
 | |
|     'index': index
 | |
|   }
 | |
| 
 | |
| def index_videos(camera_paths, cache_prefix=None):
 | |
|   """Requires that paths in camera_paths are contiguous and of the same type."""
 | |
|   if len(camera_paths) < 1:
 | |
|     raise ValueError("must provide at least one video to index")
 | |
| 
 | |
|   frame_type = fingerprint_video(camera_paths[0])
 | |
|   if frame_type == FrameType.h264_pstream:
 | |
|     index_pstream(camera_paths, "h264", cache_prefix)
 | |
|   else:
 | |
|     for fn in camera_paths:
 | |
|       index_video(fn, frame_type, cache_prefix)
 | |
| 
 | |
| def index_video(fn, frame_type=None, cache_prefix=None):
 | |
|   cache_path = cache_path_for_file_path(fn, cache_prefix)
 | |
| 
 | |
|   if os.path.exists(cache_path):
 | |
|     return
 | |
| 
 | |
|   if frame_type is None:
 | |
|     frame_type = fingerprint_video(fn[0])
 | |
| 
 | |
|   if frame_type == FrameType.h264_pstream:
 | |
|     #hack: try to index the whole route now
 | |
|     route = Route.from_file_path(fn)
 | |
| 
 | |
|     camera_paths = route.camera_paths()
 | |
|     if fn not in camera_paths:
 | |
|       raise DataUnreadableError("Not a contiguous route camera file: {}".format(fn))
 | |
| 
 | |
|     print("no pstream cache for %s, indexing route %s now" % (fn, route.name))
 | |
|     index_pstream(route.camera_paths(), "h264", cache_prefix)
 | |
|   elif frame_type == FrameType.h265_stream:
 | |
|     index_stream(fn, "hevc", cache_prefix=cache_prefix)
 | |
|   elif frame_type == FrameType.h264_mp4:
 | |
|     index_mp4(fn, cache_prefix=cache_prefix)
 | |
| 
 | |
| def get_video_index(fn, frame_type, cache_prefix=None):
 | |
|   cache_path = cache_path_for_file_path(fn, cache_prefix)
 | |
| 
 | |
|   if not os.path.exists(cache_path):
 | |
|     index_video(fn, frame_type, cache_prefix)
 | |
| 
 | |
|   if not os.path.exists(cache_path):
 | |
|     return None
 | |
|   with open(cache_path, "rb") as cache_file:
 | |
|     return pickle.load(cache_file)
 | |
| 
 | |
| def pstream_predecompress(fns, probe, indexes, global_prefix, cache_prefix, multithreaded=False):
 | |
|   assert len(fns) == len(indexes)
 | |
|   out_fns = [cache_path_for_file_path(fn, cache_prefix, extension=".predecom.mkv") for fn in fns]
 | |
|   out_exists = map(os.path.exists, out_fns)
 | |
|   if all(out_exists):
 | |
|     return
 | |
| 
 | |
|   w = probe['streams'][0]['width']
 | |
|   h = probe['streams'][0]['height']
 | |
| 
 | |
|   frame_size = w*h*3/2 # yuv420p
 | |
| 
 | |
|   decompress_proc = subprocess.Popen(
 | |
|     ["ffmpeg",
 | |
|      "-threads", "0" if multithreaded else "1",
 | |
|      "-vsync", "0",
 | |
|      "-f", "h264",
 | |
|      "-i", "pipe:0",
 | |
|      "-threads", "0" if multithreaded else "1",
 | |
|      "-f", "rawvideo",
 | |
|      "-pix_fmt", "yuv420p",
 | |
|      "pipe:1"],
 | |
|     stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=open("/dev/null", "wb"))
 | |
| 
 | |
|   def write_thread():
 | |
|     for fn in fns:
 | |
|       with FileReader(fn) as f:
 | |
|         decompress_proc.stdin.write(f.read())
 | |
|     decompress_proc.stdin.close()
 | |
| 
 | |
|   def read_frame():
 | |
|     frame = None
 | |
|     try:
 | |
|       frame = decompress_proc.stdout.read(frame_size)
 | |
|     except (IOError, ValueError):
 | |
|       pass
 | |
|     if frame is None or frame == "" or len(frame) != frame_size:
 | |
|       raise DataUnreadableError("pre-decompression failed for %s" % fn)
 | |
|     return frame
 | |
| 
 | |
|   t = threading.Thread(target=write_thread)
 | |
|   t.daemon = True
 | |
|   t.start()
 | |
| 
 | |
|   try:
 | |
|     for fn, out_fn, out_exist, index in zip(fns, out_fns, out_exists, indexes):
 | |
|       if out_exist:
 | |
|         for fi in range(index.shape[0]-1):
 | |
|           read_frame()
 | |
|         continue
 | |
| 
 | |
|       with atomic_write_in_dir(out_fn, mode="w+b", overwrite=True) as out_tmp:
 | |
|         compress_proc = subprocess.Popen(
 | |
|           ["ffmpeg",
 | |
|           "-threads", "0" if multithreaded else "1",
 | |
|            "-y",
 | |
|            "-vsync", "0",
 | |
|            "-f", "rawvideo",
 | |
|            "-pix_fmt", "yuv420p",
 | |
|            "-s", "%dx%d" % (w, h),
 | |
|            "-i", "pipe:0",
 | |
|           "-threads", "0" if multithreaded else "1",
 | |
|            "-f", "matroska",
 | |
|            "-vcodec", "ffv1",
 | |
|            "-g", "0",
 | |
|            out_tmp.name],
 | |
|           stdin=subprocess.PIPE, stderr=open("/dev/null", "wb"))
 | |
|         try:
 | |
|           for fi in range(index.shape[0]-1):
 | |
|             frame = read_frame()
 | |
|             compress_proc.stdin.write(frame)
 | |
|           compress_proc.stdin.close()
 | |
|         except:
 | |
|           compress_proc.kill()
 | |
|           raise
 | |
| 
 | |
|         assert compress_proc.wait() == 0
 | |
| 
 | |
|       cache_path = cache_path_for_file_path(fn, cache_prefix)
 | |
|       with atomic_write_in_dir(cache_path, mode="wb", overwrite=True) as cache_file:
 | |
|         pickle.dump({
 | |
|           'predecom': os.path.basename(out_fn),
 | |
|           'index': index,
 | |
|           'probe': probe,
 | |
|           'global_prefix': global_prefix,
 | |
|         }, cache_file, -1)
 | |
| 
 | |
|   except:
 | |
|     decompress_proc.kill()
 | |
|     raise
 | |
|   finally:
 | |
|     t.join()
 | |
| 
 | |
|   rc = decompress_proc.wait()
 | |
|   if rc != 0:
 | |
|     raise DataUnreadableError(fns[0])
 | |
| 
 | |
| 
 | |
| def index_pstream(fns, typ, cache_prefix=None):
 | |
|   if typ != "h264":
 | |
|     raise NotImplementedError(typ)
 | |
| 
 | |
|   if not fns:
 | |
|     raise DataUnreadableError("chffr h264 requires contiguous files")
 | |
| 
 | |
|   out_fns = [cache_path_for_file_path(fn, cache_prefix) for fn in fns]
 | |
|   out_exists = map(os.path.exists, out_fns)
 | |
|   if all(out_exists): return
 | |
| 
 | |
|   # load existing index files to avoid re-doing work
 | |
|   existing_indexes = []
 | |
|   for out_fn, exists in zip(out_fns, out_exists):
 | |
|     existing = None
 | |
|     if exists:
 | |
|       with open(out_fn, "rb") as cache_file:
 | |
|         existing = pickle.load(cache_file)
 | |
|     existing_indexes.append(existing)
 | |
| 
 | |
|   # probe the first file
 | |
|   if existing_indexes[0]:
 | |
|     probe = existing_indexes[0]['probe']
 | |
|   else:
 | |
|     with FileReader(fns[0]) as f:
 | |
|       probe = ffprobe(f.name, typ)
 | |
| 
 | |
|   global_prefix = None
 | |
| 
 | |
|   # get the video index of all the segments in this stream
 | |
|   indexes = []
 | |
|   for i, fn in enumerate(fns):
 | |
|     if existing_indexes[i]:
 | |
|       index = existing_indexes[i]['index']
 | |
|       prefix = existing_indexes[i]['global_prefix']
 | |
|     else:
 | |
|       with FileReader(fn) as f:
 | |
|         index, prefix = vidindex(f.name, typ)
 | |
|     if i == 0:
 | |
|       # assert prefix
 | |
|       if not prefix:
 | |
|         raise DataUnreadableError("vidindex failed for %s" % fn)
 | |
|       global_prefix = prefix
 | |
|     indexes.append(index)
 | |
| 
 | |
|   assert global_prefix
 | |
| 
 | |
|   if np.sum(indexes[0][:, 0] == H264_SLICE_I) <= 1:
 | |
|     print("pstream %s is unseekable. pre-decompressing all the segments..." % (fns[0]))
 | |
|     pstream_predecompress(fns, probe, indexes, global_prefix, cache_prefix)
 | |
|     return
 | |
| 
 | |
|   # generate what's required to make each segment self-contained
 | |
|   # (the partial GOP from the end of each segments are put asside to add
 | |
|   #  to the start of the following segment)
 | |
|   prefix_data = ["" for _ in fns]
 | |
|   prefix_index = [[] for _ in fns]
 | |
|   for i in range(len(fns)-1):
 | |
|     if indexes[i+1][0, 0] == H264_SLICE_I and indexes[i+1][0, 1] <= 1:
 | |
|       # next file happens to start with a i-frame, dont need use this file's end
 | |
|       continue
 | |
| 
 | |
|     index = indexes[i]
 | |
|     if i == 0 and np.sum(index[:, 0] == H264_SLICE_I) <= 1:
 | |
|       raise NotImplementedError("No I-frames in pstream.")
 | |
| 
 | |
|     # find the last GOP in the index
 | |
|     frame_b = len(index)-1
 | |
|     while frame_b > 0 and index[frame_b, 0] != H264_SLICE_I:
 | |
|       frame_b -= 1
 | |
| 
 | |
|     assert frame_b >= 0
 | |
|     assert index[frame_b, 0] == H264_SLICE_I
 | |
| 
 | |
|     end_len = len(index)-frame_b
 | |
| 
 | |
|     with FileReader(fns[i]) as vid:
 | |
|       vid.seek(index[frame_b, 1])
 | |
|       end_data = vid.read()
 | |
| 
 | |
|     prefix_data[i+1] = end_data
 | |
|     prefix_index[i+1] = index[frame_b:-1]
 | |
|     # indexes[i] = index[:frame_b]
 | |
| 
 | |
|   for i, fn in enumerate(fns):
 | |
|     cache_path = out_fns[i]
 | |
| 
 | |
|     if os.path.exists(cache_path):
 | |
|       continue
 | |
| 
 | |
|     segment_index = {
 | |
|       'index': indexes[i],
 | |
|       'global_prefix': global_prefix,
 | |
|       'probe': probe,
 | |
|       'prefix_frame_data': prefix_data[i], # data to prefix the first GOP with
 | |
|       'num_prefix_frames': len(prefix_index[i]), # number of frames to skip in the first GOP
 | |
|     }
 | |
| 
 | |
|     with atomic_write_in_dir(cache_path, mode="wb", overwrite=True) as cache_file:
 | |
|       pickle.dump(segment_index, cache_file, -1)
 | |
| 
 | |
| def gpu_info():
 | |
|   ret = []
 | |
|   for fn in glob.glob("/proc/driver/nvidia/gpus/*/information"):
 | |
|     with open(fn, "r") as f:
 | |
|       dat = f.read()
 | |
|       kvs = dat.strip().split("\n")
 | |
|       kv = {}
 | |
|       for s in kvs:
 | |
|         k, v = s.split(":", 1)
 | |
|         kv[k] = v.strip()
 | |
|       ret.append(kv)
 | |
|   return ret
 | |
| 
 | |
| def gpu_supports_hevc(gpuinfo):
 | |
|   return ("GTX 10" in gpuinfo['Model'] or "GTX 20" in gpuinfo['Model'] or gpuinfo['Model'] == "Graphics Device")
 | |
| 
 | |
| def find_hevc_gpu():
 | |
|   for gpuinfo in gpu_info():
 | |
|     if gpu_supports_hevc(gpuinfo):
 | |
|       return int(gpuinfo['Device Minor'])
 | |
|   return None
 | |
| 
 | |
| def _ffmpeg_fcamera_input_for_frame_info(frame_info):
 | |
|   st = time.time()
 | |
|   fn, num, count, cache_prefix = frame_info
 | |
| 
 | |
|   assert fn.endswith('.hevc')
 | |
|   sindex = index_stream(fn, "hevc", cache_prefix=cache_prefix)
 | |
|   index = sindex['index']
 | |
|   prefix = sindex['global_prefix']
 | |
|   probe = sindex['probe']
 | |
| 
 | |
|   frame_e = num + count
 | |
|   frame_b = num
 | |
|   # must start decoding on an i-frame
 | |
|   while index[frame_b, 0] != HEVC_SLICE_I:
 | |
|     frame_b -= 1
 | |
|   offset_b = index[frame_b, 1]
 | |
|   offset_e = index[frame_e, 1]
 | |
|   assert frame_b <= num < frame_e
 | |
|   skip = num - frame_b
 | |
| 
 | |
|   w = probe['streams'][0]['width']
 | |
|   h = probe['streams'][0]['height']
 | |
|   assert (h, w) == (874, 1164)
 | |
| 
 | |
|   st2 = time.time()
 | |
|   with FileReader(fn) as f:
 | |
|     f.seek(offset_b)
 | |
|     input_data = f.read(offset_e - offset_b)
 | |
|   et = time.time()
 | |
| 
 | |
|   get_time = et-st
 | |
|   get_time2 = et-st2
 | |
| 
 | |
|   if get_time > 10.0:
 | |
|     print("TOOK OVER 10 seconds to fetch %r %f %f" % (frame_info, get_time, get_time2))
 | |
| 
 | |
|   return prefix, input_data, skip, count
 | |
| 
 | |
| def _ffmpeg_fcamera_input_for_frame(pair):
 | |
|   cookie, frame_info = pair
 | |
|   try:
 | |
|     return cookie, _ffmpeg_fcamera_input_for_frame_info(frame_info)
 | |
|   except Exception as e:
 | |
|     # Let the caller handle exceptions.
 | |
|     return cookie, e
 | |
| 
 | |
| 
 | |
| def _feed_ffmpeg_fcamera_input_work_loop(frames, proc_stdin, select_pipe_fd, cookie_queue):
 | |
|   last_prefix = None
 | |
|   """
 | |
|   with ThreadPoolExecutor(64) as pool:
 | |
|     futures = []
 | |
|     for f in frames:
 | |
|       futures.append(pool.submit(_ffmpeg_fcamera_input_for_frame, f))
 | |
|     for f in as_completed(futures):
 | |
|       cookie, data = f.result()
 | |
|       if isinstance(data, Exception):
 | |
|         # Just print exceptions for now.
 | |
|         print(data)
 | |
|         continue
 | |
|       prefix, input_data, skip, count = data
 | |
|       cookie_queue.put((cookie, count))
 | |
| 
 | |
|       # Write zeros for skipped frames, ones for keep frames.
 | |
|       os.write(select_pipe_fd, b"\x00" * skip + b"\x01" * count)
 | |
| 
 | |
|       if prefix != last_prefix:
 | |
|         proc_stdin.write(prefix)
 | |
|         last_prefix = prefix
 | |
| 
 | |
|       proc_stdin.write(input_data)
 | |
|   """
 | |
|   num_threads = 64
 | |
|   for cookie, data in async_generator(
 | |
|       num_threads, 8 * num_threads, 8 * num_threads,
 | |
|       reliable=False)(_ffmpeg_fcamera_input_for_frame)(frames):
 | |
|     if isinstance(data, Exception):
 | |
|       # Just print exceptions for now.
 | |
|       print(data)
 | |
|       continue
 | |
|     prefix, input_data, skip, count = data
 | |
|     cookie_queue.put((cookie, count))
 | |
| 
 | |
|     # Write zeros for skipped frames, ones for keep frames.
 | |
|     os.write(select_pipe_fd, b"\x00" * skip + b"\x01" * count)
 | |
| 
 | |
|     if prefix != last_prefix:
 | |
|       proc_stdin.write(prefix)
 | |
|       last_prefix = prefix
 | |
| 
 | |
|     proc_stdin.write(input_data)
 | |
| 
 | |
| _FCAMERA_FEED_SUCCESS = object()
 | |
| def feed_ffmpeg_fcamera_input(frames, proc_stdin, select_pipe_fd, cookie_queue):
 | |
|   print("Feed started on {}".format(threading.current_thread().name))
 | |
|   try:
 | |
|     _feed_ffmpeg_fcamera_input_work_loop(frames, proc_stdin, select_pipe_fd, cookie_queue)
 | |
|     cookie_queue.put((_FCAMERA_FEED_SUCCESS, None))
 | |
|   finally:
 | |
|     # Always close ffmpeg input.
 | |
|     proc_stdin.close()
 | |
| 
 | |
| 
 | |
| def read_file_check_size(f, sz, cookie):
 | |
|   buff = bytearray(sz)
 | |
|   bytes_read = f.readinto(buff)
 | |
|   assert bytes_read == sz, (bytes_read, sz)
 | |
|   return buff
 | |
| 
 | |
| 
 | |
| import signal
 | |
| import ctypes
 | |
| def _set_pdeathsig(sig=signal.SIGTERM):
 | |
|   def f():
 | |
|     libc = ctypes.CDLL('libc.so.6')
 | |
|     return libc.prctl(1, sig)
 | |
|   return f
 | |
| 
 | |
| def vidindex_mp4(fn):
 | |
|   try:
 | |
|     xmls = subprocess.check_output(["MP4Box", fn, "-diso", "-out", "/dev/stdout"])
 | |
|   except subprocess.CalledProcessError as e:
 | |
|     raise DataUnreadableError(fn)
 | |
| 
 | |
|   tree = ET.fromstring(xmls)
 | |
| 
 | |
|   def parse_content(s):
 | |
|     assert s.startswith("data:application/octet-string,")
 | |
|     return s[len("data:application/octet-string,"):].decode("hex")
 | |
| 
 | |
|   avc_element = tree.find(".//AVCSampleEntryBox")
 | |
|   width = int(avc_element.attrib['Width'])
 | |
|   height = int(avc_element.attrib['Height'])
 | |
| 
 | |
|   sps_element = avc_element.find(".//AVCDecoderConfigurationRecord/SequenceParameterSet")
 | |
|   pps_element = avc_element.find(".//AVCDecoderConfigurationRecord/PictureParameterSet")
 | |
| 
 | |
|   sps = parse_content(sps_element.attrib['content'])
 | |
|   pps = parse_content(pps_element.attrib['content'])
 | |
| 
 | |
|   media_header = tree.find("MovieBox/TrackBox/MediaBox/MediaHeaderBox")
 | |
|   time_scale = int(media_header.attrib['TimeScale'])
 | |
| 
 | |
|   sample_sizes = [
 | |
|     int(entry.attrib['Size']) for entry in tree.findall(
 | |
|       "MovieBox/TrackBox/MediaBox/MediaInformationBox/SampleTableBox/SampleSizeBox/SampleSizeEntry")
 | |
|   ]
 | |
| 
 | |
|   sample_dependency = [
 | |
|     entry.attrib['dependsOnOther'] == "yes" for entry in tree.findall(
 | |
|       "MovieBox/TrackBox/MediaBox/MediaInformationBox/SampleTableBox/SampleDependencyTypeBox/SampleDependencyEntry")
 | |
|   ]
 | |
| 
 | |
|   assert len(sample_sizes) == len(sample_dependency)
 | |
| 
 | |
|   chunk_offsets = [
 | |
|     int(entry.attrib['offset']) for entry in tree.findall(
 | |
|       "MovieBox/TrackBox/MediaBox/MediaInformationBox/SampleTableBox/ChunkOffsetBox/ChunkEntry")
 | |
|   ]
 | |
| 
 | |
|   sample_chunk_table = [
 | |
|     (int(entry.attrib['FirstChunk'])-1, int(entry.attrib['SamplesPerChunk'])) for entry in tree.findall(
 | |
|       "MovieBox/TrackBox/MediaBox/MediaInformationBox/SampleTableBox/SampleToChunkBox/SampleToChunkEntry")
 | |
|   ]
 | |
| 
 | |
|   sample_offsets = [None for _ in sample_sizes]
 | |
| 
 | |
|   sample_i = 0
 | |
|   for i, (first_chunk, samples_per_chunk) in enumerate(sample_chunk_table):
 | |
|     if i == len(sample_chunk_table)-1:
 | |
|       last_chunk = len(chunk_offsets)-1
 | |
|     else:
 | |
|       last_chunk = sample_chunk_table[i+1][0]-1
 | |
|     for k in range(first_chunk, last_chunk+1):
 | |
|       sample_offset = chunk_offsets[k]
 | |
|       for _ in range(samples_per_chunk):
 | |
|         sample_offsets[sample_i] = sample_offset
 | |
|         sample_offset += sample_sizes[sample_i]
 | |
|         sample_i += 1
 | |
| 
 | |
|   assert sample_i == len(sample_sizes)
 | |
| 
 | |
|   pts_offset_table = [
 | |
|     ( int(entry.attrib['CompositionOffset']), int(entry.attrib['SampleCount']) ) for entry in tree.findall(
 | |
|       "MovieBox/TrackBox/MediaBox/MediaInformationBox/SampleTableBox/CompositionOffsetBox/CompositionOffsetEntry")
 | |
|   ]
 | |
|   sample_pts_offset = [0 for _ in sample_sizes]
 | |
|   sample_i = 0
 | |
|   for dt, count in pts_offset_table:
 | |
|     for _ in range(count):
 | |
|       sample_pts_offset[sample_i] = dt
 | |
|       sample_i += 1
 | |
| 
 | |
|   sample_time_table = [
 | |
|     ( int(entry.attrib['SampleDelta']), int(entry.attrib['SampleCount']) ) for entry in tree.findall(
 | |
|       "MovieBox/TrackBox/MediaBox/MediaInformationBox/SampleTableBox/TimeToSampleBox/TimeToSampleEntry")
 | |
|   ]
 | |
|   sample_time = [None for _ in sample_sizes]
 | |
|   cur_ts = 0
 | |
|   sample_i = 0
 | |
|   for dt, count in sample_time_table:
 | |
|     for _ in range(count):
 | |
|       sample_time[sample_i] = (cur_ts + sample_pts_offset[sample_i]) * 1000 / time_scale
 | |
| 
 | |
|       cur_ts += dt
 | |
|       sample_i += 1
 | |
| 
 | |
|   sample_time.sort() # because we ony decode GOPs in PTS order
 | |
| 
 | |
|   return {
 | |
|     'width': width,
 | |
|     'height': height,
 | |
|     'sample_offsets': sample_offsets,
 | |
|     'sample_sizes': sample_sizes,
 | |
|     'sample_dependency': sample_dependency,
 | |
|     'sample_time': sample_time,
 | |
|     'sps': sps,
 | |
|     'pps': pps
 | |
|   }
 | |
| 
 | |
| 
 | |
| class BaseFrameReader(object):
 | |
|   # properties: frame_type, frame_count, w, h
 | |
| 
 | |
|   def __enter__(self):
 | |
|     return self
 | |
| 
 | |
|   def __exit__(self, *args):
 | |
|     self.close()
 | |
| 
 | |
|   def close(self):
 | |
|     pass
 | |
| 
 | |
|   def get(self, num, count=1, pix_fmt="yuv420p"):
 | |
|     raise NotImplementedError
 | |
| 
 | |
| def FrameReader(fn, cache_prefix=None, readahead=False, readbehind=False, multithreaded=True):
 | |
|   frame_type = fingerprint_video(fn)
 | |
|   if frame_type == FrameType.raw:
 | |
|     return RawFrameReader(fn)
 | |
|   elif frame_type in (FrameType.h265_stream, FrameType.h264_pstream):
 | |
|     index_data = get_video_index(fn, frame_type, cache_prefix)
 | |
|     if index_data is not None and "predecom" in index_data:
 | |
|       cache_path = cache_path_for_file_path(fn, cache_prefix)
 | |
|       return MKVFrameReader(
 | |
|         os.path.join(os.path.dirname(cache_path), index_data["predecom"]))
 | |
|     else:
 | |
|       return StreamFrameReader(fn, frame_type, index_data,
 | |
|         readahead=readahead, readbehind=readbehind, multithreaded=multithreaded)
 | |
|   elif frame_type == FrameType.h264_mp4:
 | |
|     return MP4FrameReader(fn, readahead=readahead)
 | |
|   elif frame_type == FrameType.ffv1_mkv:
 | |
|     return MKVFrameReader(fn)
 | |
|   else:
 | |
|     raise NotImplementedError(frame_type)
 | |
| 
 | |
| def rgb24toyuv420(rgb):
 | |
|   yuv_from_rgb = np.array([[ 0.299     ,  0.587     ,  0.114      ],
 | |
|                            [-0.14714119, -0.28886916,  0.43601035 ],
 | |
|                            [ 0.61497538, -0.51496512, -0.10001026 ]])
 | |
|   img = np.dot(rgb.reshape(-1, 3), yuv_from_rgb.T).reshape(rgb.shape)
 | |
| 
 | |
|   y_len = img.shape[0] * img.shape[1]
 | |
|   uv_len = y_len / 4
 | |
| 
 | |
|   ys = img[:, :, 0]
 | |
|   us = (img[::2, ::2, 1] + img[1::2, ::2, 1] + img[::2, 1::2, 1] + img[1::2, 1::2, 1]) / 4 + 128
 | |
|   vs = (img[::2, ::2, 2] + img[1::2, ::2, 2] + img[::2, 1::2, 2] + img[1::2, 1::2, 2]) / 4 + 128
 | |
| 
 | |
|   yuv420 = np.empty(y_len + 2 * uv_len, dtype=img.dtype)
 | |
|   yuv420[:y_len] = ys.reshape(-1)
 | |
|   yuv420[y_len:y_len + uv_len] = us.reshape(-1)
 | |
|   yuv420[y_len + uv_len:y_len + 2 * uv_len] = vs.reshape(-1)
 | |
| 
 | |
|   return yuv420.clip(0,255).astype('uint8')
 | |
| 
 | |
| class RawData(object):
 | |
|   def __init__(self, f):
 | |
|     self.f = _io.FileIO(f, 'rb')
 | |
|     self.lenn = struct.unpack("I", self.f.read(4))[0]
 | |
|     self.count = os.path.getsize(f) / (self.lenn+4)
 | |
| 
 | |
|   def read(self, i):
 | |
|     self.f.seek((self.lenn+4)*i + 4)
 | |
|     return self.f.read(self.lenn)
 | |
| 
 | |
| class RawFrameReader(BaseFrameReader):
 | |
|   def __init__(self, fn):
 | |
|     # raw camera
 | |
|     self.fn = fn
 | |
|     self.frame_type = FrameType.raw
 | |
|     self.rawfile = RawData(self.fn)
 | |
|     self.frame_count = self.rawfile.count
 | |
|     self.w, self.h = 640, 480
 | |
| 
 | |
|   def load_and_debayer(self, img):
 | |
|     img = np.frombuffer(img, dtype='uint8').reshape(960, 1280)
 | |
|     cimg = np.dstack([img[0::2, 1::2], (
 | |
|       (img[0::2, 0::2].astype("uint16") + img[1::2, 1::2].astype("uint16"))
 | |
|       >> 1).astype("uint8"), img[1::2, 0::2]])
 | |
|     return cimg
 | |
| 
 | |
| 
 | |
|   def get(self, num, count=1, pix_fmt="yuv420p"):
 | |
|     assert self.frame_count is not None
 | |
|     assert num+count <= self.frame_count
 | |
| 
 | |
|     if pix_fmt not in ("yuv420p", "rgb24"):
 | |
|       raise ValueError("Unsupported pixel format %r" % pix_fmt)
 | |
| 
 | |
|     app = []
 | |
|     for i in range(num, num+count):
 | |
|       dat = self.rawfile.read(i)
 | |
|       rgb_dat = self.load_and_debayer(dat)
 | |
|       if pix_fmt == "rgb24":
 | |
|         app.append(rgb_dat)
 | |
|       elif pix_fmt == "yuv420p":
 | |
|         app.append(rgb24toyuv420(rgb_dat))
 | |
|       else:
 | |
|         raise NotImplementedError
 | |
| 
 | |
|     return app
 | |
| 
 | |
| def decompress_video_data(rawdat, vid_fmt, w, h, pix_fmt, multithreaded=False):
 | |
|   # using a tempfile is much faster than proc.communicate for some reason
 | |
| 
 | |
|   with tempfile.TemporaryFile() as tmpf:
 | |
|     tmpf.write(rawdat)
 | |
|     tmpf.seek(0)
 | |
| 
 | |
|     proc = subprocess.Popen(
 | |
|       ["ffmpeg",
 | |
|        "-threads", "0" if multithreaded else "1",
 | |
|        "-vsync", "0",
 | |
|        "-f", vid_fmt,
 | |
|        "-flags2", "showall",
 | |
|        "-i", "pipe:0",
 | |
|        "-threads", "0" if multithreaded else "1",
 | |
|        "-f", "rawvideo",
 | |
|        "-pix_fmt", pix_fmt,
 | |
|        "pipe:1"],
 | |
|       stdin=tmpf, stdout=subprocess.PIPE, stderr=open("/dev/null"))
 | |
| 
 | |
|     # dat = proc.communicate()[0]
 | |
|     dat = proc.stdout.read()
 | |
|     if proc.wait() != 0:
 | |
|       raise DataUnreadableError("ffmpeg failed")
 | |
| 
 | |
|   if pix_fmt == "rgb24":
 | |
|     ret = np.frombuffer(dat, dtype=np.uint8).reshape(-1, h, w, 3)
 | |
|   elif pix_fmt == "yuv420p":
 | |
|     ret = np.frombuffer(dat, dtype=np.uint8).reshape(-1, (h*w*3//2))
 | |
|   elif pix_fmt == "yuv444p":
 | |
|     ret = np.frombuffer(dat, dtype=np.uint8).reshape(-1, 3, h, w)
 | |
|   else:
 | |
|     raise NotImplementedError
 | |
| 
 | |
|   return ret
 | |
| 
 | |
| class VideoStreamDecompressor(object):
 | |
|   def __init__(self, vid_fmt, w, h, pix_fmt, multithreaded=False):
 | |
|     self.vid_fmt = vid_fmt
 | |
|     self.w = w
 | |
|     self.h = h
 | |
|     self.pix_fmt = pix_fmt
 | |
| 
 | |
|     if pix_fmt == "yuv420p":
 | |
|       self.out_size = w*h*3//2 # yuv420p
 | |
|     elif pix_fmt in ("rgb24", "yuv444p"):
 | |
|       self.out_size = w*h*3
 | |
|     else:
 | |
|       raise NotImplementedError
 | |
| 
 | |
|     self.out_q = queue.Queue()
 | |
| 
 | |
|     self.proc = subprocess.Popen(
 | |
|       ["ffmpeg",
 | |
|        "-threads", "0" if multithreaded else "1",
 | |
|        # "-avioflags", "direct",
 | |
|        "-analyzeduration", "0",
 | |
|        "-probesize", "32",
 | |
|        "-flush_packets", "0",
 | |
|        # "-fflags", "nobuffer",
 | |
|        "-vsync", "0",
 | |
|        "-f", vid_fmt,
 | |
|        "-i", "pipe:0",
 | |
|        "-threads", "0" if multithreaded else "1",
 | |
|        "-f", "rawvideo",
 | |
|        "-pix_fmt", pix_fmt,
 | |
|        "pipe:1"],
 | |
|       stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=open("/dev/null", "wb"))
 | |
| 
 | |
|     def read_thread():
 | |
|       while True:
 | |
|         r = self.proc.stdout.read(self.out_size)
 | |
|         if len(r) == 0:
 | |
|           break
 | |
|         assert len(r) == self.out_size
 | |
|         self.out_q.put(r)
 | |
| 
 | |
|     self.t = threading.Thread(target=read_thread)
 | |
|     self.t.daemon = True
 | |
|     self.t.start()
 | |
| 
 | |
|   def __enter__(self):
 | |
|     return self
 | |
| 
 | |
|   def __exit__(self, *args):
 | |
|     self.close()
 | |
| 
 | |
|   def write(self, rawdat):
 | |
|     self.proc.stdin.write(rawdat)
 | |
|     self.proc.stdin.flush()
 | |
| 
 | |
|   def read(self):
 | |
|     dat = self.out_q.get(block=True)
 | |
| 
 | |
|     if self.pix_fmt == "rgb24":
 | |
|       ret = np.frombuffer(dat, dtype=np.uint8).reshape((self.h, self.w, 3))
 | |
|     elif self.pix_fmt == "yuv420p":
 | |
|       ret = np.frombuffer(dat, dtype=np.uint8)
 | |
|     elif self.pix_fmt == "yuv444p":
 | |
|       ret = np.frombuffer(dat, dtype=np.uint8).reshape((3, self.h, self.w))
 | |
|     else:
 | |
|       assert False
 | |
| 
 | |
|     return ret
 | |
| 
 | |
|   def eos(self):
 | |
|     self.proc.stdin.close()
 | |
| 
 | |
|   def close(self):
 | |
|     self.proc.stdin.close()
 | |
|     self.t.join()
 | |
|     self.proc.wait()
 | |
|     assert self.proc.wait() == 0
 | |
| 
 | |
| 
 | |
| class MKVFrameReader(BaseFrameReader):
 | |
|   def __init__(self, fn):
 | |
|     self.fn = fn
 | |
| 
 | |
|     #print("MKVFrameReader", fn)
 | |
|     index_data = index_mkv(fn)
 | |
|     stream = index_data['probe']['streams'][0]
 | |
|     self.w = stream['width']
 | |
|     self.h = stream['height']
 | |
| 
 | |
|     if stream['codec_name'] == 'ffv1':
 | |
|       self.frame_type = FrameType.ffv1_mkv
 | |
|     elif stream['codec_name'] == 'ffvhuff':
 | |
|       self.frame_type = FrameType.ffvhuff_mkv
 | |
|     else:
 | |
|       raise NotImplementedError
 | |
| 
 | |
|     self.config_record = index_data['config_record']
 | |
|     self.index = index_data['index']
 | |
| 
 | |
|     self.frame_count = len(self.index)
 | |
| 
 | |
|   def get(self, num, count=1, pix_fmt="yuv420p"):
 | |
|     assert 0 < num+count <= self.frame_count
 | |
| 
 | |
|     frame_dats = []
 | |
|     with FileReader(self.fn) as f:
 | |
|       for i in range(num, num+count):
 | |
|         pos, length, _ = self.index[i]
 | |
|         f.seek(pos)
 | |
|         frame_dats.append(f.read(length))
 | |
| 
 | |
|     of = StringIO()
 | |
|     mkvindex.simple_gen(of, self.config_record, self.w, self.h, frame_dats)
 | |
| 
 | |
|     r = decompress_video_data(of.getvalue(), "matroska", self.w, self.h, pix_fmt)
 | |
|     assert len(r) == count
 | |
| 
 | |
|     return r
 | |
| 
 | |
| 
 | |
| class GOPReader(object):
 | |
|   def get_gop(self, num):
 | |
|     # returns (start_frame_num, num_frames, frames_to_skip, gop_data)
 | |
|     raise NotImplementedError
 | |
| 
 | |
| 
 | |
| class DoNothingContextManager(object):
 | |
|   def __enter__(self): return self
 | |
|   def __exit__(*x): pass
 | |
| 
 | |
| 
 | |
| class GOPFrameReader(BaseFrameReader):
 | |
|   #FrameReader with caching and readahead for formats that are group-of-picture based
 | |
| 
 | |
|   def __init__(self, readahead=False, readbehind=False, multithreaded=True):
 | |
|     self.open_ = True
 | |
| 
 | |
|     self.multithreaded = multithreaded
 | |
|     self.readahead = readahead
 | |
|     self.readbehind = readbehind
 | |
|     self.frame_cache = LRU(64)
 | |
| 
 | |
|     if self.readahead:
 | |
|       self.cache_lock = threading.RLock()
 | |
|       self.readahead_last = None
 | |
|       self.readahead_len = 30
 | |
|       self.readahead_c = threading.Condition()
 | |
|       self.readahead_thread = threading.Thread(target=self._readahead_thread)
 | |
|       self.readahead_thread.daemon = True
 | |
|       self.readahead_thread.start()
 | |
|     else:
 | |
|       self.cache_lock = DoNothingContextManager()
 | |
| 
 | |
|   def close(self):
 | |
|     if not self.open_:
 | |
|       return
 | |
|     self.open_ = False
 | |
| 
 | |
|     if self.readahead:
 | |
|       self.readahead_c.acquire()
 | |
|       self.readahead_c.notify()
 | |
|       self.readahead_c.release()
 | |
|       self.readahead_thread.join()
 | |
| 
 | |
|   def _readahead_thread(self):
 | |
|     while True:
 | |
|       self.readahead_c.acquire()
 | |
|       try:
 | |
|         if not self.open_:
 | |
|           break
 | |
|         self.readahead_c.wait()
 | |
|       finally:
 | |
|         self.readahead_c.release()
 | |
|       if not self.open_:
 | |
|         break
 | |
|       assert self.readahead_last
 | |
|       num, pix_fmt = self.readahead_last
 | |
| 
 | |
|       if self.readbehind:
 | |
|         for k in range(num-1, max(0, num-self.readahead_len), -1):
 | |
|           self._get_one(k, pix_fmt)
 | |
|       else:
 | |
|         for k in range(num, min(self.frame_count, num+self.readahead_len)):
 | |
|           self._get_one(k, pix_fmt)
 | |
| 
 | |
|   def _get_one(self, num, pix_fmt):
 | |
|     assert num < self.frame_count
 | |
| 
 | |
|     if (num, pix_fmt) in self.frame_cache:
 | |
|       return self.frame_cache[(num, pix_fmt)]
 | |
| 
 | |
|     with self.cache_lock:
 | |
|       if (num, pix_fmt) in self.frame_cache:
 | |
|         return self.frame_cache[(num, pix_fmt)]
 | |
| 
 | |
|       frame_b, num_frames, skip_frames, rawdat = self.get_gop(num)
 | |
| 
 | |
|       ret = decompress_video_data(rawdat, self.vid_fmt, self.w, self.h, pix_fmt,
 | |
|                                   multithreaded=self.multithreaded)
 | |
|       ret = ret[skip_frames:]
 | |
|       assert ret.shape[0] == num_frames
 | |
| 
 | |
|       for i in range(ret.shape[0]):
 | |
|         self.frame_cache[(frame_b+i, pix_fmt)] = ret[i]
 | |
| 
 | |
|       return self.frame_cache[(num, pix_fmt)]
 | |
| 
 | |
|   def get(self, num, count=1, pix_fmt="yuv420p"):
 | |
|     assert self.frame_count is not None
 | |
| 
 | |
|     if num + count > self.frame_count:
 | |
|       raise ValueError("{} > {}".format(num + count, self.frame_count))
 | |
| 
 | |
|     if pix_fmt not in ("yuv420p", "rgb24", "yuv444p"):
 | |
|       raise ValueError("Unsupported pixel format %r" % pix_fmt)
 | |
| 
 | |
|     ret = [self._get_one(num + i, pix_fmt) for i in range(count)]
 | |
| 
 | |
|     if self.readahead:
 | |
|       self.readahead_last = (num+count, pix_fmt)
 | |
|       self.readahead_c.acquire()
 | |
|       self.readahead_c.notify()
 | |
|       self.readahead_c.release()
 | |
| 
 | |
|     return ret
 | |
| 
 | |
| class MP4GOPReader(GOPReader):
 | |
|   def __init__(self, fn):
 | |
|     self.fn = fn
 | |
|     self.frame_type = FrameType.h264_mp4
 | |
| 
 | |
|     self.index = index_mp4(fn)
 | |
| 
 | |
|     self.w = self.index['width']
 | |
|     self.h = self.index['height']
 | |
|     self.sample_sizes = self.index['sample_sizes']
 | |
|     self.sample_offsets = self.index['sample_offsets']
 | |
|     self.sample_dependency = self.index['sample_dependency']
 | |
| 
 | |
|     self.vid_fmt = "h264"
 | |
| 
 | |
|     self.frame_count = len(self.sample_sizes)
 | |
| 
 | |
|     self.prefix = "\x00\x00\x00\x01"+self.index['sps']+"\x00\x00\x00\x01"+self.index['pps']
 | |
| 
 | |
|   def _lookup_gop(self, num):
 | |
|     frame_b = num
 | |
|     while frame_b > 0 and self.sample_dependency[frame_b]:
 | |
|       frame_b -= 1
 | |
| 
 | |
|     frame_e = num+1
 | |
|     while frame_e < (len(self.sample_dependency)-1) and self.sample_dependency[frame_e]:
 | |
|       frame_e += 1
 | |
| 
 | |
|     return (frame_b, frame_e)
 | |
| 
 | |
|   def get_gop(self, num):
 | |
|     frame_b, frame_e = self._lookup_gop(num)
 | |
|     assert frame_b <= num < frame_e
 | |
| 
 | |
|     num_frames = frame_e-frame_b
 | |
| 
 | |
|     with FileReader(self.fn) as f:
 | |
|       rawdat = []
 | |
| 
 | |
|       sample_i = frame_b
 | |
|       while sample_i < frame_e:
 | |
|         size = self.sample_sizes[sample_i]
 | |
|         start_offset = self.sample_offsets[sample_i]
 | |
| 
 | |
|         # try to read contiguously because a read could actually be a http request
 | |
|         sample_i += 1
 | |
|         while sample_i < frame_e and size < 10000000 and start_offset+size == self.sample_offsets[sample_i]:
 | |
|           size += self.sample_sizes[sample_i]
 | |
|           sample_i += 1
 | |
| 
 | |
|         f.seek(start_offset)
 | |
|         sampledat = f.read(size)
 | |
| 
 | |
|         # read length-prefixed NALUs and output in Annex-B
 | |
|         i = 0
 | |
|         while i < len(sampledat):
 | |
|           nal_len, = struct.unpack(">I", sampledat[i:i+4])
 | |
|           rawdat.append("\x00\x00\x00\x01"+sampledat[i+4:i+4+nal_len])
 | |
|           i = i+4+nal_len
 | |
|         assert i == len(sampledat)
 | |
| 
 | |
|     rawdat = self.prefix+''.join(rawdat)
 | |
| 
 | |
|     return frame_b, num_frames, 0, rawdat
 | |
| 
 | |
| class MP4FrameReader(MP4GOPReader, GOPFrameReader):
 | |
|   def __init__(self, fn, readahead=False):
 | |
|     MP4GOPReader.__init__(self, fn)
 | |
|     GOPFrameReader.__init__(self, readahead)
 | |
| 
 | |
| class StreamGOPReader(GOPReader):
 | |
|   def __init__(self, fn, frame_type, index_data):
 | |
|     self.fn = fn
 | |
| 
 | |
|     self.frame_type = frame_type
 | |
|     self.frame_count = None
 | |
|     self.w, self.h = None, None
 | |
| 
 | |
|     self.prefix = None
 | |
|     self.index = None
 | |
| 
 | |
|     self.index = index_data['index']
 | |
|     self.prefix = index_data['global_prefix']
 | |
|     probe = index_data['probe']
 | |
| 
 | |
|     if self.frame_type == FrameType.h265_stream:
 | |
|       self.prefix_frame_data = None
 | |
|       self.num_prefix_frames = 0
 | |
|       self.vid_fmt = "hevc"
 | |
| 
 | |
|     elif self.frame_type == FrameType.h264_pstream:
 | |
|       self.prefix_frame_data = index_data['prefix_frame_data']
 | |
|       self.num_prefix_frames = index_data['num_prefix_frames']
 | |
| 
 | |
|       self.vid_fmt = "h264"
 | |
| 
 | |
|     i = 0
 | |
|     while i < self.index.shape[0] and self.index[i, 0] != SLICE_I:
 | |
|       i += 1
 | |
|     self.first_iframe = i
 | |
| 
 | |
|     if self.frame_type == FrameType.h265_stream:
 | |
|       assert self.first_iframe == 0
 | |
| 
 | |
|     self.frame_count = len(self.index)-1
 | |
| 
 | |
|     self.w = probe['streams'][0]['width']
 | |
|     self.h = probe['streams'][0]['height']
 | |
| 
 | |
| 
 | |
|   def _lookup_gop(self, num):
 | |
|     frame_b = num
 | |
|     while frame_b > 0 and self.index[frame_b, 0] != SLICE_I:
 | |
|       frame_b -= 1
 | |
| 
 | |
|     frame_e = num+1
 | |
|     while frame_e < (len(self.index)-1) and self.index[frame_e, 0] != SLICE_I:
 | |
|       frame_e += 1
 | |
| 
 | |
|     offset_b = self.index[frame_b, 1]
 | |
|     offset_e = self.index[frame_e, 1]
 | |
| 
 | |
|     return (frame_b, frame_e, offset_b, offset_e)
 | |
| 
 | |
|   def get_gop(self, num):
 | |
|     frame_b, frame_e, offset_b, offset_e = self._lookup_gop(num)
 | |
|     assert frame_b <= num < frame_e
 | |
| 
 | |
|     num_frames = frame_e-frame_b
 | |
| 
 | |
|     with FileReader(self.fn) as f:
 | |
|       f.seek(offset_b)
 | |
|       rawdat = f.read(offset_e-offset_b)
 | |
| 
 | |
|       if num < self.first_iframe:
 | |
|         assert self.prefix_frame_data
 | |
|         rawdat = self.prefix_frame_data + rawdat
 | |
| 
 | |
|       rawdat = self.prefix + rawdat
 | |
| 
 | |
|     skip_frames = 0
 | |
|     if num < self.first_iframe:
 | |
|       skip_frames = self.num_prefix_frames
 | |
| 
 | |
|     return frame_b, num_frames, skip_frames, rawdat
 | |
| 
 | |
| class StreamFrameReader(StreamGOPReader, GOPFrameReader):
 | |
|   def __init__(self, fn, frame_type, index_data, readahead=False, readbehind=False, multithreaded=False):
 | |
|     StreamGOPReader.__init__(self, fn, frame_type, index_data)
 | |
|     GOPFrameReader.__init__(self, readahead, readbehind, multithreaded)
 | |
| 
 | |
| 
 | |
| 
 | |
| 
 | |
| def GOPFrameIterator(gop_reader, pix_fmt, multithreaded=True):
 | |
|   # this is really ugly. ill think about how to refactor it when i can think good
 | |
| 
 | |
|   IN_FLIGHT_GOPS = 6 # should be enough that the stream decompressor starts returning data
 | |
| 
 | |
|   with VideoStreamDecompressor(
 | |
|       gop_reader.vid_fmt, gop_reader.w, gop_reader.h, pix_fmt, multithreaded) as dec:
 | |
| 
 | |
|     read_work = []
 | |
| 
 | |
|     def readthing():
 | |
|       # print read_work, dec.out_q.qsize()
 | |
|       outf = dec.read()
 | |
|       read_thing = read_work[0]
 | |
|       if read_thing[0] > 0:
 | |
|         read_thing[0] -= 1
 | |
|       else:
 | |
|         assert read_thing[1] > 0
 | |
|         yield outf
 | |
|         read_thing[1] -= 1
 | |
| 
 | |
|       if read_thing[1] == 0:
 | |
|         read_work.pop(0)
 | |
| 
 | |
|     i = 0
 | |
|     while i < gop_reader.frame_count:
 | |
|       frame_b, num_frames, skip_frames, gop_data = gop_reader.get_gop(i)
 | |
|       dec.write(gop_data)
 | |
|       i += num_frames
 | |
|       read_work.append([skip_frames, num_frames])
 | |
| 
 | |
|       while len(read_work) >= IN_FLIGHT_GOPS:
 | |
|         for v in readthing(): yield v
 | |
| 
 | |
|     dec.eos()
 | |
| 
 | |
|     while read_work:
 | |
|       for v in readthing(): yield v
 | |
| 
 | |
| 
 | |
| def FrameIterator(fn, pix_fmt, **kwargs):
 | |
|   fr = FrameReader(fn, **kwargs)
 | |
|   if isinstance(fr, GOPReader):
 | |
|     for v in GOPFrameIterator(fr, pix_fmt, kwargs.get("multithreaded", True)): yield v
 | |
|   else:
 | |
|     for i in range(fr.frame_count):
 | |
|       yield fr.get(i, pix_fmt=pix_fmt)[0]
 | |
| 
 | |
| 
 | |
| def FrameWriter(ofn, frames, vid_fmt=FrameType.ffvhuff_mkv, pix_fmt="rgb24", framerate=20, multithreaded=False):
 | |
|   if pix_fmt not in ("rgb24", "yuv420p"):
 | |
|     raise NotImplementedError
 | |
| 
 | |
|   if vid_fmt == FrameType.ffv1_mkv:
 | |
|     assert ofn.endswith(".mkv")
 | |
|     vcodec = "ffv1"
 | |
|   elif vid_fmt == FrameType.ffvhuff_mkv:
 | |
|     assert ofn.endswith(".mkv")
 | |
|     vcodec = "ffvhuff"
 | |
|   else:
 | |
|     raise NotImplementedError
 | |
| 
 | |
|   frame_gen = iter(frames)
 | |
|   first_frame = next(frame_gen)
 | |
| 
 | |
|   # assert len(frames) > 1
 | |
|   if pix_fmt == "rgb24":
 | |
|     h, w = first_frame.shape[:2]
 | |
|   elif pix_fmt == "yuv420p":
 | |
|     w = first_frame.shape[1]
 | |
|     h = 2*first_frame.shape[0]//3
 | |
|   else:
 | |
|     raise NotImplementedError
 | |
| 
 | |
|   compress_proc = subprocess.Popen(
 | |
|     ["ffmpeg",
 | |
|      "-threads", "0" if multithreaded else "1",
 | |
|      "-y",
 | |
|      "-framerate", str(framerate),
 | |
|      "-vsync", "0",
 | |
|      "-f", "rawvideo",
 | |
|      "-pix_fmt", pix_fmt,
 | |
|      "-s", "%dx%d" % (w, h),
 | |
|      "-i", "pipe:0",
 | |
|      "-threads", "0" if multithreaded else "1",
 | |
|      "-f", "matroska",
 | |
|      "-vcodec", vcodec,
 | |
|      "-g", "0",
 | |
|      ofn],
 | |
|     stdin=subprocess.PIPE, stderr=open("/dev/null", "wb"))
 | |
|   try:
 | |
|     compress_proc.stdin.write(first_frame.tobytes())
 | |
|     for frame in frame_gen:
 | |
|       compress_proc.stdin.write(frame.tobytes())
 | |
|     compress_proc.stdin.close()
 | |
|   except:
 | |
|     compress_proc.kill()
 | |
|     raise
 | |
| 
 | |
|   assert compress_proc.wait() == 0
 | |
| 
 | |
| if __name__ == "__main__":
 | |
|   fn = "cd:/1c79456b0c90f15a/2017-05-10--08-17-00/2/fcamera.hevc"
 | |
|   f = FrameReader(fn)
 | |
|   # print f.get(0, 1).shape
 | |
|   # print f.get(15, 1).shape
 | |
|   for v in GOPFrameIterator(f, "yuv420p"):
 | |
|     print(v)
 | |
| 
 |