|  |  |  | #!/usr/bin/env python3
 | 
					
						
							|  |  |  | import av
 | 
					
						
							|  |  |  | import os
 | 
					
						
							|  |  |  | import sys
 | 
					
						
							|  |  |  | import argparse
 | 
					
						
							|  |  |  | import numpy as np
 | 
					
						
							|  |  |  | import multiprocessing
 | 
					
						
							|  |  |  | import time
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | import cereal.messaging as messaging
 | 
					
						
							|  |  |  | from cereal.visionipc import VisionIpcServer, VisionStreamType
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | W, H = 1928, 1208
 | 
					
						
							|  |  |  | V4L2_BUF_FLAG_KEYFRAME = 8
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | ENCODE_SOCKETS = {
 | 
					
						
							|  |  |  |   VisionStreamType.VISION_STREAM_ROAD: "roadEncodeData",
 | 
					
						
							|  |  |  |   VisionStreamType.VISION_STREAM_WIDE_ROAD: "wideRoadEncodeData",
 | 
					
						
							|  |  |  |   VisionStreamType.VISION_STREAM_DRIVER: "driverEncodeData",
 | 
					
						
							|  |  |  | }
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | def decoder(addr, vst, nvidia, debug=False):
 | 
					
						
							|  |  |  |   vipc_server = VisionIpcServer("camerad")
 | 
					
						
							|  |  |  |   vipc_server.create_buffers(vst, 4, False, W, H)
 | 
					
						
							|  |  |  |   vipc_server.start_listener()
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |   sock_name = ENCODE_SOCKETS[vst]
 | 
					
						
							|  |  |  |   if debug:
 | 
					
						
							|  |  |  |     print("start decoder for %s" % sock_name)
 | 
					
						
							|  |  |  |   if nvidia:
 | 
					
						
							|  |  |  |     os.environ["NV_LOW_LATENCY"] = "3"    # both bLowLatency and CUVID_PKT_ENDOFPICTURE
 | 
					
						
							|  |  |  |     sys.path += os.environ["LD_LIBRARY_PATH"].split(":")
 | 
					
						
							|  |  |  |     import PyNvCodec as nvc
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     nvDec = nvc.PyNvDecoder(W, H, nvc.PixelFormat.NV12, nvc.CudaVideoCodec.HEVC, 0)
 | 
					
						
							|  |  |  |     cc1 = nvc.ColorspaceConversionContext(nvc.ColorSpace.BT_709, nvc.ColorRange.JPEG)
 | 
					
						
							|  |  |  |     conv_yuv = nvc.PySurfaceConverter(W, H, nvc.PixelFormat.NV12, nvc.PixelFormat.YUV420, 0)
 | 
					
						
							|  |  |  |     nvDwn_yuv = nvc.PySurfaceDownloader(W, H, nvc.PixelFormat.YUV420, 0)
 | 
					
						
							|  |  |  |     img_yuv = np.ndarray((H*W//2*3), dtype=np.uint8)
 | 
					
						
							|  |  |  |   else:
 | 
					
						
							|  |  |  |     codec = av.CodecContext.create("hevc", "r")
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |   os.environ["ZMQ"] = "1"
 | 
					
						
							|  |  |  |   messaging.context = messaging.Context()
 | 
					
						
							|  |  |  |   sock = messaging.sub_sock(sock_name, None, addr=addr, conflate=False)
 | 
					
						
							|  |  |  |   cnt = 0
 | 
					
						
							|  |  |  |   last_idx = -1
 | 
					
						
							|  |  |  |   seen_iframe = False
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |   time_q = []
 | 
					
						
							|  |  |  |   while 1:
 | 
					
						
							|  |  |  |     msgs = messaging.drain_sock(sock, wait_for_one=True)
 | 
					
						
							|  |  |  |     for evt in msgs:
 | 
					
						
							|  |  |  |       evta = getattr(evt, evt.which())
 | 
					
						
							|  |  |  |       if debug and evta.idx.encodeId != 0 and evta.idx.encodeId != (last_idx+1):
 | 
					
						
							|  |  |  |         print("DROP PACKET!")
 | 
					
						
							| 
									
										
											  
											
												loggerd: switch to v4l encoder try 2 (#24380)
* start v4l encoder
* v4l encoder starts
* start and stop
* fill in proper controls
* it dequeued a buffer
* getting bytes
* it made a video
* it does make files
* getting close
* ahh, so that's how dequeue works
* qcam works (no remuxing)
* remuxing works
* we just need to make shutdown and rollover graceful
* graceful destruction
* switch to polling
* should work now
* fix pc build
* refactors, stop properly
* touchups, remove a copy
* add v4l encoder to release
* inlcude file
* move writing to it's own thread
* fix minor memory leak
* block instead of dropping frames
* add counter, fix tests maybe
* better debugging and test print
* print file path in assert
* format string in test
* no more oversized qlogs
* match qcam
* touchups, remove omx encoder
* remove omx include files
* checked ioctl, better debugging, open by name
* unused import
* move linux includes to third_party/linux/include
* simple encoderd
* full packet
* encoderd should be complete
* lagging print
* updates
* name dq thread
* subset idx
* video file writing works
* debug
* potential bugfix
* rotation works
* iframe
* keep writing support
* ci should pass
* loggerd, not encoderd
* remote encoder code
* support remote encoder
* cereal to master, add encoderd
* header no longer required
* put that back there
* realtime
* lower decoder latency
* don't use queue for VisionIpcBufExtra, disable realtime again
* assert all written
* hmm simpler
* only push to to_write if we are writing
* assert timestamp is right
* use at and remove assert
* revert to queue
Co-authored-by: Comma Device <device@comma.ai>
old-commit-hash: 0baa4c3e2ad9ee6f8daba8267db44c2cd44caa62
											
										 
											4 years ago
										 |  |  |       last_idx = evta.idx.encodeId
 | 
					
						
							|  |  |  |       if not seen_iframe and not (evta.idx.flags & V4L2_BUF_FLAG_KEYFRAME):
 | 
					
						
							|  |  |  |         if debug:
 | 
					
						
							|  |  |  |           print("waiting for iframe")
 | 
					
						
							|  |  |  |         continue
 | 
					
						
							|  |  |  |       time_q.append(time.monotonic())
 | 
					
						
							|  |  |  |       network_latency = (int(time.time()*1e9) - evta.unixTimestampNanos)/1e6
 | 
					
						
							|  |  |  |       frame_latency = ((evta.idx.timestampEof/1e9) - (evta.idx.timestampSof/1e9))*1000
 | 
					
						
							|  |  |  |       process_latency = ((evt.logMonoTime/1e9) - (evta.idx.timestampEof/1e9))*1000
 | 
					
						
							| 
									
										
											  
											
												loggerd: switch to v4l encoder try 2 (#24380)
* start v4l encoder
* v4l encoder starts
* start and stop
* fill in proper controls
* it dequeued a buffer
* getting bytes
* it made a video
* it does make files
* getting close
* ahh, so that's how dequeue works
* qcam works (no remuxing)
* remuxing works
* we just need to make shutdown and rollover graceful
* graceful destruction
* switch to polling
* should work now
* fix pc build
* refactors, stop properly
* touchups, remove a copy
* add v4l encoder to release
* inlcude file
* move writing to it's own thread
* fix minor memory leak
* block instead of dropping frames
* add counter, fix tests maybe
* better debugging and test print
* print file path in assert
* format string in test
* no more oversized qlogs
* match qcam
* touchups, remove omx encoder
* remove omx include files
* checked ioctl, better debugging, open by name
* unused import
* move linux includes to third_party/linux/include
* simple encoderd
* full packet
* encoderd should be complete
* lagging print
* updates
* name dq thread
* subset idx
* video file writing works
* debug
* potential bugfix
* rotation works
* iframe
* keep writing support
* ci should pass
* loggerd, not encoderd
* remote encoder code
* support remote encoder
* cereal to master, add encoderd
* header no longer required
* put that back there
* realtime
* lower decoder latency
* don't use queue for VisionIpcBufExtra, disable realtime again
* assert all written
* hmm simpler
* only push to to_write if we are writing
* assert timestamp is right
* use at and remove assert
* revert to queue
Co-authored-by: Comma Device <device@comma.ai>
old-commit-hash: 0baa4c3e2ad9ee6f8daba8267db44c2cd44caa62
											
										 
											4 years ago
										 |  |  | 
 | 
					
						
							|  |  |  |       # put in header (first)
 | 
					
						
							|  |  |  |       if not seen_iframe:
 | 
					
						
							|  |  |  |         if nvidia:
 | 
					
						
							|  |  |  |           nvDec.DecodeSurfaceFromPacket(np.frombuffer(evta.header, dtype=np.uint8))
 | 
					
						
							|  |  |  |         else:
 | 
					
						
							|  |  |  |           codec.decode(av.packet.Packet(evta.header))
 | 
					
						
							|  |  |  |         seen_iframe = True
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |       if nvidia:
 | 
					
						
							|  |  |  |         rawSurface = nvDec.DecodeSurfaceFromPacket(np.frombuffer(evta.data, dtype=np.uint8))
 | 
					
						
							|  |  |  |         if rawSurface.Empty():
 | 
					
						
							|  |  |  |           if debug:
 | 
					
						
							|  |  |  |             print("DROP SURFACE")
 | 
					
						
							|  |  |  |           continue
 | 
					
						
							|  |  |  |         convSurface = conv_yuv.Execute(rawSurface, cc1)
 | 
					
						
							|  |  |  |         nvDwn_yuv.DownloadSingleSurface(convSurface, img_yuv)
 | 
					
						
							|  |  |  |       else:
 | 
					
						
							|  |  |  |         frames = codec.decode(av.packet.Packet(evta.data))
 | 
					
						
							|  |  |  |         if len(frames) == 0:
 | 
					
						
							|  |  |  |           if debug:
 | 
					
						
							|  |  |  |             print("DROP SURFACE")
 | 
					
						
							|  |  |  |           continue
 | 
					
						
							|  |  |  |         assert len(frames) == 1
 | 
					
						
							|  |  |  |         img_yuv = frames[0].to_ndarray(format=av.video.format.VideoFormat('yuv420p')).flatten()
 | 
					
						
							|  |  |  |         uv_offset = H*W
 | 
					
						
							|  |  |  |         y = img_yuv[:uv_offset]
 | 
					
						
							|  |  |  |         uv = img_yuv[uv_offset:].reshape(2, -1).ravel('F')
 | 
					
						
							|  |  |  |         img_yuv = np.hstack((y, uv))
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |       vipc_server.send(vst, img_yuv.data, cnt, int(time_q[0]*1e9), int(time.monotonic()*1e9))
 | 
					
						
							|  |  |  |       cnt += 1
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |       pc_latency = (time.monotonic()-time_q[0])*1000
 | 
					
						
							|  |  |  |       time_q = time_q[1:]
 | 
					
						
							|  |  |  |       if debug:
 | 
					
						
							|  |  |  |         print("%2d %4d %.3f %.3f roll %6.2f ms latency %6.2f ms + %6.2f ms + %6.2f ms = %6.2f ms"
 | 
					
						
							|  |  |  |               % (len(msgs), evta.idx.encodeId, evt.logMonoTime/1e9, evta.idx.timestampEof/1e6, frame_latency,
 | 
					
						
							|  |  |  |                  process_latency, network_latency, pc_latency, process_latency+network_latency+pc_latency ), len(evta.data), sock_name)
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | class CompressedVipc:
 | 
					
						
							|  |  |  |   def __init__(self, addr, vision_streams, nvidia=False, debug=False):
 | 
					
						
							|  |  |  |     self.procs = []
 | 
					
						
							|  |  |  |     for vst in vision_streams:
 | 
					
						
							|  |  |  |       p = multiprocessing.Process(target=decoder, args=(addr, vst, nvidia, debug))
 | 
					
						
							|  |  |  |       p.start()
 | 
					
						
							|  |  |  |       self.procs.append(p)
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |   def join(self):
 | 
					
						
							|  |  |  |     for p in self.procs:
 | 
					
						
							|  |  |  |       p.join()
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |   def kill(self):
 | 
					
						
							|  |  |  |     for p in self.procs:
 | 
					
						
							|  |  |  |       p.terminate()
 | 
					
						
							|  |  |  |     self.join()
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | if __name__ == "__main__":
 | 
					
						
							|  |  |  |   parser = argparse.ArgumentParser(description="Decode video streams and broadcast on VisionIPC")
 | 
					
						
							|  |  |  |   parser.add_argument("addr", help="Address of comma three")
 | 
					
						
							|  |  |  |   parser.add_argument("--nvidia", action="store_true", help="Use nvidia instead of ffmpeg")
 | 
					
						
							|  |  |  |   parser.add_argument("--cams", default="0,1,2", help="Cameras to decode")
 | 
					
						
							|  |  |  |   parser.add_argument("--silent", action="store_true", help="Suppress debug output")
 | 
					
						
							|  |  |  |   args = parser.parse_args()
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |   vision_streams = [
 | 
					
						
							|  |  |  |     VisionStreamType.VISION_STREAM_ROAD,
 | 
					
						
							|  |  |  |     VisionStreamType.VISION_STREAM_WIDE_ROAD,
 | 
					
						
							|  |  |  |     VisionStreamType.VISION_STREAM_DRIVER,
 | 
					
						
							|  |  |  |   ]
 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |   vsts = [vision_streams[int(x)] for x in args.cams.split(",")]
 | 
					
						
							|  |  |  |   cvipc = CompressedVipc(args.addr, vsts, args.nvidia, debug=(not args.silent))
 | 
					
						
							|  |  |  |   cvipc.join()
 |