587 lines
24 KiB
Python
587 lines
24 KiB
Python
# -*- mode: python; indent-tabs-mode: nil; py-indent-offset: 4; coding: utf-8 -*-
|
|
import time
|
|
import threading
|
|
import logging
|
|
import itertools
|
|
|
|
from toxygen_wrapper.toxav_enums import *
|
|
from toxygen_wrapper.tests import support_testing as ts
|
|
from toxygen_wrapper.tests.support_testing import LOG_ERROR, LOG_WARN, LOG_INFO, LOG_DEBUG, LOG_TRACE
|
|
|
|
with ts.ignoreStderr():
|
|
import pyaudio
|
|
from av import screen_sharing
|
|
from av.call import Call
|
|
import common.tox_save
|
|
|
|
from utils import ui as util_ui
|
|
from middleware.threads import invoke_in_main_thread
|
|
# from middleware.threads import BaseThread
|
|
|
|
sleep = time.sleep
|
|
|
|
global LOG
|
|
LOG = logging.getLogger('app.'+__name__)
|
|
|
|
TIMER_TIMEOUT = 30.0
|
|
iFPS = 25
|
|
|
|
class AudioThread(BaseQThread):
|
|
def __init__(self, av, name=''):
|
|
super().__init__()
|
|
self.av = av
|
|
self._name = name
|
|
|
|
def join(self, ito=ts.iTHREAD_TIMEOUT):
|
|
LOG_DEBUG(f"AudioThread join {self}")
|
|
# dunno
|
|
|
|
def run(self):
|
|
LOG_DEBUG('AudioThread run: ')
|
|
# maybe not needed
|
|
while not self._stop_thread:
|
|
self.av.send_audio()
|
|
sleep(100.0 / 1000.0)
|
|
|
|
class VideoThread(BaseQThread):
|
|
def __init__(self, av, name=''):
|
|
super().__init__()
|
|
self.av = av
|
|
self._name = name
|
|
|
|
def join(self, ito=ts.iTHREAD_TIMEOUT):
|
|
LOG_DEBUG(f"VideoThread join {self}")
|
|
# dunno
|
|
|
|
def run(self):
|
|
LOG_DEBUG('VideoThread run: ')
|
|
# maybe not needed
|
|
while not self._stop_thread:
|
|
self.av.send_video()
|
|
sleep(100.0 / 1000.0)
|
|
|
|
class AV(common.tox_save.ToxAvSave):
|
|
|
|
def __init__(self, toxav, settings):
|
|
super().__init__(toxav)
|
|
self._toxav = toxav
|
|
self._settings = settings
|
|
self._running = True
|
|
s = settings
|
|
if 'video' not in s:
|
|
LOG.warn("AV.__init__ 'video' not in s" )
|
|
LOG.debug(f"AV.__init__ {s}" )
|
|
elif 'device' not in s['video']:
|
|
LOG.warn("AV.__init__ 'device' not in s.video" )
|
|
LOG.debug(f"AV.__init__ {s['video']}" )
|
|
|
|
self._calls = {} # dict: key - friend number, value - Call instance
|
|
|
|
self._audio = None
|
|
self._audio_stream = None
|
|
self._audio_thread = None
|
|
self._audio_running = False
|
|
self._out_stream = None
|
|
|
|
self._audio_channels = 1
|
|
self._audio_duration = 60
|
|
self._audio_rate_pa = 48000
|
|
self._audio_rate_tox = 48000
|
|
self._audio_rate_pa = 48000
|
|
self._audio_krate_tox_audio = self._audio_rate_tox // 1000
|
|
self._audio_krate_tox_video = 5000
|
|
self._audio_sample_count_pa = self._audio_rate_pa * self._audio_channels * self._audio_duration // 1000
|
|
self._audio_sample_count_tox = self._audio_rate_tox * self._audio_channels * self._audio_duration // 1000
|
|
|
|
self._video = None
|
|
self._video_thread = None
|
|
self._video_running = None
|
|
|
|
self._video_width = 320
|
|
self._video_height = 240
|
|
|
|
# was iOutput = self._settings._args.audio['output']
|
|
iInput = self._settings['audio']['input']
|
|
self.lPaSampleratesI = ts.lSdSamplerates(iInput)
|
|
iOutput = self._settings['audio']['output']
|
|
self.lPaSampleratesO = ts.lSdSamplerates(iOutput)
|
|
|
|
global oPYA
|
|
oPYA = self._audio = pyaudio.PyAudio()
|
|
|
|
def stop(self):
|
|
LOG_DEBUG(f"AV.CA stop {self._video_thread}")
|
|
self._running = False
|
|
self.stop_audio_thread()
|
|
self.stop_video_thread()
|
|
|
|
def __contains__(self, friend_number):
|
|
return friend_number in self._calls
|
|
|
|
# Calls
|
|
|
|
def __call__(self, friend_number, audio, video):
|
|
"""Call friend with specified number"""
|
|
if friend_number in self._calls:
|
|
LOG.warn(f"__call__ already has {friend_number}")
|
|
return
|
|
if self._audio_krate_tox_audio not in ts.lToxSampleratesK:
|
|
LOG.warn(f"__call__ {self._audio_krate_tox_audio} not in {ts.lToxSampleratesK}")
|
|
|
|
try:
|
|
self._toxav.call(friend_number,
|
|
self._audio_krate_tox_audio if audio else 0,
|
|
self._audio_krate_tox_video if video else 0)
|
|
except Exception as e:
|
|
LOG.warn(f"_toxav.call already has {friend_number}")
|
|
return
|
|
self._calls[friend_number] = Call(audio, video)
|
|
threading.Timer(TIMER_TIMEOUT,
|
|
lambda: self.finish_not_started_call(friend_number)).start()
|
|
|
|
def accept_call(self, friend_number, audio_enabled, video_enabled):
|
|
# obsolete
|
|
return self.call_accept_call(friend_number, audio_enabled, video_enabled)
|
|
|
|
def call_accept_call(self, friend_number, audio_enabled, video_enabled):
|
|
# called from CM.accept_call in a try:
|
|
LOG.debug(f"call_accept_call from F={friend_number} R={self._running}" +
|
|
f" A={audio_enabled} V={video_enabled}")
|
|
# import pdb; pdb.set_trace() - gets into q Qt exec_ problem
|
|
# ts.trepan_handler()
|
|
|
|
if self._audio_krate_tox_audio not in ts.lToxSampleratesK:
|
|
LOG.warn(f"__call__ {self._audio_krate_tox_audio} not in {ts.lToxSampleratesK}")
|
|
if self._running:
|
|
self._calls[friend_number] = Call(audio_enabled, video_enabled)
|
|
# audio_bit_rate: Audio bit rate in Kb/sec. Set this to 0 to disable audio sending.
|
|
# video_bit_rate: Video bit rate in Kb/sec. Set this to 0 to disable video sending.
|
|
try:
|
|
self._toxav.answer(friend_number,
|
|
self._audio_krate_tox_audio if audio_enabled else 0,
|
|
self._audio_krate_tox_video if video_enabled else 0)
|
|
except Exception as e:
|
|
LOG.error(f"AV accept_call error from {friend_number} {self._running} {e}")
|
|
raise
|
|
if video_enabled:
|
|
# may raise
|
|
self.start_video_thread()
|
|
if audio_enabled:
|
|
LOG.debug(f"calls accept_call calling start_audio_thread F={friend_number}")
|
|
# may raise
|
|
self.start_audio_thread()
|
|
|
|
def finish_call(self, friend_number, by_friend=False):
|
|
LOG.debug(f"finish_call {friend_number}")
|
|
if friend_number in self._calls:
|
|
del self._calls[friend_number]
|
|
try:
|
|
# AttributeError: 'int' object has no attribute 'out_audio'
|
|
if not len(list(filter(lambda c: c.out_audio, self._calls))):
|
|
self.stop_audio_thread()
|
|
if not len(list(filter(lambda c: c.out_video, self._calls))):
|
|
self.stop_video_thread()
|
|
except Exception as e:
|
|
LOG.error(f"finish_call FixMe: {e}")
|
|
# dunno
|
|
self.stop_audio_thread()
|
|
self.stop_video_thread()
|
|
if not by_friend:
|
|
LOG.debug(f"finish_call before call_control {friend_number}")
|
|
self._toxav.call_control(friend_number, TOXAV_CALL_CONTROL['CANCEL'])
|
|
LOG.debug(f"finish_call after call_control {friend_number}")
|
|
|
|
def finish_not_started_call(self, friend_number):
|
|
if friend_number in self:
|
|
call = self._calls[friend_number]
|
|
if not call.is_active:
|
|
self.finish_call(friend_number)
|
|
|
|
def toxav_call_state_cb(self, friend_number, state):
|
|
"""
|
|
New call state
|
|
"""
|
|
LOG.debug(f"toxav_call_state_cb {friend_number}")
|
|
call = self._calls[friend_number]
|
|
call.is_active = True
|
|
|
|
call.in_audio = state | TOXAV_FRIEND_CALL_STATE['SENDING_A'] > 0
|
|
call.in_video = state | TOXAV_FRIEND_CALL_STATE['SENDING_V'] > 0
|
|
|
|
if state | TOXAV_FRIEND_CALL_STATE['ACCEPTING_A'] and call.out_audio:
|
|
self.start_audio_thread()
|
|
|
|
if state | TOXAV_FRIEND_CALL_STATE['ACCEPTING_V'] and call.out_video:
|
|
self.start_video_thread()
|
|
|
|
def is_video_call(self, number):
|
|
return number in self and self._calls[number].in_video
|
|
|
|
# Threads
|
|
|
|
def start_audio_thread(self, bSTREAM_CALLBACK=False):
|
|
"""
|
|
Start audio sending
|
|
from a callback
|
|
"""
|
|
# called from call_accept_call in an try: from CM.accept_call
|
|
global oPYA
|
|
# was iInput = self._settings._args.audio['input']
|
|
iInput = self._settings['audio']['input']
|
|
if self._audio_thread is not None:
|
|
LOG_WARN(f"start_audio_thread device={iInput}")
|
|
return
|
|
LOG_DEBUG(f"start_audio_thread device={iInput}")
|
|
lPaSamplerates = ts.lSdSamplerates(iInput)
|
|
if not(len(lPaSamplerates)):
|
|
e = f"No sample rates for device: audio[input]={iInput}"
|
|
LOG_WARN(f"start_audio_thread {e}")
|
|
#?? dunno - cancel call? - no let the user do it
|
|
# return
|
|
# just guessing here in case that's a false negative
|
|
lPaSamplerates = [round(oPYA.get_device_info_by_index(iInput)['defaultSampleRate'])]
|
|
if lPaSamplerates and self._audio_rate_pa in lPaSamplerates:
|
|
pass
|
|
elif lPaSamplerates:
|
|
LOG_WARN(f"Setting audio_rate to: {lPaSamplerates[0]}")
|
|
self._audio_rate_pa = lPaSamplerates[0]
|
|
elif 'defaultSampleRate' in oPYA.get_device_info_by_index(iInput):
|
|
self._audio_rate_pa = oPYA.get_device_info_by_index(iInput)['defaultSampleRate']
|
|
LOG_WARN(f"setting to defaultSampleRate")
|
|
else:
|
|
LOG_WARN(f"{self._audio_rate_pa} not in {lPaSamplerates}")
|
|
# a float is in here - must it be int?
|
|
if type(self._audio_rate_pa) == float:
|
|
self._audio_rate_pa = round(self._audio_rate_pa)
|
|
try:
|
|
if self._audio_rate_pa not in lPaSamplerates:
|
|
LOG_WARN(f"PAudio sampling rate was {self._audio_rate_pa} changed to {lPaSamplerates[0]}")
|
|
LOG_DEBUG(f"lPaSamplerates={lPaSamplerates}")
|
|
self._audio_rate_pa = lPaSamplerates[0]
|
|
else:
|
|
LOG_DEBUG( f"start_audio_thread framerate: {self._audio_rate_pa}" \
|
|
+f" device: {iInput}"
|
|
+f" supported: {lPaSamplerates}")
|
|
|
|
if bSTREAM_CALLBACK:
|
|
# why would you not call a thread?
|
|
self._audio_stream = oPYA.open(format=pyaudio.paInt16,
|
|
rate=self._audio_rate_pa,
|
|
channels=self._audio_channels,
|
|
input=True,
|
|
input_device_index=iInput,
|
|
frames_per_buffer=self._audio_sample_count_pa * 10,
|
|
stream_callback=self.send_audio_data)
|
|
self._audio_running = True
|
|
self._audio_stream.start_stream()
|
|
while self._audio_stream.is_active():
|
|
sleep(0.1)
|
|
self._audio_stream.stop_stream()
|
|
self._audio_stream.close()
|
|
else:
|
|
LOG_DEBUG( f"start_audio_thread starting thread {self._audio_rate_pa}")
|
|
self._audio_stream = oPYA.open(format=pyaudio.paInt16,
|
|
rate=self._audio_rate_pa,
|
|
channels=self._audio_channels,
|
|
input=True,
|
|
input_device_index=iInput,
|
|
frames_per_buffer=self._audio_sample_count_pa * 10)
|
|
self._audio_running = True
|
|
self._audio_thread = AudioThread(self,
|
|
name='_audio_thread')
|
|
self._audio_thread.start()
|
|
LOG_DEBUG( f"start_audio_thread started thread name='_audio_thread'")
|
|
|
|
except Exception as e:
|
|
LOG_ERROR(f"Starting self._audio.open {e}")
|
|
LOG_DEBUG(repr(dict(format=pyaudio.paInt16,
|
|
rate=self._audio_rate_pa,
|
|
channels=self._audio_channels,
|
|
input=True,
|
|
input_device_index=iInput,
|
|
frames_per_buffer=self._audio_sample_count_pa * 10)))
|
|
# catcher in place in calls_manager? yes accept_call
|
|
# calls_manager._call.toxav_call_state_cb(friend_number, mask)
|
|
invoke_in_main_thread(util_ui.message_box,
|
|
str(e),
|
|
util_ui.tr("Starting self._audio.open"))
|
|
return
|
|
else:
|
|
LOG_DEBUG(f"start_audio_thread {self._audio_stream}")
|
|
|
|
def stop_audio_thread(self):
|
|
LOG_DEBUG(f"stop_audio_thread {self._audio_stream}")
|
|
|
|
if self._audio_thread is None:
|
|
return
|
|
self._audio_running = False
|
|
self._audio_thread._stop_thread = True
|
|
|
|
self._audio_thread = None
|
|
self._audio_stream = None
|
|
self._audio = None
|
|
|
|
if self._out_stream is not None:
|
|
self._out_stream.stop_stream()
|
|
self._out_stream.close()
|
|
self._out_stream = None
|
|
|
|
def start_video_thread(self):
|
|
if self._video_thread is not None:
|
|
return
|
|
s = self._settings
|
|
if 'video' not in s:
|
|
LOG.warn("AV.__init__ 'video' not in s" )
|
|
LOG.debug(f"start_video_thread {s}" )
|
|
raise RuntimeError("start_video_thread not 'video' in s)" )
|
|
elif 'device' not in s['video']:
|
|
LOG.error("start_video_thread not 'device' in s['video']" )
|
|
LOG.debug(f"start_video_thread {s['video']}" )
|
|
raise RuntimeError("start_video_thread not 'device' ins s['video']" )
|
|
self._video_width = s['video']['width']
|
|
self._video_height = s['video']['height']
|
|
|
|
# dunno
|
|
if True or s['video']['device'] == -1:
|
|
self._video = screen_sharing.DesktopGrabber(s['video']['x'],
|
|
s['video']['y'],
|
|
s['video']['width'],
|
|
s['video']['height'])
|
|
else:
|
|
with ts.ignoreStdout(): import cv2
|
|
if s['video']['device'] == 0:
|
|
# webcam
|
|
self._video = cv2.VideoCapture(s['video']['device'], cv2.DSHOW)
|
|
else:
|
|
self._video = cv2.VideoCapture(s['video']['device'])
|
|
self._video.set(cv2.CAP_PROP_FPS, iFPS)
|
|
self._video.set(cv2.CAP_PROP_FRAME_WIDTH, self._video_width)
|
|
self._video.set(cv2.CAP_PROP_FRAME_HEIGHT, self._video_height)
|
|
# self._video.set(cv2.CAP_PROP_FOURCC, cv2.VideoWriter_fourcc('M', 'J', 'P', 'G'))
|
|
if self._video is None:
|
|
LOG.error("start_video_thread " \
|
|
+f" device: {s['video']['device']}" \
|
|
+f" supported: {s['video']['width']} {s['video']['height']}")
|
|
return
|
|
LOG.info("start_video_thread " \
|
|
+f" device: {s['video']['device']}" \
|
|
+f" supported: {s['video']['width']} {s['video']['height']}")
|
|
|
|
self._video_running = True
|
|
self._video_thread = VideoThread(self,
|
|
name='_video_thread')
|
|
self._video_thread.start()
|
|
|
|
def stop_video_thread(self) -> None:
|
|
LOG_DEBUG(f"stop_video_thread {self._video_thread}")
|
|
if self._video_thread is None:
|
|
return
|
|
|
|
self._video_thread._stop_thread = True
|
|
self._video_running = False
|
|
i = 0
|
|
while i < ts.iTHREAD_JOINS:
|
|
self._video_thread.join(ts.iTHREAD_TIMEOUT)
|
|
try:
|
|
if not self._video_thread.is_alive(): break
|
|
except:
|
|
break
|
|
i = i + 1
|
|
else:
|
|
LOG.warn("self._video_thread.is_alive BLOCKED")
|
|
self._video_thread = None
|
|
self._video = None
|
|
# Incoming chunks
|
|
|
|
def audio_chunk(self, samples, channels_count, rate) -> None:
|
|
"""
|
|
Incoming chunk
|
|
"""
|
|
# from callback
|
|
if self._out_stream is None:
|
|
# was iOutput = self._settings._args.audio['output']
|
|
iOutput = self._settings['audio']['output']
|
|
if self.lPaSampleratesO and rate in self.lPaSampleratesO:
|
|
LOG_DEBUG(f"Using rate {rate} in self.lPaSampleratesO")
|
|
elif self.lPaSampleratesO:
|
|
LOG_WARN(f"{rate} not in {self.lPaSampleratesO}")
|
|
LOG_WARN(f"Setting audio_rate to: {self.lPaSampleratesO[0]}")
|
|
rate = self.lPaSampleratesO[0]
|
|
elif 'defaultSampleRate' in oPYA.get_device_info_by_index(iOutput):
|
|
rate = round(oPYA.get_device_info_by_index(iOutput)['defaultSampleRate'])
|
|
LOG_WARN(f"Setting rate to {rate} empty self.lPaSampleratesO")
|
|
else:
|
|
LOG_WARN(f"Using rate {rate} empty self.lPaSampleratesO")
|
|
if type(rate) == float:
|
|
rate = round(rate)
|
|
# test output device?
|
|
# [Errno -9985] Device unavailable
|
|
try:
|
|
with ts.ignoreStderr():
|
|
self._out_stream = oPYA.open(format=pyaudio.paInt16,
|
|
channels=channels_count,
|
|
rate=rate,
|
|
output_device_index=iOutput,
|
|
output=True)
|
|
except Exception as e:
|
|
LOG_ERROR(f"Error playing audio_chunk creating self._out_stream output_device_index={iOutput} {e}")
|
|
invoke_in_main_thread(util_ui.message_box,
|
|
str(e),
|
|
util_ui.tr("Error Chunking audio"))
|
|
# dunno
|
|
self.stop()
|
|
return
|
|
|
|
iOutput = self._settings['audio']['output']
|
|
#trace LOG_DEBUG(f"audio_chunk output_device_index={iOutput} rate={rate} channels={channels_count}")
|
|
try:
|
|
self._out_stream.write(samples)
|
|
except Exception as e:
|
|
# OSError: [Errno -9999] Unanticipated host error
|
|
LOG_WARN(f"audio_chunk output_device_index={iOutput} {e}")
|
|
|
|
# AV sending
|
|
|
|
def send_audio_data(self, data, count, *largs, **kwargs) -> None:
|
|
# callback
|
|
pcm = data
|
|
# :param sampling_rate: Audio sampling rate used in this frame.
|
|
try:
|
|
if self._toxav is None:
|
|
LOG_ERROR("_toxav not initialized")
|
|
return
|
|
if self._audio_rate_tox not in ts.lToxSamplerates:
|
|
LOG_WARN(f"ToxAudio sampling rate was {self._audio_rate_tox} changed to {ts.lToxSamplerates[0]}")
|
|
self._audio_rate_tox = ts.lToxSamplerates[0]
|
|
|
|
for friend_num in self._calls:
|
|
if self._calls[friend_num].out_audio:
|
|
# app.av.calls ERROR Error send_audio audio_send_frame: This client is currently not in a call with the friend.
|
|
self._toxav.audio_send_frame(friend_num,
|
|
pcm,
|
|
count,
|
|
self._audio_channels,
|
|
self._audio_rate_tox)
|
|
|
|
except Exception as e:
|
|
LOG.error(f"Error send_audio_data audio_send_frame: {e}")
|
|
LOG.debug(f"send_audio_data self._audio_rate_tox={self._audio_rate_tox} self._audio_channels={self._audio_channels}")
|
|
self.stop_audio_thread()
|
|
invoke_in_main_thread(util_ui.message_box,
|
|
str(e),
|
|
util_ui.tr("Error send_audio_data audio_send_frame"))
|
|
#? stop ? endcall?
|
|
|
|
def send_audio(self) -> None:
|
|
"""
|
|
This method sends audio to friends
|
|
"""
|
|
i=0
|
|
count = self._audio_sample_count_tox
|
|
LOG_DEBUG(f"send_audio stream={self._audio_stream}")
|
|
while self._audio_running:
|
|
try:
|
|
pcm = self._audio_stream.read(count, exception_on_overflow=False)
|
|
if not pcm:
|
|
sleep(0.1)
|
|
else:
|
|
self.send_audio_data(pcm, count)
|
|
except:
|
|
LOG_DEBUG(f"error send_audio {i}")
|
|
else:
|
|
LOG_TRACE(f"send_audio {i}")
|
|
i += 1
|
|
sleep(0.01)
|
|
|
|
def send_video(self) -> None:
|
|
"""
|
|
This method sends video to friends
|
|
"""
|
|
# LOG_DEBUG(f"send_video thread={threading.current_thread().name}"
|
|
# +f" self._video_running={self._video_running}"
|
|
# +f" device: {self._settings['video']['device']}" )
|
|
while self._video_running:
|
|
try:
|
|
result, frame = self._video.read()
|
|
if not result:
|
|
LOG_WARN(f"send_video video_send_frame _video.read result={result}")
|
|
break
|
|
if frame is None:
|
|
LOG_WARN(f"send_video video_send_frame _video.read result={result} frame={frame}")
|
|
continue
|
|
else:
|
|
LOG_TRACE(f"send_video video_send_frame _video.read result={result}")
|
|
height, width, channels = frame.shape
|
|
friends = []
|
|
for friend_num in self._calls:
|
|
if self._calls[friend_num].out_video:
|
|
friends.append(friend_num)
|
|
if len(friends) == 0:
|
|
LOG_WARN(f"send_video video_send_frame no friends")
|
|
else:
|
|
LOG_TRACE(f"send_video video_send_frame {friends}")
|
|
friend_num = friends[0]
|
|
try:
|
|
y, u, v = self.convert_bgr_to_yuv(frame)
|
|
self._toxav.video_send_frame(friend_num, width, height, y, u, v)
|
|
except Exception as e:
|
|
LOG_WARN(f"send_video video_send_frame ERROR {e}")
|
|
pass
|
|
|
|
except Exception as e:
|
|
LOG_ERROR(f"send_video video_send_frame {e}")
|
|
pass
|
|
|
|
sleep( 1.0/iFPS)
|
|
|
|
def convert_bgr_to_yuv(self, frame) -> tuple:
|
|
"""
|
|
:param frame: input bgr frame
|
|
:return y, u, v: y, u, v values of frame
|
|
|
|
How this function works:
|
|
OpenCV creates YUV420 frame from BGR
|
|
This frame has following structure and size:
|
|
width, height - dim of input frame
|
|
width, height * 1.5 - dim of output frame
|
|
|
|
width
|
|
-------------------------
|
|
| |
|
|
| Y | height
|
|
| |
|
|
-------------------------
|
|
| | |
|
|
| U even | U odd | height // 4
|
|
| | |
|
|
-------------------------
|
|
| | |
|
|
| V even | V odd | height // 4
|
|
| | |
|
|
-------------------------
|
|
|
|
width // 2 width // 2
|
|
|
|
Y, U, V can be extracted using slices and joined in one list using itertools.chain.from_iterable()
|
|
Function returns bytes(y), bytes(u), bytes(v), because it is required for ctypes
|
|
"""
|
|
with ts.ignoreStdout():
|
|
import cv2
|
|
frame = cv2.cvtColor(frame, cv2.COLOR_BGR2YUV_I420)
|
|
|
|
y = frame[:self._video_height, :]
|
|
y = list(itertools.chain.from_iterable(y))
|
|
|
|
import numpy as np
|
|
# was np.int
|
|
u = np.zeros((self._video_height // 2, self._video_width // 2), dtype=np.int32)
|
|
u[::2, :] = frame[self._video_height:self._video_height * 5 // 4, :self._video_width // 2]
|
|
u[1::2, :] = frame[self._video_height:self._video_height * 5 // 4, self._video_width // 2:]
|
|
u = list(itertools.chain.from_iterable(u))
|
|
v = np.zeros((self._video_height // 2, self._video_width // 2), dtype=np.int32)
|
|
v[::2, :] = frame[self._video_height * 5 // 4:, :self._video_width // 2]
|
|
v[1::2, :] = frame[self._video_height * 5 // 4:, self._video_width // 2:]
|
|
v = list(itertools.chain.from_iterable(v))
|
|
|
|
return bytes(y), bytes(u), bytes(v)
|