capture函数,设置bayer=True
啥是mmal.xxxx? 啥是camera_port?
看起来是个设置标志,产生buffer,获取针对对应格式的c数据结构
camera_port与self._camera.outputs有关
啥是mmalcamera
总之,找到Outputs有3个,disable()一下,然后配置相机参数,commit提交一下;
enable output with 回调函数,然后设置output.param【某某】为True产生buffers然后才能得到最终数据。
看完了例子,还有个encoder.start()
最后就追溯到这了。。。。
看来是多重继承
继续
又是多重
class PiRawMixin(PiEncoder):
"""
Mixin class for "raw" (unencoded) output.
This mixin class overrides the initializer of :class:`PiEncoder`, along
with :meth:`_create_resizer` and :meth:`_create_encoder` to configure the
pipeline for unencoded output. Specifically, it disables the construction
of an encoder, and sets the output port to the input port passed to the
initializer, unless resizing is required (either for actual resizing, or
for format conversion) in which case the resizer's output is used.
"""
RAW_ENCODINGS = {
# name mmal-encoding bytes-per-pixel
'yuv': (mmal.MMAL_ENCODING_I420, 1.5),
'rgb': (mmal.MMAL_ENCODING_RGB24, 3),
'rgba': (mmal.MMAL_ENCODING_RGBA, 4),
'bgr': (mmal.MMAL_ENCODING_BGR24, 3),
'bgra': (mmal.MMAL_ENCODING_BGRA, 4),
}
def __init__(
self, parent, camera_port, input_port, format, resize, **options):
encoding, bpp = self.RAW_ENCODINGS[format]
# Workaround: on older firmwares, non-YUV encodings aren't supported on
# the still port. If a non-YUV format is requested without resizing,
# test whether we can commit the requested format on the input port and
# if this fails, set resize to force resizer usage
if resize is None and encoding != mmal.MMAL_ENCODING_I420:
input_port.format = encoding
try:
input_port.commit()
except PiCameraMMALError as e:
if e.status != mmal.MMAL_EINVAL:
raise
resize = input_port.framesize
warnings.warn(
PiCameraResizerEncoding(
"using a resizer to perform non-YUV encoding; "
"upgrading your firmware with sudo rpi-update "
"may improve performance"))
# Workaround: If a non-alpha format is requested with the resizer, use
# the alpha-inclusive format and set a flag to get the callback to
# strip the alpha bytes
self._strip_alpha = False
if resize:
width, height = resize
try:
format = {
'rgb': 'rgba',
'bgr': 'bgra',
}[format]
self._strip_alpha = True
warnings.warn(
PiCameraAlphaStripping(
"using alpha-stripping to convert to non-alpha "
"format; you may find the equivalent alpha format "
"faster"))
except KeyError:
pass
else:
width, height = input_port.framesize
# Workaround (#83): when the resizer is used the width must be aligned
# (both the frame and crop values) to avoid an error when the output
# port format is set (height is aligned too, simply for consistency
# with old picamera versions). Warn the user as they're not going to
# get the resolution they expect
if not resize and format != 'yuv' and input_port.name.startswith('vc.ril.video_splitter'):
# Workaround: Expected frame size is rounded to 16x16 when splitter
# port with no resizer is used and format is not YUV
fwidth = bcm_host.VCOS_ALIGN_UP(width, 16)
else:
fwidth = bcm_host.VCOS_ALIGN_UP(width, 32)
fheight = bcm_host.VCOS_ALIGN_UP(height, 16)
if fwidth != width or fheight != height:
warnings.warn(
PiCameraResolutionRounded(
"frame size rounded up from %dx%d to %dx%d" % (
width, height, fwidth, fheight)))
if resize:
resize = (fwidth, fheight)
# Workaround: Calculate the expected frame size, to be used by the
# callback to decide when a frame ends. This is to work around a
# firmware bug that causes the raw image to be returned twice when the
# maximum camera resolution is requested
self._frame_size = int(fwidth * fheight * bpp)
super(PiRawMixin, self).__init__(
parent, camera_port, input_port, format, resize, **options)
def _create_encoder(self, format):
"""
Overridden to skip creating an encoder. Instead, this class simply uses
the resizer's port as the output port (if a resizer has been
configured) or the specified input port otherwise.
"""
if self.resizer:
self.output_port = self.resizer.outputs[0]
else:
self.output_port = self.input_port
try:
self.output_port.format = self.RAW_ENCODINGS[format][0]
except KeyError:
raise PiCameraValueError('unknown format %s' % format)
self.output_port.commit()
def _callback_write(self, buf, key=PiVideoFrameType.frame):
"""
_callback_write(buf, key=PiVideoFrameType.frame)
Overridden to strip alpha bytes when required.
"""
if self._strip_alpha:
return super(PiRawMixin, self)._callback_write(MMALBufferAlphaStrip(buf._buf), key)
else:
return super(PiRawMixin, self)._callback_write(buf, key)
class PiVideoFrame(namedtuple('PiVideoFrame', (
'index', # the frame number, where the first frame is 0
'frame_type', # a constant indicating the frame type (see PiVideoFrameType)
'frame_size', # the size (in bytes) of the frame's data
'video_size', # the size (in bytes) of the video so far
'split_size', # the size (in bytes) of the video since the last split
'timestamp', # the presentation timestamp (PTS) of the frame
'complete', # whether the frame is complete or not
))):
"""
This class is a :func:`~collections.namedtuple` derivative used to store
information about a video frame. It is recommended that you access the
information stored by this class by attribute name rather than position
(for example: ``frame.index`` rather than ``frame[0]``).
.. attribute:: index
Returns the zero-based number of the frame. This is a monotonic counter
that is simply incremented every time the camera starts outputting a
new frame. As a consequence, this attribute cannot be used to detect
dropped frames. Nor does it necessarily represent actual frames; it
will be incremented for SPS headers and motion data buffers too.
.. attribute:: frame_type
Returns a constant indicating the kind of data that the frame contains
(see :class:`PiVideoFrameType`). Please note that certain frame types
contain no image data at all.
.. attribute:: frame_size
Returns the size in bytes of the current frame. If a frame is written
in multiple chunks, this value will increment while :attr:`index`
remains static. Query :attr:`complete` to determine whether the frame
has been completely output yet.
.. attribute:: video_size
Returns the size in bytes of the entire video up to this frame. Note
that this is unlikely to match the size of the actual file/stream
written so far. This is because a stream may utilize buffering which
will cause the actual amount written (e.g. to disk) to lag behind the
value reported by this attribute.
.. attribute:: split_size
Returns the size in bytes of the video recorded since the last call to
either :meth:`~PiCamera.start_recording` or
:meth:`~PiCamera.split_recording`. For the reasons explained above,
this may differ from the size of the actual file/stream written so far.
.. attribute:: timestamp
Returns the presentation timestamp (PTS) of the frame. This represents
the point in time that the Pi received the first line of the frame from
the camera.
The timestamp is measured in microseconds (millionths of a second).
When the camera's clock mode is ``'reset'`` (the default), the
timestamp is relative to the start of the video recording. When the
camera's :attr:`~PiCamera.clock_mode` is ``'raw'``, it is relative to
the last system reboot. See :attr:`~PiCamera.timestamp` for more
information.
.. warning::
Currently, the camera occasionally returns "time unknown" values in
this field which picamera represents as ``None``. If you are
querying this property you will need to check the value is not
``None`` before using it. This happens for SPS header "frames",
for example.
.. attribute:: complete
Returns a bool indicating whether the current frame is complete or not.
If the frame is complete then :attr:`frame_size` will not increment
any further, and will reset for the next frame.
.. versionchanged:: 1.5
Deprecated :attr:`header` and :attr:`keyframe` attributes and added the
new :attr:`frame_type` attribute instead.
.. versionchanged:: 1.9
Added the :attr:`complete` attribute.
"""
__slots__ = () # workaround python issue #24931
@property
def position(self):
"""
Returns the zero-based position of the frame in the stream containing
it.
"""
return self.split_size - self.frame_size
@property
def keyframe(self):
"""
Returns a bool indicating whether the current frame is a keyframe (an
intra-frame, or I-frame in MPEG parlance).
.. deprecated:: 1.5
Please compare :attr:`frame_type` to
:attr:`PiVideoFrameType.key_frame` instead.
"""
warnings.warn(
PiCameraDeprecated(
'PiVideoFrame.keyframe is deprecated; please check '
'PiVideoFrame.frame_type for equality with '
'PiVideoFrameType.key_frame instead'))
return self.frame_type == PiVideoFrameType.key_frame
@property
def header(self):
"""
Contains a bool indicating whether the current frame is actually an
SPS/PPS header. Typically it is best to split an H.264 stream so that
it starts with an SPS/PPS header.
.. deprecated:: 1.5
Please compare :attr:`frame_type` to
:attr:`PiVideoFrameType.sps_header` instead.
"""
warnings.warn(
PiCameraDeprecated(
'PiVideoFrame.header is deprecated; please check '
'PiVideoFrame.frame_type for equality with '
'PiVideoFrameType.sps_header instead'))
return self.frame_type == PiVideoFrameType.sps_header
这个.frame就是0,用在encoder里
self.outputs就是个字典,以下是填入对象
mo.open_stream:
就是读写咯
乖乖,就是说encoder有outputs字典,可以存储多种output.
找了半天还是没找到关于buffer_count的
picamera2
库有相关设置调整,picamera貌似内置好的
buffer_count在摄像头捕获和处理图像时起着重要作用。它表示用于存储图像数据的缓冲区数量。这些缓冲区是临时的存储空间,用于在图像从摄像头传感器读取并传输到应用程序或处理器进行进一步处理时暂存图像数据。
具体来说,buffer_count的作用包括:
图像连续性:当摄像头连续捕获图像时,每个缓冲区都会存储一帧图像的数据。通过设置足够的缓冲区数量,可以确保摄像头能够连续地捕获图像,而不会因为处理速度不够快而丢失帧。这样,就能够保证图像流的连续性。
性能优化:适当的缓冲区数量可以平衡摄像头捕获速度和应用程序处理速度之间的差异。如果缓冲区数量太少,摄像头可能会因为等待应用程序处理完前一帧而减慢捕获速度。而如果缓冲区数量过多,可能会占用过多的内存资源,导致效率下降。因此,根据系统的性能和需求,设置合适的buffer_count可以提高图像处理的效率和性能。
资源管理:缓冲区数量也涉及到内存资源的管理。每个缓冲区都需要占用一定的内存空间。通过控制buffer_count,可以管理摄像头系统所使用的内存量,以确保资源的合理利用。
综上所述,buffer_count在摄像头捕获和处理图像时起到了关键的作用,它影响着图像连续性、性能优化和资源管理。通过合理设置buffer_count,可以确保摄像头系统的高效运行和图像质量。
目前未找到硬编码位置,,,,,,,,